Feb 13 19:08:54.878870 kernel: Booting Linux on physical CPU 0x0000000000 [0x413fd0c1] Feb 13 19:08:54.878898 kernel: Linux version 6.6.74-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 14.2.1_p20241221 p7) 14.2.1 20241221, GNU ld (Gentoo 2.43 p3) 2.43.1) #1 SMP PREEMPT Thu Feb 13 17:39:57 -00 2025 Feb 13 19:08:54.878911 kernel: KASLR enabled Feb 13 19:08:54.878918 kernel: efi: EFI v2.7 by Ubuntu distribution of EDK II Feb 13 19:08:54.878926 kernel: efi: SMBIOS 3.0=0x139ed0000 MEMATTR=0x1390bb018 ACPI 2.0=0x136760018 RNG=0x13676e918 MEMRESERVE=0x136b41218 Feb 13 19:08:54.878933 kernel: random: crng init done Feb 13 19:08:54.878942 kernel: secureboot: Secure boot disabled Feb 13 19:08:54.878949 kernel: ACPI: Early table checksum verification disabled Feb 13 19:08:54.878957 kernel: ACPI: RSDP 0x0000000136760018 000024 (v02 BOCHS ) Feb 13 19:08:54.878966 kernel: ACPI: XSDT 0x000000013676FE98 00006C (v01 BOCHS BXPC 00000001 01000013) Feb 13 19:08:54.878974 kernel: ACPI: FACP 0x000000013676FA98 000114 (v06 BOCHS BXPC 00000001 BXPC 00000001) Feb 13 19:08:54.878982 kernel: ACPI: DSDT 0x0000000136767518 001468 (v02 BOCHS BXPC 00000001 BXPC 00000001) Feb 13 19:08:54.878989 kernel: ACPI: APIC 0x000000013676FC18 000108 (v04 BOCHS BXPC 00000001 BXPC 00000001) Feb 13 19:08:54.878996 kernel: ACPI: PPTT 0x000000013676FD98 000060 (v02 BOCHS BXPC 00000001 BXPC 00000001) Feb 13 19:08:54.879005 kernel: ACPI: GTDT 0x000000013676D898 000060 (v02 BOCHS BXPC 00000001 BXPC 00000001) Feb 13 19:08:54.879015 kernel: ACPI: MCFG 0x000000013676FF98 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Feb 13 19:08:54.879023 kernel: ACPI: SPCR 0x000000013676E818 000050 (v02 BOCHS BXPC 00000001 BXPC 00000001) Feb 13 19:08:54.879031 kernel: ACPI: DBG2 0x000000013676E898 000057 (v00 BOCHS BXPC 00000001 BXPC 00000001) Feb 13 19:08:54.879038 kernel: ACPI: IORT 0x000000013676E418 000080 (v03 BOCHS BXPC 00000001 BXPC 00000001) Feb 13 19:08:54.879046 kernel: ACPI: BGRT 0x000000013676E798 000038 (v01 INTEL EDK2 00000002 01000013) Feb 13 19:08:54.879076 kernel: ACPI: SPCR: console: pl011,mmio32,0x9000000,9600 Feb 13 19:08:54.879085 kernel: NUMA: Failed to initialise from firmware Feb 13 19:08:54.879093 kernel: NUMA: Faking a node at [mem 0x0000000040000000-0x0000000139ffffff] Feb 13 19:08:54.879101 kernel: NUMA: NODE_DATA [mem 0x13966f800-0x139674fff] Feb 13 19:08:54.879109 kernel: Zone ranges: Feb 13 19:08:54.879118 kernel: DMA [mem 0x0000000040000000-0x00000000ffffffff] Feb 13 19:08:54.879126 kernel: DMA32 empty Feb 13 19:08:54.879134 kernel: Normal [mem 0x0000000100000000-0x0000000139ffffff] Feb 13 19:08:54.879142 kernel: Movable zone start for each node Feb 13 19:08:54.879149 kernel: Early memory node ranges Feb 13 19:08:54.879157 kernel: node 0: [mem 0x0000000040000000-0x000000013666ffff] Feb 13 19:08:54.879165 kernel: node 0: [mem 0x0000000136670000-0x000000013667ffff] Feb 13 19:08:54.879172 kernel: node 0: [mem 0x0000000136680000-0x000000013676ffff] Feb 13 19:08:54.879180 kernel: node 0: [mem 0x0000000136770000-0x0000000136b3ffff] Feb 13 19:08:54.879188 kernel: node 0: [mem 0x0000000136b40000-0x0000000139e1ffff] Feb 13 19:08:54.879195 kernel: node 0: [mem 0x0000000139e20000-0x0000000139eaffff] Feb 13 19:08:54.879203 kernel: node 0: [mem 0x0000000139eb0000-0x0000000139ebffff] Feb 13 19:08:54.879212 kernel: node 0: [mem 0x0000000139ec0000-0x0000000139fdffff] Feb 13 19:08:54.879220 kernel: node 0: [mem 0x0000000139fe0000-0x0000000139ffffff] Feb 13 19:08:54.879228 kernel: Initmem setup node 0 [mem 0x0000000040000000-0x0000000139ffffff] Feb 13 19:08:54.879239 kernel: On node 0, zone Normal: 24576 pages in unavailable ranges Feb 13 19:08:54.879247 kernel: psci: probing for conduit method from ACPI. Feb 13 19:08:54.879256 kernel: psci: PSCIv1.1 detected in firmware. Feb 13 19:08:54.879266 kernel: psci: Using standard PSCI v0.2 function IDs Feb 13 19:08:54.879274 kernel: psci: Trusted OS migration not required Feb 13 19:08:54.879282 kernel: psci: SMC Calling Convention v1.1 Feb 13 19:08:54.879290 kernel: smccc: KVM: hypervisor services detected (0x00000000 0x00000000 0x00000000 0x00000003) Feb 13 19:08:54.879299 kernel: percpu: Embedded 31 pages/cpu s86696 r8192 d32088 u126976 Feb 13 19:08:54.879307 kernel: pcpu-alloc: s86696 r8192 d32088 u126976 alloc=31*4096 Feb 13 19:08:54.879321 kernel: pcpu-alloc: [0] 0 [0] 1 Feb 13 19:08:54.879330 kernel: Detected PIPT I-cache on CPU0 Feb 13 19:08:54.879338 kernel: CPU features: detected: GIC system register CPU interface Feb 13 19:08:54.879346 kernel: CPU features: detected: Hardware dirty bit management Feb 13 19:08:54.879357 kernel: CPU features: detected: Spectre-v4 Feb 13 19:08:54.879366 kernel: CPU features: detected: Spectre-BHB Feb 13 19:08:54.879374 kernel: CPU features: kernel page table isolation forced ON by KASLR Feb 13 19:08:54.879382 kernel: CPU features: detected: Kernel page table isolation (KPTI) Feb 13 19:08:54.879390 kernel: CPU features: detected: ARM erratum 1418040 Feb 13 19:08:54.879398 kernel: CPU features: detected: SSBS not fully self-synchronizing Feb 13 19:08:54.879407 kernel: alternatives: applying boot alternatives Feb 13 19:08:54.879417 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyAMA0,115200n8 flatcar.first_boot=detected acpi=force flatcar.oem.id=hetzner verity.usrhash=f06bad36699a22ae88c1968cd72b62b3503d97da521712e50a4b744320b1ba33 Feb 13 19:08:54.879425 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Feb 13 19:08:54.879433 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Feb 13 19:08:54.879442 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Feb 13 19:08:54.879451 kernel: Fallback order for Node 0: 0 Feb 13 19:08:54.879459 kernel: Built 1 zonelists, mobility grouping on. Total pages: 1008000 Feb 13 19:08:54.879467 kernel: Policy zone: Normal Feb 13 19:08:54.879476 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Feb 13 19:08:54.879484 kernel: software IO TLB: area num 2. Feb 13 19:08:54.879492 kernel: software IO TLB: mapped [mem 0x00000000fbfff000-0x00000000fffff000] (64MB) Feb 13 19:08:54.879501 kernel: Memory: 3883896K/4096000K available (10304K kernel code, 2186K rwdata, 8092K rodata, 38336K init, 897K bss, 212104K reserved, 0K cma-reserved) Feb 13 19:08:54.879509 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Feb 13 19:08:54.879517 kernel: rcu: Preemptible hierarchical RCU implementation. Feb 13 19:08:54.879526 kernel: rcu: RCU event tracing is enabled. Feb 13 19:08:54.879535 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Feb 13 19:08:54.879543 kernel: Trampoline variant of Tasks RCU enabled. Feb 13 19:08:54.879553 kernel: Tracing variant of Tasks RCU enabled. Feb 13 19:08:54.879562 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Feb 13 19:08:54.879570 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Feb 13 19:08:54.879578 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Feb 13 19:08:54.879586 kernel: GICv3: 256 SPIs implemented Feb 13 19:08:54.879594 kernel: GICv3: 0 Extended SPIs implemented Feb 13 19:08:54.879602 kernel: Root IRQ handler: gic_handle_irq Feb 13 19:08:54.879610 kernel: GICv3: GICv3 features: 16 PPIs, DirectLPI Feb 13 19:08:54.879619 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000080a0000 Feb 13 19:08:54.879627 kernel: ITS [mem 0x08080000-0x0809ffff] Feb 13 19:08:54.879635 kernel: ITS@0x0000000008080000: allocated 8192 Devices @1000c0000 (indirect, esz 8, psz 64K, shr 1) Feb 13 19:08:54.879645 kernel: ITS@0x0000000008080000: allocated 8192 Interrupt Collections @1000d0000 (flat, esz 8, psz 64K, shr 1) Feb 13 19:08:54.879654 kernel: GICv3: using LPI property table @0x00000001000e0000 Feb 13 19:08:54.879662 kernel: GICv3: CPU0: using allocated LPI pending table @0x00000001000f0000 Feb 13 19:08:54.879670 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Feb 13 19:08:54.879678 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Feb 13 19:08:54.879687 kernel: arch_timer: cp15 timer(s) running at 25.00MHz (virt). Feb 13 19:08:54.879695 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns Feb 13 19:08:54.879704 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns Feb 13 19:08:54.879712 kernel: Console: colour dummy device 80x25 Feb 13 19:08:54.879720 kernel: ACPI: Core revision 20230628 Feb 13 19:08:54.879729 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000) Feb 13 19:08:54.879739 kernel: pid_max: default: 32768 minimum: 301 Feb 13 19:08:54.879748 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Feb 13 19:08:54.879756 kernel: landlock: Up and running. Feb 13 19:08:54.879765 kernel: SELinux: Initializing. Feb 13 19:08:54.879773 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Feb 13 19:08:54.879781 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Feb 13 19:08:54.879790 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Feb 13 19:08:54.879799 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Feb 13 19:08:54.879807 kernel: rcu: Hierarchical SRCU implementation. Feb 13 19:08:54.879818 kernel: rcu: Max phase no-delay instances is 400. Feb 13 19:08:54.879826 kernel: Platform MSI: ITS@0x8080000 domain created Feb 13 19:08:54.879841 kernel: PCI/MSI: ITS@0x8080000 domain created Feb 13 19:08:54.879852 kernel: Remapping and enabling EFI services. Feb 13 19:08:54.879860 kernel: smp: Bringing up secondary CPUs ... Feb 13 19:08:54.879868 kernel: Detected PIPT I-cache on CPU1 Feb 13 19:08:54.879877 kernel: GICv3: CPU1: found redistributor 1 region 0:0x00000000080c0000 Feb 13 19:08:54.879886 kernel: GICv3: CPU1: using allocated LPI pending table @0x0000000100100000 Feb 13 19:08:54.879895 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Feb 13 19:08:54.879906 kernel: CPU1: Booted secondary processor 0x0000000001 [0x413fd0c1] Feb 13 19:08:54.879914 kernel: smp: Brought up 1 node, 2 CPUs Feb 13 19:08:54.879929 kernel: SMP: Total of 2 processors activated. Feb 13 19:08:54.879940 kernel: CPU features: detected: 32-bit EL0 Support Feb 13 19:08:54.879949 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence Feb 13 19:08:54.879958 kernel: CPU features: detected: Common not Private translations Feb 13 19:08:54.879967 kernel: CPU features: detected: CRC32 instructions Feb 13 19:08:54.879975 kernel: CPU features: detected: Enhanced Virtualization Traps Feb 13 19:08:54.879984 kernel: CPU features: detected: RCpc load-acquire (LDAPR) Feb 13 19:08:54.879995 kernel: CPU features: detected: LSE atomic instructions Feb 13 19:08:54.880004 kernel: CPU features: detected: Privileged Access Never Feb 13 19:08:54.880013 kernel: CPU features: detected: RAS Extension Support Feb 13 19:08:54.880022 kernel: CPU features: detected: Speculative Store Bypassing Safe (SSBS) Feb 13 19:08:54.880031 kernel: CPU: All CPU(s) started at EL1 Feb 13 19:08:54.880040 kernel: alternatives: applying system-wide alternatives Feb 13 19:08:54.880049 kernel: devtmpfs: initialized Feb 13 19:08:54.880067 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Feb 13 19:08:54.880078 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Feb 13 19:08:54.880087 kernel: pinctrl core: initialized pinctrl subsystem Feb 13 19:08:54.880096 kernel: SMBIOS 3.0.0 present. Feb 13 19:08:54.880105 kernel: DMI: Hetzner vServer/KVM Virtual Machine, BIOS 20171111 11/11/2017 Feb 13 19:08:54.880114 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Feb 13 19:08:54.880124 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Feb 13 19:08:54.880132 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Feb 13 19:08:54.880142 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Feb 13 19:08:54.880150 kernel: audit: initializing netlink subsys (disabled) Feb 13 19:08:54.880161 kernel: audit: type=2000 audit(0.014:1): state=initialized audit_enabled=0 res=1 Feb 13 19:08:54.880170 kernel: thermal_sys: Registered thermal governor 'step_wise' Feb 13 19:08:54.880179 kernel: cpuidle: using governor menu Feb 13 19:08:54.880188 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Feb 13 19:08:54.880197 kernel: ASID allocator initialised with 32768 entries Feb 13 19:08:54.880206 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Feb 13 19:08:54.880215 kernel: Serial: AMBA PL011 UART driver Feb 13 19:08:54.880224 kernel: Modules: 2G module region forced by RANDOMIZE_MODULE_REGION_FULL Feb 13 19:08:54.880233 kernel: Modules: 0 pages in range for non-PLT usage Feb 13 19:08:54.880244 kernel: Modules: 509280 pages in range for PLT usage Feb 13 19:08:54.880252 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Feb 13 19:08:54.880261 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page Feb 13 19:08:54.880271 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages Feb 13 19:08:54.880280 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page Feb 13 19:08:54.880289 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Feb 13 19:08:54.880297 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page Feb 13 19:08:54.880306 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages Feb 13 19:08:54.880315 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page Feb 13 19:08:54.880326 kernel: ACPI: Added _OSI(Module Device) Feb 13 19:08:54.880335 kernel: ACPI: Added _OSI(Processor Device) Feb 13 19:08:54.880344 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Feb 13 19:08:54.880353 kernel: ACPI: Added _OSI(Processor Aggregator Device) Feb 13 19:08:54.880362 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Feb 13 19:08:54.880371 kernel: ACPI: Interpreter enabled Feb 13 19:08:54.880379 kernel: ACPI: Using GIC for interrupt routing Feb 13 19:08:54.880388 kernel: ACPI: MCFG table detected, 1 entries Feb 13 19:08:54.880397 kernel: ARMH0011:00: ttyAMA0 at MMIO 0x9000000 (irq = 12, base_baud = 0) is a SBSA Feb 13 19:08:54.880408 kernel: printk: console [ttyAMA0] enabled Feb 13 19:08:54.880417 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Feb 13 19:08:54.880590 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Feb 13 19:08:54.880701 kernel: acpi PNP0A08:00: _OSC: platform does not support [LTR] Feb 13 19:08:54.880787 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME AER PCIeCapability] Feb 13 19:08:54.880880 kernel: acpi PNP0A08:00: ECAM area [mem 0x4010000000-0x401fffffff] reserved by PNP0C02:00 Feb 13 19:08:54.880960 kernel: acpi PNP0A08:00: ECAM at [mem 0x4010000000-0x401fffffff] for [bus 00-ff] Feb 13 19:08:54.880975 kernel: ACPI: Remapped I/O 0x000000003eff0000 to [io 0x0000-0xffff window] Feb 13 19:08:54.880985 kernel: PCI host bridge to bus 0000:00 Feb 13 19:08:54.881168 kernel: pci_bus 0000:00: root bus resource [mem 0x10000000-0x3efeffff window] Feb 13 19:08:54.881249 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0xffff window] Feb 13 19:08:54.881323 kernel: pci_bus 0000:00: root bus resource [mem 0x8000000000-0xffffffffff window] Feb 13 19:08:54.881394 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Feb 13 19:08:54.881500 kernel: pci 0000:00:00.0: [1b36:0008] type 00 class 0x060000 Feb 13 19:08:54.881600 kernel: pci 0000:00:01.0: [1af4:1050] type 00 class 0x038000 Feb 13 19:08:54.881685 kernel: pci 0000:00:01.0: reg 0x14: [mem 0x11289000-0x11289fff] Feb 13 19:08:54.881768 kernel: pci 0000:00:01.0: reg 0x20: [mem 0x8000600000-0x8000603fff 64bit pref] Feb 13 19:08:54.881873 kernel: pci 0000:00:02.0: [1b36:000c] type 01 class 0x060400 Feb 13 19:08:54.881961 kernel: pci 0000:00:02.0: reg 0x10: [mem 0x11288000-0x11288fff] Feb 13 19:08:54.882050 kernel: pci 0000:00:02.1: [1b36:000c] type 01 class 0x060400 Feb 13 19:08:54.882153 kernel: pci 0000:00:02.1: reg 0x10: [mem 0x11287000-0x11287fff] Feb 13 19:08:54.882251 kernel: pci 0000:00:02.2: [1b36:000c] type 01 class 0x060400 Feb 13 19:08:54.882340 kernel: pci 0000:00:02.2: reg 0x10: [mem 0x11286000-0x11286fff] Feb 13 19:08:54.882431 kernel: pci 0000:00:02.3: [1b36:000c] type 01 class 0x060400 Feb 13 19:08:54.882513 kernel: pci 0000:00:02.3: reg 0x10: [mem 0x11285000-0x11285fff] Feb 13 19:08:54.882603 kernel: pci 0000:00:02.4: [1b36:000c] type 01 class 0x060400 Feb 13 19:08:54.882686 kernel: pci 0000:00:02.4: reg 0x10: [mem 0x11284000-0x11284fff] Feb 13 19:08:54.882783 kernel: pci 0000:00:02.5: [1b36:000c] type 01 class 0x060400 Feb 13 19:08:54.882878 kernel: pci 0000:00:02.5: reg 0x10: [mem 0x11283000-0x11283fff] Feb 13 19:08:54.882968 kernel: pci 0000:00:02.6: [1b36:000c] type 01 class 0x060400 Feb 13 19:08:54.883049 kernel: pci 0000:00:02.6: reg 0x10: [mem 0x11282000-0x11282fff] Feb 13 19:08:54.883154 kernel: pci 0000:00:02.7: [1b36:000c] type 01 class 0x060400 Feb 13 19:08:54.883242 kernel: pci 0000:00:02.7: reg 0x10: [mem 0x11281000-0x11281fff] Feb 13 19:08:54.883336 kernel: pci 0000:00:03.0: [1b36:000c] type 01 class 0x060400 Feb 13 19:08:54.883419 kernel: pci 0000:00:03.0: reg 0x10: [mem 0x11280000-0x11280fff] Feb 13 19:08:54.883509 kernel: pci 0000:00:04.0: [1b36:0002] type 00 class 0x070002 Feb 13 19:08:54.883592 kernel: pci 0000:00:04.0: reg 0x10: [io 0x0000-0x0007] Feb 13 19:08:54.883704 kernel: pci 0000:01:00.0: [1af4:1041] type 00 class 0x020000 Feb 13 19:08:54.883795 kernel: pci 0000:01:00.0: reg 0x14: [mem 0x11000000-0x11000fff] Feb 13 19:08:54.883900 kernel: pci 0000:01:00.0: reg 0x20: [mem 0x8000000000-0x8000003fff 64bit pref] Feb 13 19:08:54.883989 kernel: pci 0000:01:00.0: reg 0x30: [mem 0xfff80000-0xffffffff pref] Feb 13 19:08:54.884093 kernel: pci 0000:02:00.0: [1b36:000d] type 00 class 0x0c0330 Feb 13 19:08:54.884183 kernel: pci 0000:02:00.0: reg 0x10: [mem 0x10e00000-0x10e03fff 64bit] Feb 13 19:08:54.884274 kernel: pci 0000:03:00.0: [1af4:1043] type 00 class 0x078000 Feb 13 19:08:54.884359 kernel: pci 0000:03:00.0: reg 0x14: [mem 0x10c00000-0x10c00fff] Feb 13 19:08:54.884449 kernel: pci 0000:03:00.0: reg 0x20: [mem 0x8000100000-0x8000103fff 64bit pref] Feb 13 19:08:54.884541 kernel: pci 0000:04:00.0: [1af4:1045] type 00 class 0x00ff00 Feb 13 19:08:54.884634 kernel: pci 0000:04:00.0: reg 0x20: [mem 0x8000200000-0x8000203fff 64bit pref] Feb 13 19:08:54.884732 kernel: pci 0000:05:00.0: [1af4:1044] type 00 class 0x00ff00 Feb 13 19:08:54.884824 kernel: pci 0000:05:00.0: reg 0x14: [mem 0x10800000-0x10800fff] Feb 13 19:08:54.884937 kernel: pci 0000:05:00.0: reg 0x20: [mem 0x8000300000-0x8000303fff 64bit pref] Feb 13 19:08:54.885030 kernel: pci 0000:06:00.0: [1af4:1048] type 00 class 0x010000 Feb 13 19:08:54.885184 kernel: pci 0000:06:00.0: reg 0x14: [mem 0x10600000-0x10600fff] Feb 13 19:08:54.885270 kernel: pci 0000:06:00.0: reg 0x20: [mem 0x8000400000-0x8000403fff 64bit pref] Feb 13 19:08:54.885362 kernel: pci 0000:07:00.0: [1af4:1041] type 00 class 0x020000 Feb 13 19:08:54.885444 kernel: pci 0000:07:00.0: reg 0x14: [mem 0x10400000-0x10400fff] Feb 13 19:08:54.885526 kernel: pci 0000:07:00.0: reg 0x20: [mem 0x8000500000-0x8000503fff 64bit pref] Feb 13 19:08:54.885609 kernel: pci 0000:07:00.0: reg 0x30: [mem 0xfff80000-0xffffffff pref] Feb 13 19:08:54.885697 kernel: pci 0000:00:02.0: bridge window [io 0x1000-0x0fff] to [bus 01] add_size 1000 Feb 13 19:08:54.885779 kernel: pci 0000:00:02.0: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 01] add_size 100000 add_align 100000 Feb 13 19:08:54.885873 kernel: pci 0000:00:02.0: bridge window [mem 0x00100000-0x001fffff] to [bus 01] add_size 100000 add_align 100000 Feb 13 19:08:54.885959 kernel: pci 0000:00:02.1: bridge window [io 0x1000-0x0fff] to [bus 02] add_size 1000 Feb 13 19:08:54.886041 kernel: pci 0000:00:02.1: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 02] add_size 200000 add_align 100000 Feb 13 19:08:54.886137 kernel: pci 0000:00:02.1: bridge window [mem 0x00100000-0x001fffff] to [bus 02] add_size 100000 add_align 100000 Feb 13 19:08:54.886223 kernel: pci 0000:00:02.2: bridge window [io 0x1000-0x0fff] to [bus 03] add_size 1000 Feb 13 19:08:54.886307 kernel: pci 0000:00:02.2: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 03] add_size 100000 add_align 100000 Feb 13 19:08:54.886388 kernel: pci 0000:00:02.2: bridge window [mem 0x00100000-0x001fffff] to [bus 03] add_size 100000 add_align 100000 Feb 13 19:08:54.886472 kernel: pci 0000:00:02.3: bridge window [io 0x1000-0x0fff] to [bus 04] add_size 1000 Feb 13 19:08:54.886554 kernel: pci 0000:00:02.3: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 04] add_size 100000 add_align 100000 Feb 13 19:08:54.886634 kernel: pci 0000:00:02.3: bridge window [mem 0x00100000-0x000fffff] to [bus 04] add_size 200000 add_align 100000 Feb 13 19:08:54.886717 kernel: pci 0000:00:02.4: bridge window [io 0x1000-0x0fff] to [bus 05] add_size 1000 Feb 13 19:08:54.886798 kernel: pci 0000:00:02.4: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 05] add_size 100000 add_align 100000 Feb 13 19:08:54.886894 kernel: pci 0000:00:02.4: bridge window [mem 0x00100000-0x001fffff] to [bus 05] add_size 100000 add_align 100000 Feb 13 19:08:54.886985 kernel: pci 0000:00:02.5: bridge window [io 0x1000-0x0fff] to [bus 06] add_size 1000 Feb 13 19:08:54.887108 kernel: pci 0000:00:02.5: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 06] add_size 100000 add_align 100000 Feb 13 19:08:54.887195 kernel: pci 0000:00:02.5: bridge window [mem 0x00100000-0x001fffff] to [bus 06] add_size 100000 add_align 100000 Feb 13 19:08:54.887278 kernel: pci 0000:00:02.6: bridge window [io 0x1000-0x0fff] to [bus 07] add_size 1000 Feb 13 19:08:54.887357 kernel: pci 0000:00:02.6: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 07] add_size 100000 add_align 100000 Feb 13 19:08:54.887436 kernel: pci 0000:00:02.6: bridge window [mem 0x00100000-0x001fffff] to [bus 07] add_size 100000 add_align 100000 Feb 13 19:08:54.887520 kernel: pci 0000:00:02.7: bridge window [io 0x1000-0x0fff] to [bus 08] add_size 1000 Feb 13 19:08:54.887604 kernel: pci 0000:00:02.7: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 08] add_size 200000 add_align 100000 Feb 13 19:08:54.887685 kernel: pci 0000:00:02.7: bridge window [mem 0x00100000-0x000fffff] to [bus 08] add_size 200000 add_align 100000 Feb 13 19:08:54.887770 kernel: pci 0000:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 09] add_size 1000 Feb 13 19:08:54.887860 kernel: pci 0000:00:03.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 09] add_size 200000 add_align 100000 Feb 13 19:08:54.887945 kernel: pci 0000:00:03.0: bridge window [mem 0x00100000-0x000fffff] to [bus 09] add_size 200000 add_align 100000 Feb 13 19:08:54.888026 kernel: pci 0000:00:02.0: BAR 14: assigned [mem 0x10000000-0x101fffff] Feb 13 19:08:54.888249 kernel: pci 0000:00:02.0: BAR 15: assigned [mem 0x8000000000-0x80001fffff 64bit pref] Feb 13 19:08:54.888343 kernel: pci 0000:00:02.1: BAR 14: assigned [mem 0x10200000-0x103fffff] Feb 13 19:08:54.888422 kernel: pci 0000:00:02.1: BAR 15: assigned [mem 0x8000200000-0x80003fffff 64bit pref] Feb 13 19:08:54.888502 kernel: pci 0000:00:02.2: BAR 14: assigned [mem 0x10400000-0x105fffff] Feb 13 19:08:54.888581 kernel: pci 0000:00:02.2: BAR 15: assigned [mem 0x8000400000-0x80005fffff 64bit pref] Feb 13 19:08:54.888661 kernel: pci 0000:00:02.3: BAR 14: assigned [mem 0x10600000-0x107fffff] Feb 13 19:08:54.888740 kernel: pci 0000:00:02.3: BAR 15: assigned [mem 0x8000600000-0x80007fffff 64bit pref] Feb 13 19:08:54.888819 kernel: pci 0000:00:02.4: BAR 14: assigned [mem 0x10800000-0x109fffff] Feb 13 19:08:54.888914 kernel: pci 0000:00:02.4: BAR 15: assigned [mem 0x8000800000-0x80009fffff 64bit pref] Feb 13 19:08:54.888997 kernel: pci 0000:00:02.5: BAR 14: assigned [mem 0x10a00000-0x10bfffff] Feb 13 19:08:54.889086 kernel: pci 0000:00:02.5: BAR 15: assigned [mem 0x8000a00000-0x8000bfffff 64bit pref] Feb 13 19:08:54.889167 kernel: pci 0000:00:02.6: BAR 14: assigned [mem 0x10c00000-0x10dfffff] Feb 13 19:08:54.889246 kernel: pci 0000:00:02.6: BAR 15: assigned [mem 0x8000c00000-0x8000dfffff 64bit pref] Feb 13 19:08:54.889328 kernel: pci 0000:00:02.7: BAR 14: assigned [mem 0x10e00000-0x10ffffff] Feb 13 19:08:54.889408 kernel: pci 0000:00:02.7: BAR 15: assigned [mem 0x8000e00000-0x8000ffffff 64bit pref] Feb 13 19:08:54.889494 kernel: pci 0000:00:03.0: BAR 14: assigned [mem 0x11000000-0x111fffff] Feb 13 19:08:54.889573 kernel: pci 0000:00:03.0: BAR 15: assigned [mem 0x8001000000-0x80011fffff 64bit pref] Feb 13 19:08:54.889658 kernel: pci 0000:00:01.0: BAR 4: assigned [mem 0x8001200000-0x8001203fff 64bit pref] Feb 13 19:08:54.889738 kernel: pci 0000:00:01.0: BAR 1: assigned [mem 0x11200000-0x11200fff] Feb 13 19:08:54.889817 kernel: pci 0000:00:02.0: BAR 0: assigned [mem 0x11201000-0x11201fff] Feb 13 19:08:54.889910 kernel: pci 0000:00:02.0: BAR 13: assigned [io 0x1000-0x1fff] Feb 13 19:08:54.889991 kernel: pci 0000:00:02.1: BAR 0: assigned [mem 0x11202000-0x11202fff] Feb 13 19:08:54.890711 kernel: pci 0000:00:02.1: BAR 13: assigned [io 0x2000-0x2fff] Feb 13 19:08:54.890822 kernel: pci 0000:00:02.2: BAR 0: assigned [mem 0x11203000-0x11203fff] Feb 13 19:08:54.890931 kernel: pci 0000:00:02.2: BAR 13: assigned [io 0x3000-0x3fff] Feb 13 19:08:54.891016 kernel: pci 0000:00:02.3: BAR 0: assigned [mem 0x11204000-0x11204fff] Feb 13 19:08:54.891111 kernel: pci 0000:00:02.3: BAR 13: assigned [io 0x4000-0x4fff] Feb 13 19:08:54.891193 kernel: pci 0000:00:02.4: BAR 0: assigned [mem 0x11205000-0x11205fff] Feb 13 19:08:54.891272 kernel: pci 0000:00:02.4: BAR 13: assigned [io 0x5000-0x5fff] Feb 13 19:08:54.891352 kernel: pci 0000:00:02.5: BAR 0: assigned [mem 0x11206000-0x11206fff] Feb 13 19:08:54.891430 kernel: pci 0000:00:02.5: BAR 13: assigned [io 0x6000-0x6fff] Feb 13 19:08:54.891520 kernel: pci 0000:00:02.6: BAR 0: assigned [mem 0x11207000-0x11207fff] Feb 13 19:08:54.891600 kernel: pci 0000:00:02.6: BAR 13: assigned [io 0x7000-0x7fff] Feb 13 19:08:54.891681 kernel: pci 0000:00:02.7: BAR 0: assigned [mem 0x11208000-0x11208fff] Feb 13 19:08:54.891764 kernel: pci 0000:00:02.7: BAR 13: assigned [io 0x8000-0x8fff] Feb 13 19:08:54.891865 kernel: pci 0000:00:03.0: BAR 0: assigned [mem 0x11209000-0x11209fff] Feb 13 19:08:54.891948 kernel: pci 0000:00:03.0: BAR 13: assigned [io 0x9000-0x9fff] Feb 13 19:08:54.892033 kernel: pci 0000:00:04.0: BAR 0: assigned [io 0xa000-0xa007] Feb 13 19:08:54.892211 kernel: pci 0000:01:00.0: BAR 6: assigned [mem 0x10000000-0x1007ffff pref] Feb 13 19:08:54.892305 kernel: pci 0000:01:00.0: BAR 4: assigned [mem 0x8000000000-0x8000003fff 64bit pref] Feb 13 19:08:54.892387 kernel: pci 0000:01:00.0: BAR 1: assigned [mem 0x10080000-0x10080fff] Feb 13 19:08:54.892467 kernel: pci 0000:00:02.0: PCI bridge to [bus 01] Feb 13 19:08:54.892547 kernel: pci 0000:00:02.0: bridge window [io 0x1000-0x1fff] Feb 13 19:08:54.892627 kernel: pci 0000:00:02.0: bridge window [mem 0x10000000-0x101fffff] Feb 13 19:08:54.892705 kernel: pci 0000:00:02.0: bridge window [mem 0x8000000000-0x80001fffff 64bit pref] Feb 13 19:08:54.892792 kernel: pci 0000:02:00.0: BAR 0: assigned [mem 0x10200000-0x10203fff 64bit] Feb 13 19:08:54.892910 kernel: pci 0000:00:02.1: PCI bridge to [bus 02] Feb 13 19:08:54.892995 kernel: pci 0000:00:02.1: bridge window [io 0x2000-0x2fff] Feb 13 19:08:54.893088 kernel: pci 0000:00:02.1: bridge window [mem 0x10200000-0x103fffff] Feb 13 19:08:54.893171 kernel: pci 0000:00:02.1: bridge window [mem 0x8000200000-0x80003fffff 64bit pref] Feb 13 19:08:54.893259 kernel: pci 0000:03:00.0: BAR 4: assigned [mem 0x8000400000-0x8000403fff 64bit pref] Feb 13 19:08:54.893345 kernel: pci 0000:03:00.0: BAR 1: assigned [mem 0x10400000-0x10400fff] Feb 13 19:08:54.893445 kernel: pci 0000:00:02.2: PCI bridge to [bus 03] Feb 13 19:08:54.893525 kernel: pci 0000:00:02.2: bridge window [io 0x3000-0x3fff] Feb 13 19:08:54.893604 kernel: pci 0000:00:02.2: bridge window [mem 0x10400000-0x105fffff] Feb 13 19:08:54.893684 kernel: pci 0000:00:02.2: bridge window [mem 0x8000400000-0x80005fffff 64bit pref] Feb 13 19:08:54.893771 kernel: pci 0000:04:00.0: BAR 4: assigned [mem 0x8000600000-0x8000603fff 64bit pref] Feb 13 19:08:54.893866 kernel: pci 0000:00:02.3: PCI bridge to [bus 04] Feb 13 19:08:54.893953 kernel: pci 0000:00:02.3: bridge window [io 0x4000-0x4fff] Feb 13 19:08:54.894038 kernel: pci 0000:00:02.3: bridge window [mem 0x10600000-0x107fffff] Feb 13 19:08:54.894199 kernel: pci 0000:00:02.3: bridge window [mem 0x8000600000-0x80007fffff 64bit pref] Feb 13 19:08:54.894294 kernel: pci 0000:05:00.0: BAR 4: assigned [mem 0x8000800000-0x8000803fff 64bit pref] Feb 13 19:08:54.894379 kernel: pci 0000:05:00.0: BAR 1: assigned [mem 0x10800000-0x10800fff] Feb 13 19:08:54.894460 kernel: pci 0000:00:02.4: PCI bridge to [bus 05] Feb 13 19:08:54.894539 kernel: pci 0000:00:02.4: bridge window [io 0x5000-0x5fff] Feb 13 19:08:54.894620 kernel: pci 0000:00:02.4: bridge window [mem 0x10800000-0x109fffff] Feb 13 19:08:54.894701 kernel: pci 0000:00:02.4: bridge window [mem 0x8000800000-0x80009fffff 64bit pref] Feb 13 19:08:54.894808 kernel: pci 0000:06:00.0: BAR 4: assigned [mem 0x8000a00000-0x8000a03fff 64bit pref] Feb 13 19:08:54.894931 kernel: pci 0000:06:00.0: BAR 1: assigned [mem 0x10a00000-0x10a00fff] Feb 13 19:08:54.895018 kernel: pci 0000:00:02.5: PCI bridge to [bus 06] Feb 13 19:08:54.895111 kernel: pci 0000:00:02.5: bridge window [io 0x6000-0x6fff] Feb 13 19:08:54.895193 kernel: pci 0000:00:02.5: bridge window [mem 0x10a00000-0x10bfffff] Feb 13 19:08:54.895273 kernel: pci 0000:00:02.5: bridge window [mem 0x8000a00000-0x8000bfffff 64bit pref] Feb 13 19:08:54.895362 kernel: pci 0000:07:00.0: BAR 6: assigned [mem 0x10c00000-0x10c7ffff pref] Feb 13 19:08:54.895452 kernel: pci 0000:07:00.0: BAR 4: assigned [mem 0x8000c00000-0x8000c03fff 64bit pref] Feb 13 19:08:54.895548 kernel: pci 0000:07:00.0: BAR 1: assigned [mem 0x10c80000-0x10c80fff] Feb 13 19:08:54.895631 kernel: pci 0000:00:02.6: PCI bridge to [bus 07] Feb 13 19:08:54.895712 kernel: pci 0000:00:02.6: bridge window [io 0x7000-0x7fff] Feb 13 19:08:54.895795 kernel: pci 0000:00:02.6: bridge window [mem 0x10c00000-0x10dfffff] Feb 13 19:08:54.895894 kernel: pci 0000:00:02.6: bridge window [mem 0x8000c00000-0x8000dfffff 64bit pref] Feb 13 19:08:54.895994 kernel: pci 0000:00:02.7: PCI bridge to [bus 08] Feb 13 19:08:54.896114 kernel: pci 0000:00:02.7: bridge window [io 0x8000-0x8fff] Feb 13 19:08:54.896219 kernel: pci 0000:00:02.7: bridge window [mem 0x10e00000-0x10ffffff] Feb 13 19:08:54.896314 kernel: pci 0000:00:02.7: bridge window [mem 0x8000e00000-0x8000ffffff 64bit pref] Feb 13 19:08:54.896409 kernel: pci 0000:00:03.0: PCI bridge to [bus 09] Feb 13 19:08:54.896490 kernel: pci 0000:00:03.0: bridge window [io 0x9000-0x9fff] Feb 13 19:08:54.896570 kernel: pci 0000:00:03.0: bridge window [mem 0x11000000-0x111fffff] Feb 13 19:08:54.896651 kernel: pci 0000:00:03.0: bridge window [mem 0x8001000000-0x80011fffff 64bit pref] Feb 13 19:08:54.896734 kernel: pci_bus 0000:00: resource 4 [mem 0x10000000-0x3efeffff window] Feb 13 19:08:54.896825 kernel: pci_bus 0000:00: resource 5 [io 0x0000-0xffff window] Feb 13 19:08:54.896920 kernel: pci_bus 0000:00: resource 6 [mem 0x8000000000-0xffffffffff window] Feb 13 19:08:54.897012 kernel: pci_bus 0000:01: resource 0 [io 0x1000-0x1fff] Feb 13 19:08:54.897105 kernel: pci_bus 0000:01: resource 1 [mem 0x10000000-0x101fffff] Feb 13 19:08:54.897185 kernel: pci_bus 0000:01: resource 2 [mem 0x8000000000-0x80001fffff 64bit pref] Feb 13 19:08:54.897270 kernel: pci_bus 0000:02: resource 0 [io 0x2000-0x2fff] Feb 13 19:08:54.899173 kernel: pci_bus 0000:02: resource 1 [mem 0x10200000-0x103fffff] Feb 13 19:08:54.899282 kernel: pci_bus 0000:02: resource 2 [mem 0x8000200000-0x80003fffff 64bit pref] Feb 13 19:08:54.899382 kernel: pci_bus 0000:03: resource 0 [io 0x3000-0x3fff] Feb 13 19:08:54.899463 kernel: pci_bus 0000:03: resource 1 [mem 0x10400000-0x105fffff] Feb 13 19:08:54.899542 kernel: pci_bus 0000:03: resource 2 [mem 0x8000400000-0x80005fffff 64bit pref] Feb 13 19:08:54.899629 kernel: pci_bus 0000:04: resource 0 [io 0x4000-0x4fff] Feb 13 19:08:54.899705 kernel: pci_bus 0000:04: resource 1 [mem 0x10600000-0x107fffff] Feb 13 19:08:54.899780 kernel: pci_bus 0000:04: resource 2 [mem 0x8000600000-0x80007fffff 64bit pref] Feb 13 19:08:54.899894 kernel: pci_bus 0000:05: resource 0 [io 0x5000-0x5fff] Feb 13 19:08:54.899975 kernel: pci_bus 0000:05: resource 1 [mem 0x10800000-0x109fffff] Feb 13 19:08:54.900083 kernel: pci_bus 0000:05: resource 2 [mem 0x8000800000-0x80009fffff 64bit pref] Feb 13 19:08:54.900182 kernel: pci_bus 0000:06: resource 0 [io 0x6000-0x6fff] Feb 13 19:08:54.900281 kernel: pci_bus 0000:06: resource 1 [mem 0x10a00000-0x10bfffff] Feb 13 19:08:54.900363 kernel: pci_bus 0000:06: resource 2 [mem 0x8000a00000-0x8000bfffff 64bit pref] Feb 13 19:08:54.900447 kernel: pci_bus 0000:07: resource 0 [io 0x7000-0x7fff] Feb 13 19:08:54.900523 kernel: pci_bus 0000:07: resource 1 [mem 0x10c00000-0x10dfffff] Feb 13 19:08:54.900598 kernel: pci_bus 0000:07: resource 2 [mem 0x8000c00000-0x8000dfffff 64bit pref] Feb 13 19:08:54.900689 kernel: pci_bus 0000:08: resource 0 [io 0x8000-0x8fff] Feb 13 19:08:54.900766 kernel: pci_bus 0000:08: resource 1 [mem 0x10e00000-0x10ffffff] Feb 13 19:08:54.900856 kernel: pci_bus 0000:08: resource 2 [mem 0x8000e00000-0x8000ffffff 64bit pref] Feb 13 19:08:54.900941 kernel: pci_bus 0000:09: resource 0 [io 0x9000-0x9fff] Feb 13 19:08:54.901018 kernel: pci_bus 0000:09: resource 1 [mem 0x11000000-0x111fffff] Feb 13 19:08:54.901105 kernel: pci_bus 0000:09: resource 2 [mem 0x8001000000-0x80011fffff 64bit pref] Feb 13 19:08:54.901122 kernel: ACPI: PCI: Interrupt link GSI0 configured for IRQ 35 Feb 13 19:08:54.901132 kernel: ACPI: PCI: Interrupt link GSI1 configured for IRQ 36 Feb 13 19:08:54.901142 kernel: ACPI: PCI: Interrupt link GSI2 configured for IRQ 37 Feb 13 19:08:54.901152 kernel: ACPI: PCI: Interrupt link GSI3 configured for IRQ 38 Feb 13 19:08:54.901169 kernel: iommu: Default domain type: Translated Feb 13 19:08:54.901185 kernel: iommu: DMA domain TLB invalidation policy: strict mode Feb 13 19:08:54.901195 kernel: efivars: Registered efivars operations Feb 13 19:08:54.901205 kernel: vgaarb: loaded Feb 13 19:08:54.901214 kernel: clocksource: Switched to clocksource arch_sys_counter Feb 13 19:08:54.901224 kernel: VFS: Disk quotas dquot_6.6.0 Feb 13 19:08:54.901234 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Feb 13 19:08:54.901244 kernel: pnp: PnP ACPI init Feb 13 19:08:54.901336 kernel: system 00:00: [mem 0x4010000000-0x401fffffff window] could not be reserved Feb 13 19:08:54.901352 kernel: pnp: PnP ACPI: found 1 devices Feb 13 19:08:54.901362 kernel: NET: Registered PF_INET protocol family Feb 13 19:08:54.901372 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Feb 13 19:08:54.901381 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Feb 13 19:08:54.901391 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Feb 13 19:08:54.901400 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Feb 13 19:08:54.901410 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Feb 13 19:08:54.901420 kernel: TCP: Hash tables configured (established 32768 bind 32768) Feb 13 19:08:54.901431 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Feb 13 19:08:54.901441 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Feb 13 19:08:54.901450 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Feb 13 19:08:54.901541 kernel: pci 0000:02:00.0: enabling device (0000 -> 0002) Feb 13 19:08:54.901555 kernel: PCI: CLS 0 bytes, default 64 Feb 13 19:08:54.901565 kernel: kvm [1]: HYP mode not available Feb 13 19:08:54.901574 kernel: Initialise system trusted keyrings Feb 13 19:08:54.901584 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Feb 13 19:08:54.901593 kernel: Key type asymmetric registered Feb 13 19:08:54.901605 kernel: Asymmetric key parser 'x509' registered Feb 13 19:08:54.901614 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Feb 13 19:08:54.901623 kernel: io scheduler mq-deadline registered Feb 13 19:08:54.901633 kernel: io scheduler kyber registered Feb 13 19:08:54.901642 kernel: io scheduler bfq registered Feb 13 19:08:54.901652 kernel: ACPI: \_SB_.PCI0.GSI2: Enabled at IRQ 37 Feb 13 19:08:54.901737 kernel: pcieport 0000:00:02.0: PME: Signaling with IRQ 50 Feb 13 19:08:54.901820 kernel: pcieport 0000:00:02.0: AER: enabled with IRQ 50 Feb 13 19:08:54.901928 kernel: pcieport 0000:00:02.0: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Feb 13 19:08:54.902015 kernel: pcieport 0000:00:02.1: PME: Signaling with IRQ 51 Feb 13 19:08:54.903802 kernel: pcieport 0000:00:02.1: AER: enabled with IRQ 51 Feb 13 19:08:54.906165 kernel: pcieport 0000:00:02.1: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Feb 13 19:08:54.906290 kernel: pcieport 0000:00:02.2: PME: Signaling with IRQ 52 Feb 13 19:08:54.906375 kernel: pcieport 0000:00:02.2: AER: enabled with IRQ 52 Feb 13 19:08:54.906465 kernel: pcieport 0000:00:02.2: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Feb 13 19:08:54.906551 kernel: pcieport 0000:00:02.3: PME: Signaling with IRQ 53 Feb 13 19:08:54.906634 kernel: pcieport 0000:00:02.3: AER: enabled with IRQ 53 Feb 13 19:08:54.906714 kernel: pcieport 0000:00:02.3: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Feb 13 19:08:54.906803 kernel: pcieport 0000:00:02.4: PME: Signaling with IRQ 54 Feb 13 19:08:54.906906 kernel: pcieport 0000:00:02.4: AER: enabled with IRQ 54 Feb 13 19:08:54.906995 kernel: pcieport 0000:00:02.4: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Feb 13 19:08:54.907110 kernel: pcieport 0000:00:02.5: PME: Signaling with IRQ 55 Feb 13 19:08:54.907193 kernel: pcieport 0000:00:02.5: AER: enabled with IRQ 55 Feb 13 19:08:54.907273 kernel: pcieport 0000:00:02.5: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Feb 13 19:08:54.907358 kernel: pcieport 0000:00:02.6: PME: Signaling with IRQ 56 Feb 13 19:08:54.907439 kernel: pcieport 0000:00:02.6: AER: enabled with IRQ 56 Feb 13 19:08:54.907524 kernel: pcieport 0000:00:02.6: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Feb 13 19:08:54.907609 kernel: pcieport 0000:00:02.7: PME: Signaling with IRQ 57 Feb 13 19:08:54.907691 kernel: pcieport 0000:00:02.7: AER: enabled with IRQ 57 Feb 13 19:08:54.907775 kernel: pcieport 0000:00:02.7: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Feb 13 19:08:54.907788 kernel: ACPI: \_SB_.PCI0.GSI3: Enabled at IRQ 38 Feb 13 19:08:54.907885 kernel: pcieport 0000:00:03.0: PME: Signaling with IRQ 58 Feb 13 19:08:54.907973 kernel: pcieport 0000:00:03.0: AER: enabled with IRQ 58 Feb 13 19:08:54.910089 kernel: pcieport 0000:00:03.0: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Feb 13 19:08:54.910138 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0C:00/input/input0 Feb 13 19:08:54.910165 kernel: ACPI: button: Power Button [PWRB] Feb 13 19:08:54.910185 kernel: ACPI: \_SB_.PCI0.GSI1: Enabled at IRQ 36 Feb 13 19:08:54.910411 kernel: virtio-pci 0000:04:00.0: enabling device (0000 -> 0002) Feb 13 19:08:54.910594 kernel: virtio-pci 0000:07:00.0: enabling device (0000 -> 0002) Feb 13 19:08:54.910619 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Feb 13 19:08:54.910650 kernel: ACPI: \_SB_.PCI0.GSI0: Enabled at IRQ 35 Feb 13 19:08:54.910831 kernel: serial 0000:00:04.0: enabling device (0000 -> 0001) Feb 13 19:08:54.910928 kernel: 0000:00:04.0: ttyS0 at I/O 0xa000 (irq = 45, base_baud = 115200) is a 16550A Feb 13 19:08:54.910949 kernel: thunder_xcv, ver 1.0 Feb 13 19:08:54.910967 kernel: thunder_bgx, ver 1.0 Feb 13 19:08:54.910985 kernel: nicpf, ver 1.0 Feb 13 19:08:54.911008 kernel: nicvf, ver 1.0 Feb 13 19:08:54.911280 kernel: rtc-efi rtc-efi.0: registered as rtc0 Feb 13 19:08:54.911349 kernel: rtc-efi rtc-efi.0: setting system clock to 2025-02-13T19:08:54 UTC (1739473734) Feb 13 19:08:54.911365 kernel: hid: raw HID events driver (C) Jiri Kosina Feb 13 19:08:54.911374 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 7 counters available Feb 13 19:08:54.911382 kernel: watchdog: Delayed init of the lockup detector failed: -19 Feb 13 19:08:54.911390 kernel: watchdog: Hard watchdog permanently disabled Feb 13 19:08:54.911399 kernel: NET: Registered PF_INET6 protocol family Feb 13 19:08:54.911407 kernel: Segment Routing with IPv6 Feb 13 19:08:54.911415 kernel: In-situ OAM (IOAM) with IPv6 Feb 13 19:08:54.911423 kernel: NET: Registered PF_PACKET protocol family Feb 13 19:08:54.911433 kernel: Key type dns_resolver registered Feb 13 19:08:54.911442 kernel: registered taskstats version 1 Feb 13 19:08:54.911450 kernel: Loading compiled-in X.509 certificates Feb 13 19:08:54.911458 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.74-flatcar: 58bec1a0c6b8a133d1af4ea745973da0351f7027' Feb 13 19:08:54.911466 kernel: Key type .fscrypt registered Feb 13 19:08:54.911474 kernel: Key type fscrypt-provisioning registered Feb 13 19:08:54.911482 kernel: ima: No TPM chip found, activating TPM-bypass! Feb 13 19:08:54.911490 kernel: ima: Allocated hash algorithm: sha1 Feb 13 19:08:54.911498 kernel: ima: No architecture policies found Feb 13 19:08:54.911509 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Feb 13 19:08:54.911517 kernel: clk: Disabling unused clocks Feb 13 19:08:54.911525 kernel: Freeing unused kernel memory: 38336K Feb 13 19:08:54.911533 kernel: Run /init as init process Feb 13 19:08:54.911541 kernel: with arguments: Feb 13 19:08:54.911549 kernel: /init Feb 13 19:08:54.911557 kernel: with environment: Feb 13 19:08:54.911565 kernel: HOME=/ Feb 13 19:08:54.911572 kernel: TERM=linux Feb 13 19:08:54.911582 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Feb 13 19:08:54.911591 systemd[1]: Successfully made /usr/ read-only. Feb 13 19:08:54.911605 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Feb 13 19:08:54.911614 systemd[1]: Detected virtualization kvm. Feb 13 19:08:54.911623 systemd[1]: Detected architecture arm64. Feb 13 19:08:54.911631 systemd[1]: Running in initrd. Feb 13 19:08:54.911639 systemd[1]: No hostname configured, using default hostname. Feb 13 19:08:54.911650 systemd[1]: Hostname set to . Feb 13 19:08:54.911659 systemd[1]: Initializing machine ID from VM UUID. Feb 13 19:08:54.911667 systemd[1]: Queued start job for default target initrd.target. Feb 13 19:08:54.911676 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Feb 13 19:08:54.911685 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Feb 13 19:08:54.911695 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Feb 13 19:08:54.911703 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Feb 13 19:08:54.911712 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Feb 13 19:08:54.911724 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Feb 13 19:08:54.911734 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Feb 13 19:08:54.911743 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Feb 13 19:08:54.911751 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Feb 13 19:08:54.911760 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Feb 13 19:08:54.911776 systemd[1]: Reached target paths.target - Path Units. Feb 13 19:08:54.911785 systemd[1]: Reached target slices.target - Slice Units. Feb 13 19:08:54.911796 systemd[1]: Reached target swap.target - Swaps. Feb 13 19:08:54.911804 systemd[1]: Reached target timers.target - Timer Units. Feb 13 19:08:54.911813 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Feb 13 19:08:54.911821 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Feb 13 19:08:54.911830 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Feb 13 19:08:54.911846 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Feb 13 19:08:54.911855 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Feb 13 19:08:54.911864 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Feb 13 19:08:54.911873 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Feb 13 19:08:54.911884 systemd[1]: Reached target sockets.target - Socket Units. Feb 13 19:08:54.911893 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Feb 13 19:08:54.911901 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Feb 13 19:08:54.911910 systemd[1]: Finished network-cleanup.service - Network Cleanup. Feb 13 19:08:54.911918 systemd[1]: Starting systemd-fsck-usr.service... Feb 13 19:08:54.911927 systemd[1]: Starting systemd-journald.service - Journal Service... Feb 13 19:08:54.911936 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Feb 13 19:08:54.911945 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 19:08:54.911984 systemd-journald[237]: Collecting audit messages is disabled. Feb 13 19:08:54.912007 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Feb 13 19:08:54.912016 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Feb 13 19:08:54.912027 systemd[1]: Finished systemd-fsck-usr.service. Feb 13 19:08:54.912036 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Feb 13 19:08:54.912046 systemd-journald[237]: Journal started Feb 13 19:08:54.912330 systemd-journald[237]: Runtime Journal (/run/log/journal/7ecaf9133ce94e0d8dfa2faad6c7d0cd) is 8M, max 76.6M, 68.6M free. Feb 13 19:08:54.912382 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Feb 13 19:08:54.897467 systemd-modules-load[239]: Inserted module 'overlay' Feb 13 19:08:54.917378 systemd[1]: Started systemd-journald.service - Journal Service. Feb 13 19:08:54.917416 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Feb 13 19:08:54.919389 systemd-modules-load[239]: Inserted module 'br_netfilter' Feb 13 19:08:54.920075 kernel: Bridge firewalling registered Feb 13 19:08:54.920702 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 19:08:54.922245 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Feb 13 19:08:54.929586 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Feb 13 19:08:54.933451 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Feb 13 19:08:54.937264 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Feb 13 19:08:54.941016 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Feb 13 19:08:54.957405 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Feb 13 19:08:54.960352 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Feb 13 19:08:54.963461 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Feb 13 19:08:54.964476 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 19:08:54.972355 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Feb 13 19:08:54.978331 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Feb 13 19:08:54.989126 dracut-cmdline[273]: dracut-dracut-053 Feb 13 19:08:54.994408 dracut-cmdline[273]: Using kernel command line parameters: rd.driver.pre=btrfs BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyAMA0,115200n8 flatcar.first_boot=detected acpi=force flatcar.oem.id=hetzner verity.usrhash=f06bad36699a22ae88c1968cd72b62b3503d97da521712e50a4b744320b1ba33 Feb 13 19:08:55.027880 systemd-resolved[276]: Positive Trust Anchors: Feb 13 19:08:55.027899 systemd-resolved[276]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Feb 13 19:08:55.027930 systemd-resolved[276]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Feb 13 19:08:55.035551 systemd-resolved[276]: Defaulting to hostname 'linux'. Feb 13 19:08:55.036877 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Feb 13 19:08:55.037671 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Feb 13 19:08:55.107112 kernel: SCSI subsystem initialized Feb 13 19:08:55.111088 kernel: Loading iSCSI transport class v2.0-870. Feb 13 19:08:55.119108 kernel: iscsi: registered transport (tcp) Feb 13 19:08:55.132190 kernel: iscsi: registered transport (qla4xxx) Feb 13 19:08:55.132286 kernel: QLogic iSCSI HBA Driver Feb 13 19:08:55.180172 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Feb 13 19:08:55.186309 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Feb 13 19:08:55.203666 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Feb 13 19:08:55.203742 kernel: device-mapper: uevent: version 1.0.3 Feb 13 19:08:55.203754 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Feb 13 19:08:55.252108 kernel: raid6: neonx8 gen() 15653 MB/s Feb 13 19:08:55.269124 kernel: raid6: neonx4 gen() 15748 MB/s Feb 13 19:08:55.286128 kernel: raid6: neonx2 gen() 13132 MB/s Feb 13 19:08:55.303200 kernel: raid6: neonx1 gen() 10437 MB/s Feb 13 19:08:55.320097 kernel: raid6: int64x8 gen() 6755 MB/s Feb 13 19:08:55.337142 kernel: raid6: int64x4 gen() 7297 MB/s Feb 13 19:08:55.354143 kernel: raid6: int64x2 gen() 6039 MB/s Feb 13 19:08:55.371148 kernel: raid6: int64x1 gen() 4987 MB/s Feb 13 19:08:55.371245 kernel: raid6: using algorithm neonx4 gen() 15748 MB/s Feb 13 19:08:55.388146 kernel: raid6: .... xor() 12284 MB/s, rmw enabled Feb 13 19:08:55.388220 kernel: raid6: using neon recovery algorithm Feb 13 19:08:55.393278 kernel: xor: measuring software checksum speed Feb 13 19:08:55.393324 kernel: 8regs : 21607 MB/sec Feb 13 19:08:55.393346 kernel: 32regs : 20582 MB/sec Feb 13 19:08:55.394107 kernel: arm64_neon : 27908 MB/sec Feb 13 19:08:55.394157 kernel: xor: using function: arm64_neon (27908 MB/sec) Feb 13 19:08:55.445115 kernel: Btrfs loaded, zoned=no, fsverity=no Feb 13 19:08:55.460793 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Feb 13 19:08:55.473415 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Feb 13 19:08:55.489034 systemd-udevd[456]: Using default interface naming scheme 'v255'. Feb 13 19:08:55.493022 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Feb 13 19:08:55.502473 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Feb 13 19:08:55.520981 dracut-pre-trigger[464]: rd.md=0: removing MD RAID activation Feb 13 19:08:55.563908 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Feb 13 19:08:55.572428 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Feb 13 19:08:55.625497 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Feb 13 19:08:55.634292 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Feb 13 19:08:55.660991 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Feb 13 19:08:55.664697 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Feb 13 19:08:55.667489 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Feb 13 19:08:55.668563 systemd[1]: Reached target remote-fs.target - Remote File Systems. Feb 13 19:08:55.676258 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Feb 13 19:08:55.696792 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Feb 13 19:08:55.732291 kernel: scsi host0: Virtio SCSI HBA Feb 13 19:08:55.734169 kernel: scsi 0:0:0:0: CD-ROM QEMU QEMU CD-ROM 2.5+ PQ: 0 ANSI: 5 Feb 13 19:08:55.735413 kernel: scsi 0:0:0:1: Direct-Access QEMU QEMU HARDDISK 2.5+ PQ: 0 ANSI: 5 Feb 13 19:08:55.773381 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Feb 13 19:08:55.773501 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 19:08:55.777236 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Feb 13 19:08:55.777814 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Feb 13 19:08:55.777987 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 19:08:55.779621 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 19:08:55.788352 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 19:08:55.800091 kernel: ACPI: bus type USB registered Feb 13 19:08:55.800144 kernel: usbcore: registered new interface driver usbfs Feb 13 19:08:55.803487 kernel: usbcore: registered new interface driver hub Feb 13 19:08:55.806139 kernel: usbcore: registered new device driver usb Feb 13 19:08:55.808098 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 19:08:55.811118 kernel: sr 0:0:0:0: Power-on or device reset occurred Feb 13 19:08:55.816304 kernel: sr 0:0:0:0: [sr0] scsi3-mmc drive: 16x/50x cd/rw xa/form2 cdda tray Feb 13 19:08:55.816424 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Feb 13 19:08:55.816436 kernel: sr 0:0:0:0: Attached scsi CD-ROM sr0 Feb 13 19:08:55.816253 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Feb 13 19:08:55.819554 kernel: sd 0:0:0:1: Power-on or device reset occurred Feb 13 19:08:55.832188 kernel: sd 0:0:0:1: [sda] 80003072 512-byte logical blocks: (41.0 GB/38.1 GiB) Feb 13 19:08:55.832334 kernel: sd 0:0:0:1: [sda] Write Protect is off Feb 13 19:08:55.832427 kernel: sd 0:0:0:1: [sda] Mode Sense: 63 00 00 08 Feb 13 19:08:55.832511 kernel: sd 0:0:0:1: [sda] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA Feb 13 19:08:55.832601 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Feb 13 19:08:55.832613 kernel: GPT:17805311 != 80003071 Feb 13 19:08:55.832622 kernel: GPT:Alternate GPT header not at the end of the disk. Feb 13 19:08:55.832632 kernel: GPT:17805311 != 80003071 Feb 13 19:08:55.832641 kernel: GPT: Use GNU Parted to correct GPT errors. Feb 13 19:08:55.832650 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Feb 13 19:08:55.832660 kernel: sd 0:0:0:1: [sda] Attached SCSI disk Feb 13 19:08:55.837643 kernel: xhci_hcd 0000:02:00.0: xHCI Host Controller Feb 13 19:08:55.848354 kernel: xhci_hcd 0000:02:00.0: new USB bus registered, assigned bus number 1 Feb 13 19:08:55.849146 kernel: xhci_hcd 0000:02:00.0: hcc params 0x00087001 hci version 0x100 quirks 0x0000000000000010 Feb 13 19:08:55.849284 kernel: xhci_hcd 0000:02:00.0: xHCI Host Controller Feb 13 19:08:55.849375 kernel: xhci_hcd 0000:02:00.0: new USB bus registered, assigned bus number 2 Feb 13 19:08:55.849460 kernel: xhci_hcd 0000:02:00.0: Host supports USB 3.0 SuperSpeed Feb 13 19:08:55.849546 kernel: hub 1-0:1.0: USB hub found Feb 13 19:08:55.849651 kernel: hub 1-0:1.0: 4 ports detected Feb 13 19:08:55.849753 kernel: usb usb2: We don't know the algorithms for LPM for this host, disabling LPM. Feb 13 19:08:55.849941 kernel: hub 2-0:1.0: USB hub found Feb 13 19:08:55.850044 kernel: hub 2-0:1.0: 4 ports detected Feb 13 19:08:55.844200 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 19:08:55.889152 kernel: BTRFS: device fsid 4fff035f-dd55-45d8-9bb7-2a61f21b22d5 devid 1 transid 38 /dev/sda3 scanned by (udev-worker) (520) Feb 13 19:08:55.897089 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/sda6 scanned by (udev-worker) (507) Feb 13 19:08:55.903996 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - QEMU_HARDDISK ROOT. Feb 13 19:08:55.912547 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - QEMU_HARDDISK USR-A. Feb 13 19:08:55.913303 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - QEMU_HARDDISK USR-A. Feb 13 19:08:55.928400 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - QEMU_HARDDISK EFI-SYSTEM. Feb 13 19:08:55.950454 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - QEMU_HARDDISK OEM. Feb 13 19:08:55.965441 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Feb 13 19:08:55.974652 disk-uuid[576]: Primary Header is updated. Feb 13 19:08:55.974652 disk-uuid[576]: Secondary Entries is updated. Feb 13 19:08:55.974652 disk-uuid[576]: Secondary Header is updated. Feb 13 19:08:55.985853 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Feb 13 19:08:56.089079 kernel: usb 1-1: new high-speed USB device number 2 using xhci_hcd Feb 13 19:08:56.331223 kernel: usb 1-2: new high-speed USB device number 3 using xhci_hcd Feb 13 19:08:56.467789 kernel: input: QEMU QEMU USB Tablet as /devices/pci0000:00/0000:00:02.1/0000:02:00.0/usb1/1-1/1-1:1.0/0003:0627:0001.0001/input/input1 Feb 13 19:08:56.467868 kernel: hid-generic 0003:0627:0001.0001: input,hidraw0: USB HID v0.01 Mouse [QEMU QEMU USB Tablet] on usb-0000:02:00.0-1/input0 Feb 13 19:08:56.468101 kernel: input: QEMU QEMU USB Keyboard as /devices/pci0000:00/0000:00:02.1/0000:02:00.0/usb1/1-2/1-2:1.0/0003:0627:0001.0002/input/input2 Feb 13 19:08:56.522103 kernel: hid-generic 0003:0627:0001.0002: input,hidraw1: USB HID v1.11 Keyboard [QEMU QEMU USB Keyboard] on usb-0000:02:00.0-2/input0 Feb 13 19:08:56.522407 kernel: usbcore: registered new interface driver usbhid Feb 13 19:08:56.522446 kernel: usbhid: USB HID core driver Feb 13 19:08:57.000091 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Feb 13 19:08:57.001788 disk-uuid[577]: The operation has completed successfully. Feb 13 19:08:57.059605 systemd[1]: disk-uuid.service: Deactivated successfully. Feb 13 19:08:57.060966 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Feb 13 19:08:57.092282 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Feb 13 19:08:57.098815 sh[592]: Success Feb 13 19:08:57.111177 kernel: device-mapper: verity: sha256 using implementation "sha256-ce" Feb 13 19:08:57.167603 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Feb 13 19:08:57.174483 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Feb 13 19:08:57.179215 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Feb 13 19:08:57.201749 kernel: BTRFS info (device dm-0): first mount of filesystem 4fff035f-dd55-45d8-9bb7-2a61f21b22d5 Feb 13 19:08:57.201886 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm Feb 13 19:08:57.201911 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Feb 13 19:08:57.201946 kernel: BTRFS info (device dm-0): disabling log replay at mount time Feb 13 19:08:57.202429 kernel: BTRFS info (device dm-0): using free space tree Feb 13 19:08:57.209117 kernel: BTRFS info (device dm-0): enabling ssd optimizations Feb 13 19:08:57.211704 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Feb 13 19:08:57.212522 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Feb 13 19:08:57.222306 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Feb 13 19:08:57.227314 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Feb 13 19:08:57.243270 kernel: BTRFS info (device sda6): first mount of filesystem 843e6c1f-b3c4-44a3-b5c6-7983dd77012d Feb 13 19:08:57.243334 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Feb 13 19:08:57.243346 kernel: BTRFS info (device sda6): using free space tree Feb 13 19:08:57.248082 kernel: BTRFS info (device sda6): enabling ssd optimizations Feb 13 19:08:57.248145 kernel: BTRFS info (device sda6): auto enabling async discard Feb 13 19:08:57.261564 systemd[1]: mnt-oem.mount: Deactivated successfully. Feb 13 19:08:57.264117 kernel: BTRFS info (device sda6): last unmount of filesystem 843e6c1f-b3c4-44a3-b5c6-7983dd77012d Feb 13 19:08:57.271458 systemd[1]: Finished ignition-setup.service - Ignition (setup). Feb 13 19:08:57.277270 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Feb 13 19:08:57.360602 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Feb 13 19:08:57.371818 systemd[1]: Starting systemd-networkd.service - Network Configuration... Feb 13 19:08:57.379737 ignition[679]: Ignition 2.20.0 Feb 13 19:08:57.381724 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Feb 13 19:08:57.379750 ignition[679]: Stage: fetch-offline Feb 13 19:08:57.379788 ignition[679]: no configs at "/usr/lib/ignition/base.d" Feb 13 19:08:57.379797 ignition[679]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Feb 13 19:08:57.379963 ignition[679]: parsed url from cmdline: "" Feb 13 19:08:57.379966 ignition[679]: no config URL provided Feb 13 19:08:57.379970 ignition[679]: reading system config file "/usr/lib/ignition/user.ign" Feb 13 19:08:57.379977 ignition[679]: no config at "/usr/lib/ignition/user.ign" Feb 13 19:08:57.379982 ignition[679]: failed to fetch config: resource requires networking Feb 13 19:08:57.380261 ignition[679]: Ignition finished successfully Feb 13 19:08:57.402136 systemd-networkd[779]: lo: Link UP Feb 13 19:08:57.402146 systemd-networkd[779]: lo: Gained carrier Feb 13 19:08:57.405549 systemd-networkd[779]: Enumeration completed Feb 13 19:08:57.405681 systemd[1]: Started systemd-networkd.service - Network Configuration. Feb 13 19:08:57.407441 systemd-networkd[779]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 19:08:57.407445 systemd-networkd[779]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Feb 13 19:08:57.408247 systemd[1]: Reached target network.target - Network. Feb 13 19:08:57.409203 systemd-networkd[779]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 19:08:57.409207 systemd-networkd[779]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Feb 13 19:08:57.410216 systemd-networkd[779]: eth0: Link UP Feb 13 19:08:57.410219 systemd-networkd[779]: eth0: Gained carrier Feb 13 19:08:57.410227 systemd-networkd[779]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 19:08:57.415000 systemd-networkd[779]: eth1: Link UP Feb 13 19:08:57.415003 systemd-networkd[779]: eth1: Gained carrier Feb 13 19:08:57.415012 systemd-networkd[779]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 19:08:57.415025 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Feb 13 19:08:57.426979 ignition[785]: Ignition 2.20.0 Feb 13 19:08:57.426995 ignition[785]: Stage: fetch Feb 13 19:08:57.427196 ignition[785]: no configs at "/usr/lib/ignition/base.d" Feb 13 19:08:57.427207 ignition[785]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Feb 13 19:08:57.427317 ignition[785]: parsed url from cmdline: "" Feb 13 19:08:57.427321 ignition[785]: no config URL provided Feb 13 19:08:57.427325 ignition[785]: reading system config file "/usr/lib/ignition/user.ign" Feb 13 19:08:57.427333 ignition[785]: no config at "/usr/lib/ignition/user.ign" Feb 13 19:08:57.427416 ignition[785]: GET http://169.254.169.254/hetzner/v1/userdata: attempt #1 Feb 13 19:08:57.428096 ignition[785]: GET error: Get "http://169.254.169.254/hetzner/v1/userdata": dial tcp 169.254.169.254:80: connect: network is unreachable Feb 13 19:08:57.450218 systemd-networkd[779]: eth1: DHCPv4 address 10.0.0.3/32, gateway 10.0.0.1 acquired from 10.0.0.1 Feb 13 19:08:57.480138 systemd-networkd[779]: eth0: DHCPv4 address 78.46.147.231/32, gateway 172.31.1.1 acquired from 172.31.1.1 Feb 13 19:08:57.628307 ignition[785]: GET http://169.254.169.254/hetzner/v1/userdata: attempt #2 Feb 13 19:08:57.633654 ignition[785]: GET result: OK Feb 13 19:08:57.633785 ignition[785]: parsing config with SHA512: 121b9e36d67bfdba21fc1a71f8df707b1f0b20e414782270808c12abf7c07fd79faafc812d52b46dd569a78e25031ad6d678baf80b79dd35f9558470c0e3bb7a Feb 13 19:08:57.640238 unknown[785]: fetched base config from "system" Feb 13 19:08:57.640642 ignition[785]: fetch: fetch complete Feb 13 19:08:57.640248 unknown[785]: fetched base config from "system" Feb 13 19:08:57.640656 ignition[785]: fetch: fetch passed Feb 13 19:08:57.640253 unknown[785]: fetched user config from "hetzner" Feb 13 19:08:57.640711 ignition[785]: Ignition finished successfully Feb 13 19:08:57.643789 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Feb 13 19:08:57.654239 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Feb 13 19:08:57.671198 ignition[792]: Ignition 2.20.0 Feb 13 19:08:57.671208 ignition[792]: Stage: kargs Feb 13 19:08:57.671382 ignition[792]: no configs at "/usr/lib/ignition/base.d" Feb 13 19:08:57.671392 ignition[792]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Feb 13 19:08:57.672370 ignition[792]: kargs: kargs passed Feb 13 19:08:57.674082 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Feb 13 19:08:57.672417 ignition[792]: Ignition finished successfully Feb 13 19:08:57.681420 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Feb 13 19:08:57.694849 ignition[798]: Ignition 2.20.0 Feb 13 19:08:57.695575 ignition[798]: Stage: disks Feb 13 19:08:57.695775 ignition[798]: no configs at "/usr/lib/ignition/base.d" Feb 13 19:08:57.695785 ignition[798]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Feb 13 19:08:57.696743 ignition[798]: disks: disks passed Feb 13 19:08:57.696792 ignition[798]: Ignition finished successfully Feb 13 19:08:57.698868 systemd[1]: Finished ignition-disks.service - Ignition (disks). Feb 13 19:08:57.700922 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Feb 13 19:08:57.701621 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Feb 13 19:08:57.702997 systemd[1]: Reached target local-fs.target - Local File Systems. Feb 13 19:08:57.704180 systemd[1]: Reached target sysinit.target - System Initialization. Feb 13 19:08:57.705165 systemd[1]: Reached target basic.target - Basic System. Feb 13 19:08:57.714347 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Feb 13 19:08:57.731052 systemd-fsck[806]: ROOT: clean, 14/1628000 files, 120691/1617920 blocks Feb 13 19:08:57.736217 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Feb 13 19:08:58.205201 systemd[1]: Mounting sysroot.mount - /sysroot... Feb 13 19:08:58.271114 kernel: EXT4-fs (sda9): mounted filesystem 24882d04-b1a5-4a27-95f1-925956e69b18 r/w with ordered data mode. Quota mode: none. Feb 13 19:08:58.272404 systemd[1]: Mounted sysroot.mount - /sysroot. Feb 13 19:08:58.273752 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Feb 13 19:08:58.283213 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Feb 13 19:08:58.286789 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Feb 13 19:08:58.289776 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Feb 13 19:08:58.290697 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Feb 13 19:08:58.290734 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Feb 13 19:08:58.301288 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Feb 13 19:08:58.303579 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/sda6 scanned by mount (814) Feb 13 19:08:58.303604 kernel: BTRFS info (device sda6): first mount of filesystem 843e6c1f-b3c4-44a3-b5c6-7983dd77012d Feb 13 19:08:58.303616 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Feb 13 19:08:58.304796 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Feb 13 19:08:58.307735 kernel: BTRFS info (device sda6): using free space tree Feb 13 19:08:58.318696 kernel: BTRFS info (device sda6): enabling ssd optimizations Feb 13 19:08:58.318757 kernel: BTRFS info (device sda6): auto enabling async discard Feb 13 19:08:58.326645 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Feb 13 19:08:58.364422 coreos-metadata[816]: Feb 13 19:08:58.364 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/hostname: Attempt #1 Feb 13 19:08:58.366699 coreos-metadata[816]: Feb 13 19:08:58.366 INFO Fetch successful Feb 13 19:08:58.368148 coreos-metadata[816]: Feb 13 19:08:58.367 INFO wrote hostname ci-4230-0-1-a-2d796e8b97 to /sysroot/etc/hostname Feb 13 19:08:58.368913 initrd-setup-root[841]: cut: /sysroot/etc/passwd: No such file or directory Feb 13 19:08:58.372115 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Feb 13 19:08:58.376994 initrd-setup-root[849]: cut: /sysroot/etc/group: No such file or directory Feb 13 19:08:58.383787 initrd-setup-root[856]: cut: /sysroot/etc/shadow: No such file or directory Feb 13 19:08:58.387974 initrd-setup-root[863]: cut: /sysroot/etc/gshadow: No such file or directory Feb 13 19:08:58.499160 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Feb 13 19:08:58.505171 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Feb 13 19:08:58.508384 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Feb 13 19:08:58.517086 kernel: BTRFS info (device sda6): last unmount of filesystem 843e6c1f-b3c4-44a3-b5c6-7983dd77012d Feb 13 19:08:58.549257 ignition[931]: INFO : Ignition 2.20.0 Feb 13 19:08:58.549257 ignition[931]: INFO : Stage: mount Feb 13 19:08:58.550865 ignition[931]: INFO : no configs at "/usr/lib/ignition/base.d" Feb 13 19:08:58.550865 ignition[931]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Feb 13 19:08:58.555178 ignition[931]: INFO : mount: mount passed Feb 13 19:08:58.555178 ignition[931]: INFO : Ignition finished successfully Feb 13 19:08:58.552887 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Feb 13 19:08:58.554868 systemd[1]: Finished ignition-mount.service - Ignition (mount). Feb 13 19:08:58.560215 systemd[1]: Starting ignition-files.service - Ignition (files)... Feb 13 19:08:59.181314 systemd-networkd[779]: eth0: Gained IPv6LL Feb 13 19:08:59.200886 systemd[1]: sysroot-oem.mount: Deactivated successfully. Feb 13 19:08:59.206346 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Feb 13 19:08:59.221077 kernel: BTRFS: device label OEM devid 1 transid 16 /dev/sda6 scanned by mount (943) Feb 13 19:08:59.225183 kernel: BTRFS info (device sda6): first mount of filesystem 843e6c1f-b3c4-44a3-b5c6-7983dd77012d Feb 13 19:08:59.225241 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Feb 13 19:08:59.225256 kernel: BTRFS info (device sda6): using free space tree Feb 13 19:08:59.229259 kernel: BTRFS info (device sda6): enabling ssd optimizations Feb 13 19:08:59.229339 kernel: BTRFS info (device sda6): auto enabling async discard Feb 13 19:08:59.232414 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Feb 13 19:08:59.253575 ignition[960]: INFO : Ignition 2.20.0 Feb 13 19:08:59.253575 ignition[960]: INFO : Stage: files Feb 13 19:08:59.254851 ignition[960]: INFO : no configs at "/usr/lib/ignition/base.d" Feb 13 19:08:59.254851 ignition[960]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Feb 13 19:08:59.254851 ignition[960]: DEBUG : files: compiled without relabeling support, skipping Feb 13 19:08:59.259839 ignition[960]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Feb 13 19:08:59.259839 ignition[960]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Feb 13 19:08:59.261595 ignition[960]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Feb 13 19:08:59.261595 ignition[960]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Feb 13 19:08:59.261595 ignition[960]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Feb 13 19:08:59.261191 unknown[960]: wrote ssh authorized keys file for user: core Feb 13 19:08:59.267613 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.0-linux-arm64.tar.gz" Feb 13 19:08:59.267613 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.0-linux-arm64.tar.gz: attempt #1 Feb 13 19:08:59.360052 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Feb 13 19:08:59.437259 systemd-networkd[779]: eth1: Gained IPv6LL Feb 13 19:08:59.573353 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.0-linux-arm64.tar.gz" Feb 13 19:08:59.575348 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Feb 13 19:08:59.575348 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Feb 13 19:08:59.575348 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Feb 13 19:08:59.575348 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Feb 13 19:08:59.575348 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Feb 13 19:08:59.575348 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Feb 13 19:08:59.575348 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Feb 13 19:08:59.575348 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Feb 13 19:08:59.575348 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Feb 13 19:08:59.575348 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Feb 13 19:08:59.575348 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.0-arm64.raw" Feb 13 19:08:59.575348 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.0-arm64.raw" Feb 13 19:08:59.575348 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.0-arm64.raw" Feb 13 19:08:59.575348 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.32.0-arm64.raw: attempt #1 Feb 13 19:09:00.117643 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Feb 13 19:09:00.486896 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.0-arm64.raw" Feb 13 19:09:00.486896 ignition[960]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Feb 13 19:09:00.489730 ignition[960]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Feb 13 19:09:00.492263 ignition[960]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Feb 13 19:09:00.492263 ignition[960]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Feb 13 19:09:00.492263 ignition[960]: INFO : files: op(d): [started] processing unit "coreos-metadata.service" Feb 13 19:09:00.492263 ignition[960]: INFO : files: op(d): op(e): [started] writing systemd drop-in "00-custom-metadata.conf" at "/sysroot/etc/systemd/system/coreos-metadata.service.d/00-custom-metadata.conf" Feb 13 19:09:00.492263 ignition[960]: INFO : files: op(d): op(e): [finished] writing systemd drop-in "00-custom-metadata.conf" at "/sysroot/etc/systemd/system/coreos-metadata.service.d/00-custom-metadata.conf" Feb 13 19:09:00.492263 ignition[960]: INFO : files: op(d): [finished] processing unit "coreos-metadata.service" Feb 13 19:09:00.492263 ignition[960]: INFO : files: op(f): [started] setting preset to enabled for "prepare-helm.service" Feb 13 19:09:00.492263 ignition[960]: INFO : files: op(f): [finished] setting preset to enabled for "prepare-helm.service" Feb 13 19:09:00.492263 ignition[960]: INFO : files: createResultFile: createFiles: op(10): [started] writing file "/sysroot/etc/.ignition-result.json" Feb 13 19:09:00.492263 ignition[960]: INFO : files: createResultFile: createFiles: op(10): [finished] writing file "/sysroot/etc/.ignition-result.json" Feb 13 19:09:00.492263 ignition[960]: INFO : files: files passed Feb 13 19:09:00.492263 ignition[960]: INFO : Ignition finished successfully Feb 13 19:09:00.493583 systemd[1]: Finished ignition-files.service - Ignition (files). Feb 13 19:09:00.504786 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Feb 13 19:09:00.507050 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Feb 13 19:09:00.510087 systemd[1]: ignition-quench.service: Deactivated successfully. Feb 13 19:09:00.510751 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Feb 13 19:09:00.528388 initrd-setup-root-after-ignition[989]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Feb 13 19:09:00.528388 initrd-setup-root-after-ignition[989]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Feb 13 19:09:00.532616 initrd-setup-root-after-ignition[993]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Feb 13 19:09:00.536248 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Feb 13 19:09:00.537865 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Feb 13 19:09:00.544371 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Feb 13 19:09:00.598801 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Feb 13 19:09:00.599031 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Feb 13 19:09:00.601463 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Feb 13 19:09:00.602160 systemd[1]: Reached target initrd.target - Initrd Default Target. Feb 13 19:09:00.603363 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Feb 13 19:09:00.614452 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Feb 13 19:09:00.633479 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Feb 13 19:09:00.641301 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Feb 13 19:09:00.658767 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Feb 13 19:09:00.659839 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Feb 13 19:09:00.661252 systemd[1]: Stopped target timers.target - Timer Units. Feb 13 19:09:00.662537 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Feb 13 19:09:00.662690 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Feb 13 19:09:00.664328 systemd[1]: Stopped target initrd.target - Initrd Default Target. Feb 13 19:09:00.665042 systemd[1]: Stopped target basic.target - Basic System. Feb 13 19:09:00.666159 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Feb 13 19:09:00.667140 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Feb 13 19:09:00.668147 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Feb 13 19:09:00.669196 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Feb 13 19:09:00.670218 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Feb 13 19:09:00.671347 systemd[1]: Stopped target sysinit.target - System Initialization. Feb 13 19:09:00.672339 systemd[1]: Stopped target local-fs.target - Local File Systems. Feb 13 19:09:00.673508 systemd[1]: Stopped target swap.target - Swaps. Feb 13 19:09:00.674377 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Feb 13 19:09:00.674505 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Feb 13 19:09:00.675761 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Feb 13 19:09:00.676443 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Feb 13 19:09:00.677510 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Feb 13 19:09:00.677988 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Feb 13 19:09:00.678719 systemd[1]: dracut-initqueue.service: Deactivated successfully. Feb 13 19:09:00.678904 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Feb 13 19:09:00.681612 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Feb 13 19:09:00.681985 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Feb 13 19:09:00.684554 systemd[1]: ignition-files.service: Deactivated successfully. Feb 13 19:09:00.684784 systemd[1]: Stopped ignition-files.service - Ignition (files). Feb 13 19:09:00.686276 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Feb 13 19:09:00.686488 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Feb 13 19:09:00.695640 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Feb 13 19:09:00.703347 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Feb 13 19:09:00.708073 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Feb 13 19:09:00.708286 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Feb 13 19:09:00.709278 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Feb 13 19:09:00.709390 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Feb 13 19:09:00.718234 systemd[1]: initrd-cleanup.service: Deactivated successfully. Feb 13 19:09:00.718339 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Feb 13 19:09:00.727011 systemd[1]: sysroot-boot.mount: Deactivated successfully. Feb 13 19:09:00.727859 ignition[1013]: INFO : Ignition 2.20.0 Feb 13 19:09:00.727859 ignition[1013]: INFO : Stage: umount Feb 13 19:09:00.732624 ignition[1013]: INFO : no configs at "/usr/lib/ignition/base.d" Feb 13 19:09:00.732624 ignition[1013]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Feb 13 19:09:00.732624 ignition[1013]: INFO : umount: umount passed Feb 13 19:09:00.732624 ignition[1013]: INFO : Ignition finished successfully Feb 13 19:09:00.730662 systemd[1]: ignition-mount.service: Deactivated successfully. Feb 13 19:09:00.730769 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Feb 13 19:09:00.734499 systemd[1]: ignition-disks.service: Deactivated successfully. Feb 13 19:09:00.734634 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Feb 13 19:09:00.736120 systemd[1]: ignition-kargs.service: Deactivated successfully. Feb 13 19:09:00.736202 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Feb 13 19:09:00.737575 systemd[1]: ignition-fetch.service: Deactivated successfully. Feb 13 19:09:00.737674 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Feb 13 19:09:00.739150 systemd[1]: Stopped target network.target - Network. Feb 13 19:09:00.740440 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Feb 13 19:09:00.740514 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Feb 13 19:09:00.741600 systemd[1]: Stopped target paths.target - Path Units. Feb 13 19:09:00.742575 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Feb 13 19:09:00.746167 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Feb 13 19:09:00.746862 systemd[1]: Stopped target slices.target - Slice Units. Feb 13 19:09:00.748434 systemd[1]: Stopped target sockets.target - Socket Units. Feb 13 19:09:00.750005 systemd[1]: iscsid.socket: Deactivated successfully. Feb 13 19:09:00.750076 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Feb 13 19:09:00.751042 systemd[1]: iscsiuio.socket: Deactivated successfully. Feb 13 19:09:00.751130 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Feb 13 19:09:00.752141 systemd[1]: ignition-setup.service: Deactivated successfully. Feb 13 19:09:00.752198 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Feb 13 19:09:00.753317 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Feb 13 19:09:00.753366 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Feb 13 19:09:00.754624 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Feb 13 19:09:00.755491 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Feb 13 19:09:00.757672 systemd[1]: sysroot-boot.service: Deactivated successfully. Feb 13 19:09:00.757856 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Feb 13 19:09:00.760744 systemd[1]: initrd-setup-root.service: Deactivated successfully. Feb 13 19:09:00.760890 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Feb 13 19:09:00.763243 systemd[1]: systemd-resolved.service: Deactivated successfully. Feb 13 19:09:00.763379 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Feb 13 19:09:00.769019 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Feb 13 19:09:00.769509 systemd[1]: systemd-networkd.service: Deactivated successfully. Feb 13 19:09:00.769620 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Feb 13 19:09:00.772295 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Feb 13 19:09:00.773669 systemd[1]: systemd-networkd.socket: Deactivated successfully. Feb 13 19:09:00.773826 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Feb 13 19:09:00.780241 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Feb 13 19:09:00.780898 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Feb 13 19:09:00.780970 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Feb 13 19:09:00.781834 systemd[1]: systemd-sysctl.service: Deactivated successfully. Feb 13 19:09:00.781888 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Feb 13 19:09:00.783338 systemd[1]: systemd-modules-load.service: Deactivated successfully. Feb 13 19:09:00.783394 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Feb 13 19:09:00.785235 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Feb 13 19:09:00.785379 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Feb 13 19:09:00.787900 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Feb 13 19:09:00.792709 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Feb 13 19:09:00.792802 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Feb 13 19:09:00.808321 systemd[1]: systemd-udevd.service: Deactivated successfully. Feb 13 19:09:00.809752 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Feb 13 19:09:00.811598 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Feb 13 19:09:00.811646 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Feb 13 19:09:00.813724 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Feb 13 19:09:00.813759 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Feb 13 19:09:00.814660 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Feb 13 19:09:00.814709 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Feb 13 19:09:00.816187 systemd[1]: dracut-cmdline.service: Deactivated successfully. Feb 13 19:09:00.816239 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Feb 13 19:09:00.817774 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Feb 13 19:09:00.817836 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 19:09:00.831969 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Feb 13 19:09:00.833565 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Feb 13 19:09:00.833693 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Feb 13 19:09:00.836342 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Feb 13 19:09:00.836400 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Feb 13 19:09:00.837046 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Feb 13 19:09:00.837117 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Feb 13 19:09:00.841294 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Feb 13 19:09:00.841360 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 19:09:00.842999 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Feb 13 19:09:00.843078 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Feb 13 19:09:00.843421 systemd[1]: network-cleanup.service: Deactivated successfully. Feb 13 19:09:00.843555 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Feb 13 19:09:00.844327 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Feb 13 19:09:00.844406 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Feb 13 19:09:00.846172 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Feb 13 19:09:00.849287 systemd[1]: Starting initrd-switch-root.service - Switch Root... Feb 13 19:09:00.866901 systemd[1]: Switching root. Feb 13 19:09:00.902795 systemd-journald[237]: Journal stopped Feb 13 19:09:01.949780 systemd-journald[237]: Received SIGTERM from PID 1 (systemd). Feb 13 19:09:01.949865 kernel: SELinux: policy capability network_peer_controls=1 Feb 13 19:09:01.949880 kernel: SELinux: policy capability open_perms=1 Feb 13 19:09:01.949890 kernel: SELinux: policy capability extended_socket_class=1 Feb 13 19:09:01.949899 kernel: SELinux: policy capability always_check_network=0 Feb 13 19:09:01.949909 kernel: SELinux: policy capability cgroup_seclabel=1 Feb 13 19:09:01.949923 kernel: SELinux: policy capability nnp_nosuid_transition=1 Feb 13 19:09:01.949932 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Feb 13 19:09:01.949941 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Feb 13 19:09:01.949950 kernel: audit: type=1403 audit(1739473741.056:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Feb 13 19:09:01.949972 systemd[1]: Successfully loaded SELinux policy in 37.264ms. Feb 13 19:09:01.950000 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 10.774ms. Feb 13 19:09:01.950012 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Feb 13 19:09:01.950022 systemd[1]: Detected virtualization kvm. Feb 13 19:09:01.950034 systemd[1]: Detected architecture arm64. Feb 13 19:09:01.950044 systemd[1]: Detected first boot. Feb 13 19:09:01.954829 systemd[1]: Hostname set to . Feb 13 19:09:01.954877 systemd[1]: Initializing machine ID from VM UUID. Feb 13 19:09:01.954895 zram_generator::config[1059]: No configuration found. Feb 13 19:09:01.954908 kernel: NET: Registered PF_VSOCK protocol family Feb 13 19:09:01.954921 systemd[1]: Populated /etc with preset unit settings. Feb 13 19:09:01.954933 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Feb 13 19:09:01.954952 systemd[1]: initrd-switch-root.service: Deactivated successfully. Feb 13 19:09:01.954963 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Feb 13 19:09:01.954974 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Feb 13 19:09:01.954985 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Feb 13 19:09:01.954996 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Feb 13 19:09:01.955011 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Feb 13 19:09:01.955023 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Feb 13 19:09:01.955039 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Feb 13 19:09:01.955052 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Feb 13 19:09:01.955086 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Feb 13 19:09:01.955097 systemd[1]: Created slice user.slice - User and Session Slice. Feb 13 19:09:01.955108 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Feb 13 19:09:01.955119 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Feb 13 19:09:01.955132 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Feb 13 19:09:01.955143 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Feb 13 19:09:01.955154 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Feb 13 19:09:01.955165 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Feb 13 19:09:01.955178 systemd[1]: Expecting device dev-ttyAMA0.device - /dev/ttyAMA0... Feb 13 19:09:01.955189 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Feb 13 19:09:01.955200 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Feb 13 19:09:01.955211 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Feb 13 19:09:01.955222 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Feb 13 19:09:01.955233 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Feb 13 19:09:01.955245 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Feb 13 19:09:01.955256 systemd[1]: Reached target remote-fs.target - Remote File Systems. Feb 13 19:09:01.955267 systemd[1]: Reached target slices.target - Slice Units. Feb 13 19:09:01.955278 systemd[1]: Reached target swap.target - Swaps. Feb 13 19:09:01.955289 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Feb 13 19:09:01.955300 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Feb 13 19:09:01.955310 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Feb 13 19:09:01.955321 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Feb 13 19:09:01.955335 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Feb 13 19:09:01.955351 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Feb 13 19:09:01.955379 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Feb 13 19:09:01.955392 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Feb 13 19:09:01.955403 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Feb 13 19:09:01.955414 systemd[1]: Mounting media.mount - External Media Directory... Feb 13 19:09:01.955426 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Feb 13 19:09:01.955439 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Feb 13 19:09:01.955450 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Feb 13 19:09:01.955461 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Feb 13 19:09:01.955472 systemd[1]: Reached target machines.target - Containers. Feb 13 19:09:01.955483 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Feb 13 19:09:01.955494 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Feb 13 19:09:01.955505 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Feb 13 19:09:01.955516 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Feb 13 19:09:01.955527 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Feb 13 19:09:01.955540 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Feb 13 19:09:01.955550 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Feb 13 19:09:01.955562 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Feb 13 19:09:01.955572 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Feb 13 19:09:01.955584 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Feb 13 19:09:01.955595 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Feb 13 19:09:01.955606 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Feb 13 19:09:01.955617 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Feb 13 19:09:01.955630 systemd[1]: Stopped systemd-fsck-usr.service. Feb 13 19:09:01.955641 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Feb 13 19:09:01.955652 systemd[1]: Starting systemd-journald.service - Journal Service... Feb 13 19:09:01.955663 kernel: fuse: init (API version 7.39) Feb 13 19:09:01.955674 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Feb 13 19:09:01.955684 kernel: loop: module loaded Feb 13 19:09:01.955702 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Feb 13 19:09:01.955720 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Feb 13 19:09:01.955731 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Feb 13 19:09:01.955747 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Feb 13 19:09:01.955758 systemd[1]: verity-setup.service: Deactivated successfully. Feb 13 19:09:01.955773 systemd[1]: Stopped verity-setup.service. Feb 13 19:09:01.955786 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Feb 13 19:09:01.955809 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Feb 13 19:09:01.955825 systemd[1]: Mounted media.mount - External Media Directory. Feb 13 19:09:01.955837 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Feb 13 19:09:01.955850 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Feb 13 19:09:01.955866 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Feb 13 19:09:01.955878 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Feb 13 19:09:01.955890 systemd[1]: modprobe@configfs.service: Deactivated successfully. Feb 13 19:09:01.955901 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Feb 13 19:09:01.955912 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Feb 13 19:09:01.955925 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Feb 13 19:09:01.955936 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Feb 13 19:09:01.955947 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Feb 13 19:09:01.955958 systemd[1]: modprobe@fuse.service: Deactivated successfully. Feb 13 19:09:01.955969 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Feb 13 19:09:01.955980 systemd[1]: modprobe@loop.service: Deactivated successfully. Feb 13 19:09:01.955993 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Feb 13 19:09:01.956004 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Feb 13 19:09:01.956015 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Feb 13 19:09:01.956029 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Feb 13 19:09:01.956040 systemd[1]: Reached target network-pre.target - Preparation for Network. Feb 13 19:09:01.956051 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Feb 13 19:09:01.963077 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Feb 13 19:09:01.963109 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Feb 13 19:09:01.963121 systemd[1]: Reached target local-fs.target - Local File Systems. Feb 13 19:09:01.963132 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Feb 13 19:09:01.963142 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Feb 13 19:09:01.963160 kernel: ACPI: bus type drm_connector registered Feb 13 19:09:01.963208 systemd-journald[1127]: Collecting audit messages is disabled. Feb 13 19:09:01.963245 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Feb 13 19:09:01.963263 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Feb 13 19:09:01.963277 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Feb 13 19:09:01.963289 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Feb 13 19:09:01.963299 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Feb 13 19:09:01.963312 systemd-journald[1127]: Journal started Feb 13 19:09:01.963335 systemd-journald[1127]: Runtime Journal (/run/log/journal/7ecaf9133ce94e0d8dfa2faad6c7d0cd) is 8M, max 76.6M, 68.6M free. Feb 13 19:09:01.636699 systemd[1]: Queued start job for default target multi-user.target. Feb 13 19:09:01.647493 systemd[1]: Unnecessary job was removed for dev-sda6.device - /dev/sda6. Feb 13 19:09:01.648029 systemd[1]: systemd-journald.service: Deactivated successfully. Feb 13 19:09:01.968115 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Feb 13 19:09:01.988094 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Feb 13 19:09:01.988178 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Feb 13 19:09:01.988196 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Feb 13 19:09:02.000122 systemd[1]: Started systemd-journald.service - Journal Service. Feb 13 19:09:01.997089 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Feb 13 19:09:01.998316 systemd[1]: modprobe@drm.service: Deactivated successfully. Feb 13 19:09:01.998480 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Feb 13 19:09:02.001857 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Feb 13 19:09:02.004325 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Feb 13 19:09:02.006966 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Feb 13 19:09:02.010361 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Feb 13 19:09:02.029518 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Feb 13 19:09:02.040086 kernel: loop0: detected capacity change from 0 to 123192 Feb 13 19:09:02.051163 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Feb 13 19:09:02.062938 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Feb 13 19:09:02.065514 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Feb 13 19:09:02.076383 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Feb 13 19:09:02.083571 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Feb 13 19:09:02.085436 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Feb 13 19:09:02.091922 kernel: loop1: detected capacity change from 0 to 201592 Feb 13 19:09:02.091168 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Feb 13 19:09:02.097195 systemd-journald[1127]: Time spent on flushing to /var/log/journal/7ecaf9133ce94e0d8dfa2faad6c7d0cd is 53.690ms for 1150 entries. Feb 13 19:09:02.097195 systemd-journald[1127]: System Journal (/var/log/journal/7ecaf9133ce94e0d8dfa2faad6c7d0cd) is 8M, max 584.8M, 576.8M free. Feb 13 19:09:02.164895 systemd-journald[1127]: Received client request to flush runtime journal. Feb 13 19:09:02.165005 kernel: loop2: detected capacity change from 0 to 113512 Feb 13 19:09:02.099695 systemd-tmpfiles[1161]: ACLs are not supported, ignoring. Feb 13 19:09:02.099705 systemd-tmpfiles[1161]: ACLs are not supported, ignoring. Feb 13 19:09:02.113040 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Feb 13 19:09:02.126408 systemd[1]: Starting systemd-sysusers.service - Create System Users... Feb 13 19:09:02.143517 udevadm[1192]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Feb 13 19:09:02.169984 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Feb 13 19:09:02.180229 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Feb 13 19:09:02.207546 systemd[1]: Finished systemd-sysusers.service - Create System Users. Feb 13 19:09:02.208265 kernel: loop3: detected capacity change from 0 to 8 Feb 13 19:09:02.218547 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Feb 13 19:09:02.233086 kernel: loop4: detected capacity change from 0 to 123192 Feb 13 19:09:02.233457 systemd-tmpfiles[1206]: ACLs are not supported, ignoring. Feb 13 19:09:02.233476 systemd-tmpfiles[1206]: ACLs are not supported, ignoring. Feb 13 19:09:02.239116 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Feb 13 19:09:02.259103 kernel: loop5: detected capacity change from 0 to 201592 Feb 13 19:09:02.287146 kernel: loop6: detected capacity change from 0 to 113512 Feb 13 19:09:02.311089 kernel: loop7: detected capacity change from 0 to 8 Feb 13 19:09:02.312856 (sd-merge)[1208]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-hetzner'. Feb 13 19:09:02.314007 (sd-merge)[1208]: Merged extensions into '/usr'. Feb 13 19:09:02.321113 systemd[1]: Reload requested from client PID 1160 ('systemd-sysext') (unit systemd-sysext.service)... Feb 13 19:09:02.321289 systemd[1]: Reloading... Feb 13 19:09:02.476496 zram_generator::config[1237]: No configuration found. Feb 13 19:09:02.549271 ldconfig[1156]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Feb 13 19:09:02.624593 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 19:09:02.689409 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Feb 13 19:09:02.689604 systemd[1]: Reloading finished in 367 ms. Feb 13 19:09:02.707426 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Feb 13 19:09:02.708496 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Feb 13 19:09:02.725482 systemd[1]: Starting ensure-sysext.service... Feb 13 19:09:02.729320 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Feb 13 19:09:02.750398 systemd[1]: Reload requested from client PID 1274 ('systemctl') (unit ensure-sysext.service)... Feb 13 19:09:02.750420 systemd[1]: Reloading... Feb 13 19:09:02.769235 systemd-tmpfiles[1275]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Feb 13 19:09:02.769775 systemd-tmpfiles[1275]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Feb 13 19:09:02.770506 systemd-tmpfiles[1275]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Feb 13 19:09:02.770703 systemd-tmpfiles[1275]: ACLs are not supported, ignoring. Feb 13 19:09:02.770749 systemd-tmpfiles[1275]: ACLs are not supported, ignoring. Feb 13 19:09:02.775316 systemd-tmpfiles[1275]: Detected autofs mount point /boot during canonicalization of boot. Feb 13 19:09:02.775332 systemd-tmpfiles[1275]: Skipping /boot Feb 13 19:09:02.788608 systemd-tmpfiles[1275]: Detected autofs mount point /boot during canonicalization of boot. Feb 13 19:09:02.788630 systemd-tmpfiles[1275]: Skipping /boot Feb 13 19:09:02.827147 zram_generator::config[1300]: No configuration found. Feb 13 19:09:02.946411 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 19:09:03.010178 systemd[1]: Reloading finished in 259 ms. Feb 13 19:09:03.024778 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Feb 13 19:09:03.039404 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Feb 13 19:09:03.052507 systemd[1]: Starting audit-rules.service - Load Audit Rules... Feb 13 19:09:03.057417 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Feb 13 19:09:03.063199 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Feb 13 19:09:03.075179 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Feb 13 19:09:03.080466 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Feb 13 19:09:03.088198 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Feb 13 19:09:03.094546 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Feb 13 19:09:03.100263 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Feb 13 19:09:03.104523 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Feb 13 19:09:03.110386 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Feb 13 19:09:03.111248 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Feb 13 19:09:03.111375 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Feb 13 19:09:03.114529 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Feb 13 19:09:03.124308 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Feb 13 19:09:03.124476 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Feb 13 19:09:03.124634 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Feb 13 19:09:03.128311 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Feb 13 19:09:03.131742 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Feb 13 19:09:03.137393 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Feb 13 19:09:03.138137 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Feb 13 19:09:03.138262 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Feb 13 19:09:03.141190 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Feb 13 19:09:03.142762 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Feb 13 19:09:03.142973 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Feb 13 19:09:03.148650 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Feb 13 19:09:03.150280 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Feb 13 19:09:03.159265 systemd[1]: Finished ensure-sysext.service. Feb 13 19:09:03.166946 systemd[1]: modprobe@loop.service: Deactivated successfully. Feb 13 19:09:03.167461 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Feb 13 19:09:03.170343 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Feb 13 19:09:03.170446 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Feb 13 19:09:03.177125 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Feb 13 19:09:03.183346 systemd[1]: Starting systemd-update-done.service - Update is Completed... Feb 13 19:09:03.188638 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Feb 13 19:09:03.191274 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Feb 13 19:09:03.191675 systemd[1]: modprobe@drm.service: Deactivated successfully. Feb 13 19:09:03.191863 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Feb 13 19:09:03.196890 systemd-udevd[1353]: Using default interface naming scheme 'v255'. Feb 13 19:09:03.215694 augenrules[1383]: No rules Feb 13 19:09:03.217539 systemd[1]: audit-rules.service: Deactivated successfully. Feb 13 19:09:03.217745 systemd[1]: Finished audit-rules.service - Load Audit Rules. Feb 13 19:09:03.222490 systemd[1]: Finished systemd-update-done.service - Update is Completed. Feb 13 19:09:03.231538 systemd[1]: Started systemd-userdbd.service - User Database Manager. Feb 13 19:09:03.238551 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Feb 13 19:09:03.248762 systemd[1]: Starting systemd-networkd.service - Network Configuration... Feb 13 19:09:03.389604 systemd-resolved[1352]: Positive Trust Anchors: Feb 13 19:09:03.390538 systemd-resolved[1352]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Feb 13 19:09:03.390669 systemd-resolved[1352]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Feb 13 19:09:03.396602 systemd-resolved[1352]: Using system hostname 'ci-4230-0-1-a-2d796e8b97'. Feb 13 19:09:03.398572 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Feb 13 19:09:03.400305 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Feb 13 19:09:03.413011 systemd[1]: Condition check resulted in dev-ttyAMA0.device - /dev/ttyAMA0 being skipped. Feb 13 19:09:03.427524 systemd-networkd[1399]: lo: Link UP Feb 13 19:09:03.428361 systemd-networkd[1399]: lo: Gained carrier Feb 13 19:09:03.431903 systemd-networkd[1399]: Enumeration completed Feb 13 19:09:03.432372 systemd[1]: Started systemd-networkd.service - Network Configuration. Feb 13 19:09:03.433389 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Feb 13 19:09:03.433607 systemd-networkd[1399]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 19:09:03.433765 systemd-networkd[1399]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Feb 13 19:09:03.435130 systemd[1]: Reached target network.target - Network. Feb 13 19:09:03.435621 systemd[1]: Reached target time-set.target - System Time Set. Feb 13 19:09:03.435936 systemd-networkd[1399]: eth1: Link UP Feb 13 19:09:03.436012 systemd-networkd[1399]: eth1: Gained carrier Feb 13 19:09:03.436097 systemd-networkd[1399]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 19:09:03.443514 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Feb 13 19:09:03.453681 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Feb 13 19:09:03.466302 systemd-networkd[1399]: eth1: DHCPv4 address 10.0.0.3/32, gateway 10.0.0.1 acquired from 10.0.0.1 Feb 13 19:09:03.468130 systemd-networkd[1399]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 19:09:03.468141 systemd-networkd[1399]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Feb 13 19:09:03.468236 systemd-timesyncd[1376]: Network configuration changed, trying to establish connection. Feb 13 19:09:03.471651 systemd-networkd[1399]: eth0: Link UP Feb 13 19:09:03.471660 systemd-networkd[1399]: eth0: Gained carrier Feb 13 19:09:03.471682 systemd-networkd[1399]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 19:09:03.480543 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Feb 13 19:09:03.502169 kernel: mousedev: PS/2 mouse device common for all mice Feb 13 19:09:03.534177 systemd-networkd[1399]: eth0: DHCPv4 address 78.46.147.231/32, gateway 172.31.1.1 acquired from 172.31.1.1 Feb 13 19:09:03.534564 systemd-timesyncd[1376]: Network configuration changed, trying to establish connection. Feb 13 19:09:03.536712 systemd-timesyncd[1376]: Network configuration changed, trying to establish connection. Feb 13 19:09:03.585035 systemd[1]: Condition check resulted in dev-virtio\x2dports-org.qemu.guest_agent.0.device - /dev/virtio-ports/org.qemu.guest_agent.0 being skipped. Feb 13 19:09:03.586158 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Feb 13 19:09:03.595080 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 38 scanned by (udev-worker) (1411) Feb 13 19:09:03.595493 kernel: [drm] pci: virtio-gpu-pci detected at 0000:00:01.0 Feb 13 19:09:03.602073 kernel: [drm] features: -virgl +edid -resource_blob -host_visible Feb 13 19:09:03.602180 kernel: [drm] features: -context_init Feb 13 19:09:03.612643 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Feb 13 19:09:03.618106 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Feb 13 19:09:03.621723 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Feb 13 19:09:03.623187 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Feb 13 19:09:03.623230 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Feb 13 19:09:03.623252 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Feb 13 19:09:03.623647 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Feb 13 19:09:03.625149 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Feb 13 19:09:03.626239 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Feb 13 19:09:03.626460 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Feb 13 19:09:03.639737 systemd[1]: modprobe@loop.service: Deactivated successfully. Feb 13 19:09:03.640887 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Feb 13 19:09:03.674330 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Feb 13 19:09:03.674709 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Feb 13 19:09:03.682750 kernel: [drm] number of scanouts: 1 Feb 13 19:09:03.682829 kernel: [drm] number of cap sets: 0 Feb 13 19:09:03.685995 kernel: [drm] Initialized virtio_gpu 0.1.0 0 for 0000:00:01.0 on minor 0 Feb 13 19:09:03.685414 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 19:09:03.691228 kernel: Console: switching to colour frame buffer device 160x50 Feb 13 19:09:03.692839 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - QEMU_HARDDISK OEM. Feb 13 19:09:03.698522 kernel: virtio-pci 0000:00:01.0: [drm] fb0: virtio_gpudrmfb frame buffer device Feb 13 19:09:03.712417 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Feb 13 19:09:03.717225 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Feb 13 19:09:03.717546 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 19:09:03.719671 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Feb 13 19:09:03.728698 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 19:09:03.731223 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Feb 13 19:09:03.790196 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 19:09:03.830881 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Feb 13 19:09:03.838394 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Feb 13 19:09:03.853698 lvm[1465]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Feb 13 19:09:03.880865 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Feb 13 19:09:03.883267 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Feb 13 19:09:03.884561 systemd[1]: Reached target sysinit.target - System Initialization. Feb 13 19:09:03.885883 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Feb 13 19:09:03.887309 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Feb 13 19:09:03.888298 systemd[1]: Started logrotate.timer - Daily rotation of log files. Feb 13 19:09:03.888964 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Feb 13 19:09:03.889669 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Feb 13 19:09:03.890308 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Feb 13 19:09:03.890345 systemd[1]: Reached target paths.target - Path Units. Feb 13 19:09:03.890808 systemd[1]: Reached target timers.target - Timer Units. Feb 13 19:09:03.892183 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Feb 13 19:09:03.894557 systemd[1]: Starting docker.socket - Docker Socket for the API... Feb 13 19:09:03.897955 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Feb 13 19:09:03.899088 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Feb 13 19:09:03.899834 systemd[1]: Reached target ssh-access.target - SSH Access Available. Feb 13 19:09:03.902907 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Feb 13 19:09:03.904926 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Feb 13 19:09:03.912362 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Feb 13 19:09:03.914417 systemd[1]: Listening on docker.socket - Docker Socket for the API. Feb 13 19:09:03.915845 systemd[1]: Reached target sockets.target - Socket Units. Feb 13 19:09:03.916648 systemd[1]: Reached target basic.target - Basic System. Feb 13 19:09:03.917593 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Feb 13 19:09:03.917637 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Feb 13 19:09:03.926380 systemd[1]: Starting containerd.service - containerd container runtime... Feb 13 19:09:03.929721 lvm[1469]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Feb 13 19:09:03.936370 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Feb 13 19:09:03.939679 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Feb 13 19:09:03.944930 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Feb 13 19:09:03.948346 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Feb 13 19:09:03.951165 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Feb 13 19:09:03.955840 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Feb 13 19:09:03.958433 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Feb 13 19:09:03.964517 systemd[1]: Started qemu-guest-agent.service - QEMU Guest Agent. Feb 13 19:09:03.969940 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Feb 13 19:09:03.980249 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Feb 13 19:09:03.982195 dbus-daemon[1472]: [system] SELinux support is enabled Feb 13 19:09:03.985289 systemd[1]: Starting systemd-logind.service - User Login Management... Feb 13 19:09:03.988134 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Feb 13 19:09:03.988693 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Feb 13 19:09:03.990578 systemd[1]: Starting update-engine.service - Update Engine... Feb 13 19:09:04.001081 jq[1473]: false Feb 13 19:09:04.002200 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Feb 13 19:09:04.003426 systemd[1]: Started dbus.service - D-Bus System Message Bus. Feb 13 19:09:04.009376 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Feb 13 19:09:04.020596 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Feb 13 19:09:04.021743 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Feb 13 19:09:04.028168 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Feb 13 19:09:04.028214 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Feb 13 19:09:04.028995 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Feb 13 19:09:04.029022 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Feb 13 19:09:04.039253 jq[1487]: true Feb 13 19:09:04.046560 update_engine[1486]: I20250213 19:09:04.046408 1486 main.cc:92] Flatcar Update Engine starting Feb 13 19:09:04.050402 coreos-metadata[1471]: Feb 13 19:09:04.050 INFO Fetching http://169.254.169.254/hetzner/v1/metadata: Attempt #1 Feb 13 19:09:04.057599 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Feb 13 19:09:04.057894 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Feb 13 19:09:04.065355 update_engine[1486]: I20250213 19:09:04.064618 1486 update_check_scheduler.cc:74] Next update check in 4m14s Feb 13 19:09:04.066764 systemd[1]: Started update-engine.service - Update Engine. Feb 13 19:09:04.070427 extend-filesystems[1474]: Found loop4 Feb 13 19:09:04.070427 extend-filesystems[1474]: Found loop5 Feb 13 19:09:04.070427 extend-filesystems[1474]: Found loop6 Feb 13 19:09:04.070427 extend-filesystems[1474]: Found loop7 Feb 13 19:09:04.070427 extend-filesystems[1474]: Found sda Feb 13 19:09:04.070427 extend-filesystems[1474]: Found sda1 Feb 13 19:09:04.070427 extend-filesystems[1474]: Found sda2 Feb 13 19:09:04.070427 extend-filesystems[1474]: Found sda3 Feb 13 19:09:04.070427 extend-filesystems[1474]: Found usr Feb 13 19:09:04.070427 extend-filesystems[1474]: Found sda4 Feb 13 19:09:04.070427 extend-filesystems[1474]: Found sda6 Feb 13 19:09:04.070427 extend-filesystems[1474]: Found sda7 Feb 13 19:09:04.070427 extend-filesystems[1474]: Found sda9 Feb 13 19:09:04.070427 extend-filesystems[1474]: Checking size of /dev/sda9 Feb 13 19:09:04.113232 coreos-metadata[1471]: Feb 13 19:09:04.067 INFO Fetch successful Feb 13 19:09:04.113232 coreos-metadata[1471]: Feb 13 19:09:04.067 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/private-networks: Attempt #1 Feb 13 19:09:04.113232 coreos-metadata[1471]: Feb 13 19:09:04.083 INFO Fetch successful Feb 13 19:09:04.113439 tar[1490]: linux-arm64/LICENSE Feb 13 19:09:04.113439 tar[1490]: linux-arm64/helm Feb 13 19:09:04.077301 systemd[1]: Started locksmithd.service - Cluster reboot manager. Feb 13 19:09:04.113780 jq[1500]: true Feb 13 19:09:04.078333 systemd[1]: motdgen.service: Deactivated successfully. Feb 13 19:09:04.078541 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Feb 13 19:09:04.092383 (ntainerd)[1503]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Feb 13 19:09:04.153349 extend-filesystems[1474]: Resized partition /dev/sda9 Feb 13 19:09:04.166837 extend-filesystems[1530]: resize2fs 1.47.1 (20-May-2024) Feb 13 19:09:04.196028 kernel: EXT4-fs (sda9): resizing filesystem from 1617920 to 9393147 blocks Feb 13 19:09:04.216117 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Feb 13 19:09:04.218033 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Feb 13 19:09:04.244252 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 38 scanned by (udev-worker) (1417) Feb 13 19:09:04.308130 bash[1547]: Updated "/home/core/.ssh/authorized_keys" Feb 13 19:09:04.312143 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Feb 13 19:09:04.323256 systemd[1]: Starting sshkeys.service... Feb 13 19:09:04.341381 locksmithd[1508]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Feb 13 19:09:04.351173 systemd-logind[1485]: New seat seat0. Feb 13 19:09:04.352332 systemd-logind[1485]: Watching system buttons on /dev/input/event0 (Power Button) Feb 13 19:09:04.352355 systemd-logind[1485]: Watching system buttons on /dev/input/event2 (QEMU QEMU USB Keyboard) Feb 13 19:09:04.352769 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Feb 13 19:09:04.361357 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Feb 13 19:09:04.363639 systemd[1]: Started systemd-logind.service - User Login Management. Feb 13 19:09:04.381076 kernel: EXT4-fs (sda9): resized filesystem to 9393147 Feb 13 19:09:04.404334 extend-filesystems[1530]: Filesystem at /dev/sda9 is mounted on /; on-line resizing required Feb 13 19:09:04.404334 extend-filesystems[1530]: old_desc_blocks = 1, new_desc_blocks = 5 Feb 13 19:09:04.404334 extend-filesystems[1530]: The filesystem on /dev/sda9 is now 9393147 (4k) blocks long. Feb 13 19:09:04.413980 extend-filesystems[1474]: Resized filesystem in /dev/sda9 Feb 13 19:09:04.413980 extend-filesystems[1474]: Found sr0 Feb 13 19:09:04.407461 systemd[1]: extend-filesystems.service: Deactivated successfully. Feb 13 19:09:04.407655 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Feb 13 19:09:04.427248 coreos-metadata[1554]: Feb 13 19:09:04.426 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/public-keys: Attempt #1 Feb 13 19:09:04.428972 coreos-metadata[1554]: Feb 13 19:09:04.428 INFO Fetch successful Feb 13 19:09:04.432231 unknown[1554]: wrote ssh authorized keys file for user: core Feb 13 19:09:04.449132 containerd[1503]: time="2025-02-13T19:09:04.445622120Z" level=info msg="starting containerd" revision=9b2ad7760328148397346d10c7b2004271249db4 version=v1.7.23 Feb 13 19:09:04.472253 update-ssh-keys[1561]: Updated "/home/core/.ssh/authorized_keys" Feb 13 19:09:04.469688 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Feb 13 19:09:04.477104 systemd[1]: Finished sshkeys.service. Feb 13 19:09:04.537370 containerd[1503]: time="2025-02-13T19:09:04.531031640Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Feb 13 19:09:04.539993 containerd[1503]: time="2025-02-13T19:09:04.538641400Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.74-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Feb 13 19:09:04.539993 containerd[1503]: time="2025-02-13T19:09:04.538681200Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Feb 13 19:09:04.539993 containerd[1503]: time="2025-02-13T19:09:04.538698280Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Feb 13 19:09:04.539993 containerd[1503]: time="2025-02-13T19:09:04.538902440Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Feb 13 19:09:04.539993 containerd[1503]: time="2025-02-13T19:09:04.538923280Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Feb 13 19:09:04.539993 containerd[1503]: time="2025-02-13T19:09:04.539014240Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Feb 13 19:09:04.539993 containerd[1503]: time="2025-02-13T19:09:04.539029640Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Feb 13 19:09:04.539993 containerd[1503]: time="2025-02-13T19:09:04.539245760Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Feb 13 19:09:04.539993 containerd[1503]: time="2025-02-13T19:09:04.539260560Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Feb 13 19:09:04.539993 containerd[1503]: time="2025-02-13T19:09:04.539272920Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Feb 13 19:09:04.539993 containerd[1503]: time="2025-02-13T19:09:04.539283600Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Feb 13 19:09:04.540287 containerd[1503]: time="2025-02-13T19:09:04.539349880Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Feb 13 19:09:04.540287 containerd[1503]: time="2025-02-13T19:09:04.539541840Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Feb 13 19:09:04.540287 containerd[1503]: time="2025-02-13T19:09:04.539672800Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Feb 13 19:09:04.540287 containerd[1503]: time="2025-02-13T19:09:04.539686400Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Feb 13 19:09:04.540287 containerd[1503]: time="2025-02-13T19:09:04.539764600Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Feb 13 19:09:04.540287 containerd[1503]: time="2025-02-13T19:09:04.539823680Z" level=info msg="metadata content store policy set" policy=shared Feb 13 19:09:04.550471 containerd[1503]: time="2025-02-13T19:09:04.550403160Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Feb 13 19:09:04.550618 containerd[1503]: time="2025-02-13T19:09:04.550602040Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Feb 13 19:09:04.550713 containerd[1503]: time="2025-02-13T19:09:04.550699600Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Feb 13 19:09:04.551154 containerd[1503]: time="2025-02-13T19:09:04.550769600Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Feb 13 19:09:04.551154 containerd[1503]: time="2025-02-13T19:09:04.550826560Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Feb 13 19:09:04.551154 containerd[1503]: time="2025-02-13T19:09:04.551007160Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Feb 13 19:09:04.552361 containerd[1503]: time="2025-02-13T19:09:04.552337880Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Feb 13 19:09:04.552565 containerd[1503]: time="2025-02-13T19:09:04.552544680Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Feb 13 19:09:04.555064 containerd[1503]: time="2025-02-13T19:09:04.553097080Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Feb 13 19:09:04.555064 containerd[1503]: time="2025-02-13T19:09:04.553132160Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Feb 13 19:09:04.555064 containerd[1503]: time="2025-02-13T19:09:04.553148840Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Feb 13 19:09:04.555064 containerd[1503]: time="2025-02-13T19:09:04.553161440Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Feb 13 19:09:04.555064 containerd[1503]: time="2025-02-13T19:09:04.553174200Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Feb 13 19:09:04.555064 containerd[1503]: time="2025-02-13T19:09:04.553188600Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Feb 13 19:09:04.555064 containerd[1503]: time="2025-02-13T19:09:04.553203080Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Feb 13 19:09:04.555064 containerd[1503]: time="2025-02-13T19:09:04.553215160Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Feb 13 19:09:04.555064 containerd[1503]: time="2025-02-13T19:09:04.553227280Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Feb 13 19:09:04.555064 containerd[1503]: time="2025-02-13T19:09:04.553239600Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Feb 13 19:09:04.555064 containerd[1503]: time="2025-02-13T19:09:04.553262160Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Feb 13 19:09:04.555064 containerd[1503]: time="2025-02-13T19:09:04.553277800Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Feb 13 19:09:04.555064 containerd[1503]: time="2025-02-13T19:09:04.553289600Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Feb 13 19:09:04.555064 containerd[1503]: time="2025-02-13T19:09:04.553302600Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Feb 13 19:09:04.555353 containerd[1503]: time="2025-02-13T19:09:04.553314320Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Feb 13 19:09:04.555353 containerd[1503]: time="2025-02-13T19:09:04.553328640Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Feb 13 19:09:04.555353 containerd[1503]: time="2025-02-13T19:09:04.553340960Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Feb 13 19:09:04.555353 containerd[1503]: time="2025-02-13T19:09:04.553353920Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Feb 13 19:09:04.555353 containerd[1503]: time="2025-02-13T19:09:04.553376840Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Feb 13 19:09:04.555353 containerd[1503]: time="2025-02-13T19:09:04.553391200Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Feb 13 19:09:04.555353 containerd[1503]: time="2025-02-13T19:09:04.553401600Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Feb 13 19:09:04.555353 containerd[1503]: time="2025-02-13T19:09:04.553413800Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Feb 13 19:09:04.555353 containerd[1503]: time="2025-02-13T19:09:04.553426080Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Feb 13 19:09:04.555353 containerd[1503]: time="2025-02-13T19:09:04.553448160Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Feb 13 19:09:04.555353 containerd[1503]: time="2025-02-13T19:09:04.553469680Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Feb 13 19:09:04.555353 containerd[1503]: time="2025-02-13T19:09:04.553484160Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Feb 13 19:09:04.555353 containerd[1503]: time="2025-02-13T19:09:04.553494640Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Feb 13 19:09:04.555645 containerd[1503]: time="2025-02-13T19:09:04.555623160Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Feb 13 19:09:04.555818 containerd[1503]: time="2025-02-13T19:09:04.555767240Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Feb 13 19:09:04.557210 containerd[1503]: time="2025-02-13T19:09:04.555867760Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Feb 13 19:09:04.557210 containerd[1503]: time="2025-02-13T19:09:04.555893360Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Feb 13 19:09:04.557210 containerd[1503]: time="2025-02-13T19:09:04.555903200Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Feb 13 19:09:04.557210 containerd[1503]: time="2025-02-13T19:09:04.555924760Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Feb 13 19:09:04.557210 containerd[1503]: time="2025-02-13T19:09:04.555935120Z" level=info msg="NRI interface is disabled by configuration." Feb 13 19:09:04.557210 containerd[1503]: time="2025-02-13T19:09:04.555946680Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Feb 13 19:09:04.557399 containerd[1503]: time="2025-02-13T19:09:04.556322120Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Feb 13 19:09:04.557399 containerd[1503]: time="2025-02-13T19:09:04.556377240Z" level=info msg="Connect containerd service" Feb 13 19:09:04.557399 containerd[1503]: time="2025-02-13T19:09:04.556415320Z" level=info msg="using legacy CRI server" Feb 13 19:09:04.557399 containerd[1503]: time="2025-02-13T19:09:04.556422240Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Feb 13 19:09:04.557399 containerd[1503]: time="2025-02-13T19:09:04.556683920Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Feb 13 19:09:04.562874 containerd[1503]: time="2025-02-13T19:09:04.559418680Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Feb 13 19:09:04.562874 containerd[1503]: time="2025-02-13T19:09:04.559993800Z" level=info msg="Start subscribing containerd event" Feb 13 19:09:04.562874 containerd[1503]: time="2025-02-13T19:09:04.560492000Z" level=info msg="Start recovering state" Feb 13 19:09:04.562874 containerd[1503]: time="2025-02-13T19:09:04.560610720Z" level=info msg="Start event monitor" Feb 13 19:09:04.562874 containerd[1503]: time="2025-02-13T19:09:04.560628040Z" level=info msg="Start snapshots syncer" Feb 13 19:09:04.562874 containerd[1503]: time="2025-02-13T19:09:04.560637520Z" level=info msg="Start cni network conf syncer for default" Feb 13 19:09:04.562874 containerd[1503]: time="2025-02-13T19:09:04.560646760Z" level=info msg="Start streaming server" Feb 13 19:09:04.565726 containerd[1503]: time="2025-02-13T19:09:04.565683200Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Feb 13 19:09:04.566377 containerd[1503]: time="2025-02-13T19:09:04.566355240Z" level=info msg=serving... address=/run/containerd/containerd.sock Feb 13 19:09:04.566680 containerd[1503]: time="2025-02-13T19:09:04.566663480Z" level=info msg="containerd successfully booted in 0.125219s" Feb 13 19:09:04.566793 systemd[1]: Started containerd.service - containerd container runtime. Feb 13 19:09:04.816668 tar[1490]: linux-arm64/README.md Feb 13 19:09:04.829116 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Feb 13 19:09:04.877308 systemd-networkd[1399]: eth0: Gained IPv6LL Feb 13 19:09:04.878519 systemd-timesyncd[1376]: Network configuration changed, trying to establish connection. Feb 13 19:09:04.884117 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Feb 13 19:09:04.885428 systemd[1]: Reached target network-online.target - Network is Online. Feb 13 19:09:04.895448 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:09:04.900382 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Feb 13 19:09:04.941900 systemd-networkd[1399]: eth1: Gained IPv6LL Feb 13 19:09:04.942674 systemd-timesyncd[1376]: Network configuration changed, trying to establish connection. Feb 13 19:09:04.950508 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Feb 13 19:09:05.072727 sshd_keygen[1509]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Feb 13 19:09:05.100276 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Feb 13 19:09:05.110381 systemd[1]: Starting issuegen.service - Generate /run/issue... Feb 13 19:09:05.118716 systemd[1]: issuegen.service: Deactivated successfully. Feb 13 19:09:05.120120 systemd[1]: Finished issuegen.service - Generate /run/issue. Feb 13 19:09:05.130538 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Feb 13 19:09:05.143371 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Feb 13 19:09:05.152929 systemd[1]: Started getty@tty1.service - Getty on tty1. Feb 13 19:09:05.160557 systemd[1]: Started serial-getty@ttyAMA0.service - Serial Getty on ttyAMA0. Feb 13 19:09:05.162347 systemd[1]: Reached target getty.target - Login Prompts. Feb 13 19:09:05.687980 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:09:05.693799 systemd[1]: Reached target multi-user.target - Multi-User System. Feb 13 19:09:05.695197 systemd[1]: Startup finished in 810ms (kernel) + 6.369s (initrd) + 4.676s (userspace) = 11.857s. Feb 13 19:09:05.704607 (kubelet)[1602]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 19:09:06.218079 kubelet[1602]: E0213 19:09:06.217934 1602 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 19:09:06.222331 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 19:09:06.222503 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 19:09:06.223085 systemd[1]: kubelet.service: Consumed 842ms CPU time, 245.7M memory peak. Feb 13 19:09:16.472872 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Feb 13 19:09:16.481448 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:09:16.604129 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:09:16.609433 (kubelet)[1622]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 19:09:16.656756 kubelet[1622]: E0213 19:09:16.656685 1622 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 19:09:16.660486 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 19:09:16.660665 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 19:09:16.661569 systemd[1]: kubelet.service: Consumed 154ms CPU time, 102.8M memory peak. Feb 13 19:09:26.911732 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Feb 13 19:09:26.919424 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:09:27.051397 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:09:27.051897 (kubelet)[1637]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 19:09:27.106345 kubelet[1637]: E0213 19:09:27.106242 1637 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 19:09:27.108294 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 19:09:27.108685 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 19:09:27.109103 systemd[1]: kubelet.service: Consumed 150ms CPU time, 104.3M memory peak. Feb 13 19:09:34.127486 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Feb 13 19:09:34.138665 systemd[1]: Started sshd@0-78.46.147.231:22-117.247.111.70:32216.service - OpenSSH per-connection server daemon (117.247.111.70:32216). Feb 13 19:09:35.127427 sshd[1646]: Invalid user appserv from 117.247.111.70 port 32216 Feb 13 19:09:35.232218 systemd-timesyncd[1376]: Contacted time server 129.70.132.36:123 (2.flatcar.pool.ntp.org). Feb 13 19:09:35.232301 systemd-timesyncd[1376]: Initial clock synchronization to Thu 2025-02-13 19:09:35.412644 UTC. Feb 13 19:09:35.316277 sshd[1646]: Received disconnect from 117.247.111.70 port 32216:11: Bye Bye [preauth] Feb 13 19:09:35.316277 sshd[1646]: Disconnected from invalid user appserv 117.247.111.70 port 32216 [preauth] Feb 13 19:09:35.319954 systemd[1]: sshd@0-78.46.147.231:22-117.247.111.70:32216.service: Deactivated successfully. Feb 13 19:09:37.264972 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Feb 13 19:09:37.270332 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:09:37.391709 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:09:37.397200 (kubelet)[1658]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 19:09:37.453948 kubelet[1658]: E0213 19:09:37.453858 1658 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 19:09:37.457354 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 19:09:37.457680 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 19:09:37.458356 systemd[1]: kubelet.service: Consumed 160ms CPU time, 104.7M memory peak. Feb 13 19:09:39.708849 systemd[1]: Started sshd@1-78.46.147.231:22-103.154.87.242:39020.service - OpenSSH per-connection server daemon (103.154.87.242:39020). Feb 13 19:09:40.799566 sshd[1666]: Invalid user manchini from 103.154.87.242 port 39020 Feb 13 19:09:41.000140 sshd[1666]: Received disconnect from 103.154.87.242 port 39020:11: Bye Bye [preauth] Feb 13 19:09:41.000140 sshd[1666]: Disconnected from invalid user manchini 103.154.87.242 port 39020 [preauth] Feb 13 19:09:41.002802 systemd[1]: sshd@1-78.46.147.231:22-103.154.87.242:39020.service: Deactivated successfully. Feb 13 19:09:47.514319 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 4. Feb 13 19:09:47.522432 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:09:47.644735 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:09:47.649909 (kubelet)[1678]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 19:09:47.695320 kubelet[1678]: E0213 19:09:47.695255 1678 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 19:09:47.697534 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 19:09:47.697696 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 19:09:47.698229 systemd[1]: kubelet.service: Consumed 152ms CPU time, 100.2M memory peak. Feb 13 19:09:48.861002 update_engine[1486]: I20250213 19:09:48.860886 1486 update_attempter.cc:509] Updating boot flags... Feb 13 19:09:48.927102 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 38 scanned by (udev-worker) (1694) Feb 13 19:09:48.990126 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 38 scanned by (udev-worker) (1694) Feb 13 19:09:49.073097 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 38 scanned by (udev-worker) (1694) Feb 13 19:09:57.764860 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 5. Feb 13 19:09:57.772510 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:09:57.934339 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:09:57.943975 (kubelet)[1714]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 19:09:57.987764 kubelet[1714]: E0213 19:09:57.987613 1714 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 19:09:57.990339 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 19:09:57.990503 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 19:09:57.990934 systemd[1]: kubelet.service: Consumed 152ms CPU time, 101.5M memory peak. Feb 13 19:10:08.014122 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 6. Feb 13 19:10:08.021473 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:10:08.137828 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:10:08.152797 (kubelet)[1729]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 19:10:08.193947 kubelet[1729]: E0213 19:10:08.193817 1729 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 19:10:08.196338 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 19:10:08.196480 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 19:10:08.199133 systemd[1]: kubelet.service: Consumed 145ms CPU time, 104M memory peak. Feb 13 19:10:10.370903 systemd[1]: Started sshd@2-78.46.147.231:22-159.223.215.175:55910.service - OpenSSH per-connection server daemon (159.223.215.175:55910). Feb 13 19:10:10.557116 sshd[1737]: Invalid user hacked from 159.223.215.175 port 55910 Feb 13 19:10:10.585218 sshd[1737]: Received disconnect from 159.223.215.175 port 55910:11: Bye Bye [preauth] Feb 13 19:10:10.585218 sshd[1737]: Disconnected from invalid user hacked 159.223.215.175 port 55910 [preauth] Feb 13 19:10:10.587680 systemd[1]: sshd@2-78.46.147.231:22-159.223.215.175:55910.service: Deactivated successfully. Feb 13 19:10:18.263899 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 7. Feb 13 19:10:18.272087 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:10:18.386316 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:10:18.399590 (kubelet)[1749]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 19:10:18.444594 kubelet[1749]: E0213 19:10:18.444487 1749 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 19:10:18.446956 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 19:10:18.447146 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 19:10:18.447646 systemd[1]: kubelet.service: Consumed 146ms CPU time, 100.2M memory peak. Feb 13 19:10:28.514618 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 8. Feb 13 19:10:28.521959 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:10:28.643188 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:10:28.658107 (kubelet)[1765]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 19:10:28.714353 kubelet[1765]: E0213 19:10:28.714289 1765 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 19:10:28.718298 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 19:10:28.718559 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 19:10:28.719467 systemd[1]: kubelet.service: Consumed 168ms CPU time, 102.6M memory peak. Feb 13 19:10:38.763658 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 9. Feb 13 19:10:38.772638 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:10:38.886539 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:10:38.902814 (kubelet)[1780]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 19:10:38.949410 kubelet[1780]: E0213 19:10:38.949290 1780 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 19:10:38.950979 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 19:10:38.951147 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 19:10:38.951581 systemd[1]: kubelet.service: Consumed 151ms CPU time, 101.9M memory peak. Feb 13 19:10:49.014384 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 10. Feb 13 19:10:49.024312 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:10:49.143905 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:10:49.160887 (kubelet)[1794]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 19:10:49.206712 kubelet[1794]: E0213 19:10:49.206626 1794 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 19:10:49.211469 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 19:10:49.211928 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 19:10:49.212454 systemd[1]: kubelet.service: Consumed 158ms CPU time, 101.9M memory peak. Feb 13 19:10:58.777677 systemd[1]: Started sshd@3-78.46.147.231:22-139.178.68.195:42408.service - OpenSSH per-connection server daemon (139.178.68.195:42408). Feb 13 19:10:59.263823 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 11. Feb 13 19:10:59.276411 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:10:59.411028 systemd[1]: Started sshd@4-78.46.147.231:22-117.247.111.70:59351.service - OpenSSH per-connection server daemon (117.247.111.70:59351). Feb 13 19:10:59.422797 (kubelet)[1814]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 19:10:59.424295 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:10:59.480203 kubelet[1814]: E0213 19:10:59.480156 1814 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 19:10:59.483167 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 19:10:59.483377 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 19:10:59.483796 systemd[1]: kubelet.service: Consumed 148ms CPU time, 102.3M memory peak. Feb 13 19:10:59.765806 sshd[1803]: Accepted publickey for core from 139.178.68.195 port 42408 ssh2: RSA SHA256:GiO19KKKK6dAgXb8V1C7vI95O6t/PswdbHT7p8WkVYc Feb 13 19:10:59.769207 sshd-session[1803]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:10:59.778691 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Feb 13 19:10:59.792911 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Feb 13 19:10:59.804390 systemd-logind[1485]: New session 1 of user core. Feb 13 19:10:59.809560 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Feb 13 19:10:59.818519 systemd[1]: Starting user@500.service - User Manager for UID 500... Feb 13 19:10:59.824915 (systemd)[1825]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Feb 13 19:10:59.828025 systemd-logind[1485]: New session c1 of user core. Feb 13 19:10:59.962385 systemd[1825]: Queued start job for default target default.target. Feb 13 19:10:59.974278 systemd[1825]: Created slice app.slice - User Application Slice. Feb 13 19:10:59.974350 systemd[1825]: Reached target paths.target - Paths. Feb 13 19:10:59.974427 systemd[1825]: Reached target timers.target - Timers. Feb 13 19:10:59.976330 systemd[1825]: Starting dbus.socket - D-Bus User Message Bus Socket... Feb 13 19:10:59.990396 systemd[1825]: Listening on dbus.socket - D-Bus User Message Bus Socket. Feb 13 19:10:59.990642 systemd[1825]: Reached target sockets.target - Sockets. Feb 13 19:10:59.990720 systemd[1825]: Reached target basic.target - Basic System. Feb 13 19:10:59.990780 systemd[1825]: Reached target default.target - Main User Target. Feb 13 19:10:59.990827 systemd[1825]: Startup finished in 155ms. Feb 13 19:10:59.991093 systemd[1]: Started user@500.service - User Manager for UID 500. Feb 13 19:11:00.000377 systemd[1]: Started session-1.scope - Session 1 of User core. Feb 13 19:11:00.405764 sshd[1813]: Invalid user plesk from 117.247.111.70 port 59351 Feb 13 19:11:00.593733 sshd[1813]: Received disconnect from 117.247.111.70 port 59351:11: Bye Bye [preauth] Feb 13 19:11:00.593733 sshd[1813]: Disconnected from invalid user plesk 117.247.111.70 port 59351 [preauth] Feb 13 19:11:00.597011 systemd[1]: sshd@4-78.46.147.231:22-117.247.111.70:59351.service: Deactivated successfully. Feb 13 19:11:00.692326 systemd[1]: Started sshd@5-78.46.147.231:22-139.178.68.195:42424.service - OpenSSH per-connection server daemon (139.178.68.195:42424). Feb 13 19:11:01.683177 sshd[1838]: Accepted publickey for core from 139.178.68.195 port 42424 ssh2: RSA SHA256:GiO19KKKK6dAgXb8V1C7vI95O6t/PswdbHT7p8WkVYc Feb 13 19:11:01.685569 sshd-session[1838]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:11:01.694472 systemd-logind[1485]: New session 2 of user core. Feb 13 19:11:01.705388 systemd[1]: Started session-2.scope - Session 2 of User core. Feb 13 19:11:02.359646 sshd[1840]: Connection closed by 139.178.68.195 port 42424 Feb 13 19:11:02.359445 sshd-session[1838]: pam_unix(sshd:session): session closed for user core Feb 13 19:11:02.365335 systemd[1]: sshd@5-78.46.147.231:22-139.178.68.195:42424.service: Deactivated successfully. Feb 13 19:11:02.368969 systemd[1]: session-2.scope: Deactivated successfully. Feb 13 19:11:02.370465 systemd-logind[1485]: Session 2 logged out. Waiting for processes to exit. Feb 13 19:11:02.372774 systemd-logind[1485]: Removed session 2. Feb 13 19:11:02.539816 systemd[1]: Started sshd@6-78.46.147.231:22-139.178.68.195:42430.service - OpenSSH per-connection server daemon (139.178.68.195:42430). Feb 13 19:11:03.526681 sshd[1846]: Accepted publickey for core from 139.178.68.195 port 42430 ssh2: RSA SHA256:GiO19KKKK6dAgXb8V1C7vI95O6t/PswdbHT7p8WkVYc Feb 13 19:11:03.528525 sshd-session[1846]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:11:03.538115 systemd-logind[1485]: New session 3 of user core. Feb 13 19:11:03.547528 systemd[1]: Started session-3.scope - Session 3 of User core. Feb 13 19:11:04.042798 systemd[1]: Started sshd@7-78.46.147.231:22-103.154.87.242:59838.service - OpenSSH per-connection server daemon (103.154.87.242:59838). Feb 13 19:11:04.207820 sshd[1848]: Connection closed by 139.178.68.195 port 42430 Feb 13 19:11:04.208894 sshd-session[1846]: pam_unix(sshd:session): session closed for user core Feb 13 19:11:04.214946 systemd[1]: sshd@6-78.46.147.231:22-139.178.68.195:42430.service: Deactivated successfully. Feb 13 19:11:04.217373 systemd[1]: session-3.scope: Deactivated successfully. Feb 13 19:11:04.218365 systemd-logind[1485]: Session 3 logged out. Waiting for processes to exit. Feb 13 19:11:04.220158 systemd-logind[1485]: Removed session 3. Feb 13 19:11:04.386563 systemd[1]: Started sshd@8-78.46.147.231:22-139.178.68.195:42436.service - OpenSSH per-connection server daemon (139.178.68.195:42436). Feb 13 19:11:05.250084 sshd[1850]: Invalid user teampspeak from 103.154.87.242 port 59838 Feb 13 19:11:05.367914 sshd[1857]: Accepted publickey for core from 139.178.68.195 port 42436 ssh2: RSA SHA256:GiO19KKKK6dAgXb8V1C7vI95O6t/PswdbHT7p8WkVYc Feb 13 19:11:05.370011 sshd-session[1857]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:11:05.377175 systemd-logind[1485]: New session 4 of user core. Feb 13 19:11:05.384440 systemd[1]: Started session-4.scope - Session 4 of User core. Feb 13 19:11:05.447866 sshd[1850]: Received disconnect from 103.154.87.242 port 59838:11: Bye Bye [preauth] Feb 13 19:11:05.447866 sshd[1850]: Disconnected from invalid user teampspeak 103.154.87.242 port 59838 [preauth] Feb 13 19:11:05.450684 systemd[1]: sshd@7-78.46.147.231:22-103.154.87.242:59838.service: Deactivated successfully. Feb 13 19:11:06.048955 sshd[1859]: Connection closed by 139.178.68.195 port 42436 Feb 13 19:11:06.050181 sshd-session[1857]: pam_unix(sshd:session): session closed for user core Feb 13 19:11:06.056647 systemd[1]: sshd@8-78.46.147.231:22-139.178.68.195:42436.service: Deactivated successfully. Feb 13 19:11:06.058936 systemd[1]: session-4.scope: Deactivated successfully. Feb 13 19:11:06.062413 systemd-logind[1485]: Session 4 logged out. Waiting for processes to exit. Feb 13 19:11:06.063856 systemd-logind[1485]: Removed session 4. Feb 13 19:11:06.236643 systemd[1]: Started sshd@9-78.46.147.231:22-139.178.68.195:42450.service - OpenSSH per-connection server daemon (139.178.68.195:42450). Feb 13 19:11:07.225411 sshd[1867]: Accepted publickey for core from 139.178.68.195 port 42450 ssh2: RSA SHA256:GiO19KKKK6dAgXb8V1C7vI95O6t/PswdbHT7p8WkVYc Feb 13 19:11:07.227900 sshd-session[1867]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:11:07.233936 systemd-logind[1485]: New session 5 of user core. Feb 13 19:11:07.244482 systemd[1]: Started session-5.scope - Session 5 of User core. Feb 13 19:11:07.758977 sudo[1870]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Feb 13 19:11:07.759331 sudo[1870]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Feb 13 19:11:07.774371 sudo[1870]: pam_unix(sudo:session): session closed for user root Feb 13 19:11:07.934198 sshd[1869]: Connection closed by 139.178.68.195 port 42450 Feb 13 19:11:07.935803 sshd-session[1867]: pam_unix(sshd:session): session closed for user core Feb 13 19:11:07.940965 systemd[1]: sshd@9-78.46.147.231:22-139.178.68.195:42450.service: Deactivated successfully. Feb 13 19:11:07.943546 systemd[1]: session-5.scope: Deactivated successfully. Feb 13 19:11:07.944553 systemd-logind[1485]: Session 5 logged out. Waiting for processes to exit. Feb 13 19:11:07.945931 systemd-logind[1485]: Removed session 5. Feb 13 19:11:08.119701 systemd[1]: Started sshd@10-78.46.147.231:22-139.178.68.195:38850.service - OpenSSH per-connection server daemon (139.178.68.195:38850). Feb 13 19:11:09.108709 sshd[1876]: Accepted publickey for core from 139.178.68.195 port 38850 ssh2: RSA SHA256:GiO19KKKK6dAgXb8V1C7vI95O6t/PswdbHT7p8WkVYc Feb 13 19:11:09.111869 sshd-session[1876]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:11:09.119438 systemd-logind[1485]: New session 6 of user core. Feb 13 19:11:09.129531 systemd[1]: Started session-6.scope - Session 6 of User core. Feb 13 19:11:09.513565 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 12. Feb 13 19:11:09.520374 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:11:09.632713 sudo[1884]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Feb 13 19:11:09.633021 sudo[1884]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Feb 13 19:11:09.642325 sudo[1884]: pam_unix(sudo:session): session closed for user root Feb 13 19:11:09.643855 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:11:09.647561 (kubelet)[1889]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 19:11:09.652787 sudo[1882]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Feb 13 19:11:09.653553 sudo[1882]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Feb 13 19:11:09.670220 systemd[1]: Starting audit-rules.service - Load Audit Rules... Feb 13 19:11:09.703078 kubelet[1889]: E0213 19:11:09.702106 1889 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 19:11:09.704759 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 19:11:09.704900 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 19:11:09.705546 systemd[1]: kubelet.service: Consumed 152ms CPU time, 100.5M memory peak. Feb 13 19:11:09.708357 augenrules[1917]: No rules Feb 13 19:11:09.709304 systemd[1]: audit-rules.service: Deactivated successfully. Feb 13 19:11:09.709512 systemd[1]: Finished audit-rules.service - Load Audit Rules. Feb 13 19:11:09.710639 sudo[1882]: pam_unix(sudo:session): session closed for user root Feb 13 19:11:09.870139 sshd[1878]: Connection closed by 139.178.68.195 port 38850 Feb 13 19:11:09.872380 sshd-session[1876]: pam_unix(sshd:session): session closed for user core Feb 13 19:11:09.877005 systemd[1]: sshd@10-78.46.147.231:22-139.178.68.195:38850.service: Deactivated successfully. Feb 13 19:11:09.879315 systemd[1]: session-6.scope: Deactivated successfully. Feb 13 19:11:09.881724 systemd-logind[1485]: Session 6 logged out. Waiting for processes to exit. Feb 13 19:11:09.883910 systemd-logind[1485]: Removed session 6. Feb 13 19:11:10.046975 systemd[1]: Started sshd@11-78.46.147.231:22-139.178.68.195:38856.service - OpenSSH per-connection server daemon (139.178.68.195:38856). Feb 13 19:11:11.061225 sshd[1926]: Accepted publickey for core from 139.178.68.195 port 38856 ssh2: RSA SHA256:GiO19KKKK6dAgXb8V1C7vI95O6t/PswdbHT7p8WkVYc Feb 13 19:11:11.063513 sshd-session[1926]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:11:11.069160 systemd-logind[1485]: New session 7 of user core. Feb 13 19:11:11.077403 systemd[1]: Started session-7.scope - Session 7 of User core. Feb 13 19:11:11.587804 sudo[1929]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Feb 13 19:11:11.588133 sudo[1929]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Feb 13 19:11:11.929376 systemd[1]: Starting docker.service - Docker Application Container Engine... Feb 13 19:11:11.931849 (dockerd)[1946]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Feb 13 19:11:12.182442 dockerd[1946]: time="2025-02-13T19:11:12.182240145Z" level=info msg="Starting up" Feb 13 19:11:12.263450 systemd[1]: var-lib-docker-check\x2doverlayfs\x2dsupport1769443748-merged.mount: Deactivated successfully. Feb 13 19:11:12.284391 dockerd[1946]: time="2025-02-13T19:11:12.284343638Z" level=info msg="Loading containers: start." Feb 13 19:11:12.450147 kernel: Initializing XFRM netlink socket Feb 13 19:11:12.549073 systemd-networkd[1399]: docker0: Link UP Feb 13 19:11:12.587795 dockerd[1946]: time="2025-02-13T19:11:12.587697259Z" level=info msg="Loading containers: done." Feb 13 19:11:12.604192 dockerd[1946]: time="2025-02-13T19:11:12.603870291Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Feb 13 19:11:12.604192 dockerd[1946]: time="2025-02-13T19:11:12.603994889Z" level=info msg="Docker daemon" commit=41ca978a0a5400cc24b274137efa9f25517fcc0b containerd-snapshotter=false storage-driver=overlay2 version=27.3.1 Feb 13 19:11:12.604848 dockerd[1946]: time="2025-02-13T19:11:12.604552397Z" level=info msg="Daemon has completed initialization" Feb 13 19:11:12.636996 dockerd[1946]: time="2025-02-13T19:11:12.636924422Z" level=info msg="API listen on /run/docker.sock" Feb 13 19:11:12.638243 systemd[1]: Started docker.service - Docker Application Container Engine. Feb 13 19:11:13.356762 containerd[1503]: time="2025-02-13T19:11:13.356501424Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.2\"" Feb 13 19:11:14.056839 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2659845737.mount: Deactivated successfully. Feb 13 19:11:15.071091 containerd[1503]: time="2025-02-13T19:11:15.069765216Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.32.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:11:15.071447 containerd[1503]: time="2025-02-13T19:11:15.071365352Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.32.2: active requests=0, bytes read=26218328" Feb 13 19:11:15.072078 containerd[1503]: time="2025-02-13T19:11:15.071823866Z" level=info msg="ImageCreate event name:\"sha256:6417e1437b6d9a789e1ca789695a574e1df00a632bdbfbcae9695c9a7d500e32\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:11:15.075720 containerd[1503]: time="2025-02-13T19:11:15.075122736Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:c47449f3e751588ea0cb74e325e0f83db335a415f4f4c7fb147375dd6c84757f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:11:15.078174 containerd[1503]: time="2025-02-13T19:11:15.078120931Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.32.2\" with image id \"sha256:6417e1437b6d9a789e1ca789695a574e1df00a632bdbfbcae9695c9a7d500e32\", repo tag \"registry.k8s.io/kube-apiserver:v1.32.2\", repo digest \"registry.k8s.io/kube-apiserver@sha256:c47449f3e751588ea0cb74e325e0f83db335a415f4f4c7fb147375dd6c84757f\", size \"26215036\" in 1.721578947s" Feb 13 19:11:15.078174 containerd[1503]: time="2025-02-13T19:11:15.078175771Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.2\" returns image reference \"sha256:6417e1437b6d9a789e1ca789695a574e1df00a632bdbfbcae9695c9a7d500e32\"" Feb 13 19:11:15.081035 containerd[1503]: time="2025-02-13T19:11:15.080997688Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.2\"" Feb 13 19:11:16.394401 containerd[1503]: time="2025-02-13T19:11:16.394305489Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.32.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:11:16.395995 containerd[1503]: time="2025-02-13T19:11:16.395853508Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.32.2: active requests=0, bytes read=22528165" Feb 13 19:11:16.401179 containerd[1503]: time="2025-02-13T19:11:16.401098838Z" level=info msg="ImageCreate event name:\"sha256:3c9285acfd2ff7915bb451cc40ac060366ac519f3fef00c455f5aca0e0346c4d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:11:16.404783 containerd[1503]: time="2025-02-13T19:11:16.404580992Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:399aa50f4d1361c59dc458e634506d02de32613d03a9a614a21058741162ef90\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:11:16.407183 containerd[1503]: time="2025-02-13T19:11:16.407012280Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.32.2\" with image id \"sha256:3c9285acfd2ff7915bb451cc40ac060366ac519f3fef00c455f5aca0e0346c4d\", repo tag \"registry.k8s.io/kube-controller-manager:v1.32.2\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:399aa50f4d1361c59dc458e634506d02de32613d03a9a614a21058741162ef90\", size \"23968941\" in 1.325971792s" Feb 13 19:11:16.407183 containerd[1503]: time="2025-02-13T19:11:16.407080879Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.2\" returns image reference \"sha256:3c9285acfd2ff7915bb451cc40ac060366ac519f3fef00c455f5aca0e0346c4d\"" Feb 13 19:11:16.407948 containerd[1503]: time="2025-02-13T19:11:16.407715110Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.2\"" Feb 13 19:11:17.474097 containerd[1503]: time="2025-02-13T19:11:17.473569666Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.32.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:11:17.474866 containerd[1503]: time="2025-02-13T19:11:17.474801892Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.32.2: active requests=0, bytes read=17480820" Feb 13 19:11:17.475833 containerd[1503]: time="2025-02-13T19:11:17.475764481Z" level=info msg="ImageCreate event name:\"sha256:82dfa03f692fb5d84f66c17d6ee9126b081182152b25d28ea456d89b7d5d8911\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:11:17.483372 containerd[1503]: time="2025-02-13T19:11:17.482710279Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.32.2\" with image id \"sha256:82dfa03f692fb5d84f66c17d6ee9126b081182152b25d28ea456d89b7d5d8911\", repo tag \"registry.k8s.io/kube-scheduler:v1.32.2\", repo digest \"registry.k8s.io/kube-scheduler@sha256:45710d74cfd5aa10a001d0cf81747b77c28617444ffee0503d12f1dcd7450f76\", size \"18921614\" in 1.07495781s" Feb 13 19:11:17.483372 containerd[1503]: time="2025-02-13T19:11:17.483173754Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.2\" returns image reference \"sha256:82dfa03f692fb5d84f66c17d6ee9126b081182152b25d28ea456d89b7d5d8911\"" Feb 13 19:11:17.483372 containerd[1503]: time="2025-02-13T19:11:17.483133394Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:45710d74cfd5aa10a001d0cf81747b77c28617444ffee0503d12f1dcd7450f76\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:11:17.484595 containerd[1503]: time="2025-02-13T19:11:17.484527898Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.2\"" Feb 13 19:11:18.433866 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2500466129.mount: Deactivated successfully. Feb 13 19:11:18.790086 containerd[1503]: time="2025-02-13T19:11:18.787613015Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.32.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:11:18.791495 containerd[1503]: time="2025-02-13T19:11:18.791357617Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.32.2: active requests=0, bytes read=27363408" Feb 13 19:11:18.793350 containerd[1503]: time="2025-02-13T19:11:18.793287237Z" level=info msg="ImageCreate event name:\"sha256:e5aac5df76d9b8dc899ab8c4db25a7648e7fb25cafe7a155066247883c78f062\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:11:18.811084 containerd[1503]: time="2025-02-13T19:11:18.809331394Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:83c025f0faa6799fab6645102a98138e39a9a7db2be3bc792c79d72659b1805d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:11:18.811294 containerd[1503]: time="2025-02-13T19:11:18.811174856Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.32.2\" with image id \"sha256:e5aac5df76d9b8dc899ab8c4db25a7648e7fb25cafe7a155066247883c78f062\", repo tag \"registry.k8s.io/kube-proxy:v1.32.2\", repo digest \"registry.k8s.io/kube-proxy@sha256:83c025f0faa6799fab6645102a98138e39a9a7db2be3bc792c79d72659b1805d\", size \"27362401\" in 1.326382681s" Feb 13 19:11:18.811294 containerd[1503]: time="2025-02-13T19:11:18.811224055Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.2\" returns image reference \"sha256:e5aac5df76d9b8dc899ab8c4db25a7648e7fb25cafe7a155066247883c78f062\"" Feb 13 19:11:18.811866 containerd[1503]: time="2025-02-13T19:11:18.811828249Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Feb 13 19:11:19.409796 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3447955774.mount: Deactivated successfully. Feb 13 19:11:19.764137 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 13. Feb 13 19:11:19.770494 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:11:19.914824 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:11:19.928659 (kubelet)[2263]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 19:11:19.980696 kubelet[2263]: E0213 19:11:19.980588 2263 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 19:11:19.984515 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 19:11:19.984693 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 19:11:19.985144 systemd[1]: kubelet.service: Consumed 164ms CPU time, 104.2M memory peak. Feb 13 19:11:20.217816 containerd[1503]: time="2025-02-13T19:11:20.217088786Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:11:20.219289 containerd[1503]: time="2025-02-13T19:11:20.219204411Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=16951714" Feb 13 19:11:20.220407 containerd[1503]: time="2025-02-13T19:11:20.220349363Z" level=info msg="ImageCreate event name:\"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:11:20.229085 containerd[1503]: time="2025-02-13T19:11:20.227666150Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:11:20.229222 containerd[1503]: time="2025-02-13T19:11:20.229108500Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"16948420\" in 1.417238211s" Feb 13 19:11:20.229222 containerd[1503]: time="2025-02-13T19:11:20.229139580Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\"" Feb 13 19:11:20.229690 containerd[1503]: time="2025-02-13T19:11:20.229651856Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Feb 13 19:11:20.729743 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2974633198.mount: Deactivated successfully. Feb 13 19:11:20.736908 containerd[1503]: time="2025-02-13T19:11:20.735919166Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:11:20.737956 containerd[1503]: time="2025-02-13T19:11:20.737681234Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=268723" Feb 13 19:11:20.739291 containerd[1503]: time="2025-02-13T19:11:20.739220383Z" level=info msg="ImageCreate event name:\"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:11:20.742939 containerd[1503]: time="2025-02-13T19:11:20.742833477Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:11:20.744048 containerd[1503]: time="2025-02-13T19:11:20.743893789Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"267933\" in 514.166013ms" Feb 13 19:11:20.744048 containerd[1503]: time="2025-02-13T19:11:20.743939029Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\"" Feb 13 19:11:20.744048 containerd[1503]: time="2025-02-13T19:11:20.744545904Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\"" Feb 13 19:11:21.295354 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4177870453.mount: Deactivated successfully. Feb 13 19:11:22.707708 containerd[1503]: time="2025-02-13T19:11:22.707639270Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.16-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:11:22.709897 containerd[1503]: time="2025-02-13T19:11:22.709762500Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.16-0: active requests=0, bytes read=67812491" Feb 13 19:11:22.710652 containerd[1503]: time="2025-02-13T19:11:22.710554137Z" level=info msg="ImageCreate event name:\"sha256:7fc9d4aa817aa6a3e549f3cd49d1f7b496407be979fc36dd5f356d59ce8c3a82\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:11:22.718015 containerd[1503]: time="2025-02-13T19:11:22.717919784Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:11:22.719767 containerd[1503]: time="2025-02-13T19:11:22.719629136Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.16-0\" with image id \"sha256:7fc9d4aa817aa6a3e549f3cd49d1f7b496407be979fc36dd5f356d59ce8c3a82\", repo tag \"registry.k8s.io/etcd:3.5.16-0\", repo digest \"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\", size \"67941650\" in 1.975053992s" Feb 13 19:11:22.719767 containerd[1503]: time="2025-02-13T19:11:22.719669296Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\" returns image reference \"sha256:7fc9d4aa817aa6a3e549f3cd49d1f7b496407be979fc36dd5f356d59ce8c3a82\"" Feb 13 19:11:24.362429 systemd[1]: Started sshd@12-78.46.147.231:22-2.57.122.186:39742.service - OpenSSH per-connection server daemon (2.57.122.186:39742). Feb 13 19:11:24.636776 sshd[2350]: Invalid user master from 2.57.122.186 port 39742 Feb 13 19:11:24.700222 sshd[2350]: Connection closed by invalid user master 2.57.122.186 port 39742 [preauth] Feb 13 19:11:24.702144 systemd[1]: sshd@12-78.46.147.231:22-2.57.122.186:39742.service: Deactivated successfully. Feb 13 19:11:26.705530 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:11:26.706190 systemd[1]: kubelet.service: Consumed 164ms CPU time, 104.2M memory peak. Feb 13 19:11:26.730887 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:11:26.752681 systemd[1]: Reload requested from client PID 2361 ('systemctl') (unit session-7.scope)... Feb 13 19:11:26.752701 systemd[1]: Reloading... Feb 13 19:11:26.898094 zram_generator::config[2407]: No configuration found. Feb 13 19:11:26.998441 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 19:11:27.091240 systemd[1]: Reloading finished in 338 ms. Feb 13 19:11:27.139031 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:11:27.156734 (kubelet)[2445]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Feb 13 19:11:27.162283 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:11:27.162615 systemd[1]: kubelet.service: Deactivated successfully. Feb 13 19:11:27.162902 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:11:27.162950 systemd[1]: kubelet.service: Consumed 105ms CPU time, 91.6M memory peak. Feb 13 19:11:27.170684 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:11:27.289254 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:11:27.291813 (kubelet)[2461]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Feb 13 19:11:27.332120 kubelet[2461]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 13 19:11:27.332526 kubelet[2461]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Feb 13 19:11:27.332588 kubelet[2461]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 13 19:11:27.332746 kubelet[2461]: I0213 19:11:27.332709 2461 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 13 19:11:28.023116 kubelet[2461]: I0213 19:11:28.021402 2461 server.go:520] "Kubelet version" kubeletVersion="v1.32.0" Feb 13 19:11:28.023116 kubelet[2461]: I0213 19:11:28.021445 2461 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 13 19:11:28.023116 kubelet[2461]: I0213 19:11:28.021733 2461 server.go:954] "Client rotation is on, will bootstrap in background" Feb 13 19:11:28.054458 kubelet[2461]: E0213 19:11:28.054283 2461 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://78.46.147.231:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 78.46.147.231:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:11:28.059322 kubelet[2461]: I0213 19:11:28.059121 2461 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Feb 13 19:11:28.074592 kubelet[2461]: E0213 19:11:28.074543 2461 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Feb 13 19:11:28.074764 kubelet[2461]: I0213 19:11:28.074749 2461 server.go:1421] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Feb 13 19:11:28.078571 kubelet[2461]: I0213 19:11:28.077615 2461 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Feb 13 19:11:28.078571 kubelet[2461]: I0213 19:11:28.077846 2461 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 13 19:11:28.078571 kubelet[2461]: I0213 19:11:28.077873 2461 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4230-0-1-a-2d796e8b97","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 13 19:11:28.078571 kubelet[2461]: I0213 19:11:28.078138 2461 topology_manager.go:138] "Creating topology manager with none policy" Feb 13 19:11:28.078869 kubelet[2461]: I0213 19:11:28.078149 2461 container_manager_linux.go:304] "Creating device plugin manager" Feb 13 19:11:28.078869 kubelet[2461]: I0213 19:11:28.078353 2461 state_mem.go:36] "Initialized new in-memory state store" Feb 13 19:11:28.082124 kubelet[2461]: I0213 19:11:28.082096 2461 kubelet.go:446] "Attempting to sync node with API server" Feb 13 19:11:28.082267 kubelet[2461]: I0213 19:11:28.082254 2461 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 13 19:11:28.082332 kubelet[2461]: I0213 19:11:28.082322 2461 kubelet.go:352] "Adding apiserver pod source" Feb 13 19:11:28.082395 kubelet[2461]: I0213 19:11:28.082386 2461 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 13 19:11:28.085253 kubelet[2461]: W0213 19:11:28.085142 2461 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://78.46.147.231:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4230-0-1-a-2d796e8b97&limit=500&resourceVersion=0": dial tcp 78.46.147.231:6443: connect: connection refused Feb 13 19:11:28.085253 kubelet[2461]: E0213 19:11:28.085237 2461 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://78.46.147.231:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4230-0-1-a-2d796e8b97&limit=500&resourceVersion=0\": dial tcp 78.46.147.231:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:11:28.085819 kubelet[2461]: W0213 19:11:28.085764 2461 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://78.46.147.231:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 78.46.147.231:6443: connect: connection refused Feb 13 19:11:28.085873 kubelet[2461]: E0213 19:11:28.085845 2461 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://78.46.147.231:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 78.46.147.231:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:11:28.086566 kubelet[2461]: I0213 19:11:28.086526 2461 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v1.7.23" apiVersion="v1" Feb 13 19:11:28.087516 kubelet[2461]: I0213 19:11:28.087488 2461 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 13 19:11:28.087663 kubelet[2461]: W0213 19:11:28.087643 2461 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Feb 13 19:11:28.089376 kubelet[2461]: I0213 19:11:28.089308 2461 watchdog_linux.go:99] "Systemd watchdog is not enabled" Feb 13 19:11:28.089376 kubelet[2461]: I0213 19:11:28.089358 2461 server.go:1287] "Started kubelet" Feb 13 19:11:28.093314 kubelet[2461]: I0213 19:11:28.093278 2461 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Feb 13 19:11:28.093947 kubelet[2461]: I0213 19:11:28.093879 2461 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 13 19:11:28.094302 kubelet[2461]: I0213 19:11:28.094272 2461 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 13 19:11:28.094520 kubelet[2461]: I0213 19:11:28.094502 2461 server.go:490] "Adding debug handlers to kubelet server" Feb 13 19:11:28.094719 kubelet[2461]: E0213 19:11:28.094458 2461 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://78.46.147.231:6443/api/v1/namespaces/default/events\": dial tcp 78.46.147.231:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4230-0-1-a-2d796e8b97.1823da48fba1d407 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4230-0-1-a-2d796e8b97,UID:ci-4230-0-1-a-2d796e8b97,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4230-0-1-a-2d796e8b97,},FirstTimestamp:2025-02-13 19:11:28.089330695 +0000 UTC m=+0.793280626,LastTimestamp:2025-02-13 19:11:28.089330695 +0000 UTC m=+0.793280626,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4230-0-1-a-2d796e8b97,}" Feb 13 19:11:28.098420 kubelet[2461]: E0213 19:11:28.098351 2461 kubelet.go:1561] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Feb 13 19:11:28.098523 kubelet[2461]: I0213 19:11:28.098477 2461 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 13 19:11:28.100096 kubelet[2461]: I0213 19:11:28.098737 2461 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Feb 13 19:11:28.101841 kubelet[2461]: E0213 19:11:28.101775 2461 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ci-4230-0-1-a-2d796e8b97\" not found" Feb 13 19:11:28.102206 kubelet[2461]: I0213 19:11:28.102191 2461 volume_manager.go:297] "Starting Kubelet Volume Manager" Feb 13 19:11:28.102506 kubelet[2461]: I0213 19:11:28.102486 2461 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Feb 13 19:11:28.102635 kubelet[2461]: I0213 19:11:28.102624 2461 reconciler.go:26] "Reconciler: start to sync state" Feb 13 19:11:28.103467 kubelet[2461]: W0213 19:11:28.103420 2461 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://78.46.147.231:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 78.46.147.231:6443: connect: connection refused Feb 13 19:11:28.103590 kubelet[2461]: E0213 19:11:28.103570 2461 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://78.46.147.231:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 78.46.147.231:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:11:28.103864 kubelet[2461]: I0213 19:11:28.103841 2461 factory.go:221] Registration of the systemd container factory successfully Feb 13 19:11:28.104039 kubelet[2461]: I0213 19:11:28.104019 2461 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Feb 13 19:11:28.105604 kubelet[2461]: I0213 19:11:28.105582 2461 factory.go:221] Registration of the containerd container factory successfully Feb 13 19:11:28.108969 kubelet[2461]: E0213 19:11:28.108919 2461 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://78.46.147.231:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4230-0-1-a-2d796e8b97?timeout=10s\": dial tcp 78.46.147.231:6443: connect: connection refused" interval="200ms" Feb 13 19:11:28.119039 kubelet[2461]: I0213 19:11:28.118997 2461 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 13 19:11:28.120305 kubelet[2461]: I0213 19:11:28.120281 2461 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 13 19:11:28.120413 kubelet[2461]: I0213 19:11:28.120404 2461 status_manager.go:227] "Starting to sync pod status with apiserver" Feb 13 19:11:28.120479 kubelet[2461]: I0213 19:11:28.120469 2461 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Feb 13 19:11:28.120530 kubelet[2461]: I0213 19:11:28.120521 2461 kubelet.go:2388] "Starting kubelet main sync loop" Feb 13 19:11:28.120614 kubelet[2461]: E0213 19:11:28.120600 2461 kubelet.go:2412] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 13 19:11:28.127315 kubelet[2461]: W0213 19:11:28.127273 2461 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://78.46.147.231:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 78.46.147.231:6443: connect: connection refused Feb 13 19:11:28.127315 kubelet[2461]: E0213 19:11:28.127319 2461 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://78.46.147.231:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 78.46.147.231:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:11:28.128907 kubelet[2461]: I0213 19:11:28.128871 2461 cpu_manager.go:221] "Starting CPU manager" policy="none" Feb 13 19:11:28.128907 kubelet[2461]: I0213 19:11:28.128893 2461 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Feb 13 19:11:28.128907 kubelet[2461]: I0213 19:11:28.128912 2461 state_mem.go:36] "Initialized new in-memory state store" Feb 13 19:11:28.133336 kubelet[2461]: I0213 19:11:28.133271 2461 policy_none.go:49] "None policy: Start" Feb 13 19:11:28.133336 kubelet[2461]: I0213 19:11:28.133322 2461 memory_manager.go:186] "Starting memorymanager" policy="None" Feb 13 19:11:28.133336 kubelet[2461]: I0213 19:11:28.133346 2461 state_mem.go:35] "Initializing new in-memory state store" Feb 13 19:11:28.143207 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Feb 13 19:11:28.161633 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Feb 13 19:11:28.166370 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Feb 13 19:11:28.176082 kubelet[2461]: I0213 19:11:28.175605 2461 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 13 19:11:28.176082 kubelet[2461]: I0213 19:11:28.175924 2461 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 13 19:11:28.176082 kubelet[2461]: I0213 19:11:28.175940 2461 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 13 19:11:28.176885 kubelet[2461]: I0213 19:11:28.176376 2461 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 13 19:11:28.179121 kubelet[2461]: E0213 19:11:28.179080 2461 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Feb 13 19:11:28.179209 kubelet[2461]: E0213 19:11:28.179139 2461 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4230-0-1-a-2d796e8b97\" not found" Feb 13 19:11:28.236307 systemd[1]: Created slice kubepods-burstable-pod4ee02b111fd7daa99b33a2caa1fc3e26.slice - libcontainer container kubepods-burstable-pod4ee02b111fd7daa99b33a2caa1fc3e26.slice. Feb 13 19:11:28.255622 kubelet[2461]: E0213 19:11:28.255571 2461 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4230-0-1-a-2d796e8b97\" not found" node="ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:28.259970 systemd[1]: Created slice kubepods-burstable-podbbf2329f2f0ef18f9000f09b3f6b7011.slice - libcontainer container kubepods-burstable-podbbf2329f2f0ef18f9000f09b3f6b7011.slice. Feb 13 19:11:28.270595 kubelet[2461]: E0213 19:11:28.270535 2461 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4230-0-1-a-2d796e8b97\" not found" node="ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:28.276601 systemd[1]: Created slice kubepods-burstable-pod33f6cd87c806adfcd4f517b693e8092e.slice - libcontainer container kubepods-burstable-pod33f6cd87c806adfcd4f517b693e8092e.slice. Feb 13 19:11:28.279420 kubelet[2461]: E0213 19:11:28.279382 2461 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4230-0-1-a-2d796e8b97\" not found" node="ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:28.280494 kubelet[2461]: I0213 19:11:28.280472 2461 kubelet_node_status.go:76] "Attempting to register node" node="ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:28.281110 kubelet[2461]: E0213 19:11:28.281008 2461 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://78.46.147.231:6443/api/v1/nodes\": dial tcp 78.46.147.231:6443: connect: connection refused" node="ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:28.305159 kubelet[2461]: I0213 19:11:28.305048 2461 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/bbf2329f2f0ef18f9000f09b3f6b7011-ca-certs\") pod \"kube-controller-manager-ci-4230-0-1-a-2d796e8b97\" (UID: \"bbf2329f2f0ef18f9000f09b3f6b7011\") " pod="kube-system/kube-controller-manager-ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:28.305351 kubelet[2461]: I0213 19:11:28.305188 2461 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/bbf2329f2f0ef18f9000f09b3f6b7011-flexvolume-dir\") pod \"kube-controller-manager-ci-4230-0-1-a-2d796e8b97\" (UID: \"bbf2329f2f0ef18f9000f09b3f6b7011\") " pod="kube-system/kube-controller-manager-ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:28.305351 kubelet[2461]: I0213 19:11:28.305226 2461 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/bbf2329f2f0ef18f9000f09b3f6b7011-kubeconfig\") pod \"kube-controller-manager-ci-4230-0-1-a-2d796e8b97\" (UID: \"bbf2329f2f0ef18f9000f09b3f6b7011\") " pod="kube-system/kube-controller-manager-ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:28.305351 kubelet[2461]: I0213 19:11:28.305293 2461 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/bbf2329f2f0ef18f9000f09b3f6b7011-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4230-0-1-a-2d796e8b97\" (UID: \"bbf2329f2f0ef18f9000f09b3f6b7011\") " pod="kube-system/kube-controller-manager-ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:28.305603 kubelet[2461]: I0213 19:11:28.305326 2461 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/33f6cd87c806adfcd4f517b693e8092e-kubeconfig\") pod \"kube-scheduler-ci-4230-0-1-a-2d796e8b97\" (UID: \"33f6cd87c806adfcd4f517b693e8092e\") " pod="kube-system/kube-scheduler-ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:28.305603 kubelet[2461]: I0213 19:11:28.305520 2461 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/4ee02b111fd7daa99b33a2caa1fc3e26-ca-certs\") pod \"kube-apiserver-ci-4230-0-1-a-2d796e8b97\" (UID: \"4ee02b111fd7daa99b33a2caa1fc3e26\") " pod="kube-system/kube-apiserver-ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:28.305603 kubelet[2461]: I0213 19:11:28.305575 2461 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/4ee02b111fd7daa99b33a2caa1fc3e26-k8s-certs\") pod \"kube-apiserver-ci-4230-0-1-a-2d796e8b97\" (UID: \"4ee02b111fd7daa99b33a2caa1fc3e26\") " pod="kube-system/kube-apiserver-ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:28.305742 kubelet[2461]: I0213 19:11:28.305631 2461 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/4ee02b111fd7daa99b33a2caa1fc3e26-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4230-0-1-a-2d796e8b97\" (UID: \"4ee02b111fd7daa99b33a2caa1fc3e26\") " pod="kube-system/kube-apiserver-ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:28.305742 kubelet[2461]: I0213 19:11:28.305652 2461 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/bbf2329f2f0ef18f9000f09b3f6b7011-k8s-certs\") pod \"kube-controller-manager-ci-4230-0-1-a-2d796e8b97\" (UID: \"bbf2329f2f0ef18f9000f09b3f6b7011\") " pod="kube-system/kube-controller-manager-ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:28.311254 kubelet[2461]: E0213 19:11:28.311185 2461 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://78.46.147.231:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4230-0-1-a-2d796e8b97?timeout=10s\": dial tcp 78.46.147.231:6443: connect: connection refused" interval="400ms" Feb 13 19:11:28.485469 kubelet[2461]: I0213 19:11:28.484633 2461 kubelet_node_status.go:76] "Attempting to register node" node="ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:28.485469 kubelet[2461]: E0213 19:11:28.485389 2461 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://78.46.147.231:6443/api/v1/nodes\": dial tcp 78.46.147.231:6443: connect: connection refused" node="ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:28.559216 containerd[1503]: time="2025-02-13T19:11:28.558679015Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4230-0-1-a-2d796e8b97,Uid:4ee02b111fd7daa99b33a2caa1fc3e26,Namespace:kube-system,Attempt:0,}" Feb 13 19:11:28.572217 containerd[1503]: time="2025-02-13T19:11:28.572115534Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4230-0-1-a-2d796e8b97,Uid:bbf2329f2f0ef18f9000f09b3f6b7011,Namespace:kube-system,Attempt:0,}" Feb 13 19:11:28.580891 containerd[1503]: time="2025-02-13T19:11:28.580793679Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4230-0-1-a-2d796e8b97,Uid:33f6cd87c806adfcd4f517b693e8092e,Namespace:kube-system,Attempt:0,}" Feb 13 19:11:28.712969 kubelet[2461]: E0213 19:11:28.712860 2461 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://78.46.147.231:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4230-0-1-a-2d796e8b97?timeout=10s\": dial tcp 78.46.147.231:6443: connect: connection refused" interval="800ms" Feb 13 19:11:28.888671 kubelet[2461]: I0213 19:11:28.888390 2461 kubelet_node_status.go:76] "Attempting to register node" node="ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:28.889131 kubelet[2461]: E0213 19:11:28.889081 2461 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://78.46.147.231:6443/api/v1/nodes\": dial tcp 78.46.147.231:6443: connect: connection refused" node="ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:28.996855 kubelet[2461]: W0213 19:11:28.996747 2461 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://78.46.147.231:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 78.46.147.231:6443: connect: connection refused Feb 13 19:11:28.997102 kubelet[2461]: E0213 19:11:28.996901 2461 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://78.46.147.231:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 78.46.147.231:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:11:29.085497 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3645317920.mount: Deactivated successfully. Feb 13 19:11:29.091190 containerd[1503]: time="2025-02-13T19:11:29.090940136Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 19:11:29.092968 containerd[1503]: time="2025-02-13T19:11:29.092923384Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 19:11:29.095271 containerd[1503]: time="2025-02-13T19:11:29.095220593Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=269193" Feb 13 19:11:29.095925 containerd[1503]: time="2025-02-13T19:11:29.095854236Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Feb 13 19:11:29.097958 containerd[1503]: time="2025-02-13T19:11:29.097854404Z" level=info msg="ImageCreate event name:\"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 19:11:29.099485 containerd[1503]: time="2025-02-13T19:11:29.099386770Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Feb 13 19:11:29.099936 containerd[1503]: time="2025-02-13T19:11:29.099905572Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 19:11:29.109090 containerd[1503]: time="2025-02-13T19:11:29.108165085Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 19:11:29.109090 containerd[1503]: time="2025-02-13T19:11:29.109079369Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 528.141729ms" Feb 13 19:11:29.110871 containerd[1503]: time="2025-02-13T19:11:29.110836816Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 551.92864ms" Feb 13 19:11:29.112248 containerd[1503]: time="2025-02-13T19:11:29.112210341Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 539.958527ms" Feb 13 19:11:29.224947 containerd[1503]: time="2025-02-13T19:11:29.224605830Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:11:29.224947 containerd[1503]: time="2025-02-13T19:11:29.224821591Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:11:29.224947 containerd[1503]: time="2025-02-13T19:11:29.224834751Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:11:29.226133 containerd[1503]: time="2025-02-13T19:11:29.225628274Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:11:29.228527 containerd[1503]: time="2025-02-13T19:11:29.228244204Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:11:29.228527 containerd[1503]: time="2025-02-13T19:11:29.228309325Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:11:29.228527 containerd[1503]: time="2025-02-13T19:11:29.228324365Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:11:29.231788 containerd[1503]: time="2025-02-13T19:11:29.230530493Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:11:29.234535 containerd[1503]: time="2025-02-13T19:11:29.234188428Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:11:29.234535 containerd[1503]: time="2025-02-13T19:11:29.234241348Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:11:29.234535 containerd[1503]: time="2025-02-13T19:11:29.234255748Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:11:29.234535 containerd[1503]: time="2025-02-13T19:11:29.234334709Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:11:29.254283 systemd[1]: Started cri-containerd-571b706a80af6df754b1195ea6008135264cf47f77378086fbd48e3e6214bc78.scope - libcontainer container 571b706a80af6df754b1195ea6008135264cf47f77378086fbd48e3e6214bc78. Feb 13 19:11:29.275268 systemd[1]: Started cri-containerd-ae783dee3de5c947f40aea00fa3fe6c2a9cb8b418b4d592c4a4ff592e1338a88.scope - libcontainer container ae783dee3de5c947f40aea00fa3fe6c2a9cb8b418b4d592c4a4ff592e1338a88. Feb 13 19:11:29.277081 systemd[1]: Started cri-containerd-f62c24a5b75bacfc105a4852474f6719d3e970f7d6e3f67115e0adb23391ff83.scope - libcontainer container f62c24a5b75bacfc105a4852474f6719d3e970f7d6e3f67115e0adb23391ff83. Feb 13 19:11:29.327828 containerd[1503]: time="2025-02-13T19:11:29.327750361Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4230-0-1-a-2d796e8b97,Uid:4ee02b111fd7daa99b33a2caa1fc3e26,Namespace:kube-system,Attempt:0,} returns sandbox id \"f62c24a5b75bacfc105a4852474f6719d3e970f7d6e3f67115e0adb23391ff83\"" Feb 13 19:11:29.330294 containerd[1503]: time="2025-02-13T19:11:29.330265852Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4230-0-1-a-2d796e8b97,Uid:33f6cd87c806adfcd4f517b693e8092e,Namespace:kube-system,Attempt:0,} returns sandbox id \"ae783dee3de5c947f40aea00fa3fe6c2a9cb8b418b4d592c4a4ff592e1338a88\"" Feb 13 19:11:29.335486 containerd[1503]: time="2025-02-13T19:11:29.335402472Z" level=info msg="CreateContainer within sandbox \"f62c24a5b75bacfc105a4852474f6719d3e970f7d6e3f67115e0adb23391ff83\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Feb 13 19:11:29.337340 containerd[1503]: time="2025-02-13T19:11:29.337195399Z" level=info msg="CreateContainer within sandbox \"ae783dee3de5c947f40aea00fa3fe6c2a9cb8b418b4d592c4a4ff592e1338a88\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Feb 13 19:11:29.344613 containerd[1503]: time="2025-02-13T19:11:29.344573149Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4230-0-1-a-2d796e8b97,Uid:bbf2329f2f0ef18f9000f09b3f6b7011,Namespace:kube-system,Attempt:0,} returns sandbox id \"571b706a80af6df754b1195ea6008135264cf47f77378086fbd48e3e6214bc78\"" Feb 13 19:11:29.348508 containerd[1503]: time="2025-02-13T19:11:29.348467204Z" level=info msg="CreateContainer within sandbox \"571b706a80af6df754b1195ea6008135264cf47f77378086fbd48e3e6214bc78\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Feb 13 19:11:29.358392 containerd[1503]: time="2025-02-13T19:11:29.358310003Z" level=info msg="CreateContainer within sandbox \"ae783dee3de5c947f40aea00fa3fe6c2a9cb8b418b4d592c4a4ff592e1338a88\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"4331d2e00e884df3af15f071151f8b76e642f6b2922bff88e9eafc77b008d0af\"" Feb 13 19:11:29.359417 containerd[1503]: time="2025-02-13T19:11:29.359389528Z" level=info msg="StartContainer for \"4331d2e00e884df3af15f071151f8b76e642f6b2922bff88e9eafc77b008d0af\"" Feb 13 19:11:29.362094 containerd[1503]: time="2025-02-13T19:11:29.360836654Z" level=info msg="CreateContainer within sandbox \"f62c24a5b75bacfc105a4852474f6719d3e970f7d6e3f67115e0adb23391ff83\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"9eb726359ace23eca44e773f74b83d48bc605e6b7f4808f8370dfab2186a2b53\"" Feb 13 19:11:29.362094 containerd[1503]: time="2025-02-13T19:11:29.361438856Z" level=info msg="StartContainer for \"9eb726359ace23eca44e773f74b83d48bc605e6b7f4808f8370dfab2186a2b53\"" Feb 13 19:11:29.365507 kubelet[2461]: W0213 19:11:29.365444 2461 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://78.46.147.231:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 78.46.147.231:6443: connect: connection refused Feb 13 19:11:29.365634 kubelet[2461]: E0213 19:11:29.365512 2461 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://78.46.147.231:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 78.46.147.231:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:11:29.370048 containerd[1503]: time="2025-02-13T19:11:29.369980850Z" level=info msg="CreateContainer within sandbox \"571b706a80af6df754b1195ea6008135264cf47f77378086fbd48e3e6214bc78\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"b068163f576e0c62469fcae99d5b6f7785627c423b4030e9bc3bda75db16c086\"" Feb 13 19:11:29.370770 containerd[1503]: time="2025-02-13T19:11:29.370730013Z" level=info msg="StartContainer for \"b068163f576e0c62469fcae99d5b6f7785627c423b4030e9bc3bda75db16c086\"" Feb 13 19:11:29.387381 systemd[1]: Started cri-containerd-4331d2e00e884df3af15f071151f8b76e642f6b2922bff88e9eafc77b008d0af.scope - libcontainer container 4331d2e00e884df3af15f071151f8b76e642f6b2922bff88e9eafc77b008d0af. Feb 13 19:11:29.408321 systemd[1]: Started cri-containerd-9eb726359ace23eca44e773f74b83d48bc605e6b7f4808f8370dfab2186a2b53.scope - libcontainer container 9eb726359ace23eca44e773f74b83d48bc605e6b7f4808f8370dfab2186a2b53. Feb 13 19:11:29.422270 systemd[1]: Started cri-containerd-b068163f576e0c62469fcae99d5b6f7785627c423b4030e9bc3bda75db16c086.scope - libcontainer container b068163f576e0c62469fcae99d5b6f7785627c423b4030e9bc3bda75db16c086. Feb 13 19:11:29.463131 containerd[1503]: time="2025-02-13T19:11:29.462776380Z" level=info msg="StartContainer for \"4331d2e00e884df3af15f071151f8b76e642f6b2922bff88e9eafc77b008d0af\" returns successfully" Feb 13 19:11:29.473623 containerd[1503]: time="2025-02-13T19:11:29.473509623Z" level=info msg="StartContainer for \"9eb726359ace23eca44e773f74b83d48bc605e6b7f4808f8370dfab2186a2b53\" returns successfully" Feb 13 19:11:29.514709 kubelet[2461]: E0213 19:11:29.514510 2461 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://78.46.147.231:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4230-0-1-a-2d796e8b97?timeout=10s\": dial tcp 78.46.147.231:6443: connect: connection refused" interval="1.6s" Feb 13 19:11:29.518534 containerd[1503]: time="2025-02-13T19:11:29.518406882Z" level=info msg="StartContainer for \"b068163f576e0c62469fcae99d5b6f7785627c423b4030e9bc3bda75db16c086\" returns successfully" Feb 13 19:11:29.544392 kubelet[2461]: W0213 19:11:29.544284 2461 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://78.46.147.231:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4230-0-1-a-2d796e8b97&limit=500&resourceVersion=0": dial tcp 78.46.147.231:6443: connect: connection refused Feb 13 19:11:29.544392 kubelet[2461]: E0213 19:11:29.544359 2461 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://78.46.147.231:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4230-0-1-a-2d796e8b97&limit=500&resourceVersion=0\": dial tcp 78.46.147.231:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:11:29.691211 kubelet[2461]: I0213 19:11:29.691167 2461 kubelet_node_status.go:76] "Attempting to register node" node="ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:30.151700 kubelet[2461]: E0213 19:11:30.151526 2461 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4230-0-1-a-2d796e8b97\" not found" node="ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:30.153706 kubelet[2461]: E0213 19:11:30.153677 2461 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4230-0-1-a-2d796e8b97\" not found" node="ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:30.157441 kubelet[2461]: E0213 19:11:30.157416 2461 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4230-0-1-a-2d796e8b97\" not found" node="ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:31.162233 kubelet[2461]: E0213 19:11:31.161872 2461 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4230-0-1-a-2d796e8b97\" not found" node="ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:31.162233 kubelet[2461]: E0213 19:11:31.161915 2461 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4230-0-1-a-2d796e8b97\" not found" node="ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:31.357422 kubelet[2461]: E0213 19:11:31.357281 2461 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-4230-0-1-a-2d796e8b97\" not found" node="ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:31.425766 kubelet[2461]: E0213 19:11:31.425255 2461 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{ci-4230-0-1-a-2d796e8b97.1823da48fba1d407 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4230-0-1-a-2d796e8b97,UID:ci-4230-0-1-a-2d796e8b97,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4230-0-1-a-2d796e8b97,},FirstTimestamp:2025-02-13 19:11:28.089330695 +0000 UTC m=+0.793280626,LastTimestamp:2025-02-13 19:11:28.089330695 +0000 UTC m=+0.793280626,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4230-0-1-a-2d796e8b97,}" Feb 13 19:11:31.479514 kubelet[2461]: E0213 19:11:31.479401 2461 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{ci-4230-0-1-a-2d796e8b97.1823da48fc2b2b31 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4230-0-1-a-2d796e8b97,UID:ci-4230-0-1-a-2d796e8b97,APIVersion:,ResourceVersion:,FieldPath:,},Reason:InvalidDiskCapacity,Message:invalid capacity 0 on image filesystem,Source:EventSource{Component:kubelet,Host:ci-4230-0-1-a-2d796e8b97,},FirstTimestamp:2025-02-13 19:11:28.098331441 +0000 UTC m=+0.802281332,LastTimestamp:2025-02-13 19:11:28.098331441 +0000 UTC m=+0.802281332,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4230-0-1-a-2d796e8b97,}" Feb 13 19:11:31.513195 kubelet[2461]: I0213 19:11:31.512972 2461 kubelet_node_status.go:79] "Successfully registered node" node="ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:31.513195 kubelet[2461]: E0213 19:11:31.513016 2461 kubelet_node_status.go:549] "Error updating node status, will retry" err="error getting node \"ci-4230-0-1-a-2d796e8b97\": node \"ci-4230-0-1-a-2d796e8b97\" not found" Feb 13 19:11:31.524145 kubelet[2461]: E0213 19:11:31.523419 2461 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ci-4230-0-1-a-2d796e8b97\" not found" Feb 13 19:11:31.541727 kubelet[2461]: E0213 19:11:31.541587 2461 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{ci-4230-0-1-a-2d796e8b97.1823da48fde1159c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4230-0-1-a-2d796e8b97,UID:ci-4230-0-1-a-2d796e8b97,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node ci-4230-0-1-a-2d796e8b97 status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:ci-4230-0-1-a-2d796e8b97,},FirstTimestamp:2025-02-13 19:11:28.127030684 +0000 UTC m=+0.830980575,LastTimestamp:2025-02-13 19:11:28.127030684 +0000 UTC m=+0.830980575,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4230-0-1-a-2d796e8b97,}" Feb 13 19:11:31.602580 kubelet[2461]: E0213 19:11:31.602278 2461 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{ci-4230-0-1-a-2d796e8b97.1823da48fde12c94 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4230-0-1-a-2d796e8b97,UID:ci-4230-0-1-a-2d796e8b97,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node ci-4230-0-1-a-2d796e8b97 status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:ci-4230-0-1-a-2d796e8b97,},FirstTimestamp:2025-02-13 19:11:28.127036564 +0000 UTC m=+0.830986455,LastTimestamp:2025-02-13 19:11:28.127036564 +0000 UTC m=+0.830986455,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4230-0-1-a-2d796e8b97,}" Feb 13 19:11:31.623599 kubelet[2461]: E0213 19:11:31.623551 2461 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ci-4230-0-1-a-2d796e8b97\" not found" Feb 13 19:11:31.665142 kubelet[2461]: E0213 19:11:31.664733 2461 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{ci-4230-0-1-a-2d796e8b97.1823da48fde13c0c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4230-0-1-a-2d796e8b97,UID:ci-4230-0-1-a-2d796e8b97,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node ci-4230-0-1-a-2d796e8b97 status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:ci-4230-0-1-a-2d796e8b97,},FirstTimestamp:2025-02-13 19:11:28.127040524 +0000 UTC m=+0.830990415,LastTimestamp:2025-02-13 19:11:28.127040524 +0000 UTC m=+0.830990415,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4230-0-1-a-2d796e8b97,}" Feb 13 19:11:31.720965 kubelet[2461]: E0213 19:11:31.720766 2461 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{ci-4230-0-1-a-2d796e8b97.1823da49010d9d8e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4230-0-1-a-2d796e8b97,UID:ci-4230-0-1-a-2d796e8b97,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:ci-4230-0-1-a-2d796e8b97,},FirstTimestamp:2025-02-13 19:11:28.180280718 +0000 UTC m=+0.884230609,LastTimestamp:2025-02-13 19:11:28.180280718 +0000 UTC m=+0.884230609,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4230-0-1-a-2d796e8b97,}" Feb 13 19:11:31.724272 kubelet[2461]: E0213 19:11:31.724245 2461 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ci-4230-0-1-a-2d796e8b97\" not found" Feb 13 19:11:31.824755 kubelet[2461]: E0213 19:11:31.824715 2461 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ci-4230-0-1-a-2d796e8b97\" not found" Feb 13 19:11:31.925170 kubelet[2461]: E0213 19:11:31.925114 2461 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ci-4230-0-1-a-2d796e8b97\" not found" Feb 13 19:11:32.025857 kubelet[2461]: E0213 19:11:32.025803 2461 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ci-4230-0-1-a-2d796e8b97\" not found" Feb 13 19:11:32.126314 kubelet[2461]: E0213 19:11:32.125966 2461 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ci-4230-0-1-a-2d796e8b97\" not found" Feb 13 19:11:32.226552 kubelet[2461]: E0213 19:11:32.226507 2461 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ci-4230-0-1-a-2d796e8b97\" not found" Feb 13 19:11:32.305246 kubelet[2461]: I0213 19:11:32.304775 2461 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:32.330083 kubelet[2461]: I0213 19:11:32.329438 2461 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:32.343381 kubelet[2461]: I0213 19:11:32.343342 2461 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:33.087473 kubelet[2461]: I0213 19:11:33.087431 2461 apiserver.go:52] "Watching apiserver" Feb 13 19:11:33.103602 kubelet[2461]: I0213 19:11:33.103508 2461 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" Feb 13 19:11:33.758229 systemd[1]: Reload requested from client PID 2738 ('systemctl') (unit session-7.scope)... Feb 13 19:11:33.758375 systemd[1]: Reloading... Feb 13 19:11:33.892093 zram_generator::config[2786]: No configuration found. Feb 13 19:11:34.004525 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 19:11:34.112708 systemd[1]: Reloading finished in 353 ms. Feb 13 19:11:34.136352 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:11:34.151173 systemd[1]: kubelet.service: Deactivated successfully. Feb 13 19:11:34.151532 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:11:34.151617 systemd[1]: kubelet.service: Consumed 1.198s CPU time, 122.3M memory peak. Feb 13 19:11:34.158373 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:11:34.295636 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:11:34.296221 (kubelet)[2828]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Feb 13 19:11:34.365032 kubelet[2828]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 13 19:11:34.365032 kubelet[2828]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Feb 13 19:11:34.365032 kubelet[2828]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 13 19:11:34.365032 kubelet[2828]: I0213 19:11:34.364652 2828 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 13 19:11:34.376999 kubelet[2828]: I0213 19:11:34.375848 2828 server.go:520] "Kubelet version" kubeletVersion="v1.32.0" Feb 13 19:11:34.376999 kubelet[2828]: I0213 19:11:34.376928 2828 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 13 19:11:34.377269 kubelet[2828]: I0213 19:11:34.377228 2828 server.go:954] "Client rotation is on, will bootstrap in background" Feb 13 19:11:34.378758 kubelet[2828]: I0213 19:11:34.378730 2828 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Feb 13 19:11:34.382743 kubelet[2828]: I0213 19:11:34.382432 2828 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Feb 13 19:11:34.389585 kubelet[2828]: E0213 19:11:34.389555 2828 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Feb 13 19:11:34.390041 kubelet[2828]: I0213 19:11:34.390024 2828 server.go:1421] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Feb 13 19:11:34.392639 kubelet[2828]: I0213 19:11:34.392607 2828 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Feb 13 19:11:34.394089 kubelet[2828]: I0213 19:11:34.393016 2828 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 13 19:11:34.394089 kubelet[2828]: I0213 19:11:34.393080 2828 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4230-0-1-a-2d796e8b97","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 13 19:11:34.394089 kubelet[2828]: I0213 19:11:34.393264 2828 topology_manager.go:138] "Creating topology manager with none policy" Feb 13 19:11:34.394089 kubelet[2828]: I0213 19:11:34.393274 2828 container_manager_linux.go:304] "Creating device plugin manager" Feb 13 19:11:34.394328 kubelet[2828]: I0213 19:11:34.393319 2828 state_mem.go:36] "Initialized new in-memory state store" Feb 13 19:11:34.394328 kubelet[2828]: I0213 19:11:34.393460 2828 kubelet.go:446] "Attempting to sync node with API server" Feb 13 19:11:34.394328 kubelet[2828]: I0213 19:11:34.393473 2828 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 13 19:11:34.394328 kubelet[2828]: I0213 19:11:34.393491 2828 kubelet.go:352] "Adding apiserver pod source" Feb 13 19:11:34.394328 kubelet[2828]: I0213 19:11:34.393500 2828 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 13 19:11:34.395224 kubelet[2828]: I0213 19:11:34.395204 2828 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v1.7.23" apiVersion="v1" Feb 13 19:11:34.395771 kubelet[2828]: I0213 19:11:34.395752 2828 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 13 19:11:34.396369 kubelet[2828]: I0213 19:11:34.396348 2828 watchdog_linux.go:99] "Systemd watchdog is not enabled" Feb 13 19:11:34.396477 kubelet[2828]: I0213 19:11:34.396467 2828 server.go:1287] "Started kubelet" Feb 13 19:11:34.401862 kubelet[2828]: I0213 19:11:34.401678 2828 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 13 19:11:34.408214 kubelet[2828]: I0213 19:11:34.408153 2828 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Feb 13 19:11:34.411076 kubelet[2828]: I0213 19:11:34.409253 2828 server.go:490] "Adding debug handlers to kubelet server" Feb 13 19:11:34.411076 kubelet[2828]: I0213 19:11:34.410369 2828 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 13 19:11:34.411076 kubelet[2828]: I0213 19:11:34.410564 2828 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 13 19:11:34.411076 kubelet[2828]: I0213 19:11:34.410760 2828 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Feb 13 19:11:34.413199 kubelet[2828]: I0213 19:11:34.413171 2828 volume_manager.go:297] "Starting Kubelet Volume Manager" Feb 13 19:11:34.413463 kubelet[2828]: E0213 19:11:34.413432 2828 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ci-4230-0-1-a-2d796e8b97\" not found" Feb 13 19:11:34.413840 kubelet[2828]: I0213 19:11:34.413764 2828 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Feb 13 19:11:34.415493 kubelet[2828]: I0213 19:11:34.414379 2828 reconciler.go:26] "Reconciler: start to sync state" Feb 13 19:11:34.427442 kubelet[2828]: I0213 19:11:34.427391 2828 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 13 19:11:34.428595 kubelet[2828]: I0213 19:11:34.428570 2828 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 13 19:11:34.429020 kubelet[2828]: I0213 19:11:34.428705 2828 status_manager.go:227] "Starting to sync pod status with apiserver" Feb 13 19:11:34.429020 kubelet[2828]: I0213 19:11:34.428731 2828 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Feb 13 19:11:34.429020 kubelet[2828]: I0213 19:11:34.428739 2828 kubelet.go:2388] "Starting kubelet main sync loop" Feb 13 19:11:34.429020 kubelet[2828]: E0213 19:11:34.428784 2828 kubelet.go:2412] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 13 19:11:34.433819 kubelet[2828]: I0213 19:11:34.433768 2828 factory.go:221] Registration of the systemd container factory successfully Feb 13 19:11:34.433922 kubelet[2828]: I0213 19:11:34.433899 2828 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Feb 13 19:11:34.452362 kubelet[2828]: E0213 19:11:34.452331 2828 kubelet.go:1561] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Feb 13 19:11:34.456619 kubelet[2828]: I0213 19:11:34.456585 2828 factory.go:221] Registration of the containerd container factory successfully Feb 13 19:11:34.504101 kubelet[2828]: I0213 19:11:34.503875 2828 cpu_manager.go:221] "Starting CPU manager" policy="none" Feb 13 19:11:34.504101 kubelet[2828]: I0213 19:11:34.503898 2828 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Feb 13 19:11:34.504101 kubelet[2828]: I0213 19:11:34.503923 2828 state_mem.go:36] "Initialized new in-memory state store" Feb 13 19:11:34.504368 kubelet[2828]: I0213 19:11:34.504132 2828 state_mem.go:88] "Updated default CPUSet" cpuSet="" Feb 13 19:11:34.504368 kubelet[2828]: I0213 19:11:34.504146 2828 state_mem.go:96] "Updated CPUSet assignments" assignments={} Feb 13 19:11:34.504368 kubelet[2828]: I0213 19:11:34.504166 2828 policy_none.go:49] "None policy: Start" Feb 13 19:11:34.504368 kubelet[2828]: I0213 19:11:34.504177 2828 memory_manager.go:186] "Starting memorymanager" policy="None" Feb 13 19:11:34.504368 kubelet[2828]: I0213 19:11:34.504187 2828 state_mem.go:35] "Initializing new in-memory state store" Feb 13 19:11:34.504368 kubelet[2828]: I0213 19:11:34.504313 2828 state_mem.go:75] "Updated machine memory state" Feb 13 19:11:34.512073 kubelet[2828]: I0213 19:11:34.510853 2828 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 13 19:11:34.512073 kubelet[2828]: I0213 19:11:34.511091 2828 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 13 19:11:34.512073 kubelet[2828]: I0213 19:11:34.511106 2828 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 13 19:11:34.512073 kubelet[2828]: I0213 19:11:34.511916 2828 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 13 19:11:34.517006 kubelet[2828]: E0213 19:11:34.516981 2828 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Feb 13 19:11:34.529936 kubelet[2828]: I0213 19:11:34.529882 2828 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:34.530357 kubelet[2828]: I0213 19:11:34.530338 2828 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:34.533846 kubelet[2828]: I0213 19:11:34.532247 2828 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:34.544390 kubelet[2828]: E0213 19:11:34.543919 2828 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4230-0-1-a-2d796e8b97\" already exists" pod="kube-system/kube-controller-manager-ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:34.545039 kubelet[2828]: E0213 19:11:34.544006 2828 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4230-0-1-a-2d796e8b97\" already exists" pod="kube-system/kube-scheduler-ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:34.545402 kubelet[2828]: E0213 19:11:34.545374 2828 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4230-0-1-a-2d796e8b97\" already exists" pod="kube-system/kube-apiserver-ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:34.617180 kubelet[2828]: I0213 19:11:34.614886 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/bbf2329f2f0ef18f9000f09b3f6b7011-flexvolume-dir\") pod \"kube-controller-manager-ci-4230-0-1-a-2d796e8b97\" (UID: \"bbf2329f2f0ef18f9000f09b3f6b7011\") " pod="kube-system/kube-controller-manager-ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:34.617180 kubelet[2828]: I0213 19:11:34.614939 2828 kubelet_node_status.go:76] "Attempting to register node" node="ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:34.617180 kubelet[2828]: I0213 19:11:34.614945 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/4ee02b111fd7daa99b33a2caa1fc3e26-k8s-certs\") pod \"kube-apiserver-ci-4230-0-1-a-2d796e8b97\" (UID: \"4ee02b111fd7daa99b33a2caa1fc3e26\") " pod="kube-system/kube-apiserver-ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:34.617180 kubelet[2828]: I0213 19:11:34.614976 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/bbf2329f2f0ef18f9000f09b3f6b7011-ca-certs\") pod \"kube-controller-manager-ci-4230-0-1-a-2d796e8b97\" (UID: \"bbf2329f2f0ef18f9000f09b3f6b7011\") " pod="kube-system/kube-controller-manager-ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:34.617180 kubelet[2828]: I0213 19:11:34.615005 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/bbf2329f2f0ef18f9000f09b3f6b7011-kubeconfig\") pod \"kube-controller-manager-ci-4230-0-1-a-2d796e8b97\" (UID: \"bbf2329f2f0ef18f9000f09b3f6b7011\") " pod="kube-system/kube-controller-manager-ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:34.617381 kubelet[2828]: I0213 19:11:34.615033 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/bbf2329f2f0ef18f9000f09b3f6b7011-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4230-0-1-a-2d796e8b97\" (UID: \"bbf2329f2f0ef18f9000f09b3f6b7011\") " pod="kube-system/kube-controller-manager-ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:34.617381 kubelet[2828]: I0213 19:11:34.615151 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/33f6cd87c806adfcd4f517b693e8092e-kubeconfig\") pod \"kube-scheduler-ci-4230-0-1-a-2d796e8b97\" (UID: \"33f6cd87c806adfcd4f517b693e8092e\") " pod="kube-system/kube-scheduler-ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:34.617381 kubelet[2828]: I0213 19:11:34.615178 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/4ee02b111fd7daa99b33a2caa1fc3e26-ca-certs\") pod \"kube-apiserver-ci-4230-0-1-a-2d796e8b97\" (UID: \"4ee02b111fd7daa99b33a2caa1fc3e26\") " pod="kube-system/kube-apiserver-ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:34.617381 kubelet[2828]: I0213 19:11:34.615203 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/4ee02b111fd7daa99b33a2caa1fc3e26-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4230-0-1-a-2d796e8b97\" (UID: \"4ee02b111fd7daa99b33a2caa1fc3e26\") " pod="kube-system/kube-apiserver-ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:34.617381 kubelet[2828]: I0213 19:11:34.615233 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/bbf2329f2f0ef18f9000f09b3f6b7011-k8s-certs\") pod \"kube-controller-manager-ci-4230-0-1-a-2d796e8b97\" (UID: \"bbf2329f2f0ef18f9000f09b3f6b7011\") " pod="kube-system/kube-controller-manager-ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:34.628012 kubelet[2828]: I0213 19:11:34.627975 2828 kubelet_node_status.go:125] "Node was previously registered" node="ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:34.628185 kubelet[2828]: I0213 19:11:34.628089 2828 kubelet_node_status.go:79] "Successfully registered node" node="ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:35.395112 kubelet[2828]: I0213 19:11:35.394837 2828 apiserver.go:52] "Watching apiserver" Feb 13 19:11:35.414686 kubelet[2828]: I0213 19:11:35.414542 2828 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" Feb 13 19:11:35.489560 kubelet[2828]: I0213 19:11:35.489519 2828 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:35.510088 kubelet[2828]: E0213 19:11:35.506759 2828 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4230-0-1-a-2d796e8b97\" already exists" pod="kube-system/kube-scheduler-ci-4230-0-1-a-2d796e8b97" Feb 13 19:11:35.525224 kubelet[2828]: I0213 19:11:35.524166 2828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4230-0-1-a-2d796e8b97" podStartSLOduration=3.524123751 podStartE2EDuration="3.524123751s" podCreationTimestamp="2025-02-13 19:11:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 19:11:35.523865309 +0000 UTC m=+1.222317027" watchObservedRunningTime="2025-02-13 19:11:35.524123751 +0000 UTC m=+1.222575429" Feb 13 19:11:35.572657 kubelet[2828]: I0213 19:11:35.572595 2828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4230-0-1-a-2d796e8b97" podStartSLOduration=3.5725716690000002 podStartE2EDuration="3.572571669s" podCreationTimestamp="2025-02-13 19:11:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 19:11:35.537892807 +0000 UTC m=+1.236344485" watchObservedRunningTime="2025-02-13 19:11:35.572571669 +0000 UTC m=+1.271023347" Feb 13 19:11:35.615930 kubelet[2828]: I0213 19:11:35.615503 2828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4230-0-1-a-2d796e8b97" podStartSLOduration=3.615472573 podStartE2EDuration="3.615472573s" podCreationTimestamp="2025-02-13 19:11:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 19:11:35.574098884 +0000 UTC m=+1.272550562" watchObservedRunningTime="2025-02-13 19:11:35.615472573 +0000 UTC m=+1.313924211" Feb 13 19:11:39.821509 sudo[1929]: pam_unix(sudo:session): session closed for user root Feb 13 19:11:39.983611 sshd[1928]: Connection closed by 139.178.68.195 port 38856 Feb 13 19:11:39.985412 sshd-session[1926]: pam_unix(sshd:session): session closed for user core Feb 13 19:11:39.991749 systemd[1]: sshd@11-78.46.147.231:22-139.178.68.195:38856.service: Deactivated successfully. Feb 13 19:11:39.996128 systemd[1]: session-7.scope: Deactivated successfully. Feb 13 19:11:39.996597 systemd[1]: session-7.scope: Consumed 5.916s CPU time, 226.2M memory peak. Feb 13 19:11:39.998083 systemd-logind[1485]: Session 7 logged out. Waiting for processes to exit. Feb 13 19:11:40.000788 systemd-logind[1485]: Removed session 7. Feb 13 19:11:40.592237 kubelet[2828]: I0213 19:11:40.591658 2828 kuberuntime_manager.go:1702] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Feb 13 19:11:40.592730 containerd[1503]: time="2025-02-13T19:11:40.592135289Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Feb 13 19:11:40.593794 kubelet[2828]: I0213 19:11:40.593218 2828 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Feb 13 19:11:41.356540 kubelet[2828]: I0213 19:11:41.356193 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/81efc5a9-15e2-4e8d-acc9-46a90ca462c1-kube-proxy\") pod \"kube-proxy-pvwhg\" (UID: \"81efc5a9-15e2-4e8d-acc9-46a90ca462c1\") " pod="kube-system/kube-proxy-pvwhg" Feb 13 19:11:41.356540 kubelet[2828]: I0213 19:11:41.356250 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/81efc5a9-15e2-4e8d-acc9-46a90ca462c1-xtables-lock\") pod \"kube-proxy-pvwhg\" (UID: \"81efc5a9-15e2-4e8d-acc9-46a90ca462c1\") " pod="kube-system/kube-proxy-pvwhg" Feb 13 19:11:41.356540 kubelet[2828]: I0213 19:11:41.356272 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/81efc5a9-15e2-4e8d-acc9-46a90ca462c1-lib-modules\") pod \"kube-proxy-pvwhg\" (UID: \"81efc5a9-15e2-4e8d-acc9-46a90ca462c1\") " pod="kube-system/kube-proxy-pvwhg" Feb 13 19:11:41.356540 kubelet[2828]: I0213 19:11:41.356287 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4f9b\" (UniqueName: \"kubernetes.io/projected/81efc5a9-15e2-4e8d-acc9-46a90ca462c1-kube-api-access-p4f9b\") pod \"kube-proxy-pvwhg\" (UID: \"81efc5a9-15e2-4e8d-acc9-46a90ca462c1\") " pod="kube-system/kube-proxy-pvwhg" Feb 13 19:11:41.360454 kubelet[2828]: I0213 19:11:41.360409 2828 status_manager.go:890] "Failed to get status for pod" podUID="81efc5a9-15e2-4e8d-acc9-46a90ca462c1" pod="kube-system/kube-proxy-pvwhg" err="pods \"kube-proxy-pvwhg\" is forbidden: User \"system:node:ci-4230-0-1-a-2d796e8b97\" cannot get resource \"pods\" in API group \"\" in the namespace \"kube-system\": no relationship found between node 'ci-4230-0-1-a-2d796e8b97' and this object" Feb 13 19:11:41.367579 systemd[1]: Created slice kubepods-besteffort-pod81efc5a9_15e2_4e8d_acc9_46a90ca462c1.slice - libcontainer container kubepods-besteffort-pod81efc5a9_15e2_4e8d_acc9_46a90ca462c1.slice. Feb 13 19:11:41.679022 containerd[1503]: time="2025-02-13T19:11:41.678827817Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-pvwhg,Uid:81efc5a9-15e2-4e8d-acc9-46a90ca462c1,Namespace:kube-system,Attempt:0,}" Feb 13 19:11:41.716676 containerd[1503]: time="2025-02-13T19:11:41.716518012Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:11:41.717983 containerd[1503]: time="2025-02-13T19:11:41.717574308Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:11:41.717983 containerd[1503]: time="2025-02-13T19:11:41.717617748Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:11:41.717983 containerd[1503]: time="2025-02-13T19:11:41.717785671Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:11:41.748275 systemd[1]: Started cri-containerd-b1d39653e57d886a10d06d8a08975780e843def9e8703d11a0950a232d0e3add.scope - libcontainer container b1d39653e57d886a10d06d8a08975780e843def9e8703d11a0950a232d0e3add. Feb 13 19:11:41.758634 systemd[1]: Created slice kubepods-besteffort-pod5feef9f1_661a_4119_bebd_9e602dcd10f8.slice - libcontainer container kubepods-besteffort-pod5feef9f1_661a_4119_bebd_9e602dcd10f8.slice. Feb 13 19:11:41.830467 containerd[1503]: time="2025-02-13T19:11:41.830116884Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-pvwhg,Uid:81efc5a9-15e2-4e8d-acc9-46a90ca462c1,Namespace:kube-system,Attempt:0,} returns sandbox id \"b1d39653e57d886a10d06d8a08975780e843def9e8703d11a0950a232d0e3add\"" Feb 13 19:11:41.836048 containerd[1503]: time="2025-02-13T19:11:41.835988931Z" level=info msg="CreateContainer within sandbox \"b1d39653e57d886a10d06d8a08975780e843def9e8703d11a0950a232d0e3add\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Feb 13 19:11:41.850428 containerd[1503]: time="2025-02-13T19:11:41.850386743Z" level=info msg="CreateContainer within sandbox \"b1d39653e57d886a10d06d8a08975780e843def9e8703d11a0950a232d0e3add\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"73b690b84220db353f920206e0bb0beb9a0bfd75338de727f21143a9f3133414\"" Feb 13 19:11:41.852497 containerd[1503]: time="2025-02-13T19:11:41.852442573Z" level=info msg="StartContainer for \"73b690b84220db353f920206e0bb0beb9a0bfd75338de727f21143a9f3133414\"" Feb 13 19:11:41.858613 kubelet[2828]: I0213 19:11:41.858553 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vg455\" (UniqueName: \"kubernetes.io/projected/5feef9f1-661a-4119-bebd-9e602dcd10f8-kube-api-access-vg455\") pod \"tigera-operator-7d68577dc5-glhg9\" (UID: \"5feef9f1-661a-4119-bebd-9e602dcd10f8\") " pod="tigera-operator/tigera-operator-7d68577dc5-glhg9" Feb 13 19:11:41.858613 kubelet[2828]: I0213 19:11:41.858611 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/5feef9f1-661a-4119-bebd-9e602dcd10f8-var-lib-calico\") pod \"tigera-operator-7d68577dc5-glhg9\" (UID: \"5feef9f1-661a-4119-bebd-9e602dcd10f8\") " pod="tigera-operator/tigera-operator-7d68577dc5-glhg9" Feb 13 19:11:41.885462 systemd[1]: Started cri-containerd-73b690b84220db353f920206e0bb0beb9a0bfd75338de727f21143a9f3133414.scope - libcontainer container 73b690b84220db353f920206e0bb0beb9a0bfd75338de727f21143a9f3133414. Feb 13 19:11:41.917214 containerd[1503]: time="2025-02-13T19:11:41.917169005Z" level=info msg="StartContainer for \"73b690b84220db353f920206e0bb0beb9a0bfd75338de727f21143a9f3133414\" returns successfully" Feb 13 19:11:42.062740 containerd[1503]: time="2025-02-13T19:11:42.062395707Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-7d68577dc5-glhg9,Uid:5feef9f1-661a-4119-bebd-9e602dcd10f8,Namespace:tigera-operator,Attempt:0,}" Feb 13 19:11:42.087148 containerd[1503]: time="2025-02-13T19:11:42.086431399Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:11:42.087148 containerd[1503]: time="2025-02-13T19:11:42.086543000Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:11:42.087148 containerd[1503]: time="2025-02-13T19:11:42.086561441Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:11:42.087148 containerd[1503]: time="2025-02-13T19:11:42.086638162Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:11:42.118311 systemd[1]: Started cri-containerd-78caa98ae70d6324e6aa14f5bcf60257e08da3e9281b16881ed05a4718a471f5.scope - libcontainer container 78caa98ae70d6324e6aa14f5bcf60257e08da3e9281b16881ed05a4718a471f5. Feb 13 19:11:42.150269 containerd[1503]: time="2025-02-13T19:11:42.150210863Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-7d68577dc5-glhg9,Uid:5feef9f1-661a-4119-bebd-9e602dcd10f8,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"78caa98ae70d6324e6aa14f5bcf60257e08da3e9281b16881ed05a4718a471f5\"" Feb 13 19:11:42.153415 containerd[1503]: time="2025-02-13T19:11:42.153366472Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.36.2\"" Feb 13 19:11:46.994502 kubelet[2828]: I0213 19:11:46.994418 2828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-pvwhg" podStartSLOduration=5.994276311 podStartE2EDuration="5.994276311s" podCreationTimestamp="2025-02-13 19:11:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 19:11:42.527287646 +0000 UTC m=+8.225739364" watchObservedRunningTime="2025-02-13 19:11:46.994276311 +0000 UTC m=+12.692727989" Feb 13 19:11:48.930997 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1180627797.mount: Deactivated successfully. Feb 13 19:12:00.976079 containerd[1503]: time="2025-02-13T19:12:00.975993528Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.36.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:12:00.977370 containerd[1503]: time="2025-02-13T19:12:00.977182798Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.36.2: active requests=0, bytes read=19124160" Feb 13 19:12:00.978821 containerd[1503]: time="2025-02-13T19:12:00.978507711Z" level=info msg="ImageCreate event name:\"sha256:30d521e4e84764b396aacbb2a373ca7a573f84571e3955b34329652acccfb73c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:12:00.981815 containerd[1503]: time="2025-02-13T19:12:00.981764714Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:fc9ea45f2475fd99db1b36d2ff180a50017b1a5ea0e82a171c6b439b3a620764\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:12:00.985338 containerd[1503]: time="2025-02-13T19:12:00.985286322Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.36.2\" with image id \"sha256:30d521e4e84764b396aacbb2a373ca7a573f84571e3955b34329652acccfb73c\", repo tag \"quay.io/tigera/operator:v1.36.2\", repo digest \"quay.io/tigera/operator@sha256:fc9ea45f2475fd99db1b36d2ff180a50017b1a5ea0e82a171c6b439b3a620764\", size \"19120155\" in 18.831840529s" Feb 13 19:12:00.985507 containerd[1503]: time="2025-02-13T19:12:00.985331643Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.36.2\" returns image reference \"sha256:30d521e4e84764b396aacbb2a373ca7a573f84571e3955b34329652acccfb73c\"" Feb 13 19:12:00.990907 containerd[1503]: time="2025-02-13T19:12:00.990844462Z" level=info msg="CreateContainer within sandbox \"78caa98ae70d6324e6aa14f5bcf60257e08da3e9281b16881ed05a4718a471f5\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Feb 13 19:12:01.017482 containerd[1503]: time="2025-02-13T19:12:01.017426259Z" level=info msg="CreateContainer within sandbox \"78caa98ae70d6324e6aa14f5bcf60257e08da3e9281b16881ed05a4718a471f5\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"3518d3dda6b4af3d0174a109e8bf22d1ebb2020a4ac0f1159e0cb3834115bfad\"" Feb 13 19:12:01.018947 containerd[1503]: time="2025-02-13T19:12:01.018880536Z" level=info msg="StartContainer for \"3518d3dda6b4af3d0174a109e8bf22d1ebb2020a4ac0f1159e0cb3834115bfad\"" Feb 13 19:12:01.056379 systemd[1]: Started cri-containerd-3518d3dda6b4af3d0174a109e8bf22d1ebb2020a4ac0f1159e0cb3834115bfad.scope - libcontainer container 3518d3dda6b4af3d0174a109e8bf22d1ebb2020a4ac0f1159e0cb3834115bfad. Feb 13 19:12:01.091779 containerd[1503]: time="2025-02-13T19:12:01.091673959Z" level=info msg="StartContainer for \"3518d3dda6b4af3d0174a109e8bf22d1ebb2020a4ac0f1159e0cb3834115bfad\" returns successfully" Feb 13 19:12:01.575092 kubelet[2828]: I0213 19:12:01.574496 2828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-7d68577dc5-glhg9" podStartSLOduration=1.739045757 podStartE2EDuration="20.574473995s" podCreationTimestamp="2025-02-13 19:11:41 +0000 UTC" firstStartedPulling="2025-02-13 19:11:42.15125796 +0000 UTC m=+7.849709598" lastFinishedPulling="2025-02-13 19:12:00.986686158 +0000 UTC m=+26.685137836" observedRunningTime="2025-02-13 19:12:01.574139307 +0000 UTC m=+27.272591025" watchObservedRunningTime="2025-02-13 19:12:01.574473995 +0000 UTC m=+27.272925713" Feb 13 19:12:05.078245 systemd[1]: Created slice kubepods-besteffort-podbe757208_fa50_494b_a199_332359e0f0c1.slice - libcontainer container kubepods-besteffort-podbe757208_fa50_494b_a199_332359e0f0c1.slice. Feb 13 19:12:05.119470 kubelet[2828]: I0213 19:12:05.119350 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/be757208-fa50-494b-a199-332359e0f0c1-tigera-ca-bundle\") pod \"calico-typha-7c86f976f-v6pf2\" (UID: \"be757208-fa50-494b-a199-332359e0f0c1\") " pod="calico-system/calico-typha-7c86f976f-v6pf2" Feb 13 19:12:05.119470 kubelet[2828]: I0213 19:12:05.119399 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpbpp\" (UniqueName: \"kubernetes.io/projected/be757208-fa50-494b-a199-332359e0f0c1-kube-api-access-zpbpp\") pod \"calico-typha-7c86f976f-v6pf2\" (UID: \"be757208-fa50-494b-a199-332359e0f0c1\") " pod="calico-system/calico-typha-7c86f976f-v6pf2" Feb 13 19:12:05.119470 kubelet[2828]: I0213 19:12:05.119420 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/be757208-fa50-494b-a199-332359e0f0c1-typha-certs\") pod \"calico-typha-7c86f976f-v6pf2\" (UID: \"be757208-fa50-494b-a199-332359e0f0c1\") " pod="calico-system/calico-typha-7c86f976f-v6pf2" Feb 13 19:12:05.322106 kubelet[2828]: I0213 19:12:05.320752 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3aa63002-df02-4bd7-a4be-bc77e95f3916-lib-modules\") pod \"calico-node-8zbsj\" (UID: \"3aa63002-df02-4bd7-a4be-bc77e95f3916\") " pod="calico-system/calico-node-8zbsj" Feb 13 19:12:05.322106 kubelet[2828]: I0213 19:12:05.320794 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/3aa63002-df02-4bd7-a4be-bc77e95f3916-var-lib-calico\") pod \"calico-node-8zbsj\" (UID: \"3aa63002-df02-4bd7-a4be-bc77e95f3916\") " pod="calico-system/calico-node-8zbsj" Feb 13 19:12:05.322106 kubelet[2828]: I0213 19:12:05.320810 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/3aa63002-df02-4bd7-a4be-bc77e95f3916-policysync\") pod \"calico-node-8zbsj\" (UID: \"3aa63002-df02-4bd7-a4be-bc77e95f3916\") " pod="calico-system/calico-node-8zbsj" Feb 13 19:12:05.322106 kubelet[2828]: I0213 19:12:05.320825 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/3aa63002-df02-4bd7-a4be-bc77e95f3916-cni-net-dir\") pod \"calico-node-8zbsj\" (UID: \"3aa63002-df02-4bd7-a4be-bc77e95f3916\") " pod="calico-system/calico-node-8zbsj" Feb 13 19:12:05.322106 kubelet[2828]: I0213 19:12:05.320841 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/3aa63002-df02-4bd7-a4be-bc77e95f3916-flexvol-driver-host\") pod \"calico-node-8zbsj\" (UID: \"3aa63002-df02-4bd7-a4be-bc77e95f3916\") " pod="calico-system/calico-node-8zbsj" Feb 13 19:12:05.321380 systemd[1]: Created slice kubepods-besteffort-pod3aa63002_df02_4bd7_a4be_bc77e95f3916.slice - libcontainer container kubepods-besteffort-pod3aa63002_df02_4bd7_a4be_bc77e95f3916.slice. Feb 13 19:12:05.322474 kubelet[2828]: I0213 19:12:05.320858 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qrmx\" (UniqueName: \"kubernetes.io/projected/3aa63002-df02-4bd7-a4be-bc77e95f3916-kube-api-access-6qrmx\") pod \"calico-node-8zbsj\" (UID: \"3aa63002-df02-4bd7-a4be-bc77e95f3916\") " pod="calico-system/calico-node-8zbsj" Feb 13 19:12:05.322474 kubelet[2828]: I0213 19:12:05.320874 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/3aa63002-df02-4bd7-a4be-bc77e95f3916-cni-log-dir\") pod \"calico-node-8zbsj\" (UID: \"3aa63002-df02-4bd7-a4be-bc77e95f3916\") " pod="calico-system/calico-node-8zbsj" Feb 13 19:12:05.322474 kubelet[2828]: I0213 19:12:05.320891 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/3aa63002-df02-4bd7-a4be-bc77e95f3916-xtables-lock\") pod \"calico-node-8zbsj\" (UID: \"3aa63002-df02-4bd7-a4be-bc77e95f3916\") " pod="calico-system/calico-node-8zbsj" Feb 13 19:12:05.322474 kubelet[2828]: I0213 19:12:05.320909 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3aa63002-df02-4bd7-a4be-bc77e95f3916-tigera-ca-bundle\") pod \"calico-node-8zbsj\" (UID: \"3aa63002-df02-4bd7-a4be-bc77e95f3916\") " pod="calico-system/calico-node-8zbsj" Feb 13 19:12:05.322474 kubelet[2828]: I0213 19:12:05.320927 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/3aa63002-df02-4bd7-a4be-bc77e95f3916-node-certs\") pod \"calico-node-8zbsj\" (UID: \"3aa63002-df02-4bd7-a4be-bc77e95f3916\") " pod="calico-system/calico-node-8zbsj" Feb 13 19:12:05.322577 kubelet[2828]: I0213 19:12:05.320942 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/3aa63002-df02-4bd7-a4be-bc77e95f3916-var-run-calico\") pod \"calico-node-8zbsj\" (UID: \"3aa63002-df02-4bd7-a4be-bc77e95f3916\") " pod="calico-system/calico-node-8zbsj" Feb 13 19:12:05.322577 kubelet[2828]: I0213 19:12:05.320958 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/3aa63002-df02-4bd7-a4be-bc77e95f3916-cni-bin-dir\") pod \"calico-node-8zbsj\" (UID: \"3aa63002-df02-4bd7-a4be-bc77e95f3916\") " pod="calico-system/calico-node-8zbsj" Feb 13 19:12:05.385299 containerd[1503]: time="2025-02-13T19:12:05.384124946Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-7c86f976f-v6pf2,Uid:be757208-fa50-494b-a199-332359e0f0c1,Namespace:calico-system,Attempt:0,}" Feb 13 19:12:05.443547 kubelet[2828]: E0213 19:12:05.441629 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.443547 kubelet[2828]: W0213 19:12:05.441655 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.443547 kubelet[2828]: E0213 19:12:05.441680 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.448009 kubelet[2828]: E0213 19:12:05.447963 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.448009 kubelet[2828]: W0213 19:12:05.447989 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.448009 kubelet[2828]: E0213 19:12:05.448010 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.458551 containerd[1503]: time="2025-02-13T19:12:05.458442596Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:12:05.458939 containerd[1503]: time="2025-02-13T19:12:05.458509878Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:12:05.458939 containerd[1503]: time="2025-02-13T19:12:05.458527399Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:12:05.458939 containerd[1503]: time="2025-02-13T19:12:05.458611441Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:12:05.481806 kubelet[2828]: E0213 19:12:05.481325 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-6v9pr" podUID="b9c8452b-31d9-47bc-9697-dd4f293d71b9" Feb 13 19:12:05.492908 systemd[1]: Started cri-containerd-d4803efbd3f0d878e260db5abda856908d97888e81b4e77e0966aefbf1906381.scope - libcontainer container d4803efbd3f0d878e260db5abda856908d97888e81b4e77e0966aefbf1906381. Feb 13 19:12:05.520743 kubelet[2828]: E0213 19:12:05.520617 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.520743 kubelet[2828]: W0213 19:12:05.520661 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.520743 kubelet[2828]: E0213 19:12:05.520686 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.524023 kubelet[2828]: E0213 19:12:05.522374 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.524023 kubelet[2828]: W0213 19:12:05.522395 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.524023 kubelet[2828]: E0213 19:12:05.522447 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.524023 kubelet[2828]: E0213 19:12:05.523472 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.524023 kubelet[2828]: W0213 19:12:05.523487 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.524023 kubelet[2828]: E0213 19:12:05.523502 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.524023 kubelet[2828]: E0213 19:12:05.523670 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.524023 kubelet[2828]: W0213 19:12:05.523677 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.524023 kubelet[2828]: E0213 19:12:05.523686 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.524023 kubelet[2828]: E0213 19:12:05.523859 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.524406 kubelet[2828]: W0213 19:12:05.523867 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.524406 kubelet[2828]: E0213 19:12:05.523875 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.524546 kubelet[2828]: E0213 19:12:05.524531 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.525105 kubelet[2828]: W0213 19:12:05.524613 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.525105 kubelet[2828]: E0213 19:12:05.524634 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.525105 kubelet[2828]: E0213 19:12:05.524813 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.525105 kubelet[2828]: W0213 19:12:05.524822 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.525105 kubelet[2828]: E0213 19:12:05.524830 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.525105 kubelet[2828]: E0213 19:12:05.524966 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.525105 kubelet[2828]: W0213 19:12:05.524973 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.525105 kubelet[2828]: E0213 19:12:05.524982 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.525626 kubelet[2828]: E0213 19:12:05.525483 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.525626 kubelet[2828]: W0213 19:12:05.525498 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.525626 kubelet[2828]: E0213 19:12:05.525550 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.526226 kubelet[2828]: E0213 19:12:05.526209 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.526507 kubelet[2828]: W0213 19:12:05.526386 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.526507 kubelet[2828]: E0213 19:12:05.526411 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.526710 kubelet[2828]: E0213 19:12:05.526696 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.526793 kubelet[2828]: W0213 19:12:05.526780 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.527830 kubelet[2828]: E0213 19:12:05.526878 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.528302 kubelet[2828]: E0213 19:12:05.528096 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.528302 kubelet[2828]: W0213 19:12:05.528150 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.528302 kubelet[2828]: E0213 19:12:05.528164 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.528814 kubelet[2828]: E0213 19:12:05.528679 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.528814 kubelet[2828]: W0213 19:12:05.528711 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.528814 kubelet[2828]: E0213 19:12:05.528728 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.529364 kubelet[2828]: E0213 19:12:05.529223 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.529364 kubelet[2828]: W0213 19:12:05.529275 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.529364 kubelet[2828]: E0213 19:12:05.529291 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.529718 kubelet[2828]: E0213 19:12:05.529598 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.529718 kubelet[2828]: W0213 19:12:05.529611 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.529718 kubelet[2828]: E0213 19:12:05.529622 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.530134 kubelet[2828]: E0213 19:12:05.529968 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.530134 kubelet[2828]: W0213 19:12:05.529982 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.530134 kubelet[2828]: E0213 19:12:05.529993 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.530952 kubelet[2828]: E0213 19:12:05.530491 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.530952 kubelet[2828]: W0213 19:12:05.530503 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.530952 kubelet[2828]: E0213 19:12:05.530518 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.531283 kubelet[2828]: E0213 19:12:05.531151 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.531283 kubelet[2828]: W0213 19:12:05.531168 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.531283 kubelet[2828]: E0213 19:12:05.531180 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.533217 kubelet[2828]: E0213 19:12:05.532388 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.533217 kubelet[2828]: W0213 19:12:05.532405 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.533217 kubelet[2828]: E0213 19:12:05.532417 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.533217 kubelet[2828]: E0213 19:12:05.532682 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.533217 kubelet[2828]: W0213 19:12:05.532693 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.533217 kubelet[2828]: E0213 19:12:05.532703 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.533217 kubelet[2828]: E0213 19:12:05.533066 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.533217 kubelet[2828]: W0213 19:12:05.533076 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.533217 kubelet[2828]: E0213 19:12:05.533099 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.533605 kubelet[2828]: I0213 19:12:05.533121 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsr27\" (UniqueName: \"kubernetes.io/projected/b9c8452b-31d9-47bc-9697-dd4f293d71b9-kube-api-access-hsr27\") pod \"csi-node-driver-6v9pr\" (UID: \"b9c8452b-31d9-47bc-9697-dd4f293d71b9\") " pod="calico-system/csi-node-driver-6v9pr" Feb 13 19:12:05.534303 kubelet[2828]: E0213 19:12:05.533947 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.534303 kubelet[2828]: W0213 19:12:05.533963 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.534303 kubelet[2828]: E0213 19:12:05.533976 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.534303 kubelet[2828]: I0213 19:12:05.533998 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/b9c8452b-31d9-47bc-9697-dd4f293d71b9-registration-dir\") pod \"csi-node-driver-6v9pr\" (UID: \"b9c8452b-31d9-47bc-9697-dd4f293d71b9\") " pod="calico-system/csi-node-driver-6v9pr" Feb 13 19:12:05.534868 kubelet[2828]: E0213 19:12:05.534755 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.535183 kubelet[2828]: W0213 19:12:05.534975 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.535183 kubelet[2828]: E0213 19:12:05.535004 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.535183 kubelet[2828]: I0213 19:12:05.535024 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b9c8452b-31d9-47bc-9697-dd4f293d71b9-kubelet-dir\") pod \"csi-node-driver-6v9pr\" (UID: \"b9c8452b-31d9-47bc-9697-dd4f293d71b9\") " pod="calico-system/csi-node-driver-6v9pr" Feb 13 19:12:05.536222 kubelet[2828]: E0213 19:12:05.535819 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.536222 kubelet[2828]: W0213 19:12:05.535836 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.536222 kubelet[2828]: E0213 19:12:05.535865 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.536438 kubelet[2828]: E0213 19:12:05.536252 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.536438 kubelet[2828]: W0213 19:12:05.536264 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.536438 kubelet[2828]: E0213 19:12:05.536346 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.538567 kubelet[2828]: E0213 19:12:05.538533 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.538768 kubelet[2828]: W0213 19:12:05.538666 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.538951 kubelet[2828]: E0213 19:12:05.538831 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.539530 kubelet[2828]: E0213 19:12:05.539498 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.539530 kubelet[2828]: W0213 19:12:05.539518 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.539530 kubelet[2828]: E0213 19:12:05.539557 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.539530 kubelet[2828]: I0213 19:12:05.539588 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/b9c8452b-31d9-47bc-9697-dd4f293d71b9-socket-dir\") pod \"csi-node-driver-6v9pr\" (UID: \"b9c8452b-31d9-47bc-9697-dd4f293d71b9\") " pod="calico-system/csi-node-driver-6v9pr" Feb 13 19:12:05.539992 kubelet[2828]: E0213 19:12:05.539948 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.539992 kubelet[2828]: W0213 19:12:05.539964 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.540778 kubelet[2828]: E0213 19:12:05.540634 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.540880 kubelet[2828]: E0213 19:12:05.540862 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.540880 kubelet[2828]: W0213 19:12:05.540877 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.541176 kubelet[2828]: E0213 19:12:05.540891 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.541503 kubelet[2828]: E0213 19:12:05.541214 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.541503 kubelet[2828]: W0213 19:12:05.541228 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.541503 kubelet[2828]: E0213 19:12:05.541262 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.541834 kubelet[2828]: E0213 19:12:05.541807 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.541834 kubelet[2828]: W0213 19:12:05.541827 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.541914 kubelet[2828]: E0213 19:12:05.541846 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.541914 kubelet[2828]: I0213 19:12:05.541867 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/b9c8452b-31d9-47bc-9697-dd4f293d71b9-varrun\") pod \"csi-node-driver-6v9pr\" (UID: \"b9c8452b-31d9-47bc-9697-dd4f293d71b9\") " pod="calico-system/csi-node-driver-6v9pr" Feb 13 19:12:05.542642 kubelet[2828]: E0213 19:12:05.542614 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.542642 kubelet[2828]: W0213 19:12:05.542634 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.542642 kubelet[2828]: E0213 19:12:05.542680 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.543304 kubelet[2828]: E0213 19:12:05.543278 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.543304 kubelet[2828]: W0213 19:12:05.543299 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.543495 kubelet[2828]: E0213 19:12:05.543434 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.544293 kubelet[2828]: E0213 19:12:05.544268 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.544293 kubelet[2828]: W0213 19:12:05.544293 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.544441 kubelet[2828]: E0213 19:12:05.544308 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.544894 kubelet[2828]: E0213 19:12:05.544872 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.544894 kubelet[2828]: W0213 19:12:05.544888 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.544984 kubelet[2828]: E0213 19:12:05.544900 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.608041 containerd[1503]: time="2025-02-13T19:12:05.607974082Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-7c86f976f-v6pf2,Uid:be757208-fa50-494b-a199-332359e0f0c1,Namespace:calico-system,Attempt:0,} returns sandbox id \"d4803efbd3f0d878e260db5abda856908d97888e81b4e77e0966aefbf1906381\"" Feb 13 19:12:05.611482 containerd[1503]: time="2025-02-13T19:12:05.611231490Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.29.1\"" Feb 13 19:12:05.626691 containerd[1503]: time="2025-02-13T19:12:05.626644667Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-8zbsj,Uid:3aa63002-df02-4bd7-a4be-bc77e95f3916,Namespace:calico-system,Attempt:0,}" Feb 13 19:12:05.644258 kubelet[2828]: E0213 19:12:05.643585 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.644258 kubelet[2828]: W0213 19:12:05.643612 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.644258 kubelet[2828]: E0213 19:12:05.643651 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.644258 kubelet[2828]: E0213 19:12:05.643995 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.644258 kubelet[2828]: W0213 19:12:05.644007 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.644258 kubelet[2828]: E0213 19:12:05.644020 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.644823 kubelet[2828]: E0213 19:12:05.644541 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.644823 kubelet[2828]: W0213 19:12:05.644553 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.644823 kubelet[2828]: E0213 19:12:05.644585 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.646449 kubelet[2828]: E0213 19:12:05.644794 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.646449 kubelet[2828]: W0213 19:12:05.645309 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.646449 kubelet[2828]: E0213 19:12:05.645325 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.646449 kubelet[2828]: E0213 19:12:05.645597 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.646449 kubelet[2828]: W0213 19:12:05.645609 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.646449 kubelet[2828]: E0213 19:12:05.645625 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.646449 kubelet[2828]: E0213 19:12:05.646343 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.646449 kubelet[2828]: W0213 19:12:05.646356 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.648668 kubelet[2828]: E0213 19:12:05.646556 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.648668 kubelet[2828]: W0213 19:12:05.646565 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.648668 kubelet[2828]: E0213 19:12:05.646576 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.648668 kubelet[2828]: E0213 19:12:05.647231 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.648668 kubelet[2828]: W0213 19:12:05.647316 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.648668 kubelet[2828]: E0213 19:12:05.647332 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.648668 kubelet[2828]: E0213 19:12:05.646377 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.648668 kubelet[2828]: E0213 19:12:05.647686 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.648668 kubelet[2828]: W0213 19:12:05.647697 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.648668 kubelet[2828]: E0213 19:12:05.647709 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.648866 kubelet[2828]: E0213 19:12:05.648087 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.648866 kubelet[2828]: W0213 19:12:05.648098 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.648866 kubelet[2828]: E0213 19:12:05.648110 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.651488 kubelet[2828]: E0213 19:12:05.651361 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.651488 kubelet[2828]: W0213 19:12:05.651381 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.651606 kubelet[2828]: E0213 19:12:05.651495 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.652448 kubelet[2828]: E0213 19:12:05.652144 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.652448 kubelet[2828]: W0213 19:12:05.652162 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.652448 kubelet[2828]: E0213 19:12:05.652315 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.653260 kubelet[2828]: E0213 19:12:05.653217 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.653260 kubelet[2828]: W0213 19:12:05.653236 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.654380 kubelet[2828]: E0213 19:12:05.653290 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.654380 kubelet[2828]: E0213 19:12:05.653731 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.654380 kubelet[2828]: W0213 19:12:05.653746 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.654380 kubelet[2828]: E0213 19:12:05.654208 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.654380 kubelet[2828]: E0213 19:12:05.654336 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.654380 kubelet[2828]: W0213 19:12:05.654350 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.654715 kubelet[2828]: E0213 19:12:05.654638 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.655232 kubelet[2828]: E0213 19:12:05.655129 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.655383 kubelet[2828]: W0213 19:12:05.655314 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.655884 kubelet[2828]: E0213 19:12:05.655730 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.655884 kubelet[2828]: E0213 19:12:05.655738 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.655884 kubelet[2828]: W0213 19:12:05.655788 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.655884 kubelet[2828]: E0213 19:12:05.655862 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.656298 kubelet[2828]: E0213 19:12:05.656176 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.656298 kubelet[2828]: W0213 19:12:05.656195 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.656298 kubelet[2828]: E0213 19:12:05.656269 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.657289 kubelet[2828]: E0213 19:12:05.656823 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.657289 kubelet[2828]: W0213 19:12:05.656849 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.657289 kubelet[2828]: E0213 19:12:05.656880 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.657602 kubelet[2828]: E0213 19:12:05.657581 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.657688 kubelet[2828]: W0213 19:12:05.657675 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.657788 kubelet[2828]: E0213 19:12:05.657765 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.660555 kubelet[2828]: E0213 19:12:05.660163 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.660555 kubelet[2828]: W0213 19:12:05.660187 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.660555 kubelet[2828]: E0213 19:12:05.660341 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.660722 containerd[1503]: time="2025-02-13T19:12:05.660302658Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:12:05.660722 containerd[1503]: time="2025-02-13T19:12:05.660385220Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:12:05.660722 containerd[1503]: time="2025-02-13T19:12:05.660403420Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:12:05.660722 containerd[1503]: time="2025-02-13T19:12:05.660492303Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:12:05.660870 kubelet[2828]: E0213 19:12:05.660729 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.660870 kubelet[2828]: W0213 19:12:05.660742 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.660870 kubelet[2828]: E0213 19:12:05.660846 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.662153 kubelet[2828]: E0213 19:12:05.661133 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.662153 kubelet[2828]: W0213 19:12:05.661176 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.662153 kubelet[2828]: E0213 19:12:05.661317 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.662153 kubelet[2828]: E0213 19:12:05.661681 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.662153 kubelet[2828]: W0213 19:12:05.661694 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.662153 kubelet[2828]: E0213 19:12:05.661728 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.662414 kubelet[2828]: E0213 19:12:05.662274 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.662414 kubelet[2828]: W0213 19:12:05.662388 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.662414 kubelet[2828]: E0213 19:12:05.662403 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.674531 kubelet[2828]: E0213 19:12:05.674502 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:05.675270 kubelet[2828]: W0213 19:12:05.674680 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:05.675563 kubelet[2828]: E0213 19:12:05.675454 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:05.694259 systemd[1]: Started cri-containerd-ba2393a86293acf25a04e42925632c5262b70843267e4ab409c90b5a6834ebfa.scope - libcontainer container ba2393a86293acf25a04e42925632c5262b70843267e4ab409c90b5a6834ebfa. Feb 13 19:12:05.735514 containerd[1503]: time="2025-02-13T19:12:05.735453531Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-8zbsj,Uid:3aa63002-df02-4bd7-a4be-bc77e95f3916,Namespace:calico-system,Attempt:0,} returns sandbox id \"ba2393a86293acf25a04e42925632c5262b70843267e4ab409c90b5a6834ebfa\"" Feb 13 19:12:07.173682 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1665015116.mount: Deactivated successfully. Feb 13 19:12:07.430414 kubelet[2828]: E0213 19:12:07.429213 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-6v9pr" podUID="b9c8452b-31d9-47bc-9697-dd4f293d71b9" Feb 13 19:12:07.607535 containerd[1503]: time="2025-02-13T19:12:07.606599614Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:12:07.609091 containerd[1503]: time="2025-02-13T19:12:07.609016361Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.29.1: active requests=0, bytes read=29231308" Feb 13 19:12:07.611188 containerd[1503]: time="2025-02-13T19:12:07.611123299Z" level=info msg="ImageCreate event name:\"sha256:1d1fc316829ae1650b0b1629b54232520f297e7c3b1444eecd290ae088902a28\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:12:07.613403 containerd[1503]: time="2025-02-13T19:12:07.612489737Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:768a194e1115c73bcbf35edb7afd18a63e16e08d940c79993565b6a3cca2da7c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:12:07.613403 containerd[1503]: time="2025-02-13T19:12:07.613201517Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.29.1\" with image id \"sha256:1d1fc316829ae1650b0b1629b54232520f297e7c3b1444eecd290ae088902a28\", repo tag \"ghcr.io/flatcar/calico/typha:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:768a194e1115c73bcbf35edb7afd18a63e16e08d940c79993565b6a3cca2da7c\", size \"29231162\" in 2.001916425s" Feb 13 19:12:07.613403 containerd[1503]: time="2025-02-13T19:12:07.613233878Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.29.1\" returns image reference \"sha256:1d1fc316829ae1650b0b1629b54232520f297e7c3b1444eecd290ae088902a28\"" Feb 13 19:12:07.616267 containerd[1503]: time="2025-02-13T19:12:07.616033275Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\"" Feb 13 19:12:07.636938 containerd[1503]: time="2025-02-13T19:12:07.636756250Z" level=info msg="CreateContainer within sandbox \"d4803efbd3f0d878e260db5abda856908d97888e81b4e77e0966aefbf1906381\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Feb 13 19:12:07.654764 containerd[1503]: time="2025-02-13T19:12:07.654713987Z" level=info msg="CreateContainer within sandbox \"d4803efbd3f0d878e260db5abda856908d97888e81b4e77e0966aefbf1906381\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"020db9e9a06d47298e9f368c32bb39cc4a59fc07ea9c83c5339669e2cab7e261\"" Feb 13 19:12:07.656341 containerd[1503]: time="2025-02-13T19:12:07.655930301Z" level=info msg="StartContainer for \"020db9e9a06d47298e9f368c32bb39cc4a59fc07ea9c83c5339669e2cab7e261\"" Feb 13 19:12:07.698573 systemd[1]: Started cri-containerd-020db9e9a06d47298e9f368c32bb39cc4a59fc07ea9c83c5339669e2cab7e261.scope - libcontainer container 020db9e9a06d47298e9f368c32bb39cc4a59fc07ea9c83c5339669e2cab7e261. Feb 13 19:12:07.737500 containerd[1503]: time="2025-02-13T19:12:07.737371439Z" level=info msg="StartContainer for \"020db9e9a06d47298e9f368c32bb39cc4a59fc07ea9c83c5339669e2cab7e261\" returns successfully" Feb 13 19:12:08.622510 kubelet[2828]: I0213 19:12:08.621878 2828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-7c86f976f-v6pf2" podStartSLOduration=1.61680768 podStartE2EDuration="3.62180811s" podCreationTimestamp="2025-02-13 19:12:05 +0000 UTC" firstStartedPulling="2025-02-13 19:12:05.610193502 +0000 UTC m=+31.308645180" lastFinishedPulling="2025-02-13 19:12:07.615193932 +0000 UTC m=+33.313645610" observedRunningTime="2025-02-13 19:12:08.602756136 +0000 UTC m=+34.301207854" watchObservedRunningTime="2025-02-13 19:12:08.62180811 +0000 UTC m=+34.320259788" Feb 13 19:12:08.654598 kubelet[2828]: E0213 19:12:08.654558 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:08.654598 kubelet[2828]: W0213 19:12:08.654587 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:08.654598 kubelet[2828]: E0213 19:12:08.654610 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:08.655760 kubelet[2828]: E0213 19:12:08.655733 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:08.655836 kubelet[2828]: W0213 19:12:08.655752 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:08.655836 kubelet[2828]: E0213 19:12:08.655805 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:08.656793 kubelet[2828]: E0213 19:12:08.656697 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:08.656793 kubelet[2828]: W0213 19:12:08.656793 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:08.656982 kubelet[2828]: E0213 19:12:08.656810 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:08.657249 kubelet[2828]: E0213 19:12:08.657229 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:08.657249 kubelet[2828]: W0213 19:12:08.657244 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:08.657393 kubelet[2828]: E0213 19:12:08.657256 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:08.657764 kubelet[2828]: E0213 19:12:08.657746 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:08.657764 kubelet[2828]: W0213 19:12:08.657759 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:08.657764 kubelet[2828]: E0213 19:12:08.657771 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:08.658230 kubelet[2828]: E0213 19:12:08.658111 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:08.658230 kubelet[2828]: W0213 19:12:08.658124 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:08.658230 kubelet[2828]: E0213 19:12:08.658136 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:08.658539 kubelet[2828]: E0213 19:12:08.658518 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:08.658539 kubelet[2828]: W0213 19:12:08.658531 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:08.658539 kubelet[2828]: E0213 19:12:08.658541 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:08.659040 kubelet[2828]: E0213 19:12:08.658908 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:08.659040 kubelet[2828]: W0213 19:12:08.658919 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:08.659040 kubelet[2828]: E0213 19:12:08.658928 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:08.659406 kubelet[2828]: E0213 19:12:08.659379 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:08.659406 kubelet[2828]: W0213 19:12:08.659394 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:08.659406 kubelet[2828]: E0213 19:12:08.659404 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:08.659765 kubelet[2828]: E0213 19:12:08.659748 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:08.659765 kubelet[2828]: W0213 19:12:08.659762 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:08.659851 kubelet[2828]: E0213 19:12:08.659772 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:08.660159 kubelet[2828]: E0213 19:12:08.660140 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:08.660159 kubelet[2828]: W0213 19:12:08.660156 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:08.660249 kubelet[2828]: E0213 19:12:08.660166 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:08.660726 kubelet[2828]: E0213 19:12:08.660707 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:08.660726 kubelet[2828]: W0213 19:12:08.660722 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:08.660844 kubelet[2828]: E0213 19:12:08.660733 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:08.661467 kubelet[2828]: E0213 19:12:08.661434 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:08.661467 kubelet[2828]: W0213 19:12:08.661457 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:08.661554 kubelet[2828]: E0213 19:12:08.661474 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:08.661885 kubelet[2828]: E0213 19:12:08.661865 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:08.661885 kubelet[2828]: W0213 19:12:08.661881 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:08.662022 kubelet[2828]: E0213 19:12:08.661892 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:08.662166 kubelet[2828]: E0213 19:12:08.662119 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:08.662166 kubelet[2828]: W0213 19:12:08.662128 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:08.662166 kubelet[2828]: E0213 19:12:08.662137 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:08.671638 kubelet[2828]: E0213 19:12:08.671041 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:08.671638 kubelet[2828]: W0213 19:12:08.671249 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:08.671638 kubelet[2828]: E0213 19:12:08.671282 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:08.672869 kubelet[2828]: E0213 19:12:08.672480 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:08.673283 kubelet[2828]: W0213 19:12:08.673076 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:08.673283 kubelet[2828]: E0213 19:12:08.673158 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:08.673950 kubelet[2828]: E0213 19:12:08.673606 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:08.673950 kubelet[2828]: W0213 19:12:08.673633 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:08.673950 kubelet[2828]: E0213 19:12:08.673653 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:08.674678 kubelet[2828]: E0213 19:12:08.674555 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:08.674678 kubelet[2828]: W0213 19:12:08.674572 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:08.675077 kubelet[2828]: E0213 19:12:08.675052 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:08.675415 kubelet[2828]: E0213 19:12:08.675282 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:08.675415 kubelet[2828]: W0213 19:12:08.675297 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:08.675415 kubelet[2828]: E0213 19:12:08.675330 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:08.675674 kubelet[2828]: E0213 19:12:08.675659 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:08.675858 kubelet[2828]: W0213 19:12:08.675743 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:08.675960 kubelet[2828]: E0213 19:12:08.675929 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:08.676421 kubelet[2828]: E0213 19:12:08.676248 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:08.676421 kubelet[2828]: W0213 19:12:08.676264 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:08.676529 kubelet[2828]: E0213 19:12:08.676301 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:08.676903 kubelet[2828]: E0213 19:12:08.676714 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:08.676903 kubelet[2828]: W0213 19:12:08.676730 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:08.676903 kubelet[2828]: E0213 19:12:08.676764 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:08.678200 kubelet[2828]: E0213 19:12:08.677976 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:08.678200 kubelet[2828]: W0213 19:12:08.677999 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:08.678200 kubelet[2828]: E0213 19:12:08.678031 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:08.678690 kubelet[2828]: E0213 19:12:08.678588 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:08.678690 kubelet[2828]: W0213 19:12:08.678612 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:08.678690 kubelet[2828]: E0213 19:12:08.678631 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:08.679411 kubelet[2828]: E0213 19:12:08.679387 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:08.679411 kubelet[2828]: W0213 19:12:08.679408 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:08.679677 kubelet[2828]: E0213 19:12:08.679431 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:08.680182 kubelet[2828]: E0213 19:12:08.680151 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:08.680182 kubelet[2828]: W0213 19:12:08.680175 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:08.680666 kubelet[2828]: E0213 19:12:08.680533 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:08.680666 kubelet[2828]: E0213 19:12:08.680563 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:08.680666 kubelet[2828]: W0213 19:12:08.680579 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:08.680817 kubelet[2828]: E0213 19:12:08.680774 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:08.680817 kubelet[2828]: E0213 19:12:08.680810 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:08.680942 kubelet[2828]: W0213 19:12:08.680819 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:08.680942 kubelet[2828]: E0213 19:12:08.680852 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:08.682024 kubelet[2828]: E0213 19:12:08.681103 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:08.682024 kubelet[2828]: W0213 19:12:08.681116 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:08.682024 kubelet[2828]: E0213 19:12:08.681134 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:08.682024 kubelet[2828]: E0213 19:12:08.681709 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:08.682024 kubelet[2828]: W0213 19:12:08.681743 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:08.682024 kubelet[2828]: E0213 19:12:08.681756 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:08.682024 kubelet[2828]: E0213 19:12:08.682023 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:08.682024 kubelet[2828]: W0213 19:12:08.682034 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:08.682354 kubelet[2828]: E0213 19:12:08.682046 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:08.682462 kubelet[2828]: E0213 19:12:08.682436 2828 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:12:08.682462 kubelet[2828]: W0213 19:12:08.682448 2828 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:12:08.682462 kubelet[2828]: E0213 19:12:08.682459 2828 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:12:09.128446 containerd[1503]: time="2025-02-13T19:12:09.127523687Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:12:09.128446 containerd[1503]: time="2025-02-13T19:12:09.128288749Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1: active requests=0, bytes read=5117811" Feb 13 19:12:09.129148 containerd[1503]: time="2025-02-13T19:12:09.129118332Z" level=info msg="ImageCreate event name:\"sha256:ece9bca32e64e726de8bbfc9e175a3ca91e0881cd40352bfcd1d107411f4f348\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:12:09.132017 containerd[1503]: time="2025-02-13T19:12:09.131982893Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:a63f8b4ff531912d12d143664eb263fdbc6cd7b3ff4aa777dfb6e318a090462c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:12:09.132827 containerd[1503]: time="2025-02-13T19:12:09.132796516Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\" with image id \"sha256:ece9bca32e64e726de8bbfc9e175a3ca91e0881cd40352bfcd1d107411f4f348\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:a63f8b4ff531912d12d143664eb263fdbc6cd7b3ff4aa777dfb6e318a090462c\", size \"6487425\" in 1.516672839s" Feb 13 19:12:09.132937 containerd[1503]: time="2025-02-13T19:12:09.132922440Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\" returns image reference \"sha256:ece9bca32e64e726de8bbfc9e175a3ca91e0881cd40352bfcd1d107411f4f348\"" Feb 13 19:12:09.135654 containerd[1503]: time="2025-02-13T19:12:09.135613036Z" level=info msg="CreateContainer within sandbox \"ba2393a86293acf25a04e42925632c5262b70843267e4ab409c90b5a6834ebfa\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Feb 13 19:12:09.154762 containerd[1503]: time="2025-02-13T19:12:09.154650736Z" level=info msg="CreateContainer within sandbox \"ba2393a86293acf25a04e42925632c5262b70843267e4ab409c90b5a6834ebfa\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"239c766f94c982e8f747c635e873206735db44749fde50a44ab81a086092f6d3\"" Feb 13 19:12:09.155609 containerd[1503]: time="2025-02-13T19:12:09.155578362Z" level=info msg="StartContainer for \"239c766f94c982e8f747c635e873206735db44749fde50a44ab81a086092f6d3\"" Feb 13 19:12:09.196285 systemd[1]: Started cri-containerd-239c766f94c982e8f747c635e873206735db44749fde50a44ab81a086092f6d3.scope - libcontainer container 239c766f94c982e8f747c635e873206735db44749fde50a44ab81a086092f6d3. Feb 13 19:12:09.241890 containerd[1503]: time="2025-02-13T19:12:09.241843927Z" level=info msg="StartContainer for \"239c766f94c982e8f747c635e873206735db44749fde50a44ab81a086092f6d3\" returns successfully" Feb 13 19:12:09.283254 systemd[1]: cri-containerd-239c766f94c982e8f747c635e873206735db44749fde50a44ab81a086092f6d3.scope: Deactivated successfully. Feb 13 19:12:09.400178 containerd[1503]: time="2025-02-13T19:12:09.399300590Z" level=info msg="shim disconnected" id=239c766f94c982e8f747c635e873206735db44749fde50a44ab81a086092f6d3 namespace=k8s.io Feb 13 19:12:09.400178 containerd[1503]: time="2025-02-13T19:12:09.399390712Z" level=warning msg="cleaning up after shim disconnected" id=239c766f94c982e8f747c635e873206735db44749fde50a44ab81a086092f6d3 namespace=k8s.io Feb 13 19:12:09.400178 containerd[1503]: time="2025-02-13T19:12:09.399400193Z" level=info msg="cleaning up dead shim" namespace=k8s.io Feb 13 19:12:09.429507 kubelet[2828]: E0213 19:12:09.429465 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-6v9pr" podUID="b9c8452b-31d9-47bc-9697-dd4f293d71b9" Feb 13 19:12:09.591677 containerd[1503]: time="2025-02-13T19:12:09.590415407Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.1\"" Feb 13 19:12:10.146919 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-239c766f94c982e8f747c635e873206735db44749fde50a44ab81a086092f6d3-rootfs.mount: Deactivated successfully. Feb 13 19:12:11.429922 kubelet[2828]: E0213 19:12:11.429871 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-6v9pr" podUID="b9c8452b-31d9-47bc-9697-dd4f293d71b9" Feb 13 19:12:12.383475 containerd[1503]: time="2025-02-13T19:12:12.383417340Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:12:12.384857 containerd[1503]: time="2025-02-13T19:12:12.384604055Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.29.1: active requests=0, bytes read=89703123" Feb 13 19:12:12.386106 containerd[1503]: time="2025-02-13T19:12:12.385900933Z" level=info msg="ImageCreate event name:\"sha256:e5ca62af4ff61b88f55fe4e0d7723151103d3f6a470fd4ebb311a2de27a9597f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:12:12.389091 containerd[1503]: time="2025-02-13T19:12:12.389023024Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:21e759d51c90dfb34fc1397dc180dd3a3fb564c2b0580d2f61ffe108f2a3c94b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:12:12.390342 containerd[1503]: time="2025-02-13T19:12:12.390154937Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.29.1\" with image id \"sha256:e5ca62af4ff61b88f55fe4e0d7723151103d3f6a470fd4ebb311a2de27a9597f\", repo tag \"ghcr.io/flatcar/calico/cni:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:21e759d51c90dfb34fc1397dc180dd3a3fb564c2b0580d2f61ffe108f2a3c94b\", size \"91072777\" in 2.798982549s" Feb 13 19:12:12.390342 containerd[1503]: time="2025-02-13T19:12:12.390197538Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.1\" returns image reference \"sha256:e5ca62af4ff61b88f55fe4e0d7723151103d3f6a470fd4ebb311a2de27a9597f\"" Feb 13 19:12:12.395171 containerd[1503]: time="2025-02-13T19:12:12.395113242Z" level=info msg="CreateContainer within sandbox \"ba2393a86293acf25a04e42925632c5262b70843267e4ab409c90b5a6834ebfa\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Feb 13 19:12:12.415748 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount770085438.mount: Deactivated successfully. Feb 13 19:12:12.420644 containerd[1503]: time="2025-02-13T19:12:12.420585226Z" level=info msg="CreateContainer within sandbox \"ba2393a86293acf25a04e42925632c5262b70843267e4ab409c90b5a6834ebfa\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"32b4eecba137996277e7d03529a85bd93a5587d743a010f6771d631bf13fa0a6\"" Feb 13 19:12:12.421927 containerd[1503]: time="2025-02-13T19:12:12.421193083Z" level=info msg="StartContainer for \"32b4eecba137996277e7d03529a85bd93a5587d743a010f6771d631bf13fa0a6\"" Feb 13 19:12:12.456913 systemd[1]: run-containerd-runc-k8s.io-32b4eecba137996277e7d03529a85bd93a5587d743a010f6771d631bf13fa0a6-runc.K54jtA.mount: Deactivated successfully. Feb 13 19:12:12.466342 systemd[1]: Started cri-containerd-32b4eecba137996277e7d03529a85bd93a5587d743a010f6771d631bf13fa0a6.scope - libcontainer container 32b4eecba137996277e7d03529a85bd93a5587d743a010f6771d631bf13fa0a6. Feb 13 19:12:12.503342 containerd[1503]: time="2025-02-13T19:12:12.503286881Z" level=info msg="StartContainer for \"32b4eecba137996277e7d03529a85bd93a5587d743a010f6771d631bf13fa0a6\" returns successfully" Feb 13 19:12:12.994570 containerd[1503]: time="2025-02-13T19:12:12.994525588Z" level=error msg="failed to reload cni configuration after receiving fs change event(WRITE \"/etc/cni/net.d/calico-kubeconfig\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Feb 13 19:12:12.996819 systemd[1]: cri-containerd-32b4eecba137996277e7d03529a85bd93a5587d743a010f6771d631bf13fa0a6.scope: Deactivated successfully. Feb 13 19:12:12.997251 systemd[1]: cri-containerd-32b4eecba137996277e7d03529a85bd93a5587d743a010f6771d631bf13fa0a6.scope: Consumed 471ms CPU time, 166.2M memory peak, 147.4M written to disk. Feb 13 19:12:13.091129 kubelet[2828]: I0213 19:12:13.091093 2828 kubelet_node_status.go:502] "Fast updating node status as it just became ready" Feb 13 19:12:13.138938 containerd[1503]: time="2025-02-13T19:12:13.138437749Z" level=info msg="shim disconnected" id=32b4eecba137996277e7d03529a85bd93a5587d743a010f6771d631bf13fa0a6 namespace=k8s.io Feb 13 19:12:13.138938 containerd[1503]: time="2025-02-13T19:12:13.138526391Z" level=warning msg="cleaning up after shim disconnected" id=32b4eecba137996277e7d03529a85bd93a5587d743a010f6771d631bf13fa0a6 namespace=k8s.io Feb 13 19:12:13.138938 containerd[1503]: time="2025-02-13T19:12:13.138606234Z" level=info msg="cleaning up dead shim" namespace=k8s.io Feb 13 19:12:13.166729 systemd[1]: Created slice kubepods-besteffort-pod05e39f6f_0a83_4b62_8cff_b0d4799f8f89.slice - libcontainer container kubepods-besteffort-pod05e39f6f_0a83_4b62_8cff_b0d4799f8f89.slice. Feb 13 19:12:13.182580 systemd[1]: Created slice kubepods-besteffort-pod96785f8f_b4e3_456b_9a0d_bcc142e6beaf.slice - libcontainer container kubepods-besteffort-pod96785f8f_b4e3_456b_9a0d_bcc142e6beaf.slice. Feb 13 19:12:13.195218 systemd[1]: Created slice kubepods-burstable-pod5cf24d69_eb75_4cb3_8437_024193f516f5.slice - libcontainer container kubepods-burstable-pod5cf24d69_eb75_4cb3_8437_024193f516f5.slice. Feb 13 19:12:13.202892 systemd[1]: Created slice kubepods-burstable-podeb0c3c87_077a_4b2b_a168_5fd9131bccd3.slice - libcontainer container kubepods-burstable-podeb0c3c87_077a_4b2b_a168_5fd9131bccd3.slice. Feb 13 19:12:13.209406 systemd[1]: Created slice kubepods-besteffort-pod436348d7_7832_43e1_8780_623c434b2666.slice - libcontainer container kubepods-besteffort-pod436348d7_7832_43e1_8780_623c434b2666.slice. Feb 13 19:12:13.211732 kubelet[2828]: I0213 19:12:13.211449 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eb0c3c87-077a-4b2b-a168-5fd9131bccd3-config-volume\") pod \"coredns-668d6bf9bc-vx9cw\" (UID: \"eb0c3c87-077a-4b2b-a168-5fd9131bccd3\") " pod="kube-system/coredns-668d6bf9bc-vx9cw" Feb 13 19:12:13.211732 kubelet[2828]: I0213 19:12:13.211489 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgwvn\" (UniqueName: \"kubernetes.io/projected/eb0c3c87-077a-4b2b-a168-5fd9131bccd3-kube-api-access-hgwvn\") pod \"coredns-668d6bf9bc-vx9cw\" (UID: \"eb0c3c87-077a-4b2b-a168-5fd9131bccd3\") " pod="kube-system/coredns-668d6bf9bc-vx9cw" Feb 13 19:12:13.211732 kubelet[2828]: I0213 19:12:13.211515 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x27wd\" (UniqueName: \"kubernetes.io/projected/96785f8f-b4e3-456b-9a0d-bcc142e6beaf-kube-api-access-x27wd\") pod \"calico-apiserver-84956f4cf-72cgp\" (UID: \"96785f8f-b4e3-456b-9a0d-bcc142e6beaf\") " pod="calico-apiserver/calico-apiserver-84956f4cf-72cgp" Feb 13 19:12:13.211732 kubelet[2828]: I0213 19:12:13.211537 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nksp\" (UniqueName: \"kubernetes.io/projected/05e39f6f-0a83-4b62-8cff-b0d4799f8f89-kube-api-access-6nksp\") pod \"calico-apiserver-84956f4cf-p6mr2\" (UID: \"05e39f6f-0a83-4b62-8cff-b0d4799f8f89\") " pod="calico-apiserver/calico-apiserver-84956f4cf-p6mr2" Feb 13 19:12:13.211732 kubelet[2828]: I0213 19:12:13.211557 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5cf24d69-eb75-4cb3-8437-024193f516f5-config-volume\") pod \"coredns-668d6bf9bc-kkqg5\" (UID: \"5cf24d69-eb75-4cb3-8437-024193f516f5\") " pod="kube-system/coredns-668d6bf9bc-kkqg5" Feb 13 19:12:13.211901 kubelet[2828]: I0213 19:12:13.211572 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/05e39f6f-0a83-4b62-8cff-b0d4799f8f89-calico-apiserver-certs\") pod \"calico-apiserver-84956f4cf-p6mr2\" (UID: \"05e39f6f-0a83-4b62-8cff-b0d4799f8f89\") " pod="calico-apiserver/calico-apiserver-84956f4cf-p6mr2" Feb 13 19:12:13.211901 kubelet[2828]: I0213 19:12:13.211588 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/96785f8f-b4e3-456b-9a0d-bcc142e6beaf-calico-apiserver-certs\") pod \"calico-apiserver-84956f4cf-72cgp\" (UID: \"96785f8f-b4e3-456b-9a0d-bcc142e6beaf\") " pod="calico-apiserver/calico-apiserver-84956f4cf-72cgp" Feb 13 19:12:13.211901 kubelet[2828]: I0213 19:12:13.211607 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2cbd\" (UniqueName: \"kubernetes.io/projected/5cf24d69-eb75-4cb3-8437-024193f516f5-kube-api-access-t2cbd\") pod \"coredns-668d6bf9bc-kkqg5\" (UID: \"5cf24d69-eb75-4cb3-8437-024193f516f5\") " pod="kube-system/coredns-668d6bf9bc-kkqg5" Feb 13 19:12:13.211901 kubelet[2828]: I0213 19:12:13.211628 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7bkw\" (UniqueName: \"kubernetes.io/projected/436348d7-7832-43e1-8780-623c434b2666-kube-api-access-m7bkw\") pod \"calico-kube-controllers-78bd6d67c9-b259q\" (UID: \"436348d7-7832-43e1-8780-623c434b2666\") " pod="calico-system/calico-kube-controllers-78bd6d67c9-b259q" Feb 13 19:12:13.211901 kubelet[2828]: I0213 19:12:13.211648 2828 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/436348d7-7832-43e1-8780-623c434b2666-tigera-ca-bundle\") pod \"calico-kube-controllers-78bd6d67c9-b259q\" (UID: \"436348d7-7832-43e1-8780-623c434b2666\") " pod="calico-system/calico-kube-controllers-78bd6d67c9-b259q" Feb 13 19:12:13.419948 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-32b4eecba137996277e7d03529a85bd93a5587d743a010f6771d631bf13fa0a6-rootfs.mount: Deactivated successfully. Feb 13 19:12:13.437774 systemd[1]: Created slice kubepods-besteffort-podb9c8452b_31d9_47bc_9697_dd4f293d71b9.slice - libcontainer container kubepods-besteffort-podb9c8452b_31d9_47bc_9697_dd4f293d71b9.slice. Feb 13 19:12:13.443118 containerd[1503]: time="2025-02-13T19:12:13.442970885Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-6v9pr,Uid:b9c8452b-31d9-47bc-9697-dd4f293d71b9,Namespace:calico-system,Attempt:0,}" Feb 13 19:12:13.475350 containerd[1503]: time="2025-02-13T19:12:13.475287838Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-84956f4cf-p6mr2,Uid:05e39f6f-0a83-4b62-8cff-b0d4799f8f89,Namespace:calico-apiserver,Attempt:0,}" Feb 13 19:12:13.496248 containerd[1503]: time="2025-02-13T19:12:13.495870165Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-84956f4cf-72cgp,Uid:96785f8f-b4e3-456b-9a0d-bcc142e6beaf,Namespace:calico-apiserver,Attempt:0,}" Feb 13 19:12:13.515282 containerd[1503]: time="2025-02-13T19:12:13.515218455Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-78bd6d67c9-b259q,Uid:436348d7-7832-43e1-8780-623c434b2666,Namespace:calico-system,Attempt:0,}" Feb 13 19:12:13.518368 containerd[1503]: time="2025-02-13T19:12:13.516824862Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-kkqg5,Uid:5cf24d69-eb75-4cb3-8437-024193f516f5,Namespace:kube-system,Attempt:0,}" Feb 13 19:12:13.519216 systemd[1]: Started sshd@13-78.46.147.231:22-159.223.215.175:45010.service - OpenSSH per-connection server daemon (159.223.215.175:45010). Feb 13 19:12:13.521478 containerd[1503]: time="2025-02-13T19:12:13.516862344Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-vx9cw,Uid:eb0c3c87-077a-4b2b-a168-5fd9131bccd3,Namespace:kube-system,Attempt:0,}" Feb 13 19:12:13.636320 containerd[1503]: time="2025-02-13T19:12:13.635970534Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.1\"" Feb 13 19:12:13.718767 sshd[3616]: Invalid user testxp from 159.223.215.175 port 45010 Feb 13 19:12:13.750017 sshd[3616]: Received disconnect from 159.223.215.175 port 45010:11: Bye Bye [preauth] Feb 13 19:12:13.750017 sshd[3616]: Disconnected from invalid user testxp 159.223.215.175 port 45010 [preauth] Feb 13 19:12:13.754037 systemd[1]: sshd@13-78.46.147.231:22-159.223.215.175:45010.service: Deactivated successfully. Feb 13 19:12:13.790276 containerd[1503]: time="2025-02-13T19:12:13.789978914Z" level=error msg="Failed to destroy network for sandbox \"1f6b42d2735cedb916bc24bf2bb557d2ff72928d0c487a93a3dd46e88aa55abe\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:13.791634 containerd[1503]: time="2025-02-13T19:12:13.791552281Z" level=error msg="encountered an error cleaning up failed sandbox \"1f6b42d2735cedb916bc24bf2bb557d2ff72928d0c487a93a3dd46e88aa55abe\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:13.791850 containerd[1503]: time="2025-02-13T19:12:13.791665364Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-6v9pr,Uid:b9c8452b-31d9-47bc-9697-dd4f293d71b9,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"1f6b42d2735cedb916bc24bf2bb557d2ff72928d0c487a93a3dd46e88aa55abe\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:13.793125 kubelet[2828]: E0213 19:12:13.792345 2828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1f6b42d2735cedb916bc24bf2bb557d2ff72928d0c487a93a3dd46e88aa55abe\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:13.793125 kubelet[2828]: E0213 19:12:13.792466 2828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1f6b42d2735cedb916bc24bf2bb557d2ff72928d0c487a93a3dd46e88aa55abe\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-6v9pr" Feb 13 19:12:13.793125 kubelet[2828]: E0213 19:12:13.792488 2828 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1f6b42d2735cedb916bc24bf2bb557d2ff72928d0c487a93a3dd46e88aa55abe\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-6v9pr" Feb 13 19:12:13.793273 kubelet[2828]: E0213 19:12:13.792533 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-6v9pr_calico-system(b9c8452b-31d9-47bc-9697-dd4f293d71b9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-6v9pr_calico-system(b9c8452b-31d9-47bc-9697-dd4f293d71b9)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"1f6b42d2735cedb916bc24bf2bb557d2ff72928d0c487a93a3dd46e88aa55abe\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-6v9pr" podUID="b9c8452b-31d9-47bc-9697-dd4f293d71b9" Feb 13 19:12:13.836286 containerd[1503]: time="2025-02-13T19:12:13.835953709Z" level=error msg="Failed to destroy network for sandbox \"0c13c2a989169406cb4cea286d09a0b055b98c8bf7adaf3c864364728278a3a5\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:13.840659 containerd[1503]: time="2025-02-13T19:12:13.840581326Z" level=error msg="encountered an error cleaning up failed sandbox \"0c13c2a989169406cb4cea286d09a0b055b98c8bf7adaf3c864364728278a3a5\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:13.842588 containerd[1503]: time="2025-02-13T19:12:13.842303497Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-84956f4cf-p6mr2,Uid:05e39f6f-0a83-4b62-8cff-b0d4799f8f89,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"0c13c2a989169406cb4cea286d09a0b055b98c8bf7adaf3c864364728278a3a5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:13.844731 kubelet[2828]: E0213 19:12:13.843973 2828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0c13c2a989169406cb4cea286d09a0b055b98c8bf7adaf3c864364728278a3a5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:13.844731 kubelet[2828]: E0213 19:12:13.844037 2828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0c13c2a989169406cb4cea286d09a0b055b98c8bf7adaf3c864364728278a3a5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-84956f4cf-p6mr2" Feb 13 19:12:13.844731 kubelet[2828]: E0213 19:12:13.844072 2828 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0c13c2a989169406cb4cea286d09a0b055b98c8bf7adaf3c864364728278a3a5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-84956f4cf-p6mr2" Feb 13 19:12:13.844913 kubelet[2828]: E0213 19:12:13.844111 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-84956f4cf-p6mr2_calico-apiserver(05e39f6f-0a83-4b62-8cff-b0d4799f8f89)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-84956f4cf-p6mr2_calico-apiserver(05e39f6f-0a83-4b62-8cff-b0d4799f8f89)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"0c13c2a989169406cb4cea286d09a0b055b98c8bf7adaf3c864364728278a3a5\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-84956f4cf-p6mr2" podUID="05e39f6f-0a83-4b62-8cff-b0d4799f8f89" Feb 13 19:12:13.850036 containerd[1503]: time="2025-02-13T19:12:13.849593391Z" level=error msg="Failed to destroy network for sandbox \"0c31ed484cb1e7593659d8fd1f850350d6b3b9e01ed8530e08cf66bc47a72b73\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:13.850848 containerd[1503]: time="2025-02-13T19:12:13.850799587Z" level=error msg="encountered an error cleaning up failed sandbox \"0c31ed484cb1e7593659d8fd1f850350d6b3b9e01ed8530e08cf66bc47a72b73\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:13.852368 containerd[1503]: time="2025-02-13T19:12:13.851822217Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-vx9cw,Uid:eb0c3c87-077a-4b2b-a168-5fd9131bccd3,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"0c31ed484cb1e7593659d8fd1f850350d6b3b9e01ed8530e08cf66bc47a72b73\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:13.853731 kubelet[2828]: E0213 19:12:13.853598 2828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0c31ed484cb1e7593659d8fd1f850350d6b3b9e01ed8530e08cf66bc47a72b73\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:13.853731 kubelet[2828]: E0213 19:12:13.853663 2828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0c31ed484cb1e7593659d8fd1f850350d6b3b9e01ed8530e08cf66bc47a72b73\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-vx9cw" Feb 13 19:12:13.853731 kubelet[2828]: E0213 19:12:13.853682 2828 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0c31ed484cb1e7593659d8fd1f850350d6b3b9e01ed8530e08cf66bc47a72b73\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-vx9cw" Feb 13 19:12:13.855220 kubelet[2828]: E0213 19:12:13.853731 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-vx9cw_kube-system(eb0c3c87-077a-4b2b-a168-5fd9131bccd3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-vx9cw_kube-system(eb0c3c87-077a-4b2b-a168-5fd9131bccd3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"0c31ed484cb1e7593659d8fd1f850350d6b3b9e01ed8530e08cf66bc47a72b73\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-vx9cw" podUID="eb0c3c87-077a-4b2b-a168-5fd9131bccd3" Feb 13 19:12:13.886841 containerd[1503]: time="2025-02-13T19:12:13.886396476Z" level=error msg="Failed to destroy network for sandbox \"930563c746d8b91d524a61a614694c47905765a47fb87cb32a681d5b35060a78\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:13.887893 containerd[1503]: time="2025-02-13T19:12:13.887854519Z" level=error msg="Failed to destroy network for sandbox \"fa6e4b9a2b53f8e5e8150bdfa0fe88ef1d51405b9b5e7e6d9683c6adb0b078ab\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:13.888250 containerd[1503]: time="2025-02-13T19:12:13.888142808Z" level=error msg="encountered an error cleaning up failed sandbox \"930563c746d8b91d524a61a614694c47905765a47fb87cb32a681d5b35060a78\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:13.888346 containerd[1503]: time="2025-02-13T19:12:13.888229450Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-84956f4cf-72cgp,Uid:96785f8f-b4e3-456b-9a0d-bcc142e6beaf,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"930563c746d8b91d524a61a614694c47905765a47fb87cb32a681d5b35060a78\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:13.888581 containerd[1503]: time="2025-02-13T19:12:13.888262211Z" level=error msg="encountered an error cleaning up failed sandbox \"fa6e4b9a2b53f8e5e8150bdfa0fe88ef1d51405b9b5e7e6d9683c6adb0b078ab\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:13.888640 containerd[1503]: time="2025-02-13T19:12:13.888592501Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-78bd6d67c9-b259q,Uid:436348d7-7832-43e1-8780-623c434b2666,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"fa6e4b9a2b53f8e5e8150bdfa0fe88ef1d51405b9b5e7e6d9683c6adb0b078ab\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:13.888988 kubelet[2828]: E0213 19:12:13.888908 2828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fa6e4b9a2b53f8e5e8150bdfa0fe88ef1d51405b9b5e7e6d9683c6adb0b078ab\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:13.888988 kubelet[2828]: E0213 19:12:13.888922 2828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"930563c746d8b91d524a61a614694c47905765a47fb87cb32a681d5b35060a78\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:13.888988 kubelet[2828]: E0213 19:12:13.888971 2828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fa6e4b9a2b53f8e5e8150bdfa0fe88ef1d51405b9b5e7e6d9683c6adb0b078ab\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-78bd6d67c9-b259q" Feb 13 19:12:13.890190 kubelet[2828]: E0213 19:12:13.888991 2828 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fa6e4b9a2b53f8e5e8150bdfa0fe88ef1d51405b9b5e7e6d9683c6adb0b078ab\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-78bd6d67c9-b259q" Feb 13 19:12:13.890190 kubelet[2828]: E0213 19:12:13.889034 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-78bd6d67c9-b259q_calico-system(436348d7-7832-43e1-8780-623c434b2666)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-78bd6d67c9-b259q_calico-system(436348d7-7832-43e1-8780-623c434b2666)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"fa6e4b9a2b53f8e5e8150bdfa0fe88ef1d51405b9b5e7e6d9683c6adb0b078ab\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-78bd6d67c9-b259q" podUID="436348d7-7832-43e1-8780-623c434b2666" Feb 13 19:12:13.890190 kubelet[2828]: E0213 19:12:13.889097 2828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"930563c746d8b91d524a61a614694c47905765a47fb87cb32a681d5b35060a78\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-84956f4cf-72cgp" Feb 13 19:12:13.890315 kubelet[2828]: E0213 19:12:13.890205 2828 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"930563c746d8b91d524a61a614694c47905765a47fb87cb32a681d5b35060a78\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-84956f4cf-72cgp" Feb 13 19:12:13.890315 kubelet[2828]: E0213 19:12:13.890277 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-84956f4cf-72cgp_calico-apiserver(96785f8f-b4e3-456b-9a0d-bcc142e6beaf)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-84956f4cf-72cgp_calico-apiserver(96785f8f-b4e3-456b-9a0d-bcc142e6beaf)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"930563c746d8b91d524a61a614694c47905765a47fb87cb32a681d5b35060a78\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-84956f4cf-72cgp" podUID="96785f8f-b4e3-456b-9a0d-bcc142e6beaf" Feb 13 19:12:13.893095 containerd[1503]: time="2025-02-13T19:12:13.892615980Z" level=error msg="Failed to destroy network for sandbox \"43a5af5ee3597e852f8d0ba977ffb06562d6b40688d779a95c8d495a78df4c29\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:13.893629 containerd[1503]: time="2025-02-13T19:12:13.893582208Z" level=error msg="encountered an error cleaning up failed sandbox \"43a5af5ee3597e852f8d0ba977ffb06562d6b40688d779a95c8d495a78df4c29\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:13.893677 containerd[1503]: time="2025-02-13T19:12:13.893658770Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-kkqg5,Uid:5cf24d69-eb75-4cb3-8437-024193f516f5,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"43a5af5ee3597e852f8d0ba977ffb06562d6b40688d779a95c8d495a78df4c29\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:13.894688 kubelet[2828]: E0213 19:12:13.894625 2828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"43a5af5ee3597e852f8d0ba977ffb06562d6b40688d779a95c8d495a78df4c29\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:13.895003 kubelet[2828]: E0213 19:12:13.894838 2828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"43a5af5ee3597e852f8d0ba977ffb06562d6b40688d779a95c8d495a78df4c29\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-kkqg5" Feb 13 19:12:13.895003 kubelet[2828]: E0213 19:12:13.894866 2828 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"43a5af5ee3597e852f8d0ba977ffb06562d6b40688d779a95c8d495a78df4c29\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-kkqg5" Feb 13 19:12:13.895003 kubelet[2828]: E0213 19:12:13.894951 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-kkqg5_kube-system(5cf24d69-eb75-4cb3-8437-024193f516f5)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-kkqg5_kube-system(5cf24d69-eb75-4cb3-8437-024193f516f5)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"43a5af5ee3597e852f8d0ba977ffb06562d6b40688d779a95c8d495a78df4c29\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-kkqg5" podUID="5cf24d69-eb75-4cb3-8437-024193f516f5" Feb 13 19:12:14.412712 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-0c13c2a989169406cb4cea286d09a0b055b98c8bf7adaf3c864364728278a3a5-shm.mount: Deactivated successfully. Feb 13 19:12:14.412808 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-1f6b42d2735cedb916bc24bf2bb557d2ff72928d0c487a93a3dd46e88aa55abe-shm.mount: Deactivated successfully. Feb 13 19:12:14.637201 kubelet[2828]: I0213 19:12:14.637164 2828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c31ed484cb1e7593659d8fd1f850350d6b3b9e01ed8530e08cf66bc47a72b73" Feb 13 19:12:14.638412 containerd[1503]: time="2025-02-13T19:12:14.638365569Z" level=info msg="StopPodSandbox for \"0c31ed484cb1e7593659d8fd1f850350d6b3b9e01ed8530e08cf66bc47a72b73\"" Feb 13 19:12:14.638904 containerd[1503]: time="2025-02-13T19:12:14.638581375Z" level=info msg="Ensure that sandbox 0c31ed484cb1e7593659d8fd1f850350d6b3b9e01ed8530e08cf66bc47a72b73 in task-service has been cleanup successfully" Feb 13 19:12:14.641171 containerd[1503]: time="2025-02-13T19:12:14.641069049Z" level=info msg="TearDown network for sandbox \"0c31ed484cb1e7593659d8fd1f850350d6b3b9e01ed8530e08cf66bc47a72b73\" successfully" Feb 13 19:12:14.641171 containerd[1503]: time="2025-02-13T19:12:14.641104250Z" level=info msg="StopPodSandbox for \"0c31ed484cb1e7593659d8fd1f850350d6b3b9e01ed8530e08cf66bc47a72b73\" returns successfully" Feb 13 19:12:14.641974 containerd[1503]: time="2025-02-13T19:12:14.641935435Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-vx9cw,Uid:eb0c3c87-077a-4b2b-a168-5fd9131bccd3,Namespace:kube-system,Attempt:1,}" Feb 13 19:12:14.644413 kubelet[2828]: I0213 19:12:14.644051 2828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f6b42d2735cedb916bc24bf2bb557d2ff72928d0c487a93a3dd46e88aa55abe" Feb 13 19:12:14.644692 systemd[1]: run-netns-cni\x2db422f372\x2d8cb0\x2d062e\x2dd1b9\x2d05100f6c861c.mount: Deactivated successfully. Feb 13 19:12:14.647513 containerd[1503]: time="2025-02-13T19:12:14.645971555Z" level=info msg="StopPodSandbox for \"1f6b42d2735cedb916bc24bf2bb557d2ff72928d0c487a93a3dd46e88aa55abe\"" Feb 13 19:12:14.647513 containerd[1503]: time="2025-02-13T19:12:14.646516691Z" level=info msg="Ensure that sandbox 1f6b42d2735cedb916bc24bf2bb557d2ff72928d0c487a93a3dd46e88aa55abe in task-service has been cleanup successfully" Feb 13 19:12:14.649567 containerd[1503]: time="2025-02-13T19:12:14.649534381Z" level=info msg="TearDown network for sandbox \"1f6b42d2735cedb916bc24bf2bb557d2ff72928d0c487a93a3dd46e88aa55abe\" successfully" Feb 13 19:12:14.650077 containerd[1503]: time="2025-02-13T19:12:14.649991954Z" level=info msg="StopPodSandbox for \"1f6b42d2735cedb916bc24bf2bb557d2ff72928d0c487a93a3dd46e88aa55abe\" returns successfully" Feb 13 19:12:14.650442 systemd[1]: run-netns-cni\x2d6fca69e1\x2d3a7f\x2dcb84\x2dd936\x2d76f183d4dc8f.mount: Deactivated successfully. Feb 13 19:12:14.652146 kubelet[2828]: I0213 19:12:14.652104 2828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43a5af5ee3597e852f8d0ba977ffb06562d6b40688d779a95c8d495a78df4c29" Feb 13 19:12:14.652581 containerd[1503]: time="2025-02-13T19:12:14.652541630Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-6v9pr,Uid:b9c8452b-31d9-47bc-9697-dd4f293d71b9,Namespace:calico-system,Attempt:1,}" Feb 13 19:12:14.655104 containerd[1503]: time="2025-02-13T19:12:14.654327763Z" level=info msg="StopPodSandbox for \"43a5af5ee3597e852f8d0ba977ffb06562d6b40688d779a95c8d495a78df4c29\"" Feb 13 19:12:14.655460 containerd[1503]: time="2025-02-13T19:12:14.655256591Z" level=info msg="Ensure that sandbox 43a5af5ee3597e852f8d0ba977ffb06562d6b40688d779a95c8d495a78df4c29 in task-service has been cleanup successfully" Feb 13 19:12:14.659974 containerd[1503]: time="2025-02-13T19:12:14.659909009Z" level=info msg="TearDown network for sandbox \"43a5af5ee3597e852f8d0ba977ffb06562d6b40688d779a95c8d495a78df4c29\" successfully" Feb 13 19:12:14.660103 containerd[1503]: time="2025-02-13T19:12:14.659964011Z" level=info msg="StopPodSandbox for \"43a5af5ee3597e852f8d0ba977ffb06562d6b40688d779a95c8d495a78df4c29\" returns successfully" Feb 13 19:12:14.662154 containerd[1503]: time="2025-02-13T19:12:14.660873798Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-kkqg5,Uid:5cf24d69-eb75-4cb3-8437-024193f516f5,Namespace:kube-system,Attempt:1,}" Feb 13 19:12:14.663376 systemd[1]: run-netns-cni\x2d02442949\x2d915d\x2dc16e\x2dfda5\x2dc56041b32030.mount: Deactivated successfully. Feb 13 19:12:14.666309 kubelet[2828]: I0213 19:12:14.664879 2828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa6e4b9a2b53f8e5e8150bdfa0fe88ef1d51405b9b5e7e6d9683c6adb0b078ab" Feb 13 19:12:14.669091 containerd[1503]: time="2025-02-13T19:12:14.668531786Z" level=info msg="StopPodSandbox for \"fa6e4b9a2b53f8e5e8150bdfa0fe88ef1d51405b9b5e7e6d9683c6adb0b078ab\"" Feb 13 19:12:14.669091 containerd[1503]: time="2025-02-13T19:12:14.668795514Z" level=info msg="Ensure that sandbox fa6e4b9a2b53f8e5e8150bdfa0fe88ef1d51405b9b5e7e6d9683c6adb0b078ab in task-service has been cleanup successfully" Feb 13 19:12:14.670364 containerd[1503]: time="2025-02-13T19:12:14.670324559Z" level=info msg="TearDown network for sandbox \"fa6e4b9a2b53f8e5e8150bdfa0fe88ef1d51405b9b5e7e6d9683c6adb0b078ab\" successfully" Feb 13 19:12:14.670364 containerd[1503]: time="2025-02-13T19:12:14.670353120Z" level=info msg="StopPodSandbox for \"fa6e4b9a2b53f8e5e8150bdfa0fe88ef1d51405b9b5e7e6d9683c6adb0b078ab\" returns successfully" Feb 13 19:12:14.674074 containerd[1503]: time="2025-02-13T19:12:14.673335969Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-78bd6d67c9-b259q,Uid:436348d7-7832-43e1-8780-623c434b2666,Namespace:calico-system,Attempt:1,}" Feb 13 19:12:14.674151 kubelet[2828]: I0213 19:12:14.673900 2828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="930563c746d8b91d524a61a614694c47905765a47fb87cb32a681d5b35060a78" Feb 13 19:12:14.674653 containerd[1503]: time="2025-02-13T19:12:14.674603646Z" level=info msg="StopPodSandbox for \"930563c746d8b91d524a61a614694c47905765a47fb87cb32a681d5b35060a78\"" Feb 13 19:12:14.676249 containerd[1503]: time="2025-02-13T19:12:14.674765051Z" level=info msg="Ensure that sandbox 930563c746d8b91d524a61a614694c47905765a47fb87cb32a681d5b35060a78 in task-service has been cleanup successfully" Feb 13 19:12:14.676249 containerd[1503]: time="2025-02-13T19:12:14.674943256Z" level=info msg="TearDown network for sandbox \"930563c746d8b91d524a61a614694c47905765a47fb87cb32a681d5b35060a78\" successfully" Feb 13 19:12:14.676249 containerd[1503]: time="2025-02-13T19:12:14.674958857Z" level=info msg="StopPodSandbox for \"930563c746d8b91d524a61a614694c47905765a47fb87cb32a681d5b35060a78\" returns successfully" Feb 13 19:12:14.676249 containerd[1503]: time="2025-02-13T19:12:14.675620397Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-84956f4cf-72cgp,Uid:96785f8f-b4e3-456b-9a0d-bcc142e6beaf,Namespace:calico-apiserver,Attempt:1,}" Feb 13 19:12:14.676311 systemd[1]: run-netns-cni\x2d8fa3169e\x2dfe61\x2d71f4\x2d5c7c\x2d7970373fed49.mount: Deactivated successfully. Feb 13 19:12:14.678073 kubelet[2828]: I0213 19:12:14.677725 2828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c13c2a989169406cb4cea286d09a0b055b98c8bf7adaf3c864364728278a3a5" Feb 13 19:12:14.679377 containerd[1503]: time="2025-02-13T19:12:14.679345867Z" level=info msg="StopPodSandbox for \"0c13c2a989169406cb4cea286d09a0b055b98c8bf7adaf3c864364728278a3a5\"" Feb 13 19:12:14.679927 containerd[1503]: time="2025-02-13T19:12:14.679787681Z" level=info msg="Ensure that sandbox 0c13c2a989169406cb4cea286d09a0b055b98c8bf7adaf3c864364728278a3a5 in task-service has been cleanup successfully" Feb 13 19:12:14.680201 containerd[1503]: time="2025-02-13T19:12:14.680088089Z" level=info msg="TearDown network for sandbox \"0c13c2a989169406cb4cea286d09a0b055b98c8bf7adaf3c864364728278a3a5\" successfully" Feb 13 19:12:14.680201 containerd[1503]: time="2025-02-13T19:12:14.680107650Z" level=info msg="StopPodSandbox for \"0c13c2a989169406cb4cea286d09a0b055b98c8bf7adaf3c864364728278a3a5\" returns successfully" Feb 13 19:12:14.681679 containerd[1503]: time="2025-02-13T19:12:14.681615775Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-84956f4cf-p6mr2,Uid:05e39f6f-0a83-4b62-8cff-b0d4799f8f89,Namespace:calico-apiserver,Attempt:1,}" Feb 13 19:12:14.811755 containerd[1503]: time="2025-02-13T19:12:14.811193148Z" level=error msg="Failed to destroy network for sandbox \"07aed5c023700866ebeb2a6224ac11566918d60f5a140685f482d3a85f4cd555\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:14.811755 containerd[1503]: time="2025-02-13T19:12:14.811566119Z" level=error msg="encountered an error cleaning up failed sandbox \"07aed5c023700866ebeb2a6224ac11566918d60f5a140685f482d3a85f4cd555\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:14.811755 containerd[1503]: time="2025-02-13T19:12:14.811636322Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-vx9cw,Uid:eb0c3c87-077a-4b2b-a168-5fd9131bccd3,Namespace:kube-system,Attempt:1,} failed, error" error="failed to setup network for sandbox \"07aed5c023700866ebeb2a6224ac11566918d60f5a140685f482d3a85f4cd555\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:14.811981 kubelet[2828]: E0213 19:12:14.811917 2828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"07aed5c023700866ebeb2a6224ac11566918d60f5a140685f482d3a85f4cd555\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:14.811981 kubelet[2828]: E0213 19:12:14.811972 2828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"07aed5c023700866ebeb2a6224ac11566918d60f5a140685f482d3a85f4cd555\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-vx9cw" Feb 13 19:12:14.812042 kubelet[2828]: E0213 19:12:14.811992 2828 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"07aed5c023700866ebeb2a6224ac11566918d60f5a140685f482d3a85f4cd555\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-vx9cw" Feb 13 19:12:14.812083 kubelet[2828]: E0213 19:12:14.812034 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-vx9cw_kube-system(eb0c3c87-077a-4b2b-a168-5fd9131bccd3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-vx9cw_kube-system(eb0c3c87-077a-4b2b-a168-5fd9131bccd3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"07aed5c023700866ebeb2a6224ac11566918d60f5a140685f482d3a85f4cd555\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-vx9cw" podUID="eb0c3c87-077a-4b2b-a168-5fd9131bccd3" Feb 13 19:12:14.866019 containerd[1503]: time="2025-02-13T19:12:14.865962857Z" level=error msg="Failed to destroy network for sandbox \"ae9615caa56ac5be11c6b7cc88aac5196ade400655d7a9313d485332a9d3820e\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:14.866803 containerd[1503]: time="2025-02-13T19:12:14.866328668Z" level=error msg="encountered an error cleaning up failed sandbox \"ae9615caa56ac5be11c6b7cc88aac5196ade400655d7a9313d485332a9d3820e\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:14.866803 containerd[1503]: time="2025-02-13T19:12:14.866385510Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-84956f4cf-72cgp,Uid:96785f8f-b4e3-456b-9a0d-bcc142e6beaf,Namespace:calico-apiserver,Attempt:1,} failed, error" error="failed to setup network for sandbox \"ae9615caa56ac5be11c6b7cc88aac5196ade400655d7a9313d485332a9d3820e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:14.867135 kubelet[2828]: E0213 19:12:14.867102 2828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ae9615caa56ac5be11c6b7cc88aac5196ade400655d7a9313d485332a9d3820e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:14.867243 kubelet[2828]: E0213 19:12:14.867229 2828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ae9615caa56ac5be11c6b7cc88aac5196ade400655d7a9313d485332a9d3820e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-84956f4cf-72cgp" Feb 13 19:12:14.867323 kubelet[2828]: E0213 19:12:14.867303 2828 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ae9615caa56ac5be11c6b7cc88aac5196ade400655d7a9313d485332a9d3820e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-84956f4cf-72cgp" Feb 13 19:12:14.867418 kubelet[2828]: E0213 19:12:14.867397 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-84956f4cf-72cgp_calico-apiserver(96785f8f-b4e3-456b-9a0d-bcc142e6beaf)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-84956f4cf-72cgp_calico-apiserver(96785f8f-b4e3-456b-9a0d-bcc142e6beaf)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"ae9615caa56ac5be11c6b7cc88aac5196ade400655d7a9313d485332a9d3820e\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-84956f4cf-72cgp" podUID="96785f8f-b4e3-456b-9a0d-bcc142e6beaf" Feb 13 19:12:14.892806 containerd[1503]: time="2025-02-13T19:12:14.892745214Z" level=error msg="Failed to destroy network for sandbox \"0be20e2c65924f7b5393bc8b6a79f35d5bf8cca7f34009d6201de4852924939b\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:14.893150 containerd[1503]: time="2025-02-13T19:12:14.893121545Z" level=error msg="encountered an error cleaning up failed sandbox \"0be20e2c65924f7b5393bc8b6a79f35d5bf8cca7f34009d6201de4852924939b\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:14.893199 containerd[1503]: time="2025-02-13T19:12:14.893183747Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-78bd6d67c9-b259q,Uid:436348d7-7832-43e1-8780-623c434b2666,Namespace:calico-system,Attempt:1,} failed, error" error="failed to setup network for sandbox \"0be20e2c65924f7b5393bc8b6a79f35d5bf8cca7f34009d6201de4852924939b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:14.893413 kubelet[2828]: E0213 19:12:14.893380 2828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0be20e2c65924f7b5393bc8b6a79f35d5bf8cca7f34009d6201de4852924939b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:14.893504 kubelet[2828]: E0213 19:12:14.893449 2828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0be20e2c65924f7b5393bc8b6a79f35d5bf8cca7f34009d6201de4852924939b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-78bd6d67c9-b259q" Feb 13 19:12:14.893539 kubelet[2828]: E0213 19:12:14.893511 2828 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0be20e2c65924f7b5393bc8b6a79f35d5bf8cca7f34009d6201de4852924939b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-78bd6d67c9-b259q" Feb 13 19:12:14.894187 kubelet[2828]: E0213 19:12:14.893789 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-78bd6d67c9-b259q_calico-system(436348d7-7832-43e1-8780-623c434b2666)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-78bd6d67c9-b259q_calico-system(436348d7-7832-43e1-8780-623c434b2666)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"0be20e2c65924f7b5393bc8b6a79f35d5bf8cca7f34009d6201de4852924939b\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-78bd6d67c9-b259q" podUID="436348d7-7832-43e1-8780-623c434b2666" Feb 13 19:12:14.904999 containerd[1503]: time="2025-02-13T19:12:14.904938016Z" level=error msg="Failed to destroy network for sandbox \"4f219600df3d01ca237f481814d18c92c5997c6c501af4d70cb45cd41d1a6efc\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:14.906159 containerd[1503]: time="2025-02-13T19:12:14.906105931Z" level=error msg="encountered an error cleaning up failed sandbox \"4f219600df3d01ca237f481814d18c92c5997c6c501af4d70cb45cd41d1a6efc\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:14.906543 containerd[1503]: time="2025-02-13T19:12:14.906185893Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-6v9pr,Uid:b9c8452b-31d9-47bc-9697-dd4f293d71b9,Namespace:calico-system,Attempt:1,} failed, error" error="failed to setup network for sandbox \"4f219600df3d01ca237f481814d18c92c5997c6c501af4d70cb45cd41d1a6efc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:14.907116 containerd[1503]: time="2025-02-13T19:12:14.906837153Z" level=error msg="Failed to destroy network for sandbox \"9581f7ff6fcd5f0bba90e1d0d2984cd0efa6fe4d50a06a4c1b20f3939cb09713\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:14.907413 kubelet[2828]: E0213 19:12:14.907351 2828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4f219600df3d01ca237f481814d18c92c5997c6c501af4d70cb45cd41d1a6efc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:14.907667 kubelet[2828]: E0213 19:12:14.907626 2828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4f219600df3d01ca237f481814d18c92c5997c6c501af4d70cb45cd41d1a6efc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-6v9pr" Feb 13 19:12:14.907667 kubelet[2828]: E0213 19:12:14.907660 2828 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4f219600df3d01ca237f481814d18c92c5997c6c501af4d70cb45cd41d1a6efc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-6v9pr" Feb 13 19:12:14.908021 kubelet[2828]: E0213 19:12:14.907721 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-6v9pr_calico-system(b9c8452b-31d9-47bc-9697-dd4f293d71b9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-6v9pr_calico-system(b9c8452b-31d9-47bc-9697-dd4f293d71b9)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"4f219600df3d01ca237f481814d18c92c5997c6c501af4d70cb45cd41d1a6efc\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-6v9pr" podUID="b9c8452b-31d9-47bc-9697-dd4f293d71b9" Feb 13 19:12:14.909543 containerd[1503]: time="2025-02-13T19:12:14.908909734Z" level=error msg="encountered an error cleaning up failed sandbox \"9581f7ff6fcd5f0bba90e1d0d2984cd0efa6fe4d50a06a4c1b20f3939cb09713\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:14.909543 containerd[1503]: time="2025-02-13T19:12:14.909018938Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-84956f4cf-p6mr2,Uid:05e39f6f-0a83-4b62-8cff-b0d4799f8f89,Namespace:calico-apiserver,Attempt:1,} failed, error" error="failed to setup network for sandbox \"9581f7ff6fcd5f0bba90e1d0d2984cd0efa6fe4d50a06a4c1b20f3939cb09713\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:14.910545 kubelet[2828]: E0213 19:12:14.910002 2828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9581f7ff6fcd5f0bba90e1d0d2984cd0efa6fe4d50a06a4c1b20f3939cb09713\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:14.910545 kubelet[2828]: E0213 19:12:14.910160 2828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9581f7ff6fcd5f0bba90e1d0d2984cd0efa6fe4d50a06a4c1b20f3939cb09713\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-84956f4cf-p6mr2" Feb 13 19:12:14.910545 kubelet[2828]: E0213 19:12:14.910222 2828 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9581f7ff6fcd5f0bba90e1d0d2984cd0efa6fe4d50a06a4c1b20f3939cb09713\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-84956f4cf-p6mr2" Feb 13 19:12:14.910855 kubelet[2828]: E0213 19:12:14.910316 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-84956f4cf-p6mr2_calico-apiserver(05e39f6f-0a83-4b62-8cff-b0d4799f8f89)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-84956f4cf-p6mr2_calico-apiserver(05e39f6f-0a83-4b62-8cff-b0d4799f8f89)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"9581f7ff6fcd5f0bba90e1d0d2984cd0efa6fe4d50a06a4c1b20f3939cb09713\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-84956f4cf-p6mr2" podUID="05e39f6f-0a83-4b62-8cff-b0d4799f8f89" Feb 13 19:12:14.910904 containerd[1503]: time="2025-02-13T19:12:14.910831871Z" level=error msg="Failed to destroy network for sandbox \"10ae5b7da745ff3939219060da91f4661d2a1136a80736c3234db5b474e3599d\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:14.911698 containerd[1503]: time="2025-02-13T19:12:14.911644496Z" level=error msg="encountered an error cleaning up failed sandbox \"10ae5b7da745ff3939219060da91f4661d2a1136a80736c3234db5b474e3599d\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:14.911932 containerd[1503]: time="2025-02-13T19:12:14.911876222Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-kkqg5,Uid:5cf24d69-eb75-4cb3-8437-024193f516f5,Namespace:kube-system,Attempt:1,} failed, error" error="failed to setup network for sandbox \"10ae5b7da745ff3939219060da91f4661d2a1136a80736c3234db5b474e3599d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:14.912174 kubelet[2828]: E0213 19:12:14.912144 2828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"10ae5b7da745ff3939219060da91f4661d2a1136a80736c3234db5b474e3599d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:14.912229 kubelet[2828]: E0213 19:12:14.912200 2828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"10ae5b7da745ff3939219060da91f4661d2a1136a80736c3234db5b474e3599d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-kkqg5" Feb 13 19:12:14.912229 kubelet[2828]: E0213 19:12:14.912217 2828 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"10ae5b7da745ff3939219060da91f4661d2a1136a80736c3234db5b474e3599d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-kkqg5" Feb 13 19:12:14.912284 kubelet[2828]: E0213 19:12:14.912267 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-kkqg5_kube-system(5cf24d69-eb75-4cb3-8437-024193f516f5)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-kkqg5_kube-system(5cf24d69-eb75-4cb3-8437-024193f516f5)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"10ae5b7da745ff3939219060da91f4661d2a1136a80736c3234db5b474e3599d\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-kkqg5" podUID="5cf24d69-eb75-4cb3-8437-024193f516f5" Feb 13 19:12:15.415382 systemd[1]: run-netns-cni\x2d7d6b045f\x2d0761\x2d02d2\x2d702f\x2dbd09ff5737f5.mount: Deactivated successfully. Feb 13 19:12:15.415501 systemd[1]: run-netns-cni\x2dfb925c3f\x2d5678\x2d71df\x2d5354\x2d96af86ef4cca.mount: Deactivated successfully. Feb 13 19:12:15.682714 kubelet[2828]: I0213 19:12:15.682496 2828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07aed5c023700866ebeb2a6224ac11566918d60f5a140685f482d3a85f4cd555" Feb 13 19:12:15.684085 containerd[1503]: time="2025-02-13T19:12:15.684032518Z" level=info msg="StopPodSandbox for \"07aed5c023700866ebeb2a6224ac11566918d60f5a140685f482d3a85f4cd555\"" Feb 13 19:12:15.689204 containerd[1503]: time="2025-02-13T19:12:15.684416210Z" level=info msg="Ensure that sandbox 07aed5c023700866ebeb2a6224ac11566918d60f5a140685f482d3a85f4cd555 in task-service has been cleanup successfully" Feb 13 19:12:15.689204 containerd[1503]: time="2025-02-13T19:12:15.684589535Z" level=info msg="TearDown network for sandbox \"07aed5c023700866ebeb2a6224ac11566918d60f5a140685f482d3a85f4cd555\" successfully" Feb 13 19:12:15.689204 containerd[1503]: time="2025-02-13T19:12:15.684604935Z" level=info msg="StopPodSandbox for \"07aed5c023700866ebeb2a6224ac11566918d60f5a140685f482d3a85f4cd555\" returns successfully" Feb 13 19:12:15.689204 containerd[1503]: time="2025-02-13T19:12:15.687243614Z" level=info msg="StopPodSandbox for \"0c31ed484cb1e7593659d8fd1f850350d6b3b9e01ed8530e08cf66bc47a72b73\"" Feb 13 19:12:15.689204 containerd[1503]: time="2025-02-13T19:12:15.687328737Z" level=info msg="TearDown network for sandbox \"0c31ed484cb1e7593659d8fd1f850350d6b3b9e01ed8530e08cf66bc47a72b73\" successfully" Feb 13 19:12:15.689204 containerd[1503]: time="2025-02-13T19:12:15.687338737Z" level=info msg="StopPodSandbox for \"0c31ed484cb1e7593659d8fd1f850350d6b3b9e01ed8530e08cf66bc47a72b73\" returns successfully" Feb 13 19:12:15.688678 systemd[1]: run-netns-cni\x2decd1b742\x2d424d\x2d8225\x2dcde5\x2d610f3953637a.mount: Deactivated successfully. Feb 13 19:12:15.692717 kubelet[2828]: I0213 19:12:15.691387 2828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4f219600df3d01ca237f481814d18c92c5997c6c501af4d70cb45cd41d1a6efc" Feb 13 19:12:15.693966 containerd[1503]: time="2025-02-13T19:12:15.693775850Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-vx9cw,Uid:eb0c3c87-077a-4b2b-a168-5fd9131bccd3,Namespace:kube-system,Attempt:2,}" Feb 13 19:12:15.695887 containerd[1503]: time="2025-02-13T19:12:15.695846512Z" level=info msg="StopPodSandbox for \"4f219600df3d01ca237f481814d18c92c5997c6c501af4d70cb45cd41d1a6efc\"" Feb 13 19:12:15.696367 containerd[1503]: time="2025-02-13T19:12:15.696212443Z" level=info msg="Ensure that sandbox 4f219600df3d01ca237f481814d18c92c5997c6c501af4d70cb45cd41d1a6efc in task-service has been cleanup successfully" Feb 13 19:12:15.696451 containerd[1503]: time="2025-02-13T19:12:15.696425170Z" level=info msg="TearDown network for sandbox \"4f219600df3d01ca237f481814d18c92c5997c6c501af4d70cb45cd41d1a6efc\" successfully" Feb 13 19:12:15.696451 containerd[1503]: time="2025-02-13T19:12:15.696444770Z" level=info msg="StopPodSandbox for \"4f219600df3d01ca237f481814d18c92c5997c6c501af4d70cb45cd41d1a6efc\" returns successfully" Feb 13 19:12:15.697963 containerd[1503]: time="2025-02-13T19:12:15.697348197Z" level=info msg="StopPodSandbox for \"1f6b42d2735cedb916bc24bf2bb557d2ff72928d0c487a93a3dd46e88aa55abe\"" Feb 13 19:12:15.697963 containerd[1503]: time="2025-02-13T19:12:15.697533643Z" level=info msg="TearDown network for sandbox \"1f6b42d2735cedb916bc24bf2bb557d2ff72928d0c487a93a3dd46e88aa55abe\" successfully" Feb 13 19:12:15.697963 containerd[1503]: time="2025-02-13T19:12:15.697548083Z" level=info msg="StopPodSandbox for \"1f6b42d2735cedb916bc24bf2bb557d2ff72928d0c487a93a3dd46e88aa55abe\" returns successfully" Feb 13 19:12:15.700757 containerd[1503]: time="2025-02-13T19:12:15.699146971Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-6v9pr,Uid:b9c8452b-31d9-47bc-9697-dd4f293d71b9,Namespace:calico-system,Attempt:2,}" Feb 13 19:12:15.702423 systemd[1]: run-netns-cni\x2d7b17439f\x2dc16d\x2dea13\x2d88c4\x2d36eb56bed466.mount: Deactivated successfully. Feb 13 19:12:15.704521 kubelet[2828]: I0213 19:12:15.704456 2828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0be20e2c65924f7b5393bc8b6a79f35d5bf8cca7f34009d6201de4852924939b" Feb 13 19:12:15.707997 containerd[1503]: time="2025-02-13T19:12:15.707953316Z" level=info msg="StopPodSandbox for \"0be20e2c65924f7b5393bc8b6a79f35d5bf8cca7f34009d6201de4852924939b\"" Feb 13 19:12:15.709394 containerd[1503]: time="2025-02-13T19:12:15.709186793Z" level=info msg="Ensure that sandbox 0be20e2c65924f7b5393bc8b6a79f35d5bf8cca7f34009d6201de4852924939b in task-service has been cleanup successfully" Feb 13 19:12:15.709526 containerd[1503]: time="2025-02-13T19:12:15.709504362Z" level=info msg="TearDown network for sandbox \"0be20e2c65924f7b5393bc8b6a79f35d5bf8cca7f34009d6201de4852924939b\" successfully" Feb 13 19:12:15.709581 containerd[1503]: time="2025-02-13T19:12:15.709568244Z" level=info msg="StopPodSandbox for \"0be20e2c65924f7b5393bc8b6a79f35d5bf8cca7f34009d6201de4852924939b\" returns successfully" Feb 13 19:12:15.712989 systemd[1]: run-netns-cni\x2dddf7ac0e\x2d82b3\x2d99a7\x2db1d2\x2d8fc259c6a44c.mount: Deactivated successfully. Feb 13 19:12:15.716611 containerd[1503]: time="2025-02-13T19:12:15.715509662Z" level=info msg="StopPodSandbox for \"fa6e4b9a2b53f8e5e8150bdfa0fe88ef1d51405b9b5e7e6d9683c6adb0b078ab\"" Feb 13 19:12:15.716611 containerd[1503]: time="2025-02-13T19:12:15.715627826Z" level=info msg="TearDown network for sandbox \"fa6e4b9a2b53f8e5e8150bdfa0fe88ef1d51405b9b5e7e6d9683c6adb0b078ab\" successfully" Feb 13 19:12:15.716611 containerd[1503]: time="2025-02-13T19:12:15.715639146Z" level=info msg="StopPodSandbox for \"fa6e4b9a2b53f8e5e8150bdfa0fe88ef1d51405b9b5e7e6d9683c6adb0b078ab\" returns successfully" Feb 13 19:12:15.716611 containerd[1503]: time="2025-02-13T19:12:15.716553853Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-78bd6d67c9-b259q,Uid:436348d7-7832-43e1-8780-623c434b2666,Namespace:calico-system,Attempt:2,}" Feb 13 19:12:15.717531 kubelet[2828]: I0213 19:12:15.717441 2828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae9615caa56ac5be11c6b7cc88aac5196ade400655d7a9313d485332a9d3820e" Feb 13 19:12:15.718285 containerd[1503]: time="2025-02-13T19:12:15.718161102Z" level=info msg="StopPodSandbox for \"ae9615caa56ac5be11c6b7cc88aac5196ade400655d7a9313d485332a9d3820e\"" Feb 13 19:12:15.718525 containerd[1503]: time="2025-02-13T19:12:15.718496272Z" level=info msg="Ensure that sandbox ae9615caa56ac5be11c6b7cc88aac5196ade400655d7a9313d485332a9d3820e in task-service has been cleanup successfully" Feb 13 19:12:15.720073 containerd[1503]: time="2025-02-13T19:12:15.720021837Z" level=info msg="TearDown network for sandbox \"ae9615caa56ac5be11c6b7cc88aac5196ade400655d7a9313d485332a9d3820e\" successfully" Feb 13 19:12:15.720597 containerd[1503]: time="2025-02-13T19:12:15.720061959Z" level=info msg="StopPodSandbox for \"ae9615caa56ac5be11c6b7cc88aac5196ade400655d7a9313d485332a9d3820e\" returns successfully" Feb 13 19:12:15.721355 kubelet[2828]: I0213 19:12:15.721326 2828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9581f7ff6fcd5f0bba90e1d0d2984cd0efa6fe4d50a06a4c1b20f3939cb09713" Feb 13 19:12:15.722906 containerd[1503]: time="2025-02-13T19:12:15.722285505Z" level=info msg="StopPodSandbox for \"930563c746d8b91d524a61a614694c47905765a47fb87cb32a681d5b35060a78\"" Feb 13 19:12:15.722906 containerd[1503]: time="2025-02-13T19:12:15.722427310Z" level=info msg="TearDown network for sandbox \"930563c746d8b91d524a61a614694c47905765a47fb87cb32a681d5b35060a78\" successfully" Feb 13 19:12:15.722906 containerd[1503]: time="2025-02-13T19:12:15.722438350Z" level=info msg="StopPodSandbox for \"930563c746d8b91d524a61a614694c47905765a47fb87cb32a681d5b35060a78\" returns successfully" Feb 13 19:12:15.723334 containerd[1503]: time="2025-02-13T19:12:15.723278895Z" level=info msg="StopPodSandbox for \"9581f7ff6fcd5f0bba90e1d0d2984cd0efa6fe4d50a06a4c1b20f3939cb09713\"" Feb 13 19:12:15.723917 containerd[1503]: time="2025-02-13T19:12:15.723754429Z" level=info msg="Ensure that sandbox 9581f7ff6fcd5f0bba90e1d0d2984cd0efa6fe4d50a06a4c1b20f3939cb09713 in task-service has been cleanup successfully" Feb 13 19:12:15.724811 containerd[1503]: time="2025-02-13T19:12:15.724189562Z" level=info msg="TearDown network for sandbox \"9581f7ff6fcd5f0bba90e1d0d2984cd0efa6fe4d50a06a4c1b20f3939cb09713\" successfully" Feb 13 19:12:15.724811 containerd[1503]: time="2025-02-13T19:12:15.724793981Z" level=info msg="StopPodSandbox for \"9581f7ff6fcd5f0bba90e1d0d2984cd0efa6fe4d50a06a4c1b20f3939cb09713\" returns successfully" Feb 13 19:12:15.725450 containerd[1503]: time="2025-02-13T19:12:15.725220513Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-84956f4cf-72cgp,Uid:96785f8f-b4e3-456b-9a0d-bcc142e6beaf,Namespace:calico-apiserver,Attempt:2,}" Feb 13 19:12:15.725821 containerd[1503]: time="2025-02-13T19:12:15.725697448Z" level=info msg="StopPodSandbox for \"0c13c2a989169406cb4cea286d09a0b055b98c8bf7adaf3c864364728278a3a5\"" Feb 13 19:12:15.725821 containerd[1503]: time="2025-02-13T19:12:15.725799171Z" level=info msg="TearDown network for sandbox \"0c13c2a989169406cb4cea286d09a0b055b98c8bf7adaf3c864364728278a3a5\" successfully" Feb 13 19:12:15.725821 containerd[1503]: time="2025-02-13T19:12:15.725812411Z" level=info msg="StopPodSandbox for \"0c13c2a989169406cb4cea286d09a0b055b98c8bf7adaf3c864364728278a3a5\" returns successfully" Feb 13 19:12:15.727032 containerd[1503]: time="2025-02-13T19:12:15.726465151Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-84956f4cf-p6mr2,Uid:05e39f6f-0a83-4b62-8cff-b0d4799f8f89,Namespace:calico-apiserver,Attempt:2,}" Feb 13 19:12:15.728183 kubelet[2828]: I0213 19:12:15.728149 2828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10ae5b7da745ff3939219060da91f4661d2a1136a80736c3234db5b474e3599d" Feb 13 19:12:15.733662 containerd[1503]: time="2025-02-13T19:12:15.733623605Z" level=info msg="StopPodSandbox for \"10ae5b7da745ff3939219060da91f4661d2a1136a80736c3234db5b474e3599d\"" Feb 13 19:12:15.734299 containerd[1503]: time="2025-02-13T19:12:15.734270705Z" level=info msg="Ensure that sandbox 10ae5b7da745ff3939219060da91f4661d2a1136a80736c3234db5b474e3599d in task-service has been cleanup successfully" Feb 13 19:12:15.737724 containerd[1503]: time="2025-02-13T19:12:15.737681687Z" level=info msg="TearDown network for sandbox \"10ae5b7da745ff3939219060da91f4661d2a1136a80736c3234db5b474e3599d\" successfully" Feb 13 19:12:15.737724 containerd[1503]: time="2025-02-13T19:12:15.737719328Z" level=info msg="StopPodSandbox for \"10ae5b7da745ff3939219060da91f4661d2a1136a80736c3234db5b474e3599d\" returns successfully" Feb 13 19:12:15.738261 containerd[1503]: time="2025-02-13T19:12:15.738223583Z" level=info msg="StopPodSandbox for \"43a5af5ee3597e852f8d0ba977ffb06562d6b40688d779a95c8d495a78df4c29\"" Feb 13 19:12:15.739965 containerd[1503]: time="2025-02-13T19:12:15.739425339Z" level=info msg="TearDown network for sandbox \"43a5af5ee3597e852f8d0ba977ffb06562d6b40688d779a95c8d495a78df4c29\" successfully" Feb 13 19:12:15.739965 containerd[1503]: time="2025-02-13T19:12:15.739450140Z" level=info msg="StopPodSandbox for \"43a5af5ee3597e852f8d0ba977ffb06562d6b40688d779a95c8d495a78df4c29\" returns successfully" Feb 13 19:12:15.740986 containerd[1503]: time="2025-02-13T19:12:15.740940225Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-kkqg5,Uid:5cf24d69-eb75-4cb3-8437-024193f516f5,Namespace:kube-system,Attempt:2,}" Feb 13 19:12:15.916807 containerd[1503]: time="2025-02-13T19:12:15.916449289Z" level=error msg="Failed to destroy network for sandbox \"4642b0f0428148a51fe1afb0c6a261eff1112eb99bbdf06b50e264a81fa407b3\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:15.919865 containerd[1503]: time="2025-02-13T19:12:15.919750148Z" level=error msg="encountered an error cleaning up failed sandbox \"4642b0f0428148a51fe1afb0c6a261eff1112eb99bbdf06b50e264a81fa407b3\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:15.919865 containerd[1503]: time="2025-02-13T19:12:15.919833910Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-6v9pr,Uid:b9c8452b-31d9-47bc-9697-dd4f293d71b9,Namespace:calico-system,Attempt:2,} failed, error" error="failed to setup network for sandbox \"4642b0f0428148a51fe1afb0c6a261eff1112eb99bbdf06b50e264a81fa407b3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:15.920302 kubelet[2828]: E0213 19:12:15.920256 2828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4642b0f0428148a51fe1afb0c6a261eff1112eb99bbdf06b50e264a81fa407b3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:15.920964 kubelet[2828]: E0213 19:12:15.920323 2828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4642b0f0428148a51fe1afb0c6a261eff1112eb99bbdf06b50e264a81fa407b3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-6v9pr" Feb 13 19:12:15.920964 kubelet[2828]: E0213 19:12:15.920346 2828 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4642b0f0428148a51fe1afb0c6a261eff1112eb99bbdf06b50e264a81fa407b3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-6v9pr" Feb 13 19:12:15.921550 kubelet[2828]: E0213 19:12:15.920419 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-6v9pr_calico-system(b9c8452b-31d9-47bc-9697-dd4f293d71b9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-6v9pr_calico-system(b9c8452b-31d9-47bc-9697-dd4f293d71b9)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"4642b0f0428148a51fe1afb0c6a261eff1112eb99bbdf06b50e264a81fa407b3\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-6v9pr" podUID="b9c8452b-31d9-47bc-9697-dd4f293d71b9" Feb 13 19:12:15.948142 containerd[1503]: time="2025-02-13T19:12:15.947471739Z" level=error msg="Failed to destroy network for sandbox \"a657753a39333ef1c8ee279e3321a3d04ca609cd0499c88ae5589336776d3f50\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:15.948142 containerd[1503]: time="2025-02-13T19:12:15.947853551Z" level=error msg="encountered an error cleaning up failed sandbox \"a657753a39333ef1c8ee279e3321a3d04ca609cd0499c88ae5589336776d3f50\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:15.948142 containerd[1503]: time="2025-02-13T19:12:15.947908592Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-vx9cw,Uid:eb0c3c87-077a-4b2b-a168-5fd9131bccd3,Namespace:kube-system,Attempt:2,} failed, error" error="failed to setup network for sandbox \"a657753a39333ef1c8ee279e3321a3d04ca609cd0499c88ae5589336776d3f50\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:15.950283 kubelet[2828]: E0213 19:12:15.950120 2828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a657753a39333ef1c8ee279e3321a3d04ca609cd0499c88ae5589336776d3f50\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:15.950283 kubelet[2828]: E0213 19:12:15.950196 2828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a657753a39333ef1c8ee279e3321a3d04ca609cd0499c88ae5589336776d3f50\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-vx9cw" Feb 13 19:12:15.950283 kubelet[2828]: E0213 19:12:15.950222 2828 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a657753a39333ef1c8ee279e3321a3d04ca609cd0499c88ae5589336776d3f50\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-vx9cw" Feb 13 19:12:15.950452 kubelet[2828]: E0213 19:12:15.950275 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-vx9cw_kube-system(eb0c3c87-077a-4b2b-a168-5fd9131bccd3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-vx9cw_kube-system(eb0c3c87-077a-4b2b-a168-5fd9131bccd3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"a657753a39333ef1c8ee279e3321a3d04ca609cd0499c88ae5589336776d3f50\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-vx9cw" podUID="eb0c3c87-077a-4b2b-a168-5fd9131bccd3" Feb 13 19:12:15.980168 containerd[1503]: time="2025-02-13T19:12:15.979962874Z" level=error msg="Failed to destroy network for sandbox \"af85f8d50ee74cce2c7e2cee2af5922ba7bbd4077fc92d81b183d5575fad1329\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:15.982810 containerd[1503]: time="2025-02-13T19:12:15.981997055Z" level=error msg="encountered an error cleaning up failed sandbox \"af85f8d50ee74cce2c7e2cee2af5922ba7bbd4077fc92d81b183d5575fad1329\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:15.982810 containerd[1503]: time="2025-02-13T19:12:15.982111578Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-84956f4cf-72cgp,Uid:96785f8f-b4e3-456b-9a0d-bcc142e6beaf,Namespace:calico-apiserver,Attempt:2,} failed, error" error="failed to setup network for sandbox \"af85f8d50ee74cce2c7e2cee2af5922ba7bbd4077fc92d81b183d5575fad1329\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:15.983102 kubelet[2828]: E0213 19:12:15.982328 2828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"af85f8d50ee74cce2c7e2cee2af5922ba7bbd4077fc92d81b183d5575fad1329\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:15.983102 kubelet[2828]: E0213 19:12:15.982635 2828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"af85f8d50ee74cce2c7e2cee2af5922ba7bbd4077fc92d81b183d5575fad1329\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-84956f4cf-72cgp" Feb 13 19:12:15.983102 kubelet[2828]: E0213 19:12:15.982668 2828 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"af85f8d50ee74cce2c7e2cee2af5922ba7bbd4077fc92d81b183d5575fad1329\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-84956f4cf-72cgp" Feb 13 19:12:15.983271 kubelet[2828]: E0213 19:12:15.982724 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-84956f4cf-72cgp_calico-apiserver(96785f8f-b4e3-456b-9a0d-bcc142e6beaf)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-84956f4cf-72cgp_calico-apiserver(96785f8f-b4e3-456b-9a0d-bcc142e6beaf)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"af85f8d50ee74cce2c7e2cee2af5922ba7bbd4077fc92d81b183d5575fad1329\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-84956f4cf-72cgp" podUID="96785f8f-b4e3-456b-9a0d-bcc142e6beaf" Feb 13 19:12:16.012273 containerd[1503]: time="2025-02-13T19:12:16.012205203Z" level=error msg="Failed to destroy network for sandbox \"4ffb63334faa7203eea7b2071116973af83410f7340f4e8946194bccb25fa157\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:16.013282 containerd[1503]: time="2025-02-13T19:12:16.013087550Z" level=error msg="encountered an error cleaning up failed sandbox \"4ffb63334faa7203eea7b2071116973af83410f7340f4e8946194bccb25fa157\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:16.013389 containerd[1503]: time="2025-02-13T19:12:16.013203394Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-78bd6d67c9-b259q,Uid:436348d7-7832-43e1-8780-623c434b2666,Namespace:calico-system,Attempt:2,} failed, error" error="failed to setup network for sandbox \"4ffb63334faa7203eea7b2071116973af83410f7340f4e8946194bccb25fa157\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:16.013623 kubelet[2828]: E0213 19:12:16.013557 2828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4ffb63334faa7203eea7b2071116973af83410f7340f4e8946194bccb25fa157\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:16.013680 kubelet[2828]: E0213 19:12:16.013622 2828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4ffb63334faa7203eea7b2071116973af83410f7340f4e8946194bccb25fa157\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-78bd6d67c9-b259q" Feb 13 19:12:16.013680 kubelet[2828]: E0213 19:12:16.013644 2828 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4ffb63334faa7203eea7b2071116973af83410f7340f4e8946194bccb25fa157\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-78bd6d67c9-b259q" Feb 13 19:12:16.013869 kubelet[2828]: E0213 19:12:16.013682 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-78bd6d67c9-b259q_calico-system(436348d7-7832-43e1-8780-623c434b2666)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-78bd6d67c9-b259q_calico-system(436348d7-7832-43e1-8780-623c434b2666)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"4ffb63334faa7203eea7b2071116973af83410f7340f4e8946194bccb25fa157\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-78bd6d67c9-b259q" podUID="436348d7-7832-43e1-8780-623c434b2666" Feb 13 19:12:16.037952 containerd[1503]: time="2025-02-13T19:12:16.037899900Z" level=error msg="Failed to destroy network for sandbox \"ab072ffc918efc3594006a69ddbbd17de37178f95f1103e4fa6b31bb95fd9306\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:16.038613 containerd[1503]: time="2025-02-13T19:12:16.038428676Z" level=error msg="encountered an error cleaning up failed sandbox \"ab072ffc918efc3594006a69ddbbd17de37178f95f1103e4fa6b31bb95fd9306\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:16.038613 containerd[1503]: time="2025-02-13T19:12:16.038589681Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-84956f4cf-p6mr2,Uid:05e39f6f-0a83-4b62-8cff-b0d4799f8f89,Namespace:calico-apiserver,Attempt:2,} failed, error" error="failed to setup network for sandbox \"ab072ffc918efc3594006a69ddbbd17de37178f95f1103e4fa6b31bb95fd9306\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:16.038983 kubelet[2828]: E0213 19:12:16.038947 2828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ab072ffc918efc3594006a69ddbbd17de37178f95f1103e4fa6b31bb95fd9306\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:16.039083 kubelet[2828]: E0213 19:12:16.039001 2828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ab072ffc918efc3594006a69ddbbd17de37178f95f1103e4fa6b31bb95fd9306\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-84956f4cf-p6mr2" Feb 13 19:12:16.039083 kubelet[2828]: E0213 19:12:16.039020 2828 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ab072ffc918efc3594006a69ddbbd17de37178f95f1103e4fa6b31bb95fd9306\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-84956f4cf-p6mr2" Feb 13 19:12:16.039160 kubelet[2828]: E0213 19:12:16.039124 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-84956f4cf-p6mr2_calico-apiserver(05e39f6f-0a83-4b62-8cff-b0d4799f8f89)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-84956f4cf-p6mr2_calico-apiserver(05e39f6f-0a83-4b62-8cff-b0d4799f8f89)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"ab072ffc918efc3594006a69ddbbd17de37178f95f1103e4fa6b31bb95fd9306\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-84956f4cf-p6mr2" podUID="05e39f6f-0a83-4b62-8cff-b0d4799f8f89" Feb 13 19:12:16.049627 containerd[1503]: time="2025-02-13T19:12:16.049569733Z" level=error msg="Failed to destroy network for sandbox \"d2cb6c1d9e5335f4d8ee798e79c4a58d442740cd8f9ddfe185709b1cfbd4ea81\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:16.050186 containerd[1503]: time="2025-02-13T19:12:16.050129790Z" level=error msg="encountered an error cleaning up failed sandbox \"d2cb6c1d9e5335f4d8ee798e79c4a58d442740cd8f9ddfe185709b1cfbd4ea81\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:16.050246 containerd[1503]: time="2025-02-13T19:12:16.050215633Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-kkqg5,Uid:5cf24d69-eb75-4cb3-8437-024193f516f5,Namespace:kube-system,Attempt:2,} failed, error" error="failed to setup network for sandbox \"d2cb6c1d9e5335f4d8ee798e79c4a58d442740cd8f9ddfe185709b1cfbd4ea81\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:16.050567 kubelet[2828]: E0213 19:12:16.050530 2828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d2cb6c1d9e5335f4d8ee798e79c4a58d442740cd8f9ddfe185709b1cfbd4ea81\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:16.050634 kubelet[2828]: E0213 19:12:16.050590 2828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d2cb6c1d9e5335f4d8ee798e79c4a58d442740cd8f9ddfe185709b1cfbd4ea81\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-kkqg5" Feb 13 19:12:16.050634 kubelet[2828]: E0213 19:12:16.050613 2828 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d2cb6c1d9e5335f4d8ee798e79c4a58d442740cd8f9ddfe185709b1cfbd4ea81\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-kkqg5" Feb 13 19:12:16.050696 kubelet[2828]: E0213 19:12:16.050657 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-kkqg5_kube-system(5cf24d69-eb75-4cb3-8437-024193f516f5)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-kkqg5_kube-system(5cf24d69-eb75-4cb3-8437-024193f516f5)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"d2cb6c1d9e5335f4d8ee798e79c4a58d442740cd8f9ddfe185709b1cfbd4ea81\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-kkqg5" podUID="5cf24d69-eb75-4cb3-8437-024193f516f5" Feb 13 19:12:16.412572 systemd[1]: run-netns-cni\x2dce934d0e\x2d2230\x2dcc7b\x2d6d71\x2d17903cc9b329.mount: Deactivated successfully. Feb 13 19:12:16.412681 systemd[1]: run-netns-cni\x2d15b9a981\x2d5265\x2d32a1\x2deafc\x2d20cb11cc42fa.mount: Deactivated successfully. Feb 13 19:12:16.412728 systemd[1]: run-netns-cni\x2d2e8fa9b8\x2deb1b\x2db599\x2df423\x2dda142b1899e9.mount: Deactivated successfully. Feb 13 19:12:16.733417 kubelet[2828]: I0213 19:12:16.733249 2828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a657753a39333ef1c8ee279e3321a3d04ca609cd0499c88ae5589336776d3f50" Feb 13 19:12:16.735212 containerd[1503]: time="2025-02-13T19:12:16.734691970Z" level=info msg="StopPodSandbox for \"a657753a39333ef1c8ee279e3321a3d04ca609cd0499c88ae5589336776d3f50\"" Feb 13 19:12:16.735212 containerd[1503]: time="2025-02-13T19:12:16.735138743Z" level=info msg="Ensure that sandbox a657753a39333ef1c8ee279e3321a3d04ca609cd0499c88ae5589336776d3f50 in task-service has been cleanup successfully" Feb 13 19:12:16.737322 containerd[1503]: time="2025-02-13T19:12:16.736165854Z" level=info msg="TearDown network for sandbox \"a657753a39333ef1c8ee279e3321a3d04ca609cd0499c88ae5589336776d3f50\" successfully" Feb 13 19:12:16.737322 containerd[1503]: time="2025-02-13T19:12:16.736192775Z" level=info msg="StopPodSandbox for \"a657753a39333ef1c8ee279e3321a3d04ca609cd0499c88ae5589336776d3f50\" returns successfully" Feb 13 19:12:16.738263 systemd[1]: run-netns-cni\x2d971873c5\x2d5d0d\x2d3467\x2df6a9\x2d0a73bf3e9c31.mount: Deactivated successfully. Feb 13 19:12:16.740340 containerd[1503]: time="2025-02-13T19:12:16.739956609Z" level=info msg="StopPodSandbox for \"07aed5c023700866ebeb2a6224ac11566918d60f5a140685f482d3a85f4cd555\"" Feb 13 19:12:16.740340 containerd[1503]: time="2025-02-13T19:12:16.740090813Z" level=info msg="TearDown network for sandbox \"07aed5c023700866ebeb2a6224ac11566918d60f5a140685f482d3a85f4cd555\" successfully" Feb 13 19:12:16.740340 containerd[1503]: time="2025-02-13T19:12:16.740103013Z" level=info msg="StopPodSandbox for \"07aed5c023700866ebeb2a6224ac11566918d60f5a140685f482d3a85f4cd555\" returns successfully" Feb 13 19:12:16.741848 containerd[1503]: time="2025-02-13T19:12:16.741690101Z" level=info msg="StopPodSandbox for \"0c31ed484cb1e7593659d8fd1f850350d6b3b9e01ed8530e08cf66bc47a72b73\"" Feb 13 19:12:16.741931 containerd[1503]: time="2025-02-13T19:12:16.741857706Z" level=info msg="TearDown network for sandbox \"0c31ed484cb1e7593659d8fd1f850350d6b3b9e01ed8530e08cf66bc47a72b73\" successfully" Feb 13 19:12:16.741931 containerd[1503]: time="2025-02-13T19:12:16.741870267Z" level=info msg="StopPodSandbox for \"0c31ed484cb1e7593659d8fd1f850350d6b3b9e01ed8530e08cf66bc47a72b73\" returns successfully" Feb 13 19:12:16.743610 containerd[1503]: time="2025-02-13T19:12:16.743392113Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-vx9cw,Uid:eb0c3c87-077a-4b2b-a168-5fd9131bccd3,Namespace:kube-system,Attempt:3,}" Feb 13 19:12:16.744556 kubelet[2828]: I0213 19:12:16.744219 2828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4642b0f0428148a51fe1afb0c6a261eff1112eb99bbdf06b50e264a81fa407b3" Feb 13 19:12:16.746458 containerd[1503]: time="2025-02-13T19:12:16.746296881Z" level=info msg="StopPodSandbox for \"4642b0f0428148a51fe1afb0c6a261eff1112eb99bbdf06b50e264a81fa407b3\"" Feb 13 19:12:16.746977 containerd[1503]: time="2025-02-13T19:12:16.746881018Z" level=info msg="Ensure that sandbox 4642b0f0428148a51fe1afb0c6a261eff1112eb99bbdf06b50e264a81fa407b3 in task-service has been cleanup successfully" Feb 13 19:12:16.749731 containerd[1503]: time="2025-02-13T19:12:16.748022653Z" level=info msg="TearDown network for sandbox \"4642b0f0428148a51fe1afb0c6a261eff1112eb99bbdf06b50e264a81fa407b3\" successfully" Feb 13 19:12:16.749993 containerd[1503]: time="2025-02-13T19:12:16.749921670Z" level=info msg="StopPodSandbox for \"4642b0f0428148a51fe1afb0c6a261eff1112eb99bbdf06b50e264a81fa407b3\" returns successfully" Feb 13 19:12:16.751988 systemd[1]: run-netns-cni\x2dc4082c79\x2d9a12\x2dbfb8\x2dd52e\x2d5f340aaec782.mount: Deactivated successfully. Feb 13 19:12:16.754685 containerd[1503]: time="2025-02-13T19:12:16.754350324Z" level=info msg="StopPodSandbox for \"4f219600df3d01ca237f481814d18c92c5997c6c501af4d70cb45cd41d1a6efc\"" Feb 13 19:12:16.754685 containerd[1503]: time="2025-02-13T19:12:16.754464208Z" level=info msg="TearDown network for sandbox \"4f219600df3d01ca237f481814d18c92c5997c6c501af4d70cb45cd41d1a6efc\" successfully" Feb 13 19:12:16.754685 containerd[1503]: time="2025-02-13T19:12:16.754475008Z" level=info msg="StopPodSandbox for \"4f219600df3d01ca237f481814d18c92c5997c6c501af4d70cb45cd41d1a6efc\" returns successfully" Feb 13 19:12:16.755301 kubelet[2828]: I0213 19:12:16.755048 2828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2cb6c1d9e5335f4d8ee798e79c4a58d442740cd8f9ddfe185709b1cfbd4ea81" Feb 13 19:12:16.757904 containerd[1503]: time="2025-02-13T19:12:16.757860630Z" level=info msg="StopPodSandbox for \"1f6b42d2735cedb916bc24bf2bb557d2ff72928d0c487a93a3dd46e88aa55abe\"" Feb 13 19:12:16.758534 containerd[1503]: time="2025-02-13T19:12:16.758193880Z" level=info msg="TearDown network for sandbox \"1f6b42d2735cedb916bc24bf2bb557d2ff72928d0c487a93a3dd46e88aa55abe\" successfully" Feb 13 19:12:16.758534 containerd[1503]: time="2025-02-13T19:12:16.758214081Z" level=info msg="StopPodSandbox for \"1f6b42d2735cedb916bc24bf2bb557d2ff72928d0c487a93a3dd46e88aa55abe\" returns successfully" Feb 13 19:12:16.758622 containerd[1503]: time="2025-02-13T19:12:16.758537211Z" level=info msg="StopPodSandbox for \"d2cb6c1d9e5335f4d8ee798e79c4a58d442740cd8f9ddfe185709b1cfbd4ea81\"" Feb 13 19:12:16.759876 containerd[1503]: time="2025-02-13T19:12:16.758815019Z" level=info msg="Ensure that sandbox d2cb6c1d9e5335f4d8ee798e79c4a58d442740cd8f9ddfe185709b1cfbd4ea81 in task-service has been cleanup successfully" Feb 13 19:12:16.759876 containerd[1503]: time="2025-02-13T19:12:16.759037266Z" level=info msg="TearDown network for sandbox \"d2cb6c1d9e5335f4d8ee798e79c4a58d442740cd8f9ddfe185709b1cfbd4ea81\" successfully" Feb 13 19:12:16.759876 containerd[1503]: time="2025-02-13T19:12:16.759051706Z" level=info msg="StopPodSandbox for \"d2cb6c1d9e5335f4d8ee798e79c4a58d442740cd8f9ddfe185709b1cfbd4ea81\" returns successfully" Feb 13 19:12:16.759997 containerd[1503]: time="2025-02-13T19:12:16.759925653Z" level=info msg="StopPodSandbox for \"10ae5b7da745ff3939219060da91f4661d2a1136a80736c3234db5b474e3599d\"" Feb 13 19:12:16.761911 containerd[1503]: time="2025-02-13T19:12:16.760014855Z" level=info msg="TearDown network for sandbox \"10ae5b7da745ff3939219060da91f4661d2a1136a80736c3234db5b474e3599d\" successfully" Feb 13 19:12:16.761911 containerd[1503]: time="2025-02-13T19:12:16.760031216Z" level=info msg="StopPodSandbox for \"10ae5b7da745ff3939219060da91f4661d2a1136a80736c3234db5b474e3599d\" returns successfully" Feb 13 19:12:16.761911 containerd[1503]: time="2025-02-13T19:12:16.760286544Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-6v9pr,Uid:b9c8452b-31d9-47bc-9697-dd4f293d71b9,Namespace:calico-system,Attempt:3,}" Feb 13 19:12:16.765652 systemd[1]: run-netns-cni\x2d7c9a2d8a\x2d9d52\x2dd851\x2d28ee\x2d75968f4a6bba.mount: Deactivated successfully. Feb 13 19:12:16.768598 containerd[1503]: time="2025-02-13T19:12:16.768544833Z" level=info msg="StopPodSandbox for \"43a5af5ee3597e852f8d0ba977ffb06562d6b40688d779a95c8d495a78df4c29\"" Feb 13 19:12:16.768763 containerd[1503]: time="2025-02-13T19:12:16.768670437Z" level=info msg="TearDown network for sandbox \"43a5af5ee3597e852f8d0ba977ffb06562d6b40688d779a95c8d495a78df4c29\" successfully" Feb 13 19:12:16.768763 containerd[1503]: time="2025-02-13T19:12:16.768682278Z" level=info msg="StopPodSandbox for \"43a5af5ee3597e852f8d0ba977ffb06562d6b40688d779a95c8d495a78df4c29\" returns successfully" Feb 13 19:12:16.771113 containerd[1503]: time="2025-02-13T19:12:16.770202724Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-kkqg5,Uid:5cf24d69-eb75-4cb3-8437-024193f516f5,Namespace:kube-system,Attempt:3,}" Feb 13 19:12:16.771811 kubelet[2828]: I0213 19:12:16.771787 2828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ffb63334faa7203eea7b2071116973af83410f7340f4e8946194bccb25fa157" Feb 13 19:12:16.775280 containerd[1503]: time="2025-02-13T19:12:16.775233796Z" level=info msg="StopPodSandbox for \"4ffb63334faa7203eea7b2071116973af83410f7340f4e8946194bccb25fa157\"" Feb 13 19:12:16.775479 containerd[1503]: time="2025-02-13T19:12:16.775454322Z" level=info msg="Ensure that sandbox 4ffb63334faa7203eea7b2071116973af83410f7340f4e8946194bccb25fa157 in task-service has been cleanup successfully" Feb 13 19:12:16.779746 containerd[1503]: time="2025-02-13T19:12:16.779574727Z" level=info msg="TearDown network for sandbox \"4ffb63334faa7203eea7b2071116973af83410f7340f4e8946194bccb25fa157\" successfully" Feb 13 19:12:16.779746 containerd[1503]: time="2025-02-13T19:12:16.779606848Z" level=info msg="StopPodSandbox for \"4ffb63334faa7203eea7b2071116973af83410f7340f4e8946194bccb25fa157\" returns successfully" Feb 13 19:12:16.779788 systemd[1]: run-netns-cni\x2dff616f93\x2da177\x2d687a\x2de65c\x2d9bd4375039a0.mount: Deactivated successfully. Feb 13 19:12:16.783243 containerd[1503]: time="2025-02-13T19:12:16.783202677Z" level=info msg="StopPodSandbox for \"0be20e2c65924f7b5393bc8b6a79f35d5bf8cca7f34009d6201de4852924939b\"" Feb 13 19:12:16.783902 containerd[1503]: time="2025-02-13T19:12:16.783484085Z" level=info msg="TearDown network for sandbox \"0be20e2c65924f7b5393bc8b6a79f35d5bf8cca7f34009d6201de4852924939b\" successfully" Feb 13 19:12:16.784624 containerd[1503]: time="2025-02-13T19:12:16.784586278Z" level=info msg="StopPodSandbox for \"0be20e2c65924f7b5393bc8b6a79f35d5bf8cca7f34009d6201de4852924939b\" returns successfully" Feb 13 19:12:16.785219 kubelet[2828]: I0213 19:12:16.785189 2828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af85f8d50ee74cce2c7e2cee2af5922ba7bbd4077fc92d81b183d5575fad1329" Feb 13 19:12:16.791042 containerd[1503]: time="2025-02-13T19:12:16.788923930Z" level=info msg="StopPodSandbox for \"af85f8d50ee74cce2c7e2cee2af5922ba7bbd4077fc92d81b183d5575fad1329\"" Feb 13 19:12:16.791042 containerd[1503]: time="2025-02-13T19:12:16.789822877Z" level=info msg="Ensure that sandbox af85f8d50ee74cce2c7e2cee2af5922ba7bbd4077fc92d81b183d5575fad1329 in task-service has been cleanup successfully" Feb 13 19:12:16.791042 containerd[1503]: time="2025-02-13T19:12:16.790311252Z" level=info msg="StopPodSandbox for \"fa6e4b9a2b53f8e5e8150bdfa0fe88ef1d51405b9b5e7e6d9683c6adb0b078ab\"" Feb 13 19:12:16.792273 containerd[1503]: time="2025-02-13T19:12:16.792245030Z" level=info msg="TearDown network for sandbox \"fa6e4b9a2b53f8e5e8150bdfa0fe88ef1d51405b9b5e7e6d9683c6adb0b078ab\" successfully" Feb 13 19:12:16.792273 containerd[1503]: time="2025-02-13T19:12:16.792271511Z" level=info msg="StopPodSandbox for \"fa6e4b9a2b53f8e5e8150bdfa0fe88ef1d51405b9b5e7e6d9683c6adb0b078ab\" returns successfully" Feb 13 19:12:16.792984 containerd[1503]: time="2025-02-13T19:12:16.790406574Z" level=info msg="TearDown network for sandbox \"af85f8d50ee74cce2c7e2cee2af5922ba7bbd4077fc92d81b183d5575fad1329\" successfully" Feb 13 19:12:16.792984 containerd[1503]: time="2025-02-13T19:12:16.792867209Z" level=info msg="StopPodSandbox for \"af85f8d50ee74cce2c7e2cee2af5922ba7bbd4077fc92d81b183d5575fad1329\" returns successfully" Feb 13 19:12:16.796033 containerd[1503]: time="2025-02-13T19:12:16.795993183Z" level=info msg="StopPodSandbox for \"ae9615caa56ac5be11c6b7cc88aac5196ade400655d7a9313d485332a9d3820e\"" Feb 13 19:12:16.798750 containerd[1503]: time="2025-02-13T19:12:16.798654624Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-78bd6d67c9-b259q,Uid:436348d7-7832-43e1-8780-623c434b2666,Namespace:calico-system,Attempt:3,}" Feb 13 19:12:16.799031 containerd[1503]: time="2025-02-13T19:12:16.798933152Z" level=info msg="TearDown network for sandbox \"ae9615caa56ac5be11c6b7cc88aac5196ade400655d7a9313d485332a9d3820e\" successfully" Feb 13 19:12:16.799031 containerd[1503]: time="2025-02-13T19:12:16.798954513Z" level=info msg="StopPodSandbox for \"ae9615caa56ac5be11c6b7cc88aac5196ade400655d7a9313d485332a9d3820e\" returns successfully" Feb 13 19:12:16.803139 containerd[1503]: time="2025-02-13T19:12:16.803093638Z" level=info msg="StopPodSandbox for \"930563c746d8b91d524a61a614694c47905765a47fb87cb32a681d5b35060a78\"" Feb 13 19:12:16.803266 containerd[1503]: time="2025-02-13T19:12:16.803204801Z" level=info msg="TearDown network for sandbox \"930563c746d8b91d524a61a614694c47905765a47fb87cb32a681d5b35060a78\" successfully" Feb 13 19:12:16.803266 containerd[1503]: time="2025-02-13T19:12:16.803217562Z" level=info msg="StopPodSandbox for \"930563c746d8b91d524a61a614694c47905765a47fb87cb32a681d5b35060a78\" returns successfully" Feb 13 19:12:16.804171 containerd[1503]: time="2025-02-13T19:12:16.803868382Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-84956f4cf-72cgp,Uid:96785f8f-b4e3-456b-9a0d-bcc142e6beaf,Namespace:calico-apiserver,Attempt:3,}" Feb 13 19:12:16.805185 kubelet[2828]: I0213 19:12:16.804447 2828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab072ffc918efc3594006a69ddbbd17de37178f95f1103e4fa6b31bb95fd9306" Feb 13 19:12:16.805826 containerd[1503]: time="2025-02-13T19:12:16.805663996Z" level=info msg="StopPodSandbox for \"ab072ffc918efc3594006a69ddbbd17de37178f95f1103e4fa6b31bb95fd9306\"" Feb 13 19:12:16.806717 containerd[1503]: time="2025-02-13T19:12:16.806264254Z" level=info msg="Ensure that sandbox ab072ffc918efc3594006a69ddbbd17de37178f95f1103e4fa6b31bb95fd9306 in task-service has been cleanup successfully" Feb 13 19:12:16.807603 containerd[1503]: time="2025-02-13T19:12:16.807523532Z" level=info msg="TearDown network for sandbox \"ab072ffc918efc3594006a69ddbbd17de37178f95f1103e4fa6b31bb95fd9306\" successfully" Feb 13 19:12:16.807682 containerd[1503]: time="2025-02-13T19:12:16.807667696Z" level=info msg="StopPodSandbox for \"ab072ffc918efc3594006a69ddbbd17de37178f95f1103e4fa6b31bb95fd9306\" returns successfully" Feb 13 19:12:16.808293 containerd[1503]: time="2025-02-13T19:12:16.808257794Z" level=info msg="StopPodSandbox for \"9581f7ff6fcd5f0bba90e1d0d2984cd0efa6fe4d50a06a4c1b20f3939cb09713\"" Feb 13 19:12:16.808428 containerd[1503]: time="2025-02-13T19:12:16.808393958Z" level=info msg="TearDown network for sandbox \"9581f7ff6fcd5f0bba90e1d0d2984cd0efa6fe4d50a06a4c1b20f3939cb09713\" successfully" Feb 13 19:12:16.808428 containerd[1503]: time="2025-02-13T19:12:16.808410439Z" level=info msg="StopPodSandbox for \"9581f7ff6fcd5f0bba90e1d0d2984cd0efa6fe4d50a06a4c1b20f3939cb09713\" returns successfully" Feb 13 19:12:16.808848 containerd[1503]: time="2025-02-13T19:12:16.808819811Z" level=info msg="StopPodSandbox for \"0c13c2a989169406cb4cea286d09a0b055b98c8bf7adaf3c864364728278a3a5\"" Feb 13 19:12:16.809139 containerd[1503]: time="2025-02-13T19:12:16.809116820Z" level=info msg="TearDown network for sandbox \"0c13c2a989169406cb4cea286d09a0b055b98c8bf7adaf3c864364728278a3a5\" successfully" Feb 13 19:12:16.809468 containerd[1503]: time="2025-02-13T19:12:16.809415829Z" level=info msg="StopPodSandbox for \"0c13c2a989169406cb4cea286d09a0b055b98c8bf7adaf3c864364728278a3a5\" returns successfully" Feb 13 19:12:16.810999 containerd[1503]: time="2025-02-13T19:12:16.810746790Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-84956f4cf-p6mr2,Uid:05e39f6f-0a83-4b62-8cff-b0d4799f8f89,Namespace:calico-apiserver,Attempt:3,}" Feb 13 19:12:17.025284 containerd[1503]: time="2025-02-13T19:12:17.025239521Z" level=error msg="Failed to destroy network for sandbox \"7242b70e1006e9ac2f114fcee5b8cb6221151d9cbbf425c7398a098f6fd891ad\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:17.025971 containerd[1503]: time="2025-02-13T19:12:17.025913782Z" level=error msg="encountered an error cleaning up failed sandbox \"7242b70e1006e9ac2f114fcee5b8cb6221151d9cbbf425c7398a098f6fd891ad\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:17.026560 containerd[1503]: time="2025-02-13T19:12:17.026519040Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-kkqg5,Uid:5cf24d69-eb75-4cb3-8437-024193f516f5,Namespace:kube-system,Attempt:3,} failed, error" error="failed to setup network for sandbox \"7242b70e1006e9ac2f114fcee5b8cb6221151d9cbbf425c7398a098f6fd891ad\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:17.027525 kubelet[2828]: E0213 19:12:17.027458 2828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7242b70e1006e9ac2f114fcee5b8cb6221151d9cbbf425c7398a098f6fd891ad\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:17.027613 kubelet[2828]: E0213 19:12:17.027562 2828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7242b70e1006e9ac2f114fcee5b8cb6221151d9cbbf425c7398a098f6fd891ad\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-kkqg5" Feb 13 19:12:17.027613 kubelet[2828]: E0213 19:12:17.027587 2828 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7242b70e1006e9ac2f114fcee5b8cb6221151d9cbbf425c7398a098f6fd891ad\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-kkqg5" Feb 13 19:12:17.027680 kubelet[2828]: E0213 19:12:17.027641 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-kkqg5_kube-system(5cf24d69-eb75-4cb3-8437-024193f516f5)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-kkqg5_kube-system(5cf24d69-eb75-4cb3-8437-024193f516f5)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"7242b70e1006e9ac2f114fcee5b8cb6221151d9cbbf425c7398a098f6fd891ad\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-kkqg5" podUID="5cf24d69-eb75-4cb3-8437-024193f516f5" Feb 13 19:12:17.052151 containerd[1503]: time="2025-02-13T19:12:17.051924214Z" level=error msg="Failed to destroy network for sandbox \"c6b1e16315492197341e134074aec5720ae8f11ba5d44a3e6e3a0312f21d145d\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:17.052556 containerd[1503]: time="2025-02-13T19:12:17.051951535Z" level=error msg="Failed to destroy network for sandbox \"68ee287533ff713dee0f8970cee78f6429754d9b99d2574db1ab310eb93f45f4\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:17.053114 containerd[1503]: time="2025-02-13T19:12:17.052965806Z" level=error msg="encountered an error cleaning up failed sandbox \"c6b1e16315492197341e134074aec5720ae8f11ba5d44a3e6e3a0312f21d145d\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:17.053114 containerd[1503]: time="2025-02-13T19:12:17.053049529Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-6v9pr,Uid:b9c8452b-31d9-47bc-9697-dd4f293d71b9,Namespace:calico-system,Attempt:3,} failed, error" error="failed to setup network for sandbox \"c6b1e16315492197341e134074aec5720ae8f11ba5d44a3e6e3a0312f21d145d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:17.053326 kubelet[2828]: E0213 19:12:17.053284 2828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c6b1e16315492197341e134074aec5720ae8f11ba5d44a3e6e3a0312f21d145d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:17.053389 kubelet[2828]: E0213 19:12:17.053346 2828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c6b1e16315492197341e134074aec5720ae8f11ba5d44a3e6e3a0312f21d145d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-6v9pr" Feb 13 19:12:17.053389 kubelet[2828]: E0213 19:12:17.053365 2828 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c6b1e16315492197341e134074aec5720ae8f11ba5d44a3e6e3a0312f21d145d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-6v9pr" Feb 13 19:12:17.053438 kubelet[2828]: E0213 19:12:17.053414 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-6v9pr_calico-system(b9c8452b-31d9-47bc-9697-dd4f293d71b9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-6v9pr_calico-system(b9c8452b-31d9-47bc-9697-dd4f293d71b9)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"c6b1e16315492197341e134074aec5720ae8f11ba5d44a3e6e3a0312f21d145d\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-6v9pr" podUID="b9c8452b-31d9-47bc-9697-dd4f293d71b9" Feb 13 19:12:17.054155 containerd[1503]: time="2025-02-13T19:12:17.054001118Z" level=error msg="encountered an error cleaning up failed sandbox \"68ee287533ff713dee0f8970cee78f6429754d9b99d2574db1ab310eb93f45f4\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:17.054436 containerd[1503]: time="2025-02-13T19:12:17.054266126Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-vx9cw,Uid:eb0c3c87-077a-4b2b-a168-5fd9131bccd3,Namespace:kube-system,Attempt:3,} failed, error" error="failed to setup network for sandbox \"68ee287533ff713dee0f8970cee78f6429754d9b99d2574db1ab310eb93f45f4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:17.056461 kubelet[2828]: E0213 19:12:17.056416 2828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"68ee287533ff713dee0f8970cee78f6429754d9b99d2574db1ab310eb93f45f4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:17.056559 kubelet[2828]: E0213 19:12:17.056478 2828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"68ee287533ff713dee0f8970cee78f6429754d9b99d2574db1ab310eb93f45f4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-vx9cw" Feb 13 19:12:17.056559 kubelet[2828]: E0213 19:12:17.056497 2828 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"68ee287533ff713dee0f8970cee78f6429754d9b99d2574db1ab310eb93f45f4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-vx9cw" Feb 13 19:12:17.056640 kubelet[2828]: E0213 19:12:17.056555 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-vx9cw_kube-system(eb0c3c87-077a-4b2b-a168-5fd9131bccd3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-vx9cw_kube-system(eb0c3c87-077a-4b2b-a168-5fd9131bccd3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"68ee287533ff713dee0f8970cee78f6429754d9b99d2574db1ab310eb93f45f4\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-vx9cw" podUID="eb0c3c87-077a-4b2b-a168-5fd9131bccd3" Feb 13 19:12:17.071266 containerd[1503]: time="2025-02-13T19:12:17.071030997Z" level=error msg="Failed to destroy network for sandbox \"f1ae97b8cb19d334ba8268033afc16640078623540002cf3f5996a895113dd8b\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:17.071746 containerd[1503]: time="2025-02-13T19:12:17.071647535Z" level=error msg="encountered an error cleaning up failed sandbox \"f1ae97b8cb19d334ba8268033afc16640078623540002cf3f5996a895113dd8b\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:17.071746 containerd[1503]: time="2025-02-13T19:12:17.071723978Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-78bd6d67c9-b259q,Uid:436348d7-7832-43e1-8780-623c434b2666,Namespace:calico-system,Attempt:3,} failed, error" error="failed to setup network for sandbox \"f1ae97b8cb19d334ba8268033afc16640078623540002cf3f5996a895113dd8b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:17.072814 kubelet[2828]: E0213 19:12:17.072199 2828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f1ae97b8cb19d334ba8268033afc16640078623540002cf3f5996a895113dd8b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:17.072814 kubelet[2828]: E0213 19:12:17.072298 2828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f1ae97b8cb19d334ba8268033afc16640078623540002cf3f5996a895113dd8b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-78bd6d67c9-b259q" Feb 13 19:12:17.072814 kubelet[2828]: E0213 19:12:17.072321 2828 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f1ae97b8cb19d334ba8268033afc16640078623540002cf3f5996a895113dd8b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-78bd6d67c9-b259q" Feb 13 19:12:17.073020 kubelet[2828]: E0213 19:12:17.072375 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-78bd6d67c9-b259q_calico-system(436348d7-7832-43e1-8780-623c434b2666)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-78bd6d67c9-b259q_calico-system(436348d7-7832-43e1-8780-623c434b2666)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"f1ae97b8cb19d334ba8268033afc16640078623540002cf3f5996a895113dd8b\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-78bd6d67c9-b259q" podUID="436348d7-7832-43e1-8780-623c434b2666" Feb 13 19:12:17.088396 containerd[1503]: time="2025-02-13T19:12:17.088191920Z" level=error msg="Failed to destroy network for sandbox \"96f8dbfc99d905b595d74b0f3390d27d4b0e99b3143e0cc72edd14238522b44f\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:17.088986 containerd[1503]: time="2025-02-13T19:12:17.088838019Z" level=error msg="encountered an error cleaning up failed sandbox \"96f8dbfc99d905b595d74b0f3390d27d4b0e99b3143e0cc72edd14238522b44f\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:17.088986 containerd[1503]: time="2025-02-13T19:12:17.088916262Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-84956f4cf-p6mr2,Uid:05e39f6f-0a83-4b62-8cff-b0d4799f8f89,Namespace:calico-apiserver,Attempt:3,} failed, error" error="failed to setup network for sandbox \"96f8dbfc99d905b595d74b0f3390d27d4b0e99b3143e0cc72edd14238522b44f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:17.089400 kubelet[2828]: E0213 19:12:17.089344 2828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"96f8dbfc99d905b595d74b0f3390d27d4b0e99b3143e0cc72edd14238522b44f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:17.089484 kubelet[2828]: E0213 19:12:17.089407 2828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"96f8dbfc99d905b595d74b0f3390d27d4b0e99b3143e0cc72edd14238522b44f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-84956f4cf-p6mr2" Feb 13 19:12:17.089484 kubelet[2828]: E0213 19:12:17.089443 2828 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"96f8dbfc99d905b595d74b0f3390d27d4b0e99b3143e0cc72edd14238522b44f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-84956f4cf-p6mr2" Feb 13 19:12:17.089548 kubelet[2828]: E0213 19:12:17.089488 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-84956f4cf-p6mr2_calico-apiserver(05e39f6f-0a83-4b62-8cff-b0d4799f8f89)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-84956f4cf-p6mr2_calico-apiserver(05e39f6f-0a83-4b62-8cff-b0d4799f8f89)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"96f8dbfc99d905b595d74b0f3390d27d4b0e99b3143e0cc72edd14238522b44f\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-84956f4cf-p6mr2" podUID="05e39f6f-0a83-4b62-8cff-b0d4799f8f89" Feb 13 19:12:17.104043 containerd[1503]: time="2025-02-13T19:12:17.103991001Z" level=error msg="Failed to destroy network for sandbox \"cf3ac12d3142a48e0052c1dead1d74d6becc46ab9352d1240e51b560f73cb623\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:17.104371 containerd[1503]: time="2025-02-13T19:12:17.104345372Z" level=error msg="encountered an error cleaning up failed sandbox \"cf3ac12d3142a48e0052c1dead1d74d6becc46ab9352d1240e51b560f73cb623\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:17.104436 containerd[1503]: time="2025-02-13T19:12:17.104406934Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-84956f4cf-72cgp,Uid:96785f8f-b4e3-456b-9a0d-bcc142e6beaf,Namespace:calico-apiserver,Attempt:3,} failed, error" error="failed to setup network for sandbox \"cf3ac12d3142a48e0052c1dead1d74d6becc46ab9352d1240e51b560f73cb623\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:17.104794 kubelet[2828]: E0213 19:12:17.104753 2828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"cf3ac12d3142a48e0052c1dead1d74d6becc46ab9352d1240e51b560f73cb623\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:17.104853 kubelet[2828]: E0213 19:12:17.104815 2828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"cf3ac12d3142a48e0052c1dead1d74d6becc46ab9352d1240e51b560f73cb623\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-84956f4cf-72cgp" Feb 13 19:12:17.104853 kubelet[2828]: E0213 19:12:17.104834 2828 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"cf3ac12d3142a48e0052c1dead1d74d6becc46ab9352d1240e51b560f73cb623\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-84956f4cf-72cgp" Feb 13 19:12:17.104946 kubelet[2828]: E0213 19:12:17.104885 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-84956f4cf-72cgp_calico-apiserver(96785f8f-b4e3-456b-9a0d-bcc142e6beaf)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-84956f4cf-72cgp_calico-apiserver(96785f8f-b4e3-456b-9a0d-bcc142e6beaf)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"cf3ac12d3142a48e0052c1dead1d74d6becc46ab9352d1240e51b560f73cb623\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-84956f4cf-72cgp" podUID="96785f8f-b4e3-456b-9a0d-bcc142e6beaf" Feb 13 19:12:17.413253 systemd[1]: run-netns-cni\x2d8af4891a\x2de43c\x2db644\x2dae9b\x2d0b68783b901d.mount: Deactivated successfully. Feb 13 19:12:17.413349 systemd[1]: run-netns-cni\x2dccf6f642\x2df2a1\x2d0d01\x2d1356\x2d8405f6cbc99d.mount: Deactivated successfully. Feb 13 19:12:17.815346 kubelet[2828]: I0213 19:12:17.815296 2828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7242b70e1006e9ac2f114fcee5b8cb6221151d9cbbf425c7398a098f6fd891ad" Feb 13 19:12:17.817256 containerd[1503]: time="2025-02-13T19:12:17.816407192Z" level=info msg="StopPodSandbox for \"7242b70e1006e9ac2f114fcee5b8cb6221151d9cbbf425c7398a098f6fd891ad\"" Feb 13 19:12:17.817256 containerd[1503]: time="2025-02-13T19:12:17.816631279Z" level=info msg="Ensure that sandbox 7242b70e1006e9ac2f114fcee5b8cb6221151d9cbbf425c7398a098f6fd891ad in task-service has been cleanup successfully" Feb 13 19:12:17.821039 systemd[1]: run-netns-cni\x2dc3214a99\x2db9d5\x2ddb65\x2d3b86\x2d508808d09398.mount: Deactivated successfully. Feb 13 19:12:17.822222 containerd[1503]: time="2025-02-13T19:12:17.822144567Z" level=info msg="TearDown network for sandbox \"7242b70e1006e9ac2f114fcee5b8cb6221151d9cbbf425c7398a098f6fd891ad\" successfully" Feb 13 19:12:17.822222 containerd[1503]: time="2025-02-13T19:12:17.822181888Z" level=info msg="StopPodSandbox for \"7242b70e1006e9ac2f114fcee5b8cb6221151d9cbbf425c7398a098f6fd891ad\" returns successfully" Feb 13 19:12:17.823373 containerd[1503]: time="2025-02-13T19:12:17.822763586Z" level=info msg="StopPodSandbox for \"d2cb6c1d9e5335f4d8ee798e79c4a58d442740cd8f9ddfe185709b1cfbd4ea81\"" Feb 13 19:12:17.823373 containerd[1503]: time="2025-02-13T19:12:17.822869869Z" level=info msg="TearDown network for sandbox \"d2cb6c1d9e5335f4d8ee798e79c4a58d442740cd8f9ddfe185709b1cfbd4ea81\" successfully" Feb 13 19:12:17.823373 containerd[1503]: time="2025-02-13T19:12:17.822879510Z" level=info msg="StopPodSandbox for \"d2cb6c1d9e5335f4d8ee798e79c4a58d442740cd8f9ddfe185709b1cfbd4ea81\" returns successfully" Feb 13 19:12:17.824112 containerd[1503]: time="2025-02-13T19:12:17.823917381Z" level=info msg="StopPodSandbox for \"10ae5b7da745ff3939219060da91f4661d2a1136a80736c3234db5b474e3599d\"" Feb 13 19:12:17.824112 containerd[1503]: time="2025-02-13T19:12:17.824023065Z" level=info msg="TearDown network for sandbox \"10ae5b7da745ff3939219060da91f4661d2a1136a80736c3234db5b474e3599d\" successfully" Feb 13 19:12:17.824112 containerd[1503]: time="2025-02-13T19:12:17.824037305Z" level=info msg="StopPodSandbox for \"10ae5b7da745ff3939219060da91f4661d2a1136a80736c3234db5b474e3599d\" returns successfully" Feb 13 19:12:17.830183 containerd[1503]: time="2025-02-13T19:12:17.828922654Z" level=info msg="StopPodSandbox for \"43a5af5ee3597e852f8d0ba977ffb06562d6b40688d779a95c8d495a78df4c29\"" Feb 13 19:12:17.831630 containerd[1503]: time="2025-02-13T19:12:17.831593295Z" level=info msg="TearDown network for sandbox \"43a5af5ee3597e852f8d0ba977ffb06562d6b40688d779a95c8d495a78df4c29\" successfully" Feb 13 19:12:17.831902 containerd[1503]: time="2025-02-13T19:12:17.831772021Z" level=info msg="StopPodSandbox for \"43a5af5ee3597e852f8d0ba977ffb06562d6b40688d779a95c8d495a78df4c29\" returns successfully" Feb 13 19:12:17.832402 containerd[1503]: time="2025-02-13T19:12:17.832379279Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-kkqg5,Uid:5cf24d69-eb75-4cb3-8437-024193f516f5,Namespace:kube-system,Attempt:4,}" Feb 13 19:12:17.836220 kubelet[2828]: I0213 19:12:17.836003 2828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1ae97b8cb19d334ba8268033afc16640078623540002cf3f5996a895113dd8b" Feb 13 19:12:17.842449 containerd[1503]: time="2025-02-13T19:12:17.842157657Z" level=info msg="StopPodSandbox for \"f1ae97b8cb19d334ba8268033afc16640078623540002cf3f5996a895113dd8b\"" Feb 13 19:12:17.843706 containerd[1503]: time="2025-02-13T19:12:17.842641352Z" level=info msg="Ensure that sandbox f1ae97b8cb19d334ba8268033afc16640078623540002cf3f5996a895113dd8b in task-service has been cleanup successfully" Feb 13 19:12:17.849626 systemd[1]: run-netns-cni\x2d3c4411d8\x2d5982\x2d26b4\x2dd132\x2d37f1cb4a59e3.mount: Deactivated successfully. Feb 13 19:12:17.850882 containerd[1503]: time="2025-02-13T19:12:17.850511192Z" level=info msg="TearDown network for sandbox \"f1ae97b8cb19d334ba8268033afc16640078623540002cf3f5996a895113dd8b\" successfully" Feb 13 19:12:17.850882 containerd[1503]: time="2025-02-13T19:12:17.850561713Z" level=info msg="StopPodSandbox for \"f1ae97b8cb19d334ba8268033afc16640078623540002cf3f5996a895113dd8b\" returns successfully" Feb 13 19:12:17.851369 containerd[1503]: time="2025-02-13T19:12:17.851221373Z" level=info msg="StopPodSandbox for \"4ffb63334faa7203eea7b2071116973af83410f7340f4e8946194bccb25fa157\"" Feb 13 19:12:17.851369 containerd[1503]: time="2025-02-13T19:12:17.851308696Z" level=info msg="TearDown network for sandbox \"4ffb63334faa7203eea7b2071116973af83410f7340f4e8946194bccb25fa157\" successfully" Feb 13 19:12:17.851369 containerd[1503]: time="2025-02-13T19:12:17.851318176Z" level=info msg="StopPodSandbox for \"4ffb63334faa7203eea7b2071116973af83410f7340f4e8946194bccb25fa157\" returns successfully" Feb 13 19:12:17.852399 containerd[1503]: time="2025-02-13T19:12:17.852365768Z" level=info msg="StopPodSandbox for \"0be20e2c65924f7b5393bc8b6a79f35d5bf8cca7f34009d6201de4852924939b\"" Feb 13 19:12:17.852708 containerd[1503]: time="2025-02-13T19:12:17.852464811Z" level=info msg="TearDown network for sandbox \"0be20e2c65924f7b5393bc8b6a79f35d5bf8cca7f34009d6201de4852924939b\" successfully" Feb 13 19:12:17.852708 containerd[1503]: time="2025-02-13T19:12:17.852475412Z" level=info msg="StopPodSandbox for \"0be20e2c65924f7b5393bc8b6a79f35d5bf8cca7f34009d6201de4852924939b\" returns successfully" Feb 13 19:12:17.853706 containerd[1503]: time="2025-02-13T19:12:17.853478282Z" level=info msg="StopPodSandbox for \"fa6e4b9a2b53f8e5e8150bdfa0fe88ef1d51405b9b5e7e6d9683c6adb0b078ab\"" Feb 13 19:12:17.853993 containerd[1503]: time="2025-02-13T19:12:17.853844373Z" level=info msg="TearDown network for sandbox \"fa6e4b9a2b53f8e5e8150bdfa0fe88ef1d51405b9b5e7e6d9683c6adb0b078ab\" successfully" Feb 13 19:12:17.853993 containerd[1503]: time="2025-02-13T19:12:17.853859734Z" level=info msg="StopPodSandbox for \"fa6e4b9a2b53f8e5e8150bdfa0fe88ef1d51405b9b5e7e6d9683c6adb0b078ab\" returns successfully" Feb 13 19:12:17.855540 containerd[1503]: time="2025-02-13T19:12:17.855394461Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-78bd6d67c9-b259q,Uid:436348d7-7832-43e1-8780-623c434b2666,Namespace:calico-system,Attempt:4,}" Feb 13 19:12:17.857134 kubelet[2828]: I0213 19:12:17.856424 2828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf3ac12d3142a48e0052c1dead1d74d6becc46ab9352d1240e51b560f73cb623" Feb 13 19:12:17.857690 containerd[1503]: time="2025-02-13T19:12:17.857331880Z" level=info msg="StopPodSandbox for \"cf3ac12d3142a48e0052c1dead1d74d6becc46ab9352d1240e51b560f73cb623\"" Feb 13 19:12:17.857690 containerd[1503]: time="2025-02-13T19:12:17.857594728Z" level=info msg="Ensure that sandbox cf3ac12d3142a48e0052c1dead1d74d6becc46ab9352d1240e51b560f73cb623 in task-service has been cleanup successfully" Feb 13 19:12:17.864986 containerd[1503]: time="2025-02-13T19:12:17.863450466Z" level=info msg="TearDown network for sandbox \"cf3ac12d3142a48e0052c1dead1d74d6becc46ab9352d1240e51b560f73cb623\" successfully" Feb 13 19:12:17.864986 containerd[1503]: time="2025-02-13T19:12:17.863672633Z" level=info msg="StopPodSandbox for \"cf3ac12d3142a48e0052c1dead1d74d6becc46ab9352d1240e51b560f73cb623\" returns successfully" Feb 13 19:12:17.865648 systemd[1]: run-netns-cni\x2da2434f2a\x2d6d7d\x2db258\x2df667\x2de3765ee272b5.mount: Deactivated successfully. Feb 13 19:12:17.868197 containerd[1503]: time="2025-02-13T19:12:17.868132649Z" level=info msg="StopPodSandbox for \"af85f8d50ee74cce2c7e2cee2af5922ba7bbd4077fc92d81b183d5575fad1329\"" Feb 13 19:12:17.869274 containerd[1503]: time="2025-02-13T19:12:17.868677625Z" level=info msg="TearDown network for sandbox \"af85f8d50ee74cce2c7e2cee2af5922ba7bbd4077fc92d81b183d5575fad1329\" successfully" Feb 13 19:12:17.869274 containerd[1503]: time="2025-02-13T19:12:17.868710866Z" level=info msg="StopPodSandbox for \"af85f8d50ee74cce2c7e2cee2af5922ba7bbd4077fc92d81b183d5575fad1329\" returns successfully" Feb 13 19:12:17.874882 containerd[1503]: time="2025-02-13T19:12:17.874183433Z" level=info msg="StopPodSandbox for \"ae9615caa56ac5be11c6b7cc88aac5196ade400655d7a9313d485332a9d3820e\"" Feb 13 19:12:17.874882 containerd[1503]: time="2025-02-13T19:12:17.874290236Z" level=info msg="TearDown network for sandbox \"ae9615caa56ac5be11c6b7cc88aac5196ade400655d7a9313d485332a9d3820e\" successfully" Feb 13 19:12:17.874882 containerd[1503]: time="2025-02-13T19:12:17.874301117Z" level=info msg="StopPodSandbox for \"ae9615caa56ac5be11c6b7cc88aac5196ade400655d7a9313d485332a9d3820e\" returns successfully" Feb 13 19:12:17.877383 containerd[1503]: time="2025-02-13T19:12:17.877341689Z" level=info msg="StopPodSandbox for \"930563c746d8b91d524a61a614694c47905765a47fb87cb32a681d5b35060a78\"" Feb 13 19:12:17.877507 containerd[1503]: time="2025-02-13T19:12:17.877455773Z" level=info msg="TearDown network for sandbox \"930563c746d8b91d524a61a614694c47905765a47fb87cb32a681d5b35060a78\" successfully" Feb 13 19:12:17.877507 containerd[1503]: time="2025-02-13T19:12:17.877465853Z" level=info msg="StopPodSandbox for \"930563c746d8b91d524a61a614694c47905765a47fb87cb32a681d5b35060a78\" returns successfully" Feb 13 19:12:17.878508 kubelet[2828]: I0213 19:12:17.878083 2828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96f8dbfc99d905b595d74b0f3390d27d4b0e99b3143e0cc72edd14238522b44f" Feb 13 19:12:17.880078 containerd[1503]: time="2025-02-13T19:12:17.879243987Z" level=info msg="StopPodSandbox for \"96f8dbfc99d905b595d74b0f3390d27d4b0e99b3143e0cc72edd14238522b44f\"" Feb 13 19:12:17.881208 containerd[1503]: time="2025-02-13T19:12:17.879685041Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-84956f4cf-72cgp,Uid:96785f8f-b4e3-456b-9a0d-bcc142e6beaf,Namespace:calico-apiserver,Attempt:4,}" Feb 13 19:12:17.881860 containerd[1503]: time="2025-02-13T19:12:17.881825706Z" level=info msg="Ensure that sandbox 96f8dbfc99d905b595d74b0f3390d27d4b0e99b3143e0cc72edd14238522b44f in task-service has been cleanup successfully" Feb 13 19:12:17.884741 containerd[1503]: time="2025-02-13T19:12:17.884624151Z" level=info msg="TearDown network for sandbox \"96f8dbfc99d905b595d74b0f3390d27d4b0e99b3143e0cc72edd14238522b44f\" successfully" Feb 13 19:12:17.884741 containerd[1503]: time="2025-02-13T19:12:17.884654192Z" level=info msg="StopPodSandbox for \"96f8dbfc99d905b595d74b0f3390d27d4b0e99b3143e0cc72edd14238522b44f\" returns successfully" Feb 13 19:12:17.886701 containerd[1503]: time="2025-02-13T19:12:17.886665694Z" level=info msg="StopPodSandbox for \"ab072ffc918efc3594006a69ddbbd17de37178f95f1103e4fa6b31bb95fd9306\"" Feb 13 19:12:17.886825 containerd[1503]: time="2025-02-13T19:12:17.886767177Z" level=info msg="TearDown network for sandbox \"ab072ffc918efc3594006a69ddbbd17de37178f95f1103e4fa6b31bb95fd9306\" successfully" Feb 13 19:12:17.886825 containerd[1503]: time="2025-02-13T19:12:17.886781057Z" level=info msg="StopPodSandbox for \"ab072ffc918efc3594006a69ddbbd17de37178f95f1103e4fa6b31bb95fd9306\" returns successfully" Feb 13 19:12:17.888259 kubelet[2828]: I0213 19:12:17.887967 2828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="68ee287533ff713dee0f8970cee78f6429754d9b99d2574db1ab310eb93f45f4" Feb 13 19:12:17.889106 containerd[1503]: time="2025-02-13T19:12:17.888998245Z" level=info msg="StopPodSandbox for \"9581f7ff6fcd5f0bba90e1d0d2984cd0efa6fe4d50a06a4c1b20f3939cb09713\"" Feb 13 19:12:17.889252 containerd[1503]: time="2025-02-13T19:12:17.889173450Z" level=info msg="TearDown network for sandbox \"9581f7ff6fcd5f0bba90e1d0d2984cd0efa6fe4d50a06a4c1b20f3939cb09713\" successfully" Feb 13 19:12:17.889252 containerd[1503]: time="2025-02-13T19:12:17.889207171Z" level=info msg="StopPodSandbox for \"9581f7ff6fcd5f0bba90e1d0d2984cd0efa6fe4d50a06a4c1b20f3939cb09713\" returns successfully" Feb 13 19:12:17.893493 containerd[1503]: time="2025-02-13T19:12:17.892043337Z" level=info msg="StopPodSandbox for \"0c13c2a989169406cb4cea286d09a0b055b98c8bf7adaf3c864364728278a3a5\"" Feb 13 19:12:17.894026 containerd[1503]: time="2025-02-13T19:12:17.893616545Z" level=info msg="TearDown network for sandbox \"0c13c2a989169406cb4cea286d09a0b055b98c8bf7adaf3c864364728278a3a5\" successfully" Feb 13 19:12:17.894026 containerd[1503]: time="2025-02-13T19:12:17.893638906Z" level=info msg="StopPodSandbox for \"0c13c2a989169406cb4cea286d09a0b055b98c8bf7adaf3c864364728278a3a5\" returns successfully" Feb 13 19:12:17.895384 containerd[1503]: time="2025-02-13T19:12:17.894860343Z" level=info msg="StopPodSandbox for \"68ee287533ff713dee0f8970cee78f6429754d9b99d2574db1ab310eb93f45f4\"" Feb 13 19:12:17.895649 containerd[1503]: time="2025-02-13T19:12:17.895445761Z" level=info msg="Ensure that sandbox 68ee287533ff713dee0f8970cee78f6429754d9b99d2574db1ab310eb93f45f4 in task-service has been cleanup successfully" Feb 13 19:12:17.896639 containerd[1503]: time="2025-02-13T19:12:17.896421551Z" level=info msg="TearDown network for sandbox \"68ee287533ff713dee0f8970cee78f6429754d9b99d2574db1ab310eb93f45f4\" successfully" Feb 13 19:12:17.896639 containerd[1503]: time="2025-02-13T19:12:17.896636837Z" level=info msg="StopPodSandbox for \"68ee287533ff713dee0f8970cee78f6429754d9b99d2574db1ab310eb93f45f4\" returns successfully" Feb 13 19:12:17.900461 containerd[1503]: time="2025-02-13T19:12:17.898292248Z" level=info msg="StopPodSandbox for \"a657753a39333ef1c8ee279e3321a3d04ca609cd0499c88ae5589336776d3f50\"" Feb 13 19:12:17.900461 containerd[1503]: time="2025-02-13T19:12:17.898384611Z" level=info msg="TearDown network for sandbox \"a657753a39333ef1c8ee279e3321a3d04ca609cd0499c88ae5589336776d3f50\" successfully" Feb 13 19:12:17.900461 containerd[1503]: time="2025-02-13T19:12:17.898395331Z" level=info msg="StopPodSandbox for \"a657753a39333ef1c8ee279e3321a3d04ca609cd0499c88ae5589336776d3f50\" returns successfully" Feb 13 19:12:17.900461 containerd[1503]: time="2025-02-13T19:12:17.898504574Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-84956f4cf-p6mr2,Uid:05e39f6f-0a83-4b62-8cff-b0d4799f8f89,Namespace:calico-apiserver,Attempt:4,}" Feb 13 19:12:17.901212 containerd[1503]: time="2025-02-13T19:12:17.901173416Z" level=info msg="StopPodSandbox for \"07aed5c023700866ebeb2a6224ac11566918d60f5a140685f482d3a85f4cd555\"" Feb 13 19:12:17.901295 containerd[1503]: time="2025-02-13T19:12:17.901274779Z" level=info msg="TearDown network for sandbox \"07aed5c023700866ebeb2a6224ac11566918d60f5a140685f482d3a85f4cd555\" successfully" Feb 13 19:12:17.901295 containerd[1503]: time="2025-02-13T19:12:17.901289659Z" level=info msg="StopPodSandbox for \"07aed5c023700866ebeb2a6224ac11566918d60f5a140685f482d3a85f4cd555\" returns successfully" Feb 13 19:12:17.904296 containerd[1503]: time="2025-02-13T19:12:17.904259310Z" level=info msg="StopPodSandbox for \"0c31ed484cb1e7593659d8fd1f850350d6b3b9e01ed8530e08cf66bc47a72b73\"" Feb 13 19:12:17.904476 containerd[1503]: time="2025-02-13T19:12:17.904350553Z" level=info msg="TearDown network for sandbox \"0c31ed484cb1e7593659d8fd1f850350d6b3b9e01ed8530e08cf66bc47a72b73\" successfully" Feb 13 19:12:17.904476 containerd[1503]: time="2025-02-13T19:12:17.904362193Z" level=info msg="StopPodSandbox for \"0c31ed484cb1e7593659d8fd1f850350d6b3b9e01ed8530e08cf66bc47a72b73\" returns successfully" Feb 13 19:12:17.905419 kubelet[2828]: I0213 19:12:17.905392 2828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6b1e16315492197341e134074aec5720ae8f11ba5d44a3e6e3a0312f21d145d" Feb 13 19:12:17.905664 containerd[1503]: time="2025-02-13T19:12:17.905617711Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-vx9cw,Uid:eb0c3c87-077a-4b2b-a168-5fd9131bccd3,Namespace:kube-system,Attempt:4,}" Feb 13 19:12:17.907604 containerd[1503]: time="2025-02-13T19:12:17.906818948Z" level=info msg="StopPodSandbox for \"c6b1e16315492197341e134074aec5720ae8f11ba5d44a3e6e3a0312f21d145d\"" Feb 13 19:12:17.907688 containerd[1503]: time="2025-02-13T19:12:17.907633493Z" level=info msg="Ensure that sandbox c6b1e16315492197341e134074aec5720ae8f11ba5d44a3e6e3a0312f21d145d in task-service has been cleanup successfully" Feb 13 19:12:17.910671 containerd[1503]: time="2025-02-13T19:12:17.909935723Z" level=info msg="TearDown network for sandbox \"c6b1e16315492197341e134074aec5720ae8f11ba5d44a3e6e3a0312f21d145d\" successfully" Feb 13 19:12:17.910671 containerd[1503]: time="2025-02-13T19:12:17.910446018Z" level=info msg="StopPodSandbox for \"c6b1e16315492197341e134074aec5720ae8f11ba5d44a3e6e3a0312f21d145d\" returns successfully" Feb 13 19:12:17.913000 containerd[1503]: time="2025-02-13T19:12:17.912951295Z" level=info msg="StopPodSandbox for \"4642b0f0428148a51fe1afb0c6a261eff1112eb99bbdf06b50e264a81fa407b3\"" Feb 13 19:12:17.913179 containerd[1503]: time="2025-02-13T19:12:17.913157261Z" level=info msg="TearDown network for sandbox \"4642b0f0428148a51fe1afb0c6a261eff1112eb99bbdf06b50e264a81fa407b3\" successfully" Feb 13 19:12:17.913179 containerd[1503]: time="2025-02-13T19:12:17.913175702Z" level=info msg="StopPodSandbox for \"4642b0f0428148a51fe1afb0c6a261eff1112eb99bbdf06b50e264a81fa407b3\" returns successfully" Feb 13 19:12:17.913994 containerd[1503]: time="2025-02-13T19:12:17.913913764Z" level=info msg="StopPodSandbox for \"4f219600df3d01ca237f481814d18c92c5997c6c501af4d70cb45cd41d1a6efc\"" Feb 13 19:12:17.914839 containerd[1503]: time="2025-02-13T19:12:17.914714668Z" level=info msg="TearDown network for sandbox \"4f219600df3d01ca237f481814d18c92c5997c6c501af4d70cb45cd41d1a6efc\" successfully" Feb 13 19:12:17.914922 containerd[1503]: time="2025-02-13T19:12:17.914854273Z" level=info msg="StopPodSandbox for \"4f219600df3d01ca237f481814d18c92c5997c6c501af4d70cb45cd41d1a6efc\" returns successfully" Feb 13 19:12:17.919713 containerd[1503]: time="2025-02-13T19:12:17.919136683Z" level=info msg="StopPodSandbox for \"1f6b42d2735cedb916bc24bf2bb557d2ff72928d0c487a93a3dd46e88aa55abe\"" Feb 13 19:12:17.919713 containerd[1503]: time="2025-02-13T19:12:17.919246727Z" level=info msg="TearDown network for sandbox \"1f6b42d2735cedb916bc24bf2bb557d2ff72928d0c487a93a3dd46e88aa55abe\" successfully" Feb 13 19:12:17.919713 containerd[1503]: time="2025-02-13T19:12:17.919256487Z" level=info msg="StopPodSandbox for \"1f6b42d2735cedb916bc24bf2bb557d2ff72928d0c487a93a3dd46e88aa55abe\" returns successfully" Feb 13 19:12:17.921545 containerd[1503]: time="2025-02-13T19:12:17.921485395Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-6v9pr,Uid:b9c8452b-31d9-47bc-9697-dd4f293d71b9,Namespace:calico-system,Attempt:4,}" Feb 13 19:12:18.134594 containerd[1503]: time="2025-02-13T19:12:18.134461156Z" level=error msg="Failed to destroy network for sandbox \"931e447af991e370f310777638932ea566d0d3e47f8ec0fb6ba2b0fe9e6ff401\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:18.136725 containerd[1503]: time="2025-02-13T19:12:18.136680624Z" level=error msg="encountered an error cleaning up failed sandbox \"931e447af991e370f310777638932ea566d0d3e47f8ec0fb6ba2b0fe9e6ff401\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:18.137715 containerd[1503]: time="2025-02-13T19:12:18.137403926Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-kkqg5,Uid:5cf24d69-eb75-4cb3-8437-024193f516f5,Namespace:kube-system,Attempt:4,} failed, error" error="failed to setup network for sandbox \"931e447af991e370f310777638932ea566d0d3e47f8ec0fb6ba2b0fe9e6ff401\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:18.138291 kubelet[2828]: E0213 19:12:18.138153 2828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"931e447af991e370f310777638932ea566d0d3e47f8ec0fb6ba2b0fe9e6ff401\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:18.138525 kubelet[2828]: E0213 19:12:18.138487 2828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"931e447af991e370f310777638932ea566d0d3e47f8ec0fb6ba2b0fe9e6ff401\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-kkqg5" Feb 13 19:12:18.138593 kubelet[2828]: E0213 19:12:18.138524 2828 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"931e447af991e370f310777638932ea566d0d3e47f8ec0fb6ba2b0fe9e6ff401\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-kkqg5" Feb 13 19:12:18.139154 kubelet[2828]: E0213 19:12:18.138631 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-kkqg5_kube-system(5cf24d69-eb75-4cb3-8437-024193f516f5)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-kkqg5_kube-system(5cf24d69-eb75-4cb3-8437-024193f516f5)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"931e447af991e370f310777638932ea566d0d3e47f8ec0fb6ba2b0fe9e6ff401\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-kkqg5" podUID="5cf24d69-eb75-4cb3-8437-024193f516f5" Feb 13 19:12:18.169447 containerd[1503]: time="2025-02-13T19:12:18.169393029Z" level=error msg="Failed to destroy network for sandbox \"f101bce244f680aa9fab77bf6802cf73b54ffba4caa118c44f2e7b203fb0dc84\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:18.169995 containerd[1503]: time="2025-02-13T19:12:18.169955206Z" level=error msg="encountered an error cleaning up failed sandbox \"f101bce244f680aa9fab77bf6802cf73b54ffba4caa118c44f2e7b203fb0dc84\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:18.170068 containerd[1503]: time="2025-02-13T19:12:18.170037529Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-78bd6d67c9-b259q,Uid:436348d7-7832-43e1-8780-623c434b2666,Namespace:calico-system,Attempt:4,} failed, error" error="failed to setup network for sandbox \"f101bce244f680aa9fab77bf6802cf73b54ffba4caa118c44f2e7b203fb0dc84\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:18.170684 kubelet[2828]: E0213 19:12:18.170647 2828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f101bce244f680aa9fab77bf6802cf73b54ffba4caa118c44f2e7b203fb0dc84\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:18.170839 kubelet[2828]: E0213 19:12:18.170818 2828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f101bce244f680aa9fab77bf6802cf73b54ffba4caa118c44f2e7b203fb0dc84\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-78bd6d67c9-b259q" Feb 13 19:12:18.170957 kubelet[2828]: E0213 19:12:18.170940 2828 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f101bce244f680aa9fab77bf6802cf73b54ffba4caa118c44f2e7b203fb0dc84\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-78bd6d67c9-b259q" Feb 13 19:12:18.171178 kubelet[2828]: E0213 19:12:18.171052 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-78bd6d67c9-b259q_calico-system(436348d7-7832-43e1-8780-623c434b2666)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-78bd6d67c9-b259q_calico-system(436348d7-7832-43e1-8780-623c434b2666)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"f101bce244f680aa9fab77bf6802cf73b54ffba4caa118c44f2e7b203fb0dc84\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-78bd6d67c9-b259q" podUID="436348d7-7832-43e1-8780-623c434b2666" Feb 13 19:12:18.195308 containerd[1503]: time="2025-02-13T19:12:18.195248743Z" level=error msg="Failed to destroy network for sandbox \"728705ac6957c83c18f350153e00e11e14122291646534ceb675f3fda8f70cdb\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:18.196046 containerd[1503]: time="2025-02-13T19:12:18.195860601Z" level=error msg="encountered an error cleaning up failed sandbox \"728705ac6957c83c18f350153e00e11e14122291646534ceb675f3fda8f70cdb\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:18.196046 containerd[1503]: time="2025-02-13T19:12:18.195933964Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-6v9pr,Uid:b9c8452b-31d9-47bc-9697-dd4f293d71b9,Namespace:calico-system,Attempt:4,} failed, error" error="failed to setup network for sandbox \"728705ac6957c83c18f350153e00e11e14122291646534ceb675f3fda8f70cdb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:18.196349 kubelet[2828]: E0213 19:12:18.196305 2828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"728705ac6957c83c18f350153e00e11e14122291646534ceb675f3fda8f70cdb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:18.196923 kubelet[2828]: E0213 19:12:18.196628 2828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"728705ac6957c83c18f350153e00e11e14122291646534ceb675f3fda8f70cdb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-6v9pr" Feb 13 19:12:18.196923 kubelet[2828]: E0213 19:12:18.196657 2828 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"728705ac6957c83c18f350153e00e11e14122291646534ceb675f3fda8f70cdb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-6v9pr" Feb 13 19:12:18.196923 kubelet[2828]: E0213 19:12:18.196718 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-6v9pr_calico-system(b9c8452b-31d9-47bc-9697-dd4f293d71b9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-6v9pr_calico-system(b9c8452b-31d9-47bc-9697-dd4f293d71b9)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"728705ac6957c83c18f350153e00e11e14122291646534ceb675f3fda8f70cdb\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-6v9pr" podUID="b9c8452b-31d9-47bc-9697-dd4f293d71b9" Feb 13 19:12:18.202832 containerd[1503]: time="2025-02-13T19:12:18.202143434Z" level=error msg="Failed to destroy network for sandbox \"4d5614a8446f6178d266e8f2147ca224323d91c476899b1e9e66b08c962ad2a3\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:18.204945 containerd[1503]: time="2025-02-13T19:12:18.204900679Z" level=error msg="encountered an error cleaning up failed sandbox \"4d5614a8446f6178d266e8f2147ca224323d91c476899b1e9e66b08c962ad2a3\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:18.206724 containerd[1503]: time="2025-02-13T19:12:18.205154207Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-84956f4cf-72cgp,Uid:96785f8f-b4e3-456b-9a0d-bcc142e6beaf,Namespace:calico-apiserver,Attempt:4,} failed, error" error="failed to setup network for sandbox \"4d5614a8446f6178d266e8f2147ca224323d91c476899b1e9e66b08c962ad2a3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:18.206874 kubelet[2828]: E0213 19:12:18.205653 2828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4d5614a8446f6178d266e8f2147ca224323d91c476899b1e9e66b08c962ad2a3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:18.206874 kubelet[2828]: E0213 19:12:18.205748 2828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4d5614a8446f6178d266e8f2147ca224323d91c476899b1e9e66b08c962ad2a3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-84956f4cf-72cgp" Feb 13 19:12:18.206874 kubelet[2828]: E0213 19:12:18.205772 2828 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4d5614a8446f6178d266e8f2147ca224323d91c476899b1e9e66b08c962ad2a3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-84956f4cf-72cgp" Feb 13 19:12:18.206961 kubelet[2828]: E0213 19:12:18.205824 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-84956f4cf-72cgp_calico-apiserver(96785f8f-b4e3-456b-9a0d-bcc142e6beaf)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-84956f4cf-72cgp_calico-apiserver(96785f8f-b4e3-456b-9a0d-bcc142e6beaf)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"4d5614a8446f6178d266e8f2147ca224323d91c476899b1e9e66b08c962ad2a3\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-84956f4cf-72cgp" podUID="96785f8f-b4e3-456b-9a0d-bcc142e6beaf" Feb 13 19:12:18.219334 containerd[1503]: time="2025-02-13T19:12:18.219289441Z" level=error msg="Failed to destroy network for sandbox \"b3134abe6d3978218dbaec771f16517d173324bf774a94fc2b6f62660051520d\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:18.219975 containerd[1503]: time="2025-02-13T19:12:18.219942781Z" level=error msg="encountered an error cleaning up failed sandbox \"b3134abe6d3978218dbaec771f16517d173324bf774a94fc2b6f62660051520d\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:18.220374 containerd[1503]: time="2025-02-13T19:12:18.220345313Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-84956f4cf-p6mr2,Uid:05e39f6f-0a83-4b62-8cff-b0d4799f8f89,Namespace:calico-apiserver,Attempt:4,} failed, error" error="failed to setup network for sandbox \"b3134abe6d3978218dbaec771f16517d173324bf774a94fc2b6f62660051520d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:18.222093 kubelet[2828]: E0213 19:12:18.221706 2828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b3134abe6d3978218dbaec771f16517d173324bf774a94fc2b6f62660051520d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:18.222093 kubelet[2828]: E0213 19:12:18.221780 2828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b3134abe6d3978218dbaec771f16517d173324bf774a94fc2b6f62660051520d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-84956f4cf-p6mr2" Feb 13 19:12:18.222093 kubelet[2828]: E0213 19:12:18.221798 2828 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b3134abe6d3978218dbaec771f16517d173324bf774a94fc2b6f62660051520d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-84956f4cf-p6mr2" Feb 13 19:12:18.222414 kubelet[2828]: E0213 19:12:18.221843 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-84956f4cf-p6mr2_calico-apiserver(05e39f6f-0a83-4b62-8cff-b0d4799f8f89)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-84956f4cf-p6mr2_calico-apiserver(05e39f6f-0a83-4b62-8cff-b0d4799f8f89)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"b3134abe6d3978218dbaec771f16517d173324bf774a94fc2b6f62660051520d\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-84956f4cf-p6mr2" podUID="05e39f6f-0a83-4b62-8cff-b0d4799f8f89" Feb 13 19:12:18.227966 containerd[1503]: time="2025-02-13T19:12:18.227830703Z" level=error msg="Failed to destroy network for sandbox \"602a514f998cb50210ddc95537cfa35ed74b37741d39df2aa29785c3996317a9\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:18.228668 containerd[1503]: time="2025-02-13T19:12:18.228327798Z" level=error msg="encountered an error cleaning up failed sandbox \"602a514f998cb50210ddc95537cfa35ed74b37741d39df2aa29785c3996317a9\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:18.228668 containerd[1503]: time="2025-02-13T19:12:18.228451842Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-vx9cw,Uid:eb0c3c87-077a-4b2b-a168-5fd9131bccd3,Namespace:kube-system,Attempt:4,} failed, error" error="failed to setup network for sandbox \"602a514f998cb50210ddc95537cfa35ed74b37741d39df2aa29785c3996317a9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:18.228772 kubelet[2828]: E0213 19:12:18.228698 2828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"602a514f998cb50210ddc95537cfa35ed74b37741d39df2aa29785c3996317a9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:18.228772 kubelet[2828]: E0213 19:12:18.228753 2828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"602a514f998cb50210ddc95537cfa35ed74b37741d39df2aa29785c3996317a9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-vx9cw" Feb 13 19:12:18.228833 kubelet[2828]: E0213 19:12:18.228780 2828 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"602a514f998cb50210ddc95537cfa35ed74b37741d39df2aa29785c3996317a9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-vx9cw" Feb 13 19:12:18.228833 kubelet[2828]: E0213 19:12:18.228815 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-vx9cw_kube-system(eb0c3c87-077a-4b2b-a168-5fd9131bccd3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-vx9cw_kube-system(eb0c3c87-077a-4b2b-a168-5fd9131bccd3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"602a514f998cb50210ddc95537cfa35ed74b37741d39df2aa29785c3996317a9\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-vx9cw" podUID="eb0c3c87-077a-4b2b-a168-5fd9131bccd3" Feb 13 19:12:18.412289 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-931e447af991e370f310777638932ea566d0d3e47f8ec0fb6ba2b0fe9e6ff401-shm.mount: Deactivated successfully. Feb 13 19:12:18.412394 systemd[1]: run-netns-cni\x2da78054b8\x2da61b\x2da0b2\x2dec52\x2de948bb550c37.mount: Deactivated successfully. Feb 13 19:12:18.412450 systemd[1]: run-netns-cni\x2daac0dbc2\x2d141b\x2de6ed\x2df125\x2d360305159177.mount: Deactivated successfully. Feb 13 19:12:18.412504 systemd[1]: run-netns-cni\x2d65d502f2\x2da00f\x2d5b80\x2d592d\x2dcb2e8d6806e2.mount: Deactivated successfully. Feb 13 19:12:18.914100 kubelet[2828]: I0213 19:12:18.910810 2828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="602a514f998cb50210ddc95537cfa35ed74b37741d39df2aa29785c3996317a9" Feb 13 19:12:18.913839 systemd[1]: run-netns-cni\x2d55b69c77\x2d3862\x2d767c\x2de941\x2dcc8dfaae7e25.mount: Deactivated successfully. Feb 13 19:12:18.914614 containerd[1503]: time="2025-02-13T19:12:18.911308330Z" level=info msg="StopPodSandbox for \"602a514f998cb50210ddc95537cfa35ed74b37741d39df2aa29785c3996317a9\"" Feb 13 19:12:18.914614 containerd[1503]: time="2025-02-13T19:12:18.911485096Z" level=info msg="Ensure that sandbox 602a514f998cb50210ddc95537cfa35ed74b37741d39df2aa29785c3996317a9 in task-service has been cleanup successfully" Feb 13 19:12:18.916949 containerd[1503]: time="2025-02-13T19:12:18.916698816Z" level=info msg="TearDown network for sandbox \"602a514f998cb50210ddc95537cfa35ed74b37741d39df2aa29785c3996317a9\" successfully" Feb 13 19:12:18.916949 containerd[1503]: time="2025-02-13T19:12:18.916731497Z" level=info msg="StopPodSandbox for \"602a514f998cb50210ddc95537cfa35ed74b37741d39df2aa29785c3996317a9\" returns successfully" Feb 13 19:12:18.917620 containerd[1503]: time="2025-02-13T19:12:18.917545042Z" level=info msg="StopPodSandbox for \"68ee287533ff713dee0f8970cee78f6429754d9b99d2574db1ab310eb93f45f4\"" Feb 13 19:12:18.918106 containerd[1503]: time="2025-02-13T19:12:18.918006536Z" level=info msg="TearDown network for sandbox \"68ee287533ff713dee0f8970cee78f6429754d9b99d2574db1ab310eb93f45f4\" successfully" Feb 13 19:12:18.918106 containerd[1503]: time="2025-02-13T19:12:18.918052257Z" level=info msg="StopPodSandbox for \"68ee287533ff713dee0f8970cee78f6429754d9b99d2574db1ab310eb93f45f4\" returns successfully" Feb 13 19:12:18.918989 containerd[1503]: time="2025-02-13T19:12:18.918959645Z" level=info msg="StopPodSandbox for \"a657753a39333ef1c8ee279e3321a3d04ca609cd0499c88ae5589336776d3f50\"" Feb 13 19:12:18.919114 containerd[1503]: time="2025-02-13T19:12:18.919050408Z" level=info msg="TearDown network for sandbox \"a657753a39333ef1c8ee279e3321a3d04ca609cd0499c88ae5589336776d3f50\" successfully" Feb 13 19:12:18.919198 containerd[1503]: time="2025-02-13T19:12:18.919111890Z" level=info msg="StopPodSandbox for \"a657753a39333ef1c8ee279e3321a3d04ca609cd0499c88ae5589336776d3f50\" returns successfully" Feb 13 19:12:18.919835 containerd[1503]: time="2025-02-13T19:12:18.919766630Z" level=info msg="StopPodSandbox for \"07aed5c023700866ebeb2a6224ac11566918d60f5a140685f482d3a85f4cd555\"" Feb 13 19:12:18.919892 containerd[1503]: time="2025-02-13T19:12:18.919855913Z" level=info msg="TearDown network for sandbox \"07aed5c023700866ebeb2a6224ac11566918d60f5a140685f482d3a85f4cd555\" successfully" Feb 13 19:12:18.919892 containerd[1503]: time="2025-02-13T19:12:18.919867673Z" level=info msg="StopPodSandbox for \"07aed5c023700866ebeb2a6224ac11566918d60f5a140685f482d3a85f4cd555\" returns successfully" Feb 13 19:12:18.920361 containerd[1503]: time="2025-02-13T19:12:18.920327847Z" level=info msg="StopPodSandbox for \"0c31ed484cb1e7593659d8fd1f850350d6b3b9e01ed8530e08cf66bc47a72b73\"" Feb 13 19:12:18.920432 containerd[1503]: time="2025-02-13T19:12:18.920403249Z" level=info msg="TearDown network for sandbox \"0c31ed484cb1e7593659d8fd1f850350d6b3b9e01ed8530e08cf66bc47a72b73\" successfully" Feb 13 19:12:18.920432 containerd[1503]: time="2025-02-13T19:12:18.920412770Z" level=info msg="StopPodSandbox for \"0c31ed484cb1e7593659d8fd1f850350d6b3b9e01ed8530e08cf66bc47a72b73\" returns successfully" Feb 13 19:12:18.921979 containerd[1503]: time="2025-02-13T19:12:18.921596246Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-vx9cw,Uid:eb0c3c87-077a-4b2b-a168-5fd9131bccd3,Namespace:kube-system,Attempt:5,}" Feb 13 19:12:18.923329 kubelet[2828]: I0213 19:12:18.923305 2828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="728705ac6957c83c18f350153e00e11e14122291646534ceb675f3fda8f70cdb" Feb 13 19:12:18.924344 containerd[1503]: time="2025-02-13T19:12:18.924319130Z" level=info msg="StopPodSandbox for \"728705ac6957c83c18f350153e00e11e14122291646534ceb675f3fda8f70cdb\"" Feb 13 19:12:18.924961 containerd[1503]: time="2025-02-13T19:12:18.924850986Z" level=info msg="Ensure that sandbox 728705ac6957c83c18f350153e00e11e14122291646534ceb675f3fda8f70cdb in task-service has been cleanup successfully" Feb 13 19:12:18.925387 containerd[1503]: time="2025-02-13T19:12:18.925353441Z" level=info msg="TearDown network for sandbox \"728705ac6957c83c18f350153e00e11e14122291646534ceb675f3fda8f70cdb\" successfully" Feb 13 19:12:18.925682 containerd[1503]: time="2025-02-13T19:12:18.925500926Z" level=info msg="StopPodSandbox for \"728705ac6957c83c18f350153e00e11e14122291646534ceb675f3fda8f70cdb\" returns successfully" Feb 13 19:12:18.928175 systemd[1]: run-netns-cni\x2da89773f8\x2da83d\x2d751e\x2d1a7a\x2dfd859dcd63d6.mount: Deactivated successfully. Feb 13 19:12:18.935879 containerd[1503]: time="2025-02-13T19:12:18.935605236Z" level=info msg="StopPodSandbox for \"c6b1e16315492197341e134074aec5720ae8f11ba5d44a3e6e3a0312f21d145d\"" Feb 13 19:12:18.936437 containerd[1503]: time="2025-02-13T19:12:18.936079971Z" level=info msg="TearDown network for sandbox \"c6b1e16315492197341e134074aec5720ae8f11ba5d44a3e6e3a0312f21d145d\" successfully" Feb 13 19:12:18.936437 containerd[1503]: time="2025-02-13T19:12:18.936100971Z" level=info msg="StopPodSandbox for \"c6b1e16315492197341e134074aec5720ae8f11ba5d44a3e6e3a0312f21d145d\" returns successfully" Feb 13 19:12:18.936898 containerd[1503]: time="2025-02-13T19:12:18.936873755Z" level=info msg="StopPodSandbox for \"4642b0f0428148a51fe1afb0c6a261eff1112eb99bbdf06b50e264a81fa407b3\"" Feb 13 19:12:18.937530 containerd[1503]: time="2025-02-13T19:12:18.937509415Z" level=info msg="TearDown network for sandbox \"4642b0f0428148a51fe1afb0c6a261eff1112eb99bbdf06b50e264a81fa407b3\" successfully" Feb 13 19:12:18.938026 containerd[1503]: time="2025-02-13T19:12:18.938002750Z" level=info msg="StopPodSandbox for \"4642b0f0428148a51fe1afb0c6a261eff1112eb99bbdf06b50e264a81fa407b3\" returns successfully" Feb 13 19:12:18.940836 containerd[1503]: time="2025-02-13T19:12:18.940619150Z" level=info msg="StopPodSandbox for \"4f219600df3d01ca237f481814d18c92c5997c6c501af4d70cb45cd41d1a6efc\"" Feb 13 19:12:18.942540 containerd[1503]: time="2025-02-13T19:12:18.942496888Z" level=info msg="TearDown network for sandbox \"4f219600df3d01ca237f481814d18c92c5997c6c501af4d70cb45cd41d1a6efc\" successfully" Feb 13 19:12:18.944343 containerd[1503]: time="2025-02-13T19:12:18.943613642Z" level=info msg="StopPodSandbox for \"4f219600df3d01ca237f481814d18c92c5997c6c501af4d70cb45cd41d1a6efc\" returns successfully" Feb 13 19:12:18.944490 containerd[1503]: time="2025-02-13T19:12:18.944430067Z" level=info msg="StopPodSandbox for \"1f6b42d2735cedb916bc24bf2bb557d2ff72928d0c487a93a3dd46e88aa55abe\"" Feb 13 19:12:18.944583 containerd[1503]: time="2025-02-13T19:12:18.944523030Z" level=info msg="TearDown network for sandbox \"1f6b42d2735cedb916bc24bf2bb557d2ff72928d0c487a93a3dd46e88aa55abe\" successfully" Feb 13 19:12:18.944583 containerd[1503]: time="2025-02-13T19:12:18.944553191Z" level=info msg="StopPodSandbox for \"1f6b42d2735cedb916bc24bf2bb557d2ff72928d0c487a93a3dd46e88aa55abe\" returns successfully" Feb 13 19:12:18.945592 kubelet[2828]: I0213 19:12:18.944958 2828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="931e447af991e370f310777638932ea566d0d3e47f8ec0fb6ba2b0fe9e6ff401" Feb 13 19:12:18.946525 containerd[1503]: time="2025-02-13T19:12:18.946418448Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-6v9pr,Uid:b9c8452b-31d9-47bc-9697-dd4f293d71b9,Namespace:calico-system,Attempt:5,}" Feb 13 19:12:18.950377 containerd[1503]: time="2025-02-13T19:12:18.950336489Z" level=info msg="StopPodSandbox for \"931e447af991e370f310777638932ea566d0d3e47f8ec0fb6ba2b0fe9e6ff401\"" Feb 13 19:12:18.950525 containerd[1503]: time="2025-02-13T19:12:18.950502174Z" level=info msg="Ensure that sandbox 931e447af991e370f310777638932ea566d0d3e47f8ec0fb6ba2b0fe9e6ff401 in task-service has been cleanup successfully" Feb 13 19:12:18.952075 containerd[1503]: time="2025-02-13T19:12:18.951090752Z" level=info msg="TearDown network for sandbox \"931e447af991e370f310777638932ea566d0d3e47f8ec0fb6ba2b0fe9e6ff401\" successfully" Feb 13 19:12:18.952075 containerd[1503]: time="2025-02-13T19:12:18.952041861Z" level=info msg="StopPodSandbox for \"931e447af991e370f310777638932ea566d0d3e47f8ec0fb6ba2b0fe9e6ff401\" returns successfully" Feb 13 19:12:18.956651 systemd[1]: run-netns-cni\x2da00db4d7\x2df613\x2d0ee7\x2d4fff\x2d91ce28d4aa25.mount: Deactivated successfully. Feb 13 19:12:18.958598 containerd[1503]: time="2025-02-13T19:12:18.956848848Z" level=info msg="StopPodSandbox for \"7242b70e1006e9ac2f114fcee5b8cb6221151d9cbbf425c7398a098f6fd891ad\"" Feb 13 19:12:18.958598 containerd[1503]: time="2025-02-13T19:12:18.956952252Z" level=info msg="TearDown network for sandbox \"7242b70e1006e9ac2f114fcee5b8cb6221151d9cbbf425c7398a098f6fd891ad\" successfully" Feb 13 19:12:18.958598 containerd[1503]: time="2025-02-13T19:12:18.956962372Z" level=info msg="StopPodSandbox for \"7242b70e1006e9ac2f114fcee5b8cb6221151d9cbbf425c7398a098f6fd891ad\" returns successfully" Feb 13 19:12:18.960003 containerd[1503]: time="2025-02-13T19:12:18.959968104Z" level=info msg="StopPodSandbox for \"d2cb6c1d9e5335f4d8ee798e79c4a58d442740cd8f9ddfe185709b1cfbd4ea81\"" Feb 13 19:12:18.960159 containerd[1503]: time="2025-02-13T19:12:18.960128029Z" level=info msg="TearDown network for sandbox \"d2cb6c1d9e5335f4d8ee798e79c4a58d442740cd8f9ddfe185709b1cfbd4ea81\" successfully" Feb 13 19:12:18.960159 containerd[1503]: time="2025-02-13T19:12:18.960140950Z" level=info msg="StopPodSandbox for \"d2cb6c1d9e5335f4d8ee798e79c4a58d442740cd8f9ddfe185709b1cfbd4ea81\" returns successfully" Feb 13 19:12:18.960918 containerd[1503]: time="2025-02-13T19:12:18.960889253Z" level=info msg="StopPodSandbox for \"10ae5b7da745ff3939219060da91f4661d2a1136a80736c3234db5b474e3599d\"" Feb 13 19:12:18.961616 containerd[1503]: time="2025-02-13T19:12:18.961368947Z" level=info msg="TearDown network for sandbox \"10ae5b7da745ff3939219060da91f4661d2a1136a80736c3234db5b474e3599d\" successfully" Feb 13 19:12:18.961616 containerd[1503]: time="2025-02-13T19:12:18.961395948Z" level=info msg="StopPodSandbox for \"10ae5b7da745ff3939219060da91f4661d2a1136a80736c3234db5b474e3599d\" returns successfully" Feb 13 19:12:18.961737 containerd[1503]: time="2025-02-13T19:12:18.961716278Z" level=info msg="StopPodSandbox for \"43a5af5ee3597e852f8d0ba977ffb06562d6b40688d779a95c8d495a78df4c29\"" Feb 13 19:12:18.962540 containerd[1503]: time="2025-02-13T19:12:18.962240334Z" level=info msg="TearDown network for sandbox \"43a5af5ee3597e852f8d0ba977ffb06562d6b40688d779a95c8d495a78df4c29\" successfully" Feb 13 19:12:18.962540 containerd[1503]: time="2025-02-13T19:12:18.962264415Z" level=info msg="StopPodSandbox for \"43a5af5ee3597e852f8d0ba977ffb06562d6b40688d779a95c8d495a78df4c29\" returns successfully" Feb 13 19:12:18.963291 containerd[1503]: time="2025-02-13T19:12:18.963254125Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-kkqg5,Uid:5cf24d69-eb75-4cb3-8437-024193f516f5,Namespace:kube-system,Attempt:5,}" Feb 13 19:12:18.963782 kubelet[2828]: I0213 19:12:18.963754 2828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f101bce244f680aa9fab77bf6802cf73b54ffba4caa118c44f2e7b203fb0dc84" Feb 13 19:12:18.965456 containerd[1503]: time="2025-02-13T19:12:18.965263867Z" level=info msg="StopPodSandbox for \"f101bce244f680aa9fab77bf6802cf73b54ffba4caa118c44f2e7b203fb0dc84\"" Feb 13 19:12:18.965779 containerd[1503]: time="2025-02-13T19:12:18.965664919Z" level=info msg="Ensure that sandbox f101bce244f680aa9fab77bf6802cf73b54ffba4caa118c44f2e7b203fb0dc84 in task-service has been cleanup successfully" Feb 13 19:12:18.971379 containerd[1503]: time="2025-02-13T19:12:18.970840438Z" level=info msg="TearDown network for sandbox \"f101bce244f680aa9fab77bf6802cf73b54ffba4caa118c44f2e7b203fb0dc84\" successfully" Feb 13 19:12:18.972288 containerd[1503]: time="2025-02-13T19:12:18.972214240Z" level=info msg="StopPodSandbox for \"f101bce244f680aa9fab77bf6802cf73b54ffba4caa118c44f2e7b203fb0dc84\" returns successfully" Feb 13 19:12:18.975877 containerd[1503]: time="2025-02-13T19:12:18.975299055Z" level=info msg="StopPodSandbox for \"f1ae97b8cb19d334ba8268033afc16640078623540002cf3f5996a895113dd8b\"" Feb 13 19:12:18.975877 containerd[1503]: time="2025-02-13T19:12:18.975398298Z" level=info msg="TearDown network for sandbox \"f1ae97b8cb19d334ba8268033afc16640078623540002cf3f5996a895113dd8b\" successfully" Feb 13 19:12:18.975877 containerd[1503]: time="2025-02-13T19:12:18.975407938Z" level=info msg="StopPodSandbox for \"f1ae97b8cb19d334ba8268033afc16640078623540002cf3f5996a895113dd8b\" returns successfully" Feb 13 19:12:18.978052 containerd[1503]: time="2025-02-13T19:12:18.977906055Z" level=info msg="StopPodSandbox for \"4ffb63334faa7203eea7b2071116973af83410f7340f4e8946194bccb25fa157\"" Feb 13 19:12:18.978457 containerd[1503]: time="2025-02-13T19:12:18.978421351Z" level=info msg="TearDown network for sandbox \"4ffb63334faa7203eea7b2071116973af83410f7340f4e8946194bccb25fa157\" successfully" Feb 13 19:12:18.979765 containerd[1503]: time="2025-02-13T19:12:18.979722231Z" level=info msg="StopPodSandbox for \"4ffb63334faa7203eea7b2071116973af83410f7340f4e8946194bccb25fa157\" returns successfully" Feb 13 19:12:18.982542 containerd[1503]: time="2025-02-13T19:12:18.981261958Z" level=info msg="StopPodSandbox for \"0be20e2c65924f7b5393bc8b6a79f35d5bf8cca7f34009d6201de4852924939b\"" Feb 13 19:12:18.982542 containerd[1503]: time="2025-02-13T19:12:18.981835896Z" level=info msg="TearDown network for sandbox \"0be20e2c65924f7b5393bc8b6a79f35d5bf8cca7f34009d6201de4852924939b\" successfully" Feb 13 19:12:18.982542 containerd[1503]: time="2025-02-13T19:12:18.981958460Z" level=info msg="StopPodSandbox for \"0be20e2c65924f7b5393bc8b6a79f35d5bf8cca7f34009d6201de4852924939b\" returns successfully" Feb 13 19:12:18.984004 containerd[1503]: time="2025-02-13T19:12:18.983967041Z" level=info msg="StopPodSandbox for \"fa6e4b9a2b53f8e5e8150bdfa0fe88ef1d51405b9b5e7e6d9683c6adb0b078ab\"" Feb 13 19:12:18.984125 containerd[1503]: time="2025-02-13T19:12:18.984078765Z" level=info msg="TearDown network for sandbox \"fa6e4b9a2b53f8e5e8150bdfa0fe88ef1d51405b9b5e7e6d9683c6adb0b078ab\" successfully" Feb 13 19:12:18.984125 containerd[1503]: time="2025-02-13T19:12:18.984090685Z" level=info msg="StopPodSandbox for \"fa6e4b9a2b53f8e5e8150bdfa0fe88ef1d51405b9b5e7e6d9683c6adb0b078ab\" returns successfully" Feb 13 19:12:18.984686 kubelet[2828]: I0213 19:12:18.984501 2828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d5614a8446f6178d266e8f2147ca224323d91c476899b1e9e66b08c962ad2a3" Feb 13 19:12:18.986036 containerd[1503]: time="2025-02-13T19:12:18.986000784Z" level=info msg="StopPodSandbox for \"4d5614a8446f6178d266e8f2147ca224323d91c476899b1e9e66b08c962ad2a3\"" Feb 13 19:12:18.986501 containerd[1503]: time="2025-02-13T19:12:18.986473318Z" level=info msg="Ensure that sandbox 4d5614a8446f6178d266e8f2147ca224323d91c476899b1e9e66b08c962ad2a3 in task-service has been cleanup successfully" Feb 13 19:12:18.987565 containerd[1503]: time="2025-02-13T19:12:18.987352105Z" level=info msg="TearDown network for sandbox \"4d5614a8446f6178d266e8f2147ca224323d91c476899b1e9e66b08c962ad2a3\" successfully" Feb 13 19:12:18.987565 containerd[1503]: time="2025-02-13T19:12:18.987474269Z" level=info msg="StopPodSandbox for \"4d5614a8446f6178d266e8f2147ca224323d91c476899b1e9e66b08c962ad2a3\" returns successfully" Feb 13 19:12:18.988376 containerd[1503]: time="2025-02-13T19:12:18.987381666Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-78bd6d67c9-b259q,Uid:436348d7-7832-43e1-8780-623c434b2666,Namespace:calico-system,Attempt:5,}" Feb 13 19:12:18.988901 containerd[1503]: time="2025-02-13T19:12:18.988870272Z" level=info msg="StopPodSandbox for \"cf3ac12d3142a48e0052c1dead1d74d6becc46ab9352d1240e51b560f73cb623\"" Feb 13 19:12:18.989141 containerd[1503]: time="2025-02-13T19:12:18.989104359Z" level=info msg="TearDown network for sandbox \"cf3ac12d3142a48e0052c1dead1d74d6becc46ab9352d1240e51b560f73cb623\" successfully" Feb 13 19:12:18.989141 containerd[1503]: time="2025-02-13T19:12:18.989117479Z" level=info msg="StopPodSandbox for \"cf3ac12d3142a48e0052c1dead1d74d6becc46ab9352d1240e51b560f73cb623\" returns successfully" Feb 13 19:12:18.990261 containerd[1503]: time="2025-02-13T19:12:18.990163951Z" level=info msg="StopPodSandbox for \"af85f8d50ee74cce2c7e2cee2af5922ba7bbd4077fc92d81b183d5575fad1329\"" Feb 13 19:12:18.990261 containerd[1503]: time="2025-02-13T19:12:18.990260114Z" level=info msg="TearDown network for sandbox \"af85f8d50ee74cce2c7e2cee2af5922ba7bbd4077fc92d81b183d5575fad1329\" successfully" Feb 13 19:12:18.990828 containerd[1503]: time="2025-02-13T19:12:18.990271835Z" level=info msg="StopPodSandbox for \"af85f8d50ee74cce2c7e2cee2af5922ba7bbd4077fc92d81b183d5575fad1329\" returns successfully" Feb 13 19:12:18.993356 containerd[1503]: time="2025-02-13T19:12:18.993322608Z" level=info msg="StopPodSandbox for \"ae9615caa56ac5be11c6b7cc88aac5196ade400655d7a9313d485332a9d3820e\"" Feb 13 19:12:18.994153 containerd[1503]: time="2025-02-13T19:12:18.993424772Z" level=info msg="TearDown network for sandbox \"ae9615caa56ac5be11c6b7cc88aac5196ade400655d7a9313d485332a9d3820e\" successfully" Feb 13 19:12:18.994153 containerd[1503]: time="2025-02-13T19:12:18.993441252Z" level=info msg="StopPodSandbox for \"ae9615caa56ac5be11c6b7cc88aac5196ade400655d7a9313d485332a9d3820e\" returns successfully" Feb 13 19:12:18.995326 containerd[1503]: time="2025-02-13T19:12:18.995021821Z" level=info msg="StopPodSandbox for \"930563c746d8b91d524a61a614694c47905765a47fb87cb32a681d5b35060a78\"" Feb 13 19:12:18.995798 containerd[1503]: time="2025-02-13T19:12:18.995722762Z" level=info msg="TearDown network for sandbox \"930563c746d8b91d524a61a614694c47905765a47fb87cb32a681d5b35060a78\" successfully" Feb 13 19:12:18.997192 containerd[1503]: time="2025-02-13T19:12:18.997156926Z" level=info msg="StopPodSandbox for \"930563c746d8b91d524a61a614694c47905765a47fb87cb32a681d5b35060a78\" returns successfully" Feb 13 19:12:19.001184 containerd[1503]: time="2025-02-13T19:12:19.000193779Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-84956f4cf-72cgp,Uid:96785f8f-b4e3-456b-9a0d-bcc142e6beaf,Namespace:calico-apiserver,Attempt:5,}" Feb 13 19:12:19.005438 kubelet[2828]: I0213 19:12:19.005404 2828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3134abe6d3978218dbaec771f16517d173324bf774a94fc2b6f62660051520d" Feb 13 19:12:19.007526 containerd[1503]: time="2025-02-13T19:12:19.006924987Z" level=info msg="StopPodSandbox for \"b3134abe6d3978218dbaec771f16517d173324bf774a94fc2b6f62660051520d\"" Feb 13 19:12:19.007526 containerd[1503]: time="2025-02-13T19:12:19.007119473Z" level=info msg="Ensure that sandbox b3134abe6d3978218dbaec771f16517d173324bf774a94fc2b6f62660051520d in task-service has been cleanup successfully" Feb 13 19:12:19.013378 containerd[1503]: time="2025-02-13T19:12:19.012995855Z" level=info msg="TearDown network for sandbox \"b3134abe6d3978218dbaec771f16517d173324bf774a94fc2b6f62660051520d\" successfully" Feb 13 19:12:19.013780 containerd[1503]: time="2025-02-13T19:12:19.013748839Z" level=info msg="StopPodSandbox for \"b3134abe6d3978218dbaec771f16517d173324bf774a94fc2b6f62660051520d\" returns successfully" Feb 13 19:12:19.015397 containerd[1503]: time="2025-02-13T19:12:19.015351128Z" level=info msg="StopPodSandbox for \"96f8dbfc99d905b595d74b0f3390d27d4b0e99b3143e0cc72edd14238522b44f\"" Feb 13 19:12:19.015720 containerd[1503]: time="2025-02-13T19:12:19.015701539Z" level=info msg="TearDown network for sandbox \"96f8dbfc99d905b595d74b0f3390d27d4b0e99b3143e0cc72edd14238522b44f\" successfully" Feb 13 19:12:19.016104 containerd[1503]: time="2025-02-13T19:12:19.016079511Z" level=info msg="StopPodSandbox for \"96f8dbfc99d905b595d74b0f3390d27d4b0e99b3143e0cc72edd14238522b44f\" returns successfully" Feb 13 19:12:19.017368 containerd[1503]: time="2025-02-13T19:12:19.017095382Z" level=info msg="StopPodSandbox for \"ab072ffc918efc3594006a69ddbbd17de37178f95f1103e4fa6b31bb95fd9306\"" Feb 13 19:12:19.017368 containerd[1503]: time="2025-02-13T19:12:19.017192905Z" level=info msg="TearDown network for sandbox \"ab072ffc918efc3594006a69ddbbd17de37178f95f1103e4fa6b31bb95fd9306\" successfully" Feb 13 19:12:19.017368 containerd[1503]: time="2025-02-13T19:12:19.017203665Z" level=info msg="StopPodSandbox for \"ab072ffc918efc3594006a69ddbbd17de37178f95f1103e4fa6b31bb95fd9306\" returns successfully" Feb 13 19:12:19.021356 containerd[1503]: time="2025-02-13T19:12:19.021311392Z" level=info msg="StopPodSandbox for \"9581f7ff6fcd5f0bba90e1d0d2984cd0efa6fe4d50a06a4c1b20f3939cb09713\"" Feb 13 19:12:19.026116 containerd[1503]: time="2025-02-13T19:12:19.026036619Z" level=info msg="TearDown network for sandbox \"9581f7ff6fcd5f0bba90e1d0d2984cd0efa6fe4d50a06a4c1b20f3939cb09713\" successfully" Feb 13 19:12:19.026116 containerd[1503]: time="2025-02-13T19:12:19.026090220Z" level=info msg="StopPodSandbox for \"9581f7ff6fcd5f0bba90e1d0d2984cd0efa6fe4d50a06a4c1b20f3939cb09713\" returns successfully" Feb 13 19:12:19.028078 containerd[1503]: time="2025-02-13T19:12:19.028031160Z" level=info msg="StopPodSandbox for \"0c13c2a989169406cb4cea286d09a0b055b98c8bf7adaf3c864364728278a3a5\"" Feb 13 19:12:19.028174 containerd[1503]: time="2025-02-13T19:12:19.028144604Z" level=info msg="TearDown network for sandbox \"0c13c2a989169406cb4cea286d09a0b055b98c8bf7adaf3c864364728278a3a5\" successfully" Feb 13 19:12:19.028174 containerd[1503]: time="2025-02-13T19:12:19.028156404Z" level=info msg="StopPodSandbox for \"0c13c2a989169406cb4cea286d09a0b055b98c8bf7adaf3c864364728278a3a5\" returns successfully" Feb 13 19:12:19.029635 containerd[1503]: time="2025-02-13T19:12:19.029120114Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-84956f4cf-p6mr2,Uid:05e39f6f-0a83-4b62-8cff-b0d4799f8f89,Namespace:calico-apiserver,Attempt:5,}" Feb 13 19:12:19.208451 containerd[1503]: time="2025-02-13T19:12:19.208311376Z" level=error msg="Failed to destroy network for sandbox \"b86651f4cc54f520cab1a18f98b7a70986a4aa77bdc28bce93539ac71deee2a2\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:19.212887 containerd[1503]: time="2025-02-13T19:12:19.212825756Z" level=error msg="encountered an error cleaning up failed sandbox \"b86651f4cc54f520cab1a18f98b7a70986a4aa77bdc28bce93539ac71deee2a2\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:19.213153 containerd[1503]: time="2025-02-13T19:12:19.212916999Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-84956f4cf-72cgp,Uid:96785f8f-b4e3-456b-9a0d-bcc142e6beaf,Namespace:calico-apiserver,Attempt:5,} failed, error" error="failed to setup network for sandbox \"b86651f4cc54f520cab1a18f98b7a70986a4aa77bdc28bce93539ac71deee2a2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:19.214483 containerd[1503]: time="2025-02-13T19:12:19.214374484Z" level=error msg="Failed to destroy network for sandbox \"76c8fc7995e25f780bca1521611986ac1d0cb18b34d8d5db637de52885bbe0b7\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:19.216288 kubelet[2828]: E0213 19:12:19.215862 2828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b86651f4cc54f520cab1a18f98b7a70986a4aa77bdc28bce93539ac71deee2a2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:19.216288 kubelet[2828]: E0213 19:12:19.215969 2828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b86651f4cc54f520cab1a18f98b7a70986a4aa77bdc28bce93539ac71deee2a2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-84956f4cf-72cgp" Feb 13 19:12:19.216288 kubelet[2828]: E0213 19:12:19.215999 2828 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b86651f4cc54f520cab1a18f98b7a70986a4aa77bdc28bce93539ac71deee2a2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-84956f4cf-72cgp" Feb 13 19:12:19.216520 containerd[1503]: time="2025-02-13T19:12:19.216456508Z" level=error msg="encountered an error cleaning up failed sandbox \"76c8fc7995e25f780bca1521611986ac1d0cb18b34d8d5db637de52885bbe0b7\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:19.217101 containerd[1503]: time="2025-02-13T19:12:19.216543311Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-vx9cw,Uid:eb0c3c87-077a-4b2b-a168-5fd9131bccd3,Namespace:kube-system,Attempt:5,} failed, error" error="failed to setup network for sandbox \"76c8fc7995e25f780bca1521611986ac1d0cb18b34d8d5db637de52885bbe0b7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:19.217217 kubelet[2828]: E0213 19:12:19.216908 2828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"76c8fc7995e25f780bca1521611986ac1d0cb18b34d8d5db637de52885bbe0b7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:19.217217 kubelet[2828]: E0213 19:12:19.216941 2828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"76c8fc7995e25f780bca1521611986ac1d0cb18b34d8d5db637de52885bbe0b7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-vx9cw" Feb 13 19:12:19.217217 kubelet[2828]: E0213 19:12:19.216957 2828 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"76c8fc7995e25f780bca1521611986ac1d0cb18b34d8d5db637de52885bbe0b7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-vx9cw" Feb 13 19:12:19.219081 kubelet[2828]: E0213 19:12:19.218471 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-84956f4cf-72cgp_calico-apiserver(96785f8f-b4e3-456b-9a0d-bcc142e6beaf)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-84956f4cf-72cgp_calico-apiserver(96785f8f-b4e3-456b-9a0d-bcc142e6beaf)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"b86651f4cc54f520cab1a18f98b7a70986a4aa77bdc28bce93539ac71deee2a2\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-84956f4cf-72cgp" podUID="96785f8f-b4e3-456b-9a0d-bcc142e6beaf" Feb 13 19:12:19.219250 kubelet[2828]: E0213 19:12:19.219203 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-vx9cw_kube-system(eb0c3c87-077a-4b2b-a168-5fd9131bccd3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-vx9cw_kube-system(eb0c3c87-077a-4b2b-a168-5fd9131bccd3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"76c8fc7995e25f780bca1521611986ac1d0cb18b34d8d5db637de52885bbe0b7\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-vx9cw" podUID="eb0c3c87-077a-4b2b-a168-5fd9131bccd3" Feb 13 19:12:19.227529 containerd[1503]: time="2025-02-13T19:12:19.227475769Z" level=error msg="Failed to destroy network for sandbox \"77460032187adb446e84e042dcc0956c8b578caedef00c1e669bc2016d01bf85\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:19.229371 containerd[1503]: time="2025-02-13T19:12:19.229318946Z" level=error msg="encountered an error cleaning up failed sandbox \"77460032187adb446e84e042dcc0956c8b578caedef00c1e669bc2016d01bf85\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:19.229483 containerd[1503]: time="2025-02-13T19:12:19.229396268Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-6v9pr,Uid:b9c8452b-31d9-47bc-9697-dd4f293d71b9,Namespace:calico-system,Attempt:5,} failed, error" error="failed to setup network for sandbox \"77460032187adb446e84e042dcc0956c8b578caedef00c1e669bc2016d01bf85\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:19.230006 kubelet[2828]: E0213 19:12:19.229971 2828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"77460032187adb446e84e042dcc0956c8b578caedef00c1e669bc2016d01bf85\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:19.230176 kubelet[2828]: E0213 19:12:19.230152 2828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"77460032187adb446e84e042dcc0956c8b578caedef00c1e669bc2016d01bf85\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-6v9pr" Feb 13 19:12:19.230257 kubelet[2828]: E0213 19:12:19.230241 2828 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"77460032187adb446e84e042dcc0956c8b578caedef00c1e669bc2016d01bf85\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-6v9pr" Feb 13 19:12:19.230393 kubelet[2828]: E0213 19:12:19.230358 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-6v9pr_calico-system(b9c8452b-31d9-47bc-9697-dd4f293d71b9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-6v9pr_calico-system(b9c8452b-31d9-47bc-9697-dd4f293d71b9)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"77460032187adb446e84e042dcc0956c8b578caedef00c1e669bc2016d01bf85\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-6v9pr" podUID="b9c8452b-31d9-47bc-9697-dd4f293d71b9" Feb 13 19:12:19.236007 containerd[1503]: time="2025-02-13T19:12:19.235952871Z" level=error msg="Failed to destroy network for sandbox \"d485d108247618e01bc6b75342565a89c0d97dadc8e9f0f8b8a03154fa8d72b1\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:19.241120 containerd[1503]: time="2025-02-13T19:12:19.241070389Z" level=error msg="encountered an error cleaning up failed sandbox \"d485d108247618e01bc6b75342565a89c0d97dadc8e9f0f8b8a03154fa8d72b1\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:19.241416 containerd[1503]: time="2025-02-13T19:12:19.241304477Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-kkqg5,Uid:5cf24d69-eb75-4cb3-8437-024193f516f5,Namespace:kube-system,Attempt:5,} failed, error" error="failed to setup network for sandbox \"d485d108247618e01bc6b75342565a89c0d97dadc8e9f0f8b8a03154fa8d72b1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:19.241907 kubelet[2828]: E0213 19:12:19.241539 2828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d485d108247618e01bc6b75342565a89c0d97dadc8e9f0f8b8a03154fa8d72b1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:19.241907 kubelet[2828]: E0213 19:12:19.241614 2828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d485d108247618e01bc6b75342565a89c0d97dadc8e9f0f8b8a03154fa8d72b1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-kkqg5" Feb 13 19:12:19.241907 kubelet[2828]: E0213 19:12:19.241634 2828 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d485d108247618e01bc6b75342565a89c0d97dadc8e9f0f8b8a03154fa8d72b1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-kkqg5" Feb 13 19:12:19.242081 kubelet[2828]: E0213 19:12:19.241683 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-kkqg5_kube-system(5cf24d69-eb75-4cb3-8437-024193f516f5)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-kkqg5_kube-system(5cf24d69-eb75-4cb3-8437-024193f516f5)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"d485d108247618e01bc6b75342565a89c0d97dadc8e9f0f8b8a03154fa8d72b1\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-kkqg5" podUID="5cf24d69-eb75-4cb3-8437-024193f516f5" Feb 13 19:12:19.300709 containerd[1503]: time="2025-02-13T19:12:19.300643672Z" level=error msg="Failed to destroy network for sandbox \"6ca053ac29abc2425e4b004e866bb9d2dbabd9431e4a54c02f8b398f38c0da08\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:19.305035 containerd[1503]: time="2025-02-13T19:12:19.304986966Z" level=error msg="encountered an error cleaning up failed sandbox \"6ca053ac29abc2425e4b004e866bb9d2dbabd9431e4a54c02f8b398f38c0da08\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:19.305270 containerd[1503]: time="2025-02-13T19:12:19.305244214Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-78bd6d67c9-b259q,Uid:436348d7-7832-43e1-8780-623c434b2666,Namespace:calico-system,Attempt:5,} failed, error" error="failed to setup network for sandbox \"6ca053ac29abc2425e4b004e866bb9d2dbabd9431e4a54c02f8b398f38c0da08\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:19.307761 kubelet[2828]: E0213 19:12:19.307261 2828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6ca053ac29abc2425e4b004e866bb9d2dbabd9431e4a54c02f8b398f38c0da08\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:19.307761 kubelet[2828]: E0213 19:12:19.307329 2828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6ca053ac29abc2425e4b004e866bb9d2dbabd9431e4a54c02f8b398f38c0da08\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-78bd6d67c9-b259q" Feb 13 19:12:19.307761 kubelet[2828]: E0213 19:12:19.307348 2828 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6ca053ac29abc2425e4b004e866bb9d2dbabd9431e4a54c02f8b398f38c0da08\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-78bd6d67c9-b259q" Feb 13 19:12:19.308049 kubelet[2828]: E0213 19:12:19.307384 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-78bd6d67c9-b259q_calico-system(436348d7-7832-43e1-8780-623c434b2666)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-78bd6d67c9-b259q_calico-system(436348d7-7832-43e1-8780-623c434b2666)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"6ca053ac29abc2425e4b004e866bb9d2dbabd9431e4a54c02f8b398f38c0da08\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-78bd6d67c9-b259q" podUID="436348d7-7832-43e1-8780-623c434b2666" Feb 13 19:12:19.309774 containerd[1503]: time="2025-02-13T19:12:19.309726953Z" level=error msg="Failed to destroy network for sandbox \"15ac125800531efb9e148c216d3f053334f90d5dd6675f1df06b0c64e6f022c5\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:19.310238 containerd[1503]: time="2025-02-13T19:12:19.310211208Z" level=error msg="encountered an error cleaning up failed sandbox \"15ac125800531efb9e148c216d3f053334f90d5dd6675f1df06b0c64e6f022c5\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:19.310514 containerd[1503]: time="2025-02-13T19:12:19.310346012Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-84956f4cf-p6mr2,Uid:05e39f6f-0a83-4b62-8cff-b0d4799f8f89,Namespace:calico-apiserver,Attempt:5,} failed, error" error="failed to setup network for sandbox \"15ac125800531efb9e148c216d3f053334f90d5dd6675f1df06b0c64e6f022c5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:19.312131 kubelet[2828]: E0213 19:12:19.310827 2828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"15ac125800531efb9e148c216d3f053334f90d5dd6675f1df06b0c64e6f022c5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:12:19.312131 kubelet[2828]: E0213 19:12:19.310907 2828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"15ac125800531efb9e148c216d3f053334f90d5dd6675f1df06b0c64e6f022c5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-84956f4cf-p6mr2" Feb 13 19:12:19.312131 kubelet[2828]: E0213 19:12:19.310928 2828 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"15ac125800531efb9e148c216d3f053334f90d5dd6675f1df06b0c64e6f022c5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-84956f4cf-p6mr2" Feb 13 19:12:19.312305 kubelet[2828]: E0213 19:12:19.310973 2828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-84956f4cf-p6mr2_calico-apiserver(05e39f6f-0a83-4b62-8cff-b0d4799f8f89)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-84956f4cf-p6mr2_calico-apiserver(05e39f6f-0a83-4b62-8cff-b0d4799f8f89)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"15ac125800531efb9e148c216d3f053334f90d5dd6675f1df06b0c64e6f022c5\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-84956f4cf-p6mr2" podUID="05e39f6f-0a83-4b62-8cff-b0d4799f8f89" Feb 13 19:12:19.338358 containerd[1503]: time="2025-02-13T19:12:19.338260956Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.29.1: active requests=0, bytes read=137671762" Feb 13 19:12:19.342841 containerd[1503]: time="2025-02-13T19:12:19.342776255Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.29.1\" with image id \"sha256:680b8c280812d12c035ca9f0deedea7c761afe0f1cc65109ea2f96bf63801758\", repo tag \"ghcr.io/flatcar/calico/node:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/node@sha256:99c3917516efe1f807a0cfdf2d14b628b7c5cc6bd8a9ee5a253154f31756bea1\", size \"137671624\" in 5.70676372s" Feb 13 19:12:19.342841 containerd[1503]: time="2025-02-13T19:12:19.342820217Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.1\" returns image reference \"sha256:680b8c280812d12c035ca9f0deedea7c761afe0f1cc65109ea2f96bf63801758\"" Feb 13 19:12:19.354141 containerd[1503]: time="2025-02-13T19:12:19.354102405Z" level=info msg="CreateContainer within sandbox \"ba2393a86293acf25a04e42925632c5262b70843267e4ab409c90b5a6834ebfa\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Feb 13 19:12:19.363535 containerd[1503]: time="2025-02-13T19:12:19.363481856Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:12:19.364962 containerd[1503]: time="2025-02-13T19:12:19.364929860Z" level=info msg="ImageCreate event name:\"sha256:680b8c280812d12c035ca9f0deedea7c761afe0f1cc65109ea2f96bf63801758\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:12:19.365729 containerd[1503]: time="2025-02-13T19:12:19.365698484Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:99c3917516efe1f807a0cfdf2d14b628b7c5cc6bd8a9ee5a253154f31756bea1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:12:19.375871 containerd[1503]: time="2025-02-13T19:12:19.375800597Z" level=info msg="CreateContainer within sandbox \"ba2393a86293acf25a04e42925632c5262b70843267e4ab409c90b5a6834ebfa\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"913aa84377122e4ce9c88bd802d7c06ff5a888db011d36c546246f89f0494806\"" Feb 13 19:12:19.377533 containerd[1503]: time="2025-02-13T19:12:19.377491409Z" level=info msg="StartContainer for \"913aa84377122e4ce9c88bd802d7c06ff5a888db011d36c546246f89f0494806\"" Feb 13 19:12:19.410331 systemd[1]: Started cri-containerd-913aa84377122e4ce9c88bd802d7c06ff5a888db011d36c546246f89f0494806.scope - libcontainer container 913aa84377122e4ce9c88bd802d7c06ff5a888db011d36c546246f89f0494806. Feb 13 19:12:19.421549 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-76c8fc7995e25f780bca1521611986ac1d0cb18b34d8d5db637de52885bbe0b7-shm.mount: Deactivated successfully. Feb 13 19:12:19.421850 systemd[1]: run-netns-cni\x2d44071d6e\x2d3bf9\x2d8490\x2d7bf9\x2dad760ec8af8c.mount: Deactivated successfully. Feb 13 19:12:19.421912 systemd[1]: run-netns-cni\x2d3ef779ea\x2dcb96\x2d1507\x2d73c0\x2df5013845ba66.mount: Deactivated successfully. Feb 13 19:12:19.421967 systemd[1]: run-netns-cni\x2d71ace8fd\x2d046c\x2d56c1\x2dfc03\x2de36e5204bd6a.mount: Deactivated successfully. Feb 13 19:12:19.422021 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1071094234.mount: Deactivated successfully. Feb 13 19:12:19.459557 containerd[1503]: time="2025-02-13T19:12:19.458545836Z" level=info msg="StartContainer for \"913aa84377122e4ce9c88bd802d7c06ff5a888db011d36c546246f89f0494806\" returns successfully" Feb 13 19:12:19.576088 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Feb 13 19:12:19.576239 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Feb 13 19:12:20.013873 kubelet[2828]: I0213 19:12:20.012349 2828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d485d108247618e01bc6b75342565a89c0d97dadc8e9f0f8b8a03154fa8d72b1" Feb 13 19:12:20.014309 containerd[1503]: time="2025-02-13T19:12:20.013338638Z" level=info msg="StopPodSandbox for \"d485d108247618e01bc6b75342565a89c0d97dadc8e9f0f8b8a03154fa8d72b1\"" Feb 13 19:12:20.016155 containerd[1503]: time="2025-02-13T19:12:20.016095683Z" level=info msg="Ensure that sandbox d485d108247618e01bc6b75342565a89c0d97dadc8e9f0f8b8a03154fa8d72b1 in task-service has been cleanup successfully" Feb 13 19:12:20.019868 containerd[1503]: time="2025-02-13T19:12:20.016836826Z" level=info msg="TearDown network for sandbox \"d485d108247618e01bc6b75342565a89c0d97dadc8e9f0f8b8a03154fa8d72b1\" successfully" Feb 13 19:12:20.019868 containerd[1503]: time="2025-02-13T19:12:20.016864947Z" level=info msg="StopPodSandbox for \"d485d108247618e01bc6b75342565a89c0d97dadc8e9f0f8b8a03154fa8d72b1\" returns successfully" Feb 13 19:12:20.024535 containerd[1503]: time="2025-02-13T19:12:20.022337158Z" level=info msg="StopPodSandbox for \"931e447af991e370f310777638932ea566d0d3e47f8ec0fb6ba2b0fe9e6ff401\"" Feb 13 19:12:20.024535 containerd[1503]: time="2025-02-13T19:12:20.022474242Z" level=info msg="TearDown network for sandbox \"931e447af991e370f310777638932ea566d0d3e47f8ec0fb6ba2b0fe9e6ff401\" successfully" Feb 13 19:12:20.024535 containerd[1503]: time="2025-02-13T19:12:20.022494883Z" level=info msg="StopPodSandbox for \"931e447af991e370f310777638932ea566d0d3e47f8ec0fb6ba2b0fe9e6ff401\" returns successfully" Feb 13 19:12:20.024535 containerd[1503]: time="2025-02-13T19:12:20.023138703Z" level=info msg="StopPodSandbox for \"7242b70e1006e9ac2f114fcee5b8cb6221151d9cbbf425c7398a098f6fd891ad\"" Feb 13 19:12:20.024535 containerd[1503]: time="2025-02-13T19:12:20.023255946Z" level=info msg="TearDown network for sandbox \"7242b70e1006e9ac2f114fcee5b8cb6221151d9cbbf425c7398a098f6fd891ad\" successfully" Feb 13 19:12:20.024535 containerd[1503]: time="2025-02-13T19:12:20.023268467Z" level=info msg="StopPodSandbox for \"7242b70e1006e9ac2f114fcee5b8cb6221151d9cbbf425c7398a098f6fd891ad\" returns successfully" Feb 13 19:12:20.024535 containerd[1503]: time="2025-02-13T19:12:20.023806444Z" level=info msg="StopPodSandbox for \"d2cb6c1d9e5335f4d8ee798e79c4a58d442740cd8f9ddfe185709b1cfbd4ea81\"" Feb 13 19:12:20.024535 containerd[1503]: time="2025-02-13T19:12:20.023924447Z" level=info msg="TearDown network for sandbox \"d2cb6c1d9e5335f4d8ee798e79c4a58d442740cd8f9ddfe185709b1cfbd4ea81\" successfully" Feb 13 19:12:20.024535 containerd[1503]: time="2025-02-13T19:12:20.024458784Z" level=info msg="StopPodSandbox for \"d2cb6c1d9e5335f4d8ee798e79c4a58d442740cd8f9ddfe185709b1cfbd4ea81\" returns successfully" Feb 13 19:12:20.023743 systemd[1]: run-netns-cni\x2dc1885a6f\x2d9c8e\x2d1fb4\x2d87ae\x2d2f75b8385755.mount: Deactivated successfully. Feb 13 19:12:20.027326 containerd[1503]: time="2025-02-13T19:12:20.026315562Z" level=info msg="StopPodSandbox for \"10ae5b7da745ff3939219060da91f4661d2a1136a80736c3234db5b474e3599d\"" Feb 13 19:12:20.027326 containerd[1503]: time="2025-02-13T19:12:20.026445406Z" level=info msg="TearDown network for sandbox \"10ae5b7da745ff3939219060da91f4661d2a1136a80736c3234db5b474e3599d\" successfully" Feb 13 19:12:20.027326 containerd[1503]: time="2025-02-13T19:12:20.026456206Z" level=info msg="StopPodSandbox for \"10ae5b7da745ff3939219060da91f4661d2a1136a80736c3234db5b474e3599d\" returns successfully" Feb 13 19:12:20.028114 containerd[1503]: time="2025-02-13T19:12:20.027640163Z" level=info msg="StopPodSandbox for \"43a5af5ee3597e852f8d0ba977ffb06562d6b40688d779a95c8d495a78df4c29\"" Feb 13 19:12:20.028114 containerd[1503]: time="2025-02-13T19:12:20.027762207Z" level=info msg="TearDown network for sandbox \"43a5af5ee3597e852f8d0ba977ffb06562d6b40688d779a95c8d495a78df4c29\" successfully" Feb 13 19:12:20.028114 containerd[1503]: time="2025-02-13T19:12:20.027795448Z" level=info msg="StopPodSandbox for \"43a5af5ee3597e852f8d0ba977ffb06562d6b40688d779a95c8d495a78df4c29\" returns successfully" Feb 13 19:12:20.028555 containerd[1503]: time="2025-02-13T19:12:20.028428148Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-kkqg5,Uid:5cf24d69-eb75-4cb3-8437-024193f516f5,Namespace:kube-system,Attempt:6,}" Feb 13 19:12:20.030408 kubelet[2828]: I0213 19:12:20.029692 2828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ca053ac29abc2425e4b004e866bb9d2dbabd9431e4a54c02f8b398f38c0da08" Feb 13 19:12:20.031882 containerd[1503]: time="2025-02-13T19:12:20.031849974Z" level=info msg="StopPodSandbox for \"6ca053ac29abc2425e4b004e866bb9d2dbabd9431e4a54c02f8b398f38c0da08\"" Feb 13 19:12:20.032512 containerd[1503]: time="2025-02-13T19:12:20.032480314Z" level=info msg="Ensure that sandbox 6ca053ac29abc2425e4b004e866bb9d2dbabd9431e4a54c02f8b398f38c0da08 in task-service has been cleanup successfully" Feb 13 19:12:20.035119 containerd[1503]: time="2025-02-13T19:12:20.033193056Z" level=info msg="TearDown network for sandbox \"6ca053ac29abc2425e4b004e866bb9d2dbabd9431e4a54c02f8b398f38c0da08\" successfully" Feb 13 19:12:20.035119 containerd[1503]: time="2025-02-13T19:12:20.033218097Z" level=info msg="StopPodSandbox for \"6ca053ac29abc2425e4b004e866bb9d2dbabd9431e4a54c02f8b398f38c0da08\" returns successfully" Feb 13 19:12:20.037805 systemd[1]: run-netns-cni\x2d7194d3b8\x2d1442\x2dd71f\x2d216f\x2d7789c378a0ec.mount: Deactivated successfully. Feb 13 19:12:20.040250 containerd[1503]: time="2025-02-13T19:12:20.039567694Z" level=info msg="StopPodSandbox for \"f101bce244f680aa9fab77bf6802cf73b54ffba4caa118c44f2e7b203fb0dc84\"" Feb 13 19:12:20.040250 containerd[1503]: time="2025-02-13T19:12:20.039700539Z" level=info msg="TearDown network for sandbox \"f101bce244f680aa9fab77bf6802cf73b54ffba4caa118c44f2e7b203fb0dc84\" successfully" Feb 13 19:12:20.040250 containerd[1503]: time="2025-02-13T19:12:20.039710899Z" level=info msg="StopPodSandbox for \"f101bce244f680aa9fab77bf6802cf73b54ffba4caa118c44f2e7b203fb0dc84\" returns successfully" Feb 13 19:12:20.041793 containerd[1503]: time="2025-02-13T19:12:20.041433673Z" level=info msg="StopPodSandbox for \"f1ae97b8cb19d334ba8268033afc16640078623540002cf3f5996a895113dd8b\"" Feb 13 19:12:20.041793 containerd[1503]: time="2025-02-13T19:12:20.041560077Z" level=info msg="TearDown network for sandbox \"f1ae97b8cb19d334ba8268033afc16640078623540002cf3f5996a895113dd8b\" successfully" Feb 13 19:12:20.041793 containerd[1503]: time="2025-02-13T19:12:20.041572837Z" level=info msg="StopPodSandbox for \"f1ae97b8cb19d334ba8268033afc16640078623540002cf3f5996a895113dd8b\" returns successfully" Feb 13 19:12:20.043863 containerd[1503]: time="2025-02-13T19:12:20.043463656Z" level=info msg="StopPodSandbox for \"4ffb63334faa7203eea7b2071116973af83410f7340f4e8946194bccb25fa157\"" Feb 13 19:12:20.043863 containerd[1503]: time="2025-02-13T19:12:20.043568299Z" level=info msg="TearDown network for sandbox \"4ffb63334faa7203eea7b2071116973af83410f7340f4e8946194bccb25fa157\" successfully" Feb 13 19:12:20.043863 containerd[1503]: time="2025-02-13T19:12:20.043577899Z" level=info msg="StopPodSandbox for \"4ffb63334faa7203eea7b2071116973af83410f7340f4e8946194bccb25fa157\" returns successfully" Feb 13 19:12:20.044468 containerd[1503]: time="2025-02-13T19:12:20.044445646Z" level=info msg="StopPodSandbox for \"0be20e2c65924f7b5393bc8b6a79f35d5bf8cca7f34009d6201de4852924939b\"" Feb 13 19:12:20.044559 containerd[1503]: time="2025-02-13T19:12:20.044542689Z" level=info msg="TearDown network for sandbox \"0be20e2c65924f7b5393bc8b6a79f35d5bf8cca7f34009d6201de4852924939b\" successfully" Feb 13 19:12:20.044559 containerd[1503]: time="2025-02-13T19:12:20.044557130Z" level=info msg="StopPodSandbox for \"0be20e2c65924f7b5393bc8b6a79f35d5bf8cca7f34009d6201de4852924939b\" returns successfully" Feb 13 19:12:20.046036 containerd[1503]: time="2025-02-13T19:12:20.045880331Z" level=info msg="StopPodSandbox for \"fa6e4b9a2b53f8e5e8150bdfa0fe88ef1d51405b9b5e7e6d9683c6adb0b078ab\"" Feb 13 19:12:20.046036 containerd[1503]: time="2025-02-13T19:12:20.045973574Z" level=info msg="TearDown network for sandbox \"fa6e4b9a2b53f8e5e8150bdfa0fe88ef1d51405b9b5e7e6d9683c6adb0b078ab\" successfully" Feb 13 19:12:20.046036 containerd[1503]: time="2025-02-13T19:12:20.045983134Z" level=info msg="StopPodSandbox for \"fa6e4b9a2b53f8e5e8150bdfa0fe88ef1d51405b9b5e7e6d9683c6adb0b078ab\" returns successfully" Feb 13 19:12:20.047439 kubelet[2828]: I0213 19:12:20.045980 2828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b86651f4cc54f520cab1a18f98b7a70986a4aa77bdc28bce93539ac71deee2a2" Feb 13 19:12:20.062638 containerd[1503]: time="2025-02-13T19:12:20.062597492Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-78bd6d67c9-b259q,Uid:436348d7-7832-43e1-8780-623c434b2666,Namespace:calico-system,Attempt:6,}" Feb 13 19:12:20.063993 containerd[1503]: time="2025-02-13T19:12:20.063869291Z" level=info msg="StopPodSandbox for \"b86651f4cc54f520cab1a18f98b7a70986a4aa77bdc28bce93539ac71deee2a2\"" Feb 13 19:12:20.067975 containerd[1503]: time="2025-02-13T19:12:20.067236676Z" level=info msg="Ensure that sandbox b86651f4cc54f520cab1a18f98b7a70986a4aa77bdc28bce93539ac71deee2a2 in task-service has been cleanup successfully" Feb 13 19:12:20.067975 containerd[1503]: time="2025-02-13T19:12:20.067966339Z" level=info msg="TearDown network for sandbox \"b86651f4cc54f520cab1a18f98b7a70986a4aa77bdc28bce93539ac71deee2a2\" successfully" Feb 13 19:12:20.068247 containerd[1503]: time="2025-02-13T19:12:20.067988420Z" level=info msg="StopPodSandbox for \"b86651f4cc54f520cab1a18f98b7a70986a4aa77bdc28bce93539ac71deee2a2\" returns successfully" Feb 13 19:12:20.069296 containerd[1503]: time="2025-02-13T19:12:20.069267900Z" level=info msg="StopPodSandbox for \"4d5614a8446f6178d266e8f2147ca224323d91c476899b1e9e66b08c962ad2a3\"" Feb 13 19:12:20.069477 containerd[1503]: time="2025-02-13T19:12:20.069461386Z" level=info msg="TearDown network for sandbox \"4d5614a8446f6178d266e8f2147ca224323d91c476899b1e9e66b08c962ad2a3\" successfully" Feb 13 19:12:20.069545 containerd[1503]: time="2025-02-13T19:12:20.069531628Z" level=info msg="StopPodSandbox for \"4d5614a8446f6178d266e8f2147ca224323d91c476899b1e9e66b08c962ad2a3\" returns successfully" Feb 13 19:12:20.069897 containerd[1503]: time="2025-02-13T19:12:20.069873838Z" level=info msg="StopPodSandbox for \"cf3ac12d3142a48e0052c1dead1d74d6becc46ab9352d1240e51b560f73cb623\"" Feb 13 19:12:20.070135 containerd[1503]: time="2025-02-13T19:12:20.070116846Z" level=info msg="TearDown network for sandbox \"cf3ac12d3142a48e0052c1dead1d74d6becc46ab9352d1240e51b560f73cb623\" successfully" Feb 13 19:12:20.070238 containerd[1503]: time="2025-02-13T19:12:20.070223929Z" level=info msg="StopPodSandbox for \"cf3ac12d3142a48e0052c1dead1d74d6becc46ab9352d1240e51b560f73cb623\" returns successfully" Feb 13 19:12:20.071046 containerd[1503]: time="2025-02-13T19:12:20.071015994Z" level=info msg="StopPodSandbox for \"af85f8d50ee74cce2c7e2cee2af5922ba7bbd4077fc92d81b183d5575fad1329\"" Feb 13 19:12:20.071339 containerd[1503]: time="2025-02-13T19:12:20.071320723Z" level=info msg="TearDown network for sandbox \"af85f8d50ee74cce2c7e2cee2af5922ba7bbd4077fc92d81b183d5575fad1329\" successfully" Feb 13 19:12:20.071427 containerd[1503]: time="2025-02-13T19:12:20.071411086Z" level=info msg="StopPodSandbox for \"af85f8d50ee74cce2c7e2cee2af5922ba7bbd4077fc92d81b183d5575fad1329\" returns successfully" Feb 13 19:12:20.071971 systemd[1]: run-netns-cni\x2d4edfd26b\x2da194\x2dc1f8\x2d9fec\x2d5dc195d5b8be.mount: Deactivated successfully. Feb 13 19:12:20.073340 containerd[1503]: time="2025-02-13T19:12:20.073197742Z" level=info msg="StopPodSandbox for \"ae9615caa56ac5be11c6b7cc88aac5196ade400655d7a9313d485332a9d3820e\"" Feb 13 19:12:20.074532 containerd[1503]: time="2025-02-13T19:12:20.074114810Z" level=info msg="TearDown network for sandbox \"ae9615caa56ac5be11c6b7cc88aac5196ade400655d7a9313d485332a9d3820e\" successfully" Feb 13 19:12:20.074532 containerd[1503]: time="2025-02-13T19:12:20.074323257Z" level=info msg="StopPodSandbox for \"ae9615caa56ac5be11c6b7cc88aac5196ade400655d7a9313d485332a9d3820e\" returns successfully" Feb 13 19:12:20.075335 containerd[1503]: time="2025-02-13T19:12:20.075312968Z" level=info msg="StopPodSandbox for \"930563c746d8b91d524a61a614694c47905765a47fb87cb32a681d5b35060a78\"" Feb 13 19:12:20.075427 kubelet[2828]: I0213 19:12:20.075409 2828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15ac125800531efb9e148c216d3f053334f90d5dd6675f1df06b0c64e6f022c5" Feb 13 19:12:20.075911 containerd[1503]: time="2025-02-13T19:12:20.075777302Z" level=info msg="TearDown network for sandbox \"930563c746d8b91d524a61a614694c47905765a47fb87cb32a681d5b35060a78\" successfully" Feb 13 19:12:20.076129 containerd[1503]: time="2025-02-13T19:12:20.075797383Z" level=info msg="StopPodSandbox for \"930563c746d8b91d524a61a614694c47905765a47fb87cb32a681d5b35060a78\" returns successfully" Feb 13 19:12:20.077078 containerd[1503]: time="2025-02-13T19:12:20.076521605Z" level=info msg="StopPodSandbox for \"15ac125800531efb9e148c216d3f053334f90d5dd6675f1df06b0c64e6f022c5\"" Feb 13 19:12:20.078301 containerd[1503]: time="2025-02-13T19:12:20.078127975Z" level=info msg="Ensure that sandbox 15ac125800531efb9e148c216d3f053334f90d5dd6675f1df06b0c64e6f022c5 in task-service has been cleanup successfully" Feb 13 19:12:20.079545 containerd[1503]: time="2025-02-13T19:12:20.079326213Z" level=info msg="TearDown network for sandbox \"15ac125800531efb9e148c216d3f053334f90d5dd6675f1df06b0c64e6f022c5\" successfully" Feb 13 19:12:20.079698 containerd[1503]: time="2025-02-13T19:12:20.079666063Z" level=info msg="StopPodSandbox for \"15ac125800531efb9e148c216d3f053334f90d5dd6675f1df06b0c64e6f022c5\" returns successfully" Feb 13 19:12:20.081533 containerd[1503]: time="2025-02-13T19:12:20.081146749Z" level=info msg="StopPodSandbox for \"b3134abe6d3978218dbaec771f16517d173324bf774a94fc2b6f62660051520d\"" Feb 13 19:12:20.081533 containerd[1503]: time="2025-02-13T19:12:20.081459479Z" level=info msg="TearDown network for sandbox \"b3134abe6d3978218dbaec771f16517d173324bf774a94fc2b6f62660051520d\" successfully" Feb 13 19:12:20.081533 containerd[1503]: time="2025-02-13T19:12:20.081474200Z" level=info msg="StopPodSandbox for \"b3134abe6d3978218dbaec771f16517d173324bf774a94fc2b6f62660051520d\" returns successfully" Feb 13 19:12:20.082075 containerd[1503]: time="2025-02-13T19:12:20.081976415Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-84956f4cf-72cgp,Uid:96785f8f-b4e3-456b-9a0d-bcc142e6beaf,Namespace:calico-apiserver,Attempt:6,}" Feb 13 19:12:20.084493 containerd[1503]: time="2025-02-13T19:12:20.084469173Z" level=info msg="StopPodSandbox for \"96f8dbfc99d905b595d74b0f3390d27d4b0e99b3143e0cc72edd14238522b44f\"" Feb 13 19:12:20.084850 containerd[1503]: time="2025-02-13T19:12:20.084764542Z" level=info msg="TearDown network for sandbox \"96f8dbfc99d905b595d74b0f3390d27d4b0e99b3143e0cc72edd14238522b44f\" successfully" Feb 13 19:12:20.084850 containerd[1503]: time="2025-02-13T19:12:20.084792903Z" level=info msg="StopPodSandbox for \"96f8dbfc99d905b595d74b0f3390d27d4b0e99b3143e0cc72edd14238522b44f\" returns successfully" Feb 13 19:12:20.085404 containerd[1503]: time="2025-02-13T19:12:20.085361721Z" level=info msg="StopPodSandbox for \"ab072ffc918efc3594006a69ddbbd17de37178f95f1103e4fa6b31bb95fd9306\"" Feb 13 19:12:20.085730 containerd[1503]: time="2025-02-13T19:12:20.085659250Z" level=info msg="TearDown network for sandbox \"ab072ffc918efc3594006a69ddbbd17de37178f95f1103e4fa6b31bb95fd9306\" successfully" Feb 13 19:12:20.085730 containerd[1503]: time="2025-02-13T19:12:20.085679851Z" level=info msg="StopPodSandbox for \"ab072ffc918efc3594006a69ddbbd17de37178f95f1103e4fa6b31bb95fd9306\" returns successfully" Feb 13 19:12:20.086459 containerd[1503]: time="2025-02-13T19:12:20.086399313Z" level=info msg="StopPodSandbox for \"9581f7ff6fcd5f0bba90e1d0d2984cd0efa6fe4d50a06a4c1b20f3939cb09713\"" Feb 13 19:12:20.086990 containerd[1503]: time="2025-02-13T19:12:20.086888008Z" level=info msg="TearDown network for sandbox \"9581f7ff6fcd5f0bba90e1d0d2984cd0efa6fe4d50a06a4c1b20f3939cb09713\" successfully" Feb 13 19:12:20.086990 containerd[1503]: time="2025-02-13T19:12:20.086937210Z" level=info msg="StopPodSandbox for \"9581f7ff6fcd5f0bba90e1d0d2984cd0efa6fe4d50a06a4c1b20f3939cb09713\" returns successfully" Feb 13 19:12:20.094459 containerd[1503]: time="2025-02-13T19:12:20.092180973Z" level=info msg="StopPodSandbox for \"0c13c2a989169406cb4cea286d09a0b055b98c8bf7adaf3c864364728278a3a5\"" Feb 13 19:12:20.094459 containerd[1503]: time="2025-02-13T19:12:20.094341840Z" level=info msg="TearDown network for sandbox \"0c13c2a989169406cb4cea286d09a0b055b98c8bf7adaf3c864364728278a3a5\" successfully" Feb 13 19:12:20.094459 containerd[1503]: time="2025-02-13T19:12:20.094369401Z" level=info msg="StopPodSandbox for \"0c13c2a989169406cb4cea286d09a0b055b98c8bf7adaf3c864364728278a3a5\" returns successfully" Feb 13 19:12:20.101339 containerd[1503]: time="2025-02-13T19:12:20.100235344Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-84956f4cf-p6mr2,Uid:05e39f6f-0a83-4b62-8cff-b0d4799f8f89,Namespace:calico-apiserver,Attempt:6,}" Feb 13 19:12:20.117782 kubelet[2828]: I0213 19:12:20.115396 2828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76c8fc7995e25f780bca1521611986ac1d0cb18b34d8d5db637de52885bbe0b7" Feb 13 19:12:20.123395 containerd[1503]: time="2025-02-13T19:12:20.122952652Z" level=info msg="StopPodSandbox for \"76c8fc7995e25f780bca1521611986ac1d0cb18b34d8d5db637de52885bbe0b7\"" Feb 13 19:12:20.126507 containerd[1503]: time="2025-02-13T19:12:20.126435000Z" level=info msg="Ensure that sandbox 76c8fc7995e25f780bca1521611986ac1d0cb18b34d8d5db637de52885bbe0b7 in task-service has been cleanup successfully" Feb 13 19:12:20.129723 containerd[1503]: time="2025-02-13T19:12:20.129508816Z" level=info msg="TearDown network for sandbox \"76c8fc7995e25f780bca1521611986ac1d0cb18b34d8d5db637de52885bbe0b7\" successfully" Feb 13 19:12:20.129723 containerd[1503]: time="2025-02-13T19:12:20.129549457Z" level=info msg="StopPodSandbox for \"76c8fc7995e25f780bca1521611986ac1d0cb18b34d8d5db637de52885bbe0b7\" returns successfully" Feb 13 19:12:20.133104 kubelet[2828]: I0213 19:12:20.132104 2828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-8zbsj" podStartSLOduration=1.526173967 podStartE2EDuration="15.132082336s" podCreationTimestamp="2025-02-13 19:12:05 +0000 UTC" firstStartedPulling="2025-02-13 19:12:05.737699032 +0000 UTC m=+31.436150710" lastFinishedPulling="2025-02-13 19:12:19.343607441 +0000 UTC m=+45.042059079" observedRunningTime="2025-02-13 19:12:20.125688377 +0000 UTC m=+45.824140055" watchObservedRunningTime="2025-02-13 19:12:20.132082336 +0000 UTC m=+45.830534014" Feb 13 19:12:20.134028 containerd[1503]: time="2025-02-13T19:12:20.133973555Z" level=info msg="StopPodSandbox for \"602a514f998cb50210ddc95537cfa35ed74b37741d39df2aa29785c3996317a9\"" Feb 13 19:12:20.135025 containerd[1503]: time="2025-02-13T19:12:20.134980826Z" level=info msg="TearDown network for sandbox \"602a514f998cb50210ddc95537cfa35ed74b37741d39df2aa29785c3996317a9\" successfully" Feb 13 19:12:20.139141 containerd[1503]: time="2025-02-13T19:12:20.138992311Z" level=info msg="StopPodSandbox for \"602a514f998cb50210ddc95537cfa35ed74b37741d39df2aa29785c3996317a9\" returns successfully" Feb 13 19:12:20.148755 containerd[1503]: time="2025-02-13T19:12:20.148695293Z" level=info msg="StopPodSandbox for \"68ee287533ff713dee0f8970cee78f6429754d9b99d2574db1ab310eb93f45f4\"" Feb 13 19:12:20.149811 containerd[1503]: time="2025-02-13T19:12:20.149781407Z" level=info msg="TearDown network for sandbox \"68ee287533ff713dee0f8970cee78f6429754d9b99d2574db1ab310eb93f45f4\" successfully" Feb 13 19:12:20.149936 containerd[1503]: time="2025-02-13T19:12:20.149921772Z" level=info msg="StopPodSandbox for \"68ee287533ff713dee0f8970cee78f6429754d9b99d2574db1ab310eb93f45f4\" returns successfully" Feb 13 19:12:20.152385 containerd[1503]: time="2025-02-13T19:12:20.152326966Z" level=info msg="StopPodSandbox for \"a657753a39333ef1c8ee279e3321a3d04ca609cd0499c88ae5589336776d3f50\"" Feb 13 19:12:20.153282 containerd[1503]: time="2025-02-13T19:12:20.153256515Z" level=info msg="TearDown network for sandbox \"a657753a39333ef1c8ee279e3321a3d04ca609cd0499c88ae5589336776d3f50\" successfully" Feb 13 19:12:20.153442 containerd[1503]: time="2025-02-13T19:12:20.153425521Z" level=info msg="StopPodSandbox for \"a657753a39333ef1c8ee279e3321a3d04ca609cd0499c88ae5589336776d3f50\" returns successfully" Feb 13 19:12:20.154075 containerd[1503]: time="2025-02-13T19:12:20.153981018Z" level=info msg="StopPodSandbox for \"07aed5c023700866ebeb2a6224ac11566918d60f5a140685f482d3a85f4cd555\"" Feb 13 19:12:20.154329 containerd[1503]: time="2025-02-13T19:12:20.154252706Z" level=info msg="TearDown network for sandbox \"07aed5c023700866ebeb2a6224ac11566918d60f5a140685f482d3a85f4cd555\" successfully" Feb 13 19:12:20.154329 containerd[1503]: time="2025-02-13T19:12:20.154269307Z" level=info msg="StopPodSandbox for \"07aed5c023700866ebeb2a6224ac11566918d60f5a140685f482d3a85f4cd555\" returns successfully" Feb 13 19:12:20.154882 containerd[1503]: time="2025-02-13T19:12:20.154747242Z" level=info msg="StopPodSandbox for \"0c31ed484cb1e7593659d8fd1f850350d6b3b9e01ed8530e08cf66bc47a72b73\"" Feb 13 19:12:20.154882 containerd[1503]: time="2025-02-13T19:12:20.154823844Z" level=info msg="TearDown network for sandbox \"0c31ed484cb1e7593659d8fd1f850350d6b3b9e01ed8530e08cf66bc47a72b73\" successfully" Feb 13 19:12:20.154882 containerd[1503]: time="2025-02-13T19:12:20.154833205Z" level=info msg="StopPodSandbox for \"0c31ed484cb1e7593659d8fd1f850350d6b3b9e01ed8530e08cf66bc47a72b73\" returns successfully" Feb 13 19:12:20.156641 containerd[1503]: time="2025-02-13T19:12:20.156474416Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-vx9cw,Uid:eb0c3c87-077a-4b2b-a168-5fd9131bccd3,Namespace:kube-system,Attempt:6,}" Feb 13 19:12:20.157997 kubelet[2828]: I0213 19:12:20.157961 2828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77460032187adb446e84e042dcc0956c8b578caedef00c1e669bc2016d01bf85" Feb 13 19:12:20.159953 containerd[1503]: time="2025-02-13T19:12:20.159313224Z" level=info msg="StopPodSandbox for \"77460032187adb446e84e042dcc0956c8b578caedef00c1e669bc2016d01bf85\"" Feb 13 19:12:20.161307 containerd[1503]: time="2025-02-13T19:12:20.161268165Z" level=info msg="Ensure that sandbox 77460032187adb446e84e042dcc0956c8b578caedef00c1e669bc2016d01bf85 in task-service has been cleanup successfully" Feb 13 19:12:20.163400 containerd[1503]: time="2025-02-13T19:12:20.163320589Z" level=info msg="TearDown network for sandbox \"77460032187adb446e84e042dcc0956c8b578caedef00c1e669bc2016d01bf85\" successfully" Feb 13 19:12:20.165786 containerd[1503]: time="2025-02-13T19:12:20.165462696Z" level=info msg="StopPodSandbox for \"77460032187adb446e84e042dcc0956c8b578caedef00c1e669bc2016d01bf85\" returns successfully" Feb 13 19:12:20.168960 containerd[1503]: time="2025-02-13T19:12:20.168377506Z" level=info msg="StopPodSandbox for \"728705ac6957c83c18f350153e00e11e14122291646534ceb675f3fda8f70cdb\"" Feb 13 19:12:20.170367 containerd[1503]: time="2025-02-13T19:12:20.170334607Z" level=info msg="TearDown network for sandbox \"728705ac6957c83c18f350153e00e11e14122291646534ceb675f3fda8f70cdb\" successfully" Feb 13 19:12:20.170673 containerd[1503]: time="2025-02-13T19:12:20.170649177Z" level=info msg="StopPodSandbox for \"728705ac6957c83c18f350153e00e11e14122291646534ceb675f3fda8f70cdb\" returns successfully" Feb 13 19:12:20.177340 containerd[1503]: time="2025-02-13T19:12:20.177287584Z" level=info msg="StopPodSandbox for \"c6b1e16315492197341e134074aec5720ae8f11ba5d44a3e6e3a0312f21d145d\"" Feb 13 19:12:20.177666 containerd[1503]: time="2025-02-13T19:12:20.177642435Z" level=info msg="TearDown network for sandbox \"c6b1e16315492197341e134074aec5720ae8f11ba5d44a3e6e3a0312f21d145d\" successfully" Feb 13 19:12:20.177761 containerd[1503]: time="2025-02-13T19:12:20.177744038Z" level=info msg="StopPodSandbox for \"c6b1e16315492197341e134074aec5720ae8f11ba5d44a3e6e3a0312f21d145d\" returns successfully" Feb 13 19:12:20.179352 containerd[1503]: time="2025-02-13T19:12:20.179318327Z" level=info msg="StopPodSandbox for \"4642b0f0428148a51fe1afb0c6a261eff1112eb99bbdf06b50e264a81fa407b3\"" Feb 13 19:12:20.180072 containerd[1503]: time="2025-02-13T19:12:20.180030149Z" level=info msg="TearDown network for sandbox \"4642b0f0428148a51fe1afb0c6a261eff1112eb99bbdf06b50e264a81fa407b3\" successfully" Feb 13 19:12:20.180307 containerd[1503]: time="2025-02-13T19:12:20.180154713Z" level=info msg="StopPodSandbox for \"4642b0f0428148a51fe1afb0c6a261eff1112eb99bbdf06b50e264a81fa407b3\" returns successfully" Feb 13 19:12:20.185971 containerd[1503]: time="2025-02-13T19:12:20.185920413Z" level=info msg="StopPodSandbox for \"4f219600df3d01ca237f481814d18c92c5997c6c501af4d70cb45cd41d1a6efc\"" Feb 13 19:12:20.187503 containerd[1503]: time="2025-02-13T19:12:20.187455981Z" level=info msg="TearDown network for sandbox \"4f219600df3d01ca237f481814d18c92c5997c6c501af4d70cb45cd41d1a6efc\" successfully" Feb 13 19:12:20.187978 containerd[1503]: time="2025-02-13T19:12:20.187886514Z" level=info msg="StopPodSandbox for \"4f219600df3d01ca237f481814d18c92c5997c6c501af4d70cb45cd41d1a6efc\" returns successfully" Feb 13 19:12:20.190300 containerd[1503]: time="2025-02-13T19:12:20.189866016Z" level=info msg="StopPodSandbox for \"1f6b42d2735cedb916bc24bf2bb557d2ff72928d0c487a93a3dd46e88aa55abe\"" Feb 13 19:12:20.190920 containerd[1503]: time="2025-02-13T19:12:20.190783564Z" level=info msg="TearDown network for sandbox \"1f6b42d2735cedb916bc24bf2bb557d2ff72928d0c487a93a3dd46e88aa55abe\" successfully" Feb 13 19:12:20.191111 containerd[1503]: time="2025-02-13T19:12:20.191089934Z" level=info msg="StopPodSandbox for \"1f6b42d2735cedb916bc24bf2bb557d2ff72928d0c487a93a3dd46e88aa55abe\" returns successfully" Feb 13 19:12:20.192445 containerd[1503]: time="2025-02-13T19:12:20.192420055Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-6v9pr,Uid:b9c8452b-31d9-47bc-9697-dd4f293d71b9,Namespace:calico-system,Attempt:6,}" Feb 13 19:12:20.429177 systemd[1]: run-netns-cni\x2d9fc9b6e4\x2d5a91\x2d2ed0\x2d8dfe\x2d360626d3e939.mount: Deactivated successfully. Feb 13 19:12:20.429263 systemd[1]: run-netns-cni\x2d4ca4f5fd\x2d88be\x2d61cd\x2db8a4\x2db2d012f44e04.mount: Deactivated successfully. Feb 13 19:12:20.429309 systemd[1]: run-netns-cni\x2d5f6c9487\x2daecb\x2d05b7\x2dda45\x2dd7120b7c0e9e.mount: Deactivated successfully. Feb 13 19:12:20.564193 systemd-networkd[1399]: calicc1bce0eade: Link UP Feb 13 19:12:20.565737 systemd-networkd[1399]: calicc1bce0eade: Gained carrier Feb 13 19:12:20.612098 containerd[1503]: 2025-02-13 19:12:20.122 [INFO][4774] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Feb 13 19:12:20.612098 containerd[1503]: 2025-02-13 19:12:20.200 [INFO][4774] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4230--0--1--a--2d796e8b97-k8s-coredns--668d6bf9bc--kkqg5-eth0 coredns-668d6bf9bc- kube-system 5cf24d69-eb75-4cb3-8437-024193f516f5 716 0 2025-02-13 19:11:41 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4230-0-1-a-2d796e8b97 coredns-668d6bf9bc-kkqg5 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] calicc1bce0eade [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="7c498096de4d518aa8367480279c86a097a53744f8a34963cdf1593cc7f82c21" Namespace="kube-system" Pod="coredns-668d6bf9bc-kkqg5" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-coredns--668d6bf9bc--kkqg5-" Feb 13 19:12:20.612098 containerd[1503]: 2025-02-13 19:12:20.201 [INFO][4774] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="7c498096de4d518aa8367480279c86a097a53744f8a34963cdf1593cc7f82c21" Namespace="kube-system" Pod="coredns-668d6bf9bc-kkqg5" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-coredns--668d6bf9bc--kkqg5-eth0" Feb 13 19:12:20.612098 containerd[1503]: 2025-02-13 19:12:20.395 [INFO][4843] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="7c498096de4d518aa8367480279c86a097a53744f8a34963cdf1593cc7f82c21" HandleID="k8s-pod-network.7c498096de4d518aa8367480279c86a097a53744f8a34963cdf1593cc7f82c21" Workload="ci--4230--0--1--a--2d796e8b97-k8s-coredns--668d6bf9bc--kkqg5-eth0" Feb 13 19:12:20.612098 containerd[1503]: 2025-02-13 19:12:20.433 [INFO][4843] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="7c498096de4d518aa8367480279c86a097a53744f8a34963cdf1593cc7f82c21" HandleID="k8s-pod-network.7c498096de4d518aa8367480279c86a097a53744f8a34963cdf1593cc7f82c21" Workload="ci--4230--0--1--a--2d796e8b97-k8s-coredns--668d6bf9bc--kkqg5-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40003f2a00), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4230-0-1-a-2d796e8b97", "pod":"coredns-668d6bf9bc-kkqg5", "timestamp":"2025-02-13 19:12:20.393710925 +0000 UTC"}, Hostname:"ci-4230-0-1-a-2d796e8b97", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Feb 13 19:12:20.612098 containerd[1503]: 2025-02-13 19:12:20.433 [INFO][4843] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Feb 13 19:12:20.612098 containerd[1503]: 2025-02-13 19:12:20.433 [INFO][4843] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Feb 13 19:12:20.612098 containerd[1503]: 2025-02-13 19:12:20.433 [INFO][4843] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4230-0-1-a-2d796e8b97' Feb 13 19:12:20.612098 containerd[1503]: 2025-02-13 19:12:20.440 [INFO][4843] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.7c498096de4d518aa8367480279c86a097a53744f8a34963cdf1593cc7f82c21" host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:20.612098 containerd[1503]: 2025-02-13 19:12:20.460 [INFO][4843] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:20.612098 containerd[1503]: 2025-02-13 19:12:20.476 [INFO][4843] ipam/ipam.go 489: Trying affinity for 192.168.40.64/26 host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:20.612098 containerd[1503]: 2025-02-13 19:12:20.479 [INFO][4843] ipam/ipam.go 155: Attempting to load block cidr=192.168.40.64/26 host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:20.612098 containerd[1503]: 2025-02-13 19:12:20.484 [INFO][4843] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.40.64/26 host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:20.612098 containerd[1503]: 2025-02-13 19:12:20.484 [INFO][4843] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.40.64/26 handle="k8s-pod-network.7c498096de4d518aa8367480279c86a097a53744f8a34963cdf1593cc7f82c21" host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:20.612098 containerd[1503]: 2025-02-13 19:12:20.488 [INFO][4843] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.7c498096de4d518aa8367480279c86a097a53744f8a34963cdf1593cc7f82c21 Feb 13 19:12:20.612098 containerd[1503]: 2025-02-13 19:12:20.502 [INFO][4843] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.40.64/26 handle="k8s-pod-network.7c498096de4d518aa8367480279c86a097a53744f8a34963cdf1593cc7f82c21" host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:20.612098 containerd[1503]: 2025-02-13 19:12:20.515 [INFO][4843] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.40.65/26] block=192.168.40.64/26 handle="k8s-pod-network.7c498096de4d518aa8367480279c86a097a53744f8a34963cdf1593cc7f82c21" host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:20.612098 containerd[1503]: 2025-02-13 19:12:20.515 [INFO][4843] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.40.65/26] handle="k8s-pod-network.7c498096de4d518aa8367480279c86a097a53744f8a34963cdf1593cc7f82c21" host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:20.612098 containerd[1503]: 2025-02-13 19:12:20.515 [INFO][4843] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Feb 13 19:12:20.612098 containerd[1503]: 2025-02-13 19:12:20.515 [INFO][4843] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.40.65/26] IPv6=[] ContainerID="7c498096de4d518aa8367480279c86a097a53744f8a34963cdf1593cc7f82c21" HandleID="k8s-pod-network.7c498096de4d518aa8367480279c86a097a53744f8a34963cdf1593cc7f82c21" Workload="ci--4230--0--1--a--2d796e8b97-k8s-coredns--668d6bf9bc--kkqg5-eth0" Feb 13 19:12:20.614159 containerd[1503]: 2025-02-13 19:12:20.524 [INFO][4774] cni-plugin/k8s.go 386: Populated endpoint ContainerID="7c498096de4d518aa8367480279c86a097a53744f8a34963cdf1593cc7f82c21" Namespace="kube-system" Pod="coredns-668d6bf9bc-kkqg5" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-coredns--668d6bf9bc--kkqg5-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4230--0--1--a--2d796e8b97-k8s-coredns--668d6bf9bc--kkqg5-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"5cf24d69-eb75-4cb3-8437-024193f516f5", ResourceVersion:"716", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 19, 11, 41, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4230-0-1-a-2d796e8b97", ContainerID:"", Pod:"coredns-668d6bf9bc-kkqg5", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.40.65/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calicc1bce0eade", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 19:12:20.614159 containerd[1503]: 2025-02-13 19:12:20.528 [INFO][4774] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.40.65/32] ContainerID="7c498096de4d518aa8367480279c86a097a53744f8a34963cdf1593cc7f82c21" Namespace="kube-system" Pod="coredns-668d6bf9bc-kkqg5" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-coredns--668d6bf9bc--kkqg5-eth0" Feb 13 19:12:20.614159 containerd[1503]: 2025-02-13 19:12:20.528 [INFO][4774] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calicc1bce0eade ContainerID="7c498096de4d518aa8367480279c86a097a53744f8a34963cdf1593cc7f82c21" Namespace="kube-system" Pod="coredns-668d6bf9bc-kkqg5" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-coredns--668d6bf9bc--kkqg5-eth0" Feb 13 19:12:20.614159 containerd[1503]: 2025-02-13 19:12:20.569 [INFO][4774] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="7c498096de4d518aa8367480279c86a097a53744f8a34963cdf1593cc7f82c21" Namespace="kube-system" Pod="coredns-668d6bf9bc-kkqg5" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-coredns--668d6bf9bc--kkqg5-eth0" Feb 13 19:12:20.614159 containerd[1503]: 2025-02-13 19:12:20.569 [INFO][4774] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="7c498096de4d518aa8367480279c86a097a53744f8a34963cdf1593cc7f82c21" Namespace="kube-system" Pod="coredns-668d6bf9bc-kkqg5" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-coredns--668d6bf9bc--kkqg5-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4230--0--1--a--2d796e8b97-k8s-coredns--668d6bf9bc--kkqg5-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"5cf24d69-eb75-4cb3-8437-024193f516f5", ResourceVersion:"716", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 19, 11, 41, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4230-0-1-a-2d796e8b97", ContainerID:"7c498096de4d518aa8367480279c86a097a53744f8a34963cdf1593cc7f82c21", Pod:"coredns-668d6bf9bc-kkqg5", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.40.65/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calicc1bce0eade", MAC:"be:72:ae:84:ac:2d", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 19:12:20.614159 containerd[1503]: 2025-02-13 19:12:20.607 [INFO][4774] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="7c498096de4d518aa8367480279c86a097a53744f8a34963cdf1593cc7f82c21" Namespace="kube-system" Pod="coredns-668d6bf9bc-kkqg5" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-coredns--668d6bf9bc--kkqg5-eth0" Feb 13 19:12:20.660889 containerd[1503]: time="2025-02-13T19:12:20.660707960Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:12:20.660889 containerd[1503]: time="2025-02-13T19:12:20.660778003Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:12:20.660889 containerd[1503]: time="2025-02-13T19:12:20.660793523Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:12:20.661805 containerd[1503]: time="2025-02-13T19:12:20.661704631Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:12:20.684941 systemd-networkd[1399]: caliac205cc4341: Link UP Feb 13 19:12:20.686587 systemd-networkd[1399]: caliac205cc4341: Gained carrier Feb 13 19:12:20.719411 containerd[1503]: 2025-02-13 19:12:20.281 [INFO][4815] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Feb 13 19:12:20.719411 containerd[1503]: 2025-02-13 19:12:20.327 [INFO][4815] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4230--0--1--a--2d796e8b97-k8s-calico--apiserver--84956f4cf--72cgp-eth0 calico-apiserver-84956f4cf- calico-apiserver 96785f8f-b4e3-456b-9a0d-bcc142e6beaf 712 0 2025-02-13 19:12:05 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:84956f4cf projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4230-0-1-a-2d796e8b97 calico-apiserver-84956f4cf-72cgp eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] caliac205cc4341 [] []}} ContainerID="d366cd2cc4c7c16e754d70bbe9025807e19a94be019d47fbac45a766f5cecf79" Namespace="calico-apiserver" Pod="calico-apiserver-84956f4cf-72cgp" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-calico--apiserver--84956f4cf--72cgp-" Feb 13 19:12:20.719411 containerd[1503]: 2025-02-13 19:12:20.327 [INFO][4815] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="d366cd2cc4c7c16e754d70bbe9025807e19a94be019d47fbac45a766f5cecf79" Namespace="calico-apiserver" Pod="calico-apiserver-84956f4cf-72cgp" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-calico--apiserver--84956f4cf--72cgp-eth0" Feb 13 19:12:20.719411 containerd[1503]: 2025-02-13 19:12:20.489 [INFO][4877] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="d366cd2cc4c7c16e754d70bbe9025807e19a94be019d47fbac45a766f5cecf79" HandleID="k8s-pod-network.d366cd2cc4c7c16e754d70bbe9025807e19a94be019d47fbac45a766f5cecf79" Workload="ci--4230--0--1--a--2d796e8b97-k8s-calico--apiserver--84956f4cf--72cgp-eth0" Feb 13 19:12:20.719411 containerd[1503]: 2025-02-13 19:12:20.532 [INFO][4877] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="d366cd2cc4c7c16e754d70bbe9025807e19a94be019d47fbac45a766f5cecf79" HandleID="k8s-pod-network.d366cd2cc4c7c16e754d70bbe9025807e19a94be019d47fbac45a766f5cecf79" Workload="ci--4230--0--1--a--2d796e8b97-k8s-calico--apiserver--84956f4cf--72cgp-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002b4460), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4230-0-1-a-2d796e8b97", "pod":"calico-apiserver-84956f4cf-72cgp", "timestamp":"2025-02-13 19:12:20.489848679 +0000 UTC"}, Hostname:"ci-4230-0-1-a-2d796e8b97", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Feb 13 19:12:20.719411 containerd[1503]: 2025-02-13 19:12:20.533 [INFO][4877] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Feb 13 19:12:20.719411 containerd[1503]: 2025-02-13 19:12:20.534 [INFO][4877] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Feb 13 19:12:20.719411 containerd[1503]: 2025-02-13 19:12:20.535 [INFO][4877] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4230-0-1-a-2d796e8b97' Feb 13 19:12:20.719411 containerd[1503]: 2025-02-13 19:12:20.567 [INFO][4877] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.d366cd2cc4c7c16e754d70bbe9025807e19a94be019d47fbac45a766f5cecf79" host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:20.719411 containerd[1503]: 2025-02-13 19:12:20.616 [INFO][4877] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:20.719411 containerd[1503]: 2025-02-13 19:12:20.630 [INFO][4877] ipam/ipam.go 489: Trying affinity for 192.168.40.64/26 host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:20.719411 containerd[1503]: 2025-02-13 19:12:20.634 [INFO][4877] ipam/ipam.go 155: Attempting to load block cidr=192.168.40.64/26 host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:20.719411 containerd[1503]: 2025-02-13 19:12:20.642 [INFO][4877] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.40.64/26 host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:20.719411 containerd[1503]: 2025-02-13 19:12:20.642 [INFO][4877] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.40.64/26 handle="k8s-pod-network.d366cd2cc4c7c16e754d70bbe9025807e19a94be019d47fbac45a766f5cecf79" host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:20.719411 containerd[1503]: 2025-02-13 19:12:20.645 [INFO][4877] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.d366cd2cc4c7c16e754d70bbe9025807e19a94be019d47fbac45a766f5cecf79 Feb 13 19:12:20.719411 containerd[1503]: 2025-02-13 19:12:20.655 [INFO][4877] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.40.64/26 handle="k8s-pod-network.d366cd2cc4c7c16e754d70bbe9025807e19a94be019d47fbac45a766f5cecf79" host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:20.719411 containerd[1503]: 2025-02-13 19:12:20.668 [INFO][4877] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.40.66/26] block=192.168.40.64/26 handle="k8s-pod-network.d366cd2cc4c7c16e754d70bbe9025807e19a94be019d47fbac45a766f5cecf79" host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:20.719411 containerd[1503]: 2025-02-13 19:12:20.669 [INFO][4877] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.40.66/26] handle="k8s-pod-network.d366cd2cc4c7c16e754d70bbe9025807e19a94be019d47fbac45a766f5cecf79" host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:20.719411 containerd[1503]: 2025-02-13 19:12:20.669 [INFO][4877] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Feb 13 19:12:20.719411 containerd[1503]: 2025-02-13 19:12:20.669 [INFO][4877] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.40.66/26] IPv6=[] ContainerID="d366cd2cc4c7c16e754d70bbe9025807e19a94be019d47fbac45a766f5cecf79" HandleID="k8s-pod-network.d366cd2cc4c7c16e754d70bbe9025807e19a94be019d47fbac45a766f5cecf79" Workload="ci--4230--0--1--a--2d796e8b97-k8s-calico--apiserver--84956f4cf--72cgp-eth0" Feb 13 19:12:20.724078 containerd[1503]: 2025-02-13 19:12:20.677 [INFO][4815] cni-plugin/k8s.go 386: Populated endpoint ContainerID="d366cd2cc4c7c16e754d70bbe9025807e19a94be019d47fbac45a766f5cecf79" Namespace="calico-apiserver" Pod="calico-apiserver-84956f4cf-72cgp" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-calico--apiserver--84956f4cf--72cgp-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4230--0--1--a--2d796e8b97-k8s-calico--apiserver--84956f4cf--72cgp-eth0", GenerateName:"calico-apiserver-84956f4cf-", Namespace:"calico-apiserver", SelfLink:"", UID:"96785f8f-b4e3-456b-9a0d-bcc142e6beaf", ResourceVersion:"712", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 19, 12, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"84956f4cf", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4230-0-1-a-2d796e8b97", ContainerID:"", Pod:"calico-apiserver-84956f4cf-72cgp", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.40.66/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"caliac205cc4341", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 19:12:20.724078 containerd[1503]: 2025-02-13 19:12:20.677 [INFO][4815] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.40.66/32] ContainerID="d366cd2cc4c7c16e754d70bbe9025807e19a94be019d47fbac45a766f5cecf79" Namespace="calico-apiserver" Pod="calico-apiserver-84956f4cf-72cgp" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-calico--apiserver--84956f4cf--72cgp-eth0" Feb 13 19:12:20.724078 containerd[1503]: 2025-02-13 19:12:20.677 [INFO][4815] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to caliac205cc4341 ContainerID="d366cd2cc4c7c16e754d70bbe9025807e19a94be019d47fbac45a766f5cecf79" Namespace="calico-apiserver" Pod="calico-apiserver-84956f4cf-72cgp" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-calico--apiserver--84956f4cf--72cgp-eth0" Feb 13 19:12:20.724078 containerd[1503]: 2025-02-13 19:12:20.686 [INFO][4815] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="d366cd2cc4c7c16e754d70bbe9025807e19a94be019d47fbac45a766f5cecf79" Namespace="calico-apiserver" Pod="calico-apiserver-84956f4cf-72cgp" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-calico--apiserver--84956f4cf--72cgp-eth0" Feb 13 19:12:20.724078 containerd[1503]: 2025-02-13 19:12:20.688 [INFO][4815] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="d366cd2cc4c7c16e754d70bbe9025807e19a94be019d47fbac45a766f5cecf79" Namespace="calico-apiserver" Pod="calico-apiserver-84956f4cf-72cgp" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-calico--apiserver--84956f4cf--72cgp-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4230--0--1--a--2d796e8b97-k8s-calico--apiserver--84956f4cf--72cgp-eth0", GenerateName:"calico-apiserver-84956f4cf-", Namespace:"calico-apiserver", SelfLink:"", UID:"96785f8f-b4e3-456b-9a0d-bcc142e6beaf", ResourceVersion:"712", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 19, 12, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"84956f4cf", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4230-0-1-a-2d796e8b97", ContainerID:"d366cd2cc4c7c16e754d70bbe9025807e19a94be019d47fbac45a766f5cecf79", Pod:"calico-apiserver-84956f4cf-72cgp", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.40.66/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"caliac205cc4341", MAC:"16:fe:93:90:6b:8a", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 19:12:20.724078 containerd[1503]: 2025-02-13 19:12:20.711 [INFO][4815] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="d366cd2cc4c7c16e754d70bbe9025807e19a94be019d47fbac45a766f5cecf79" Namespace="calico-apiserver" Pod="calico-apiserver-84956f4cf-72cgp" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-calico--apiserver--84956f4cf--72cgp-eth0" Feb 13 19:12:20.719671 systemd[1]: Started cri-containerd-7c498096de4d518aa8367480279c86a097a53744f8a34963cdf1593cc7f82c21.scope - libcontainer container 7c498096de4d518aa8367480279c86a097a53744f8a34963cdf1593cc7f82c21. Feb 13 19:12:20.761453 containerd[1503]: time="2025-02-13T19:12:20.761341455Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:12:20.761660 containerd[1503]: time="2025-02-13T19:12:20.761480099Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:12:20.761660 containerd[1503]: time="2025-02-13T19:12:20.761530101Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:12:20.761895 containerd[1503]: time="2025-02-13T19:12:20.761739347Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:12:20.791968 systemd-networkd[1399]: cali4d5c8ae055c: Link UP Feb 13 19:12:20.793930 systemd-networkd[1399]: cali4d5c8ae055c: Gained carrier Feb 13 19:12:20.812167 containerd[1503]: time="2025-02-13T19:12:20.812119356Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-kkqg5,Uid:5cf24d69-eb75-4cb3-8437-024193f516f5,Namespace:kube-system,Attempt:6,} returns sandbox id \"7c498096de4d518aa8367480279c86a097a53744f8a34963cdf1593cc7f82c21\"" Feb 13 19:12:20.822243 containerd[1503]: time="2025-02-13T19:12:20.821504529Z" level=info msg="CreateContainer within sandbox \"7c498096de4d518aa8367480279c86a097a53744f8a34963cdf1593cc7f82c21\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Feb 13 19:12:20.825404 systemd[1]: Started cri-containerd-d366cd2cc4c7c16e754d70bbe9025807e19a94be019d47fbac45a766f5cecf79.scope - libcontainer container d366cd2cc4c7c16e754d70bbe9025807e19a94be019d47fbac45a766f5cecf79. Feb 13 19:12:20.829521 containerd[1503]: 2025-02-13 19:12:20.255 [INFO][4792] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Feb 13 19:12:20.829521 containerd[1503]: 2025-02-13 19:12:20.314 [INFO][4792] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4230--0--1--a--2d796e8b97-k8s-calico--kube--controllers--78bd6d67c9--b259q-eth0 calico-kube-controllers-78bd6d67c9- calico-system 436348d7-7832-43e1-8780-623c434b2666 714 0 2025-02-13 19:12:05 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:78bd6d67c9 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s ci-4230-0-1-a-2d796e8b97 calico-kube-controllers-78bd6d67c9-b259q eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali4d5c8ae055c [] []}} ContainerID="9b637346f8653af89f0cb9d44b0723cb41c71875e3fcf79734935232fb4eb95b" Namespace="calico-system" Pod="calico-kube-controllers-78bd6d67c9-b259q" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-calico--kube--controllers--78bd6d67c9--b259q-" Feb 13 19:12:20.829521 containerd[1503]: 2025-02-13 19:12:20.315 [INFO][4792] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="9b637346f8653af89f0cb9d44b0723cb41c71875e3fcf79734935232fb4eb95b" Namespace="calico-system" Pod="calico-kube-controllers-78bd6d67c9-b259q" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-calico--kube--controllers--78bd6d67c9--b259q-eth0" Feb 13 19:12:20.829521 containerd[1503]: 2025-02-13 19:12:20.511 [INFO][4873] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="9b637346f8653af89f0cb9d44b0723cb41c71875e3fcf79734935232fb4eb95b" HandleID="k8s-pod-network.9b637346f8653af89f0cb9d44b0723cb41c71875e3fcf79734935232fb4eb95b" Workload="ci--4230--0--1--a--2d796e8b97-k8s-calico--kube--controllers--78bd6d67c9--b259q-eth0" Feb 13 19:12:20.829521 containerd[1503]: 2025-02-13 19:12:20.555 [INFO][4873] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="9b637346f8653af89f0cb9d44b0723cb41c71875e3fcf79734935232fb4eb95b" HandleID="k8s-pod-network.9b637346f8653af89f0cb9d44b0723cb41c71875e3fcf79734935232fb4eb95b" Workload="ci--4230--0--1--a--2d796e8b97-k8s-calico--kube--controllers--78bd6d67c9--b259q-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x400039bb90), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4230-0-1-a-2d796e8b97", "pod":"calico-kube-controllers-78bd6d67c9-b259q", "timestamp":"2025-02-13 19:12:20.511796482 +0000 UTC"}, Hostname:"ci-4230-0-1-a-2d796e8b97", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Feb 13 19:12:20.829521 containerd[1503]: 2025-02-13 19:12:20.557 [INFO][4873] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Feb 13 19:12:20.829521 containerd[1503]: 2025-02-13 19:12:20.669 [INFO][4873] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Feb 13 19:12:20.829521 containerd[1503]: 2025-02-13 19:12:20.670 [INFO][4873] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4230-0-1-a-2d796e8b97' Feb 13 19:12:20.829521 containerd[1503]: 2025-02-13 19:12:20.678 [INFO][4873] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.9b637346f8653af89f0cb9d44b0723cb41c71875e3fcf79734935232fb4eb95b" host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:20.829521 containerd[1503]: 2025-02-13 19:12:20.718 [INFO][4873] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:20.829521 containerd[1503]: 2025-02-13 19:12:20.732 [INFO][4873] ipam/ipam.go 489: Trying affinity for 192.168.40.64/26 host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:20.829521 containerd[1503]: 2025-02-13 19:12:20.735 [INFO][4873] ipam/ipam.go 155: Attempting to load block cidr=192.168.40.64/26 host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:20.829521 containerd[1503]: 2025-02-13 19:12:20.739 [INFO][4873] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.40.64/26 host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:20.829521 containerd[1503]: 2025-02-13 19:12:20.739 [INFO][4873] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.40.64/26 handle="k8s-pod-network.9b637346f8653af89f0cb9d44b0723cb41c71875e3fcf79734935232fb4eb95b" host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:20.829521 containerd[1503]: 2025-02-13 19:12:20.743 [INFO][4873] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.9b637346f8653af89f0cb9d44b0723cb41c71875e3fcf79734935232fb4eb95b Feb 13 19:12:20.829521 containerd[1503]: 2025-02-13 19:12:20.753 [INFO][4873] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.40.64/26 handle="k8s-pod-network.9b637346f8653af89f0cb9d44b0723cb41c71875e3fcf79734935232fb4eb95b" host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:20.829521 containerd[1503]: 2025-02-13 19:12:20.767 [INFO][4873] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.40.67/26] block=192.168.40.64/26 handle="k8s-pod-network.9b637346f8653af89f0cb9d44b0723cb41c71875e3fcf79734935232fb4eb95b" host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:20.829521 containerd[1503]: 2025-02-13 19:12:20.767 [INFO][4873] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.40.67/26] handle="k8s-pod-network.9b637346f8653af89f0cb9d44b0723cb41c71875e3fcf79734935232fb4eb95b" host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:20.829521 containerd[1503]: 2025-02-13 19:12:20.767 [INFO][4873] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Feb 13 19:12:20.829521 containerd[1503]: 2025-02-13 19:12:20.767 [INFO][4873] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.40.67/26] IPv6=[] ContainerID="9b637346f8653af89f0cb9d44b0723cb41c71875e3fcf79734935232fb4eb95b" HandleID="k8s-pod-network.9b637346f8653af89f0cb9d44b0723cb41c71875e3fcf79734935232fb4eb95b" Workload="ci--4230--0--1--a--2d796e8b97-k8s-calico--kube--controllers--78bd6d67c9--b259q-eth0" Feb 13 19:12:20.830241 containerd[1503]: 2025-02-13 19:12:20.774 [INFO][4792] cni-plugin/k8s.go 386: Populated endpoint ContainerID="9b637346f8653af89f0cb9d44b0723cb41c71875e3fcf79734935232fb4eb95b" Namespace="calico-system" Pod="calico-kube-controllers-78bd6d67c9-b259q" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-calico--kube--controllers--78bd6d67c9--b259q-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4230--0--1--a--2d796e8b97-k8s-calico--kube--controllers--78bd6d67c9--b259q-eth0", GenerateName:"calico-kube-controllers-78bd6d67c9-", Namespace:"calico-system", SelfLink:"", UID:"436348d7-7832-43e1-8780-623c434b2666", ResourceVersion:"714", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 19, 12, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"78bd6d67c9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4230-0-1-a-2d796e8b97", ContainerID:"", Pod:"calico-kube-controllers-78bd6d67c9-b259q", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.40.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali4d5c8ae055c", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 19:12:20.830241 containerd[1503]: 2025-02-13 19:12:20.776 [INFO][4792] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.40.67/32] ContainerID="9b637346f8653af89f0cb9d44b0723cb41c71875e3fcf79734935232fb4eb95b" Namespace="calico-system" Pod="calico-kube-controllers-78bd6d67c9-b259q" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-calico--kube--controllers--78bd6d67c9--b259q-eth0" Feb 13 19:12:20.830241 containerd[1503]: 2025-02-13 19:12:20.776 [INFO][4792] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali4d5c8ae055c ContainerID="9b637346f8653af89f0cb9d44b0723cb41c71875e3fcf79734935232fb4eb95b" Namespace="calico-system" Pod="calico-kube-controllers-78bd6d67c9-b259q" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-calico--kube--controllers--78bd6d67c9--b259q-eth0" Feb 13 19:12:20.830241 containerd[1503]: 2025-02-13 19:12:20.797 [INFO][4792] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="9b637346f8653af89f0cb9d44b0723cb41c71875e3fcf79734935232fb4eb95b" Namespace="calico-system" Pod="calico-kube-controllers-78bd6d67c9-b259q" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-calico--kube--controllers--78bd6d67c9--b259q-eth0" Feb 13 19:12:20.830241 containerd[1503]: 2025-02-13 19:12:20.800 [INFO][4792] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="9b637346f8653af89f0cb9d44b0723cb41c71875e3fcf79734935232fb4eb95b" Namespace="calico-system" Pod="calico-kube-controllers-78bd6d67c9-b259q" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-calico--kube--controllers--78bd6d67c9--b259q-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4230--0--1--a--2d796e8b97-k8s-calico--kube--controllers--78bd6d67c9--b259q-eth0", GenerateName:"calico-kube-controllers-78bd6d67c9-", Namespace:"calico-system", SelfLink:"", UID:"436348d7-7832-43e1-8780-623c434b2666", ResourceVersion:"714", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 19, 12, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"78bd6d67c9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4230-0-1-a-2d796e8b97", ContainerID:"9b637346f8653af89f0cb9d44b0723cb41c71875e3fcf79734935232fb4eb95b", Pod:"calico-kube-controllers-78bd6d67c9-b259q", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.40.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali4d5c8ae055c", MAC:"ca:ec:90:30:fa:9d", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 19:12:20.830241 containerd[1503]: 2025-02-13 19:12:20.823 [INFO][4792] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="9b637346f8653af89f0cb9d44b0723cb41c71875e3fcf79734935232fb4eb95b" Namespace="calico-system" Pod="calico-kube-controllers-78bd6d67c9-b259q" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-calico--kube--controllers--78bd6d67c9--b259q-eth0" Feb 13 19:12:20.853618 containerd[1503]: time="2025-02-13T19:12:20.853179275Z" level=info msg="CreateContainer within sandbox \"7c498096de4d518aa8367480279c86a097a53744f8a34963cdf1593cc7f82c21\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"0df3bd8dea4896d8646702388fb8e1e02d80c32a131f7f66024278e7798d8c7f\"" Feb 13 19:12:20.854452 containerd[1503]: time="2025-02-13T19:12:20.854028062Z" level=info msg="StartContainer for \"0df3bd8dea4896d8646702388fb8e1e02d80c32a131f7f66024278e7798d8c7f\"" Feb 13 19:12:20.882148 containerd[1503]: time="2025-02-13T19:12:20.881564999Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:12:20.882148 containerd[1503]: time="2025-02-13T19:12:20.882017773Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:12:20.882148 containerd[1503]: time="2025-02-13T19:12:20.882031854Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:12:20.882655 containerd[1503]: time="2025-02-13T19:12:20.882492748Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:12:20.911374 systemd-networkd[1399]: calicde3fabc286: Link UP Feb 13 19:12:20.913889 systemd-networkd[1399]: calicde3fabc286: Gained carrier Feb 13 19:12:20.928029 systemd[1]: Started cri-containerd-9b637346f8653af89f0cb9d44b0723cb41c71875e3fcf79734935232fb4eb95b.scope - libcontainer container 9b637346f8653af89f0cb9d44b0723cb41c71875e3fcf79734935232fb4eb95b. Feb 13 19:12:20.944603 systemd[1]: Started cri-containerd-0df3bd8dea4896d8646702388fb8e1e02d80c32a131f7f66024278e7798d8c7f.scope - libcontainer container 0df3bd8dea4896d8646702388fb8e1e02d80c32a131f7f66024278e7798d8c7f. Feb 13 19:12:20.975169 containerd[1503]: 2025-02-13 19:12:20.322 [INFO][4814] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Feb 13 19:12:20.975169 containerd[1503]: 2025-02-13 19:12:20.370 [INFO][4814] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4230--0--1--a--2d796e8b97-k8s-calico--apiserver--84956f4cf--p6mr2-eth0 calico-apiserver-84956f4cf- calico-apiserver 05e39f6f-0a83-4b62-8cff-b0d4799f8f89 710 0 2025-02-13 19:12:05 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:84956f4cf projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4230-0-1-a-2d796e8b97 calico-apiserver-84956f4cf-p6mr2 eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] calicde3fabc286 [] []}} ContainerID="f2541446bafb170e13b958b7bea1fd0f66a6984abb6abdb77c88a352ed058c69" Namespace="calico-apiserver" Pod="calico-apiserver-84956f4cf-p6mr2" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-calico--apiserver--84956f4cf--p6mr2-" Feb 13 19:12:20.975169 containerd[1503]: 2025-02-13 19:12:20.371 [INFO][4814] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="f2541446bafb170e13b958b7bea1fd0f66a6984abb6abdb77c88a352ed058c69" Namespace="calico-apiserver" Pod="calico-apiserver-84956f4cf-p6mr2" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-calico--apiserver--84956f4cf--p6mr2-eth0" Feb 13 19:12:20.975169 containerd[1503]: 2025-02-13 19:12:20.526 [INFO][4886] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="f2541446bafb170e13b958b7bea1fd0f66a6984abb6abdb77c88a352ed058c69" HandleID="k8s-pod-network.f2541446bafb170e13b958b7bea1fd0f66a6984abb6abdb77c88a352ed058c69" Workload="ci--4230--0--1--a--2d796e8b97-k8s-calico--apiserver--84956f4cf--p6mr2-eth0" Feb 13 19:12:20.975169 containerd[1503]: 2025-02-13 19:12:20.615 [INFO][4886] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="f2541446bafb170e13b958b7bea1fd0f66a6984abb6abdb77c88a352ed058c69" HandleID="k8s-pod-network.f2541446bafb170e13b958b7bea1fd0f66a6984abb6abdb77c88a352ed058c69" Workload="ci--4230--0--1--a--2d796e8b97-k8s-calico--apiserver--84956f4cf--p6mr2-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40003e0f20), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4230-0-1-a-2d796e8b97", "pod":"calico-apiserver-84956f4cf-p6mr2", "timestamp":"2025-02-13 19:12:20.526115688 +0000 UTC"}, Hostname:"ci-4230-0-1-a-2d796e8b97", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Feb 13 19:12:20.975169 containerd[1503]: 2025-02-13 19:12:20.615 [INFO][4886] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Feb 13 19:12:20.975169 containerd[1503]: 2025-02-13 19:12:20.767 [INFO][4886] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Feb 13 19:12:20.975169 containerd[1503]: 2025-02-13 19:12:20.768 [INFO][4886] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4230-0-1-a-2d796e8b97' Feb 13 19:12:20.975169 containerd[1503]: 2025-02-13 19:12:20.788 [INFO][4886] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.f2541446bafb170e13b958b7bea1fd0f66a6984abb6abdb77c88a352ed058c69" host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:20.975169 containerd[1503]: 2025-02-13 19:12:20.821 [INFO][4886] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:20.975169 containerd[1503]: 2025-02-13 19:12:20.837 [INFO][4886] ipam/ipam.go 489: Trying affinity for 192.168.40.64/26 host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:20.975169 containerd[1503]: 2025-02-13 19:12:20.843 [INFO][4886] ipam/ipam.go 155: Attempting to load block cidr=192.168.40.64/26 host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:20.975169 containerd[1503]: 2025-02-13 19:12:20.850 [INFO][4886] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.40.64/26 host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:20.975169 containerd[1503]: 2025-02-13 19:12:20.851 [INFO][4886] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.40.64/26 handle="k8s-pod-network.f2541446bafb170e13b958b7bea1fd0f66a6984abb6abdb77c88a352ed058c69" host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:20.975169 containerd[1503]: 2025-02-13 19:12:20.855 [INFO][4886] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.f2541446bafb170e13b958b7bea1fd0f66a6984abb6abdb77c88a352ed058c69 Feb 13 19:12:20.975169 containerd[1503]: 2025-02-13 19:12:20.867 [INFO][4886] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.40.64/26 handle="k8s-pod-network.f2541446bafb170e13b958b7bea1fd0f66a6984abb6abdb77c88a352ed058c69" host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:20.975169 containerd[1503]: 2025-02-13 19:12:20.888 [INFO][4886] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.40.68/26] block=192.168.40.64/26 handle="k8s-pod-network.f2541446bafb170e13b958b7bea1fd0f66a6984abb6abdb77c88a352ed058c69" host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:20.975169 containerd[1503]: 2025-02-13 19:12:20.889 [INFO][4886] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.40.68/26] handle="k8s-pod-network.f2541446bafb170e13b958b7bea1fd0f66a6984abb6abdb77c88a352ed058c69" host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:20.975169 containerd[1503]: 2025-02-13 19:12:20.889 [INFO][4886] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Feb 13 19:12:20.975169 containerd[1503]: 2025-02-13 19:12:20.890 [INFO][4886] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.40.68/26] IPv6=[] ContainerID="f2541446bafb170e13b958b7bea1fd0f66a6984abb6abdb77c88a352ed058c69" HandleID="k8s-pod-network.f2541446bafb170e13b958b7bea1fd0f66a6984abb6abdb77c88a352ed058c69" Workload="ci--4230--0--1--a--2d796e8b97-k8s-calico--apiserver--84956f4cf--p6mr2-eth0" Feb 13 19:12:20.976621 containerd[1503]: 2025-02-13 19:12:20.901 [INFO][4814] cni-plugin/k8s.go 386: Populated endpoint ContainerID="f2541446bafb170e13b958b7bea1fd0f66a6984abb6abdb77c88a352ed058c69" Namespace="calico-apiserver" Pod="calico-apiserver-84956f4cf-p6mr2" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-calico--apiserver--84956f4cf--p6mr2-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4230--0--1--a--2d796e8b97-k8s-calico--apiserver--84956f4cf--p6mr2-eth0", GenerateName:"calico-apiserver-84956f4cf-", Namespace:"calico-apiserver", SelfLink:"", UID:"05e39f6f-0a83-4b62-8cff-b0d4799f8f89", ResourceVersion:"710", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 19, 12, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"84956f4cf", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4230-0-1-a-2d796e8b97", ContainerID:"", Pod:"calico-apiserver-84956f4cf-p6mr2", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.40.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calicde3fabc286", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 19:12:20.976621 containerd[1503]: 2025-02-13 19:12:20.902 [INFO][4814] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.40.68/32] ContainerID="f2541446bafb170e13b958b7bea1fd0f66a6984abb6abdb77c88a352ed058c69" Namespace="calico-apiserver" Pod="calico-apiserver-84956f4cf-p6mr2" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-calico--apiserver--84956f4cf--p6mr2-eth0" Feb 13 19:12:20.976621 containerd[1503]: 2025-02-13 19:12:20.902 [INFO][4814] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calicde3fabc286 ContainerID="f2541446bafb170e13b958b7bea1fd0f66a6984abb6abdb77c88a352ed058c69" Namespace="calico-apiserver" Pod="calico-apiserver-84956f4cf-p6mr2" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-calico--apiserver--84956f4cf--p6mr2-eth0" Feb 13 19:12:20.976621 containerd[1503]: 2025-02-13 19:12:20.911 [INFO][4814] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="f2541446bafb170e13b958b7bea1fd0f66a6984abb6abdb77c88a352ed058c69" Namespace="calico-apiserver" Pod="calico-apiserver-84956f4cf-p6mr2" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-calico--apiserver--84956f4cf--p6mr2-eth0" Feb 13 19:12:20.976621 containerd[1503]: 2025-02-13 19:12:20.918 [INFO][4814] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="f2541446bafb170e13b958b7bea1fd0f66a6984abb6abdb77c88a352ed058c69" Namespace="calico-apiserver" Pod="calico-apiserver-84956f4cf-p6mr2" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-calico--apiserver--84956f4cf--p6mr2-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4230--0--1--a--2d796e8b97-k8s-calico--apiserver--84956f4cf--p6mr2-eth0", GenerateName:"calico-apiserver-84956f4cf-", Namespace:"calico-apiserver", SelfLink:"", UID:"05e39f6f-0a83-4b62-8cff-b0d4799f8f89", ResourceVersion:"710", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 19, 12, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"84956f4cf", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4230-0-1-a-2d796e8b97", ContainerID:"f2541446bafb170e13b958b7bea1fd0f66a6984abb6abdb77c88a352ed058c69", Pod:"calico-apiserver-84956f4cf-p6mr2", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.40.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calicde3fabc286", MAC:"d2:47:7a:4b:f3:70", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 19:12:20.976621 containerd[1503]: 2025-02-13 19:12:20.944 [INFO][4814] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="f2541446bafb170e13b958b7bea1fd0f66a6984abb6abdb77c88a352ed058c69" Namespace="calico-apiserver" Pod="calico-apiserver-84956f4cf-p6mr2" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-calico--apiserver--84956f4cf--p6mr2-eth0" Feb 13 19:12:20.985111 containerd[1503]: time="2025-02-13T19:12:20.985045622Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-84956f4cf-72cgp,Uid:96785f8f-b4e3-456b-9a0d-bcc142e6beaf,Namespace:calico-apiserver,Attempt:6,} returns sandbox id \"d366cd2cc4c7c16e754d70bbe9025807e19a94be019d47fbac45a766f5cecf79\"" Feb 13 19:12:20.990962 containerd[1503]: time="2025-02-13T19:12:20.990894564Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.1\"" Feb 13 19:12:21.021778 systemd-networkd[1399]: calic2e7ce92ea9: Link UP Feb 13 19:12:21.022107 systemd-networkd[1399]: calic2e7ce92ea9: Gained carrier Feb 13 19:12:21.087868 containerd[1503]: 2025-02-13 19:12:20.358 [INFO][4839] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Feb 13 19:12:21.087868 containerd[1503]: 2025-02-13 19:12:20.429 [INFO][4839] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4230--0--1--a--2d796e8b97-k8s-coredns--668d6bf9bc--vx9cw-eth0 coredns-668d6bf9bc- kube-system eb0c3c87-077a-4b2b-a168-5fd9131bccd3 715 0 2025-02-13 19:11:41 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4230-0-1-a-2d796e8b97 coredns-668d6bf9bc-vx9cw eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] calic2e7ce92ea9 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="45225a5d0a993e64267075f2bbdc29d9ae95a25629a61da4ce1da8b6ff614881" Namespace="kube-system" Pod="coredns-668d6bf9bc-vx9cw" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-coredns--668d6bf9bc--vx9cw-" Feb 13 19:12:21.087868 containerd[1503]: 2025-02-13 19:12:20.429 [INFO][4839] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="45225a5d0a993e64267075f2bbdc29d9ae95a25629a61da4ce1da8b6ff614881" Namespace="kube-system" Pod="coredns-668d6bf9bc-vx9cw" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-coredns--668d6bf9bc--vx9cw-eth0" Feb 13 19:12:21.087868 containerd[1503]: 2025-02-13 19:12:20.520 [INFO][4891] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="45225a5d0a993e64267075f2bbdc29d9ae95a25629a61da4ce1da8b6ff614881" HandleID="k8s-pod-network.45225a5d0a993e64267075f2bbdc29d9ae95a25629a61da4ce1da8b6ff614881" Workload="ci--4230--0--1--a--2d796e8b97-k8s-coredns--668d6bf9bc--vx9cw-eth0" Feb 13 19:12:21.087868 containerd[1503]: 2025-02-13 19:12:20.615 [INFO][4891] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="45225a5d0a993e64267075f2bbdc29d9ae95a25629a61da4ce1da8b6ff614881" HandleID="k8s-pod-network.45225a5d0a993e64267075f2bbdc29d9ae95a25629a61da4ce1da8b6ff614881" Workload="ci--4230--0--1--a--2d796e8b97-k8s-coredns--668d6bf9bc--vx9cw-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40003164f0), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4230-0-1-a-2d796e8b97", "pod":"coredns-668d6bf9bc-vx9cw", "timestamp":"2025-02-13 19:12:20.520492393 +0000 UTC"}, Hostname:"ci-4230-0-1-a-2d796e8b97", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Feb 13 19:12:21.087868 containerd[1503]: 2025-02-13 19:12:20.615 [INFO][4891] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Feb 13 19:12:21.087868 containerd[1503]: 2025-02-13 19:12:20.889 [INFO][4891] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Feb 13 19:12:21.087868 containerd[1503]: 2025-02-13 19:12:20.890 [INFO][4891] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4230-0-1-a-2d796e8b97' Feb 13 19:12:21.087868 containerd[1503]: 2025-02-13 19:12:20.895 [INFO][4891] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.45225a5d0a993e64267075f2bbdc29d9ae95a25629a61da4ce1da8b6ff614881" host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:21.087868 containerd[1503]: 2025-02-13 19:12:20.917 [INFO][4891] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:21.087868 containerd[1503]: 2025-02-13 19:12:20.941 [INFO][4891] ipam/ipam.go 489: Trying affinity for 192.168.40.64/26 host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:21.087868 containerd[1503]: 2025-02-13 19:12:20.949 [INFO][4891] ipam/ipam.go 155: Attempting to load block cidr=192.168.40.64/26 host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:21.087868 containerd[1503]: 2025-02-13 19:12:20.957 [INFO][4891] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.40.64/26 host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:21.087868 containerd[1503]: 2025-02-13 19:12:20.957 [INFO][4891] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.40.64/26 handle="k8s-pod-network.45225a5d0a993e64267075f2bbdc29d9ae95a25629a61da4ce1da8b6ff614881" host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:21.087868 containerd[1503]: 2025-02-13 19:12:20.961 [INFO][4891] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.45225a5d0a993e64267075f2bbdc29d9ae95a25629a61da4ce1da8b6ff614881 Feb 13 19:12:21.087868 containerd[1503]: 2025-02-13 19:12:20.974 [INFO][4891] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.40.64/26 handle="k8s-pod-network.45225a5d0a993e64267075f2bbdc29d9ae95a25629a61da4ce1da8b6ff614881" host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:21.087868 containerd[1503]: 2025-02-13 19:12:20.999 [INFO][4891] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.40.69/26] block=192.168.40.64/26 handle="k8s-pod-network.45225a5d0a993e64267075f2bbdc29d9ae95a25629a61da4ce1da8b6ff614881" host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:21.087868 containerd[1503]: 2025-02-13 19:12:20.999 [INFO][4891] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.40.69/26] handle="k8s-pod-network.45225a5d0a993e64267075f2bbdc29d9ae95a25629a61da4ce1da8b6ff614881" host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:21.087868 containerd[1503]: 2025-02-13 19:12:20.999 [INFO][4891] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Feb 13 19:12:21.087868 containerd[1503]: 2025-02-13 19:12:20.999 [INFO][4891] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.40.69/26] IPv6=[] ContainerID="45225a5d0a993e64267075f2bbdc29d9ae95a25629a61da4ce1da8b6ff614881" HandleID="k8s-pod-network.45225a5d0a993e64267075f2bbdc29d9ae95a25629a61da4ce1da8b6ff614881" Workload="ci--4230--0--1--a--2d796e8b97-k8s-coredns--668d6bf9bc--vx9cw-eth0" Feb 13 19:12:21.088751 containerd[1503]: 2025-02-13 19:12:21.010 [INFO][4839] cni-plugin/k8s.go 386: Populated endpoint ContainerID="45225a5d0a993e64267075f2bbdc29d9ae95a25629a61da4ce1da8b6ff614881" Namespace="kube-system" Pod="coredns-668d6bf9bc-vx9cw" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-coredns--668d6bf9bc--vx9cw-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4230--0--1--a--2d796e8b97-k8s-coredns--668d6bf9bc--vx9cw-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"eb0c3c87-077a-4b2b-a168-5fd9131bccd3", ResourceVersion:"715", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 19, 11, 41, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4230-0-1-a-2d796e8b97", ContainerID:"", Pod:"coredns-668d6bf9bc-vx9cw", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.40.69/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calic2e7ce92ea9", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 19:12:21.088751 containerd[1503]: 2025-02-13 19:12:21.010 [INFO][4839] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.40.69/32] ContainerID="45225a5d0a993e64267075f2bbdc29d9ae95a25629a61da4ce1da8b6ff614881" Namespace="kube-system" Pod="coredns-668d6bf9bc-vx9cw" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-coredns--668d6bf9bc--vx9cw-eth0" Feb 13 19:12:21.088751 containerd[1503]: 2025-02-13 19:12:21.010 [INFO][4839] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calic2e7ce92ea9 ContainerID="45225a5d0a993e64267075f2bbdc29d9ae95a25629a61da4ce1da8b6ff614881" Namespace="kube-system" Pod="coredns-668d6bf9bc-vx9cw" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-coredns--668d6bf9bc--vx9cw-eth0" Feb 13 19:12:21.088751 containerd[1503]: 2025-02-13 19:12:21.035 [INFO][4839] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="45225a5d0a993e64267075f2bbdc29d9ae95a25629a61da4ce1da8b6ff614881" Namespace="kube-system" Pod="coredns-668d6bf9bc-vx9cw" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-coredns--668d6bf9bc--vx9cw-eth0" Feb 13 19:12:21.088751 containerd[1503]: 2025-02-13 19:12:21.054 [INFO][4839] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="45225a5d0a993e64267075f2bbdc29d9ae95a25629a61da4ce1da8b6ff614881" Namespace="kube-system" Pod="coredns-668d6bf9bc-vx9cw" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-coredns--668d6bf9bc--vx9cw-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4230--0--1--a--2d796e8b97-k8s-coredns--668d6bf9bc--vx9cw-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"eb0c3c87-077a-4b2b-a168-5fd9131bccd3", ResourceVersion:"715", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 19, 11, 41, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4230-0-1-a-2d796e8b97", ContainerID:"45225a5d0a993e64267075f2bbdc29d9ae95a25629a61da4ce1da8b6ff614881", Pod:"coredns-668d6bf9bc-vx9cw", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.40.69/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calic2e7ce92ea9", MAC:"4e:cd:89:d5:90:7e", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 19:12:21.088751 containerd[1503]: 2025-02-13 19:12:21.081 [INFO][4839] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="45225a5d0a993e64267075f2bbdc29d9ae95a25629a61da4ce1da8b6ff614881" Namespace="kube-system" Pod="coredns-668d6bf9bc-vx9cw" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-coredns--668d6bf9bc--vx9cw-eth0" Feb 13 19:12:21.112991 containerd[1503]: time="2025-02-13T19:12:21.112926029Z" level=info msg="StartContainer for \"0df3bd8dea4896d8646702388fb8e1e02d80c32a131f7f66024278e7798d8c7f\" returns successfully" Feb 13 19:12:21.188897 systemd-networkd[1399]: cali05f501d3577: Link UP Feb 13 19:12:21.190920 systemd-networkd[1399]: cali05f501d3577: Gained carrier Feb 13 19:12:21.196326 containerd[1503]: time="2025-02-13T19:12:21.182940344Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:12:21.196326 containerd[1503]: time="2025-02-13T19:12:21.182999346Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:12:21.196326 containerd[1503]: time="2025-02-13T19:12:21.183014626Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:12:21.196326 containerd[1503]: time="2025-02-13T19:12:21.183129150Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:12:21.211014 containerd[1503]: time="2025-02-13T19:12:21.208692311Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-78bd6d67c9-b259q,Uid:436348d7-7832-43e1-8780-623c434b2666,Namespace:calico-system,Attempt:6,} returns sandbox id \"9b637346f8653af89f0cb9d44b0723cb41c71875e3fcf79734935232fb4eb95b\"" Feb 13 19:12:21.226513 containerd[1503]: time="2025-02-13T19:12:21.226187540Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:12:21.226513 containerd[1503]: time="2025-02-13T19:12:21.226262342Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:12:21.226513 containerd[1503]: time="2025-02-13T19:12:21.226287183Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:12:21.226513 containerd[1503]: time="2025-02-13T19:12:21.226372986Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:12:21.271842 containerd[1503]: 2025-02-13 19:12:20.425 [INFO][4849] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Feb 13 19:12:21.271842 containerd[1503]: 2025-02-13 19:12:20.478 [INFO][4849] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4230--0--1--a--2d796e8b97-k8s-csi--node--driver--6v9pr-eth0 csi-node-driver- calico-system b9c8452b-31d9-47bc-9697-dd4f293d71b9 634 0 2025-02-13 19:12:05 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:84cddb44f k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s ci-4230-0-1-a-2d796e8b97 csi-node-driver-6v9pr eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali05f501d3577 [] []}} ContainerID="54d08303486449cd8591710ce79b951324a215caf30205c5f317c06267df138e" Namespace="calico-system" Pod="csi-node-driver-6v9pr" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-csi--node--driver--6v9pr-" Feb 13 19:12:21.271842 containerd[1503]: 2025-02-13 19:12:20.478 [INFO][4849] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="54d08303486449cd8591710ce79b951324a215caf30205c5f317c06267df138e" Namespace="calico-system" Pod="csi-node-driver-6v9pr" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-csi--node--driver--6v9pr-eth0" Feb 13 19:12:21.271842 containerd[1503]: 2025-02-13 19:12:20.562 [INFO][4899] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="54d08303486449cd8591710ce79b951324a215caf30205c5f317c06267df138e" HandleID="k8s-pod-network.54d08303486449cd8591710ce79b951324a215caf30205c5f317c06267df138e" Workload="ci--4230--0--1--a--2d796e8b97-k8s-csi--node--driver--6v9pr-eth0" Feb 13 19:12:21.271842 containerd[1503]: 2025-02-13 19:12:20.617 [INFO][4899] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="54d08303486449cd8591710ce79b951324a215caf30205c5f317c06267df138e" HandleID="k8s-pod-network.54d08303486449cd8591710ce79b951324a215caf30205c5f317c06267df138e" Workload="ci--4230--0--1--a--2d796e8b97-k8s-csi--node--driver--6v9pr-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x400028cab0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4230-0-1-a-2d796e8b97", "pod":"csi-node-driver-6v9pr", "timestamp":"2025-02-13 19:12:20.561954965 +0000 UTC"}, Hostname:"ci-4230-0-1-a-2d796e8b97", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Feb 13 19:12:21.271842 containerd[1503]: 2025-02-13 19:12:20.617 [INFO][4899] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Feb 13 19:12:21.271842 containerd[1503]: 2025-02-13 19:12:20.999 [INFO][4899] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Feb 13 19:12:21.271842 containerd[1503]: 2025-02-13 19:12:20.999 [INFO][4899] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4230-0-1-a-2d796e8b97' Feb 13 19:12:21.271842 containerd[1503]: 2025-02-13 19:12:21.008 [INFO][4899] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.54d08303486449cd8591710ce79b951324a215caf30205c5f317c06267df138e" host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:21.271842 containerd[1503]: 2025-02-13 19:12:21.040 [INFO][4899] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:21.271842 containerd[1503]: 2025-02-13 19:12:21.084 [INFO][4899] ipam/ipam.go 489: Trying affinity for 192.168.40.64/26 host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:21.271842 containerd[1503]: 2025-02-13 19:12:21.092 [INFO][4899] ipam/ipam.go 155: Attempting to load block cidr=192.168.40.64/26 host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:21.271842 containerd[1503]: 2025-02-13 19:12:21.105 [INFO][4899] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.40.64/26 host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:21.271842 containerd[1503]: 2025-02-13 19:12:21.109 [INFO][4899] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.40.64/26 handle="k8s-pod-network.54d08303486449cd8591710ce79b951324a215caf30205c5f317c06267df138e" host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:21.271842 containerd[1503]: 2025-02-13 19:12:21.116 [INFO][4899] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.54d08303486449cd8591710ce79b951324a215caf30205c5f317c06267df138e Feb 13 19:12:21.271842 containerd[1503]: 2025-02-13 19:12:21.130 [INFO][4899] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.40.64/26 handle="k8s-pod-network.54d08303486449cd8591710ce79b951324a215caf30205c5f317c06267df138e" host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:21.271842 containerd[1503]: 2025-02-13 19:12:21.154 [INFO][4899] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.40.70/26] block=192.168.40.64/26 handle="k8s-pod-network.54d08303486449cd8591710ce79b951324a215caf30205c5f317c06267df138e" host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:21.271842 containerd[1503]: 2025-02-13 19:12:21.155 [INFO][4899] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.40.70/26] handle="k8s-pod-network.54d08303486449cd8591710ce79b951324a215caf30205c5f317c06267df138e" host="ci-4230-0-1-a-2d796e8b97" Feb 13 19:12:21.271842 containerd[1503]: 2025-02-13 19:12:21.155 [INFO][4899] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Feb 13 19:12:21.271842 containerd[1503]: 2025-02-13 19:12:21.155 [INFO][4899] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.40.70/26] IPv6=[] ContainerID="54d08303486449cd8591710ce79b951324a215caf30205c5f317c06267df138e" HandleID="k8s-pod-network.54d08303486449cd8591710ce79b951324a215caf30205c5f317c06267df138e" Workload="ci--4230--0--1--a--2d796e8b97-k8s-csi--node--driver--6v9pr-eth0" Feb 13 19:12:21.275809 containerd[1503]: 2025-02-13 19:12:21.160 [INFO][4849] cni-plugin/k8s.go 386: Populated endpoint ContainerID="54d08303486449cd8591710ce79b951324a215caf30205c5f317c06267df138e" Namespace="calico-system" Pod="csi-node-driver-6v9pr" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-csi--node--driver--6v9pr-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4230--0--1--a--2d796e8b97-k8s-csi--node--driver--6v9pr-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"b9c8452b-31d9-47bc-9697-dd4f293d71b9", ResourceVersion:"634", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 19, 12, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"84cddb44f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4230-0-1-a-2d796e8b97", ContainerID:"", Pod:"csi-node-driver-6v9pr", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.40.70/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali05f501d3577", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 19:12:21.275809 containerd[1503]: 2025-02-13 19:12:21.161 [INFO][4849] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.40.70/32] ContainerID="54d08303486449cd8591710ce79b951324a215caf30205c5f317c06267df138e" Namespace="calico-system" Pod="csi-node-driver-6v9pr" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-csi--node--driver--6v9pr-eth0" Feb 13 19:12:21.275809 containerd[1503]: 2025-02-13 19:12:21.163 [INFO][4849] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali05f501d3577 ContainerID="54d08303486449cd8591710ce79b951324a215caf30205c5f317c06267df138e" Namespace="calico-system" Pod="csi-node-driver-6v9pr" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-csi--node--driver--6v9pr-eth0" Feb 13 19:12:21.275809 containerd[1503]: 2025-02-13 19:12:21.197 [INFO][4849] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="54d08303486449cd8591710ce79b951324a215caf30205c5f317c06267df138e" Namespace="calico-system" Pod="csi-node-driver-6v9pr" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-csi--node--driver--6v9pr-eth0" Feb 13 19:12:21.275809 containerd[1503]: 2025-02-13 19:12:21.202 [INFO][4849] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="54d08303486449cd8591710ce79b951324a215caf30205c5f317c06267df138e" Namespace="calico-system" Pod="csi-node-driver-6v9pr" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-csi--node--driver--6v9pr-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4230--0--1--a--2d796e8b97-k8s-csi--node--driver--6v9pr-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"b9c8452b-31d9-47bc-9697-dd4f293d71b9", ResourceVersion:"634", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 19, 12, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"84cddb44f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4230-0-1-a-2d796e8b97", ContainerID:"54d08303486449cd8591710ce79b951324a215caf30205c5f317c06267df138e", Pod:"csi-node-driver-6v9pr", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.40.70/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali05f501d3577", MAC:"ca:0c:ac:cd:6c:8e", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 19:12:21.275809 containerd[1503]: 2025-02-13 19:12:21.258 [INFO][4849] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="54d08303486449cd8591710ce79b951324a215caf30205c5f317c06267df138e" Namespace="calico-system" Pod="csi-node-driver-6v9pr" WorkloadEndpoint="ci--4230--0--1--a--2d796e8b97-k8s-csi--node--driver--6v9pr-eth0" Feb 13 19:12:21.319489 containerd[1503]: time="2025-02-13T19:12:21.318418752Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:12:21.319489 containerd[1503]: time="2025-02-13T19:12:21.318500874Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:12:21.319489 containerd[1503]: time="2025-02-13T19:12:21.318517395Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:12:21.319489 containerd[1503]: time="2025-02-13T19:12:21.318639039Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:12:21.343371 systemd[1]: Started cri-containerd-f2541446bafb170e13b958b7bea1fd0f66a6984abb6abdb77c88a352ed058c69.scope - libcontainer container f2541446bafb170e13b958b7bea1fd0f66a6984abb6abdb77c88a352ed058c69. Feb 13 19:12:21.355225 systemd[1]: Started cri-containerd-45225a5d0a993e64267075f2bbdc29d9ae95a25629a61da4ce1da8b6ff614881.scope - libcontainer container 45225a5d0a993e64267075f2bbdc29d9ae95a25629a61da4ce1da8b6ff614881. Feb 13 19:12:21.357712 systemd[1]: Started cri-containerd-54d08303486449cd8591710ce79b951324a215caf30205c5f317c06267df138e.scope - libcontainer container 54d08303486449cd8591710ce79b951324a215caf30205c5f317c06267df138e. Feb 13 19:12:21.457370 containerd[1503]: time="2025-02-13T19:12:21.456957576Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-vx9cw,Uid:eb0c3c87-077a-4b2b-a168-5fd9131bccd3,Namespace:kube-system,Attempt:6,} returns sandbox id \"45225a5d0a993e64267075f2bbdc29d9ae95a25629a61da4ce1da8b6ff614881\"" Feb 13 19:12:21.466383 containerd[1503]: time="2025-02-13T19:12:21.466224266Z" level=info msg="CreateContainer within sandbox \"45225a5d0a993e64267075f2bbdc29d9ae95a25629a61da4ce1da8b6ff614881\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Feb 13 19:12:21.495992 containerd[1503]: time="2025-02-13T19:12:21.495492864Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-6v9pr,Uid:b9c8452b-31d9-47bc-9697-dd4f293d71b9,Namespace:calico-system,Attempt:6,} returns sandbox id \"54d08303486449cd8591710ce79b951324a215caf30205c5f317c06267df138e\"" Feb 13 19:12:21.517390 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1641137175.mount: Deactivated successfully. Feb 13 19:12:21.530364 containerd[1503]: time="2025-02-13T19:12:21.530318356Z" level=info msg="CreateContainer within sandbox \"45225a5d0a993e64267075f2bbdc29d9ae95a25629a61da4ce1da8b6ff614881\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"39226d9ade63afbac281a0a36d24cb907365064c790120785a6efba9f0ec5cfd\"" Feb 13 19:12:21.532712 containerd[1503]: time="2025-02-13T19:12:21.532669670Z" level=info msg="StartContainer for \"39226d9ade63afbac281a0a36d24cb907365064c790120785a6efba9f0ec5cfd\"" Feb 13 19:12:21.611341 systemd[1]: Started cri-containerd-39226d9ade63afbac281a0a36d24cb907365064c790120785a6efba9f0ec5cfd.scope - libcontainer container 39226d9ade63afbac281a0a36d24cb907365064c790120785a6efba9f0ec5cfd. Feb 13 19:12:21.632041 containerd[1503]: time="2025-02-13T19:12:21.631642253Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-84956f4cf-p6mr2,Uid:05e39f6f-0a83-4b62-8cff-b0d4799f8f89,Namespace:calico-apiserver,Attempt:6,} returns sandbox id \"f2541446bafb170e13b958b7bea1fd0f66a6984abb6abdb77c88a352ed058c69\"" Feb 13 19:12:21.683101 containerd[1503]: time="2025-02-13T19:12:21.682762256Z" level=info msg="StartContainer for \"39226d9ade63afbac281a0a36d24cb907365064c790120785a6efba9f0ec5cfd\" returns successfully" Feb 13 19:12:21.933201 kernel: bpftool[5438]: memfd_create() called without MFD_EXEC or MFD_NOEXEC_SEAL set Feb 13 19:12:22.062409 systemd-networkd[1399]: calicc1bce0eade: Gained IPv6LL Feb 13 19:12:22.143829 systemd-networkd[1399]: vxlan.calico: Link UP Feb 13 19:12:22.143837 systemd-networkd[1399]: vxlan.calico: Gained carrier Feb 13 19:12:22.312450 kubelet[2828]: I0213 19:12:22.311570 2828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-kkqg5" podStartSLOduration=41.311548155 podStartE2EDuration="41.311548155s" podCreationTimestamp="2025-02-13 19:11:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 19:12:21.296720192 +0000 UTC m=+46.995171870" watchObservedRunningTime="2025-02-13 19:12:22.311548155 +0000 UTC m=+48.009999833" Feb 13 19:12:22.334161 kubelet[2828]: I0213 19:12:22.334090 2828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-vx9cw" podStartSLOduration=41.333798657 podStartE2EDuration="41.333798657s" podCreationTimestamp="2025-02-13 19:11:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 19:12:22.311820164 +0000 UTC m=+48.010271842" watchObservedRunningTime="2025-02-13 19:12:22.333798657 +0000 UTC m=+48.032250335" Feb 13 19:12:22.637463 systemd-networkd[1399]: caliac205cc4341: Gained IPv6LL Feb 13 19:12:22.637871 systemd-networkd[1399]: cali4d5c8ae055c: Gained IPv6LL Feb 13 19:12:22.829593 systemd-networkd[1399]: calicde3fabc286: Gained IPv6LL Feb 13 19:12:22.831265 systemd-networkd[1399]: calic2e7ce92ea9: Gained IPv6LL Feb 13 19:12:23.213462 systemd-networkd[1399]: cali05f501d3577: Gained IPv6LL Feb 13 19:12:23.875848 containerd[1503]: time="2025-02-13T19:12:23.875737770Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:12:23.877357 containerd[1503]: time="2025-02-13T19:12:23.877314020Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.29.1: active requests=0, bytes read=39298409" Feb 13 19:12:23.877658 containerd[1503]: time="2025-02-13T19:12:23.877615470Z" level=info msg="ImageCreate event name:\"sha256:5451b31bd8d0784796fa1204c4ec22975a270e21feadf2c5095fe41a38524c6c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:12:23.881672 containerd[1503]: time="2025-02-13T19:12:23.881590996Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:b8c43e264fe52e0c327b0bf3ac882a0224b33bdd7f4ff58a74242da7d9b00486\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:12:23.882899 containerd[1503]: time="2025-02-13T19:12:23.882243577Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.29.1\" with image id \"sha256:5451b31bd8d0784796fa1204c4ec22975a270e21feadf2c5095fe41a38524c6c\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:b8c43e264fe52e0c327b0bf3ac882a0224b33bdd7f4ff58a74242da7d9b00486\", size \"40668079\" in 2.891291611s" Feb 13 19:12:23.882899 containerd[1503]: time="2025-02-13T19:12:23.882278938Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.1\" returns image reference \"sha256:5451b31bd8d0784796fa1204c4ec22975a270e21feadf2c5095fe41a38524c6c\"" Feb 13 19:12:23.892085 containerd[1503]: time="2025-02-13T19:12:23.891620595Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.29.1\"" Feb 13 19:12:23.900452 containerd[1503]: time="2025-02-13T19:12:23.900301310Z" level=info msg="CreateContainer within sandbox \"d366cd2cc4c7c16e754d70bbe9025807e19a94be019d47fbac45a766f5cecf79\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Feb 13 19:12:23.917970 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount584831258.mount: Deactivated successfully. Feb 13 19:12:23.919920 containerd[1503]: time="2025-02-13T19:12:23.919577202Z" level=info msg="CreateContainer within sandbox \"d366cd2cc4c7c16e754d70bbe9025807e19a94be019d47fbac45a766f5cecf79\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"c2f3d36c369254779ff9bfa51f9d620c903d7be7339233302fc66308da26596a\"" Feb 13 19:12:23.922085 containerd[1503]: time="2025-02-13T19:12:23.921993759Z" level=info msg="StartContainer for \"c2f3d36c369254779ff9bfa51f9d620c903d7be7339233302fc66308da26596a\"" Feb 13 19:12:23.960584 systemd[1]: Started cri-containerd-c2f3d36c369254779ff9bfa51f9d620c903d7be7339233302fc66308da26596a.scope - libcontainer container c2f3d36c369254779ff9bfa51f9d620c903d7be7339233302fc66308da26596a. Feb 13 19:12:24.016607 containerd[1503]: time="2025-02-13T19:12:24.015848462Z" level=info msg="StartContainer for \"c2f3d36c369254779ff9bfa51f9d620c903d7be7339233302fc66308da26596a\" returns successfully" Feb 13 19:12:24.109549 systemd-networkd[1399]: vxlan.calico: Gained IPv6LL Feb 13 19:12:24.369382 systemd[1]: Started sshd@14-78.46.147.231:22-117.247.111.70:18477.service - OpenSSH per-connection server daemon (117.247.111.70:18477). Feb 13 19:12:25.320349 kubelet[2828]: I0213 19:12:25.320238 2828 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 13 19:12:25.371082 sshd[5571]: Invalid user aguie from 117.247.111.70 port 18477 Feb 13 19:12:25.562087 sshd[5571]: Received disconnect from 117.247.111.70 port 18477:11: Bye Bye [preauth] Feb 13 19:12:25.562087 sshd[5571]: Disconnected from invalid user aguie 117.247.111.70 port 18477 [preauth] Feb 13 19:12:25.565132 systemd[1]: sshd@14-78.46.147.231:22-117.247.111.70:18477.service: Deactivated successfully. Feb 13 19:12:25.768866 containerd[1503]: time="2025-02-13T19:12:25.767611725Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:12:25.768866 containerd[1503]: time="2025-02-13T19:12:25.768805643Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.29.1: active requests=0, bytes read=31953828" Feb 13 19:12:25.769435 containerd[1503]: time="2025-02-13T19:12:25.769269298Z" level=info msg="ImageCreate event name:\"sha256:32c335fdb9d757e7ba6a76a9cfa8d292a5a229101ae7ea37b42f53c28adf2db1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:12:25.771629 containerd[1503]: time="2025-02-13T19:12:25.771540571Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:1072d6a98167a14ca361e9ce757733f9bae36d1f1c6a9621ea10934b6b1e10d9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:12:25.773103 containerd[1503]: time="2025-02-13T19:12:25.773027659Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.29.1\" with image id \"sha256:32c335fdb9d757e7ba6a76a9cfa8d292a5a229101ae7ea37b42f53c28adf2db1\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:1072d6a98167a14ca361e9ce757733f9bae36d1f1c6a9621ea10934b6b1e10d9\", size \"33323450\" in 1.881339383s" Feb 13 19:12:25.773388 containerd[1503]: time="2025-02-13T19:12:25.773360190Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.29.1\" returns image reference \"sha256:32c335fdb9d757e7ba6a76a9cfa8d292a5a229101ae7ea37b42f53c28adf2db1\"" Feb 13 19:12:25.775161 containerd[1503]: time="2025-02-13T19:12:25.775107686Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.1\"" Feb 13 19:12:25.797322 containerd[1503]: time="2025-02-13T19:12:25.797285959Z" level=info msg="CreateContainer within sandbox \"9b637346f8653af89f0cb9d44b0723cb41c71875e3fcf79734935232fb4eb95b\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Feb 13 19:12:25.824674 containerd[1503]: time="2025-02-13T19:12:25.824608356Z" level=info msg="CreateContainer within sandbox \"9b637346f8653af89f0cb9d44b0723cb41c71875e3fcf79734935232fb4eb95b\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"2f4e0a0e052dfdd1d169970003434489eb5dfea1c524dfeb21d458ec6e8d5d49\"" Feb 13 19:12:25.826782 containerd[1503]: time="2025-02-13T19:12:25.826564619Z" level=info msg="StartContainer for \"2f4e0a0e052dfdd1d169970003434489eb5dfea1c524dfeb21d458ec6e8d5d49\"" Feb 13 19:12:25.864336 systemd[1]: Started cri-containerd-2f4e0a0e052dfdd1d169970003434489eb5dfea1c524dfeb21d458ec6e8d5d49.scope - libcontainer container 2f4e0a0e052dfdd1d169970003434489eb5dfea1c524dfeb21d458ec6e8d5d49. Feb 13 19:12:25.912166 containerd[1503]: time="2025-02-13T19:12:25.911884041Z" level=info msg="StartContainer for \"2f4e0a0e052dfdd1d169970003434489eb5dfea1c524dfeb21d458ec6e8d5d49\" returns successfully" Feb 13 19:12:26.355801 kubelet[2828]: I0213 19:12:26.355467 2828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-78bd6d67c9-b259q" podStartSLOduration=16.795877717 podStartE2EDuration="21.355447756s" podCreationTimestamp="2025-02-13 19:12:05 +0000 UTC" firstStartedPulling="2025-02-13 19:12:21.215210396 +0000 UTC m=+46.913662074" lastFinishedPulling="2025-02-13 19:12:25.774780355 +0000 UTC m=+51.473232113" observedRunningTime="2025-02-13 19:12:26.355178787 +0000 UTC m=+52.053630505" watchObservedRunningTime="2025-02-13 19:12:26.355447756 +0000 UTC m=+52.053899474" Feb 13 19:12:26.357191 kubelet[2828]: I0213 19:12:26.356415 2828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-84956f4cf-72cgp" podStartSLOduration=18.455040617 podStartE2EDuration="21.356401907s" podCreationTimestamp="2025-02-13 19:12:05 +0000 UTC" firstStartedPulling="2025-02-13 19:12:20.989918534 +0000 UTC m=+46.688370212" lastFinishedPulling="2025-02-13 19:12:23.891279824 +0000 UTC m=+49.589731502" observedRunningTime="2025-02-13 19:12:24.329184352 +0000 UTC m=+50.027636070" watchObservedRunningTime="2025-02-13 19:12:26.356401907 +0000 UTC m=+52.054853585" Feb 13 19:12:26.784215 systemd[1]: run-containerd-runc-k8s.io-2f4e0a0e052dfdd1d169970003434489eb5dfea1c524dfeb21d458ec6e8d5d49-runc.ZpwNWq.mount: Deactivated successfully. Feb 13 19:12:27.248465 containerd[1503]: time="2025-02-13T19:12:27.248323567Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:12:27.250090 containerd[1503]: time="2025-02-13T19:12:27.249546767Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.29.1: active requests=0, bytes read=7464730" Feb 13 19:12:27.250769 containerd[1503]: time="2025-02-13T19:12:27.250723045Z" level=info msg="ImageCreate event name:\"sha256:3c11734f3001b7070e7e2b5e64938f89891cf8c44f8997e86aa23c5d5bf70163\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:12:27.254120 containerd[1503]: time="2025-02-13T19:12:27.254043113Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:eaa7e01fb16b603c155a67b81f16992281db7f831684c7b2081d3434587a7ff3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:12:27.255250 containerd[1503]: time="2025-02-13T19:12:27.255217271Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.29.1\" with image id \"sha256:3c11734f3001b7070e7e2b5e64938f89891cf8c44f8997e86aa23c5d5bf70163\", repo tag \"ghcr.io/flatcar/calico/csi:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:eaa7e01fb16b603c155a67b81f16992281db7f831684c7b2081d3434587a7ff3\", size \"8834384\" in 1.480066743s" Feb 13 19:12:27.255303 containerd[1503]: time="2025-02-13T19:12:27.255257392Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.1\" returns image reference \"sha256:3c11734f3001b7070e7e2b5e64938f89891cf8c44f8997e86aa23c5d5bf70163\"" Feb 13 19:12:27.258903 containerd[1503]: time="2025-02-13T19:12:27.257480904Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.1\"" Feb 13 19:12:27.269448 containerd[1503]: time="2025-02-13T19:12:27.269383771Z" level=info msg="CreateContainer within sandbox \"54d08303486449cd8591710ce79b951324a215caf30205c5f317c06267df138e\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Feb 13 19:12:27.294950 containerd[1503]: time="2025-02-13T19:12:27.294823357Z" level=info msg="CreateContainer within sandbox \"54d08303486449cd8591710ce79b951324a215caf30205c5f317c06267df138e\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"0eb5e6ba84794c0b861066aae225ea2067ce731a1c51c28b2e4b0045fb829077\"" Feb 13 19:12:27.295768 containerd[1503]: time="2025-02-13T19:12:27.295742547Z" level=info msg="StartContainer for \"0eb5e6ba84794c0b861066aae225ea2067ce731a1c51c28b2e4b0045fb829077\"" Feb 13 19:12:27.334427 systemd[1]: Started cri-containerd-0eb5e6ba84794c0b861066aae225ea2067ce731a1c51c28b2e4b0045fb829077.scope - libcontainer container 0eb5e6ba84794c0b861066aae225ea2067ce731a1c51c28b2e4b0045fb829077. Feb 13 19:12:27.376947 containerd[1503]: time="2025-02-13T19:12:27.375972513Z" level=info msg="StartContainer for \"0eb5e6ba84794c0b861066aae225ea2067ce731a1c51c28b2e4b0045fb829077\" returns successfully" Feb 13 19:12:27.654359 containerd[1503]: time="2025-02-13T19:12:27.654296954Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:12:27.655124 containerd[1503]: time="2025-02-13T19:12:27.655050218Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.29.1: active requests=0, bytes read=77" Feb 13 19:12:27.658930 containerd[1503]: time="2025-02-13T19:12:27.658311724Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.29.1\" with image id \"sha256:5451b31bd8d0784796fa1204c4ec22975a270e21feadf2c5095fe41a38524c6c\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:b8c43e264fe52e0c327b0bf3ac882a0224b33bdd7f4ff58a74242da7d9b00486\", size \"40668079\" in 400.790339ms" Feb 13 19:12:27.658930 containerd[1503]: time="2025-02-13T19:12:27.658407607Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.1\" returns image reference \"sha256:5451b31bd8d0784796fa1204c4ec22975a270e21feadf2c5095fe41a38524c6c\"" Feb 13 19:12:27.659937 containerd[1503]: time="2025-02-13T19:12:27.659884535Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\"" Feb 13 19:12:27.664896 containerd[1503]: time="2025-02-13T19:12:27.664815816Z" level=info msg="CreateContainer within sandbox \"f2541446bafb170e13b958b7bea1fd0f66a6984abb6abdb77c88a352ed058c69\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Feb 13 19:12:27.683035 containerd[1503]: time="2025-02-13T19:12:27.682990126Z" level=info msg="CreateContainer within sandbox \"f2541446bafb170e13b958b7bea1fd0f66a6984abb6abdb77c88a352ed058c69\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"f004fb42a78b1f3145f561f918b66ce7b875d89943fd4bbb09f3e69cc82749ef\"" Feb 13 19:12:27.684482 containerd[1503]: time="2025-02-13T19:12:27.684451213Z" level=info msg="StartContainer for \"f004fb42a78b1f3145f561f918b66ce7b875d89943fd4bbb09f3e69cc82749ef\"" Feb 13 19:12:27.733522 systemd[1]: Started cri-containerd-f004fb42a78b1f3145f561f918b66ce7b875d89943fd4bbb09f3e69cc82749ef.scope - libcontainer container f004fb42a78b1f3145f561f918b66ce7b875d89943fd4bbb09f3e69cc82749ef. Feb 13 19:12:27.775002 containerd[1503]: time="2025-02-13T19:12:27.774942673Z" level=info msg="StartContainer for \"f004fb42a78b1f3145f561f918b66ce7b875d89943fd4bbb09f3e69cc82749ef\" returns successfully" Feb 13 19:12:28.384604 kubelet[2828]: I0213 19:12:28.384530 2828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-84956f4cf-p6mr2" podStartSLOduration=17.360514788 podStartE2EDuration="23.384512177s" podCreationTimestamp="2025-02-13 19:12:05 +0000 UTC" firstStartedPulling="2025-02-13 19:12:21.635719141 +0000 UTC m=+47.334170819" lastFinishedPulling="2025-02-13 19:12:27.65971653 +0000 UTC m=+53.358168208" observedRunningTime="2025-02-13 19:12:28.383873676 +0000 UTC m=+54.082325354" watchObservedRunningTime="2025-02-13 19:12:28.384512177 +0000 UTC m=+54.082963815" Feb 13 19:12:29.274645 containerd[1503]: time="2025-02-13T19:12:29.274584402Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:12:29.276790 containerd[1503]: time="2025-02-13T19:12:29.276742393Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1: active requests=0, bytes read=9883368" Feb 13 19:12:29.277624 containerd[1503]: time="2025-02-13T19:12:29.277595101Z" level=info msg="ImageCreate event name:\"sha256:3eb557f7694f230afd24a75a691bcda4c0a7bfe87a981386dcd4ecf2b0701349\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:12:29.280734 containerd[1503]: time="2025-02-13T19:12:29.280675042Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:a338da9488cbaa83c78457c3d7354d84149969c0480e88dd768e036632ff5b76\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:12:29.284134 containerd[1503]: time="2025-02-13T19:12:29.283923149Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\" with image id \"sha256:3eb557f7694f230afd24a75a691bcda4c0a7bfe87a981386dcd4ecf2b0701349\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:a338da9488cbaa83c78457c3d7354d84149969c0480e88dd768e036632ff5b76\", size \"11252974\" in 1.623980011s" Feb 13 19:12:29.284134 containerd[1503]: time="2025-02-13T19:12:29.284003551Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\" returns image reference \"sha256:3eb557f7694f230afd24a75a691bcda4c0a7bfe87a981386dcd4ecf2b0701349\"" Feb 13 19:12:29.293150 containerd[1503]: time="2025-02-13T19:12:29.293003567Z" level=info msg="CreateContainer within sandbox \"54d08303486449cd8591710ce79b951324a215caf30205c5f317c06267df138e\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Feb 13 19:12:29.317605 containerd[1503]: time="2025-02-13T19:12:29.317539332Z" level=info msg="CreateContainer within sandbox \"54d08303486449cd8591710ce79b951324a215caf30205c5f317c06267df138e\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"c0029ed1ead687fbf0a5c6846e6d05b9dee1ba5cec334e02c9e176a3528b2df5\"" Feb 13 19:12:29.318447 containerd[1503]: time="2025-02-13T19:12:29.318380679Z" level=info msg="StartContainer for \"c0029ed1ead687fbf0a5c6846e6d05b9dee1ba5cec334e02c9e176a3528b2df5\"" Feb 13 19:12:29.359531 systemd[1]: Started cri-containerd-c0029ed1ead687fbf0a5c6846e6d05b9dee1ba5cec334e02c9e176a3528b2df5.scope - libcontainer container c0029ed1ead687fbf0a5c6846e6d05b9dee1ba5cec334e02c9e176a3528b2df5. Feb 13 19:12:29.386706 kubelet[2828]: I0213 19:12:29.386649 2828 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 13 19:12:29.411387 containerd[1503]: time="2025-02-13T19:12:29.411179604Z" level=info msg="StartContainer for \"c0029ed1ead687fbf0a5c6846e6d05b9dee1ba5cec334e02c9e176a3528b2df5\" returns successfully" Feb 13 19:12:29.575084 kubelet[2828]: I0213 19:12:29.574914 2828 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Feb 13 19:12:29.579246 kubelet[2828]: I0213 19:12:29.579189 2828 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Feb 13 19:12:30.417821 kubelet[2828]: I0213 19:12:30.417751 2828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-6v9pr" podStartSLOduration=17.628725986 podStartE2EDuration="25.417391646s" podCreationTimestamp="2025-02-13 19:12:05 +0000 UTC" firstStartedPulling="2025-02-13 19:12:21.500367297 +0000 UTC m=+47.198818975" lastFinishedPulling="2025-02-13 19:12:29.289032957 +0000 UTC m=+54.987484635" observedRunningTime="2025-02-13 19:12:30.41477272 +0000 UTC m=+56.113224398" watchObservedRunningTime="2025-02-13 19:12:30.417391646 +0000 UTC m=+56.115843324" Feb 13 19:12:32.951571 systemd[1]: Started sshd@15-78.46.147.231:22-103.154.87.242:44580.service - OpenSSH per-connection server daemon (103.154.87.242:44580). Feb 13 19:12:34.044625 sshd[5774]: Invalid user hasin from 103.154.87.242 port 44580 Feb 13 19:12:34.273502 sshd[5774]: Received disconnect from 103.154.87.242 port 44580:11: Bye Bye [preauth] Feb 13 19:12:34.273502 sshd[5774]: Disconnected from invalid user hasin 103.154.87.242 port 44580 [preauth] Feb 13 19:12:34.276821 systemd[1]: sshd@15-78.46.147.231:22-103.154.87.242:44580.service: Deactivated successfully. Feb 13 19:12:34.462418 containerd[1503]: time="2025-02-13T19:12:34.461557668Z" level=info msg="StopPodSandbox for \"0c13c2a989169406cb4cea286d09a0b055b98c8bf7adaf3c864364728278a3a5\"" Feb 13 19:12:34.462418 containerd[1503]: time="2025-02-13T19:12:34.461738234Z" level=info msg="TearDown network for sandbox \"0c13c2a989169406cb4cea286d09a0b055b98c8bf7adaf3c864364728278a3a5\" successfully" Feb 13 19:12:34.462418 containerd[1503]: time="2025-02-13T19:12:34.461756154Z" level=info msg="StopPodSandbox for \"0c13c2a989169406cb4cea286d09a0b055b98c8bf7adaf3c864364728278a3a5\" returns successfully" Feb 13 19:12:34.463637 containerd[1503]: time="2025-02-13T19:12:34.463579495Z" level=info msg="RemovePodSandbox for \"0c13c2a989169406cb4cea286d09a0b055b98c8bf7adaf3c864364728278a3a5\"" Feb 13 19:12:34.463749 containerd[1503]: time="2025-02-13T19:12:34.463648978Z" level=info msg="Forcibly stopping sandbox \"0c13c2a989169406cb4cea286d09a0b055b98c8bf7adaf3c864364728278a3a5\"" Feb 13 19:12:34.463812 containerd[1503]: time="2025-02-13T19:12:34.463785942Z" level=info msg="TearDown network for sandbox \"0c13c2a989169406cb4cea286d09a0b055b98c8bf7adaf3c864364728278a3a5\" successfully" Feb 13 19:12:34.468284 containerd[1503]: time="2025-02-13T19:12:34.467921281Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"0c13c2a989169406cb4cea286d09a0b055b98c8bf7adaf3c864364728278a3a5\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:12:34.468433 containerd[1503]: time="2025-02-13T19:12:34.468347335Z" level=info msg="RemovePodSandbox \"0c13c2a989169406cb4cea286d09a0b055b98c8bf7adaf3c864364728278a3a5\" returns successfully" Feb 13 19:12:34.469345 containerd[1503]: time="2025-02-13T19:12:34.469314768Z" level=info msg="StopPodSandbox for \"9581f7ff6fcd5f0bba90e1d0d2984cd0efa6fe4d50a06a4c1b20f3939cb09713\"" Feb 13 19:12:34.469449 containerd[1503]: time="2025-02-13T19:12:34.469430412Z" level=info msg="TearDown network for sandbox \"9581f7ff6fcd5f0bba90e1d0d2984cd0efa6fe4d50a06a4c1b20f3939cb09713\" successfully" Feb 13 19:12:34.469512 containerd[1503]: time="2025-02-13T19:12:34.469448572Z" level=info msg="StopPodSandbox for \"9581f7ff6fcd5f0bba90e1d0d2984cd0efa6fe4d50a06a4c1b20f3939cb09713\" returns successfully" Feb 13 19:12:34.469892 containerd[1503]: time="2025-02-13T19:12:34.469869266Z" level=info msg="RemovePodSandbox for \"9581f7ff6fcd5f0bba90e1d0d2984cd0efa6fe4d50a06a4c1b20f3939cb09713\"" Feb 13 19:12:34.469938 containerd[1503]: time="2025-02-13T19:12:34.469895867Z" level=info msg="Forcibly stopping sandbox \"9581f7ff6fcd5f0bba90e1d0d2984cd0efa6fe4d50a06a4c1b20f3939cb09713\"" Feb 13 19:12:34.470009 containerd[1503]: time="2025-02-13T19:12:34.469964270Z" level=info msg="TearDown network for sandbox \"9581f7ff6fcd5f0bba90e1d0d2984cd0efa6fe4d50a06a4c1b20f3939cb09713\" successfully" Feb 13 19:12:34.473868 containerd[1503]: time="2025-02-13T19:12:34.473802358Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"9581f7ff6fcd5f0bba90e1d0d2984cd0efa6fe4d50a06a4c1b20f3939cb09713\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:12:34.474023 containerd[1503]: time="2025-02-13T19:12:34.473896762Z" level=info msg="RemovePodSandbox \"9581f7ff6fcd5f0bba90e1d0d2984cd0efa6fe4d50a06a4c1b20f3939cb09713\" returns successfully" Feb 13 19:12:34.474413 containerd[1503]: time="2025-02-13T19:12:34.474386178Z" level=info msg="StopPodSandbox for \"ab072ffc918efc3594006a69ddbbd17de37178f95f1103e4fa6b31bb95fd9306\"" Feb 13 19:12:34.474516 containerd[1503]: time="2025-02-13T19:12:34.474498142Z" level=info msg="TearDown network for sandbox \"ab072ffc918efc3594006a69ddbbd17de37178f95f1103e4fa6b31bb95fd9306\" successfully" Feb 13 19:12:34.474516 containerd[1503]: time="2025-02-13T19:12:34.474513542Z" level=info msg="StopPodSandbox for \"ab072ffc918efc3594006a69ddbbd17de37178f95f1103e4fa6b31bb95fd9306\" returns successfully" Feb 13 19:12:34.474911 containerd[1503]: time="2025-02-13T19:12:34.474844153Z" level=info msg="RemovePodSandbox for \"ab072ffc918efc3594006a69ddbbd17de37178f95f1103e4fa6b31bb95fd9306\"" Feb 13 19:12:34.474973 containerd[1503]: time="2025-02-13T19:12:34.474916916Z" level=info msg="Forcibly stopping sandbox \"ab072ffc918efc3594006a69ddbbd17de37178f95f1103e4fa6b31bb95fd9306\"" Feb 13 19:12:34.475115 containerd[1503]: time="2025-02-13T19:12:34.475093682Z" level=info msg="TearDown network for sandbox \"ab072ffc918efc3594006a69ddbbd17de37178f95f1103e4fa6b31bb95fd9306\" successfully" Feb 13 19:12:34.479518 containerd[1503]: time="2025-02-13T19:12:34.479455668Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"ab072ffc918efc3594006a69ddbbd17de37178f95f1103e4fa6b31bb95fd9306\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:12:34.479655 containerd[1503]: time="2025-02-13T19:12:34.479534151Z" level=info msg="RemovePodSandbox \"ab072ffc918efc3594006a69ddbbd17de37178f95f1103e4fa6b31bb95fd9306\" returns successfully" Feb 13 19:12:34.481661 containerd[1503]: time="2025-02-13T19:12:34.480948438Z" level=info msg="StopPodSandbox for \"96f8dbfc99d905b595d74b0f3390d27d4b0e99b3143e0cc72edd14238522b44f\"" Feb 13 19:12:34.481661 containerd[1503]: time="2025-02-13T19:12:34.481303650Z" level=info msg="TearDown network for sandbox \"96f8dbfc99d905b595d74b0f3390d27d4b0e99b3143e0cc72edd14238522b44f\" successfully" Feb 13 19:12:34.481661 containerd[1503]: time="2025-02-13T19:12:34.481327051Z" level=info msg="StopPodSandbox for \"96f8dbfc99d905b595d74b0f3390d27d4b0e99b3143e0cc72edd14238522b44f\" returns successfully" Feb 13 19:12:34.484328 containerd[1503]: time="2025-02-13T19:12:34.484280910Z" level=info msg="RemovePodSandbox for \"96f8dbfc99d905b595d74b0f3390d27d4b0e99b3143e0cc72edd14238522b44f\"" Feb 13 19:12:34.484328 containerd[1503]: time="2025-02-13T19:12:34.484320551Z" level=info msg="Forcibly stopping sandbox \"96f8dbfc99d905b595d74b0f3390d27d4b0e99b3143e0cc72edd14238522b44f\"" Feb 13 19:12:34.484495 containerd[1503]: time="2025-02-13T19:12:34.484422595Z" level=info msg="TearDown network for sandbox \"96f8dbfc99d905b595d74b0f3390d27d4b0e99b3143e0cc72edd14238522b44f\" successfully" Feb 13 19:12:34.487725 containerd[1503]: time="2025-02-13T19:12:34.487660703Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"96f8dbfc99d905b595d74b0f3390d27d4b0e99b3143e0cc72edd14238522b44f\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:12:34.487843 containerd[1503]: time="2025-02-13T19:12:34.487768387Z" level=info msg="RemovePodSandbox \"96f8dbfc99d905b595d74b0f3390d27d4b0e99b3143e0cc72edd14238522b44f\" returns successfully" Feb 13 19:12:34.489098 containerd[1503]: time="2025-02-13T19:12:34.488464090Z" level=info msg="StopPodSandbox for \"b3134abe6d3978218dbaec771f16517d173324bf774a94fc2b6f62660051520d\"" Feb 13 19:12:34.489098 containerd[1503]: time="2025-02-13T19:12:34.488658937Z" level=info msg="TearDown network for sandbox \"b3134abe6d3978218dbaec771f16517d173324bf774a94fc2b6f62660051520d\" successfully" Feb 13 19:12:34.489098 containerd[1503]: time="2025-02-13T19:12:34.488677417Z" level=info msg="StopPodSandbox for \"b3134abe6d3978218dbaec771f16517d173324bf774a94fc2b6f62660051520d\" returns successfully" Feb 13 19:12:34.489098 containerd[1503]: time="2025-02-13T19:12:34.489092031Z" level=info msg="RemovePodSandbox for \"b3134abe6d3978218dbaec771f16517d173324bf774a94fc2b6f62660051520d\"" Feb 13 19:12:34.489271 containerd[1503]: time="2025-02-13T19:12:34.489124152Z" level=info msg="Forcibly stopping sandbox \"b3134abe6d3978218dbaec771f16517d173324bf774a94fc2b6f62660051520d\"" Feb 13 19:12:34.489271 containerd[1503]: time="2025-02-13T19:12:34.489218956Z" level=info msg="TearDown network for sandbox \"b3134abe6d3978218dbaec771f16517d173324bf774a94fc2b6f62660051520d\" successfully" Feb 13 19:12:34.494548 containerd[1503]: time="2025-02-13T19:12:34.494495293Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"b3134abe6d3978218dbaec771f16517d173324bf774a94fc2b6f62660051520d\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:12:34.494548 containerd[1503]: time="2025-02-13T19:12:34.494560855Z" level=info msg="RemovePodSandbox \"b3134abe6d3978218dbaec771f16517d173324bf774a94fc2b6f62660051520d\" returns successfully" Feb 13 19:12:34.495722 containerd[1503]: time="2025-02-13T19:12:34.495697813Z" level=info msg="StopPodSandbox for \"15ac125800531efb9e148c216d3f053334f90d5dd6675f1df06b0c64e6f022c5\"" Feb 13 19:12:34.495823 containerd[1503]: time="2025-02-13T19:12:34.495807337Z" level=info msg="TearDown network for sandbox \"15ac125800531efb9e148c216d3f053334f90d5dd6675f1df06b0c64e6f022c5\" successfully" Feb 13 19:12:34.495823 containerd[1503]: time="2025-02-13T19:12:34.495821857Z" level=info msg="StopPodSandbox for \"15ac125800531efb9e148c216d3f053334f90d5dd6675f1df06b0c64e6f022c5\" returns successfully" Feb 13 19:12:34.496201 containerd[1503]: time="2025-02-13T19:12:34.496176829Z" level=info msg="RemovePodSandbox for \"15ac125800531efb9e148c216d3f053334f90d5dd6675f1df06b0c64e6f022c5\"" Feb 13 19:12:34.496201 containerd[1503]: time="2025-02-13T19:12:34.496205110Z" level=info msg="Forcibly stopping sandbox \"15ac125800531efb9e148c216d3f053334f90d5dd6675f1df06b0c64e6f022c5\"" Feb 13 19:12:34.496281 containerd[1503]: time="2025-02-13T19:12:34.496264872Z" level=info msg="TearDown network for sandbox \"15ac125800531efb9e148c216d3f053334f90d5dd6675f1df06b0c64e6f022c5\" successfully" Feb 13 19:12:34.500560 containerd[1503]: time="2025-02-13T19:12:34.500503534Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"15ac125800531efb9e148c216d3f053334f90d5dd6675f1df06b0c64e6f022c5\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:12:34.500777 containerd[1503]: time="2025-02-13T19:12:34.500588217Z" level=info msg="RemovePodSandbox \"15ac125800531efb9e148c216d3f053334f90d5dd6675f1df06b0c64e6f022c5\" returns successfully" Feb 13 19:12:34.502202 containerd[1503]: time="2025-02-13T19:12:34.501479807Z" level=info msg="StopPodSandbox for \"fa6e4b9a2b53f8e5e8150bdfa0fe88ef1d51405b9b5e7e6d9683c6adb0b078ab\"" Feb 13 19:12:34.502202 containerd[1503]: time="2025-02-13T19:12:34.501817258Z" level=info msg="TearDown network for sandbox \"fa6e4b9a2b53f8e5e8150bdfa0fe88ef1d51405b9b5e7e6d9683c6adb0b078ab\" successfully" Feb 13 19:12:34.502202 containerd[1503]: time="2025-02-13T19:12:34.501835419Z" level=info msg="StopPodSandbox for \"fa6e4b9a2b53f8e5e8150bdfa0fe88ef1d51405b9b5e7e6d9683c6adb0b078ab\" returns successfully" Feb 13 19:12:34.502750 containerd[1503]: time="2025-02-13T19:12:34.502267513Z" level=info msg="RemovePodSandbox for \"fa6e4b9a2b53f8e5e8150bdfa0fe88ef1d51405b9b5e7e6d9683c6adb0b078ab\"" Feb 13 19:12:34.502750 containerd[1503]: time="2025-02-13T19:12:34.502294754Z" level=info msg="Forcibly stopping sandbox \"fa6e4b9a2b53f8e5e8150bdfa0fe88ef1d51405b9b5e7e6d9683c6adb0b078ab\"" Feb 13 19:12:34.502750 containerd[1503]: time="2025-02-13T19:12:34.502373037Z" level=info msg="TearDown network for sandbox \"fa6e4b9a2b53f8e5e8150bdfa0fe88ef1d51405b9b5e7e6d9683c6adb0b078ab\" successfully" Feb 13 19:12:34.506778 containerd[1503]: time="2025-02-13T19:12:34.506728263Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"fa6e4b9a2b53f8e5e8150bdfa0fe88ef1d51405b9b5e7e6d9683c6adb0b078ab\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:12:34.506937 containerd[1503]: time="2025-02-13T19:12:34.506808866Z" level=info msg="RemovePodSandbox \"fa6e4b9a2b53f8e5e8150bdfa0fe88ef1d51405b9b5e7e6d9683c6adb0b078ab\" returns successfully" Feb 13 19:12:34.507321 containerd[1503]: time="2025-02-13T19:12:34.507288762Z" level=info msg="StopPodSandbox for \"0be20e2c65924f7b5393bc8b6a79f35d5bf8cca7f34009d6201de4852924939b\"" Feb 13 19:12:34.507484 containerd[1503]: time="2025-02-13T19:12:34.507398406Z" level=info msg="TearDown network for sandbox \"0be20e2c65924f7b5393bc8b6a79f35d5bf8cca7f34009d6201de4852924939b\" successfully" Feb 13 19:12:34.507484 containerd[1503]: time="2025-02-13T19:12:34.507413526Z" level=info msg="StopPodSandbox for \"0be20e2c65924f7b5393bc8b6a79f35d5bf8cca7f34009d6201de4852924939b\" returns successfully" Feb 13 19:12:34.507782 containerd[1503]: time="2025-02-13T19:12:34.507758258Z" level=info msg="RemovePodSandbox for \"0be20e2c65924f7b5393bc8b6a79f35d5bf8cca7f34009d6201de4852924939b\"" Feb 13 19:12:34.507818 containerd[1503]: time="2025-02-13T19:12:34.507783979Z" level=info msg="Forcibly stopping sandbox \"0be20e2c65924f7b5393bc8b6a79f35d5bf8cca7f34009d6201de4852924939b\"" Feb 13 19:12:34.507864 containerd[1503]: time="2025-02-13T19:12:34.507843261Z" level=info msg="TearDown network for sandbox \"0be20e2c65924f7b5393bc8b6a79f35d5bf8cca7f34009d6201de4852924939b\" successfully" Feb 13 19:12:34.512038 containerd[1503]: time="2025-02-13T19:12:34.511937878Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"0be20e2c65924f7b5393bc8b6a79f35d5bf8cca7f34009d6201de4852924939b\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:12:34.512240 containerd[1503]: time="2025-02-13T19:12:34.512111884Z" level=info msg="RemovePodSandbox \"0be20e2c65924f7b5393bc8b6a79f35d5bf8cca7f34009d6201de4852924939b\" returns successfully" Feb 13 19:12:34.512665 containerd[1503]: time="2025-02-13T19:12:34.512629261Z" level=info msg="StopPodSandbox for \"4ffb63334faa7203eea7b2071116973af83410f7340f4e8946194bccb25fa157\"" Feb 13 19:12:34.512804 containerd[1503]: time="2025-02-13T19:12:34.512783386Z" level=info msg="TearDown network for sandbox \"4ffb63334faa7203eea7b2071116973af83410f7340f4e8946194bccb25fa157\" successfully" Feb 13 19:12:34.512843 containerd[1503]: time="2025-02-13T19:12:34.512802507Z" level=info msg="StopPodSandbox for \"4ffb63334faa7203eea7b2071116973af83410f7340f4e8946194bccb25fa157\" returns successfully" Feb 13 19:12:34.513388 containerd[1503]: time="2025-02-13T19:12:34.513358286Z" level=info msg="RemovePodSandbox for \"4ffb63334faa7203eea7b2071116973af83410f7340f4e8946194bccb25fa157\"" Feb 13 19:12:34.513453 containerd[1503]: time="2025-02-13T19:12:34.513389887Z" level=info msg="Forcibly stopping sandbox \"4ffb63334faa7203eea7b2071116973af83410f7340f4e8946194bccb25fa157\"" Feb 13 19:12:34.513537 containerd[1503]: time="2025-02-13T19:12:34.513466009Z" level=info msg="TearDown network for sandbox \"4ffb63334faa7203eea7b2071116973af83410f7340f4e8946194bccb25fa157\" successfully" Feb 13 19:12:34.517818 containerd[1503]: time="2025-02-13T19:12:34.517720472Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"4ffb63334faa7203eea7b2071116973af83410f7340f4e8946194bccb25fa157\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:12:34.518011 containerd[1503]: time="2025-02-13T19:12:34.517831876Z" level=info msg="RemovePodSandbox \"4ffb63334faa7203eea7b2071116973af83410f7340f4e8946194bccb25fa157\" returns successfully" Feb 13 19:12:34.518422 containerd[1503]: time="2025-02-13T19:12:34.518382854Z" level=info msg="StopPodSandbox for \"f1ae97b8cb19d334ba8268033afc16640078623540002cf3f5996a895113dd8b\"" Feb 13 19:12:34.518512 containerd[1503]: time="2025-02-13T19:12:34.518496698Z" level=info msg="TearDown network for sandbox \"f1ae97b8cb19d334ba8268033afc16640078623540002cf3f5996a895113dd8b\" successfully" Feb 13 19:12:34.518564 containerd[1503]: time="2025-02-13T19:12:34.518511098Z" level=info msg="StopPodSandbox for \"f1ae97b8cb19d334ba8268033afc16640078623540002cf3f5996a895113dd8b\" returns successfully" Feb 13 19:12:34.519106 containerd[1503]: time="2025-02-13T19:12:34.518896911Z" level=info msg="RemovePodSandbox for \"f1ae97b8cb19d334ba8268033afc16640078623540002cf3f5996a895113dd8b\"" Feb 13 19:12:34.519106 containerd[1503]: time="2025-02-13T19:12:34.518924912Z" level=info msg="Forcibly stopping sandbox \"f1ae97b8cb19d334ba8268033afc16640078623540002cf3f5996a895113dd8b\"" Feb 13 19:12:34.519106 containerd[1503]: time="2025-02-13T19:12:34.519007315Z" level=info msg="TearDown network for sandbox \"f1ae97b8cb19d334ba8268033afc16640078623540002cf3f5996a895113dd8b\" successfully" Feb 13 19:12:34.523764 containerd[1503]: time="2025-02-13T19:12:34.523719313Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"f1ae97b8cb19d334ba8268033afc16640078623540002cf3f5996a895113dd8b\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:12:34.524528 containerd[1503]: time="2025-02-13T19:12:34.524009083Z" level=info msg="RemovePodSandbox \"f1ae97b8cb19d334ba8268033afc16640078623540002cf3f5996a895113dd8b\" returns successfully" Feb 13 19:12:34.525479 containerd[1503]: time="2025-02-13T19:12:34.525187843Z" level=info msg="StopPodSandbox for \"f101bce244f680aa9fab77bf6802cf73b54ffba4caa118c44f2e7b203fb0dc84\"" Feb 13 19:12:34.525479 containerd[1503]: time="2025-02-13T19:12:34.525285566Z" level=info msg="TearDown network for sandbox \"f101bce244f680aa9fab77bf6802cf73b54ffba4caa118c44f2e7b203fb0dc84\" successfully" Feb 13 19:12:34.525479 containerd[1503]: time="2025-02-13T19:12:34.525294886Z" level=info msg="StopPodSandbox for \"f101bce244f680aa9fab77bf6802cf73b54ffba4caa118c44f2e7b203fb0dc84\" returns successfully" Feb 13 19:12:34.526748 containerd[1503]: time="2025-02-13T19:12:34.526516967Z" level=info msg="RemovePodSandbox for \"f101bce244f680aa9fab77bf6802cf73b54ffba4caa118c44f2e7b203fb0dc84\"" Feb 13 19:12:34.526748 containerd[1503]: time="2025-02-13T19:12:34.526548168Z" level=info msg="Forcibly stopping sandbox \"f101bce244f680aa9fab77bf6802cf73b54ffba4caa118c44f2e7b203fb0dc84\"" Feb 13 19:12:34.526748 containerd[1503]: time="2025-02-13T19:12:34.526632011Z" level=info msg="TearDown network for sandbox \"f101bce244f680aa9fab77bf6802cf73b54ffba4caa118c44f2e7b203fb0dc84\" successfully" Feb 13 19:12:34.532709 containerd[1503]: time="2025-02-13T19:12:34.532639893Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"f101bce244f680aa9fab77bf6802cf73b54ffba4caa118c44f2e7b203fb0dc84\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:12:34.532885 containerd[1503]: time="2025-02-13T19:12:34.532718055Z" level=info msg="RemovePodSandbox \"f101bce244f680aa9fab77bf6802cf73b54ffba4caa118c44f2e7b203fb0dc84\" returns successfully" Feb 13 19:12:34.534091 containerd[1503]: time="2025-02-13T19:12:34.534035179Z" level=info msg="StopPodSandbox for \"6ca053ac29abc2425e4b004e866bb9d2dbabd9431e4a54c02f8b398f38c0da08\"" Feb 13 19:12:34.534515 containerd[1503]: time="2025-02-13T19:12:34.534425752Z" level=info msg="TearDown network for sandbox \"6ca053ac29abc2425e4b004e866bb9d2dbabd9431e4a54c02f8b398f38c0da08\" successfully" Feb 13 19:12:34.534515 containerd[1503]: time="2025-02-13T19:12:34.534448513Z" level=info msg="StopPodSandbox for \"6ca053ac29abc2425e4b004e866bb9d2dbabd9431e4a54c02f8b398f38c0da08\" returns successfully" Feb 13 19:12:34.535242 containerd[1503]: time="2025-02-13T19:12:34.535213579Z" level=info msg="RemovePodSandbox for \"6ca053ac29abc2425e4b004e866bb9d2dbabd9431e4a54c02f8b398f38c0da08\"" Feb 13 19:12:34.535242 containerd[1503]: time="2025-02-13T19:12:34.535245540Z" level=info msg="Forcibly stopping sandbox \"6ca053ac29abc2425e4b004e866bb9d2dbabd9431e4a54c02f8b398f38c0da08\"" Feb 13 19:12:34.535347 containerd[1503]: time="2025-02-13T19:12:34.535314502Z" level=info msg="TearDown network for sandbox \"6ca053ac29abc2425e4b004e866bb9d2dbabd9431e4a54c02f8b398f38c0da08\" successfully" Feb 13 19:12:34.539421 containerd[1503]: time="2025-02-13T19:12:34.539364358Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"6ca053ac29abc2425e4b004e866bb9d2dbabd9431e4a54c02f8b398f38c0da08\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:12:34.539582 containerd[1503]: time="2025-02-13T19:12:34.539468202Z" level=info msg="RemovePodSandbox \"6ca053ac29abc2425e4b004e866bb9d2dbabd9431e4a54c02f8b398f38c0da08\" returns successfully" Feb 13 19:12:34.540612 containerd[1503]: time="2025-02-13T19:12:34.540570559Z" level=info msg="StopPodSandbox for \"43a5af5ee3597e852f8d0ba977ffb06562d6b40688d779a95c8d495a78df4c29\"" Feb 13 19:12:34.540777 containerd[1503]: time="2025-02-13T19:12:34.540745845Z" level=info msg="TearDown network for sandbox \"43a5af5ee3597e852f8d0ba977ffb06562d6b40688d779a95c8d495a78df4c29\" successfully" Feb 13 19:12:34.540828 containerd[1503]: time="2025-02-13T19:12:34.540778006Z" level=info msg="StopPodSandbox for \"43a5af5ee3597e852f8d0ba977ffb06562d6b40688d779a95c8d495a78df4c29\" returns successfully" Feb 13 19:12:34.541672 containerd[1503]: time="2025-02-13T19:12:34.541560272Z" level=info msg="RemovePodSandbox for \"43a5af5ee3597e852f8d0ba977ffb06562d6b40688d779a95c8d495a78df4c29\"" Feb 13 19:12:34.541672 containerd[1503]: time="2025-02-13T19:12:34.541626514Z" level=info msg="Forcibly stopping sandbox \"43a5af5ee3597e852f8d0ba977ffb06562d6b40688d779a95c8d495a78df4c29\"" Feb 13 19:12:34.542050 containerd[1503]: time="2025-02-13T19:12:34.542002247Z" level=info msg="TearDown network for sandbox \"43a5af5ee3597e852f8d0ba977ffb06562d6b40688d779a95c8d495a78df4c29\" successfully" Feb 13 19:12:34.546833 containerd[1503]: time="2025-02-13T19:12:34.546655283Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"43a5af5ee3597e852f8d0ba977ffb06562d6b40688d779a95c8d495a78df4c29\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:12:34.546833 containerd[1503]: time="2025-02-13T19:12:34.546733205Z" level=info msg="RemovePodSandbox \"43a5af5ee3597e852f8d0ba977ffb06562d6b40688d779a95c8d495a78df4c29\" returns successfully" Feb 13 19:12:34.547708 containerd[1503]: time="2025-02-13T19:12:34.547661837Z" level=info msg="StopPodSandbox for \"10ae5b7da745ff3939219060da91f4661d2a1136a80736c3234db5b474e3599d\"" Feb 13 19:12:34.547956 containerd[1503]: time="2025-02-13T19:12:34.547928326Z" level=info msg="TearDown network for sandbox \"10ae5b7da745ff3939219060da91f4661d2a1136a80736c3234db5b474e3599d\" successfully" Feb 13 19:12:34.548046 containerd[1503]: time="2025-02-13T19:12:34.547957526Z" level=info msg="StopPodSandbox for \"10ae5b7da745ff3939219060da91f4661d2a1136a80736c3234db5b474e3599d\" returns successfully" Feb 13 19:12:34.548616 containerd[1503]: time="2025-02-13T19:12:34.548571387Z" level=info msg="RemovePodSandbox for \"10ae5b7da745ff3939219060da91f4661d2a1136a80736c3234db5b474e3599d\"" Feb 13 19:12:34.548673 containerd[1503]: time="2025-02-13T19:12:34.548621709Z" level=info msg="Forcibly stopping sandbox \"10ae5b7da745ff3939219060da91f4661d2a1136a80736c3234db5b474e3599d\"" Feb 13 19:12:34.548723 containerd[1503]: time="2025-02-13T19:12:34.548703191Z" level=info msg="TearDown network for sandbox \"10ae5b7da745ff3939219060da91f4661d2a1136a80736c3234db5b474e3599d\" successfully" Feb 13 19:12:34.552983 containerd[1503]: time="2025-02-13T19:12:34.552928933Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"10ae5b7da745ff3939219060da91f4661d2a1136a80736c3234db5b474e3599d\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:12:34.553115 containerd[1503]: time="2025-02-13T19:12:34.553037057Z" level=info msg="RemovePodSandbox \"10ae5b7da745ff3939219060da91f4661d2a1136a80736c3234db5b474e3599d\" returns successfully" Feb 13 19:12:34.553473 containerd[1503]: time="2025-02-13T19:12:34.553448031Z" level=info msg="StopPodSandbox for \"d2cb6c1d9e5335f4d8ee798e79c4a58d442740cd8f9ddfe185709b1cfbd4ea81\"" Feb 13 19:12:34.553564 containerd[1503]: time="2025-02-13T19:12:34.553544674Z" level=info msg="TearDown network for sandbox \"d2cb6c1d9e5335f4d8ee798e79c4a58d442740cd8f9ddfe185709b1cfbd4ea81\" successfully" Feb 13 19:12:34.553564 containerd[1503]: time="2025-02-13T19:12:34.553560874Z" level=info msg="StopPodSandbox for \"d2cb6c1d9e5335f4d8ee798e79c4a58d442740cd8f9ddfe185709b1cfbd4ea81\" returns successfully" Feb 13 19:12:34.554004 containerd[1503]: time="2025-02-13T19:12:34.553969968Z" level=info msg="RemovePodSandbox for \"d2cb6c1d9e5335f4d8ee798e79c4a58d442740cd8f9ddfe185709b1cfbd4ea81\"" Feb 13 19:12:34.554137 containerd[1503]: time="2025-02-13T19:12:34.554008970Z" level=info msg="Forcibly stopping sandbox \"d2cb6c1d9e5335f4d8ee798e79c4a58d442740cd8f9ddfe185709b1cfbd4ea81\"" Feb 13 19:12:34.554137 containerd[1503]: time="2025-02-13T19:12:34.554116813Z" level=info msg="TearDown network for sandbox \"d2cb6c1d9e5335f4d8ee798e79c4a58d442740cd8f9ddfe185709b1cfbd4ea81\" successfully" Feb 13 19:12:34.557268 containerd[1503]: time="2025-02-13T19:12:34.557217517Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"d2cb6c1d9e5335f4d8ee798e79c4a58d442740cd8f9ddfe185709b1cfbd4ea81\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:12:34.557365 containerd[1503]: time="2025-02-13T19:12:34.557323001Z" level=info msg="RemovePodSandbox \"d2cb6c1d9e5335f4d8ee798e79c4a58d442740cd8f9ddfe185709b1cfbd4ea81\" returns successfully" Feb 13 19:12:34.558656 containerd[1503]: time="2025-02-13T19:12:34.558417237Z" level=info msg="StopPodSandbox for \"7242b70e1006e9ac2f114fcee5b8cb6221151d9cbbf425c7398a098f6fd891ad\"" Feb 13 19:12:34.558656 containerd[1503]: time="2025-02-13T19:12:34.558532841Z" level=info msg="TearDown network for sandbox \"7242b70e1006e9ac2f114fcee5b8cb6221151d9cbbf425c7398a098f6fd891ad\" successfully" Feb 13 19:12:34.558656 containerd[1503]: time="2025-02-13T19:12:34.558545202Z" level=info msg="StopPodSandbox for \"7242b70e1006e9ac2f114fcee5b8cb6221151d9cbbf425c7398a098f6fd891ad\" returns successfully" Feb 13 19:12:34.559624 containerd[1503]: time="2025-02-13T19:12:34.559339188Z" level=info msg="RemovePodSandbox for \"7242b70e1006e9ac2f114fcee5b8cb6221151d9cbbf425c7398a098f6fd891ad\"" Feb 13 19:12:34.559624 containerd[1503]: time="2025-02-13T19:12:34.559381950Z" level=info msg="Forcibly stopping sandbox \"7242b70e1006e9ac2f114fcee5b8cb6221151d9cbbf425c7398a098f6fd891ad\"" Feb 13 19:12:34.559624 containerd[1503]: time="2025-02-13T19:12:34.559491273Z" level=info msg="TearDown network for sandbox \"7242b70e1006e9ac2f114fcee5b8cb6221151d9cbbf425c7398a098f6fd891ad\" successfully" Feb 13 19:12:34.563309 containerd[1503]: time="2025-02-13T19:12:34.563154196Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"7242b70e1006e9ac2f114fcee5b8cb6221151d9cbbf425c7398a098f6fd891ad\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:12:34.563309 containerd[1503]: time="2025-02-13T19:12:34.563235119Z" level=info msg="RemovePodSandbox \"7242b70e1006e9ac2f114fcee5b8cb6221151d9cbbf425c7398a098f6fd891ad\" returns successfully" Feb 13 19:12:34.564512 containerd[1503]: time="2025-02-13T19:12:34.564477401Z" level=info msg="StopPodSandbox for \"931e447af991e370f310777638932ea566d0d3e47f8ec0fb6ba2b0fe9e6ff401\"" Feb 13 19:12:34.564783 containerd[1503]: time="2025-02-13T19:12:34.564765210Z" level=info msg="TearDown network for sandbox \"931e447af991e370f310777638932ea566d0d3e47f8ec0fb6ba2b0fe9e6ff401\" successfully" Feb 13 19:12:34.564937 containerd[1503]: time="2025-02-13T19:12:34.564873014Z" level=info msg="StopPodSandbox for \"931e447af991e370f310777638932ea566d0d3e47f8ec0fb6ba2b0fe9e6ff401\" returns successfully" Feb 13 19:12:34.565737 containerd[1503]: time="2025-02-13T19:12:34.565698242Z" level=info msg="RemovePodSandbox for \"931e447af991e370f310777638932ea566d0d3e47f8ec0fb6ba2b0fe9e6ff401\"" Feb 13 19:12:34.565856 containerd[1503]: time="2025-02-13T19:12:34.565746283Z" level=info msg="Forcibly stopping sandbox \"931e447af991e370f310777638932ea566d0d3e47f8ec0fb6ba2b0fe9e6ff401\"" Feb 13 19:12:34.565899 containerd[1503]: time="2025-02-13T19:12:34.565860007Z" level=info msg="TearDown network for sandbox \"931e447af991e370f310777638932ea566d0d3e47f8ec0fb6ba2b0fe9e6ff401\" successfully" Feb 13 19:12:34.569499 containerd[1503]: time="2025-02-13T19:12:34.569319723Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"931e447af991e370f310777638932ea566d0d3e47f8ec0fb6ba2b0fe9e6ff401\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:12:34.569499 containerd[1503]: time="2025-02-13T19:12:34.569400326Z" level=info msg="RemovePodSandbox \"931e447af991e370f310777638932ea566d0d3e47f8ec0fb6ba2b0fe9e6ff401\" returns successfully" Feb 13 19:12:34.570169 containerd[1503]: time="2025-02-13T19:12:34.570140871Z" level=info msg="StopPodSandbox for \"d485d108247618e01bc6b75342565a89c0d97dadc8e9f0f8b8a03154fa8d72b1\"" Feb 13 19:12:34.570474 containerd[1503]: time="2025-02-13T19:12:34.570450601Z" level=info msg="TearDown network for sandbox \"d485d108247618e01bc6b75342565a89c0d97dadc8e9f0f8b8a03154fa8d72b1\" successfully" Feb 13 19:12:34.570474 containerd[1503]: time="2025-02-13T19:12:34.570471442Z" level=info msg="StopPodSandbox for \"d485d108247618e01bc6b75342565a89c0d97dadc8e9f0f8b8a03154fa8d72b1\" returns successfully" Feb 13 19:12:34.574086 containerd[1503]: time="2025-02-13T19:12:34.573622428Z" level=info msg="RemovePodSandbox for \"d485d108247618e01bc6b75342565a89c0d97dadc8e9f0f8b8a03154fa8d72b1\"" Feb 13 19:12:34.574086 containerd[1503]: time="2025-02-13T19:12:34.573664109Z" level=info msg="Forcibly stopping sandbox \"d485d108247618e01bc6b75342565a89c0d97dadc8e9f0f8b8a03154fa8d72b1\"" Feb 13 19:12:34.574086 containerd[1503]: time="2025-02-13T19:12:34.573739752Z" level=info msg="TearDown network for sandbox \"d485d108247618e01bc6b75342565a89c0d97dadc8e9f0f8b8a03154fa8d72b1\" successfully" Feb 13 19:12:34.577134 containerd[1503]: time="2025-02-13T19:12:34.577087864Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"d485d108247618e01bc6b75342565a89c0d97dadc8e9f0f8b8a03154fa8d72b1\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:12:34.577246 containerd[1503]: time="2025-02-13T19:12:34.577172987Z" level=info msg="RemovePodSandbox \"d485d108247618e01bc6b75342565a89c0d97dadc8e9f0f8b8a03154fa8d72b1\" returns successfully" Feb 13 19:12:34.577765 containerd[1503]: time="2025-02-13T19:12:34.577731205Z" level=info msg="StopPodSandbox for \"1f6b42d2735cedb916bc24bf2bb557d2ff72928d0c487a93a3dd46e88aa55abe\"" Feb 13 19:12:34.577859 containerd[1503]: time="2025-02-13T19:12:34.577838609Z" level=info msg="TearDown network for sandbox \"1f6b42d2735cedb916bc24bf2bb557d2ff72928d0c487a93a3dd46e88aa55abe\" successfully" Feb 13 19:12:34.577859 containerd[1503]: time="2025-02-13T19:12:34.577854930Z" level=info msg="StopPodSandbox for \"1f6b42d2735cedb916bc24bf2bb557d2ff72928d0c487a93a3dd46e88aa55abe\" returns successfully" Feb 13 19:12:34.578175 containerd[1503]: time="2025-02-13T19:12:34.578156420Z" level=info msg="RemovePodSandbox for \"1f6b42d2735cedb916bc24bf2bb557d2ff72928d0c487a93a3dd46e88aa55abe\"" Feb 13 19:12:34.578225 containerd[1503]: time="2025-02-13T19:12:34.578178460Z" level=info msg="Forcibly stopping sandbox \"1f6b42d2735cedb916bc24bf2bb557d2ff72928d0c487a93a3dd46e88aa55abe\"" Feb 13 19:12:34.578257 containerd[1503]: time="2025-02-13T19:12:34.578236302Z" level=info msg="TearDown network for sandbox \"1f6b42d2735cedb916bc24bf2bb557d2ff72928d0c487a93a3dd46e88aa55abe\" successfully" Feb 13 19:12:34.582109 containerd[1503]: time="2025-02-13T19:12:34.581975188Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"1f6b42d2735cedb916bc24bf2bb557d2ff72928d0c487a93a3dd46e88aa55abe\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:12:34.582109 containerd[1503]: time="2025-02-13T19:12:34.582112872Z" level=info msg="RemovePodSandbox \"1f6b42d2735cedb916bc24bf2bb557d2ff72928d0c487a93a3dd46e88aa55abe\" returns successfully" Feb 13 19:12:34.582626 containerd[1503]: time="2025-02-13T19:12:34.582548767Z" level=info msg="StopPodSandbox for \"4f219600df3d01ca237f481814d18c92c5997c6c501af4d70cb45cd41d1a6efc\"" Feb 13 19:12:34.582669 containerd[1503]: time="2025-02-13T19:12:34.582652651Z" level=info msg="TearDown network for sandbox \"4f219600df3d01ca237f481814d18c92c5997c6c501af4d70cb45cd41d1a6efc\" successfully" Feb 13 19:12:34.582669 containerd[1503]: time="2025-02-13T19:12:34.582663211Z" level=info msg="StopPodSandbox for \"4f219600df3d01ca237f481814d18c92c5997c6c501af4d70cb45cd41d1a6efc\" returns successfully" Feb 13 19:12:34.582903 containerd[1503]: time="2025-02-13T19:12:34.582878978Z" level=info msg="RemovePodSandbox for \"4f219600df3d01ca237f481814d18c92c5997c6c501af4d70cb45cd41d1a6efc\"" Feb 13 19:12:34.582938 containerd[1503]: time="2025-02-13T19:12:34.582926700Z" level=info msg="Forcibly stopping sandbox \"4f219600df3d01ca237f481814d18c92c5997c6c501af4d70cb45cd41d1a6efc\"" Feb 13 19:12:34.583182 containerd[1503]: time="2025-02-13T19:12:34.583025423Z" level=info msg="TearDown network for sandbox \"4f219600df3d01ca237f481814d18c92c5997c6c501af4d70cb45cd41d1a6efc\" successfully" Feb 13 19:12:34.586630 containerd[1503]: time="2025-02-13T19:12:34.586571142Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"4f219600df3d01ca237f481814d18c92c5997c6c501af4d70cb45cd41d1a6efc\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:12:34.586929 containerd[1503]: time="2025-02-13T19:12:34.586655985Z" level=info msg="RemovePodSandbox \"4f219600df3d01ca237f481814d18c92c5997c6c501af4d70cb45cd41d1a6efc\" returns successfully" Feb 13 19:12:34.588190 containerd[1503]: time="2025-02-13T19:12:34.587573816Z" level=info msg="StopPodSandbox for \"4642b0f0428148a51fe1afb0c6a261eff1112eb99bbdf06b50e264a81fa407b3\"" Feb 13 19:12:34.588190 containerd[1503]: time="2025-02-13T19:12:34.587705500Z" level=info msg="TearDown network for sandbox \"4642b0f0428148a51fe1afb0c6a261eff1112eb99bbdf06b50e264a81fa407b3\" successfully" Feb 13 19:12:34.588190 containerd[1503]: time="2025-02-13T19:12:34.587718301Z" level=info msg="StopPodSandbox for \"4642b0f0428148a51fe1afb0c6a261eff1112eb99bbdf06b50e264a81fa407b3\" returns successfully" Feb 13 19:12:34.589278 containerd[1503]: time="2025-02-13T19:12:34.589222751Z" level=info msg="RemovePodSandbox for \"4642b0f0428148a51fe1afb0c6a261eff1112eb99bbdf06b50e264a81fa407b3\"" Feb 13 19:12:34.589278 containerd[1503]: time="2025-02-13T19:12:34.589259152Z" level=info msg="Forcibly stopping sandbox \"4642b0f0428148a51fe1afb0c6a261eff1112eb99bbdf06b50e264a81fa407b3\"" Feb 13 19:12:34.589377 containerd[1503]: time="2025-02-13T19:12:34.589331595Z" level=info msg="TearDown network for sandbox \"4642b0f0428148a51fe1afb0c6a261eff1112eb99bbdf06b50e264a81fa407b3\" successfully" Feb 13 19:12:34.592868 containerd[1503]: time="2025-02-13T19:12:34.592824232Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"4642b0f0428148a51fe1afb0c6a261eff1112eb99bbdf06b50e264a81fa407b3\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:12:34.592868 containerd[1503]: time="2025-02-13T19:12:34.592893594Z" level=info msg="RemovePodSandbox \"4642b0f0428148a51fe1afb0c6a261eff1112eb99bbdf06b50e264a81fa407b3\" returns successfully" Feb 13 19:12:34.593966 containerd[1503]: time="2025-02-13T19:12:34.593601378Z" level=info msg="StopPodSandbox for \"c6b1e16315492197341e134074aec5720ae8f11ba5d44a3e6e3a0312f21d145d\"" Feb 13 19:12:34.593966 containerd[1503]: time="2025-02-13T19:12:34.593731302Z" level=info msg="TearDown network for sandbox \"c6b1e16315492197341e134074aec5720ae8f11ba5d44a3e6e3a0312f21d145d\" successfully" Feb 13 19:12:34.593966 containerd[1503]: time="2025-02-13T19:12:34.593744703Z" level=info msg="StopPodSandbox for \"c6b1e16315492197341e134074aec5720ae8f11ba5d44a3e6e3a0312f21d145d\" returns successfully" Feb 13 19:12:34.594787 containerd[1503]: time="2025-02-13T19:12:34.594511288Z" level=info msg="RemovePodSandbox for \"c6b1e16315492197341e134074aec5720ae8f11ba5d44a3e6e3a0312f21d145d\"" Feb 13 19:12:34.594787 containerd[1503]: time="2025-02-13T19:12:34.594745176Z" level=info msg="Forcibly stopping sandbox \"c6b1e16315492197341e134074aec5720ae8f11ba5d44a3e6e3a0312f21d145d\"" Feb 13 19:12:34.595322 containerd[1503]: time="2025-02-13T19:12:34.595218352Z" level=info msg="TearDown network for sandbox \"c6b1e16315492197341e134074aec5720ae8f11ba5d44a3e6e3a0312f21d145d\" successfully" Feb 13 19:12:34.605947 containerd[1503]: time="2025-02-13T19:12:34.605745825Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"c6b1e16315492197341e134074aec5720ae8f11ba5d44a3e6e3a0312f21d145d\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:12:34.605947 containerd[1503]: time="2025-02-13T19:12:34.605824948Z" level=info msg="RemovePodSandbox \"c6b1e16315492197341e134074aec5720ae8f11ba5d44a3e6e3a0312f21d145d\" returns successfully" Feb 13 19:12:34.607130 containerd[1503]: time="2025-02-13T19:12:34.606664736Z" level=info msg="StopPodSandbox for \"728705ac6957c83c18f350153e00e11e14122291646534ceb675f3fda8f70cdb\"" Feb 13 19:12:34.607130 containerd[1503]: time="2025-02-13T19:12:34.606812741Z" level=info msg="TearDown network for sandbox \"728705ac6957c83c18f350153e00e11e14122291646534ceb675f3fda8f70cdb\" successfully" Feb 13 19:12:34.607130 containerd[1503]: time="2025-02-13T19:12:34.606826582Z" level=info msg="StopPodSandbox for \"728705ac6957c83c18f350153e00e11e14122291646534ceb675f3fda8f70cdb\" returns successfully" Feb 13 19:12:34.609099 containerd[1503]: time="2025-02-13T19:12:34.607500644Z" level=info msg="RemovePodSandbox for \"728705ac6957c83c18f350153e00e11e14122291646534ceb675f3fda8f70cdb\"" Feb 13 19:12:34.609099 containerd[1503]: time="2025-02-13T19:12:34.607529565Z" level=info msg="Forcibly stopping sandbox \"728705ac6957c83c18f350153e00e11e14122291646534ceb675f3fda8f70cdb\"" Feb 13 19:12:34.609099 containerd[1503]: time="2025-02-13T19:12:34.607594007Z" level=info msg="TearDown network for sandbox \"728705ac6957c83c18f350153e00e11e14122291646534ceb675f3fda8f70cdb\" successfully" Feb 13 19:12:34.611658 containerd[1503]: time="2025-02-13T19:12:34.611618982Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"728705ac6957c83c18f350153e00e11e14122291646534ceb675f3fda8f70cdb\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:12:34.611906 containerd[1503]: time="2025-02-13T19:12:34.611887071Z" level=info msg="RemovePodSandbox \"728705ac6957c83c18f350153e00e11e14122291646534ceb675f3fda8f70cdb\" returns successfully" Feb 13 19:12:34.612721 containerd[1503]: time="2025-02-13T19:12:34.612693219Z" level=info msg="StopPodSandbox for \"77460032187adb446e84e042dcc0956c8b578caedef00c1e669bc2016d01bf85\"" Feb 13 19:12:34.612924 containerd[1503]: time="2025-02-13T19:12:34.612901706Z" level=info msg="TearDown network for sandbox \"77460032187adb446e84e042dcc0956c8b578caedef00c1e669bc2016d01bf85\" successfully" Feb 13 19:12:34.613099 containerd[1503]: time="2025-02-13T19:12:34.613078671Z" level=info msg="StopPodSandbox for \"77460032187adb446e84e042dcc0956c8b578caedef00c1e669bc2016d01bf85\" returns successfully" Feb 13 19:12:34.613563 containerd[1503]: time="2025-02-13T19:12:34.613541167Z" level=info msg="RemovePodSandbox for \"77460032187adb446e84e042dcc0956c8b578caedef00c1e669bc2016d01bf85\"" Feb 13 19:12:34.613713 containerd[1503]: time="2025-02-13T19:12:34.613675171Z" level=info msg="Forcibly stopping sandbox \"77460032187adb446e84e042dcc0956c8b578caedef00c1e669bc2016d01bf85\"" Feb 13 19:12:34.613816 containerd[1503]: time="2025-02-13T19:12:34.613794215Z" level=info msg="TearDown network for sandbox \"77460032187adb446e84e042dcc0956c8b578caedef00c1e669bc2016d01bf85\" successfully" Feb 13 19:12:34.619433 containerd[1503]: time="2025-02-13T19:12:34.619345242Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"77460032187adb446e84e042dcc0956c8b578caedef00c1e669bc2016d01bf85\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:12:34.620041 containerd[1503]: time="2025-02-13T19:12:34.619461766Z" level=info msg="RemovePodSandbox \"77460032187adb446e84e042dcc0956c8b578caedef00c1e669bc2016d01bf85\" returns successfully" Feb 13 19:12:34.620912 containerd[1503]: time="2025-02-13T19:12:34.620415198Z" level=info msg="StopPodSandbox for \"930563c746d8b91d524a61a614694c47905765a47fb87cb32a681d5b35060a78\"" Feb 13 19:12:34.620912 containerd[1503]: time="2025-02-13T19:12:34.620568203Z" level=info msg="TearDown network for sandbox \"930563c746d8b91d524a61a614694c47905765a47fb87cb32a681d5b35060a78\" successfully" Feb 13 19:12:34.620912 containerd[1503]: time="2025-02-13T19:12:34.620587003Z" level=info msg="StopPodSandbox for \"930563c746d8b91d524a61a614694c47905765a47fb87cb32a681d5b35060a78\" returns successfully" Feb 13 19:12:34.622348 containerd[1503]: time="2025-02-13T19:12:34.621903568Z" level=info msg="RemovePodSandbox for \"930563c746d8b91d524a61a614694c47905765a47fb87cb32a681d5b35060a78\"" Feb 13 19:12:34.622348 containerd[1503]: time="2025-02-13T19:12:34.622013691Z" level=info msg="Forcibly stopping sandbox \"930563c746d8b91d524a61a614694c47905765a47fb87cb32a681d5b35060a78\"" Feb 13 19:12:34.622348 containerd[1503]: time="2025-02-13T19:12:34.622223218Z" level=info msg="TearDown network for sandbox \"930563c746d8b91d524a61a614694c47905765a47fb87cb32a681d5b35060a78\" successfully" Feb 13 19:12:34.627209 containerd[1503]: time="2025-02-13T19:12:34.627148944Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"930563c746d8b91d524a61a614694c47905765a47fb87cb32a681d5b35060a78\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:12:34.627860 containerd[1503]: time="2025-02-13T19:12:34.627498635Z" level=info msg="RemovePodSandbox \"930563c746d8b91d524a61a614694c47905765a47fb87cb32a681d5b35060a78\" returns successfully" Feb 13 19:12:34.628495 containerd[1503]: time="2025-02-13T19:12:34.628152457Z" level=info msg="StopPodSandbox for \"ae9615caa56ac5be11c6b7cc88aac5196ade400655d7a9313d485332a9d3820e\"" Feb 13 19:12:34.628495 containerd[1503]: time="2025-02-13T19:12:34.628246060Z" level=info msg="TearDown network for sandbox \"ae9615caa56ac5be11c6b7cc88aac5196ade400655d7a9313d485332a9d3820e\" successfully" Feb 13 19:12:34.628495 containerd[1503]: time="2025-02-13T19:12:34.628255221Z" level=info msg="StopPodSandbox for \"ae9615caa56ac5be11c6b7cc88aac5196ade400655d7a9313d485332a9d3820e\" returns successfully" Feb 13 19:12:34.629630 containerd[1503]: time="2025-02-13T19:12:34.629436500Z" level=info msg="RemovePodSandbox for \"ae9615caa56ac5be11c6b7cc88aac5196ade400655d7a9313d485332a9d3820e\"" Feb 13 19:12:34.629630 containerd[1503]: time="2025-02-13T19:12:34.629478942Z" level=info msg="Forcibly stopping sandbox \"ae9615caa56ac5be11c6b7cc88aac5196ade400655d7a9313d485332a9d3820e\"" Feb 13 19:12:34.629630 containerd[1503]: time="2025-02-13T19:12:34.629574185Z" level=info msg="TearDown network for sandbox \"ae9615caa56ac5be11c6b7cc88aac5196ade400655d7a9313d485332a9d3820e\" successfully" Feb 13 19:12:34.635260 containerd[1503]: time="2025-02-13T19:12:34.634749959Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"ae9615caa56ac5be11c6b7cc88aac5196ade400655d7a9313d485332a9d3820e\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:12:34.635260 containerd[1503]: time="2025-02-13T19:12:34.634835601Z" level=info msg="RemovePodSandbox \"ae9615caa56ac5be11c6b7cc88aac5196ade400655d7a9313d485332a9d3820e\" returns successfully" Feb 13 19:12:34.635818 containerd[1503]: time="2025-02-13T19:12:34.635638268Z" level=info msg="StopPodSandbox for \"af85f8d50ee74cce2c7e2cee2af5922ba7bbd4077fc92d81b183d5575fad1329\"" Feb 13 19:12:34.635818 containerd[1503]: time="2025-02-13T19:12:34.635761353Z" level=info msg="TearDown network for sandbox \"af85f8d50ee74cce2c7e2cee2af5922ba7bbd4077fc92d81b183d5575fad1329\" successfully" Feb 13 19:12:34.635818 containerd[1503]: time="2025-02-13T19:12:34.635773953Z" level=info msg="StopPodSandbox for \"af85f8d50ee74cce2c7e2cee2af5922ba7bbd4077fc92d81b183d5575fad1329\" returns successfully" Feb 13 19:12:34.637252 containerd[1503]: time="2025-02-13T19:12:34.636398574Z" level=info msg="RemovePodSandbox for \"af85f8d50ee74cce2c7e2cee2af5922ba7bbd4077fc92d81b183d5575fad1329\"" Feb 13 19:12:34.637252 containerd[1503]: time="2025-02-13T19:12:34.636422535Z" level=info msg="Forcibly stopping sandbox \"af85f8d50ee74cce2c7e2cee2af5922ba7bbd4077fc92d81b183d5575fad1329\"" Feb 13 19:12:34.637252 containerd[1503]: time="2025-02-13T19:12:34.636497657Z" level=info msg="TearDown network for sandbox \"af85f8d50ee74cce2c7e2cee2af5922ba7bbd4077fc92d81b183d5575fad1329\" successfully" Feb 13 19:12:34.639875 containerd[1503]: time="2025-02-13T19:12:34.639808888Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"af85f8d50ee74cce2c7e2cee2af5922ba7bbd4077fc92d81b183d5575fad1329\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:12:34.639875 containerd[1503]: time="2025-02-13T19:12:34.639878451Z" level=info msg="RemovePodSandbox \"af85f8d50ee74cce2c7e2cee2af5922ba7bbd4077fc92d81b183d5575fad1329\" returns successfully" Feb 13 19:12:34.641246 containerd[1503]: time="2025-02-13T19:12:34.640443550Z" level=info msg="StopPodSandbox for \"cf3ac12d3142a48e0052c1dead1d74d6becc46ab9352d1240e51b560f73cb623\"" Feb 13 19:12:34.641246 containerd[1503]: time="2025-02-13T19:12:34.640583394Z" level=info msg="TearDown network for sandbox \"cf3ac12d3142a48e0052c1dead1d74d6becc46ab9352d1240e51b560f73cb623\" successfully" Feb 13 19:12:34.641246 containerd[1503]: time="2025-02-13T19:12:34.640599995Z" level=info msg="StopPodSandbox for \"cf3ac12d3142a48e0052c1dead1d74d6becc46ab9352d1240e51b560f73cb623\" returns successfully" Feb 13 19:12:34.641246 containerd[1503]: time="2025-02-13T19:12:34.641028649Z" level=info msg="RemovePodSandbox for \"cf3ac12d3142a48e0052c1dead1d74d6becc46ab9352d1240e51b560f73cb623\"" Feb 13 19:12:34.641246 containerd[1503]: time="2025-02-13T19:12:34.641072691Z" level=info msg="Forcibly stopping sandbox \"cf3ac12d3142a48e0052c1dead1d74d6becc46ab9352d1240e51b560f73cb623\"" Feb 13 19:12:34.641246 containerd[1503]: time="2025-02-13T19:12:34.641151853Z" level=info msg="TearDown network for sandbox \"cf3ac12d3142a48e0052c1dead1d74d6becc46ab9352d1240e51b560f73cb623\" successfully" Feb 13 19:12:34.644194 containerd[1503]: time="2025-02-13T19:12:34.644138474Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"cf3ac12d3142a48e0052c1dead1d74d6becc46ab9352d1240e51b560f73cb623\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:12:34.644537 containerd[1503]: time="2025-02-13T19:12:34.644216876Z" level=info msg="RemovePodSandbox \"cf3ac12d3142a48e0052c1dead1d74d6becc46ab9352d1240e51b560f73cb623\" returns successfully" Feb 13 19:12:34.644637 containerd[1503]: time="2025-02-13T19:12:34.644603929Z" level=info msg="StopPodSandbox for \"4d5614a8446f6178d266e8f2147ca224323d91c476899b1e9e66b08c962ad2a3\"" Feb 13 19:12:34.644748 containerd[1503]: time="2025-02-13T19:12:34.644702813Z" level=info msg="TearDown network for sandbox \"4d5614a8446f6178d266e8f2147ca224323d91c476899b1e9e66b08c962ad2a3\" successfully" Feb 13 19:12:34.644748 containerd[1503]: time="2025-02-13T19:12:34.644718733Z" level=info msg="StopPodSandbox for \"4d5614a8446f6178d266e8f2147ca224323d91c476899b1e9e66b08c962ad2a3\" returns successfully" Feb 13 19:12:34.645245 containerd[1503]: time="2025-02-13T19:12:34.645197589Z" level=info msg="RemovePodSandbox for \"4d5614a8446f6178d266e8f2147ca224323d91c476899b1e9e66b08c962ad2a3\"" Feb 13 19:12:34.645245 containerd[1503]: time="2025-02-13T19:12:34.645227510Z" level=info msg="Forcibly stopping sandbox \"4d5614a8446f6178d266e8f2147ca224323d91c476899b1e9e66b08c962ad2a3\"" Feb 13 19:12:34.645349 containerd[1503]: time="2025-02-13T19:12:34.645284792Z" level=info msg="TearDown network for sandbox \"4d5614a8446f6178d266e8f2147ca224323d91c476899b1e9e66b08c962ad2a3\" successfully" Feb 13 19:12:34.648210 containerd[1503]: time="2025-02-13T19:12:34.648168489Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"4d5614a8446f6178d266e8f2147ca224323d91c476899b1e9e66b08c962ad2a3\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:12:34.648304 containerd[1503]: time="2025-02-13T19:12:34.648233171Z" level=info msg="RemovePodSandbox \"4d5614a8446f6178d266e8f2147ca224323d91c476899b1e9e66b08c962ad2a3\" returns successfully" Feb 13 19:12:34.648960 containerd[1503]: time="2025-02-13T19:12:34.648854392Z" level=info msg="StopPodSandbox for \"b86651f4cc54f520cab1a18f98b7a70986a4aa77bdc28bce93539ac71deee2a2\"" Feb 13 19:12:34.648960 containerd[1503]: time="2025-02-13T19:12:34.648942155Z" level=info msg="TearDown network for sandbox \"b86651f4cc54f520cab1a18f98b7a70986a4aa77bdc28bce93539ac71deee2a2\" successfully" Feb 13 19:12:34.648960 containerd[1503]: time="2025-02-13T19:12:34.648952795Z" level=info msg="StopPodSandbox for \"b86651f4cc54f520cab1a18f98b7a70986a4aa77bdc28bce93539ac71deee2a2\" returns successfully" Feb 13 19:12:34.652317 containerd[1503]: time="2025-02-13T19:12:34.649726501Z" level=info msg="RemovePodSandbox for \"b86651f4cc54f520cab1a18f98b7a70986a4aa77bdc28bce93539ac71deee2a2\"" Feb 13 19:12:34.652317 containerd[1503]: time="2025-02-13T19:12:34.649766102Z" level=info msg="Forcibly stopping sandbox \"b86651f4cc54f520cab1a18f98b7a70986a4aa77bdc28bce93539ac71deee2a2\"" Feb 13 19:12:34.652317 containerd[1503]: time="2025-02-13T19:12:34.649872426Z" level=info msg="TearDown network for sandbox \"b86651f4cc54f520cab1a18f98b7a70986a4aa77bdc28bce93539ac71deee2a2\" successfully" Feb 13 19:12:34.654962 containerd[1503]: time="2025-02-13T19:12:34.654909355Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"b86651f4cc54f520cab1a18f98b7a70986a4aa77bdc28bce93539ac71deee2a2\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:12:34.655096 containerd[1503]: time="2025-02-13T19:12:34.655012398Z" level=info msg="RemovePodSandbox \"b86651f4cc54f520cab1a18f98b7a70986a4aa77bdc28bce93539ac71deee2a2\" returns successfully" Feb 13 19:12:34.656105 containerd[1503]: time="2025-02-13T19:12:34.655851947Z" level=info msg="StopPodSandbox for \"0c31ed484cb1e7593659d8fd1f850350d6b3b9e01ed8530e08cf66bc47a72b73\"" Feb 13 19:12:34.656105 containerd[1503]: time="2025-02-13T19:12:34.655972391Z" level=info msg="TearDown network for sandbox \"0c31ed484cb1e7593659d8fd1f850350d6b3b9e01ed8530e08cf66bc47a72b73\" successfully" Feb 13 19:12:34.656105 containerd[1503]: time="2025-02-13T19:12:34.656014872Z" level=info msg="StopPodSandbox for \"0c31ed484cb1e7593659d8fd1f850350d6b3b9e01ed8530e08cf66bc47a72b73\" returns successfully" Feb 13 19:12:34.657097 containerd[1503]: time="2025-02-13T19:12:34.656738256Z" level=info msg="RemovePodSandbox for \"0c31ed484cb1e7593659d8fd1f850350d6b3b9e01ed8530e08cf66bc47a72b73\"" Feb 13 19:12:34.657097 containerd[1503]: time="2025-02-13T19:12:34.656802298Z" level=info msg="Forcibly stopping sandbox \"0c31ed484cb1e7593659d8fd1f850350d6b3b9e01ed8530e08cf66bc47a72b73\"" Feb 13 19:12:34.657097 containerd[1503]: time="2025-02-13T19:12:34.656894382Z" level=info msg="TearDown network for sandbox \"0c31ed484cb1e7593659d8fd1f850350d6b3b9e01ed8530e08cf66bc47a72b73\" successfully" Feb 13 19:12:34.661379 containerd[1503]: time="2025-02-13T19:12:34.661314050Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"0c31ed484cb1e7593659d8fd1f850350d6b3b9e01ed8530e08cf66bc47a72b73\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:12:34.661499 containerd[1503]: time="2025-02-13T19:12:34.661400533Z" level=info msg="RemovePodSandbox \"0c31ed484cb1e7593659d8fd1f850350d6b3b9e01ed8530e08cf66bc47a72b73\" returns successfully" Feb 13 19:12:34.662135 containerd[1503]: time="2025-02-13T19:12:34.662043474Z" level=info msg="StopPodSandbox for \"07aed5c023700866ebeb2a6224ac11566918d60f5a140685f482d3a85f4cd555\"" Feb 13 19:12:34.662264 containerd[1503]: time="2025-02-13T19:12:34.662195679Z" level=info msg="TearDown network for sandbox \"07aed5c023700866ebeb2a6224ac11566918d60f5a140685f482d3a85f4cd555\" successfully" Feb 13 19:12:34.662264 containerd[1503]: time="2025-02-13T19:12:34.662210480Z" level=info msg="StopPodSandbox for \"07aed5c023700866ebeb2a6224ac11566918d60f5a140685f482d3a85f4cd555\" returns successfully" Feb 13 19:12:34.662600 containerd[1503]: time="2025-02-13T19:12:34.662563052Z" level=info msg="RemovePodSandbox for \"07aed5c023700866ebeb2a6224ac11566918d60f5a140685f482d3a85f4cd555\"" Feb 13 19:12:34.662708 containerd[1503]: time="2025-02-13T19:12:34.662677136Z" level=info msg="Forcibly stopping sandbox \"07aed5c023700866ebeb2a6224ac11566918d60f5a140685f482d3a85f4cd555\"" Feb 13 19:12:34.662792 containerd[1503]: time="2025-02-13T19:12:34.662769099Z" level=info msg="TearDown network for sandbox \"07aed5c023700866ebeb2a6224ac11566918d60f5a140685f482d3a85f4cd555\" successfully" Feb 13 19:12:34.669145 containerd[1503]: time="2025-02-13T19:12:34.669026749Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"07aed5c023700866ebeb2a6224ac11566918d60f5a140685f482d3a85f4cd555\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:12:34.669145 containerd[1503]: time="2025-02-13T19:12:34.669150473Z" level=info msg="RemovePodSandbox \"07aed5c023700866ebeb2a6224ac11566918d60f5a140685f482d3a85f4cd555\" returns successfully" Feb 13 19:12:34.669659 containerd[1503]: time="2025-02-13T19:12:34.669541126Z" level=info msg="StopPodSandbox for \"a657753a39333ef1c8ee279e3321a3d04ca609cd0499c88ae5589336776d3f50\"" Feb 13 19:12:34.669659 containerd[1503]: time="2025-02-13T19:12:34.669639969Z" level=info msg="TearDown network for sandbox \"a657753a39333ef1c8ee279e3321a3d04ca609cd0499c88ae5589336776d3f50\" successfully" Feb 13 19:12:34.669659 containerd[1503]: time="2025-02-13T19:12:34.669650490Z" level=info msg="StopPodSandbox for \"a657753a39333ef1c8ee279e3321a3d04ca609cd0499c88ae5589336776d3f50\" returns successfully" Feb 13 19:12:34.670907 containerd[1503]: time="2025-02-13T19:12:34.670374834Z" level=info msg="RemovePodSandbox for \"a657753a39333ef1c8ee279e3321a3d04ca609cd0499c88ae5589336776d3f50\"" Feb 13 19:12:34.670907 containerd[1503]: time="2025-02-13T19:12:34.670404595Z" level=info msg="Forcibly stopping sandbox \"a657753a39333ef1c8ee279e3321a3d04ca609cd0499c88ae5589336776d3f50\"" Feb 13 19:12:34.670907 containerd[1503]: time="2025-02-13T19:12:34.670479797Z" level=info msg="TearDown network for sandbox \"a657753a39333ef1c8ee279e3321a3d04ca609cd0499c88ae5589336776d3f50\" successfully" Feb 13 19:12:34.676509 containerd[1503]: time="2025-02-13T19:12:34.676394156Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"a657753a39333ef1c8ee279e3321a3d04ca609cd0499c88ae5589336776d3f50\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:12:34.676622 containerd[1503]: time="2025-02-13T19:12:34.676571762Z" level=info msg="RemovePodSandbox \"a657753a39333ef1c8ee279e3321a3d04ca609cd0499c88ae5589336776d3f50\" returns successfully" Feb 13 19:12:34.677130 containerd[1503]: time="2025-02-13T19:12:34.677108860Z" level=info msg="StopPodSandbox for \"68ee287533ff713dee0f8970cee78f6429754d9b99d2574db1ab310eb93f45f4\"" Feb 13 19:12:34.677643 containerd[1503]: time="2025-02-13T19:12:34.677414710Z" level=info msg="TearDown network for sandbox \"68ee287533ff713dee0f8970cee78f6429754d9b99d2574db1ab310eb93f45f4\" successfully" Feb 13 19:12:34.677643 containerd[1503]: time="2025-02-13T19:12:34.677433791Z" level=info msg="StopPodSandbox for \"68ee287533ff713dee0f8970cee78f6429754d9b99d2574db1ab310eb93f45f4\" returns successfully" Feb 13 19:12:34.678587 containerd[1503]: time="2025-02-13T19:12:34.678484826Z" level=info msg="RemovePodSandbox for \"68ee287533ff713dee0f8970cee78f6429754d9b99d2574db1ab310eb93f45f4\"" Feb 13 19:12:34.678587 containerd[1503]: time="2025-02-13T19:12:34.678518027Z" level=info msg="Forcibly stopping sandbox \"68ee287533ff713dee0f8970cee78f6429754d9b99d2574db1ab310eb93f45f4\"" Feb 13 19:12:34.678667 containerd[1503]: time="2025-02-13T19:12:34.678604990Z" level=info msg="TearDown network for sandbox \"68ee287533ff713dee0f8970cee78f6429754d9b99d2574db1ab310eb93f45f4\" successfully" Feb 13 19:12:34.683182 containerd[1503]: time="2025-02-13T19:12:34.683143342Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"68ee287533ff713dee0f8970cee78f6429754d9b99d2574db1ab310eb93f45f4\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:12:34.683289 containerd[1503]: time="2025-02-13T19:12:34.683214385Z" level=info msg="RemovePodSandbox \"68ee287533ff713dee0f8970cee78f6429754d9b99d2574db1ab310eb93f45f4\" returns successfully" Feb 13 19:12:34.683897 containerd[1503]: time="2025-02-13T19:12:34.683872087Z" level=info msg="StopPodSandbox for \"602a514f998cb50210ddc95537cfa35ed74b37741d39df2aa29785c3996317a9\"" Feb 13 19:12:34.684004 containerd[1503]: time="2025-02-13T19:12:34.683972250Z" level=info msg="TearDown network for sandbox \"602a514f998cb50210ddc95537cfa35ed74b37741d39df2aa29785c3996317a9\" successfully" Feb 13 19:12:34.684004 containerd[1503]: time="2025-02-13T19:12:34.683996771Z" level=info msg="StopPodSandbox for \"602a514f998cb50210ddc95537cfa35ed74b37741d39df2aa29785c3996317a9\" returns successfully" Feb 13 19:12:34.684642 containerd[1503]: time="2025-02-13T19:12:34.684620272Z" level=info msg="RemovePodSandbox for \"602a514f998cb50210ddc95537cfa35ed74b37741d39df2aa29785c3996317a9\"" Feb 13 19:12:34.684696 containerd[1503]: time="2025-02-13T19:12:34.684646593Z" level=info msg="Forcibly stopping sandbox \"602a514f998cb50210ddc95537cfa35ed74b37741d39df2aa29785c3996317a9\"" Feb 13 19:12:34.684747 containerd[1503]: time="2025-02-13T19:12:34.684706475Z" level=info msg="TearDown network for sandbox \"602a514f998cb50210ddc95537cfa35ed74b37741d39df2aa29785c3996317a9\" successfully" Feb 13 19:12:34.689707 containerd[1503]: time="2025-02-13T19:12:34.689646240Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"602a514f998cb50210ddc95537cfa35ed74b37741d39df2aa29785c3996317a9\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:12:34.689952 containerd[1503]: time="2025-02-13T19:12:34.689733003Z" level=info msg="RemovePodSandbox \"602a514f998cb50210ddc95537cfa35ed74b37741d39df2aa29785c3996317a9\" returns successfully" Feb 13 19:12:34.691422 containerd[1503]: time="2025-02-13T19:12:34.691378779Z" level=info msg="StopPodSandbox for \"76c8fc7995e25f780bca1521611986ac1d0cb18b34d8d5db637de52885bbe0b7\"" Feb 13 19:12:34.691512 containerd[1503]: time="2025-02-13T19:12:34.691487902Z" level=info msg="TearDown network for sandbox \"76c8fc7995e25f780bca1521611986ac1d0cb18b34d8d5db637de52885bbe0b7\" successfully" Feb 13 19:12:34.691512 containerd[1503]: time="2025-02-13T19:12:34.691498063Z" level=info msg="StopPodSandbox for \"76c8fc7995e25f780bca1521611986ac1d0cb18b34d8d5db637de52885bbe0b7\" returns successfully" Feb 13 19:12:34.693035 containerd[1503]: time="2025-02-13T19:12:34.691929437Z" level=info msg="RemovePodSandbox for \"76c8fc7995e25f780bca1521611986ac1d0cb18b34d8d5db637de52885bbe0b7\"" Feb 13 19:12:34.693035 containerd[1503]: time="2025-02-13T19:12:34.691957318Z" level=info msg="Forcibly stopping sandbox \"76c8fc7995e25f780bca1521611986ac1d0cb18b34d8d5db637de52885bbe0b7\"" Feb 13 19:12:34.693035 containerd[1503]: time="2025-02-13T19:12:34.692036961Z" level=info msg="TearDown network for sandbox \"76c8fc7995e25f780bca1521611986ac1d0cb18b34d8d5db637de52885bbe0b7\" successfully" Feb 13 19:12:34.696977 containerd[1503]: time="2025-02-13T19:12:34.696905444Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"76c8fc7995e25f780bca1521611986ac1d0cb18b34d8d5db637de52885bbe0b7\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:12:34.697160 containerd[1503]: time="2025-02-13T19:12:34.697050689Z" level=info msg="RemovePodSandbox \"76c8fc7995e25f780bca1521611986ac1d0cb18b34d8d5db637de52885bbe0b7\" returns successfully" Feb 13 19:12:47.528123 kubelet[2828]: I0213 19:12:47.527851 2828 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 13 19:13:10.733393 kubelet[2828]: I0213 19:13:10.733231 2828 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 13 19:13:17.883282 update_engine[1486]: I20250213 19:13:17.883195 1486 prefs.cc:52] certificate-report-to-send-update not present in /var/lib/update_engine/prefs Feb 13 19:13:17.883282 update_engine[1486]: I20250213 19:13:17.883260 1486 prefs.cc:52] certificate-report-to-send-download not present in /var/lib/update_engine/prefs Feb 13 19:13:17.884027 update_engine[1486]: I20250213 19:13:17.883587 1486 prefs.cc:52] aleph-version not present in /var/lib/update_engine/prefs Feb 13 19:13:17.884819 update_engine[1486]: I20250213 19:13:17.884656 1486 omaha_request_params.cc:62] Current group set to alpha Feb 13 19:13:17.885882 update_engine[1486]: I20250213 19:13:17.885730 1486 update_attempter.cc:499] Already updated boot flags. Skipping. Feb 13 19:13:17.885882 update_engine[1486]: I20250213 19:13:17.885762 1486 update_attempter.cc:643] Scheduling an action processor start. Feb 13 19:13:17.885882 update_engine[1486]: I20250213 19:13:17.885787 1486 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction Feb 13 19:13:17.885882 update_engine[1486]: I20250213 19:13:17.885840 1486 prefs.cc:52] previous-version not present in /var/lib/update_engine/prefs Feb 13 19:13:17.886556 update_engine[1486]: I20250213 19:13:17.886516 1486 omaha_request_action.cc:271] Posting an Omaha request to disabled Feb 13 19:13:17.887092 update_engine[1486]: I20250213 19:13:17.886646 1486 omaha_request_action.cc:272] Request: Feb 13 19:13:17.887092 update_engine[1486]: Feb 13 19:13:17.887092 update_engine[1486]: Feb 13 19:13:17.887092 update_engine[1486]: Feb 13 19:13:17.887092 update_engine[1486]: Feb 13 19:13:17.887092 update_engine[1486]: Feb 13 19:13:17.887092 update_engine[1486]: Feb 13 19:13:17.887092 update_engine[1486]: Feb 13 19:13:17.887092 update_engine[1486]: Feb 13 19:13:17.887092 update_engine[1486]: I20250213 19:13:17.886668 1486 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Feb 13 19:13:17.892048 update_engine[1486]: I20250213 19:13:17.890637 1486 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Feb 13 19:13:17.892048 update_engine[1486]: I20250213 19:13:17.891161 1486 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Feb 13 19:13:17.892630 locksmithd[1508]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_CHECKING_FOR_UPDATE" NewVersion=0.0.0 NewSize=0 Feb 13 19:13:17.893710 update_engine[1486]: E20250213 19:13:17.893599 1486 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Feb 13 19:13:17.893710 update_engine[1486]: I20250213 19:13:17.893675 1486 libcurl_http_fetcher.cc:283] No HTTP response, retry 1 Feb 13 19:13:27.848602 update_engine[1486]: I20250213 19:13:27.848100 1486 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Feb 13 19:13:27.848602 update_engine[1486]: I20250213 19:13:27.848335 1486 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Feb 13 19:13:27.848602 update_engine[1486]: I20250213 19:13:27.848558 1486 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Feb 13 19:13:27.849310 update_engine[1486]: E20250213 19:13:27.849282 1486 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Feb 13 19:13:27.849408 update_engine[1486]: I20250213 19:13:27.849391 1486 libcurl_http_fetcher.cc:283] No HTTP response, retry 2 Feb 13 19:13:37.843129 update_engine[1486]: I20250213 19:13:37.842769 1486 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Feb 13 19:13:37.843129 update_engine[1486]: I20250213 19:13:37.843031 1486 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Feb 13 19:13:37.843569 update_engine[1486]: I20250213 19:13:37.843375 1486 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Feb 13 19:13:37.843925 update_engine[1486]: E20250213 19:13:37.843876 1486 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Feb 13 19:13:37.843978 update_engine[1486]: I20250213 19:13:37.843935 1486 libcurl_http_fetcher.cc:283] No HTTP response, retry 3 Feb 13 19:13:46.473438 systemd[1]: Started sshd@16-78.46.147.231:22-117.247.111.70:38239.service - OpenSSH per-connection server daemon (117.247.111.70:38239). Feb 13 19:13:47.466954 sshd[5925]: Invalid user user1 from 117.247.111.70 port 38239 Feb 13 19:13:47.656435 sshd[5925]: Received disconnect from 117.247.111.70 port 38239:11: Bye Bye [preauth] Feb 13 19:13:47.656435 sshd[5925]: Disconnected from invalid user user1 117.247.111.70 port 38239 [preauth] Feb 13 19:13:47.660635 systemd[1]: sshd@16-78.46.147.231:22-117.247.111.70:38239.service: Deactivated successfully. Feb 13 19:13:47.850048 update_engine[1486]: I20250213 19:13:47.849098 1486 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Feb 13 19:13:47.850048 update_engine[1486]: I20250213 19:13:47.849547 1486 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Feb 13 19:13:47.850048 update_engine[1486]: I20250213 19:13:47.849946 1486 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Feb 13 19:13:47.851141 update_engine[1486]: E20250213 19:13:47.851087 1486 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Feb 13 19:13:47.851510 update_engine[1486]: I20250213 19:13:47.851475 1486 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded Feb 13 19:13:47.851730 update_engine[1486]: I20250213 19:13:47.851681 1486 omaha_request_action.cc:617] Omaha request response: Feb 13 19:13:47.851939 update_engine[1486]: E20250213 19:13:47.851910 1486 omaha_request_action.cc:636] Omaha request network transfer failed. Feb 13 19:13:47.852093 update_engine[1486]: I20250213 19:13:47.852041 1486 action_processor.cc:68] ActionProcessor::ActionComplete: OmahaRequestAction action failed. Aborting processing. Feb 13 19:13:47.852229 update_engine[1486]: I20250213 19:13:47.852174 1486 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Feb 13 19:13:47.852806 update_engine[1486]: I20250213 19:13:47.852304 1486 update_attempter.cc:306] Processing Done. Feb 13 19:13:47.852806 update_engine[1486]: E20250213 19:13:47.852339 1486 update_attempter.cc:619] Update failed. Feb 13 19:13:47.852806 update_engine[1486]: I20250213 19:13:47.852352 1486 utils.cc:600] Converting error code 2000 to kActionCodeOmahaErrorInHTTPResponse Feb 13 19:13:47.852806 update_engine[1486]: I20250213 19:13:47.852362 1486 payload_state.cc:97] Updating payload state for error code: 37 (kActionCodeOmahaErrorInHTTPResponse) Feb 13 19:13:47.852806 update_engine[1486]: I20250213 19:13:47.852372 1486 payload_state.cc:103] Ignoring failures until we get a valid Omaha response. Feb 13 19:13:47.852806 update_engine[1486]: I20250213 19:13:47.852475 1486 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction Feb 13 19:13:47.852806 update_engine[1486]: I20250213 19:13:47.852541 1486 omaha_request_action.cc:271] Posting an Omaha request to disabled Feb 13 19:13:47.852806 update_engine[1486]: I20250213 19:13:47.852555 1486 omaha_request_action.cc:272] Request: Feb 13 19:13:47.852806 update_engine[1486]: Feb 13 19:13:47.852806 update_engine[1486]: Feb 13 19:13:47.852806 update_engine[1486]: Feb 13 19:13:47.852806 update_engine[1486]: Feb 13 19:13:47.852806 update_engine[1486]: Feb 13 19:13:47.852806 update_engine[1486]: Feb 13 19:13:47.852806 update_engine[1486]: I20250213 19:13:47.852566 1486 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Feb 13 19:13:47.853571 locksmithd[1508]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_REPORTING_ERROR_EVENT" NewVersion=0.0.0 NewSize=0 Feb 13 19:13:47.854762 update_engine[1486]: I20250213 19:13:47.854282 1486 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Feb 13 19:13:47.854762 update_engine[1486]: I20250213 19:13:47.854702 1486 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Feb 13 19:13:47.855613 update_engine[1486]: E20250213 19:13:47.855355 1486 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Feb 13 19:13:47.855613 update_engine[1486]: I20250213 19:13:47.855401 1486 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded Feb 13 19:13:47.855613 update_engine[1486]: I20250213 19:13:47.855407 1486 omaha_request_action.cc:617] Omaha request response: Feb 13 19:13:47.855613 update_engine[1486]: I20250213 19:13:47.855413 1486 action_processor.cc:65] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Feb 13 19:13:47.855613 update_engine[1486]: I20250213 19:13:47.855417 1486 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Feb 13 19:13:47.855613 update_engine[1486]: I20250213 19:13:47.855492 1486 update_attempter.cc:306] Processing Done. Feb 13 19:13:47.855613 update_engine[1486]: I20250213 19:13:47.855500 1486 update_attempter.cc:310] Error event sent. Feb 13 19:13:47.855613 update_engine[1486]: I20250213 19:13:47.855509 1486 update_check_scheduler.cc:74] Next update check in 44m34s Feb 13 19:13:47.856262 locksmithd[1508]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_IDLE" NewVersion=0.0.0 NewSize=0 Feb 13 19:14:00.340308 systemd[1]: Started sshd@17-78.46.147.231:22-103.154.87.242:38438.service - OpenSSH per-connection server daemon (103.154.87.242:38438). Feb 13 19:14:01.364114 sshd[5990]: Invalid user admire from 103.154.87.242 port 38438 Feb 13 19:14:01.565766 sshd[5990]: Received disconnect from 103.154.87.242 port 38438:11: Bye Bye [preauth] Feb 13 19:14:01.565766 sshd[5990]: Disconnected from invalid user admire 103.154.87.242 port 38438 [preauth] Feb 13 19:14:01.569847 systemd[1]: sshd@17-78.46.147.231:22-103.154.87.242:38438.service: Deactivated successfully. Feb 13 19:14:16.273173 systemd[1]: Started sshd@18-78.46.147.231:22-159.223.215.175:48708.service - OpenSSH per-connection server daemon (159.223.215.175:48708). Feb 13 19:14:16.436289 sshd[6020]: Invalid user ambulator from 159.223.215.175 port 48708 Feb 13 19:14:16.457631 sshd[6020]: Received disconnect from 159.223.215.175 port 48708:11: Bye Bye [preauth] Feb 13 19:14:16.457631 sshd[6020]: Disconnected from invalid user ambulator 159.223.215.175 port 48708 [preauth] Feb 13 19:14:16.461442 systemd[1]: sshd@18-78.46.147.231:22-159.223.215.175:48708.service: Deactivated successfully. Feb 13 19:14:56.355407 systemd[1]: run-containerd-runc-k8s.io-2f4e0a0e052dfdd1d169970003434489eb5dfea1c524dfeb21d458ec6e8d5d49-runc.58yf4B.mount: Deactivated successfully. Feb 13 19:15:05.092391 systemd[1]: Started sshd@19-78.46.147.231:22-117.247.111.70:43989.service - OpenSSH per-connection server daemon (117.247.111.70:43989). Feb 13 19:15:06.084102 sshd[6121]: Invalid user tft105 from 117.247.111.70 port 43989 Feb 13 19:15:06.269899 sshd[6121]: Received disconnect from 117.247.111.70 port 43989:11: Bye Bye [preauth] Feb 13 19:15:06.269899 sshd[6121]: Disconnected from invalid user tft105 117.247.111.70 port 43989 [preauth] Feb 13 19:15:06.271344 systemd[1]: sshd@19-78.46.147.231:22-117.247.111.70:43989.service: Deactivated successfully. Feb 13 19:15:23.872585 systemd[1]: Started sshd@20-78.46.147.231:22-103.154.87.242:42158.service - OpenSSH per-connection server daemon (103.154.87.242:42158). Feb 13 19:15:24.920279 sshd[6169]: Invalid user indigo from 103.154.87.242 port 42158 Feb 13 19:15:25.113762 sshd[6169]: Received disconnect from 103.154.87.242 port 42158:11: Bye Bye [preauth] Feb 13 19:15:25.113762 sshd[6169]: Disconnected from invalid user indigo 103.154.87.242 port 42158 [preauth] Feb 13 19:15:25.117159 systemd[1]: sshd@20-78.46.147.231:22-103.154.87.242:42158.service: Deactivated successfully. Feb 13 19:16:18.288571 systemd[1]: Started sshd@21-78.46.147.231:22-159.223.215.175:49384.service - OpenSSH per-connection server daemon (159.223.215.175:49384). Feb 13 19:16:18.451879 sshd[6280]: Invalid user letter from 159.223.215.175 port 49384 Feb 13 19:16:18.471616 sshd[6280]: Received disconnect from 159.223.215.175 port 49384:11: Bye Bye [preauth] Feb 13 19:16:18.471616 sshd[6280]: Disconnected from invalid user letter 159.223.215.175 port 49384 [preauth] Feb 13 19:16:18.473247 systemd[1]: sshd@21-78.46.147.231:22-159.223.215.175:49384.service: Deactivated successfully. Feb 13 19:16:24.087271 systemd[1]: Started sshd@22-78.46.147.231:22-117.247.111.70:27642.service - OpenSSH per-connection server daemon (117.247.111.70:27642). Feb 13 19:16:25.085093 sshd[6308]: Invalid user rowan from 117.247.111.70 port 27642 Feb 13 19:16:25.278546 sshd[6308]: Received disconnect from 117.247.111.70 port 27642:11: Bye Bye [preauth] Feb 13 19:16:25.278546 sshd[6308]: Disconnected from invalid user rowan 117.247.111.70 port 27642 [preauth] Feb 13 19:16:25.280005 systemd[1]: sshd@22-78.46.147.231:22-117.247.111.70:27642.service: Deactivated successfully. Feb 13 19:16:25.908784 systemd[1]: Started sshd@23-78.46.147.231:22-139.178.68.195:55948.service - OpenSSH per-connection server daemon (139.178.68.195:55948). Feb 13 19:16:26.895663 sshd[6316]: Accepted publickey for core from 139.178.68.195 port 55948 ssh2: RSA SHA256:GiO19KKKK6dAgXb8V1C7vI95O6t/PswdbHT7p8WkVYc Feb 13 19:16:26.899218 sshd-session[6316]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:16:26.907305 systemd-logind[1485]: New session 8 of user core. Feb 13 19:16:26.921856 systemd[1]: Started session-8.scope - Session 8 of User core. Feb 13 19:16:27.673006 sshd[6335]: Connection closed by 139.178.68.195 port 55948 Feb 13 19:16:27.672710 sshd-session[6316]: pam_unix(sshd:session): session closed for user core Feb 13 19:16:27.677453 systemd[1]: sshd@23-78.46.147.231:22-139.178.68.195:55948.service: Deactivated successfully. Feb 13 19:16:27.679914 systemd[1]: session-8.scope: Deactivated successfully. Feb 13 19:16:27.681548 systemd-logind[1485]: Session 8 logged out. Waiting for processes to exit. Feb 13 19:16:27.683249 systemd-logind[1485]: Removed session 8. Feb 13 19:16:32.858604 systemd[1]: Started sshd@24-78.46.147.231:22-139.178.68.195:47042.service - OpenSSH per-connection server daemon (139.178.68.195:47042). Feb 13 19:16:33.843423 sshd[6348]: Accepted publickey for core from 139.178.68.195 port 47042 ssh2: RSA SHA256:GiO19KKKK6dAgXb8V1C7vI95O6t/PswdbHT7p8WkVYc Feb 13 19:16:33.845279 sshd-session[6348]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:16:33.851877 systemd-logind[1485]: New session 9 of user core. Feb 13 19:16:33.862368 systemd[1]: Started session-9.scope - Session 9 of User core. Feb 13 19:16:34.616602 sshd[6350]: Connection closed by 139.178.68.195 port 47042 Feb 13 19:16:34.617446 sshd-session[6348]: pam_unix(sshd:session): session closed for user core Feb 13 19:16:34.623793 systemd[1]: sshd@24-78.46.147.231:22-139.178.68.195:47042.service: Deactivated successfully. Feb 13 19:16:34.626441 systemd[1]: session-9.scope: Deactivated successfully. Feb 13 19:16:34.627650 systemd-logind[1485]: Session 9 logged out. Waiting for processes to exit. Feb 13 19:16:34.629539 systemd-logind[1485]: Removed session 9. Feb 13 19:16:39.797196 systemd[1]: Started sshd@25-78.46.147.231:22-139.178.68.195:42332.service - OpenSSH per-connection server daemon (139.178.68.195:42332). Feb 13 19:16:40.779932 sshd[6365]: Accepted publickey for core from 139.178.68.195 port 42332 ssh2: RSA SHA256:GiO19KKKK6dAgXb8V1C7vI95O6t/PswdbHT7p8WkVYc Feb 13 19:16:40.781226 sshd-session[6365]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:16:40.788204 systemd-logind[1485]: New session 10 of user core. Feb 13 19:16:40.793409 systemd[1]: Started session-10.scope - Session 10 of User core. Feb 13 19:16:41.542140 sshd[6367]: Connection closed by 139.178.68.195 port 42332 Feb 13 19:16:41.543076 sshd-session[6365]: pam_unix(sshd:session): session closed for user core Feb 13 19:16:41.550208 systemd-logind[1485]: Session 10 logged out. Waiting for processes to exit. Feb 13 19:16:41.550618 systemd[1]: sshd@25-78.46.147.231:22-139.178.68.195:42332.service: Deactivated successfully. Feb 13 19:16:41.553480 systemd[1]: session-10.scope: Deactivated successfully. Feb 13 19:16:41.556859 systemd-logind[1485]: Removed session 10. Feb 13 19:16:41.724830 systemd[1]: Started sshd@26-78.46.147.231:22-139.178.68.195:42342.service - OpenSSH per-connection server daemon (139.178.68.195:42342). Feb 13 19:16:42.713668 sshd[6380]: Accepted publickey for core from 139.178.68.195 port 42342 ssh2: RSA SHA256:GiO19KKKK6dAgXb8V1C7vI95O6t/PswdbHT7p8WkVYc Feb 13 19:16:42.717461 sshd-session[6380]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:16:42.727407 systemd-logind[1485]: New session 11 of user core. Feb 13 19:16:42.734409 systemd[1]: Started session-11.scope - Session 11 of User core. Feb 13 19:16:43.525471 sshd[6384]: Connection closed by 139.178.68.195 port 42342 Feb 13 19:16:43.527585 sshd-session[6380]: pam_unix(sshd:session): session closed for user core Feb 13 19:16:43.532195 systemd[1]: sshd@26-78.46.147.231:22-139.178.68.195:42342.service: Deactivated successfully. Feb 13 19:16:43.536317 systemd[1]: session-11.scope: Deactivated successfully. Feb 13 19:16:43.537540 systemd-logind[1485]: Session 11 logged out. Waiting for processes to exit. Feb 13 19:16:43.538860 systemd-logind[1485]: Removed session 11. Feb 13 19:16:43.703462 systemd[1]: Started sshd@27-78.46.147.231:22-139.178.68.195:42358.service - OpenSSH per-connection server daemon (139.178.68.195:42358). Feb 13 19:16:44.691129 sshd[6394]: Accepted publickey for core from 139.178.68.195 port 42358 ssh2: RSA SHA256:GiO19KKKK6dAgXb8V1C7vI95O6t/PswdbHT7p8WkVYc Feb 13 19:16:44.693446 sshd-session[6394]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:16:44.701648 systemd-logind[1485]: New session 12 of user core. Feb 13 19:16:44.710336 systemd[1]: Started session-12.scope - Session 12 of User core. Feb 13 19:16:45.136606 systemd[1]: Started sshd@28-78.46.147.231:22-103.154.87.242:45782.service - OpenSSH per-connection server daemon (103.154.87.242:45782). Feb 13 19:16:45.466963 sshd[6399]: Connection closed by 139.178.68.195 port 42358 Feb 13 19:16:45.467758 sshd-session[6394]: pam_unix(sshd:session): session closed for user core Feb 13 19:16:45.476555 systemd[1]: sshd@27-78.46.147.231:22-139.178.68.195:42358.service: Deactivated successfully. Feb 13 19:16:45.483873 systemd[1]: session-12.scope: Deactivated successfully. Feb 13 19:16:45.486689 systemd-logind[1485]: Session 12 logged out. Waiting for processes to exit. Feb 13 19:16:45.488017 systemd-logind[1485]: Removed session 12. Feb 13 19:16:46.404007 sshd[6401]: Invalid user netstat from 103.154.87.242 port 45782 Feb 13 19:16:46.660312 sshd[6401]: Received disconnect from 103.154.87.242 port 45782:11: Bye Bye [preauth] Feb 13 19:16:46.660312 sshd[6401]: Disconnected from invalid user netstat 103.154.87.242 port 45782 [preauth] Feb 13 19:16:46.663286 systemd[1]: sshd@28-78.46.147.231:22-103.154.87.242:45782.service: Deactivated successfully. Feb 13 19:16:50.653439 systemd[1]: Started sshd@29-78.46.147.231:22-139.178.68.195:42952.service - OpenSSH per-connection server daemon (139.178.68.195:42952). Feb 13 19:16:51.647249 sshd[6416]: Accepted publickey for core from 139.178.68.195 port 42952 ssh2: RSA SHA256:GiO19KKKK6dAgXb8V1C7vI95O6t/PswdbHT7p8WkVYc Feb 13 19:16:51.649103 sshd-session[6416]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:16:51.659684 systemd-logind[1485]: New session 13 of user core. Feb 13 19:16:51.662959 systemd[1]: Started session-13.scope - Session 13 of User core. Feb 13 19:16:52.414310 sshd[6440]: Connection closed by 139.178.68.195 port 42952 Feb 13 19:16:52.415491 sshd-session[6416]: pam_unix(sshd:session): session closed for user core Feb 13 19:16:52.420971 systemd[1]: sshd@29-78.46.147.231:22-139.178.68.195:42952.service: Deactivated successfully. Feb 13 19:16:52.424025 systemd[1]: session-13.scope: Deactivated successfully. Feb 13 19:16:52.425624 systemd-logind[1485]: Session 13 logged out. Waiting for processes to exit. Feb 13 19:16:52.426787 systemd-logind[1485]: Removed session 13. Feb 13 19:16:52.596369 systemd[1]: Started sshd@30-78.46.147.231:22-139.178.68.195:42958.service - OpenSSH per-connection server daemon (139.178.68.195:42958). Feb 13 19:16:53.605542 sshd[6452]: Accepted publickey for core from 139.178.68.195 port 42958 ssh2: RSA SHA256:GiO19KKKK6dAgXb8V1C7vI95O6t/PswdbHT7p8WkVYc Feb 13 19:16:53.607584 sshd-session[6452]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:16:53.615552 systemd-logind[1485]: New session 14 of user core. Feb 13 19:16:53.621559 systemd[1]: Started session-14.scope - Session 14 of User core. Feb 13 19:16:54.523420 sshd[6454]: Connection closed by 139.178.68.195 port 42958 Feb 13 19:16:54.524031 sshd-session[6452]: pam_unix(sshd:session): session closed for user core Feb 13 19:16:54.529417 systemd[1]: sshd@30-78.46.147.231:22-139.178.68.195:42958.service: Deactivated successfully. Feb 13 19:16:54.533313 systemd[1]: session-14.scope: Deactivated successfully. Feb 13 19:16:54.535185 systemd-logind[1485]: Session 14 logged out. Waiting for processes to exit. Feb 13 19:16:54.536829 systemd-logind[1485]: Removed session 14. Feb 13 19:16:54.699628 systemd[1]: Started sshd@31-78.46.147.231:22-139.178.68.195:42962.service - OpenSSH per-connection server daemon (139.178.68.195:42962). Feb 13 19:16:55.684594 sshd[6464]: Accepted publickey for core from 139.178.68.195 port 42962 ssh2: RSA SHA256:GiO19KKKK6dAgXb8V1C7vI95O6t/PswdbHT7p8WkVYc Feb 13 19:16:55.686555 sshd-session[6464]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:16:55.692996 systemd-logind[1485]: New session 15 of user core. Feb 13 19:16:55.695613 systemd[1]: Started session-15.scope - Session 15 of User core. Feb 13 19:16:57.433760 sshd[6466]: Connection closed by 139.178.68.195 port 42962 Feb 13 19:16:57.437085 sshd-session[6464]: pam_unix(sshd:session): session closed for user core Feb 13 19:16:57.442877 systemd[1]: sshd@31-78.46.147.231:22-139.178.68.195:42962.service: Deactivated successfully. Feb 13 19:16:57.446559 systemd[1]: session-15.scope: Deactivated successfully. Feb 13 19:16:57.449005 systemd-logind[1485]: Session 15 logged out. Waiting for processes to exit. Feb 13 19:16:57.450351 systemd-logind[1485]: Removed session 15. Feb 13 19:16:57.612652 systemd[1]: Started sshd@32-78.46.147.231:22-139.178.68.195:47966.service - OpenSSH per-connection server daemon (139.178.68.195:47966). Feb 13 19:16:58.615096 sshd[6503]: Accepted publickey for core from 139.178.68.195 port 47966 ssh2: RSA SHA256:GiO19KKKK6dAgXb8V1C7vI95O6t/PswdbHT7p8WkVYc Feb 13 19:16:58.616178 sshd-session[6503]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:16:58.625897 systemd-logind[1485]: New session 16 of user core. Feb 13 19:16:58.629862 systemd[1]: Started session-16.scope - Session 16 of User core. Feb 13 19:16:59.519602 sshd[6505]: Connection closed by 139.178.68.195 port 47966 Feb 13 19:16:59.519413 sshd-session[6503]: pam_unix(sshd:session): session closed for user core Feb 13 19:16:59.526739 systemd[1]: sshd@32-78.46.147.231:22-139.178.68.195:47966.service: Deactivated successfully. Feb 13 19:16:59.530027 systemd[1]: session-16.scope: Deactivated successfully. Feb 13 19:16:59.531362 systemd-logind[1485]: Session 16 logged out. Waiting for processes to exit. Feb 13 19:16:59.532501 systemd-logind[1485]: Removed session 16. Feb 13 19:16:59.698415 systemd[1]: Started sshd@33-78.46.147.231:22-139.178.68.195:47980.service - OpenSSH per-connection server daemon (139.178.68.195:47980). Feb 13 19:17:00.698430 sshd[6514]: Accepted publickey for core from 139.178.68.195 port 47980 ssh2: RSA SHA256:GiO19KKKK6dAgXb8V1C7vI95O6t/PswdbHT7p8WkVYc Feb 13 19:17:00.700416 sshd-session[6514]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:17:00.705919 systemd-logind[1485]: New session 17 of user core. Feb 13 19:17:00.712720 systemd[1]: Started session-17.scope - Session 17 of User core. Feb 13 19:17:01.476161 sshd[6516]: Connection closed by 139.178.68.195 port 47980 Feb 13 19:17:01.477780 sshd-session[6514]: pam_unix(sshd:session): session closed for user core Feb 13 19:17:01.484984 systemd[1]: sshd@33-78.46.147.231:22-139.178.68.195:47980.service: Deactivated successfully. Feb 13 19:17:01.490180 systemd[1]: session-17.scope: Deactivated successfully. Feb 13 19:17:01.491708 systemd-logind[1485]: Session 17 logged out. Waiting for processes to exit. Feb 13 19:17:01.492766 systemd-logind[1485]: Removed session 17. Feb 13 19:17:06.661617 systemd[1]: Started sshd@34-78.46.147.231:22-139.178.68.195:48168.service - OpenSSH per-connection server daemon (139.178.68.195:48168). Feb 13 19:17:07.664973 sshd[6561]: Accepted publickey for core from 139.178.68.195 port 48168 ssh2: RSA SHA256:GiO19KKKK6dAgXb8V1C7vI95O6t/PswdbHT7p8WkVYc Feb 13 19:17:07.666906 sshd-session[6561]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:17:07.674418 systemd-logind[1485]: New session 18 of user core. Feb 13 19:17:07.679414 systemd[1]: Started session-18.scope - Session 18 of User core. Feb 13 19:17:08.431145 sshd[6563]: Connection closed by 139.178.68.195 port 48168 Feb 13 19:17:08.432239 sshd-session[6561]: pam_unix(sshd:session): session closed for user core Feb 13 19:17:08.439000 systemd[1]: sshd@34-78.46.147.231:22-139.178.68.195:48168.service: Deactivated successfully. Feb 13 19:17:08.442197 systemd[1]: session-18.scope: Deactivated successfully. Feb 13 19:17:08.444124 systemd-logind[1485]: Session 18 logged out. Waiting for processes to exit. Feb 13 19:17:08.447241 systemd-logind[1485]: Removed session 18. Feb 13 19:17:13.608815 systemd[1]: Started sshd@35-78.46.147.231:22-139.178.68.195:48178.service - OpenSSH per-connection server daemon (139.178.68.195:48178). Feb 13 19:17:14.618315 sshd[6583]: Accepted publickey for core from 139.178.68.195 port 48178 ssh2: RSA SHA256:GiO19KKKK6dAgXb8V1C7vI95O6t/PswdbHT7p8WkVYc Feb 13 19:17:14.620828 sshd-session[6583]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:17:14.627876 systemd-logind[1485]: New session 19 of user core. Feb 13 19:17:14.636292 systemd[1]: Started session-19.scope - Session 19 of User core. Feb 13 19:17:15.381180 sshd[6585]: Connection closed by 139.178.68.195 port 48178 Feb 13 19:17:15.382739 sshd-session[6583]: pam_unix(sshd:session): session closed for user core Feb 13 19:17:15.389201 systemd-logind[1485]: Session 19 logged out. Waiting for processes to exit. Feb 13 19:17:15.390184 systemd[1]: sshd@35-78.46.147.231:22-139.178.68.195:48178.service: Deactivated successfully. Feb 13 19:17:15.393377 systemd[1]: session-19.scope: Deactivated successfully. Feb 13 19:17:15.395533 systemd-logind[1485]: Removed session 19. Feb 13 19:17:31.094283 systemd[1]: cri-containerd-3518d3dda6b4af3d0174a109e8bf22d1ebb2020a4ac0f1159e0cb3834115bfad.scope: Deactivated successfully. Feb 13 19:17:31.095401 systemd[1]: cri-containerd-3518d3dda6b4af3d0174a109e8bf22d1ebb2020a4ac0f1159e0cb3834115bfad.scope: Consumed 6.194s CPU time, 46.8M memory peak. Feb 13 19:17:31.116020 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-3518d3dda6b4af3d0174a109e8bf22d1ebb2020a4ac0f1159e0cb3834115bfad-rootfs.mount: Deactivated successfully. Feb 13 19:17:31.116969 containerd[1503]: time="2025-02-13T19:17:31.116897162Z" level=info msg="shim disconnected" id=3518d3dda6b4af3d0174a109e8bf22d1ebb2020a4ac0f1159e0cb3834115bfad namespace=k8s.io Feb 13 19:17:31.117625 containerd[1503]: time="2025-02-13T19:17:31.117450097Z" level=warning msg="cleaning up after shim disconnected" id=3518d3dda6b4af3d0174a109e8bf22d1ebb2020a4ac0f1159e0cb3834115bfad namespace=k8s.io Feb 13 19:17:31.117625 containerd[1503]: time="2025-02-13T19:17:31.117474458Z" level=info msg="cleaning up dead shim" namespace=k8s.io Feb 13 19:17:31.184007 systemd[1]: cri-containerd-b068163f576e0c62469fcae99d5b6f7785627c423b4030e9bc3bda75db16c086.scope: Deactivated successfully. Feb 13 19:17:31.184837 systemd[1]: cri-containerd-b068163f576e0c62469fcae99d5b6f7785627c423b4030e9bc3bda75db16c086.scope: Consumed 6.605s CPU time, 62.2M memory peak, 3.7M read from disk. Feb 13 19:17:31.208594 containerd[1503]: time="2025-02-13T19:17:31.208516309Z" level=info msg="shim disconnected" id=b068163f576e0c62469fcae99d5b6f7785627c423b4030e9bc3bda75db16c086 namespace=k8s.io Feb 13 19:17:31.208594 containerd[1503]: time="2025-02-13T19:17:31.208586311Z" level=warning msg="cleaning up after shim disconnected" id=b068163f576e0c62469fcae99d5b6f7785627c423b4030e9bc3bda75db16c086 namespace=k8s.io Feb 13 19:17:31.208594 containerd[1503]: time="2025-02-13T19:17:31.208595631Z" level=info msg="cleaning up dead shim" namespace=k8s.io Feb 13 19:17:31.211533 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-b068163f576e0c62469fcae99d5b6f7785627c423b4030e9bc3bda75db16c086-rootfs.mount: Deactivated successfully. Feb 13 19:17:31.250469 kubelet[2828]: E0213 19:17:31.250424 2828 controller.go:195] "Failed to update lease" err="rpc error: code = Unavailable desc = error reading from server: read tcp 10.0.0.3:44088->10.0.0.2:2379: read: connection timed out" Feb 13 19:17:31.256556 systemd[1]: cri-containerd-4331d2e00e884df3af15f071151f8b76e642f6b2922bff88e9eafc77b008d0af.scope: Deactivated successfully. Feb 13 19:17:31.256954 systemd[1]: cri-containerd-4331d2e00e884df3af15f071151f8b76e642f6b2922bff88e9eafc77b008d0af.scope: Consumed 5.959s CPU time, 26.2M memory peak, 3M read from disk. Feb 13 19:17:31.272976 kubelet[2828]: I0213 19:17:31.272916 2828 scope.go:117] "RemoveContainer" containerID="3518d3dda6b4af3d0174a109e8bf22d1ebb2020a4ac0f1159e0cb3834115bfad" Feb 13 19:17:31.279857 kubelet[2828]: I0213 19:17:31.279804 2828 scope.go:117] "RemoveContainer" containerID="b068163f576e0c62469fcae99d5b6f7785627c423b4030e9bc3bda75db16c086" Feb 13 19:17:31.281204 containerd[1503]: time="2025-02-13T19:17:31.280930267Z" level=info msg="CreateContainer within sandbox \"78caa98ae70d6324e6aa14f5bcf60257e08da3e9281b16881ed05a4718a471f5\" for container &ContainerMetadata{Name:tigera-operator,Attempt:1,}" Feb 13 19:17:31.284220 containerd[1503]: time="2025-02-13T19:17:31.283851305Z" level=info msg="CreateContainer within sandbox \"571b706a80af6df754b1195ea6008135264cf47f77378086fbd48e3e6214bc78\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:1,}" Feb 13 19:17:31.291779 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-4331d2e00e884df3af15f071151f8b76e642f6b2922bff88e9eafc77b008d0af-rootfs.mount: Deactivated successfully. Feb 13 19:17:31.294559 containerd[1503]: time="2025-02-13T19:17:31.294340143Z" level=info msg="shim disconnected" id=4331d2e00e884df3af15f071151f8b76e642f6b2922bff88e9eafc77b008d0af namespace=k8s.io Feb 13 19:17:31.294559 containerd[1503]: time="2025-02-13T19:17:31.294418425Z" level=warning msg="cleaning up after shim disconnected" id=4331d2e00e884df3af15f071151f8b76e642f6b2922bff88e9eafc77b008d0af namespace=k8s.io Feb 13 19:17:31.294559 containerd[1503]: time="2025-02-13T19:17:31.294427705Z" level=info msg="cleaning up dead shim" namespace=k8s.io Feb 13 19:17:31.319713 containerd[1503]: time="2025-02-13T19:17:31.319654773Z" level=info msg="CreateContainer within sandbox \"571b706a80af6df754b1195ea6008135264cf47f77378086fbd48e3e6214bc78\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:1,} returns container id \"b2137abf92b7bab4c2f87e6d2e166d182a1c0f9c45b61865f91c7b30f79dddb6\"" Feb 13 19:17:31.320777 containerd[1503]: time="2025-02-13T19:17:31.320719761Z" level=info msg="StartContainer for \"b2137abf92b7bab4c2f87e6d2e166d182a1c0f9c45b61865f91c7b30f79dddb6\"" Feb 13 19:17:31.329227 containerd[1503]: time="2025-02-13T19:17:31.329047782Z" level=info msg="CreateContainer within sandbox \"78caa98ae70d6324e6aa14f5bcf60257e08da3e9281b16881ed05a4718a471f5\" for &ContainerMetadata{Name:tigera-operator,Attempt:1,} returns container id \"d3a0a698ed78bb356a0b2de9c98b8b3b5cf01096f569e3e83fe323be73c7ccf0\"" Feb 13 19:17:31.330242 containerd[1503]: time="2025-02-13T19:17:31.330094690Z" level=info msg="StartContainer for \"d3a0a698ed78bb356a0b2de9c98b8b3b5cf01096f569e3e83fe323be73c7ccf0\"" Feb 13 19:17:31.373689 systemd[1]: Started cri-containerd-b2137abf92b7bab4c2f87e6d2e166d182a1c0f9c45b61865f91c7b30f79dddb6.scope - libcontainer container b2137abf92b7bab4c2f87e6d2e166d182a1c0f9c45b61865f91c7b30f79dddb6. Feb 13 19:17:31.378772 systemd[1]: Started cri-containerd-d3a0a698ed78bb356a0b2de9c98b8b3b5cf01096f569e3e83fe323be73c7ccf0.scope - libcontainer container d3a0a698ed78bb356a0b2de9c98b8b3b5cf01096f569e3e83fe323be73c7ccf0. Feb 13 19:17:31.430393 containerd[1503]: time="2025-02-13T19:17:31.429283557Z" level=info msg="StartContainer for \"d3a0a698ed78bb356a0b2de9c98b8b3b5cf01096f569e3e83fe323be73c7ccf0\" returns successfully" Feb 13 19:17:31.430711 containerd[1503]: time="2025-02-13T19:17:31.430689594Z" level=info msg="StartContainer for \"b2137abf92b7bab4c2f87e6d2e166d182a1c0f9c45b61865f91c7b30f79dddb6\" returns successfully" Feb 13 19:17:32.080007 kubelet[2828]: I0213 19:17:32.079786 2828 status_manager.go:890] "Failed to get status for pod" podUID="5feef9f1-661a-4119-bebd-9e602dcd10f8" pod="tigera-operator/tigera-operator-7d68577dc5-glhg9" err="rpc error: code = Unavailable desc = error reading from server: read tcp 10.0.0.3:44018->10.0.0.2:2379: read: connection timed out" Feb 13 19:17:32.120995 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4272407054.mount: Deactivated successfully. Feb 13 19:17:32.288737 kubelet[2828]: I0213 19:17:32.288213 2828 scope.go:117] "RemoveContainer" containerID="4331d2e00e884df3af15f071151f8b76e642f6b2922bff88e9eafc77b008d0af" Feb 13 19:17:32.292450 containerd[1503]: time="2025-02-13T19:17:32.291958415Z" level=info msg="CreateContainer within sandbox \"ae783dee3de5c947f40aea00fa3fe6c2a9cb8b418b4d592c4a4ff592e1338a88\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:1,}" Feb 13 19:17:32.318429 containerd[1503]: time="2025-02-13T19:17:32.318377596Z" level=info msg="CreateContainer within sandbox \"ae783dee3de5c947f40aea00fa3fe6c2a9cb8b418b4d592c4a4ff592e1338a88\" for &ContainerMetadata{Name:kube-scheduler,Attempt:1,} returns container id \"d803fba7c9f4f37fce9c7cdc4b8024637406dd9d88ac5fec305ee89e394235c7\"" Feb 13 19:17:32.319340 containerd[1503]: time="2025-02-13T19:17:32.318974171Z" level=info msg="StartContainer for \"d803fba7c9f4f37fce9c7cdc4b8024637406dd9d88ac5fec305ee89e394235c7\"" Feb 13 19:17:32.362248 systemd[1]: Started cri-containerd-d803fba7c9f4f37fce9c7cdc4b8024637406dd9d88ac5fec305ee89e394235c7.scope - libcontainer container d803fba7c9f4f37fce9c7cdc4b8024637406dd9d88ac5fec305ee89e394235c7. Feb 13 19:17:32.410494 containerd[1503]: time="2025-02-13T19:17:32.409966464Z" level=info msg="StartContainer for \"d803fba7c9f4f37fce9c7cdc4b8024637406dd9d88ac5fec305ee89e394235c7\" returns successfully" Feb 13 19:17:33.807692 kubelet[2828]: E0213 19:17:33.807210 2828 event.go:359] "Server rejected event (will not retry!)" err="rpc error: code = Unavailable desc = error reading from server: read tcp 10.0.0.3:43914->10.0.0.2:2379: read: connection timed out" event="&Event{ObjectMeta:{kube-apiserver-ci-4230-0-1-a-2d796e8b97.1823da9bb392ad43 kube-system 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:kube-system,Name:kube-apiserver-ci-4230-0-1-a-2d796e8b97,UID:4ee02b111fd7daa99b33a2caa1fc3e26,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Liveness probe failed: HTTP probe failed with statuscode: 500,Source:EventSource{Component:kubelet,Host:ci-4230-0-1-a-2d796e8b97,},FirstTimestamp:2025-02-13 19:17:23.362663747 +0000 UTC m=+349.061115465,LastTimestamp:2025-02-13 19:17:23.362663747 +0000 UTC m=+349.061115465,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4230-0-1-a-2d796e8b97,}"