Feb 13 19:49:55.923790 kernel: Linux version 6.6.74-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p1) 13.3.1 20240614, GNU ld (Gentoo 2.42 p6) 2.42.0) #1 SMP PREEMPT_DYNAMIC Thu Feb 13 17:44:05 -00 2025 Feb 13 19:49:55.923815 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=ed9b5d8ea73d2e47b8decea8124089e04dd398ef43013c1b1a5809314044b1c3 Feb 13 19:49:55.923826 kernel: BIOS-provided physical RAM map: Feb 13 19:49:55.923832 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable Feb 13 19:49:55.923838 kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000007fffff] usable Feb 13 19:49:55.923845 kernel: BIOS-e820: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS Feb 13 19:49:55.923852 kernel: BIOS-e820: [mem 0x0000000000808000-0x000000000080afff] usable Feb 13 19:49:55.923858 kernel: BIOS-e820: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS Feb 13 19:49:55.923864 kernel: BIOS-e820: [mem 0x000000000080c000-0x0000000000810fff] usable Feb 13 19:49:55.923870 kernel: BIOS-e820: [mem 0x0000000000811000-0x00000000008fffff] ACPI NVS Feb 13 19:49:55.923879 kernel: BIOS-e820: [mem 0x0000000000900000-0x000000009bd3efff] usable Feb 13 19:49:55.923885 kernel: BIOS-e820: [mem 0x000000009bd3f000-0x000000009bdfffff] reserved Feb 13 19:49:55.923891 kernel: BIOS-e820: [mem 0x000000009be00000-0x000000009c8ecfff] usable Feb 13 19:49:55.923897 kernel: BIOS-e820: [mem 0x000000009c8ed000-0x000000009cb6cfff] reserved Feb 13 19:49:55.923905 kernel: BIOS-e820: [mem 0x000000009cb6d000-0x000000009cb7efff] ACPI data Feb 13 19:49:55.923912 kernel: BIOS-e820: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS Feb 13 19:49:55.923921 kernel: BIOS-e820: [mem 0x000000009cbff000-0x000000009ce91fff] usable Feb 13 19:49:55.923927 kernel: BIOS-e820: [mem 0x000000009ce92000-0x000000009ce95fff] reserved Feb 13 19:49:55.923944 kernel: BIOS-e820: [mem 0x000000009ce96000-0x000000009ce97fff] ACPI NVS Feb 13 19:49:55.923950 kernel: BIOS-e820: [mem 0x000000009ce98000-0x000000009cedbfff] usable Feb 13 19:49:55.923957 kernel: BIOS-e820: [mem 0x000000009cedc000-0x000000009cf5ffff] reserved Feb 13 19:49:55.923963 kernel: BIOS-e820: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS Feb 13 19:49:55.923970 kernel: BIOS-e820: [mem 0x00000000e0000000-0x00000000efffffff] reserved Feb 13 19:49:55.923976 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Feb 13 19:49:55.923983 kernel: BIOS-e820: [mem 0x00000000ffc00000-0x00000000ffffffff] reserved Feb 13 19:49:55.923990 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Feb 13 19:49:55.923996 kernel: NX (Execute Disable) protection: active Feb 13 19:49:55.924005 kernel: APIC: Static calls initialized Feb 13 19:49:55.924012 kernel: e820: update [mem 0x9b351018-0x9b35ac57] usable ==> usable Feb 13 19:49:55.924019 kernel: e820: update [mem 0x9b351018-0x9b35ac57] usable ==> usable Feb 13 19:49:55.924025 kernel: e820: update [mem 0x9b314018-0x9b350e57] usable ==> usable Feb 13 19:49:55.924032 kernel: e820: update [mem 0x9b314018-0x9b350e57] usable ==> usable Feb 13 19:49:55.924038 kernel: extended physical RAM map: Feb 13 19:49:55.924045 kernel: reserve setup_data: [mem 0x0000000000000000-0x000000000009ffff] usable Feb 13 19:49:55.924052 kernel: reserve setup_data: [mem 0x0000000000100000-0x00000000007fffff] usable Feb 13 19:49:55.924058 kernel: reserve setup_data: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS Feb 13 19:49:55.924065 kernel: reserve setup_data: [mem 0x0000000000808000-0x000000000080afff] usable Feb 13 19:49:55.924072 kernel: reserve setup_data: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS Feb 13 19:49:55.924081 kernel: reserve setup_data: [mem 0x000000000080c000-0x0000000000810fff] usable Feb 13 19:49:55.924087 kernel: reserve setup_data: [mem 0x0000000000811000-0x00000000008fffff] ACPI NVS Feb 13 19:49:55.924099 kernel: reserve setup_data: [mem 0x0000000000900000-0x000000009b314017] usable Feb 13 19:49:55.924111 kernel: reserve setup_data: [mem 0x000000009b314018-0x000000009b350e57] usable Feb 13 19:49:55.924123 kernel: reserve setup_data: [mem 0x000000009b350e58-0x000000009b351017] usable Feb 13 19:49:55.924132 kernel: reserve setup_data: [mem 0x000000009b351018-0x000000009b35ac57] usable Feb 13 19:49:55.924141 kernel: reserve setup_data: [mem 0x000000009b35ac58-0x000000009bd3efff] usable Feb 13 19:49:55.924154 kernel: reserve setup_data: [mem 0x000000009bd3f000-0x000000009bdfffff] reserved Feb 13 19:49:55.924163 kernel: reserve setup_data: [mem 0x000000009be00000-0x000000009c8ecfff] usable Feb 13 19:49:55.924172 kernel: reserve setup_data: [mem 0x000000009c8ed000-0x000000009cb6cfff] reserved Feb 13 19:49:55.924181 kernel: reserve setup_data: [mem 0x000000009cb6d000-0x000000009cb7efff] ACPI data Feb 13 19:49:55.924190 kernel: reserve setup_data: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS Feb 13 19:49:55.924199 kernel: reserve setup_data: [mem 0x000000009cbff000-0x000000009ce91fff] usable Feb 13 19:49:55.924206 kernel: reserve setup_data: [mem 0x000000009ce92000-0x000000009ce95fff] reserved Feb 13 19:49:55.924213 kernel: reserve setup_data: [mem 0x000000009ce96000-0x000000009ce97fff] ACPI NVS Feb 13 19:49:55.924220 kernel: reserve setup_data: [mem 0x000000009ce98000-0x000000009cedbfff] usable Feb 13 19:49:55.924229 kernel: reserve setup_data: [mem 0x000000009cedc000-0x000000009cf5ffff] reserved Feb 13 19:49:55.924239 kernel: reserve setup_data: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS Feb 13 19:49:55.924248 kernel: reserve setup_data: [mem 0x00000000e0000000-0x00000000efffffff] reserved Feb 13 19:49:55.924258 kernel: reserve setup_data: [mem 0x00000000feffc000-0x00000000feffffff] reserved Feb 13 19:49:55.924267 kernel: reserve setup_data: [mem 0x00000000ffc00000-0x00000000ffffffff] reserved Feb 13 19:49:55.924277 kernel: reserve setup_data: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Feb 13 19:49:55.924285 kernel: efi: EFI v2.7 by EDK II Feb 13 19:49:55.924293 kernel: efi: SMBIOS=0x9c988000 ACPI=0x9cb7e000 ACPI 2.0=0x9cb7e014 MEMATTR=0x9ba0d198 RNG=0x9cb73018 Feb 13 19:49:55.924300 kernel: random: crng init done Feb 13 19:49:55.924307 kernel: efi: Remove mem142: MMIO range=[0xffc00000-0xffffffff] (4MB) from e820 map Feb 13 19:49:55.924314 kernel: e820: remove [mem 0xffc00000-0xffffffff] reserved Feb 13 19:49:55.924324 kernel: secureboot: Secure boot disabled Feb 13 19:49:55.924331 kernel: SMBIOS 2.8 present. Feb 13 19:49:55.924338 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS unknown 02/02/2022 Feb 13 19:49:55.924345 kernel: Hypervisor detected: KVM Feb 13 19:49:55.924352 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Feb 13 19:49:55.924359 kernel: kvm-clock: using sched offset of 2604326259 cycles Feb 13 19:49:55.924366 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Feb 13 19:49:55.924376 kernel: tsc: Detected 2794.750 MHz processor Feb 13 19:49:55.924386 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Feb 13 19:49:55.924396 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Feb 13 19:49:55.924406 kernel: last_pfn = 0x9cedc max_arch_pfn = 0x400000000 Feb 13 19:49:55.924418 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs Feb 13 19:49:55.924426 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Feb 13 19:49:55.924433 kernel: Using GB pages for direct mapping Feb 13 19:49:55.924440 kernel: ACPI: Early table checksum verification disabled Feb 13 19:49:55.924447 kernel: ACPI: RSDP 0x000000009CB7E014 000024 (v02 BOCHS ) Feb 13 19:49:55.924454 kernel: ACPI: XSDT 0x000000009CB7D0E8 000054 (v01 BOCHS BXPC 00000001 01000013) Feb 13 19:49:55.924461 kernel: ACPI: FACP 0x000000009CB79000 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Feb 13 19:49:55.924469 kernel: ACPI: DSDT 0x000000009CB7A000 0021A8 (v01 BOCHS BXPC 00000001 BXPC 00000001) Feb 13 19:49:55.924476 kernel: ACPI: FACS 0x000000009CBDD000 000040 Feb 13 19:49:55.924485 kernel: ACPI: APIC 0x000000009CB78000 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) Feb 13 19:49:55.924492 kernel: ACPI: HPET 0x000000009CB77000 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Feb 13 19:49:55.924499 kernel: ACPI: MCFG 0x000000009CB76000 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Feb 13 19:49:55.924507 kernel: ACPI: WAET 0x000000009CB75000 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Feb 13 19:49:55.924514 kernel: ACPI: BGRT 0x000000009CB74000 000038 (v01 INTEL EDK2 00000002 01000013) Feb 13 19:49:55.924521 kernel: ACPI: Reserving FACP table memory at [mem 0x9cb79000-0x9cb790f3] Feb 13 19:49:55.924528 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cb7a000-0x9cb7c1a7] Feb 13 19:49:55.924535 kernel: ACPI: Reserving FACS table memory at [mem 0x9cbdd000-0x9cbdd03f] Feb 13 19:49:55.924544 kernel: ACPI: Reserving APIC table memory at [mem 0x9cb78000-0x9cb7808f] Feb 13 19:49:55.924551 kernel: ACPI: Reserving HPET table memory at [mem 0x9cb77000-0x9cb77037] Feb 13 19:49:55.924558 kernel: ACPI: Reserving MCFG table memory at [mem 0x9cb76000-0x9cb7603b] Feb 13 19:49:55.924565 kernel: ACPI: Reserving WAET table memory at [mem 0x9cb75000-0x9cb75027] Feb 13 19:49:55.924572 kernel: ACPI: Reserving BGRT table memory at [mem 0x9cb74000-0x9cb74037] Feb 13 19:49:55.924579 kernel: No NUMA configuration found Feb 13 19:49:55.924587 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cedbfff] Feb 13 19:49:55.924597 kernel: NODE_DATA(0) allocated [mem 0x9ce3a000-0x9ce3ffff] Feb 13 19:49:55.924606 kernel: Zone ranges: Feb 13 19:49:55.924629 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Feb 13 19:49:55.924643 kernel: DMA32 [mem 0x0000000001000000-0x000000009cedbfff] Feb 13 19:49:55.924653 kernel: Normal empty Feb 13 19:49:55.924663 kernel: Movable zone start for each node Feb 13 19:49:55.924670 kernel: Early memory node ranges Feb 13 19:49:55.924677 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] Feb 13 19:49:55.924684 kernel: node 0: [mem 0x0000000000100000-0x00000000007fffff] Feb 13 19:49:55.924691 kernel: node 0: [mem 0x0000000000808000-0x000000000080afff] Feb 13 19:49:55.924698 kernel: node 0: [mem 0x000000000080c000-0x0000000000810fff] Feb 13 19:49:55.924705 kernel: node 0: [mem 0x0000000000900000-0x000000009bd3efff] Feb 13 19:49:55.924715 kernel: node 0: [mem 0x000000009be00000-0x000000009c8ecfff] Feb 13 19:49:55.924722 kernel: node 0: [mem 0x000000009cbff000-0x000000009ce91fff] Feb 13 19:49:55.924729 kernel: node 0: [mem 0x000000009ce98000-0x000000009cedbfff] Feb 13 19:49:55.924736 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cedbfff] Feb 13 19:49:55.924743 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Feb 13 19:49:55.924750 kernel: On node 0, zone DMA: 96 pages in unavailable ranges Feb 13 19:49:55.924766 kernel: On node 0, zone DMA: 8 pages in unavailable ranges Feb 13 19:49:55.924786 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Feb 13 19:49:55.924796 kernel: On node 0, zone DMA: 239 pages in unavailable ranges Feb 13 19:49:55.924803 kernel: On node 0, zone DMA32: 193 pages in unavailable ranges Feb 13 19:49:55.924811 kernel: On node 0, zone DMA32: 786 pages in unavailable ranges Feb 13 19:49:55.924818 kernel: On node 0, zone DMA32: 6 pages in unavailable ranges Feb 13 19:49:55.924829 kernel: On node 0, zone DMA32: 12580 pages in unavailable ranges Feb 13 19:49:55.924836 kernel: ACPI: PM-Timer IO Port: 0x608 Feb 13 19:49:55.924844 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Feb 13 19:49:55.924851 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Feb 13 19:49:55.924859 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Feb 13 19:49:55.924868 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Feb 13 19:49:55.924877 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Feb 13 19:49:55.924886 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Feb 13 19:49:55.924895 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Feb 13 19:49:55.924907 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Feb 13 19:49:55.924922 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Feb 13 19:49:55.924945 kernel: TSC deadline timer available Feb 13 19:49:55.924954 kernel: smpboot: Allowing 4 CPUs, 0 hotplug CPUs Feb 13 19:49:55.924964 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Feb 13 19:49:55.924977 kernel: kvm-guest: KVM setup pv remote TLB flush Feb 13 19:49:55.924987 kernel: kvm-guest: setup PV sched yield Feb 13 19:49:55.924997 kernel: [mem 0x9d000000-0xdfffffff] available for PCI devices Feb 13 19:49:55.925006 kernel: Booting paravirtualized kernel on KVM Feb 13 19:49:55.925016 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Feb 13 19:49:55.925026 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:4 nr_cpu_ids:4 nr_node_ids:1 Feb 13 19:49:55.925035 kernel: percpu: Embedded 58 pages/cpu s197032 r8192 d32344 u524288 Feb 13 19:49:55.925042 kernel: pcpu-alloc: s197032 r8192 d32344 u524288 alloc=1*2097152 Feb 13 19:49:55.925050 kernel: pcpu-alloc: [0] 0 1 2 3 Feb 13 19:49:55.925059 kernel: kvm-guest: PV spinlocks enabled Feb 13 19:49:55.925067 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Feb 13 19:49:55.925076 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=ed9b5d8ea73d2e47b8decea8124089e04dd398ef43013c1b1a5809314044b1c3 Feb 13 19:49:55.925084 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Feb 13 19:49:55.925093 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Feb 13 19:49:55.925103 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Feb 13 19:49:55.925111 kernel: Fallback order for Node 0: 0 Feb 13 19:49:55.925120 kernel: Built 1 zonelists, mobility grouping on. Total pages: 629460 Feb 13 19:49:55.925133 kernel: Policy zone: DMA32 Feb 13 19:49:55.925143 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Feb 13 19:49:55.925154 kernel: Memory: 2389768K/2565800K available (12288K kernel code, 2301K rwdata, 22736K rodata, 42976K init, 2216K bss, 175776K reserved, 0K cma-reserved) Feb 13 19:49:55.925164 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Feb 13 19:49:55.925174 kernel: ftrace: allocating 37923 entries in 149 pages Feb 13 19:49:55.925184 kernel: ftrace: allocated 149 pages with 4 groups Feb 13 19:49:55.925193 kernel: Dynamic Preempt: voluntary Feb 13 19:49:55.925200 kernel: rcu: Preemptible hierarchical RCU implementation. Feb 13 19:49:55.925208 kernel: rcu: RCU event tracing is enabled. Feb 13 19:49:55.925219 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Feb 13 19:49:55.925226 kernel: Trampoline variant of Tasks RCU enabled. Feb 13 19:49:55.925234 kernel: Rude variant of Tasks RCU enabled. Feb 13 19:49:55.925241 kernel: Tracing variant of Tasks RCU enabled. Feb 13 19:49:55.925249 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Feb 13 19:49:55.925256 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Feb 13 19:49:55.925264 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 Feb 13 19:49:55.925271 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Feb 13 19:49:55.925279 kernel: Console: colour dummy device 80x25 Feb 13 19:49:55.925288 kernel: printk: console [ttyS0] enabled Feb 13 19:49:55.925298 kernel: ACPI: Core revision 20230628 Feb 13 19:49:55.925308 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Feb 13 19:49:55.925319 kernel: APIC: Switch to symmetric I/O mode setup Feb 13 19:49:55.925328 kernel: x2apic enabled Feb 13 19:49:55.925335 kernel: APIC: Switched APIC routing to: physical x2apic Feb 13 19:49:55.925344 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Feb 13 19:49:55.925354 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Feb 13 19:49:55.925363 kernel: kvm-guest: setup PV IPIs Feb 13 19:49:55.925373 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Feb 13 19:49:55.925383 kernel: tsc: Marking TSC unstable due to TSCs unsynchronized Feb 13 19:49:55.925391 kernel: Calibrating delay loop (skipped) preset value.. 5589.50 BogoMIPS (lpj=2794750) Feb 13 19:49:55.925398 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Feb 13 19:49:55.925406 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Feb 13 19:49:55.925414 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Feb 13 19:49:55.925673 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Feb 13 19:49:55.925685 kernel: Spectre V2 : Mitigation: Retpolines Feb 13 19:49:55.925692 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch Feb 13 19:49:55.925700 kernel: Spectre V2 : Spectre v2 / SpectreRSB : Filling RSB on VMEXIT Feb 13 19:49:55.925711 kernel: Spectre V2 : Enabling Speculation Barrier for firmware calls Feb 13 19:49:55.925718 kernel: RETBleed: Mitigation: untrained return thunk Feb 13 19:49:55.925726 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Feb 13 19:49:55.925734 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Feb 13 19:49:55.925741 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! Feb 13 19:49:55.925749 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. Feb 13 19:49:55.925757 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode Feb 13 19:49:55.925765 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Feb 13 19:49:55.925779 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Feb 13 19:49:55.925789 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Feb 13 19:49:55.925799 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Feb 13 19:49:55.925807 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Feb 13 19:49:55.925815 kernel: Freeing SMP alternatives memory: 32K Feb 13 19:49:55.925822 kernel: pid_max: default: 32768 minimum: 301 Feb 13 19:49:55.925830 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Feb 13 19:49:55.925837 kernel: landlock: Up and running. Feb 13 19:49:55.925845 kernel: SELinux: Initializing. Feb 13 19:49:55.925855 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Feb 13 19:49:55.925863 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Feb 13 19:49:55.925870 kernel: smpboot: CPU0: AMD EPYC 7402P 24-Core Processor (family: 0x17, model: 0x31, stepping: 0x0) Feb 13 19:49:55.925878 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Feb 13 19:49:55.925886 kernel: RCU Tasks Rude: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Feb 13 19:49:55.925893 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Feb 13 19:49:55.925901 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. Feb 13 19:49:55.925908 kernel: ... version: 0 Feb 13 19:49:55.925915 kernel: ... bit width: 48 Feb 13 19:49:55.925925 kernel: ... generic registers: 6 Feb 13 19:49:55.925940 kernel: ... value mask: 0000ffffffffffff Feb 13 19:49:55.925948 kernel: ... max period: 00007fffffffffff Feb 13 19:49:55.925956 kernel: ... fixed-purpose events: 0 Feb 13 19:49:55.925963 kernel: ... event mask: 000000000000003f Feb 13 19:49:55.925971 kernel: signal: max sigframe size: 1776 Feb 13 19:49:55.925978 kernel: rcu: Hierarchical SRCU implementation. Feb 13 19:49:55.925986 kernel: rcu: Max phase no-delay instances is 400. Feb 13 19:49:55.925993 kernel: smp: Bringing up secondary CPUs ... Feb 13 19:49:55.926004 kernel: smpboot: x86: Booting SMP configuration: Feb 13 19:49:55.926011 kernel: .... node #0, CPUs: #1 #2 #3 Feb 13 19:49:55.926018 kernel: smp: Brought up 1 node, 4 CPUs Feb 13 19:49:55.926026 kernel: smpboot: Max logical packages: 1 Feb 13 19:49:55.926033 kernel: smpboot: Total of 4 processors activated (22358.00 BogoMIPS) Feb 13 19:49:55.926050 kernel: devtmpfs: initialized Feb 13 19:49:55.926058 kernel: x86/mm: Memory block size: 128MB Feb 13 19:49:55.926065 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00800000-0x00807fff] (32768 bytes) Feb 13 19:49:55.926073 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x0080b000-0x0080bfff] (4096 bytes) Feb 13 19:49:55.926092 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00811000-0x008fffff] (978944 bytes) Feb 13 19:49:55.926113 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cb7f000-0x9cbfefff] (524288 bytes) Feb 13 19:49:55.926122 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9ce96000-0x9ce97fff] (8192 bytes) Feb 13 19:49:55.926137 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cf60000-0x9cffffff] (655360 bytes) Feb 13 19:49:55.926158 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Feb 13 19:49:55.926173 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Feb 13 19:49:55.926181 kernel: pinctrl core: initialized pinctrl subsystem Feb 13 19:49:55.926201 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Feb 13 19:49:55.926210 kernel: audit: initializing netlink subsys (disabled) Feb 13 19:49:55.926220 kernel: audit: type=2000 audit(1739476195.852:1): state=initialized audit_enabled=0 res=1 Feb 13 19:49:55.926227 kernel: thermal_sys: Registered thermal governor 'step_wise' Feb 13 19:49:55.926234 kernel: thermal_sys: Registered thermal governor 'user_space' Feb 13 19:49:55.926242 kernel: cpuidle: using governor menu Feb 13 19:49:55.926249 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Feb 13 19:49:55.926256 kernel: dca service started, version 1.12.1 Feb 13 19:49:55.926264 kernel: PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xe0000000-0xefffffff] (base 0xe0000000) Feb 13 19:49:55.926271 kernel: PCI: Using configuration type 1 for base access Feb 13 19:49:55.926279 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Feb 13 19:49:55.926289 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Feb 13 19:49:55.926296 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Feb 13 19:49:55.926303 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Feb 13 19:49:55.926311 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Feb 13 19:49:55.926318 kernel: ACPI: Added _OSI(Module Device) Feb 13 19:49:55.926325 kernel: ACPI: Added _OSI(Processor Device) Feb 13 19:49:55.926333 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Feb 13 19:49:55.926340 kernel: ACPI: Added _OSI(Processor Aggregator Device) Feb 13 19:49:55.926348 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Feb 13 19:49:55.926357 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Feb 13 19:49:55.926365 kernel: ACPI: Interpreter enabled Feb 13 19:49:55.926372 kernel: ACPI: PM: (supports S0 S3 S5) Feb 13 19:49:55.926379 kernel: ACPI: Using IOAPIC for interrupt routing Feb 13 19:49:55.926387 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Feb 13 19:49:55.926395 kernel: PCI: Using E820 reservations for host bridge windows Feb 13 19:49:55.926402 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Feb 13 19:49:55.926409 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Feb 13 19:49:55.926593 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Feb 13 19:49:55.926766 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Feb 13 19:49:55.926903 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Feb 13 19:49:55.926914 kernel: PCI host bridge to bus 0000:00 Feb 13 19:49:55.927056 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Feb 13 19:49:55.927178 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Feb 13 19:49:55.927297 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Feb 13 19:49:55.927511 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xdfffffff window] Feb 13 19:49:55.927682 kernel: pci_bus 0000:00: root bus resource [mem 0xf0000000-0xfebfffff window] Feb 13 19:49:55.927835 kernel: pci_bus 0000:00: root bus resource [mem 0x380000000000-0x3807ffffffff window] Feb 13 19:49:55.927970 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Feb 13 19:49:55.928132 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 Feb 13 19:49:55.928275 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 Feb 13 19:49:55.928444 kernel: pci 0000:00:01.0: reg 0x10: [mem 0xc0000000-0xc0ffffff pref] Feb 13 19:49:55.928606 kernel: pci 0000:00:01.0: reg 0x18: [mem 0xc1044000-0xc1044fff] Feb 13 19:49:55.928788 kernel: pci 0000:00:01.0: reg 0x30: [mem 0xffff0000-0xffffffff pref] Feb 13 19:49:55.929003 kernel: pci 0000:00:01.0: BAR 0: assigned to efifb Feb 13 19:49:55.929176 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Feb 13 19:49:55.929340 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 Feb 13 19:49:55.929471 kernel: pci 0000:00:02.0: reg 0x10: [io 0x6100-0x611f] Feb 13 19:49:55.929665 kernel: pci 0000:00:02.0: reg 0x14: [mem 0xc1043000-0xc1043fff] Feb 13 19:49:55.929819 kernel: pci 0000:00:02.0: reg 0x20: [mem 0x380000000000-0x380000003fff 64bit pref] Feb 13 19:49:55.930024 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 Feb 13 19:49:55.930167 kernel: pci 0000:00:03.0: reg 0x10: [io 0x6000-0x607f] Feb 13 19:49:55.930320 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xc1042000-0xc1042fff] Feb 13 19:49:55.930453 kernel: pci 0000:00:03.0: reg 0x20: [mem 0x380000004000-0x380000007fff 64bit pref] Feb 13 19:49:55.930637 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 Feb 13 19:49:55.930804 kernel: pci 0000:00:04.0: reg 0x10: [io 0x60e0-0x60ff] Feb 13 19:49:55.930966 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xc1041000-0xc1041fff] Feb 13 19:49:55.931115 kernel: pci 0000:00:04.0: reg 0x20: [mem 0x380000008000-0x38000000bfff 64bit pref] Feb 13 19:49:55.931264 kernel: pci 0000:00:04.0: reg 0x30: [mem 0xfffc0000-0xffffffff pref] Feb 13 19:49:55.931445 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 Feb 13 19:49:55.931602 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Feb 13 19:49:55.931821 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 Feb 13 19:49:55.931998 kernel: pci 0000:00:1f.2: reg 0x20: [io 0x60c0-0x60df] Feb 13 19:49:55.932171 kernel: pci 0000:00:1f.2: reg 0x24: [mem 0xc1040000-0xc1040fff] Feb 13 19:49:55.932339 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 Feb 13 19:49:55.932495 kernel: pci 0000:00:1f.3: reg 0x20: [io 0x6080-0x60bf] Feb 13 19:49:55.932512 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Feb 13 19:49:55.932521 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Feb 13 19:49:55.932531 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Feb 13 19:49:55.932546 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Feb 13 19:49:55.932557 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Feb 13 19:49:55.932567 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Feb 13 19:49:55.932578 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Feb 13 19:49:55.932587 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Feb 13 19:49:55.932595 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Feb 13 19:49:55.932603 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Feb 13 19:49:55.932610 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Feb 13 19:49:55.932631 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Feb 13 19:49:55.932642 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Feb 13 19:49:55.932650 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Feb 13 19:49:55.932658 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Feb 13 19:49:55.932666 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Feb 13 19:49:55.932673 kernel: iommu: Default domain type: Translated Feb 13 19:49:55.932681 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Feb 13 19:49:55.932688 kernel: efivars: Registered efivars operations Feb 13 19:49:55.932696 kernel: PCI: Using ACPI for IRQ routing Feb 13 19:49:55.932703 kernel: PCI: pci_cache_line_size set to 64 bytes Feb 13 19:49:55.932713 kernel: e820: reserve RAM buffer [mem 0x0080b000-0x008fffff] Feb 13 19:49:55.932721 kernel: e820: reserve RAM buffer [mem 0x00811000-0x008fffff] Feb 13 19:49:55.932728 kernel: e820: reserve RAM buffer [mem 0x9b314018-0x9bffffff] Feb 13 19:49:55.932736 kernel: e820: reserve RAM buffer [mem 0x9b351018-0x9bffffff] Feb 13 19:49:55.932743 kernel: e820: reserve RAM buffer [mem 0x9bd3f000-0x9bffffff] Feb 13 19:49:55.932751 kernel: e820: reserve RAM buffer [mem 0x9c8ed000-0x9fffffff] Feb 13 19:49:55.932759 kernel: e820: reserve RAM buffer [mem 0x9ce92000-0x9fffffff] Feb 13 19:49:55.932766 kernel: e820: reserve RAM buffer [mem 0x9cedc000-0x9fffffff] Feb 13 19:49:55.932893 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Feb 13 19:49:55.933028 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Feb 13 19:49:55.933149 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Feb 13 19:49:55.933159 kernel: vgaarb: loaded Feb 13 19:49:55.933167 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Feb 13 19:49:55.933175 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Feb 13 19:49:55.933183 kernel: clocksource: Switched to clocksource kvm-clock Feb 13 19:49:55.933190 kernel: VFS: Disk quotas dquot_6.6.0 Feb 13 19:49:55.933198 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Feb 13 19:49:55.933209 kernel: pnp: PnP ACPI init Feb 13 19:49:55.933338 kernel: system 00:05: [mem 0xe0000000-0xefffffff window] has been reserved Feb 13 19:49:55.933349 kernel: pnp: PnP ACPI: found 6 devices Feb 13 19:49:55.933357 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Feb 13 19:49:55.933365 kernel: NET: Registered PF_INET protocol family Feb 13 19:49:55.933392 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Feb 13 19:49:55.933402 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Feb 13 19:49:55.933410 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Feb 13 19:49:55.933420 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Feb 13 19:49:55.933428 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Feb 13 19:49:55.933436 kernel: TCP: Hash tables configured (established 32768 bind 32768) Feb 13 19:49:55.933444 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Feb 13 19:49:55.933452 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Feb 13 19:49:55.933459 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Feb 13 19:49:55.933467 kernel: NET: Registered PF_XDP protocol family Feb 13 19:49:55.933589 kernel: pci 0000:00:04.0: can't claim BAR 6 [mem 0xfffc0000-0xffffffff pref]: no compatible bridge window Feb 13 19:49:55.933724 kernel: pci 0000:00:04.0: BAR 6: assigned [mem 0x9d000000-0x9d03ffff pref] Feb 13 19:49:55.933843 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Feb 13 19:49:55.933969 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Feb 13 19:49:55.934080 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Feb 13 19:49:55.934190 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xdfffffff window] Feb 13 19:49:55.934299 kernel: pci_bus 0000:00: resource 8 [mem 0xf0000000-0xfebfffff window] Feb 13 19:49:55.934418 kernel: pci_bus 0000:00: resource 9 [mem 0x380000000000-0x3807ffffffff window] Feb 13 19:49:55.934429 kernel: PCI: CLS 0 bytes, default 64 Feb 13 19:49:55.934437 kernel: Initialise system trusted keyrings Feb 13 19:49:55.934449 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Feb 13 19:49:55.934457 kernel: Key type asymmetric registered Feb 13 19:49:55.934465 kernel: Asymmetric key parser 'x509' registered Feb 13 19:49:55.934473 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Feb 13 19:49:55.934481 kernel: io scheduler mq-deadline registered Feb 13 19:49:55.934488 kernel: io scheduler kyber registered Feb 13 19:49:55.934496 kernel: io scheduler bfq registered Feb 13 19:49:55.934504 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Feb 13 19:49:55.934512 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Feb 13 19:49:55.934523 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Feb 13 19:49:55.934533 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Feb 13 19:49:55.934541 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Feb 13 19:49:55.934549 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Feb 13 19:49:55.934557 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Feb 13 19:49:55.934565 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Feb 13 19:49:55.934575 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Feb 13 19:49:55.934714 kernel: rtc_cmos 00:04: RTC can wake from S4 Feb 13 19:49:55.934726 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Feb 13 19:49:55.934842 kernel: rtc_cmos 00:04: registered as rtc0 Feb 13 19:49:55.934968 kernel: rtc_cmos 00:04: setting system clock to 2025-02-13T19:49:55 UTC (1739476195) Feb 13 19:49:55.935081 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram Feb 13 19:49:55.935091 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Feb 13 19:49:55.935102 kernel: efifb: probing for efifb Feb 13 19:49:55.935110 kernel: efifb: framebuffer at 0xc0000000, using 4000k, total 4000k Feb 13 19:49:55.935118 kernel: efifb: mode is 1280x800x32, linelength=5120, pages=1 Feb 13 19:49:55.935126 kernel: efifb: scrolling: redraw Feb 13 19:49:55.935134 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Feb 13 19:49:55.935142 kernel: Console: switching to colour frame buffer device 160x50 Feb 13 19:49:55.935150 kernel: fb0: EFI VGA frame buffer device Feb 13 19:49:55.935158 kernel: pstore: Using crash dump compression: deflate Feb 13 19:49:55.935167 kernel: pstore: Registered efi_pstore as persistent store backend Feb 13 19:49:55.935177 kernel: NET: Registered PF_INET6 protocol family Feb 13 19:49:55.935185 kernel: Segment Routing with IPv6 Feb 13 19:49:55.935193 kernel: In-situ OAM (IOAM) with IPv6 Feb 13 19:49:55.935201 kernel: NET: Registered PF_PACKET protocol family Feb 13 19:49:55.935209 kernel: Key type dns_resolver registered Feb 13 19:49:55.935216 kernel: IPI shorthand broadcast: enabled Feb 13 19:49:55.935224 kernel: sched_clock: Marking stable (689002720, 161023875)->(878694091, -28667496) Feb 13 19:49:55.935232 kernel: registered taskstats version 1 Feb 13 19:49:55.935240 kernel: Loading compiled-in X.509 certificates Feb 13 19:49:55.935248 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.74-flatcar: 0cc219a306b9e46e583adebba1820decbdc4307b' Feb 13 19:49:55.935259 kernel: Key type .fscrypt registered Feb 13 19:49:55.935266 kernel: Key type fscrypt-provisioning registered Feb 13 19:49:55.935274 kernel: ima: No TPM chip found, activating TPM-bypass! Feb 13 19:49:55.935282 kernel: ima: Allocated hash algorithm: sha1 Feb 13 19:49:55.935291 kernel: ima: No architecture policies found Feb 13 19:49:55.935298 kernel: clk: Disabling unused clocks Feb 13 19:49:55.935306 kernel: Freeing unused kernel image (initmem) memory: 42976K Feb 13 19:49:55.935314 kernel: Write protecting the kernel read-only data: 36864k Feb 13 19:49:55.935325 kernel: Freeing unused kernel image (rodata/data gap) memory: 1840K Feb 13 19:49:55.935333 kernel: Run /init as init process Feb 13 19:49:55.935341 kernel: with arguments: Feb 13 19:49:55.935349 kernel: /init Feb 13 19:49:55.935357 kernel: with environment: Feb 13 19:49:55.935364 kernel: HOME=/ Feb 13 19:49:55.935372 kernel: TERM=linux Feb 13 19:49:55.935380 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Feb 13 19:49:55.935390 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Feb 13 19:49:55.935403 systemd[1]: Detected virtualization kvm. Feb 13 19:49:55.935411 systemd[1]: Detected architecture x86-64. Feb 13 19:49:55.935420 systemd[1]: Running in initrd. Feb 13 19:49:55.935428 systemd[1]: No hostname configured, using default hostname. Feb 13 19:49:55.935436 systemd[1]: Hostname set to . Feb 13 19:49:55.935445 systemd[1]: Initializing machine ID from VM UUID. Feb 13 19:49:55.935453 systemd[1]: Queued start job for default target initrd.target. Feb 13 19:49:55.935463 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Feb 13 19:49:55.935472 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Feb 13 19:49:55.935481 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Feb 13 19:49:55.935490 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Feb 13 19:49:55.935498 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Feb 13 19:49:55.935507 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Feb 13 19:49:55.935517 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Feb 13 19:49:55.935527 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Feb 13 19:49:55.935536 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Feb 13 19:49:55.935544 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Feb 13 19:49:55.935552 systemd[1]: Reached target paths.target - Path Units. Feb 13 19:49:55.935561 systemd[1]: Reached target slices.target - Slice Units. Feb 13 19:49:55.935569 systemd[1]: Reached target swap.target - Swaps. Feb 13 19:49:55.935577 systemd[1]: Reached target timers.target - Timer Units. Feb 13 19:49:55.935586 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Feb 13 19:49:55.935594 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Feb 13 19:49:55.935605 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Feb 13 19:49:55.935674 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Feb 13 19:49:55.935684 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Feb 13 19:49:55.935693 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Feb 13 19:49:55.935704 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Feb 13 19:49:55.935714 systemd[1]: Reached target sockets.target - Socket Units. Feb 13 19:49:55.935725 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Feb 13 19:49:55.935736 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Feb 13 19:49:55.935748 systemd[1]: Finished network-cleanup.service - Network Cleanup. Feb 13 19:49:55.935756 systemd[1]: Starting systemd-fsck-usr.service... Feb 13 19:49:55.935765 systemd[1]: Starting systemd-journald.service - Journal Service... Feb 13 19:49:55.935773 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Feb 13 19:49:55.935781 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 19:49:55.935790 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Feb 13 19:49:55.935798 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Feb 13 19:49:55.935825 systemd-journald[194]: Collecting audit messages is disabled. Feb 13 19:49:55.935848 systemd[1]: Finished systemd-fsck-usr.service. Feb 13 19:49:55.935857 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Feb 13 19:49:55.935868 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 19:49:55.935877 systemd-journald[194]: Journal started Feb 13 19:49:55.935894 systemd-journald[194]: Runtime Journal (/run/log/journal/746b73c0bb8e4cadb9b74db0f0f2dea7) is 6.0M, max 48.3M, 42.2M free. Feb 13 19:49:55.927786 systemd-modules-load[195]: Inserted module 'overlay' Feb 13 19:49:55.938926 systemd[1]: Started systemd-journald.service - Journal Service. Feb 13 19:49:55.943443 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Feb 13 19:49:55.946425 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Feb 13 19:49:55.949415 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Feb 13 19:49:55.953570 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Feb 13 19:49:55.958650 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Feb 13 19:49:55.962853 kernel: Bridge firewalling registered Feb 13 19:49:55.961991 systemd-modules-load[195]: Inserted module 'br_netfilter' Feb 13 19:49:55.964395 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Feb 13 19:49:55.965914 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Feb 13 19:49:55.968097 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Feb 13 19:49:55.971551 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 19:49:55.974388 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Feb 13 19:49:55.978290 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Feb 13 19:49:55.982513 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Feb 13 19:49:55.984063 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Feb 13 19:49:55.999590 dracut-cmdline[226]: dracut-dracut-053 Feb 13 19:49:56.002815 dracut-cmdline[226]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=ed9b5d8ea73d2e47b8decea8124089e04dd398ef43013c1b1a5809314044b1c3 Feb 13 19:49:56.024593 systemd-resolved[229]: Positive Trust Anchors: Feb 13 19:49:56.024623 systemd-resolved[229]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Feb 13 19:49:56.024656 systemd-resolved[229]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Feb 13 19:49:56.027133 systemd-resolved[229]: Defaulting to hostname 'linux'. Feb 13 19:49:56.028175 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Feb 13 19:49:56.037017 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Feb 13 19:49:56.089665 kernel: SCSI subsystem initialized Feb 13 19:49:56.098659 kernel: Loading iSCSI transport class v2.0-870. Feb 13 19:49:56.109642 kernel: iscsi: registered transport (tcp) Feb 13 19:49:56.129947 kernel: iscsi: registered transport (qla4xxx) Feb 13 19:49:56.129971 kernel: QLogic iSCSI HBA Driver Feb 13 19:49:56.176577 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Feb 13 19:49:56.183816 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Feb 13 19:49:56.206933 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Feb 13 19:49:56.206982 kernel: device-mapper: uevent: version 1.0.3 Feb 13 19:49:56.207959 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Feb 13 19:49:56.248654 kernel: raid6: avx2x4 gen() 23887 MB/s Feb 13 19:49:56.265640 kernel: raid6: avx2x2 gen() 26973 MB/s Feb 13 19:49:56.282721 kernel: raid6: avx2x1 gen() 25949 MB/s Feb 13 19:49:56.282766 kernel: raid6: using algorithm avx2x2 gen() 26973 MB/s Feb 13 19:49:56.300712 kernel: raid6: .... xor() 19783 MB/s, rmw enabled Feb 13 19:49:56.300734 kernel: raid6: using avx2x2 recovery algorithm Feb 13 19:49:56.322639 kernel: xor: automatically using best checksumming function avx Feb 13 19:49:56.472643 kernel: Btrfs loaded, zoned=no, fsverity=no Feb 13 19:49:56.484006 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Feb 13 19:49:56.490790 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Feb 13 19:49:56.502689 systemd-udevd[412]: Using default interface naming scheme 'v255'. Feb 13 19:49:56.507280 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Feb 13 19:49:56.513739 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Feb 13 19:49:56.525201 dracut-pre-trigger[416]: rd.md=0: removing MD RAID activation Feb 13 19:49:56.552554 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Feb 13 19:49:56.563802 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Feb 13 19:49:56.624997 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Feb 13 19:49:56.632014 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Feb 13 19:49:56.646383 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Feb 13 19:49:56.649366 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Feb 13 19:49:56.653005 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Feb 13 19:49:56.655262 systemd[1]: Reached target remote-fs.target - Remote File Systems. Feb 13 19:49:56.662928 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Feb 13 19:49:56.673003 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Feb 13 19:49:56.679359 kernel: libata version 3.00 loaded. Feb 13 19:49:56.682845 kernel: virtio_blk virtio1: 4/0/0 default/read/poll queues Feb 13 19:49:56.711121 kernel: cryptd: max_cpu_qlen set to 1000 Feb 13 19:49:56.711139 kernel: ahci 0000:00:1f.2: version 3.0 Feb 13 19:49:56.720198 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Feb 13 19:49:56.720222 kernel: ahci 0000:00:1f.2: AHCI 0001.0000 32 slots 6 ports 1.5 Gbps 0x3f impl SATA mode Feb 13 19:49:56.720376 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Feb 13 19:49:56.720670 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Feb 13 19:49:56.721158 kernel: AVX2 version of gcm_enc/dec engaged. Feb 13 19:49:56.721170 kernel: AES CTR mode by8 optimization enabled Feb 13 19:49:56.721185 kernel: scsi host0: ahci Feb 13 19:49:56.721341 kernel: scsi host1: ahci Feb 13 19:49:56.721498 kernel: scsi host2: ahci Feb 13 19:49:56.721673 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Feb 13 19:49:56.721685 kernel: GPT:9289727 != 19775487 Feb 13 19:49:56.721711 kernel: GPT:Alternate GPT header not at the end of the disk. Feb 13 19:49:56.721723 kernel: GPT:9289727 != 19775487 Feb 13 19:49:56.721733 kernel: GPT: Use GNU Parted to correct GPT errors. Feb 13 19:49:56.721743 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Feb 13 19:49:56.721754 kernel: scsi host3: ahci Feb 13 19:49:56.721919 kernel: scsi host4: ahci Feb 13 19:49:56.722066 kernel: scsi host5: ahci Feb 13 19:49:56.722206 kernel: ata1: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040100 irq 31 Feb 13 19:49:56.722217 kernel: ata2: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040180 irq 31 Feb 13 19:49:56.722227 kernel: ata3: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040200 irq 31 Feb 13 19:49:56.722237 kernel: ata4: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040280 irq 31 Feb 13 19:49:56.722247 kernel: ata5: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040300 irq 31 Feb 13 19:49:56.722260 kernel: ata6: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040380 irq 31 Feb 13 19:49:56.700314 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Feb 13 19:49:56.700865 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 19:49:56.731719 kernel: BTRFS: device fsid e9c87d9f-3864-4b45-9be4-80a5397f1fc6 devid 1 transid 38 /dev/vda3 scanned by (udev-worker) (455) Feb 13 19:49:56.702548 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Feb 13 19:49:56.704695 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Feb 13 19:49:56.704843 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 19:49:56.709118 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 19:49:56.724006 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 19:49:56.737640 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/vda6 scanned by (udev-worker) (470) Feb 13 19:49:56.748459 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Feb 13 19:49:56.756188 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Feb 13 19:49:56.764152 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Feb 13 19:49:56.764229 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Feb 13 19:49:56.774128 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Feb 13 19:49:56.786821 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Feb 13 19:49:56.788021 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Feb 13 19:49:56.788100 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 19:49:56.790452 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 19:49:56.793428 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 19:49:56.810405 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 19:49:56.821800 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Feb 13 19:49:56.825362 disk-uuid[555]: Primary Header is updated. Feb 13 19:49:56.825362 disk-uuid[555]: Secondary Entries is updated. Feb 13 19:49:56.825362 disk-uuid[555]: Secondary Header is updated. Feb 13 19:49:56.829657 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Feb 13 19:49:56.834639 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Feb 13 19:49:56.842460 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 19:49:57.025660 kernel: ata2: SATA link down (SStatus 0 SControl 300) Feb 13 19:49:57.034249 kernel: ata5: SATA link down (SStatus 0 SControl 300) Feb 13 19:49:57.034327 kernel: ata1: SATA link down (SStatus 0 SControl 300) Feb 13 19:49:57.034338 kernel: ata4: SATA link down (SStatus 0 SControl 300) Feb 13 19:49:57.035652 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Feb 13 19:49:57.036643 kernel: ata6: SATA link down (SStatus 0 SControl 300) Feb 13 19:49:57.037648 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Feb 13 19:49:57.039141 kernel: ata3.00: applying bridge limits Feb 13 19:49:57.039159 kernel: ata3.00: configured for UDMA/100 Feb 13 19:49:57.039651 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Feb 13 19:49:57.085662 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Feb 13 19:49:57.102456 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Feb 13 19:49:57.102478 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Feb 13 19:49:57.851654 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Feb 13 19:49:57.851990 disk-uuid[561]: The operation has completed successfully. Feb 13 19:49:57.883115 systemd[1]: disk-uuid.service: Deactivated successfully. Feb 13 19:49:57.883258 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Feb 13 19:49:57.912784 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Feb 13 19:49:57.921977 sh[598]: Success Feb 13 19:49:57.937648 kernel: device-mapper: verity: sha256 using implementation "sha256-ni" Feb 13 19:49:57.969000 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Feb 13 19:49:57.991066 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Feb 13 19:49:57.993286 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Feb 13 19:49:58.005932 kernel: BTRFS info (device dm-0): first mount of filesystem e9c87d9f-3864-4b45-9be4-80a5397f1fc6 Feb 13 19:49:58.005962 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Feb 13 19:49:58.005973 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Feb 13 19:49:58.006940 kernel: BTRFS info (device dm-0): disabling log replay at mount time Feb 13 19:49:58.008631 kernel: BTRFS info (device dm-0): using free space tree Feb 13 19:49:58.012093 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Feb 13 19:49:58.024137 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Feb 13 19:49:58.033747 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Feb 13 19:49:58.035366 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Feb 13 19:49:58.047657 kernel: BTRFS info (device vda6): first mount of filesystem 84d576e4-038f-4c76-aa8e-6cfd81e812ea Feb 13 19:49:58.047700 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Feb 13 19:49:58.047716 kernel: BTRFS info (device vda6): using free space tree Feb 13 19:49:58.050663 kernel: BTRFS info (device vda6): auto enabling async discard Feb 13 19:49:58.059270 systemd[1]: mnt-oem.mount: Deactivated successfully. Feb 13 19:49:58.061036 kernel: BTRFS info (device vda6): last unmount of filesystem 84d576e4-038f-4c76-aa8e-6cfd81e812ea Feb 13 19:49:58.081080 systemd[1]: Finished ignition-setup.service - Ignition (setup). Feb 13 19:49:58.089897 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Feb 13 19:49:58.153092 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Feb 13 19:49:58.188721 systemd[1]: Starting systemd-networkd.service - Network Configuration... Feb 13 19:49:58.214480 systemd-networkd[779]: lo: Link UP Feb 13 19:49:58.214492 systemd-networkd[779]: lo: Gained carrier Feb 13 19:49:58.214602 ignition[713]: Ignition 2.20.0 Feb 13 19:49:58.216105 systemd-networkd[779]: Enumeration completed Feb 13 19:49:58.214609 ignition[713]: Stage: fetch-offline Feb 13 19:49:58.216453 systemd[1]: Started systemd-networkd.service - Network Configuration. Feb 13 19:49:58.214656 ignition[713]: no configs at "/usr/lib/ignition/base.d" Feb 13 19:49:58.216500 systemd-networkd[779]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 19:49:58.214667 ignition[713]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Feb 13 19:49:58.216504 systemd-networkd[779]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Feb 13 19:49:58.214748 ignition[713]: parsed url from cmdline: "" Feb 13 19:49:58.218003 systemd[1]: Reached target network.target - Network. Feb 13 19:49:58.214752 ignition[713]: no config URL provided Feb 13 19:49:58.219129 systemd-networkd[779]: eth0: Link UP Feb 13 19:49:58.214757 ignition[713]: reading system config file "/usr/lib/ignition/user.ign" Feb 13 19:49:58.219133 systemd-networkd[779]: eth0: Gained carrier Feb 13 19:49:58.214766 ignition[713]: no config at "/usr/lib/ignition/user.ign" Feb 13 19:49:58.219141 systemd-networkd[779]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 19:49:58.214795 ignition[713]: op(1): [started] loading QEMU firmware config module Feb 13 19:49:58.214800 ignition[713]: op(1): executing: "modprobe" "qemu_fw_cfg" Feb 13 19:49:58.237886 ignition[713]: op(1): [finished] loading QEMU firmware config module Feb 13 19:49:58.243690 systemd-networkd[779]: eth0: DHCPv4 address 10.0.0.104/16, gateway 10.0.0.1 acquired from 10.0.0.1 Feb 13 19:49:58.280797 ignition[713]: parsing config with SHA512: 3bf034c6e7c564c5b37e273571dbef442d3cccc5031e10fd6dd9c178cb91f0ac5d1332904f7bdb8bbec0cf0be514dacc2942e99d3a163b87ff8ed7c978690156 Feb 13 19:49:58.284839 unknown[713]: fetched base config from "system" Feb 13 19:49:58.284913 unknown[713]: fetched user config from "qemu" Feb 13 19:49:58.285273 ignition[713]: fetch-offline: fetch-offline passed Feb 13 19:49:58.288137 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Feb 13 19:49:58.285341 ignition[713]: Ignition finished successfully Feb 13 19:49:58.289838 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Feb 13 19:49:58.303811 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Feb 13 19:49:58.321579 ignition[792]: Ignition 2.20.0 Feb 13 19:49:58.321590 ignition[792]: Stage: kargs Feb 13 19:49:58.321763 ignition[792]: no configs at "/usr/lib/ignition/base.d" Feb 13 19:49:58.321774 ignition[792]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Feb 13 19:49:58.322596 ignition[792]: kargs: kargs passed Feb 13 19:49:58.322650 ignition[792]: Ignition finished successfully Feb 13 19:49:58.326126 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Feb 13 19:49:58.342816 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Feb 13 19:49:58.355883 ignition[801]: Ignition 2.20.0 Feb 13 19:49:58.355896 ignition[801]: Stage: disks Feb 13 19:49:58.356075 ignition[801]: no configs at "/usr/lib/ignition/base.d" Feb 13 19:49:58.356086 ignition[801]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Feb 13 19:49:58.356888 ignition[801]: disks: disks passed Feb 13 19:49:58.359265 systemd[1]: Finished ignition-disks.service - Ignition (disks). Feb 13 19:49:58.356932 ignition[801]: Ignition finished successfully Feb 13 19:49:58.360581 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Feb 13 19:49:58.362149 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Feb 13 19:49:58.364277 systemd[1]: Reached target local-fs.target - Local File Systems. Feb 13 19:49:58.365425 systemd[1]: Reached target sysinit.target - System Initialization. Feb 13 19:49:58.367164 systemd[1]: Reached target basic.target - Basic System. Feb 13 19:49:58.378816 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Feb 13 19:49:58.390757 systemd-resolved[229]: Detected conflict on linux IN A 10.0.0.104 Feb 13 19:49:58.390777 systemd-resolved[229]: Hostname conflict, changing published hostname from 'linux' to 'linux6'. Feb 13 19:49:58.394282 systemd-fsck[811]: ROOT: clean, 14/553520 files, 52654/553472 blocks Feb 13 19:49:58.432689 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Feb 13 19:49:58.440712 systemd[1]: Mounting sysroot.mount - /sysroot... Feb 13 19:49:58.530647 kernel: EXT4-fs (vda9): mounted filesystem c5993b0e-9201-4b44-aa01-79dc9d6c9fc9 r/w with ordered data mode. Quota mode: none. Feb 13 19:49:58.530875 systemd[1]: Mounted sysroot.mount - /sysroot. Feb 13 19:49:58.531548 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Feb 13 19:49:58.541740 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Feb 13 19:49:58.543991 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Feb 13 19:49:58.544429 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Feb 13 19:49:58.550872 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/vda6 scanned by mount (819) Feb 13 19:49:58.550895 kernel: BTRFS info (device vda6): first mount of filesystem 84d576e4-038f-4c76-aa8e-6cfd81e812ea Feb 13 19:49:58.544480 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Feb 13 19:49:58.558512 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Feb 13 19:49:58.558541 kernel: BTRFS info (device vda6): using free space tree Feb 13 19:49:58.558556 kernel: BTRFS info (device vda6): auto enabling async discard Feb 13 19:49:58.544508 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Feb 13 19:49:58.552748 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Feb 13 19:49:58.560030 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Feb 13 19:49:58.571963 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Feb 13 19:49:58.608742 initrd-setup-root[843]: cut: /sysroot/etc/passwd: No such file or directory Feb 13 19:49:58.614743 initrd-setup-root[850]: cut: /sysroot/etc/group: No such file or directory Feb 13 19:49:58.619832 initrd-setup-root[857]: cut: /sysroot/etc/shadow: No such file or directory Feb 13 19:49:58.624132 initrd-setup-root[864]: cut: /sysroot/etc/gshadow: No such file or directory Feb 13 19:49:58.723741 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Feb 13 19:49:58.730774 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Feb 13 19:49:58.733345 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Feb 13 19:49:58.740657 kernel: BTRFS info (device vda6): last unmount of filesystem 84d576e4-038f-4c76-aa8e-6cfd81e812ea Feb 13 19:49:58.762197 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Feb 13 19:49:58.813452 ignition[933]: INFO : Ignition 2.20.0 Feb 13 19:49:58.813452 ignition[933]: INFO : Stage: mount Feb 13 19:49:58.815379 ignition[933]: INFO : no configs at "/usr/lib/ignition/base.d" Feb 13 19:49:58.815379 ignition[933]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Feb 13 19:49:58.815379 ignition[933]: INFO : mount: mount passed Feb 13 19:49:58.815379 ignition[933]: INFO : Ignition finished successfully Feb 13 19:49:58.817288 systemd[1]: Finished ignition-mount.service - Ignition (mount). Feb 13 19:49:58.825740 systemd[1]: Starting ignition-files.service - Ignition (files)... Feb 13 19:49:59.005853 systemd[1]: sysroot-oem.mount: Deactivated successfully. Feb 13 19:49:59.017976 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Feb 13 19:49:59.025030 kernel: BTRFS: device label OEM devid 1 transid 16 /dev/vda6 scanned by mount (944) Feb 13 19:49:59.025076 kernel: BTRFS info (device vda6): first mount of filesystem 84d576e4-038f-4c76-aa8e-6cfd81e812ea Feb 13 19:49:59.025090 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Feb 13 19:49:59.025878 kernel: BTRFS info (device vda6): using free space tree Feb 13 19:49:59.029625 kernel: BTRFS info (device vda6): auto enabling async discard Feb 13 19:49:59.031046 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Feb 13 19:49:59.054538 ignition[961]: INFO : Ignition 2.20.0 Feb 13 19:49:59.054538 ignition[961]: INFO : Stage: files Feb 13 19:49:59.056340 ignition[961]: INFO : no configs at "/usr/lib/ignition/base.d" Feb 13 19:49:59.056340 ignition[961]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Feb 13 19:49:59.056340 ignition[961]: DEBUG : files: compiled without relabeling support, skipping Feb 13 19:49:59.059800 ignition[961]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Feb 13 19:49:59.059800 ignition[961]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Feb 13 19:49:59.063363 ignition[961]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Feb 13 19:49:59.063363 ignition[961]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Feb 13 19:49:59.063363 ignition[961]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Feb 13 19:49:59.063363 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Feb 13 19:49:59.063363 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.0-linux-amd64.tar.gz: attempt #1 Feb 13 19:49:59.060768 unknown[961]: wrote ssh authorized keys file for user: core Feb 13 19:49:59.119900 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Feb 13 19:49:59.257827 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" Feb 13 19:49:59.257827 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Feb 13 19:49:59.261932 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Feb 13 19:49:59.261932 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Feb 13 19:49:59.261932 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Feb 13 19:49:59.261932 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Feb 13 19:49:59.261932 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Feb 13 19:49:59.261932 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Feb 13 19:49:59.261932 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Feb 13 19:49:59.261932 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Feb 13 19:49:59.261932 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Feb 13 19:49:59.261932 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.0-x86-64.raw" Feb 13 19:49:59.261932 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.0-x86-64.raw" Feb 13 19:49:59.261932 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.0-x86-64.raw" Feb 13 19:49:59.261932 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.32.0-x86-64.raw: attempt #1 Feb 13 19:49:59.814960 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Feb 13 19:50:00.154993 systemd-networkd[779]: eth0: Gained IPv6LL Feb 13 19:50:00.434949 ignition[961]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.0-x86-64.raw" Feb 13 19:50:00.434949 ignition[961]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Feb 13 19:50:00.457639 ignition[961]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Feb 13 19:50:00.457639 ignition[961]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Feb 13 19:50:00.457639 ignition[961]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Feb 13 19:50:00.457639 ignition[961]: INFO : files: op(d): [started] processing unit "coreos-metadata.service" Feb 13 19:50:00.457639 ignition[961]: INFO : files: op(d): op(e): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Feb 13 19:50:00.457639 ignition[961]: INFO : files: op(d): op(e): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Feb 13 19:50:00.457639 ignition[961]: INFO : files: op(d): [finished] processing unit "coreos-metadata.service" Feb 13 19:50:00.457639 ignition[961]: INFO : files: op(f): [started] setting preset to disabled for "coreos-metadata.service" Feb 13 19:50:00.518916 ignition[961]: INFO : files: op(f): op(10): [started] removing enablement symlink(s) for "coreos-metadata.service" Feb 13 19:50:00.524211 ignition[961]: INFO : files: op(f): op(10): [finished] removing enablement symlink(s) for "coreos-metadata.service" Feb 13 19:50:00.526068 ignition[961]: INFO : files: op(f): [finished] setting preset to disabled for "coreos-metadata.service" Feb 13 19:50:00.526068 ignition[961]: INFO : files: op(11): [started] setting preset to enabled for "prepare-helm.service" Feb 13 19:50:00.526068 ignition[961]: INFO : files: op(11): [finished] setting preset to enabled for "prepare-helm.service" Feb 13 19:50:00.526068 ignition[961]: INFO : files: createResultFile: createFiles: op(12): [started] writing file "/sysroot/etc/.ignition-result.json" Feb 13 19:50:00.526068 ignition[961]: INFO : files: createResultFile: createFiles: op(12): [finished] writing file "/sysroot/etc/.ignition-result.json" Feb 13 19:50:00.526068 ignition[961]: INFO : files: files passed Feb 13 19:50:00.526068 ignition[961]: INFO : Ignition finished successfully Feb 13 19:50:00.539076 systemd[1]: Finished ignition-files.service - Ignition (files). Feb 13 19:50:00.556887 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Feb 13 19:50:00.594913 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Feb 13 19:50:00.598196 systemd[1]: ignition-quench.service: Deactivated successfully. Feb 13 19:50:00.598324 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Feb 13 19:50:00.610965 initrd-setup-root-after-ignition[990]: grep: /sysroot/oem/oem-release: No such file or directory Feb 13 19:50:00.614496 initrd-setup-root-after-ignition[992]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Feb 13 19:50:00.614496 initrd-setup-root-after-ignition[992]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Feb 13 19:50:00.618149 initrd-setup-root-after-ignition[996]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Feb 13 19:50:00.620811 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Feb 13 19:50:00.621068 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Feb 13 19:50:00.665754 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Feb 13 19:50:00.693329 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Feb 13 19:50:00.693452 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Feb 13 19:50:00.744778 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Feb 13 19:50:00.746840 systemd[1]: Reached target initrd.target - Initrd Default Target. Feb 13 19:50:00.747858 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Feb 13 19:50:00.754922 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Feb 13 19:50:00.769829 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Feb 13 19:50:00.781919 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Feb 13 19:50:00.790579 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Feb 13 19:50:00.819430 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Feb 13 19:50:00.823223 systemd[1]: Stopped target timers.target - Timer Units. Feb 13 19:50:00.823363 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Feb 13 19:50:00.823498 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Feb 13 19:50:00.825965 systemd[1]: Stopped target initrd.target - Initrd Default Target. Feb 13 19:50:00.829223 systemd[1]: Stopped target basic.target - Basic System. Feb 13 19:50:00.829641 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Feb 13 19:50:00.830241 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Feb 13 19:50:00.830682 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Feb 13 19:50:00.839405 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Feb 13 19:50:00.840735 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Feb 13 19:50:00.841128 systemd[1]: Stopped target sysinit.target - System Initialization. Feb 13 19:50:00.841526 systemd[1]: Stopped target local-fs.target - Local File Systems. Feb 13 19:50:00.841945 systemd[1]: Stopped target swap.target - Swaps. Feb 13 19:50:00.842327 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Feb 13 19:50:00.842443 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Feb 13 19:50:00.873022 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Feb 13 19:50:00.873183 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Feb 13 19:50:00.873563 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Feb 13 19:50:00.873967 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Feb 13 19:50:00.878154 systemd[1]: dracut-initqueue.service: Deactivated successfully. Feb 13 19:50:00.878289 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Feb 13 19:50:00.880883 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Feb 13 19:50:00.881006 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Feb 13 19:50:00.883477 systemd[1]: Stopped target paths.target - Path Units. Feb 13 19:50:00.883996 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Feb 13 19:50:00.887940 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Feb 13 19:50:00.889164 systemd[1]: Stopped target slices.target - Slice Units. Feb 13 19:50:00.891802 systemd[1]: Stopped target sockets.target - Socket Units. Feb 13 19:50:00.892131 systemd[1]: iscsid.socket: Deactivated successfully. Feb 13 19:50:00.892220 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Feb 13 19:50:00.895256 systemd[1]: iscsiuio.socket: Deactivated successfully. Feb 13 19:50:00.895334 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Feb 13 19:50:00.919034 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Feb 13 19:50:00.919141 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Feb 13 19:50:00.920096 systemd[1]: ignition-files.service: Deactivated successfully. Feb 13 19:50:00.920209 systemd[1]: Stopped ignition-files.service - Ignition (files). Feb 13 19:50:00.944906 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Feb 13 19:50:00.947115 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Feb 13 19:50:00.947281 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Feb 13 19:50:00.949463 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Feb 13 19:50:00.953231 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Feb 13 19:50:00.954553 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Feb 13 19:50:00.958067 ignition[1016]: INFO : Ignition 2.20.0 Feb 13 19:50:00.958067 ignition[1016]: INFO : Stage: umount Feb 13 19:50:00.988424 ignition[1016]: INFO : no configs at "/usr/lib/ignition/base.d" Feb 13 19:50:00.988424 ignition[1016]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Feb 13 19:50:00.988424 ignition[1016]: INFO : umount: umount passed Feb 13 19:50:00.988424 ignition[1016]: INFO : Ignition finished successfully Feb 13 19:50:00.990721 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Feb 13 19:50:01.025458 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Feb 13 19:50:01.031736 systemd[1]: sysroot-boot.mount: Deactivated successfully. Feb 13 19:50:01.033692 systemd[1]: ignition-mount.service: Deactivated successfully. Feb 13 19:50:01.034925 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Feb 13 19:50:01.039400 systemd[1]: initrd-cleanup.service: Deactivated successfully. Feb 13 19:50:01.040642 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Feb 13 19:50:01.043997 systemd[1]: Stopped target network.target - Network. Feb 13 19:50:01.045727 systemd[1]: ignition-disks.service: Deactivated successfully. Feb 13 19:50:01.046681 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Feb 13 19:50:01.048759 systemd[1]: ignition-kargs.service: Deactivated successfully. Feb 13 19:50:01.048820 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Feb 13 19:50:01.052260 systemd[1]: ignition-setup.service: Deactivated successfully. Feb 13 19:50:01.052327 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Feb 13 19:50:01.055683 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Feb 13 19:50:01.080284 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Feb 13 19:50:01.082525 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Feb 13 19:50:01.084736 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Feb 13 19:50:01.089658 systemd-networkd[779]: eth0: DHCPv6 lease lost Feb 13 19:50:01.091679 systemd[1]: systemd-networkd.service: Deactivated successfully. Feb 13 19:50:01.091846 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Feb 13 19:50:01.093203 systemd[1]: systemd-networkd.socket: Deactivated successfully. Feb 13 19:50:01.093245 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Feb 13 19:50:01.109723 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Feb 13 19:50:01.110676 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Feb 13 19:50:01.110741 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Feb 13 19:50:01.111097 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Feb 13 19:50:01.111723 systemd[1]: systemd-resolved.service: Deactivated successfully. Feb 13 19:50:01.111875 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Feb 13 19:50:01.161584 systemd[1]: systemd-sysctl.service: Deactivated successfully. Feb 13 19:50:01.162155 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Feb 13 19:50:01.165172 systemd[1]: systemd-modules-load.service: Deactivated successfully. Feb 13 19:50:01.165240 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Feb 13 19:50:01.167537 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Feb 13 19:50:01.167597 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Feb 13 19:50:01.173024 systemd[1]: network-cleanup.service: Deactivated successfully. Feb 13 19:50:01.173181 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Feb 13 19:50:01.215793 systemd[1]: systemd-udevd.service: Deactivated successfully. Feb 13 19:50:01.216023 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Feb 13 19:50:01.218495 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Feb 13 19:50:01.218561 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Feb 13 19:50:01.221076 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Feb 13 19:50:01.221118 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Feb 13 19:50:01.223471 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Feb 13 19:50:01.223526 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Feb 13 19:50:01.225743 systemd[1]: dracut-cmdline.service: Deactivated successfully. Feb 13 19:50:01.225800 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Feb 13 19:50:01.227746 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Feb 13 19:50:01.227801 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 19:50:01.240753 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Feb 13 19:50:01.241851 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Feb 13 19:50:01.241907 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Feb 13 19:50:01.244187 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Feb 13 19:50:01.244237 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 19:50:01.250426 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Feb 13 19:50:01.250539 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Feb 13 19:50:01.409569 systemd[1]: sysroot-boot.service: Deactivated successfully. Feb 13 19:50:01.409799 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Feb 13 19:50:01.412471 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Feb 13 19:50:01.413890 systemd[1]: initrd-setup-root.service: Deactivated successfully. Feb 13 19:50:01.413990 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Feb 13 19:50:01.429018 systemd[1]: Starting initrd-switch-root.service - Switch Root... Feb 13 19:50:01.437395 systemd[1]: Switching root. Feb 13 19:50:01.477841 systemd-journald[194]: Journal stopped Feb 13 19:50:03.115974 systemd-journald[194]: Received SIGTERM from PID 1 (systemd). Feb 13 19:50:03.116041 kernel: SELinux: policy capability network_peer_controls=1 Feb 13 19:50:03.116057 kernel: SELinux: policy capability open_perms=1 Feb 13 19:50:03.116068 kernel: SELinux: policy capability extended_socket_class=1 Feb 13 19:50:03.116079 kernel: SELinux: policy capability always_check_network=0 Feb 13 19:50:03.116090 kernel: SELinux: policy capability cgroup_seclabel=1 Feb 13 19:50:03.116106 kernel: SELinux: policy capability nnp_nosuid_transition=1 Feb 13 19:50:03.116117 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Feb 13 19:50:03.116129 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Feb 13 19:50:03.116140 kernel: audit: type=1403 audit(1739476202.252:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Feb 13 19:50:03.116158 systemd[1]: Successfully loaded SELinux policy in 51.110ms. Feb 13 19:50:03.116182 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 14.632ms. Feb 13 19:50:03.116195 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Feb 13 19:50:03.116207 systemd[1]: Detected virtualization kvm. Feb 13 19:50:03.116219 systemd[1]: Detected architecture x86-64. Feb 13 19:50:03.116232 systemd[1]: Detected first boot. Feb 13 19:50:03.116244 systemd[1]: Initializing machine ID from VM UUID. Feb 13 19:50:03.116255 zram_generator::config[1059]: No configuration found. Feb 13 19:50:03.116272 systemd[1]: Populated /etc with preset unit settings. Feb 13 19:50:03.116284 systemd[1]: initrd-switch-root.service: Deactivated successfully. Feb 13 19:50:03.116296 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Feb 13 19:50:03.116308 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Feb 13 19:50:03.116320 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Feb 13 19:50:03.116336 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Feb 13 19:50:03.116349 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Feb 13 19:50:03.116361 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Feb 13 19:50:03.116373 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Feb 13 19:50:03.116385 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Feb 13 19:50:03.116397 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Feb 13 19:50:03.116409 systemd[1]: Created slice user.slice - User and Session Slice. Feb 13 19:50:03.116420 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Feb 13 19:50:03.116433 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Feb 13 19:50:03.116447 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Feb 13 19:50:03.116463 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Feb 13 19:50:03.116475 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Feb 13 19:50:03.116487 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Feb 13 19:50:03.116498 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Feb 13 19:50:03.116510 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Feb 13 19:50:03.116522 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Feb 13 19:50:03.116534 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Feb 13 19:50:03.116549 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Feb 13 19:50:03.116561 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Feb 13 19:50:03.116572 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Feb 13 19:50:03.116588 systemd[1]: Reached target remote-fs.target - Remote File Systems. Feb 13 19:50:03.116602 systemd[1]: Reached target slices.target - Slice Units. Feb 13 19:50:03.116625 systemd[1]: Reached target swap.target - Swaps. Feb 13 19:50:03.116638 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Feb 13 19:50:03.116650 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Feb 13 19:50:03.116664 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Feb 13 19:50:03.116676 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Feb 13 19:50:03.116688 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Feb 13 19:50:03.116699 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Feb 13 19:50:03.116711 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Feb 13 19:50:03.116723 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Feb 13 19:50:03.116744 systemd[1]: Mounting media.mount - External Media Directory... Feb 13 19:50:03.116756 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 13 19:50:03.116768 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Feb 13 19:50:03.116782 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Feb 13 19:50:03.116794 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Feb 13 19:50:03.116807 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Feb 13 19:50:03.116819 systemd[1]: Reached target machines.target - Containers. Feb 13 19:50:03.116831 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Feb 13 19:50:03.116847 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Feb 13 19:50:03.116859 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Feb 13 19:50:03.116872 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Feb 13 19:50:03.116890 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Feb 13 19:50:03.116904 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Feb 13 19:50:03.116916 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Feb 13 19:50:03.116928 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Feb 13 19:50:03.116940 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Feb 13 19:50:03.116951 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Feb 13 19:50:03.116963 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Feb 13 19:50:03.116975 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Feb 13 19:50:03.116987 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Feb 13 19:50:03.117001 systemd[1]: Stopped systemd-fsck-usr.service. Feb 13 19:50:03.117012 kernel: fuse: init (API version 7.39) Feb 13 19:50:03.117024 systemd[1]: Starting systemd-journald.service - Journal Service... Feb 13 19:50:03.117035 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Feb 13 19:50:03.117047 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Feb 13 19:50:03.117059 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Feb 13 19:50:03.117070 kernel: loop: module loaded Feb 13 19:50:03.117103 systemd-journald[1133]: Collecting audit messages is disabled. Feb 13 19:50:03.117131 kernel: ACPI: bus type drm_connector registered Feb 13 19:50:03.117143 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Feb 13 19:50:03.117155 systemd-journald[1133]: Journal started Feb 13 19:50:03.117176 systemd-journald[1133]: Runtime Journal (/run/log/journal/746b73c0bb8e4cadb9b74db0f0f2dea7) is 6.0M, max 48.3M, 42.2M free. Feb 13 19:50:02.874095 systemd[1]: Queued start job for default target multi-user.target. Feb 13 19:50:02.899463 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Feb 13 19:50:02.900064 systemd[1]: systemd-journald.service: Deactivated successfully. Feb 13 19:50:03.118644 systemd[1]: verity-setup.service: Deactivated successfully. Feb 13 19:50:03.118666 systemd[1]: Stopped verity-setup.service. Feb 13 19:50:03.122694 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 13 19:50:03.126037 systemd[1]: Started systemd-journald.service - Journal Service. Feb 13 19:50:03.126893 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Feb 13 19:50:03.128147 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Feb 13 19:50:03.129399 systemd[1]: Mounted media.mount - External Media Directory. Feb 13 19:50:03.130534 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Feb 13 19:50:03.131797 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Feb 13 19:50:03.133173 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Feb 13 19:50:03.134474 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Feb 13 19:50:03.136066 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Feb 13 19:50:03.137740 systemd[1]: modprobe@configfs.service: Deactivated successfully. Feb 13 19:50:03.137913 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Feb 13 19:50:03.139415 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Feb 13 19:50:03.139582 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Feb 13 19:50:03.141073 systemd[1]: modprobe@drm.service: Deactivated successfully. Feb 13 19:50:03.141274 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Feb 13 19:50:03.142686 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Feb 13 19:50:03.142864 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Feb 13 19:50:03.144396 systemd[1]: modprobe@fuse.service: Deactivated successfully. Feb 13 19:50:03.144559 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Feb 13 19:50:03.145998 systemd[1]: modprobe@loop.service: Deactivated successfully. Feb 13 19:50:03.146165 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Feb 13 19:50:03.147570 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Feb 13 19:50:03.149006 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Feb 13 19:50:03.150790 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Feb 13 19:50:03.167950 systemd[1]: Reached target network-pre.target - Preparation for Network. Feb 13 19:50:03.176791 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Feb 13 19:50:03.178531 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Feb 13 19:50:03.179897 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Feb 13 19:50:03.179943 systemd[1]: Reached target local-fs.target - Local File Systems. Feb 13 19:50:03.182423 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Feb 13 19:50:03.185798 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Feb 13 19:50:03.190081 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Feb 13 19:50:03.191337 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Feb 13 19:50:03.194902 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Feb 13 19:50:03.199221 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Feb 13 19:50:03.201072 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Feb 13 19:50:03.203251 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Feb 13 19:50:03.204904 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Feb 13 19:50:03.206502 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Feb 13 19:50:03.215894 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Feb 13 19:50:03.244161 systemd[1]: Starting systemd-sysusers.service - Create System Users... Feb 13 19:50:03.251508 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Feb 13 19:50:03.260087 systemd-journald[1133]: Time spent on flushing to /var/log/journal/746b73c0bb8e4cadb9b74db0f0f2dea7 is 39.509ms for 1045 entries. Feb 13 19:50:03.260087 systemd-journald[1133]: System Journal (/var/log/journal/746b73c0bb8e4cadb9b74db0f0f2dea7) is 8.0M, max 195.6M, 187.6M free. Feb 13 19:50:03.330096 systemd-journald[1133]: Received client request to flush runtime journal. Feb 13 19:50:03.330187 kernel: loop0: detected capacity change from 0 to 140992 Feb 13 19:50:03.330239 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Feb 13 19:50:03.256985 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Feb 13 19:50:03.260115 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Feb 13 19:50:03.292065 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Feb 13 19:50:03.296826 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Feb 13 19:50:03.311070 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Feb 13 19:50:03.333152 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Feb 13 19:50:03.339860 systemd[1]: Finished systemd-sysusers.service - Create System Users. Feb 13 19:50:03.352023 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Feb 13 19:50:03.354542 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Feb 13 19:50:03.355556 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Feb 13 19:50:03.357670 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Feb 13 19:50:03.359757 kernel: loop1: detected capacity change from 0 to 138184 Feb 13 19:50:03.359979 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Feb 13 19:50:03.375986 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Feb 13 19:50:03.387203 systemd-tmpfiles[1188]: ACLs are not supported, ignoring. Feb 13 19:50:03.387228 systemd-tmpfiles[1188]: ACLs are not supported, ignoring. Feb 13 19:50:03.395306 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Feb 13 19:50:03.413530 udevadm[1194]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Feb 13 19:50:03.428654 kernel: loop2: detected capacity change from 0 to 218376 Feb 13 19:50:03.491197 kernel: loop3: detected capacity change from 0 to 140992 Feb 13 19:50:03.508053 kernel: loop4: detected capacity change from 0 to 138184 Feb 13 19:50:03.526656 kernel: loop5: detected capacity change from 0 to 218376 Feb 13 19:50:03.533749 (sd-merge)[1198]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. Feb 13 19:50:03.534365 (sd-merge)[1198]: Merged extensions into '/usr'. Feb 13 19:50:03.538444 systemd[1]: Reloading requested from client PID 1173 ('systemd-sysext') (unit systemd-sysext.service)... Feb 13 19:50:03.538537 systemd[1]: Reloading... Feb 13 19:50:03.622642 zram_generator::config[1223]: No configuration found. Feb 13 19:50:03.765403 ldconfig[1168]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Feb 13 19:50:03.775321 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 19:50:03.832105 systemd[1]: Reloading finished in 293 ms. Feb 13 19:50:03.866680 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Feb 13 19:50:03.868574 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Feb 13 19:50:03.882970 systemd[1]: Starting ensure-sysext.service... Feb 13 19:50:03.885431 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Feb 13 19:50:03.892869 systemd[1]: Reloading requested from client PID 1261 ('systemctl') (unit ensure-sysext.service)... Feb 13 19:50:03.892892 systemd[1]: Reloading... Feb 13 19:50:03.913370 systemd-tmpfiles[1262]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Feb 13 19:50:03.913844 systemd-tmpfiles[1262]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Feb 13 19:50:03.914949 systemd-tmpfiles[1262]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Feb 13 19:50:03.915374 systemd-tmpfiles[1262]: ACLs are not supported, ignoring. Feb 13 19:50:03.915484 systemd-tmpfiles[1262]: ACLs are not supported, ignoring. Feb 13 19:50:03.920090 systemd-tmpfiles[1262]: Detected autofs mount point /boot during canonicalization of boot. Feb 13 19:50:03.920110 systemd-tmpfiles[1262]: Skipping /boot Feb 13 19:50:03.935486 systemd-tmpfiles[1262]: Detected autofs mount point /boot during canonicalization of boot. Feb 13 19:50:03.935506 systemd-tmpfiles[1262]: Skipping /boot Feb 13 19:50:03.962643 zram_generator::config[1292]: No configuration found. Feb 13 19:50:04.067764 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 19:50:04.118557 systemd[1]: Reloading finished in 225 ms. Feb 13 19:50:04.140819 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Feb 13 19:50:04.159177 systemd[1]: Starting audit-rules.service - Load Audit Rules... Feb 13 19:50:04.182027 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Feb 13 19:50:04.193805 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Feb 13 19:50:04.198891 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Feb 13 19:50:04.211185 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Feb 13 19:50:04.232305 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Feb 13 19:50:04.236635 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 13 19:50:04.236841 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Feb 13 19:50:04.238195 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Feb 13 19:50:04.240277 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Feb 13 19:50:04.248869 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Feb 13 19:50:04.250040 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Feb 13 19:50:04.252089 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Feb 13 19:50:04.253322 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 13 19:50:04.254506 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Feb 13 19:50:04.254723 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Feb 13 19:50:04.256395 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Feb 13 19:50:04.256607 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Feb 13 19:50:04.262208 systemd[1]: modprobe@loop.service: Deactivated successfully. Feb 13 19:50:04.262391 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Feb 13 19:50:04.265158 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 13 19:50:04.265442 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Feb 13 19:50:04.276832 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Feb 13 19:50:04.295036 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Feb 13 19:50:04.300069 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Feb 13 19:50:04.301845 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Feb 13 19:50:04.302143 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 13 19:50:04.303267 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Feb 13 19:50:04.303473 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Feb 13 19:50:04.319176 systemd[1]: Started systemd-userdbd.service - User Database Manager. Feb 13 19:50:04.321262 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Feb 13 19:50:04.323862 systemd[1]: modprobe@drm.service: Deactivated successfully. Feb 13 19:50:04.324250 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Feb 13 19:50:04.326367 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Feb 13 19:50:04.326580 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Feb 13 19:50:04.330376 systemd[1]: Finished ensure-sysext.service. Feb 13 19:50:04.339947 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Feb 13 19:50:04.340024 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Feb 13 19:50:04.348875 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Feb 13 19:50:04.355315 augenrules[1373]: No rules Feb 13 19:50:04.358492 systemd[1]: audit-rules.service: Deactivated successfully. Feb 13 19:50:04.359524 systemd[1]: Finished audit-rules.service - Load Audit Rules. Feb 13 19:50:04.375268 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Feb 13 19:50:04.378410 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Feb 13 19:50:04.388047 systemd-resolved[1331]: Positive Trust Anchors: Feb 13 19:50:04.388392 systemd-resolved[1331]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Feb 13 19:50:04.388463 systemd-resolved[1331]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Feb 13 19:50:04.391962 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Feb 13 19:50:04.394020 systemd-resolved[1331]: Defaulting to hostname 'linux'. Feb 13 19:50:04.403994 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Feb 13 19:50:04.407858 systemd[1]: Starting systemd-update-done.service - Update is Completed... Feb 13 19:50:04.409545 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Feb 13 19:50:04.411326 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Feb 13 19:50:04.425104 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Feb 13 19:50:04.426894 systemd[1]: Reached target time-set.target - System Time Set. Feb 13 19:50:04.432195 systemd[1]: Finished systemd-update-done.service - Update is Completed. Feb 13 19:50:04.433286 systemd-udevd[1381]: Using default interface naming scheme 'v255'. Feb 13 19:50:04.453185 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Feb 13 19:50:04.463881 systemd[1]: Starting systemd-networkd.service - Network Configuration... Feb 13 19:50:04.492106 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Feb 13 19:50:04.515780 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 38 scanned by (udev-worker) (1394) Feb 13 19:50:04.555067 systemd-networkd[1392]: lo: Link UP Feb 13 19:50:04.555078 systemd-networkd[1392]: lo: Gained carrier Feb 13 19:50:04.556739 systemd-networkd[1392]: Enumeration completed Feb 13 19:50:04.557169 systemd[1]: Started systemd-networkd.service - Network Configuration. Feb 13 19:50:04.558083 systemd-networkd[1392]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 19:50:04.558095 systemd-networkd[1392]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Feb 13 19:50:04.559077 systemd[1]: Reached target network.target - Network. Feb 13 19:50:04.559192 systemd-networkd[1392]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 19:50:04.559226 systemd-networkd[1392]: eth0: Link UP Feb 13 19:50:04.559230 systemd-networkd[1392]: eth0: Gained carrier Feb 13 19:50:04.559240 systemd-networkd[1392]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 19:50:04.565671 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input2 Feb 13 19:50:04.569938 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Feb 13 19:50:04.571714 systemd-networkd[1392]: eth0: DHCPv4 address 10.0.0.104/16, gateway 10.0.0.1 acquired from 10.0.0.1 Feb 13 19:50:04.573110 systemd-timesyncd[1371]: Network configuration changed, trying to establish connection. Feb 13 19:50:05.647290 systemd-resolved[1331]: Clock change detected. Flushing caches. Feb 13 19:50:05.647425 systemd-timesyncd[1371]: Contacted time server 10.0.0.1:123 (10.0.0.1). Feb 13 19:50:05.647489 systemd-timesyncd[1371]: Initial clock synchronization to Thu 2025-02-13 19:50:05.647250 UTC. Feb 13 19:50:05.663025 kernel: ACPI: button: Power Button [PWRF] Feb 13 19:50:05.666916 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input3 Feb 13 19:50:05.671038 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Feb 13 19:50:05.678111 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Feb 13 19:50:05.717540 kernel: i801_smbus 0000:00:1f.3: Enabling SMBus device Feb 13 19:50:05.717928 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Feb 13 19:50:05.718138 kernel: i2c i2c-0: 1/1 memory slots populated (from DMI) Feb 13 19:50:05.718354 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Feb 13 19:50:05.718522 kernel: mousedev: PS/2 mouse device common for all mice Feb 13 19:50:05.713296 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 19:50:05.715935 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Feb 13 19:50:05.735690 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Feb 13 19:50:05.736019 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 19:50:05.779720 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 19:50:05.813518 kernel: kvm_amd: TSC scaling supported Feb 13 19:50:05.813648 kernel: kvm_amd: Nested Virtualization enabled Feb 13 19:50:05.813670 kernel: kvm_amd: Nested Paging enabled Feb 13 19:50:05.814114 kernel: kvm_amd: LBR virtualization supported Feb 13 19:50:05.815479 kernel: kvm_amd: Virtual VMLOAD VMSAVE supported Feb 13 19:50:05.815508 kernel: kvm_amd: Virtual GIF supported Feb 13 19:50:05.840165 kernel: EDAC MC: Ver: 3.0.0 Feb 13 19:50:05.849846 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 19:50:05.869268 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Feb 13 19:50:05.880083 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Feb 13 19:50:05.890462 lvm[1433]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Feb 13 19:50:05.924333 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Feb 13 19:50:05.926805 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Feb 13 19:50:05.928109 systemd[1]: Reached target sysinit.target - System Initialization. Feb 13 19:50:05.929462 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Feb 13 19:50:05.930893 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Feb 13 19:50:05.932523 systemd[1]: Started logrotate.timer - Daily rotation of log files. Feb 13 19:50:05.933829 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Feb 13 19:50:05.935251 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Feb 13 19:50:05.936649 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Feb 13 19:50:05.936677 systemd[1]: Reached target paths.target - Path Units. Feb 13 19:50:05.937688 systemd[1]: Reached target timers.target - Timer Units. Feb 13 19:50:05.939627 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Feb 13 19:50:05.942549 systemd[1]: Starting docker.socket - Docker Socket for the API... Feb 13 19:50:05.952546 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Feb 13 19:50:05.955605 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Feb 13 19:50:05.957595 systemd[1]: Listening on docker.socket - Docker Socket for the API. Feb 13 19:50:05.958984 systemd[1]: Reached target sockets.target - Socket Units. Feb 13 19:50:05.960113 systemd[1]: Reached target basic.target - Basic System. Feb 13 19:50:05.961117 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Feb 13 19:50:05.961148 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Feb 13 19:50:05.962434 systemd[1]: Starting containerd.service - containerd container runtime... Feb 13 19:50:05.964825 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Feb 13 19:50:05.969461 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Feb 13 19:50:05.972419 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Feb 13 19:50:05.973919 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Feb 13 19:50:05.976300 lvm[1437]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Feb 13 19:50:05.978036 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Feb 13 19:50:05.981485 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Feb 13 19:50:05.983013 jq[1440]: false Feb 13 19:50:05.984015 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Feb 13 19:50:05.988016 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Feb 13 19:50:05.992229 systemd[1]: Starting systemd-logind.service - User Login Management... Feb 13 19:50:05.998532 extend-filesystems[1441]: Found loop3 Feb 13 19:50:06.053784 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Feb 13 19:50:06.053819 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 38 scanned by (udev-worker) (1390) Feb 13 19:50:06.045125 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Feb 13 19:50:06.001289 dbus-daemon[1439]: [system] SELinux support is enabled Feb 13 19:50:06.054284 extend-filesystems[1441]: Found loop4 Feb 13 19:50:06.054284 extend-filesystems[1441]: Found loop5 Feb 13 19:50:06.054284 extend-filesystems[1441]: Found sr0 Feb 13 19:50:06.054284 extend-filesystems[1441]: Found vda Feb 13 19:50:06.054284 extend-filesystems[1441]: Found vda1 Feb 13 19:50:06.054284 extend-filesystems[1441]: Found vda2 Feb 13 19:50:06.054284 extend-filesystems[1441]: Found vda3 Feb 13 19:50:06.054284 extend-filesystems[1441]: Found usr Feb 13 19:50:06.054284 extend-filesystems[1441]: Found vda4 Feb 13 19:50:06.054284 extend-filesystems[1441]: Found vda6 Feb 13 19:50:06.054284 extend-filesystems[1441]: Found vda7 Feb 13 19:50:06.054284 extend-filesystems[1441]: Found vda9 Feb 13 19:50:06.054284 extend-filesystems[1441]: Checking size of /dev/vda9 Feb 13 19:50:06.054284 extend-filesystems[1441]: Resized partition /dev/vda9 Feb 13 19:50:06.101328 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Feb 13 19:50:06.045864 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Feb 13 19:50:06.102744 extend-filesystems[1457]: resize2fs 1.47.1 (20-May-2024) Feb 13 19:50:06.102744 extend-filesystems[1457]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Feb 13 19:50:06.102744 extend-filesystems[1457]: old_desc_blocks = 1, new_desc_blocks = 1 Feb 13 19:50:06.102744 extend-filesystems[1457]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Feb 13 19:50:06.048274 systemd[1]: Starting update-engine.service - Update Engine... Feb 13 19:50:06.115105 extend-filesystems[1441]: Resized filesystem in /dev/vda9 Feb 13 19:50:06.056002 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Feb 13 19:50:06.117761 update_engine[1459]: I20250213 19:50:06.083340 1459 main.cc:92] Flatcar Update Engine starting Feb 13 19:50:06.117761 update_engine[1459]: I20250213 19:50:06.087101 1459 update_check_scheduler.cc:74] Next update check in 9m2s Feb 13 19:50:06.067628 systemd[1]: Started dbus.service - D-Bus System Message Bus. Feb 13 19:50:06.120411 jq[1460]: true Feb 13 19:50:06.072287 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Feb 13 19:50:06.076368 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Feb 13 19:50:06.120837 tar[1464]: linux-amd64/LICENSE Feb 13 19:50:06.120837 tar[1464]: linux-amd64/helm Feb 13 19:50:06.076614 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Feb 13 19:50:06.077181 systemd[1]: motdgen.service: Deactivated successfully. Feb 13 19:50:06.123088 jq[1466]: true Feb 13 19:50:06.077377 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Feb 13 19:50:06.080316 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Feb 13 19:50:06.080546 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Feb 13 19:50:06.085834 systemd[1]: extend-filesystems.service: Deactivated successfully. Feb 13 19:50:06.086132 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Feb 13 19:50:06.100014 (ntainerd)[1468]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Feb 13 19:50:06.133999 systemd[1]: Started update-engine.service - Update Engine. Feb 13 19:50:06.135542 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Feb 13 19:50:06.135571 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Feb 13 19:50:06.137010 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Feb 13 19:50:06.137029 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Feb 13 19:50:06.137446 systemd-logind[1447]: Watching system buttons on /dev/input/event1 (Power Button) Feb 13 19:50:06.137706 systemd-logind[1447]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Feb 13 19:50:06.137996 systemd-logind[1447]: New seat seat0. Feb 13 19:50:06.147075 systemd[1]: Started locksmithd.service - Cluster reboot manager. Feb 13 19:50:06.148393 systemd[1]: Started systemd-logind.service - User Login Management. Feb 13 19:50:06.231591 sshd_keygen[1467]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Feb 13 19:50:06.251310 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Feb 13 19:50:06.267548 locksmithd[1492]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Feb 13 19:50:06.288654 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Feb 13 19:50:06.298105 systemd[1]: Starting issuegen.service - Generate /run/issue... Feb 13 19:50:06.302333 systemd[1]: Started sshd@0-10.0.0.104:22-10.0.0.1:54668.service - OpenSSH per-connection server daemon (10.0.0.1:54668). Feb 13 19:50:06.308578 systemd[1]: issuegen.service: Deactivated successfully. Feb 13 19:50:06.308807 systemd[1]: Finished issuegen.service - Generate /run/issue. Feb 13 19:50:06.314231 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Feb 13 19:50:06.378367 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Feb 13 19:50:06.387257 systemd[1]: Started getty@tty1.service - Getty on tty1. Feb 13 19:50:06.390119 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Feb 13 19:50:06.391502 systemd[1]: Reached target getty.target - Login Prompts. Feb 13 19:50:06.478212 sshd[1510]: Connection closed by authenticating user core 10.0.0.1 port 54668 [preauth] Feb 13 19:50:06.481691 systemd[1]: sshd@0-10.0.0.104:22-10.0.0.1:54668.service: Deactivated successfully. Feb 13 19:50:06.504059 bash[1495]: Updated "/home/core/.ssh/authorized_keys" Feb 13 19:50:06.505556 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Feb 13 19:50:06.508660 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Feb 13 19:50:06.627215 containerd[1468]: time="2025-02-13T19:50:06.627110799Z" level=info msg="starting containerd" revision=9b2ad7760328148397346d10c7b2004271249db4 version=v1.7.23 Feb 13 19:50:06.652830 containerd[1468]: time="2025-02-13T19:50:06.652764971Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Feb 13 19:50:06.654835 containerd[1468]: time="2025-02-13T19:50:06.654774549Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.74-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Feb 13 19:50:06.654835 containerd[1468]: time="2025-02-13T19:50:06.654817920Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Feb 13 19:50:06.654835 containerd[1468]: time="2025-02-13T19:50:06.654837196Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Feb 13 19:50:06.655081 containerd[1468]: time="2025-02-13T19:50:06.655054594Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Feb 13 19:50:06.655081 containerd[1468]: time="2025-02-13T19:50:06.655077386Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Feb 13 19:50:06.655164 containerd[1468]: time="2025-02-13T19:50:06.655145194Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Feb 13 19:50:06.655164 containerd[1468]: time="2025-02-13T19:50:06.655162065Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Feb 13 19:50:06.655382 containerd[1468]: time="2025-02-13T19:50:06.655354736Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Feb 13 19:50:06.655382 containerd[1468]: time="2025-02-13T19:50:06.655375906Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Feb 13 19:50:06.655420 containerd[1468]: time="2025-02-13T19:50:06.655391305Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Feb 13 19:50:06.655420 containerd[1468]: time="2025-02-13T19:50:06.655401765Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Feb 13 19:50:06.655509 containerd[1468]: time="2025-02-13T19:50:06.655492445Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Feb 13 19:50:06.655753 containerd[1468]: time="2025-02-13T19:50:06.655726924Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Feb 13 19:50:06.655885 containerd[1468]: time="2025-02-13T19:50:06.655846689Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Feb 13 19:50:06.655885 containerd[1468]: time="2025-02-13T19:50:06.655863661Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Feb 13 19:50:06.656016 containerd[1468]: time="2025-02-13T19:50:06.655991360Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Feb 13 19:50:06.656069 containerd[1468]: time="2025-02-13T19:50:06.656052525Z" level=info msg="metadata content store policy set" policy=shared Feb 13 19:50:06.756618 tar[1464]: linux-amd64/README.md Feb 13 19:50:06.783514 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Feb 13 19:50:06.849523 containerd[1468]: time="2025-02-13T19:50:06.849443864Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Feb 13 19:50:06.849523 containerd[1468]: time="2025-02-13T19:50:06.849529745Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Feb 13 19:50:06.849685 containerd[1468]: time="2025-02-13T19:50:06.849546907Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Feb 13 19:50:06.849685 containerd[1468]: time="2025-02-13T19:50:06.849564710Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Feb 13 19:50:06.849685 containerd[1468]: time="2025-02-13T19:50:06.849578736Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Feb 13 19:50:06.849837 containerd[1468]: time="2025-02-13T19:50:06.849807425Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Feb 13 19:50:06.850229 containerd[1468]: time="2025-02-13T19:50:06.850167811Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Feb 13 19:50:06.850448 containerd[1468]: time="2025-02-13T19:50:06.850420224Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Feb 13 19:50:06.850448 containerd[1468]: time="2025-02-13T19:50:06.850443047Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Feb 13 19:50:06.850494 containerd[1468]: time="2025-02-13T19:50:06.850461221Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Feb 13 19:50:06.850494 containerd[1468]: time="2025-02-13T19:50:06.850479135Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Feb 13 19:50:06.850494 containerd[1468]: time="2025-02-13T19:50:06.850492890Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Feb 13 19:50:06.850547 containerd[1468]: time="2025-02-13T19:50:06.850507147Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Feb 13 19:50:06.850547 containerd[1468]: time="2025-02-13T19:50:06.850523047Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Feb 13 19:50:06.850586 containerd[1468]: time="2025-02-13T19:50:06.850546952Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Feb 13 19:50:06.850586 containerd[1468]: time="2025-02-13T19:50:06.850562100Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Feb 13 19:50:06.850586 containerd[1468]: time="2025-02-13T19:50:06.850574804Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Feb 13 19:50:06.850645 containerd[1468]: time="2025-02-13T19:50:06.850586135Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Feb 13 19:50:06.850645 containerd[1468]: time="2025-02-13T19:50:06.850608147Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Feb 13 19:50:06.850645 containerd[1468]: time="2025-02-13T19:50:06.850621371Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Feb 13 19:50:06.850645 containerd[1468]: time="2025-02-13T19:50:06.850633835Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Feb 13 19:50:06.850645 containerd[1468]: time="2025-02-13T19:50:06.850646078Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Feb 13 19:50:06.850741 containerd[1468]: time="2025-02-13T19:50:06.850658301Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Feb 13 19:50:06.850741 containerd[1468]: time="2025-02-13T19:50:06.850676094Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Feb 13 19:50:06.850741 containerd[1468]: time="2025-02-13T19:50:06.850687185Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Feb 13 19:50:06.850741 containerd[1468]: time="2025-02-13T19:50:06.850699638Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Feb 13 19:50:06.850741 containerd[1468]: time="2025-02-13T19:50:06.850712683Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Feb 13 19:50:06.850741 containerd[1468]: time="2025-02-13T19:50:06.850728111Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Feb 13 19:50:06.850741 containerd[1468]: time="2025-02-13T19:50:06.850739693Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Feb 13 19:50:06.850888 containerd[1468]: time="2025-02-13T19:50:06.850753008Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Feb 13 19:50:06.850888 containerd[1468]: time="2025-02-13T19:50:06.850765301Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Feb 13 19:50:06.850888 containerd[1468]: time="2025-02-13T19:50:06.850778726Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Feb 13 19:50:06.850888 containerd[1468]: time="2025-02-13T19:50:06.850798904Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Feb 13 19:50:06.850888 containerd[1468]: time="2025-02-13T19:50:06.850811297Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Feb 13 19:50:06.850888 containerd[1468]: time="2025-02-13T19:50:06.850821757Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Feb 13 19:50:06.851596 containerd[1468]: time="2025-02-13T19:50:06.851565161Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Feb 13 19:50:06.851622 containerd[1468]: time="2025-02-13T19:50:06.851593955Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Feb 13 19:50:06.851622 containerd[1468]: time="2025-02-13T19:50:06.851606999Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Feb 13 19:50:06.851622 containerd[1468]: time="2025-02-13T19:50:06.851619763Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Feb 13 19:50:06.851687 containerd[1468]: time="2025-02-13T19:50:06.851629181Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Feb 13 19:50:06.851687 containerd[1468]: time="2025-02-13T19:50:06.851641334Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Feb 13 19:50:06.851687 containerd[1468]: time="2025-02-13T19:50:06.851652074Z" level=info msg="NRI interface is disabled by configuration." Feb 13 19:50:06.851687 containerd[1468]: time="2025-02-13T19:50:06.851661972Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Feb 13 19:50:06.852017 containerd[1468]: time="2025-02-13T19:50:06.851966233Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Feb 13 19:50:06.852017 containerd[1468]: time="2025-02-13T19:50:06.852011838Z" level=info msg="Connect containerd service" Feb 13 19:50:06.852185 containerd[1468]: time="2025-02-13T19:50:06.852043869Z" level=info msg="using legacy CRI server" Feb 13 19:50:06.852185 containerd[1468]: time="2025-02-13T19:50:06.852054178Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Feb 13 19:50:06.852185 containerd[1468]: time="2025-02-13T19:50:06.852162812Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Feb 13 19:50:06.852808 containerd[1468]: time="2025-02-13T19:50:06.852774008Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Feb 13 19:50:06.853042 containerd[1468]: time="2025-02-13T19:50:06.852961699Z" level=info msg="Start subscribing containerd event" Feb 13 19:50:06.853042 containerd[1468]: time="2025-02-13T19:50:06.853052390Z" level=info msg="Start recovering state" Feb 13 19:50:06.853327 containerd[1468]: time="2025-02-13T19:50:06.853124304Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Feb 13 19:50:06.853327 containerd[1468]: time="2025-02-13T19:50:06.853145865Z" level=info msg="Start event monitor" Feb 13 19:50:06.853327 containerd[1468]: time="2025-02-13T19:50:06.853172445Z" level=info msg="Start snapshots syncer" Feb 13 19:50:06.853327 containerd[1468]: time="2025-02-13T19:50:06.853183235Z" level=info msg=serving... address=/run/containerd/containerd.sock Feb 13 19:50:06.853327 containerd[1468]: time="2025-02-13T19:50:06.853186571Z" level=info msg="Start cni network conf syncer for default" Feb 13 19:50:06.853327 containerd[1468]: time="2025-02-13T19:50:06.853244400Z" level=info msg="Start streaming server" Feb 13 19:50:06.853440 systemd[1]: Started containerd.service - containerd container runtime. Feb 13 19:50:06.853956 containerd[1468]: time="2025-02-13T19:50:06.853927390Z" level=info msg="containerd successfully booted in 0.228648s" Feb 13 19:50:07.051133 systemd-networkd[1392]: eth0: Gained IPv6LL Feb 13 19:50:07.054544 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Feb 13 19:50:07.056496 systemd[1]: Reached target network-online.target - Network is Online. Feb 13 19:50:07.069267 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Feb 13 19:50:07.072680 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:50:07.075304 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Feb 13 19:50:07.096602 systemd[1]: coreos-metadata.service: Deactivated successfully. Feb 13 19:50:07.096907 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Feb 13 19:50:07.098666 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Feb 13 19:50:07.104064 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Feb 13 19:50:08.381382 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:50:08.383496 systemd[1]: Reached target multi-user.target - Multi-User System. Feb 13 19:50:08.385228 systemd[1]: Startup finished in 843ms (kernel) + 6.512s (initrd) + 5.109s (userspace) = 12.465s. Feb 13 19:50:08.390537 (kubelet)[1556]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 19:50:08.958146 kubelet[1556]: E0213 19:50:08.958067 1556 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 19:50:08.962752 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 19:50:08.963051 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 19:50:08.963482 systemd[1]: kubelet.service: Consumed 1.671s CPU time. Feb 13 19:50:16.462440 systemd[1]: Started sshd@1-10.0.0.104:22-10.0.0.1:54502.service - OpenSSH per-connection server daemon (10.0.0.1:54502). Feb 13 19:50:16.515466 sshd[1570]: Accepted publickey for core from 10.0.0.1 port 54502 ssh2: RSA SHA256:8WP2kqV5KzwZsuVRMXRFZkAHZWbkdD5kizbT2H+wOcw Feb 13 19:50:16.517506 sshd-session[1570]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:50:16.529227 systemd-logind[1447]: New session 1 of user core. Feb 13 19:50:16.531099 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Feb 13 19:50:16.545207 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Feb 13 19:50:16.559220 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Feb 13 19:50:16.571299 systemd[1]: Starting user@500.service - User Manager for UID 500... Feb 13 19:50:16.574618 (systemd)[1574]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Feb 13 19:50:16.710267 systemd[1574]: Queued start job for default target default.target. Feb 13 19:50:16.722466 systemd[1574]: Created slice app.slice - User Application Slice. Feb 13 19:50:16.722498 systemd[1574]: Reached target paths.target - Paths. Feb 13 19:50:16.722513 systemd[1574]: Reached target timers.target - Timers. Feb 13 19:50:16.724285 systemd[1574]: Starting dbus.socket - D-Bus User Message Bus Socket... Feb 13 19:50:16.736338 systemd[1574]: Listening on dbus.socket - D-Bus User Message Bus Socket. Feb 13 19:50:16.736522 systemd[1574]: Reached target sockets.target - Sockets. Feb 13 19:50:16.736550 systemd[1574]: Reached target basic.target - Basic System. Feb 13 19:50:16.736606 systemd[1574]: Reached target default.target - Main User Target. Feb 13 19:50:16.736664 systemd[1574]: Startup finished in 154ms. Feb 13 19:50:16.737066 systemd[1]: Started user@500.service - User Manager for UID 500. Feb 13 19:50:16.738647 systemd[1]: Started session-1.scope - Session 1 of User core. Feb 13 19:50:16.800299 systemd[1]: Started sshd@2-10.0.0.104:22-10.0.0.1:54510.service - OpenSSH per-connection server daemon (10.0.0.1:54510). Feb 13 19:50:16.845767 sshd[1586]: Accepted publickey for core from 10.0.0.1 port 54510 ssh2: RSA SHA256:8WP2kqV5KzwZsuVRMXRFZkAHZWbkdD5kizbT2H+wOcw Feb 13 19:50:16.847385 sshd-session[1586]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:50:16.851827 systemd-logind[1447]: New session 2 of user core. Feb 13 19:50:16.867162 systemd[1]: Started session-2.scope - Session 2 of User core. Feb 13 19:50:16.921793 sshd[1588]: Connection closed by 10.0.0.1 port 54510 Feb 13 19:50:16.922183 sshd-session[1586]: pam_unix(sshd:session): session closed for user core Feb 13 19:50:16.933171 systemd[1]: sshd@2-10.0.0.104:22-10.0.0.1:54510.service: Deactivated successfully. Feb 13 19:50:16.935509 systemd[1]: session-2.scope: Deactivated successfully. Feb 13 19:50:16.937613 systemd-logind[1447]: Session 2 logged out. Waiting for processes to exit. Feb 13 19:50:16.954375 systemd[1]: Started sshd@3-10.0.0.104:22-10.0.0.1:54516.service - OpenSSH per-connection server daemon (10.0.0.1:54516). Feb 13 19:50:16.955889 systemd-logind[1447]: Removed session 2. Feb 13 19:50:16.993476 sshd[1593]: Accepted publickey for core from 10.0.0.1 port 54516 ssh2: RSA SHA256:8WP2kqV5KzwZsuVRMXRFZkAHZWbkdD5kizbT2H+wOcw Feb 13 19:50:16.994945 sshd-session[1593]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:50:16.999168 systemd-logind[1447]: New session 3 of user core. Feb 13 19:50:17.009031 systemd[1]: Started session-3.scope - Session 3 of User core. Feb 13 19:50:17.060025 sshd[1595]: Connection closed by 10.0.0.1 port 54516 Feb 13 19:50:17.060410 sshd-session[1593]: pam_unix(sshd:session): session closed for user core Feb 13 19:50:17.068169 systemd[1]: sshd@3-10.0.0.104:22-10.0.0.1:54516.service: Deactivated successfully. Feb 13 19:50:17.070224 systemd[1]: session-3.scope: Deactivated successfully. Feb 13 19:50:17.072515 systemd-logind[1447]: Session 3 logged out. Waiting for processes to exit. Feb 13 19:50:17.085300 systemd[1]: Started sshd@4-10.0.0.104:22-10.0.0.1:54522.service - OpenSSH per-connection server daemon (10.0.0.1:54522). Feb 13 19:50:17.086251 systemd-logind[1447]: Removed session 3. Feb 13 19:50:17.125041 sshd[1600]: Accepted publickey for core from 10.0.0.1 port 54522 ssh2: RSA SHA256:8WP2kqV5KzwZsuVRMXRFZkAHZWbkdD5kizbT2H+wOcw Feb 13 19:50:17.127155 sshd-session[1600]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:50:17.131594 systemd-logind[1447]: New session 4 of user core. Feb 13 19:50:17.145201 systemd[1]: Started session-4.scope - Session 4 of User core. Feb 13 19:50:17.200504 sshd[1602]: Connection closed by 10.0.0.1 port 54522 Feb 13 19:50:17.200935 sshd-session[1600]: pam_unix(sshd:session): session closed for user core Feb 13 19:50:17.211793 systemd[1]: sshd@4-10.0.0.104:22-10.0.0.1:54522.service: Deactivated successfully. Feb 13 19:50:17.213692 systemd[1]: session-4.scope: Deactivated successfully. Feb 13 19:50:17.215489 systemd-logind[1447]: Session 4 logged out. Waiting for processes to exit. Feb 13 19:50:17.225169 systemd[1]: Started sshd@5-10.0.0.104:22-10.0.0.1:54536.service - OpenSSH per-connection server daemon (10.0.0.1:54536). Feb 13 19:50:17.226026 systemd-logind[1447]: Removed session 4. Feb 13 19:50:17.267306 sshd[1607]: Accepted publickey for core from 10.0.0.1 port 54536 ssh2: RSA SHA256:8WP2kqV5KzwZsuVRMXRFZkAHZWbkdD5kizbT2H+wOcw Feb 13 19:50:17.268982 sshd-session[1607]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:50:17.273338 systemd-logind[1447]: New session 5 of user core. Feb 13 19:50:17.283101 systemd[1]: Started session-5.scope - Session 5 of User core. Feb 13 19:50:17.342988 sudo[1610]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Feb 13 19:50:17.343335 sudo[1610]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Feb 13 19:50:17.374566 sudo[1610]: pam_unix(sudo:session): session closed for user root Feb 13 19:50:17.376427 sshd[1609]: Connection closed by 10.0.0.1 port 54536 Feb 13 19:50:17.376808 sshd-session[1607]: pam_unix(sshd:session): session closed for user core Feb 13 19:50:17.390143 systemd[1]: sshd@5-10.0.0.104:22-10.0.0.1:54536.service: Deactivated successfully. Feb 13 19:50:17.392102 systemd[1]: session-5.scope: Deactivated successfully. Feb 13 19:50:17.393844 systemd-logind[1447]: Session 5 logged out. Waiting for processes to exit. Feb 13 19:50:17.408309 systemd[1]: Started sshd@6-10.0.0.104:22-10.0.0.1:54548.service - OpenSSH per-connection server daemon (10.0.0.1:54548). Feb 13 19:50:17.409621 systemd-logind[1447]: Removed session 5. Feb 13 19:50:17.448735 sshd[1615]: Accepted publickey for core from 10.0.0.1 port 54548 ssh2: RSA SHA256:8WP2kqV5KzwZsuVRMXRFZkAHZWbkdD5kizbT2H+wOcw Feb 13 19:50:17.450454 sshd-session[1615]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:50:17.455152 systemd-logind[1447]: New session 6 of user core. Feb 13 19:50:17.466134 systemd[1]: Started session-6.scope - Session 6 of User core. Feb 13 19:50:17.523228 sudo[1619]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Feb 13 19:50:17.523679 sudo[1619]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Feb 13 19:50:17.528800 sudo[1619]: pam_unix(sudo:session): session closed for user root Feb 13 19:50:17.535224 sudo[1618]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Feb 13 19:50:17.535619 sudo[1618]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Feb 13 19:50:17.559231 systemd[1]: Starting audit-rules.service - Load Audit Rules... Feb 13 19:50:17.586462 augenrules[1641]: No rules Feb 13 19:50:17.587257 systemd[1]: audit-rules.service: Deactivated successfully. Feb 13 19:50:17.587486 systemd[1]: Finished audit-rules.service - Load Audit Rules. Feb 13 19:50:17.588625 sudo[1618]: pam_unix(sudo:session): session closed for user root Feb 13 19:50:17.590005 sshd[1617]: Connection closed by 10.0.0.1 port 54548 Feb 13 19:50:17.590358 sshd-session[1615]: pam_unix(sshd:session): session closed for user core Feb 13 19:50:17.609568 systemd[1]: sshd@6-10.0.0.104:22-10.0.0.1:54548.service: Deactivated successfully. Feb 13 19:50:17.611272 systemd[1]: session-6.scope: Deactivated successfully. Feb 13 19:50:17.612621 systemd-logind[1447]: Session 6 logged out. Waiting for processes to exit. Feb 13 19:50:17.613833 systemd[1]: Started sshd@7-10.0.0.104:22-10.0.0.1:54552.service - OpenSSH per-connection server daemon (10.0.0.1:54552). Feb 13 19:50:17.614636 systemd-logind[1447]: Removed session 6. Feb 13 19:50:17.656346 sshd[1649]: Accepted publickey for core from 10.0.0.1 port 54552 ssh2: RSA SHA256:8WP2kqV5KzwZsuVRMXRFZkAHZWbkdD5kizbT2H+wOcw Feb 13 19:50:17.657867 sshd-session[1649]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:50:17.661823 systemd-logind[1447]: New session 7 of user core. Feb 13 19:50:17.668995 systemd[1]: Started session-7.scope - Session 7 of User core. Feb 13 19:50:17.721666 sudo[1652]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Feb 13 19:50:17.722013 sudo[1652]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Feb 13 19:50:17.998176 systemd[1]: Starting docker.service - Docker Application Container Engine... Feb 13 19:50:17.998310 (dockerd)[1672]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Feb 13 19:50:18.262712 dockerd[1672]: time="2025-02-13T19:50:18.262545436Z" level=info msg="Starting up" Feb 13 19:50:18.380960 dockerd[1672]: time="2025-02-13T19:50:18.380888409Z" level=info msg="Loading containers: start." Feb 13 19:50:18.592184 kernel: Initializing XFRM netlink socket Feb 13 19:50:18.694554 systemd-networkd[1392]: docker0: Link UP Feb 13 19:50:18.737729 dockerd[1672]: time="2025-02-13T19:50:18.737672209Z" level=info msg="Loading containers: done." Feb 13 19:50:18.755163 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck4134757260-merged.mount: Deactivated successfully. Feb 13 19:50:18.756708 dockerd[1672]: time="2025-02-13T19:50:18.756643690Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Feb 13 19:50:18.756815 dockerd[1672]: time="2025-02-13T19:50:18.756783222Z" level=info msg="Docker daemon" commit=8b539b8df24032dabeaaa099cf1d0535ef0286a3 containerd-snapshotter=false storage-driver=overlay2 version=27.2.1 Feb 13 19:50:18.756999 dockerd[1672]: time="2025-02-13T19:50:18.756969361Z" level=info msg="Daemon has completed initialization" Feb 13 19:50:18.799476 dockerd[1672]: time="2025-02-13T19:50:18.799186428Z" level=info msg="API listen on /run/docker.sock" Feb 13 19:50:18.799752 systemd[1]: Started docker.service - Docker Application Container Engine. Feb 13 19:50:19.210508 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Feb 13 19:50:19.224216 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:50:19.370315 containerd[1468]: time="2025-02-13T19:50:19.370253659Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.2\"" Feb 13 19:50:19.427545 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:50:19.433370 (kubelet)[1877]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 19:50:19.611163 kubelet[1877]: E0213 19:50:19.610960 1877 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 19:50:19.617558 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 19:50:19.617804 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 19:50:20.144423 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1607241866.mount: Deactivated successfully. Feb 13 19:50:22.568891 containerd[1468]: time="2025-02-13T19:50:22.568810372Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.32.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:50:22.569749 containerd[1468]: time="2025-02-13T19:50:22.569702725Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.32.2: active requests=0, bytes read=28673931" Feb 13 19:50:22.571393 containerd[1468]: time="2025-02-13T19:50:22.571362807Z" level=info msg="ImageCreate event name:\"sha256:85b7a174738baecbc53029b7913cd430a2060e0cbdb5f56c7957d32ff7f241ef\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:50:22.574705 containerd[1468]: time="2025-02-13T19:50:22.574676681Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:c47449f3e751588ea0cb74e325e0f83db335a415f4f4c7fb147375dd6c84757f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:50:22.576111 containerd[1468]: time="2025-02-13T19:50:22.576053082Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.32.2\" with image id \"sha256:85b7a174738baecbc53029b7913cd430a2060e0cbdb5f56c7957d32ff7f241ef\", repo tag \"registry.k8s.io/kube-apiserver:v1.32.2\", repo digest \"registry.k8s.io/kube-apiserver@sha256:c47449f3e751588ea0cb74e325e0f83db335a415f4f4c7fb147375dd6c84757f\", size \"28670731\" in 3.20574437s" Feb 13 19:50:22.576168 containerd[1468]: time="2025-02-13T19:50:22.576117723Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.2\" returns image reference \"sha256:85b7a174738baecbc53029b7913cd430a2060e0cbdb5f56c7957d32ff7f241ef\"" Feb 13 19:50:22.576987 containerd[1468]: time="2025-02-13T19:50:22.576947238Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.2\"" Feb 13 19:50:24.129424 containerd[1468]: time="2025-02-13T19:50:24.129348650Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.32.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:50:24.130488 containerd[1468]: time="2025-02-13T19:50:24.130406043Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.32.2: active requests=0, bytes read=24771784" Feb 13 19:50:24.132219 containerd[1468]: time="2025-02-13T19:50:24.132183195Z" level=info msg="ImageCreate event name:\"sha256:b6a454c5a800d201daacead6ff195ec6049fe6dc086621b0670bca912efaf389\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:50:24.135316 containerd[1468]: time="2025-02-13T19:50:24.135258381Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:399aa50f4d1361c59dc458e634506d02de32613d03a9a614a21058741162ef90\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:50:24.136728 containerd[1468]: time="2025-02-13T19:50:24.136682351Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.32.2\" with image id \"sha256:b6a454c5a800d201daacead6ff195ec6049fe6dc086621b0670bca912efaf389\", repo tag \"registry.k8s.io/kube-controller-manager:v1.32.2\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:399aa50f4d1361c59dc458e634506d02de32613d03a9a614a21058741162ef90\", size \"26259392\" in 1.559687243s" Feb 13 19:50:24.136728 containerd[1468]: time="2025-02-13T19:50:24.136725883Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.2\" returns image reference \"sha256:b6a454c5a800d201daacead6ff195ec6049fe6dc086621b0670bca912efaf389\"" Feb 13 19:50:24.137435 containerd[1468]: time="2025-02-13T19:50:24.137227163Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.2\"" Feb 13 19:50:26.234049 containerd[1468]: time="2025-02-13T19:50:26.233962300Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.32.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:50:26.235328 containerd[1468]: time="2025-02-13T19:50:26.235242100Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.32.2: active requests=0, bytes read=19170276" Feb 13 19:50:26.237039 containerd[1468]: time="2025-02-13T19:50:26.236981742Z" level=info msg="ImageCreate event name:\"sha256:d8e673e7c9983f1f53569a9d2ba786c8abb42e3f744f77dc97a595f3caf9435d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:50:26.240628 containerd[1468]: time="2025-02-13T19:50:26.240514566Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:45710d74cfd5aa10a001d0cf81747b77c28617444ffee0503d12f1dcd7450f76\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:50:26.241854 containerd[1468]: time="2025-02-13T19:50:26.241797011Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.32.2\" with image id \"sha256:d8e673e7c9983f1f53569a9d2ba786c8abb42e3f744f77dc97a595f3caf9435d\", repo tag \"registry.k8s.io/kube-scheduler:v1.32.2\", repo digest \"registry.k8s.io/kube-scheduler@sha256:45710d74cfd5aa10a001d0cf81747b77c28617444ffee0503d12f1dcd7450f76\", size \"20657902\" in 2.104534121s" Feb 13 19:50:26.241854 containerd[1468]: time="2025-02-13T19:50:26.241833319Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.2\" returns image reference \"sha256:d8e673e7c9983f1f53569a9d2ba786c8abb42e3f744f77dc97a595f3caf9435d\"" Feb 13 19:50:26.242448 containerd[1468]: time="2025-02-13T19:50:26.242344738Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.2\"" Feb 13 19:50:27.437113 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount855109235.mount: Deactivated successfully. Feb 13 19:50:28.834912 containerd[1468]: time="2025-02-13T19:50:28.834419798Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.32.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:50:28.835708 containerd[1468]: time="2025-02-13T19:50:28.835664481Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.32.2: active requests=0, bytes read=30908839" Feb 13 19:50:28.837645 containerd[1468]: time="2025-02-13T19:50:28.837584882Z" level=info msg="ImageCreate event name:\"sha256:f1332858868e1c6a905123b21e2e322ab45a5b99a3532e68ff49a87c2266ebc5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:50:28.841989 containerd[1468]: time="2025-02-13T19:50:28.841939227Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:83c025f0faa6799fab6645102a98138e39a9a7db2be3bc792c79d72659b1805d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:50:28.842515 containerd[1468]: time="2025-02-13T19:50:28.842466104Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.32.2\" with image id \"sha256:f1332858868e1c6a905123b21e2e322ab45a5b99a3532e68ff49a87c2266ebc5\", repo tag \"registry.k8s.io/kube-proxy:v1.32.2\", repo digest \"registry.k8s.io/kube-proxy@sha256:83c025f0faa6799fab6645102a98138e39a9a7db2be3bc792c79d72659b1805d\", size \"30907858\" in 2.600016059s" Feb 13 19:50:28.842515 containerd[1468]: time="2025-02-13T19:50:28.842511890Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.2\" returns image reference \"sha256:f1332858868e1c6a905123b21e2e322ab45a5b99a3532e68ff49a87c2266ebc5\"" Feb 13 19:50:28.843167 containerd[1468]: time="2025-02-13T19:50:28.843108218Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Feb 13 19:50:29.492238 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3045464632.mount: Deactivated successfully. Feb 13 19:50:29.710480 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Feb 13 19:50:29.718214 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:50:30.001954 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:50:30.009025 (kubelet)[1977]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 19:50:30.127158 kubelet[1977]: E0213 19:50:30.127069 1977 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 19:50:30.131879 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 19:50:30.132161 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 19:50:32.508700 containerd[1468]: time="2025-02-13T19:50:32.508609923Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:50:32.534449 containerd[1468]: time="2025-02-13T19:50:32.534346559Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=18565241" Feb 13 19:50:32.546930 containerd[1468]: time="2025-02-13T19:50:32.546858409Z" level=info msg="ImageCreate event name:\"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:50:32.554639 containerd[1468]: time="2025-02-13T19:50:32.554548979Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:50:32.556082 containerd[1468]: time="2025-02-13T19:50:32.556020648Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"18562039\" in 3.712841136s" Feb 13 19:50:32.556145 containerd[1468]: time="2025-02-13T19:50:32.556091261Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\"" Feb 13 19:50:32.556668 containerd[1468]: time="2025-02-13T19:50:32.556639709Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Feb 13 19:50:33.074564 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount526828744.mount: Deactivated successfully. Feb 13 19:50:33.082485 containerd[1468]: time="2025-02-13T19:50:33.082415869Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:50:33.083559 containerd[1468]: time="2025-02-13T19:50:33.083510181Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Feb 13 19:50:33.085331 containerd[1468]: time="2025-02-13T19:50:33.085278296Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:50:33.088014 containerd[1468]: time="2025-02-13T19:50:33.087966747Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:50:33.088730 containerd[1468]: time="2025-02-13T19:50:33.088673933Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 531.995001ms" Feb 13 19:50:33.088730 containerd[1468]: time="2025-02-13T19:50:33.088724428Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Feb 13 19:50:33.089340 containerd[1468]: time="2025-02-13T19:50:33.089309314Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\"" Feb 13 19:50:34.212993 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2809578183.mount: Deactivated successfully. Feb 13 19:50:35.899489 containerd[1468]: time="2025-02-13T19:50:35.899406887Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.16-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:50:35.900724 containerd[1468]: time="2025-02-13T19:50:35.900644928Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.16-0: active requests=0, bytes read=57551320" Feb 13 19:50:35.904093 containerd[1468]: time="2025-02-13T19:50:35.904027731Z" level=info msg="ImageCreate event name:\"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:50:35.910833 containerd[1468]: time="2025-02-13T19:50:35.910744114Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:50:35.912108 containerd[1468]: time="2025-02-13T19:50:35.912034133Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.16-0\" with image id \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\", repo tag \"registry.k8s.io/etcd:3.5.16-0\", repo digest \"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\", size \"57680541\" in 2.822684132s" Feb 13 19:50:35.912108 containerd[1468]: time="2025-02-13T19:50:35.912095387Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\" returns image reference \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\"" Feb 13 19:50:38.548174 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:50:38.559193 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:50:38.582590 systemd[1]: Reloading requested from client PID 2113 ('systemctl') (unit session-7.scope)... Feb 13 19:50:38.582605 systemd[1]: Reloading... Feb 13 19:50:38.714906 zram_generator::config[2155]: No configuration found. Feb 13 19:50:39.534897 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 19:50:39.617078 systemd[1]: Reloading finished in 1034 ms. Feb 13 19:50:39.674095 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:50:39.678009 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:50:39.679861 systemd[1]: kubelet.service: Deactivated successfully. Feb 13 19:50:39.680205 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:50:39.682211 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:50:39.854707 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:50:39.859987 (kubelet)[2202]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Feb 13 19:50:39.947183 kubelet[2202]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 13 19:50:39.947183 kubelet[2202]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Feb 13 19:50:39.947183 kubelet[2202]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 13 19:50:39.947749 kubelet[2202]: I0213 19:50:39.947172 2202 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 13 19:50:40.435171 kubelet[2202]: I0213 19:50:40.435100 2202 server.go:520] "Kubelet version" kubeletVersion="v1.32.0" Feb 13 19:50:40.435171 kubelet[2202]: I0213 19:50:40.435155 2202 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 13 19:50:40.435572 kubelet[2202]: I0213 19:50:40.435545 2202 server.go:954] "Client rotation is on, will bootstrap in background" Feb 13 19:50:40.534390 kubelet[2202]: I0213 19:50:40.534348 2202 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Feb 13 19:50:40.539955 kubelet[2202]: E0213 19:50:40.539912 2202 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.0.0.104:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.104:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:50:40.550908 kubelet[2202]: E0213 19:50:40.550809 2202 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Feb 13 19:50:40.550908 kubelet[2202]: I0213 19:50:40.550898 2202 server.go:1421] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Feb 13 19:50:40.556826 kubelet[2202]: I0213 19:50:40.556784 2202 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Feb 13 19:50:40.557197 kubelet[2202]: I0213 19:50:40.557133 2202 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 13 19:50:40.557387 kubelet[2202]: I0213 19:50:40.557181 2202 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 13 19:50:40.557387 kubelet[2202]: I0213 19:50:40.557387 2202 topology_manager.go:138] "Creating topology manager with none policy" Feb 13 19:50:40.557566 kubelet[2202]: I0213 19:50:40.557398 2202 container_manager_linux.go:304] "Creating device plugin manager" Feb 13 19:50:40.557599 kubelet[2202]: I0213 19:50:40.557576 2202 state_mem.go:36] "Initialized new in-memory state store" Feb 13 19:50:40.561124 kubelet[2202]: I0213 19:50:40.561089 2202 kubelet.go:446] "Attempting to sync node with API server" Feb 13 19:50:40.561124 kubelet[2202]: I0213 19:50:40.561109 2202 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 13 19:50:40.561194 kubelet[2202]: I0213 19:50:40.561130 2202 kubelet.go:352] "Adding apiserver pod source" Feb 13 19:50:40.561194 kubelet[2202]: I0213 19:50:40.561143 2202 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 13 19:50:40.564584 kubelet[2202]: W0213 19:50:40.564419 2202 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.104:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.104:6443: connect: connection refused Feb 13 19:50:40.564584 kubelet[2202]: E0213 19:50:40.564479 2202 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.0.0.104:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.104:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:50:40.564584 kubelet[2202]: W0213 19:50:40.564493 2202 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.104:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.0.0.104:6443: connect: connection refused Feb 13 19:50:40.564584 kubelet[2202]: E0213 19:50:40.564559 2202 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.0.0.104:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.104:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:50:40.565406 kubelet[2202]: I0213 19:50:40.565359 2202 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v1.7.23" apiVersion="v1" Feb 13 19:50:40.565865 kubelet[2202]: I0213 19:50:40.565839 2202 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 13 19:50:40.566784 kubelet[2202]: W0213 19:50:40.566742 2202 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Feb 13 19:50:40.570835 kubelet[2202]: I0213 19:50:40.570799 2202 watchdog_linux.go:99] "Systemd watchdog is not enabled" Feb 13 19:50:40.570907 kubelet[2202]: I0213 19:50:40.570849 2202 server.go:1287] "Started kubelet" Feb 13 19:50:40.574120 kubelet[2202]: I0213 19:50:40.573076 2202 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 13 19:50:40.574120 kubelet[2202]: I0213 19:50:40.573095 2202 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Feb 13 19:50:40.574120 kubelet[2202]: I0213 19:50:40.573111 2202 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 13 19:50:40.574120 kubelet[2202]: I0213 19:50:40.573395 2202 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 13 19:50:40.574120 kubelet[2202]: I0213 19:50:40.573930 2202 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Feb 13 19:50:40.575466 kubelet[2202]: I0213 19:50:40.574633 2202 server.go:490] "Adding debug handlers to kubelet server" Feb 13 19:50:40.576321 kubelet[2202]: E0213 19:50:40.576297 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:40.576389 kubelet[2202]: I0213 19:50:40.576341 2202 volume_manager.go:297] "Starting Kubelet Volume Manager" Feb 13 19:50:40.576699 kubelet[2202]: I0213 19:50:40.576681 2202 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Feb 13 19:50:40.576757 kubelet[2202]: I0213 19:50:40.576742 2202 reconciler.go:26] "Reconciler: start to sync state" Feb 13 19:50:40.577772 kubelet[2202]: W0213 19:50:40.577136 2202 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.104:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.104:6443: connect: connection refused Feb 13 19:50:40.577772 kubelet[2202]: E0213 19:50:40.577194 2202 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.0.0.104:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.104:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:50:40.577772 kubelet[2202]: I0213 19:50:40.577380 2202 factory.go:221] Registration of the systemd container factory successfully Feb 13 19:50:40.577772 kubelet[2202]: I0213 19:50:40.577454 2202 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Feb 13 19:50:40.577772 kubelet[2202]: E0213 19:50:40.577546 2202 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.104:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.104:6443: connect: connection refused" interval="200ms" Feb 13 19:50:40.577772 kubelet[2202]: E0213 19:50:40.577687 2202 kubelet.go:1561] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Feb 13 19:50:40.578014 kubelet[2202]: E0213 19:50:40.575905 2202 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.104:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.104:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.1823dc6cb6749933 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-02-13 19:50:40.570816819 +0000 UTC m=+0.703894047,LastTimestamp:2025-02-13 19:50:40.570816819 +0000 UTC m=+0.703894047,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Feb 13 19:50:40.581473 kubelet[2202]: I0213 19:50:40.579542 2202 factory.go:221] Registration of the containerd container factory successfully Feb 13 19:50:40.594550 kubelet[2202]: I0213 19:50:40.594503 2202 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 13 19:50:40.596943 kubelet[2202]: I0213 19:50:40.596593 2202 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 13 19:50:40.596943 kubelet[2202]: I0213 19:50:40.596631 2202 status_manager.go:227] "Starting to sync pod status with apiserver" Feb 13 19:50:40.596943 kubelet[2202]: I0213 19:50:40.596630 2202 cpu_manager.go:221] "Starting CPU manager" policy="none" Feb 13 19:50:40.596943 kubelet[2202]: I0213 19:50:40.596678 2202 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Feb 13 19:50:40.596943 kubelet[2202]: I0213 19:50:40.596701 2202 state_mem.go:36] "Initialized new in-memory state store" Feb 13 19:50:40.597492 kubelet[2202]: W0213 19:50:40.597448 2202 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.104:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.104:6443: connect: connection refused Feb 13 19:50:40.597598 kubelet[2202]: E0213 19:50:40.597574 2202 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.0.0.104:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.104:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:50:40.597769 kubelet[2202]: I0213 19:50:40.596665 2202 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Feb 13 19:50:40.597835 kubelet[2202]: I0213 19:50:40.597824 2202 kubelet.go:2388] "Starting kubelet main sync loop" Feb 13 19:50:40.598000 kubelet[2202]: E0213 19:50:40.597979 2202 kubelet.go:2412] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 13 19:50:40.677505 kubelet[2202]: E0213 19:50:40.677445 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:40.698773 kubelet[2202]: E0213 19:50:40.698625 2202 kubelet.go:2412] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Feb 13 19:50:40.778099 kubelet[2202]: E0213 19:50:40.778027 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:40.778812 kubelet[2202]: E0213 19:50:40.778593 2202 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.104:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.104:6443: connect: connection refused" interval="400ms" Feb 13 19:50:40.879210 kubelet[2202]: E0213 19:50:40.879149 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:40.899470 kubelet[2202]: E0213 19:50:40.899401 2202 kubelet.go:2412] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Feb 13 19:50:40.961454 kubelet[2202]: I0213 19:50:40.961398 2202 policy_none.go:49] "None policy: Start" Feb 13 19:50:40.961454 kubelet[2202]: I0213 19:50:40.961454 2202 memory_manager.go:186] "Starting memorymanager" policy="None" Feb 13 19:50:40.961948 kubelet[2202]: I0213 19:50:40.961485 2202 state_mem.go:35] "Initializing new in-memory state store" Feb 13 19:50:40.969706 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Feb 13 19:50:40.980255 kubelet[2202]: E0213 19:50:40.980207 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:40.981280 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Feb 13 19:50:40.997853 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Feb 13 19:50:40.999603 kubelet[2202]: I0213 19:50:40.999434 2202 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 13 19:50:41.000046 kubelet[2202]: I0213 19:50:40.999689 2202 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 13 19:50:41.000046 kubelet[2202]: I0213 19:50:40.999705 2202 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 13 19:50:41.000046 kubelet[2202]: I0213 19:50:40.999986 2202 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 13 19:50:41.001348 kubelet[2202]: E0213 19:50:41.001151 2202 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Feb 13 19:50:41.001348 kubelet[2202]: E0213 19:50:41.001216 2202 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Feb 13 19:50:41.101588 kubelet[2202]: I0213 19:50:41.101534 2202 kubelet_node_status.go:76] "Attempting to register node" node="localhost" Feb 13 19:50:41.102118 kubelet[2202]: E0213 19:50:41.102062 2202 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://10.0.0.104:6443/api/v1/nodes\": dial tcp 10.0.0.104:6443: connect: connection refused" node="localhost" Feb 13 19:50:41.180088 kubelet[2202]: E0213 19:50:41.180021 2202 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.104:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.104:6443: connect: connection refused" interval="800ms" Feb 13 19:50:41.303583 kubelet[2202]: I0213 19:50:41.303004 2202 kubelet_node_status.go:76] "Attempting to register node" node="localhost" Feb 13 19:50:41.303692 kubelet[2202]: E0213 19:50:41.303637 2202 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://10.0.0.104:6443/api/v1/nodes\": dial tcp 10.0.0.104:6443: connect: connection refused" node="localhost" Feb 13 19:50:41.308832 systemd[1]: Created slice kubepods-burstable-podaa404ccfad8dad97a6ff6b78cbbb56a3.slice - libcontainer container kubepods-burstable-podaa404ccfad8dad97a6ff6b78cbbb56a3.slice. Feb 13 19:50:41.325008 kubelet[2202]: E0213 19:50:41.324964 2202 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Feb 13 19:50:41.328859 systemd[1]: Created slice kubepods-burstable-podc72911152bbceda2f57fd8d59261e015.slice - libcontainer container kubepods-burstable-podc72911152bbceda2f57fd8d59261e015.slice. Feb 13 19:50:41.336261 kubelet[2202]: E0213 19:50:41.336213 2202 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Feb 13 19:50:41.339297 systemd[1]: Created slice kubepods-burstable-pod95ef9ac46cd4dbaadc63cb713310ae59.slice - libcontainer container kubepods-burstable-pod95ef9ac46cd4dbaadc63cb713310ae59.slice. Feb 13 19:50:41.341355 kubelet[2202]: E0213 19:50:41.341325 2202 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Feb 13 19:50:41.381817 kubelet[2202]: I0213 19:50:41.381752 2202 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/95ef9ac46cd4dbaadc63cb713310ae59-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"95ef9ac46cd4dbaadc63cb713310ae59\") " pod="kube-system/kube-scheduler-localhost" Feb 13 19:50:41.381817 kubelet[2202]: I0213 19:50:41.381803 2202 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/aa404ccfad8dad97a6ff6b78cbbb56a3-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"aa404ccfad8dad97a6ff6b78cbbb56a3\") " pod="kube-system/kube-apiserver-localhost" Feb 13 19:50:41.382039 kubelet[2202]: I0213 19:50:41.381852 2202 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/aa404ccfad8dad97a6ff6b78cbbb56a3-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"aa404ccfad8dad97a6ff6b78cbbb56a3\") " pod="kube-system/kube-apiserver-localhost" Feb 13 19:50:41.382039 kubelet[2202]: I0213 19:50:41.381921 2202 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/c72911152bbceda2f57fd8d59261e015-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"c72911152bbceda2f57fd8d59261e015\") " pod="kube-system/kube-controller-manager-localhost" Feb 13 19:50:41.382039 kubelet[2202]: I0213 19:50:41.381947 2202 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/c72911152bbceda2f57fd8d59261e015-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"c72911152bbceda2f57fd8d59261e015\") " pod="kube-system/kube-controller-manager-localhost" Feb 13 19:50:41.382147 kubelet[2202]: I0213 19:50:41.382063 2202 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/aa404ccfad8dad97a6ff6b78cbbb56a3-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"aa404ccfad8dad97a6ff6b78cbbb56a3\") " pod="kube-system/kube-apiserver-localhost" Feb 13 19:50:41.382182 kubelet[2202]: I0213 19:50:41.382148 2202 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/c72911152bbceda2f57fd8d59261e015-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"c72911152bbceda2f57fd8d59261e015\") " pod="kube-system/kube-controller-manager-localhost" Feb 13 19:50:41.382216 kubelet[2202]: I0213 19:50:41.382181 2202 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/c72911152bbceda2f57fd8d59261e015-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"c72911152bbceda2f57fd8d59261e015\") " pod="kube-system/kube-controller-manager-localhost" Feb 13 19:50:41.382216 kubelet[2202]: I0213 19:50:41.382208 2202 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/c72911152bbceda2f57fd8d59261e015-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"c72911152bbceda2f57fd8d59261e015\") " pod="kube-system/kube-controller-manager-localhost" Feb 13 19:50:41.530860 kubelet[2202]: W0213 19:50:41.530769 2202 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.104:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.0.0.104:6443: connect: connection refused Feb 13 19:50:41.530998 kubelet[2202]: E0213 19:50:41.530861 2202 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.0.0.104:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.104:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:50:41.626719 kubelet[2202]: E0213 19:50:41.626576 2202 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:50:41.627784 containerd[1468]: time="2025-02-13T19:50:41.627739857Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:aa404ccfad8dad97a6ff6b78cbbb56a3,Namespace:kube-system,Attempt:0,}" Feb 13 19:50:41.638149 kubelet[2202]: E0213 19:50:41.638076 2202 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:50:41.638642 containerd[1468]: time="2025-02-13T19:50:41.638588572Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:c72911152bbceda2f57fd8d59261e015,Namespace:kube-system,Attempt:0,}" Feb 13 19:50:41.640229 kubelet[2202]: W0213 19:50:41.640163 2202 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.104:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.104:6443: connect: connection refused Feb 13 19:50:41.640289 kubelet[2202]: E0213 19:50:41.640240 2202 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.0.0.104:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.104:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:50:41.642705 kubelet[2202]: E0213 19:50:41.642676 2202 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:50:41.643164 containerd[1468]: time="2025-02-13T19:50:41.643133782Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:95ef9ac46cd4dbaadc63cb713310ae59,Namespace:kube-system,Attempt:0,}" Feb 13 19:50:41.666926 kubelet[2202]: W0213 19:50:41.666857 2202 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.104:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.104:6443: connect: connection refused Feb 13 19:50:41.666997 kubelet[2202]: E0213 19:50:41.666930 2202 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.0.0.104:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.104:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:50:41.694217 kubelet[2202]: E0213 19:50:41.694085 2202 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.104:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.104:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.1823dc6cb6749933 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-02-13 19:50:40.570816819 +0000 UTC m=+0.703894047,LastTimestamp:2025-02-13 19:50:40.570816819 +0000 UTC m=+0.703894047,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Feb 13 19:50:41.705321 kubelet[2202]: I0213 19:50:41.705290 2202 kubelet_node_status.go:76] "Attempting to register node" node="localhost" Feb 13 19:50:41.705727 kubelet[2202]: E0213 19:50:41.705684 2202 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://10.0.0.104:6443/api/v1/nodes\": dial tcp 10.0.0.104:6443: connect: connection refused" node="localhost" Feb 13 19:50:41.866094 kubelet[2202]: W0213 19:50:41.866046 2202 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.104:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.104:6443: connect: connection refused Feb 13 19:50:41.866094 kubelet[2202]: E0213 19:50:41.866102 2202 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.0.0.104:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.104:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:50:41.980939 kubelet[2202]: E0213 19:50:41.980892 2202 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.104:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.104:6443: connect: connection refused" interval="1.6s" Feb 13 19:50:42.507148 kubelet[2202]: I0213 19:50:42.507098 2202 kubelet_node_status.go:76] "Attempting to register node" node="localhost" Feb 13 19:50:42.507574 kubelet[2202]: E0213 19:50:42.507527 2202 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://10.0.0.104:6443/api/v1/nodes\": dial tcp 10.0.0.104:6443: connect: connection refused" node="localhost" Feb 13 19:50:42.559010 kubelet[2202]: E0213 19:50:42.558948 2202 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.0.0.104:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.104:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:50:43.349742 kubelet[2202]: W0213 19:50:43.349689 2202 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.104:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.104:6443: connect: connection refused Feb 13 19:50:43.349742 kubelet[2202]: E0213 19:50:43.349744 2202 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.0.0.104:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.104:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:50:43.582399 kubelet[2202]: E0213 19:50:43.582346 2202 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.104:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.104:6443: connect: connection refused" interval="3.2s" Feb 13 19:50:43.669072 kubelet[2202]: W0213 19:50:43.668916 2202 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.104:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.104:6443: connect: connection refused Feb 13 19:50:43.669072 kubelet[2202]: E0213 19:50:43.668986 2202 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.0.0.104:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.104:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:50:44.109054 kubelet[2202]: I0213 19:50:44.109007 2202 kubelet_node_status.go:76] "Attempting to register node" node="localhost" Feb 13 19:50:44.109400 kubelet[2202]: E0213 19:50:44.109362 2202 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://10.0.0.104:6443/api/v1/nodes\": dial tcp 10.0.0.104:6443: connect: connection refused" node="localhost" Feb 13 19:50:44.202694 kubelet[2202]: W0213 19:50:44.202645 2202 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.104:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.0.0.104:6443: connect: connection refused Feb 13 19:50:44.202694 kubelet[2202]: E0213 19:50:44.202695 2202 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.0.0.104:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.104:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:50:44.538008 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount531931558.mount: Deactivated successfully. Feb 13 19:50:44.582782 kubelet[2202]: W0213 19:50:44.582702 2202 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.104:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.104:6443: connect: connection refused Feb 13 19:50:44.582782 kubelet[2202]: E0213 19:50:44.582773 2202 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.0.0.104:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.104:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:50:45.750201 containerd[1468]: time="2025-02-13T19:50:45.750116068Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 19:50:45.859600 containerd[1468]: time="2025-02-13T19:50:45.859491596Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Feb 13 19:50:45.981303 containerd[1468]: time="2025-02-13T19:50:45.981228295Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 19:50:46.020431 containerd[1468]: time="2025-02-13T19:50:46.020174648Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 19:50:46.101485 containerd[1468]: time="2025-02-13T19:50:46.101366941Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 19:50:46.153376 containerd[1468]: time="2025-02-13T19:50:46.153297193Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312056" Feb 13 19:50:46.206303 containerd[1468]: time="2025-02-13T19:50:46.206241412Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 19:50:46.207360 containerd[1468]: time="2025-02-13T19:50:46.207311242Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 4.56861336s" Feb 13 19:50:46.223463 containerd[1468]: time="2025-02-13T19:50:46.223378799Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Feb 13 19:50:46.293432 containerd[1468]: time="2025-02-13T19:50:46.292905464Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 4.649700976s" Feb 13 19:50:46.295607 containerd[1468]: time="2025-02-13T19:50:46.294173263Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 4.666305801s" Feb 13 19:50:46.622745 kubelet[2202]: E0213 19:50:46.622564 2202 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.0.0.104:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.104:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:50:46.680065 containerd[1468]: time="2025-02-13T19:50:46.677133337Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:50:46.680065 containerd[1468]: time="2025-02-13T19:50:46.679608820Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:50:46.680065 containerd[1468]: time="2025-02-13T19:50:46.679622576Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:50:46.680065 containerd[1468]: time="2025-02-13T19:50:46.679731845Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:50:46.686981 containerd[1468]: time="2025-02-13T19:50:46.686081730Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:50:46.686981 containerd[1468]: time="2025-02-13T19:50:46.686205586Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:50:46.686981 containerd[1468]: time="2025-02-13T19:50:46.686220815Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:50:46.686981 containerd[1468]: time="2025-02-13T19:50:46.686400859Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:50:46.781299 systemd[1]: Started cri-containerd-d48606ca4694dc3e824cb1a3d6960ff3595b96b0dc06931ad9676291d322a1cb.scope - libcontainer container d48606ca4694dc3e824cb1a3d6960ff3595b96b0dc06931ad9676291d322a1cb. Feb 13 19:50:46.783255 kubelet[2202]: E0213 19:50:46.783069 2202 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.104:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.104:6443: connect: connection refused" interval="6.4s" Feb 13 19:50:46.789006 systemd[1]: Started cri-containerd-1a62faa312e5553c79b265b7c37914d81a00fbdaa22dc364deae468ed6a8700d.scope - libcontainer container 1a62faa312e5553c79b265b7c37914d81a00fbdaa22dc364deae468ed6a8700d. Feb 13 19:50:46.791564 kubelet[2202]: W0213 19:50:46.789689 2202 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.104:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.104:6443: connect: connection refused Feb 13 19:50:46.791564 kubelet[2202]: E0213 19:50:46.789751 2202 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.0.0.104:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.104:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:50:46.807377 containerd[1468]: time="2025-02-13T19:50:46.807211203Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:50:46.807377 containerd[1468]: time="2025-02-13T19:50:46.807305754Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:50:46.807377 containerd[1468]: time="2025-02-13T19:50:46.807317236Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:50:46.807955 containerd[1468]: time="2025-02-13T19:50:46.807420753Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:50:46.835094 systemd[1]: Started cri-containerd-7736dec51466e58eda2c9f897ef8660248bbbead82bb23c86d017e7397bc3467.scope - libcontainer container 7736dec51466e58eda2c9f897ef8660248bbbead82bb23c86d017e7397bc3467. Feb 13 19:50:46.857657 containerd[1468]: time="2025-02-13T19:50:46.857484303Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:c72911152bbceda2f57fd8d59261e015,Namespace:kube-system,Attempt:0,} returns sandbox id \"d48606ca4694dc3e824cb1a3d6960ff3595b96b0dc06931ad9676291d322a1cb\"" Feb 13 19:50:46.859168 kubelet[2202]: E0213 19:50:46.859132 2202 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:50:46.862259 containerd[1468]: time="2025-02-13T19:50:46.862094560Z" level=info msg="CreateContainer within sandbox \"d48606ca4694dc3e824cb1a3d6960ff3595b96b0dc06931ad9676291d322a1cb\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Feb 13 19:50:46.864454 containerd[1468]: time="2025-02-13T19:50:46.864418234Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:aa404ccfad8dad97a6ff6b78cbbb56a3,Namespace:kube-system,Attempt:0,} returns sandbox id \"1a62faa312e5553c79b265b7c37914d81a00fbdaa22dc364deae468ed6a8700d\"" Feb 13 19:50:46.865759 kubelet[2202]: E0213 19:50:46.865707 2202 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:50:46.869860 containerd[1468]: time="2025-02-13T19:50:46.869807467Z" level=info msg="CreateContainer within sandbox \"1a62faa312e5553c79b265b7c37914d81a00fbdaa22dc364deae468ed6a8700d\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Feb 13 19:50:46.891003 containerd[1468]: time="2025-02-13T19:50:46.890881144Z" level=info msg="CreateContainer within sandbox \"d48606ca4694dc3e824cb1a3d6960ff3595b96b0dc06931ad9676291d322a1cb\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"c08acdd0c13a33a2da2f76308d045331493a06158ba5c7651151e66f8cba7b65\"" Feb 13 19:50:46.892025 containerd[1468]: time="2025-02-13T19:50:46.891939814Z" level=info msg="StartContainer for \"c08acdd0c13a33a2da2f76308d045331493a06158ba5c7651151e66f8cba7b65\"" Feb 13 19:50:46.898260 containerd[1468]: time="2025-02-13T19:50:46.898148721Z" level=info msg="CreateContainer within sandbox \"1a62faa312e5553c79b265b7c37914d81a00fbdaa22dc364deae468ed6a8700d\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"3db7a7a0f9834a40265d975d46d6da10121ce02cbd9ec40557df5622d727d367\"" Feb 13 19:50:46.898802 containerd[1468]: time="2025-02-13T19:50:46.898771449Z" level=info msg="StartContainer for \"3db7a7a0f9834a40265d975d46d6da10121ce02cbd9ec40557df5622d727d367\"" Feb 13 19:50:46.920004 containerd[1468]: time="2025-02-13T19:50:46.919961910Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:95ef9ac46cd4dbaadc63cb713310ae59,Namespace:kube-system,Attempt:0,} returns sandbox id \"7736dec51466e58eda2c9f897ef8660248bbbead82bb23c86d017e7397bc3467\"" Feb 13 19:50:46.920978 kubelet[2202]: E0213 19:50:46.920923 2202 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:50:46.923647 containerd[1468]: time="2025-02-13T19:50:46.923610201Z" level=info msg="CreateContainer within sandbox \"7736dec51466e58eda2c9f897ef8660248bbbead82bb23c86d017e7397bc3467\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Feb 13 19:50:46.940504 containerd[1468]: time="2025-02-13T19:50:46.940449280Z" level=info msg="CreateContainer within sandbox \"7736dec51466e58eda2c9f897ef8660248bbbead82bb23c86d017e7397bc3467\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"dcf28b923a77615628bbb7e9f68fbd51d5482fa71ed07073ba0911cb44170856\"" Feb 13 19:50:46.940894 containerd[1468]: time="2025-02-13T19:50:46.940852590Z" level=info msg="StartContainer for \"dcf28b923a77615628bbb7e9f68fbd51d5482fa71ed07073ba0911cb44170856\"" Feb 13 19:50:46.944080 systemd[1]: Started cri-containerd-3db7a7a0f9834a40265d975d46d6da10121ce02cbd9ec40557df5622d727d367.scope - libcontainer container 3db7a7a0f9834a40265d975d46d6da10121ce02cbd9ec40557df5622d727d367. Feb 13 19:50:46.949828 systemd[1]: Started cri-containerd-c08acdd0c13a33a2da2f76308d045331493a06158ba5c7651151e66f8cba7b65.scope - libcontainer container c08acdd0c13a33a2da2f76308d045331493a06158ba5c7651151e66f8cba7b65. Feb 13 19:50:46.989210 systemd[1]: Started cri-containerd-dcf28b923a77615628bbb7e9f68fbd51d5482fa71ed07073ba0911cb44170856.scope - libcontainer container dcf28b923a77615628bbb7e9f68fbd51d5482fa71ed07073ba0911cb44170856. Feb 13 19:50:47.008809 containerd[1468]: time="2025-02-13T19:50:47.007637951Z" level=info msg="StartContainer for \"3db7a7a0f9834a40265d975d46d6da10121ce02cbd9ec40557df5622d727d367\" returns successfully" Feb 13 19:50:47.017547 containerd[1468]: time="2025-02-13T19:50:47.017249891Z" level=info msg="StartContainer for \"c08acdd0c13a33a2da2f76308d045331493a06158ba5c7651151e66f8cba7b65\" returns successfully" Feb 13 19:50:47.047445 containerd[1468]: time="2025-02-13T19:50:47.047174034Z" level=info msg="StartContainer for \"dcf28b923a77615628bbb7e9f68fbd51d5482fa71ed07073ba0911cb44170856\" returns successfully" Feb 13 19:50:47.313256 kubelet[2202]: I0213 19:50:47.312620 2202 kubelet_node_status.go:76] "Attempting to register node" node="localhost" Feb 13 19:50:47.627142 kubelet[2202]: E0213 19:50:47.627017 2202 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Feb 13 19:50:47.628589 kubelet[2202]: E0213 19:50:47.628248 2202 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:50:47.632569 kubelet[2202]: E0213 19:50:47.632283 2202 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Feb 13 19:50:47.632569 kubelet[2202]: E0213 19:50:47.632395 2202 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:50:47.635431 kubelet[2202]: E0213 19:50:47.635231 2202 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Feb 13 19:50:47.635431 kubelet[2202]: E0213 19:50:47.635351 2202 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:50:48.683448 kubelet[2202]: E0213 19:50:48.682661 2202 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Feb 13 19:50:48.683448 kubelet[2202]: E0213 19:50:48.682790 2202 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:50:48.683448 kubelet[2202]: E0213 19:50:48.682987 2202 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Feb 13 19:50:48.683448 kubelet[2202]: E0213 19:50:48.683070 2202 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:50:48.683448 kubelet[2202]: E0213 19:50:48.683254 2202 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Feb 13 19:50:48.683448 kubelet[2202]: E0213 19:50:48.683321 2202 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:50:48.921348 kubelet[2202]: I0213 19:50:48.921241 2202 kubelet_node_status.go:79] "Successfully registered node" node="localhost" Feb 13 19:50:48.921348 kubelet[2202]: E0213 19:50:48.921285 2202 kubelet_node_status.go:549] "Error updating node status, will retry" err="error getting node \"localhost\": node \"localhost\" not found" Feb 13 19:50:48.933660 kubelet[2202]: E0213 19:50:48.933549 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:49.034510 kubelet[2202]: E0213 19:50:49.034450 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:49.134714 kubelet[2202]: E0213 19:50:49.134621 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:49.235713 kubelet[2202]: E0213 19:50:49.235633 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:49.336794 kubelet[2202]: E0213 19:50:49.336711 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:49.437705 kubelet[2202]: E0213 19:50:49.437611 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:49.537888 kubelet[2202]: E0213 19:50:49.537730 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:49.638035 kubelet[2202]: E0213 19:50:49.637969 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:49.738967 kubelet[2202]: E0213 19:50:49.738892 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:49.839604 kubelet[2202]: E0213 19:50:49.839470 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:49.940258 kubelet[2202]: E0213 19:50:49.940198 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:50.040509 kubelet[2202]: E0213 19:50:50.040458 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:50.141183 kubelet[2202]: E0213 19:50:50.141033 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:50.241707 kubelet[2202]: E0213 19:50:50.241638 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:50.342423 kubelet[2202]: E0213 19:50:50.342347 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:50.443138 kubelet[2202]: E0213 19:50:50.442998 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:50.544089 kubelet[2202]: E0213 19:50:50.544022 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:50.644927 kubelet[2202]: E0213 19:50:50.644848 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:50.746089 kubelet[2202]: E0213 19:50:50.746029 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:50.846716 kubelet[2202]: E0213 19:50:50.846652 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:50.923066 update_engine[1459]: I20250213 19:50:50.922967 1459 update_attempter.cc:509] Updating boot flags... Feb 13 19:50:50.949245 kubelet[2202]: E0213 19:50:50.948923 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:50.987526 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 38 scanned by (udev-worker) (2482) Feb 13 19:50:51.002343 kubelet[2202]: E0213 19:50:51.002021 2202 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Feb 13 19:50:51.042922 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 38 scanned by (udev-worker) (2485) Feb 13 19:50:51.049390 kubelet[2202]: E0213 19:50:51.049354 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:51.156056 kubelet[2202]: E0213 19:50:51.155974 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:51.256289 kubelet[2202]: E0213 19:50:51.256130 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:51.356817 kubelet[2202]: E0213 19:50:51.356747 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:51.457545 kubelet[2202]: E0213 19:50:51.457476 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:51.558556 kubelet[2202]: E0213 19:50:51.558384 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:51.658768 kubelet[2202]: E0213 19:50:51.658715 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:51.759336 kubelet[2202]: E0213 19:50:51.759158 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:51.859464 kubelet[2202]: E0213 19:50:51.859296 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:51.959890 kubelet[2202]: E0213 19:50:51.959782 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:52.060546 kubelet[2202]: E0213 19:50:52.060488 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:52.161515 kubelet[2202]: E0213 19:50:52.161312 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:52.261927 kubelet[2202]: E0213 19:50:52.261809 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:52.362890 kubelet[2202]: E0213 19:50:52.362826 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:52.463406 kubelet[2202]: E0213 19:50:52.463355 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:52.564310 kubelet[2202]: E0213 19:50:52.564251 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:52.665072 kubelet[2202]: E0213 19:50:52.665006 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:52.765984 kubelet[2202]: E0213 19:50:52.765816 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:52.836848 systemd[1]: Reloading requested from client PID 2491 ('systemctl') (unit session-7.scope)... Feb 13 19:50:52.836866 systemd[1]: Reloading... Feb 13 19:50:52.866607 kubelet[2202]: E0213 19:50:52.866559 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:52.934058 zram_generator::config[2533]: No configuration found. Feb 13 19:50:52.967575 kubelet[2202]: E0213 19:50:52.967513 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:53.068137 kubelet[2202]: E0213 19:50:53.067991 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:53.093699 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 19:50:53.168744 kubelet[2202]: E0213 19:50:53.168684 2202 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:53.198805 systemd[1]: Reloading finished in 361 ms. Feb 13 19:50:53.244839 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:50:53.261406 systemd[1]: kubelet.service: Deactivated successfully. Feb 13 19:50:53.261710 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:50:53.261760 systemd[1]: kubelet.service: Consumed 1.308s CPU time, 125.4M memory peak, 0B memory swap peak. Feb 13 19:50:53.273110 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:50:53.444295 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:50:53.449059 (kubelet)[2575]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Feb 13 19:50:53.495729 kubelet[2575]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 13 19:50:53.495729 kubelet[2575]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Feb 13 19:50:53.495729 kubelet[2575]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 13 19:50:53.496223 kubelet[2575]: I0213 19:50:53.495809 2575 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 13 19:50:53.502814 kubelet[2575]: I0213 19:50:53.502741 2575 server.go:520] "Kubelet version" kubeletVersion="v1.32.0" Feb 13 19:50:53.502814 kubelet[2575]: I0213 19:50:53.502793 2575 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 13 19:50:53.503187 kubelet[2575]: I0213 19:50:53.503158 2575 server.go:954] "Client rotation is on, will bootstrap in background" Feb 13 19:50:53.504563 kubelet[2575]: I0213 19:50:53.504518 2575 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Feb 13 19:50:53.506774 kubelet[2575]: I0213 19:50:53.506734 2575 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Feb 13 19:50:53.595748 kubelet[2575]: E0213 19:50:53.595690 2575 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Feb 13 19:50:53.595748 kubelet[2575]: I0213 19:50:53.595732 2575 server.go:1421] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Feb 13 19:50:53.601894 kubelet[2575]: I0213 19:50:53.601842 2575 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Feb 13 19:50:53.602189 kubelet[2575]: I0213 19:50:53.602150 2575 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 13 19:50:53.602364 kubelet[2575]: I0213 19:50:53.602182 2575 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 13 19:50:53.602485 kubelet[2575]: I0213 19:50:53.602365 2575 topology_manager.go:138] "Creating topology manager with none policy" Feb 13 19:50:53.602485 kubelet[2575]: I0213 19:50:53.602373 2575 container_manager_linux.go:304] "Creating device plugin manager" Feb 13 19:50:53.602485 kubelet[2575]: I0213 19:50:53.602412 2575 state_mem.go:36] "Initialized new in-memory state store" Feb 13 19:50:53.602591 kubelet[2575]: I0213 19:50:53.602568 2575 kubelet.go:446] "Attempting to sync node with API server" Feb 13 19:50:53.602591 kubelet[2575]: I0213 19:50:53.602589 2575 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 13 19:50:53.602639 kubelet[2575]: I0213 19:50:53.602610 2575 kubelet.go:352] "Adding apiserver pod source" Feb 13 19:50:53.602639 kubelet[2575]: I0213 19:50:53.602622 2575 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 13 19:50:53.604080 kubelet[2575]: I0213 19:50:53.603403 2575 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v1.7.23" apiVersion="v1" Feb 13 19:50:53.604080 kubelet[2575]: I0213 19:50:53.603863 2575 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 13 19:50:53.604522 kubelet[2575]: I0213 19:50:53.604397 2575 watchdog_linux.go:99] "Systemd watchdog is not enabled" Feb 13 19:50:53.604522 kubelet[2575]: I0213 19:50:53.604440 2575 server.go:1287] "Started kubelet" Feb 13 19:50:53.606429 kubelet[2575]: I0213 19:50:53.606406 2575 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 13 19:50:53.608627 kubelet[2575]: I0213 19:50:53.608587 2575 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Feb 13 19:50:53.609687 kubelet[2575]: I0213 19:50:53.609449 2575 server.go:490] "Adding debug handlers to kubelet server" Feb 13 19:50:53.613383 kubelet[2575]: I0213 19:50:53.610200 2575 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 13 19:50:53.613383 kubelet[2575]: I0213 19:50:53.610407 2575 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 13 19:50:53.613383 kubelet[2575]: I0213 19:50:53.610572 2575 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Feb 13 19:50:53.613383 kubelet[2575]: I0213 19:50:53.610779 2575 volume_manager.go:297] "Starting Kubelet Volume Manager" Feb 13 19:50:53.613383 kubelet[2575]: I0213 19:50:53.610845 2575 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Feb 13 19:50:53.613383 kubelet[2575]: I0213 19:50:53.611014 2575 reconciler.go:26] "Reconciler: start to sync state" Feb 13 19:50:53.613383 kubelet[2575]: E0213 19:50:53.611345 2575 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:50:53.614603 kubelet[2575]: I0213 19:50:53.614577 2575 factory.go:221] Registration of the systemd container factory successfully Feb 13 19:50:53.614727 kubelet[2575]: I0213 19:50:53.614695 2575 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Feb 13 19:50:53.618782 kubelet[2575]: E0213 19:50:53.618382 2575 kubelet.go:1561] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Feb 13 19:50:53.619705 kubelet[2575]: I0213 19:50:53.619683 2575 factory.go:221] Registration of the containerd container factory successfully Feb 13 19:50:53.625367 kubelet[2575]: I0213 19:50:53.625324 2575 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 13 19:50:53.627586 kubelet[2575]: I0213 19:50:53.627553 2575 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 13 19:50:53.627718 kubelet[2575]: I0213 19:50:53.627706 2575 status_manager.go:227] "Starting to sync pod status with apiserver" Feb 13 19:50:53.627783 kubelet[2575]: I0213 19:50:53.627773 2575 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Feb 13 19:50:53.627911 kubelet[2575]: I0213 19:50:53.627900 2575 kubelet.go:2388] "Starting kubelet main sync loop" Feb 13 19:50:53.628057 kubelet[2575]: E0213 19:50:53.628013 2575 kubelet.go:2412] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 13 19:50:53.651165 kubelet[2575]: I0213 19:50:53.651116 2575 cpu_manager.go:221] "Starting CPU manager" policy="none" Feb 13 19:50:53.651165 kubelet[2575]: I0213 19:50:53.651150 2575 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Feb 13 19:50:53.651165 kubelet[2575]: I0213 19:50:53.651174 2575 state_mem.go:36] "Initialized new in-memory state store" Feb 13 19:50:53.651360 kubelet[2575]: I0213 19:50:53.651340 2575 state_mem.go:88] "Updated default CPUSet" cpuSet="" Feb 13 19:50:53.651381 kubelet[2575]: I0213 19:50:53.651354 2575 state_mem.go:96] "Updated CPUSet assignments" assignments={} Feb 13 19:50:53.651381 kubelet[2575]: I0213 19:50:53.651374 2575 policy_none.go:49] "None policy: Start" Feb 13 19:50:53.651429 kubelet[2575]: I0213 19:50:53.651383 2575 memory_manager.go:186] "Starting memorymanager" policy="None" Feb 13 19:50:53.651429 kubelet[2575]: I0213 19:50:53.651393 2575 state_mem.go:35] "Initializing new in-memory state store" Feb 13 19:50:53.651495 kubelet[2575]: I0213 19:50:53.651484 2575 state_mem.go:75] "Updated machine memory state" Feb 13 19:50:53.655405 kubelet[2575]: I0213 19:50:53.655389 2575 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 13 19:50:53.655565 kubelet[2575]: I0213 19:50:53.655547 2575 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 13 19:50:53.655596 kubelet[2575]: I0213 19:50:53.655564 2575 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 13 19:50:53.655774 kubelet[2575]: I0213 19:50:53.655760 2575 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 13 19:50:53.656799 kubelet[2575]: E0213 19:50:53.656772 2575 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Feb 13 19:50:53.729488 kubelet[2575]: I0213 19:50:53.729452 2575 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Feb 13 19:50:53.729648 kubelet[2575]: I0213 19:50:53.729452 2575 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Feb 13 19:50:53.729677 kubelet[2575]: I0213 19:50:53.729641 2575 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Feb 13 19:50:53.760626 kubelet[2575]: I0213 19:50:53.760586 2575 kubelet_node_status.go:76] "Attempting to register node" node="localhost" Feb 13 19:50:53.812592 kubelet[2575]: I0213 19:50:53.812540 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/aa404ccfad8dad97a6ff6b78cbbb56a3-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"aa404ccfad8dad97a6ff6b78cbbb56a3\") " pod="kube-system/kube-apiserver-localhost" Feb 13 19:50:53.812592 kubelet[2575]: I0213 19:50:53.812580 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/c72911152bbceda2f57fd8d59261e015-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"c72911152bbceda2f57fd8d59261e015\") " pod="kube-system/kube-controller-manager-localhost" Feb 13 19:50:53.812796 kubelet[2575]: I0213 19:50:53.812605 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/c72911152bbceda2f57fd8d59261e015-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"c72911152bbceda2f57fd8d59261e015\") " pod="kube-system/kube-controller-manager-localhost" Feb 13 19:50:53.812796 kubelet[2575]: I0213 19:50:53.812628 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/95ef9ac46cd4dbaadc63cb713310ae59-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"95ef9ac46cd4dbaadc63cb713310ae59\") " pod="kube-system/kube-scheduler-localhost" Feb 13 19:50:53.812796 kubelet[2575]: I0213 19:50:53.812650 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/aa404ccfad8dad97a6ff6b78cbbb56a3-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"aa404ccfad8dad97a6ff6b78cbbb56a3\") " pod="kube-system/kube-apiserver-localhost" Feb 13 19:50:53.812796 kubelet[2575]: I0213 19:50:53.812679 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/aa404ccfad8dad97a6ff6b78cbbb56a3-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"aa404ccfad8dad97a6ff6b78cbbb56a3\") " pod="kube-system/kube-apiserver-localhost" Feb 13 19:50:53.812796 kubelet[2575]: I0213 19:50:53.812699 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/c72911152bbceda2f57fd8d59261e015-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"c72911152bbceda2f57fd8d59261e015\") " pod="kube-system/kube-controller-manager-localhost" Feb 13 19:50:53.812936 kubelet[2575]: I0213 19:50:53.812721 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/c72911152bbceda2f57fd8d59261e015-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"c72911152bbceda2f57fd8d59261e015\") " pod="kube-system/kube-controller-manager-localhost" Feb 13 19:50:53.812936 kubelet[2575]: I0213 19:50:53.812742 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/c72911152bbceda2f57fd8d59261e015-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"c72911152bbceda2f57fd8d59261e015\") " pod="kube-system/kube-controller-manager-localhost" Feb 13 19:50:54.051574 kubelet[2575]: E0213 19:50:54.049034 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:50:54.051574 kubelet[2575]: E0213 19:50:54.049209 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:50:54.051574 kubelet[2575]: I0213 19:50:54.049257 2575 kubelet_node_status.go:125] "Node was previously registered" node="localhost" Feb 13 19:50:54.051574 kubelet[2575]: I0213 19:50:54.049357 2575 kubelet_node_status.go:79] "Successfully registered node" node="localhost" Feb 13 19:50:54.055114 kubelet[2575]: E0213 19:50:54.055083 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:50:54.603516 kubelet[2575]: I0213 19:50:54.603444 2575 apiserver.go:52] "Watching apiserver" Feb 13 19:50:54.611013 kubelet[2575]: I0213 19:50:54.610974 2575 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" Feb 13 19:50:54.638433 kubelet[2575]: I0213 19:50:54.637238 2575 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Feb 13 19:50:54.638433 kubelet[2575]: E0213 19:50:54.637333 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:50:54.638433 kubelet[2575]: I0213 19:50:54.637513 2575 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Feb 13 19:50:54.680172 kubelet[2575]: E0213 19:50:54.680017 2575 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost" Feb 13 19:50:54.680897 kubelet[2575]: E0213 19:50:54.680418 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:50:54.680897 kubelet[2575]: E0213 19:50:54.680648 2575 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Feb 13 19:50:54.680897 kubelet[2575]: E0213 19:50:54.680792 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:50:54.705746 kubelet[2575]: I0213 19:50:54.705673 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.705647301 podStartE2EDuration="1.705647301s" podCreationTimestamp="2025-02-13 19:50:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 19:50:54.705390645 +0000 UTC m=+1.251707081" watchObservedRunningTime="2025-02-13 19:50:54.705647301 +0000 UTC m=+1.251963737" Feb 13 19:50:54.706018 kubelet[2575]: I0213 19:50:54.705821 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=1.705813005 podStartE2EDuration="1.705813005s" podCreationTimestamp="2025-02-13 19:50:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 19:50:54.680189467 +0000 UTC m=+1.226505903" watchObservedRunningTime="2025-02-13 19:50:54.705813005 +0000 UTC m=+1.252129431" Feb 13 19:50:54.787411 kubelet[2575]: I0213 19:50:54.787273 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.787247137 podStartE2EDuration="1.787247137s" podCreationTimestamp="2025-02-13 19:50:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 19:50:54.745335859 +0000 UTC m=+1.291652285" watchObservedRunningTime="2025-02-13 19:50:54.787247137 +0000 UTC m=+1.333563563" Feb 13 19:50:55.644624 kubelet[2575]: E0213 19:50:55.644567 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:50:55.646352 kubelet[2575]: E0213 19:50:55.645365 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:50:55.646352 kubelet[2575]: E0213 19:50:55.645761 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:50:58.102277 kubelet[2575]: E0213 19:50:58.102243 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:50:58.650732 kubelet[2575]: E0213 19:50:58.650695 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:50:58.660409 kubelet[2575]: I0213 19:50:58.660372 2575 kuberuntime_manager.go:1702] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Feb 13 19:50:58.660835 containerd[1468]: time="2025-02-13T19:50:58.660763105Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Feb 13 19:50:58.661353 kubelet[2575]: I0213 19:50:58.660941 2575 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Feb 13 19:50:59.344947 systemd[1]: Created slice kubepods-besteffort-pod9402f824_5625_407a_8e63_82e168dc95ef.slice - libcontainer container kubepods-besteffort-pod9402f824_5625_407a_8e63_82e168dc95ef.slice. Feb 13 19:50:59.382986 kubelet[2575]: I0213 19:50:59.382944 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/9402f824-5625-407a-8e63-82e168dc95ef-kube-proxy\") pod \"kube-proxy-wfsjx\" (UID: \"9402f824-5625-407a-8e63-82e168dc95ef\") " pod="kube-system/kube-proxy-wfsjx" Feb 13 19:50:59.382986 kubelet[2575]: I0213 19:50:59.382982 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/9402f824-5625-407a-8e63-82e168dc95ef-xtables-lock\") pod \"kube-proxy-wfsjx\" (UID: \"9402f824-5625-407a-8e63-82e168dc95ef\") " pod="kube-system/kube-proxy-wfsjx" Feb 13 19:50:59.383481 kubelet[2575]: I0213 19:50:59.383005 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9402f824-5625-407a-8e63-82e168dc95ef-lib-modules\") pod \"kube-proxy-wfsjx\" (UID: \"9402f824-5625-407a-8e63-82e168dc95ef\") " pod="kube-system/kube-proxy-wfsjx" Feb 13 19:50:59.383481 kubelet[2575]: I0213 19:50:59.383100 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gv59\" (UniqueName: \"kubernetes.io/projected/9402f824-5625-407a-8e63-82e168dc95ef-kube-api-access-6gv59\") pod \"kube-proxy-wfsjx\" (UID: \"9402f824-5625-407a-8e63-82e168dc95ef\") " pod="kube-system/kube-proxy-wfsjx" Feb 13 19:50:59.525248 sudo[1652]: pam_unix(sudo:session): session closed for user root Feb 13 19:50:59.527725 sshd[1651]: Connection closed by 10.0.0.1 port 54552 Feb 13 19:50:59.528584 sshd-session[1649]: pam_unix(sshd:session): session closed for user core Feb 13 19:50:59.546287 systemd[1]: sshd@7-10.0.0.104:22-10.0.0.1:54552.service: Deactivated successfully. Feb 13 19:50:59.548674 systemd[1]: session-7.scope: Deactivated successfully. Feb 13 19:50:59.548919 systemd[1]: session-7.scope: Consumed 5.415s CPU time, 152.9M memory peak, 0B memory swap peak. Feb 13 19:50:59.550194 systemd-logind[1447]: Session 7 logged out. Waiting for processes to exit. Feb 13 19:50:59.552475 systemd-logind[1447]: Removed session 7. Feb 13 19:50:59.662410 kubelet[2575]: E0213 19:50:59.662268 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:50:59.663137 containerd[1468]: time="2025-02-13T19:50:59.663079881Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-wfsjx,Uid:9402f824-5625-407a-8e63-82e168dc95ef,Namespace:kube-system,Attempt:0,}" Feb 13 19:50:59.860161 kubelet[2575]: E0213 19:50:59.860113 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:00.261176 containerd[1468]: time="2025-02-13T19:51:00.261047253Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:51:00.261387 containerd[1468]: time="2025-02-13T19:51:00.261318165Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:51:00.261387 containerd[1468]: time="2025-02-13T19:51:00.261358631Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:51:00.263145 containerd[1468]: time="2025-02-13T19:51:00.263010040Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:51:00.287623 systemd[1]: run-containerd-runc-k8s.io-f8c176a5364282a52f3edbdebd96ce292022c0582d71187f8d1bd7effbec4c69-runc.tz8Nih.mount: Deactivated successfully. Feb 13 19:51:00.293380 kubelet[2575]: I0213 19:51:00.293332 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9q6f\" (UniqueName: \"kubernetes.io/projected/aead79df-4d05-45b0-8fa7-54d0db61c09e-kube-api-access-d9q6f\") pod \"tigera-operator-7d68577dc5-gz24s\" (UID: \"aead79df-4d05-45b0-8fa7-54d0db61c09e\") " pod="tigera-operator/tigera-operator-7d68577dc5-gz24s" Feb 13 19:51:00.293380 kubelet[2575]: I0213 19:51:00.293384 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/aead79df-4d05-45b0-8fa7-54d0db61c09e-var-lib-calico\") pod \"tigera-operator-7d68577dc5-gz24s\" (UID: \"aead79df-4d05-45b0-8fa7-54d0db61c09e\") " pod="tigera-operator/tigera-operator-7d68577dc5-gz24s" Feb 13 19:51:00.297286 systemd[1]: Started cri-containerd-f8c176a5364282a52f3edbdebd96ce292022c0582d71187f8d1bd7effbec4c69.scope - libcontainer container f8c176a5364282a52f3edbdebd96ce292022c0582d71187f8d1bd7effbec4c69. Feb 13 19:51:00.303920 systemd[1]: Created slice kubepods-besteffort-podaead79df_4d05_45b0_8fa7_54d0db61c09e.slice - libcontainer container kubepods-besteffort-podaead79df_4d05_45b0_8fa7_54d0db61c09e.slice. Feb 13 19:51:00.322013 containerd[1468]: time="2025-02-13T19:51:00.321965469Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-wfsjx,Uid:9402f824-5625-407a-8e63-82e168dc95ef,Namespace:kube-system,Attempt:0,} returns sandbox id \"f8c176a5364282a52f3edbdebd96ce292022c0582d71187f8d1bd7effbec4c69\"" Feb 13 19:51:00.322861 kubelet[2575]: E0213 19:51:00.322772 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:00.325115 containerd[1468]: time="2025-02-13T19:51:00.325035627Z" level=info msg="CreateContainer within sandbox \"f8c176a5364282a52f3edbdebd96ce292022c0582d71187f8d1bd7effbec4c69\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Feb 13 19:51:00.345465 containerd[1468]: time="2025-02-13T19:51:00.345401545Z" level=info msg="CreateContainer within sandbox \"f8c176a5364282a52f3edbdebd96ce292022c0582d71187f8d1bd7effbec4c69\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"fa29b8c6a0840766ed626939f5f5311b00b791cd836f80f53ea75bedd00bcd1c\"" Feb 13 19:51:00.346258 containerd[1468]: time="2025-02-13T19:51:00.346208099Z" level=info msg="StartContainer for \"fa29b8c6a0840766ed626939f5f5311b00b791cd836f80f53ea75bedd00bcd1c\"" Feb 13 19:51:00.381047 systemd[1]: Started cri-containerd-fa29b8c6a0840766ed626939f5f5311b00b791cd836f80f53ea75bedd00bcd1c.scope - libcontainer container fa29b8c6a0840766ed626939f5f5311b00b791cd836f80f53ea75bedd00bcd1c. Feb 13 19:51:00.492855 containerd[1468]: time="2025-02-13T19:51:00.492810317Z" level=info msg="StartContainer for \"fa29b8c6a0840766ed626939f5f5311b00b791cd836f80f53ea75bedd00bcd1c\" returns successfully" Feb 13 19:51:00.607810 containerd[1468]: time="2025-02-13T19:51:00.607657229Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-7d68577dc5-gz24s,Uid:aead79df-4d05-45b0-8fa7-54d0db61c09e,Namespace:tigera-operator,Attempt:0,}" Feb 13 19:51:00.655626 kubelet[2575]: E0213 19:51:00.655594 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:00.656156 kubelet[2575]: E0213 19:51:00.655672 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:00.762387 kubelet[2575]: I0213 19:51:00.761830 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-wfsjx" podStartSLOduration=1.761805297 podStartE2EDuration="1.761805297s" podCreationTimestamp="2025-02-13 19:50:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 19:51:00.761770291 +0000 UTC m=+7.308086717" watchObservedRunningTime="2025-02-13 19:51:00.761805297 +0000 UTC m=+7.308121723" Feb 13 19:51:00.858992 containerd[1468]: time="2025-02-13T19:51:00.858511153Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:51:00.858992 containerd[1468]: time="2025-02-13T19:51:00.858585924Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:51:00.858992 containerd[1468]: time="2025-02-13T19:51:00.858600202Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:51:00.858992 containerd[1468]: time="2025-02-13T19:51:00.858789889Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:51:00.883138 systemd[1]: Started cri-containerd-40b7a9d867d1abbcc4dc45d319873a2ac32e906ad831889257ab5ebfcec7d3e8.scope - libcontainer container 40b7a9d867d1abbcc4dc45d319873a2ac32e906ad831889257ab5ebfcec7d3e8. Feb 13 19:51:00.926243 containerd[1468]: time="2025-02-13T19:51:00.926182021Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-7d68577dc5-gz24s,Uid:aead79df-4d05-45b0-8fa7-54d0db61c09e,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"40b7a9d867d1abbcc4dc45d319873a2ac32e906ad831889257ab5ebfcec7d3e8\"" Feb 13 19:51:00.928080 containerd[1468]: time="2025-02-13T19:51:00.928038015Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.36.2\"" Feb 13 19:51:02.533198 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2062893004.mount: Deactivated successfully. Feb 13 19:51:02.821756 containerd[1468]: time="2025-02-13T19:51:02.821622742Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.36.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:51:02.822934 containerd[1468]: time="2025-02-13T19:51:02.822892608Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.36.2: active requests=0, bytes read=21762497" Feb 13 19:51:02.824378 containerd[1468]: time="2025-02-13T19:51:02.824317035Z" level=info msg="ImageCreate event name:\"sha256:3045aa4a360d468ed15090f280e94c54bf4678269a6e863a9ebcf5b31534a346\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:51:02.827097 containerd[1468]: time="2025-02-13T19:51:02.827053399Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:fc9ea45f2475fd99db1b36d2ff180a50017b1a5ea0e82a171c6b439b3a620764\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:51:02.827837 containerd[1468]: time="2025-02-13T19:51:02.827791370Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.36.2\" with image id \"sha256:3045aa4a360d468ed15090f280e94c54bf4678269a6e863a9ebcf5b31534a346\", repo tag \"quay.io/tigera/operator:v1.36.2\", repo digest \"quay.io/tigera/operator@sha256:fc9ea45f2475fd99db1b36d2ff180a50017b1a5ea0e82a171c6b439b3a620764\", size \"21758492\" in 1.899701718s" Feb 13 19:51:02.827837 containerd[1468]: time="2025-02-13T19:51:02.827831016Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.36.2\" returns image reference \"sha256:3045aa4a360d468ed15090f280e94c54bf4678269a6e863a9ebcf5b31534a346\"" Feb 13 19:51:02.829960 containerd[1468]: time="2025-02-13T19:51:02.829927131Z" level=info msg="CreateContainer within sandbox \"40b7a9d867d1abbcc4dc45d319873a2ac32e906ad831889257ab5ebfcec7d3e8\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Feb 13 19:51:02.847817 containerd[1468]: time="2025-02-13T19:51:02.847758604Z" level=info msg="CreateContainer within sandbox \"40b7a9d867d1abbcc4dc45d319873a2ac32e906ad831889257ab5ebfcec7d3e8\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"2c192b09d719c529e99736b268433b7be357fa388df72dce7987ef27e56fe60a\"" Feb 13 19:51:02.848363 containerd[1468]: time="2025-02-13T19:51:02.848300056Z" level=info msg="StartContainer for \"2c192b09d719c529e99736b268433b7be357fa388df72dce7987ef27e56fe60a\"" Feb 13 19:51:02.884188 systemd[1]: Started cri-containerd-2c192b09d719c529e99736b268433b7be357fa388df72dce7987ef27e56fe60a.scope - libcontainer container 2c192b09d719c529e99736b268433b7be357fa388df72dce7987ef27e56fe60a. Feb 13 19:51:02.913072 containerd[1468]: time="2025-02-13T19:51:02.912985170Z" level=info msg="StartContainer for \"2c192b09d719c529e99736b268433b7be357fa388df72dce7987ef27e56fe60a\" returns successfully" Feb 13 19:51:03.671450 kubelet[2575]: I0213 19:51:03.671383 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-7d68577dc5-gz24s" podStartSLOduration=1.77014204 podStartE2EDuration="3.671362745s" podCreationTimestamp="2025-02-13 19:51:00 +0000 UTC" firstStartedPulling="2025-02-13 19:51:00.927468139 +0000 UTC m=+7.473784565" lastFinishedPulling="2025-02-13 19:51:02.828688844 +0000 UTC m=+9.375005270" observedRunningTime="2025-02-13 19:51:03.670661413 +0000 UTC m=+10.216977839" watchObservedRunningTime="2025-02-13 19:51:03.671362745 +0000 UTC m=+10.217679161" Feb 13 19:51:04.960902 kubelet[2575]: E0213 19:51:04.960852 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:05.967836 systemd[1]: Created slice kubepods-besteffort-podaf55f01b_e53e_49b2_9c2f_ac2bb56b4927.slice - libcontainer container kubepods-besteffort-podaf55f01b_e53e_49b2_9c2f_ac2bb56b4927.slice. Feb 13 19:51:05.983153 systemd[1]: Created slice kubepods-besteffort-pod2e17297f_a259_4a64_8bc5_1fa1a46a7c4c.slice - libcontainer container kubepods-besteffort-pod2e17297f_a259_4a64_8bc5_1fa1a46a7c4c.slice. Feb 13 19:51:06.023344 kubelet[2575]: I0213 19:51:06.023219 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2e17297f-a259-4a64-8bc5-1fa1a46a7c4c-lib-modules\") pod \"calico-node-x7j5b\" (UID: \"2e17297f-a259-4a64-8bc5-1fa1a46a7c4c\") " pod="calico-system/calico-node-x7j5b" Feb 13 19:51:06.023344 kubelet[2575]: I0213 19:51:06.023270 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/2e17297f-a259-4a64-8bc5-1fa1a46a7c4c-policysync\") pod \"calico-node-x7j5b\" (UID: \"2e17297f-a259-4a64-8bc5-1fa1a46a7c4c\") " pod="calico-system/calico-node-x7j5b" Feb 13 19:51:06.023344 kubelet[2575]: I0213 19:51:06.023289 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/2e17297f-a259-4a64-8bc5-1fa1a46a7c4c-node-certs\") pod \"calico-node-x7j5b\" (UID: \"2e17297f-a259-4a64-8bc5-1fa1a46a7c4c\") " pod="calico-system/calico-node-x7j5b" Feb 13 19:51:06.023344 kubelet[2575]: I0213 19:51:06.023307 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/af55f01b-e53e-49b2-9c2f-ac2bb56b4927-typha-certs\") pod \"calico-typha-6b7676bc7f-2tjsj\" (UID: \"af55f01b-e53e-49b2-9c2f-ac2bb56b4927\") " pod="calico-system/calico-typha-6b7676bc7f-2tjsj" Feb 13 19:51:06.023344 kubelet[2575]: I0213 19:51:06.023324 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2e17297f-a259-4a64-8bc5-1fa1a46a7c4c-tigera-ca-bundle\") pod \"calico-node-x7j5b\" (UID: \"2e17297f-a259-4a64-8bc5-1fa1a46a7c4c\") " pod="calico-system/calico-node-x7j5b" Feb 13 19:51:06.023905 kubelet[2575]: I0213 19:51:06.023339 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/2e17297f-a259-4a64-8bc5-1fa1a46a7c4c-var-run-calico\") pod \"calico-node-x7j5b\" (UID: \"2e17297f-a259-4a64-8bc5-1fa1a46a7c4c\") " pod="calico-system/calico-node-x7j5b" Feb 13 19:51:06.023905 kubelet[2575]: I0213 19:51:06.023356 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af55f01b-e53e-49b2-9c2f-ac2bb56b4927-tigera-ca-bundle\") pod \"calico-typha-6b7676bc7f-2tjsj\" (UID: \"af55f01b-e53e-49b2-9c2f-ac2bb56b4927\") " pod="calico-system/calico-typha-6b7676bc7f-2tjsj" Feb 13 19:51:06.023905 kubelet[2575]: I0213 19:51:06.023376 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/2e17297f-a259-4a64-8bc5-1fa1a46a7c4c-cni-net-dir\") pod \"calico-node-x7j5b\" (UID: \"2e17297f-a259-4a64-8bc5-1fa1a46a7c4c\") " pod="calico-system/calico-node-x7j5b" Feb 13 19:51:06.023905 kubelet[2575]: I0213 19:51:06.023394 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/2e17297f-a259-4a64-8bc5-1fa1a46a7c4c-flexvol-driver-host\") pod \"calico-node-x7j5b\" (UID: \"2e17297f-a259-4a64-8bc5-1fa1a46a7c4c\") " pod="calico-system/calico-node-x7j5b" Feb 13 19:51:06.023905 kubelet[2575]: I0213 19:51:06.023412 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/2e17297f-a259-4a64-8bc5-1fa1a46a7c4c-cni-bin-dir\") pod \"calico-node-x7j5b\" (UID: \"2e17297f-a259-4a64-8bc5-1fa1a46a7c4c\") " pod="calico-system/calico-node-x7j5b" Feb 13 19:51:06.024040 kubelet[2575]: I0213 19:51:06.023430 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/2e17297f-a259-4a64-8bc5-1fa1a46a7c4c-var-lib-calico\") pod \"calico-node-x7j5b\" (UID: \"2e17297f-a259-4a64-8bc5-1fa1a46a7c4c\") " pod="calico-system/calico-node-x7j5b" Feb 13 19:51:06.024040 kubelet[2575]: I0213 19:51:06.023447 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/2e17297f-a259-4a64-8bc5-1fa1a46a7c4c-cni-log-dir\") pod \"calico-node-x7j5b\" (UID: \"2e17297f-a259-4a64-8bc5-1fa1a46a7c4c\") " pod="calico-system/calico-node-x7j5b" Feb 13 19:51:06.024040 kubelet[2575]: I0213 19:51:06.023464 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4m2ww\" (UniqueName: \"kubernetes.io/projected/2e17297f-a259-4a64-8bc5-1fa1a46a7c4c-kube-api-access-4m2ww\") pod \"calico-node-x7j5b\" (UID: \"2e17297f-a259-4a64-8bc5-1fa1a46a7c4c\") " pod="calico-system/calico-node-x7j5b" Feb 13 19:51:06.024040 kubelet[2575]: I0213 19:51:06.023533 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bs5l\" (UniqueName: \"kubernetes.io/projected/af55f01b-e53e-49b2-9c2f-ac2bb56b4927-kube-api-access-4bs5l\") pod \"calico-typha-6b7676bc7f-2tjsj\" (UID: \"af55f01b-e53e-49b2-9c2f-ac2bb56b4927\") " pod="calico-system/calico-typha-6b7676bc7f-2tjsj" Feb 13 19:51:06.024040 kubelet[2575]: I0213 19:51:06.023584 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/2e17297f-a259-4a64-8bc5-1fa1a46a7c4c-xtables-lock\") pod \"calico-node-x7j5b\" (UID: \"2e17297f-a259-4a64-8bc5-1fa1a46a7c4c\") " pod="calico-system/calico-node-x7j5b" Feb 13 19:51:06.066352 kubelet[2575]: E0213 19:51:06.066181 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-sxwqd" podUID="bbd75eeb-e06b-4812-b2c6-522804c5ef2f" Feb 13 19:51:06.124748 kubelet[2575]: I0213 19:51:06.123798 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/bbd75eeb-e06b-4812-b2c6-522804c5ef2f-registration-dir\") pod \"csi-node-driver-sxwqd\" (UID: \"bbd75eeb-e06b-4812-b2c6-522804c5ef2f\") " pod="calico-system/csi-node-driver-sxwqd" Feb 13 19:51:06.124748 kubelet[2575]: I0213 19:51:06.123892 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/bbd75eeb-e06b-4812-b2c6-522804c5ef2f-varrun\") pod \"csi-node-driver-sxwqd\" (UID: \"bbd75eeb-e06b-4812-b2c6-522804c5ef2f\") " pod="calico-system/csi-node-driver-sxwqd" Feb 13 19:51:06.124748 kubelet[2575]: I0213 19:51:06.123917 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bbd75eeb-e06b-4812-b2c6-522804c5ef2f-kubelet-dir\") pod \"csi-node-driver-sxwqd\" (UID: \"bbd75eeb-e06b-4812-b2c6-522804c5ef2f\") " pod="calico-system/csi-node-driver-sxwqd" Feb 13 19:51:06.124748 kubelet[2575]: I0213 19:51:06.124068 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/bbd75eeb-e06b-4812-b2c6-522804c5ef2f-socket-dir\") pod \"csi-node-driver-sxwqd\" (UID: \"bbd75eeb-e06b-4812-b2c6-522804c5ef2f\") " pod="calico-system/csi-node-driver-sxwqd" Feb 13 19:51:06.124748 kubelet[2575]: I0213 19:51:06.124094 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5sjd\" (UniqueName: \"kubernetes.io/projected/bbd75eeb-e06b-4812-b2c6-522804c5ef2f-kube-api-access-t5sjd\") pod \"csi-node-driver-sxwqd\" (UID: \"bbd75eeb-e06b-4812-b2c6-522804c5ef2f\") " pod="calico-system/csi-node-driver-sxwqd" Feb 13 19:51:06.125067 kubelet[2575]: E0213 19:51:06.124996 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.125067 kubelet[2575]: W0213 19:51:06.125012 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.125067 kubelet[2575]: E0213 19:51:06.125037 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.125342 kubelet[2575]: E0213 19:51:06.125315 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.125342 kubelet[2575]: W0213 19:51:06.125329 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.125419 kubelet[2575]: E0213 19:51:06.125347 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.125618 kubelet[2575]: E0213 19:51:06.125595 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.125618 kubelet[2575]: W0213 19:51:06.125609 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.125722 kubelet[2575]: E0213 19:51:06.125626 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.126142 kubelet[2575]: E0213 19:51:06.126112 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.126142 kubelet[2575]: W0213 19:51:06.126127 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.126242 kubelet[2575]: E0213 19:51:06.126144 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.126526 kubelet[2575]: E0213 19:51:06.126466 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.126526 kubelet[2575]: W0213 19:51:06.126482 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.126765 kubelet[2575]: E0213 19:51:06.126605 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.126863 kubelet[2575]: E0213 19:51:06.126834 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.126863 kubelet[2575]: W0213 19:51:06.126857 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.127179 kubelet[2575]: E0213 19:51:06.127083 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.127444 kubelet[2575]: E0213 19:51:06.127423 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.127444 kubelet[2575]: W0213 19:51:06.127440 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.127663 kubelet[2575]: E0213 19:51:06.127485 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.127775 kubelet[2575]: E0213 19:51:06.127757 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.127775 kubelet[2575]: W0213 19:51:06.127773 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.128416 kubelet[2575]: E0213 19:51:06.127864 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.128746 kubelet[2575]: E0213 19:51:06.128726 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.128793 kubelet[2575]: W0213 19:51:06.128745 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.128935 kubelet[2575]: E0213 19:51:06.128915 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.129084 kubelet[2575]: E0213 19:51:06.129069 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.129129 kubelet[2575]: W0213 19:51:06.129083 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.129225 kubelet[2575]: E0213 19:51:06.129201 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.129323 kubelet[2575]: E0213 19:51:06.129308 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.129418 kubelet[2575]: W0213 19:51:06.129322 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.129418 kubelet[2575]: E0213 19:51:06.129392 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.129594 kubelet[2575]: E0213 19:51:06.129564 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.129594 kubelet[2575]: W0213 19:51:06.129593 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.129594 kubelet[2575]: E0213 19:51:06.129804 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.129594 kubelet[2575]: W0213 19:51:06.129818 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.129594 kubelet[2575]: E0213 19:51:06.130098 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.129594 kubelet[2575]: W0213 19:51:06.130108 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.129594 kubelet[2575]: E0213 19:51:06.130330 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.129594 kubelet[2575]: W0213 19:51:06.130340 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.130736 kubelet[2575]: E0213 19:51:06.130588 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.130736 kubelet[2575]: W0213 19:51:06.130596 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.131118 kubelet[2575]: E0213 19:51:06.130804 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.131118 kubelet[2575]: W0213 19:51:06.130812 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.131118 kubelet[2575]: E0213 19:51:06.130972 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.131118 kubelet[2575]: E0213 19:51:06.131008 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.131118 kubelet[2575]: E0213 19:51:06.131017 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.131118 kubelet[2575]: E0213 19:51:06.131029 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.131118 kubelet[2575]: E0213 19:51:06.131040 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.131118 kubelet[2575]: E0213 19:51:06.131042 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.131118 kubelet[2575]: E0213 19:51:06.131047 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.131118 kubelet[2575]: W0213 19:51:06.131052 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.131426 kubelet[2575]: E0213 19:51:06.131106 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.131426 kubelet[2575]: E0213 19:51:06.131253 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.131426 kubelet[2575]: W0213 19:51:06.131261 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.131426 kubelet[2575]: E0213 19:51:06.131271 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.131544 kubelet[2575]: E0213 19:51:06.131481 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.131544 kubelet[2575]: W0213 19:51:06.131488 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.131544 kubelet[2575]: E0213 19:51:06.131499 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.131941 kubelet[2575]: E0213 19:51:06.131911 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.131941 kubelet[2575]: W0213 19:51:06.131929 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.132029 kubelet[2575]: E0213 19:51:06.131989 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.132998 kubelet[2575]: E0213 19:51:06.132969 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.132998 kubelet[2575]: W0213 19:51:06.132985 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.133068 kubelet[2575]: E0213 19:51:06.133027 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.133366 kubelet[2575]: E0213 19:51:06.133345 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.133366 kubelet[2575]: W0213 19:51:06.133361 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.133436 kubelet[2575]: E0213 19:51:06.133417 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.133912 kubelet[2575]: E0213 19:51:06.133824 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.133912 kubelet[2575]: W0213 19:51:06.133908 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.137383 kubelet[2575]: E0213 19:51:06.137349 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.137503 kubelet[2575]: E0213 19:51:06.137449 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.137503 kubelet[2575]: W0213 19:51:06.137466 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.137929 kubelet[2575]: E0213 19:51:06.137843 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.138766 kubelet[2575]: E0213 19:51:06.138750 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.138766 kubelet[2575]: W0213 19:51:06.138765 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.139257 kubelet[2575]: E0213 19:51:06.139063 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.139257 kubelet[2575]: W0213 19:51:06.139077 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.139628 kubelet[2575]: E0213 19:51:06.139529 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.139628 kubelet[2575]: W0213 19:51:06.139541 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.139947 kubelet[2575]: E0213 19:51:06.139917 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.139991 kubelet[2575]: E0213 19:51:06.139947 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.140013 kubelet[2575]: E0213 19:51:06.139998 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.141146 kubelet[2575]: E0213 19:51:06.141051 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.141146 kubelet[2575]: W0213 19:51:06.141066 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.141146 kubelet[2575]: E0213 19:51:06.141103 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.141350 kubelet[2575]: E0213 19:51:06.141334 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.141350 kubelet[2575]: W0213 19:51:06.141348 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.141795 kubelet[2575]: E0213 19:51:06.141415 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.141795 kubelet[2575]: E0213 19:51:06.141603 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.141795 kubelet[2575]: W0213 19:51:06.141614 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.141795 kubelet[2575]: E0213 19:51:06.141737 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.141954 kubelet[2575]: E0213 19:51:06.141821 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.141954 kubelet[2575]: W0213 19:51:06.141833 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.141954 kubelet[2575]: E0213 19:51:06.141907 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.143121 kubelet[2575]: E0213 19:51:06.143095 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.143121 kubelet[2575]: W0213 19:51:06.143108 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.143224 kubelet[2575]: E0213 19:51:06.143205 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.144192 kubelet[2575]: E0213 19:51:06.144170 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.144192 kubelet[2575]: W0213 19:51:06.144188 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.144412 kubelet[2575]: E0213 19:51:06.144225 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.146449 kubelet[2575]: E0213 19:51:06.146419 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.146449 kubelet[2575]: W0213 19:51:06.146434 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.146583 kubelet[2575]: E0213 19:51:06.146545 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.148022 kubelet[2575]: E0213 19:51:06.148005 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.148022 kubelet[2575]: W0213 19:51:06.148020 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.148182 kubelet[2575]: E0213 19:51:06.148151 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.148378 kubelet[2575]: E0213 19:51:06.148364 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.148378 kubelet[2575]: W0213 19:51:06.148376 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.148574 kubelet[2575]: E0213 19:51:06.148481 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.148660 kubelet[2575]: E0213 19:51:06.148638 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.148660 kubelet[2575]: W0213 19:51:06.148651 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.149255 kubelet[2575]: E0213 19:51:06.149228 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.149523 kubelet[2575]: E0213 19:51:06.149495 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.149781 kubelet[2575]: W0213 19:51:06.149608 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.149956 kubelet[2575]: E0213 19:51:06.149901 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.151646 kubelet[2575]: E0213 19:51:06.151625 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.151646 kubelet[2575]: W0213 19:51:06.151642 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.152273 kubelet[2575]: E0213 19:51:06.151895 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.152273 kubelet[2575]: W0213 19:51:06.151909 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.152273 kubelet[2575]: E0213 19:51:06.151995 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.154354 kubelet[2575]: E0213 19:51:06.152674 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.154656 kubelet[2575]: E0213 19:51:06.154632 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.154656 kubelet[2575]: W0213 19:51:06.154652 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.155018 kubelet[2575]: E0213 19:51:06.154741 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.155018 kubelet[2575]: E0213 19:51:06.154971 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.155018 kubelet[2575]: W0213 19:51:06.154982 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.155307 kubelet[2575]: E0213 19:51:06.155068 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.155740 kubelet[2575]: E0213 19:51:06.155647 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.155740 kubelet[2575]: W0213 19:51:06.155663 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.155740 kubelet[2575]: E0213 19:51:06.155708 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.156261 kubelet[2575]: E0213 19:51:06.156152 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.156261 kubelet[2575]: W0213 19:51:06.156166 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.156738 kubelet[2575]: E0213 19:51:06.156717 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.158139 kubelet[2575]: E0213 19:51:06.158122 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.158255 kubelet[2575]: W0213 19:51:06.158139 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.158255 kubelet[2575]: E0213 19:51:06.158152 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.225705 kubelet[2575]: E0213 19:51:06.225029 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.225705 kubelet[2575]: W0213 19:51:06.225055 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.225705 kubelet[2575]: E0213 19:51:06.225077 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.225705 kubelet[2575]: E0213 19:51:06.225416 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.225705 kubelet[2575]: W0213 19:51:06.225447 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.225705 kubelet[2575]: E0213 19:51:06.225485 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.226095 kubelet[2575]: E0213 19:51:06.226077 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.226135 kubelet[2575]: W0213 19:51:06.226095 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.226135 kubelet[2575]: E0213 19:51:06.226118 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.226518 kubelet[2575]: E0213 19:51:06.226486 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.226564 kubelet[2575]: W0213 19:51:06.226520 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.226564 kubelet[2575]: E0213 19:51:06.226555 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.226833 kubelet[2575]: E0213 19:51:06.226817 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.226833 kubelet[2575]: W0213 19:51:06.226831 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.227096 kubelet[2575]: E0213 19:51:06.227030 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.228008 kubelet[2575]: E0213 19:51:06.227461 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.228008 kubelet[2575]: W0213 19:51:06.227484 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.228008 kubelet[2575]: E0213 19:51:06.227864 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.228197 kubelet[2575]: E0213 19:51:06.228168 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.228197 kubelet[2575]: W0213 19:51:06.228189 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.228264 kubelet[2575]: E0213 19:51:06.228238 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.228533 kubelet[2575]: E0213 19:51:06.228506 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.228533 kubelet[2575]: W0213 19:51:06.228522 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.228731 kubelet[2575]: E0213 19:51:06.228560 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.228769 kubelet[2575]: E0213 19:51:06.228749 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.228769 kubelet[2575]: W0213 19:51:06.228761 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.229025 kubelet[2575]: E0213 19:51:06.228912 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.229092 kubelet[2575]: E0213 19:51:06.229054 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.229092 kubelet[2575]: W0213 19:51:06.229064 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.229092 kubelet[2575]: E0213 19:51:06.229079 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.229335 kubelet[2575]: E0213 19:51:06.229313 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.229335 kubelet[2575]: W0213 19:51:06.229331 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.229428 kubelet[2575]: E0213 19:51:06.229354 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.229693 kubelet[2575]: E0213 19:51:06.229674 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.229693 kubelet[2575]: W0213 19:51:06.229689 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.229806 kubelet[2575]: E0213 19:51:06.229780 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.230065 kubelet[2575]: E0213 19:51:06.230048 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.230065 kubelet[2575]: W0213 19:51:06.230063 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.230149 kubelet[2575]: E0213 19:51:06.230097 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.230345 kubelet[2575]: E0213 19:51:06.230326 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.230345 kubelet[2575]: W0213 19:51:06.230342 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.230415 kubelet[2575]: E0213 19:51:06.230386 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.230555 kubelet[2575]: E0213 19:51:06.230537 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.230555 kubelet[2575]: W0213 19:51:06.230550 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.230714 kubelet[2575]: E0213 19:51:06.230655 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.230796 kubelet[2575]: E0213 19:51:06.230781 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.230796 kubelet[2575]: W0213 19:51:06.230794 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.230912 kubelet[2575]: E0213 19:51:06.230881 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.231037 kubelet[2575]: E0213 19:51:06.231022 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.231070 kubelet[2575]: W0213 19:51:06.231033 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.231070 kubelet[2575]: E0213 19:51:06.231059 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.231284 kubelet[2575]: E0213 19:51:06.231268 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.231284 kubelet[2575]: W0213 19:51:06.231279 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.231384 kubelet[2575]: E0213 19:51:06.231296 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.231505 kubelet[2575]: E0213 19:51:06.231488 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.231505 kubelet[2575]: W0213 19:51:06.231503 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.231599 kubelet[2575]: E0213 19:51:06.231540 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.231724 kubelet[2575]: E0213 19:51:06.231707 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.231724 kubelet[2575]: W0213 19:51:06.231719 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.231821 kubelet[2575]: E0213 19:51:06.231750 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.232006 kubelet[2575]: E0213 19:51:06.231986 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.232006 kubelet[2575]: W0213 19:51:06.232000 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.232160 kubelet[2575]: E0213 19:51:06.232092 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.232273 kubelet[2575]: E0213 19:51:06.232254 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.232273 kubelet[2575]: W0213 19:51:06.232268 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.232356 kubelet[2575]: E0213 19:51:06.232286 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.232663 kubelet[2575]: E0213 19:51:06.232638 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.232663 kubelet[2575]: W0213 19:51:06.232658 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.232720 kubelet[2575]: E0213 19:51:06.232676 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.233211 kubelet[2575]: E0213 19:51:06.233174 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.233211 kubelet[2575]: W0213 19:51:06.233201 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.233293 kubelet[2575]: E0213 19:51:06.233235 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.233563 kubelet[2575]: E0213 19:51:06.233520 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.233563 kubelet[2575]: W0213 19:51:06.233537 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.233563 kubelet[2575]: E0213 19:51:06.233549 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.242591 kubelet[2575]: E0213 19:51:06.242551 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:06.242591 kubelet[2575]: W0213 19:51:06.242571 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:06.242591 kubelet[2575]: E0213 19:51:06.242587 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:06.272733 kubelet[2575]: E0213 19:51:06.272694 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:06.273315 containerd[1468]: time="2025-02-13T19:51:06.273272019Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-6b7676bc7f-2tjsj,Uid:af55f01b-e53e-49b2-9c2f-ac2bb56b4927,Namespace:calico-system,Attempt:0,}" Feb 13 19:51:06.286889 kubelet[2575]: E0213 19:51:06.286808 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:06.287498 containerd[1468]: time="2025-02-13T19:51:06.287461403Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-x7j5b,Uid:2e17297f-a259-4a64-8bc5-1fa1a46a7c4c,Namespace:calico-system,Attempt:0,}" Feb 13 19:51:06.301407 containerd[1468]: time="2025-02-13T19:51:06.300905983Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:51:06.301407 containerd[1468]: time="2025-02-13T19:51:06.300996544Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:51:06.301407 containerd[1468]: time="2025-02-13T19:51:06.301011682Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:51:06.303053 containerd[1468]: time="2025-02-13T19:51:06.301854531Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:51:06.322045 systemd[1]: Started cri-containerd-32bf9ff275af19b849c126591477ff8ecf2177781f0eee08dfea940b94eac498.scope - libcontainer container 32bf9ff275af19b849c126591477ff8ecf2177781f0eee08dfea940b94eac498. Feb 13 19:51:06.346567 containerd[1468]: time="2025-02-13T19:51:06.346367978Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:51:06.346810 containerd[1468]: time="2025-02-13T19:51:06.346584596Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:51:06.346810 containerd[1468]: time="2025-02-13T19:51:06.346609192Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:51:06.347066 containerd[1468]: time="2025-02-13T19:51:06.347010779Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:51:06.371154 containerd[1468]: time="2025-02-13T19:51:06.371067823Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-6b7676bc7f-2tjsj,Uid:af55f01b-e53e-49b2-9c2f-ac2bb56b4927,Namespace:calico-system,Attempt:0,} returns sandbox id \"32bf9ff275af19b849c126591477ff8ecf2177781f0eee08dfea940b94eac498\"" Feb 13 19:51:06.372447 kubelet[2575]: E0213 19:51:06.372236 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:06.373122 containerd[1468]: time="2025-02-13T19:51:06.373098740Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.29.1\"" Feb 13 19:51:06.379069 systemd[1]: Started cri-containerd-f9006b0c7095feabafe4193f0ce61148f5b8f4c8d7f93c8d390cefc8f6f6014f.scope - libcontainer container f9006b0c7095feabafe4193f0ce61148f5b8f4c8d7f93c8d390cefc8f6f6014f. Feb 13 19:51:06.406783 containerd[1468]: time="2025-02-13T19:51:06.406723952Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-x7j5b,Uid:2e17297f-a259-4a64-8bc5-1fa1a46a7c4c,Namespace:calico-system,Attempt:0,} returns sandbox id \"f9006b0c7095feabafe4193f0ce61148f5b8f4c8d7f93c8d390cefc8f6f6014f\"" Feb 13 19:51:06.407701 kubelet[2575]: E0213 19:51:06.407671 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:07.628823 kubelet[2575]: E0213 19:51:07.628743 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-sxwqd" podUID="bbd75eeb-e06b-4812-b2c6-522804c5ef2f" Feb 13 19:51:08.496729 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3560306338.mount: Deactivated successfully. Feb 13 19:51:09.629722 kubelet[2575]: E0213 19:51:09.629662 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-sxwqd" podUID="bbd75eeb-e06b-4812-b2c6-522804c5ef2f" Feb 13 19:51:09.734691 containerd[1468]: time="2025-02-13T19:51:09.734603014Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:51:09.735641 containerd[1468]: time="2025-02-13T19:51:09.735582678Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.29.1: active requests=0, bytes read=31343363" Feb 13 19:51:09.736935 containerd[1468]: time="2025-02-13T19:51:09.736855204Z" level=info msg="ImageCreate event name:\"sha256:4cb3738506f5a9c530033d1e24fd6b9ec618518a2ec8b012ded33572be06ab44\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:51:09.738991 containerd[1468]: time="2025-02-13T19:51:09.738965076Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:768a194e1115c73bcbf35edb7afd18a63e16e08d940c79993565b6a3cca2da7c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:51:09.739584 containerd[1468]: time="2025-02-13T19:51:09.739539006Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.29.1\" with image id \"sha256:4cb3738506f5a9c530033d1e24fd6b9ec618518a2ec8b012ded33572be06ab44\", repo tag \"ghcr.io/flatcar/calico/typha:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:768a194e1115c73bcbf35edb7afd18a63e16e08d940c79993565b6a3cca2da7c\", size \"31343217\" in 3.366408886s" Feb 13 19:51:09.739626 containerd[1468]: time="2025-02-13T19:51:09.739585723Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.29.1\" returns image reference \"sha256:4cb3738506f5a9c530033d1e24fd6b9ec618518a2ec8b012ded33572be06ab44\"" Feb 13 19:51:09.740909 containerd[1468]: time="2025-02-13T19:51:09.740848982Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\"" Feb 13 19:51:09.749577 containerd[1468]: time="2025-02-13T19:51:09.749532238Z" level=info msg="CreateContainer within sandbox \"32bf9ff275af19b849c126591477ff8ecf2177781f0eee08dfea940b94eac498\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Feb 13 19:51:09.822538 containerd[1468]: time="2025-02-13T19:51:09.822468028Z" level=info msg="CreateContainer within sandbox \"32bf9ff275af19b849c126591477ff8ecf2177781f0eee08dfea940b94eac498\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"94afa009c912ece554ab9cc372c03221aae552a8dd07cc7bb03f630c78b5880f\"" Feb 13 19:51:09.823174 containerd[1468]: time="2025-02-13T19:51:09.823136676Z" level=info msg="StartContainer for \"94afa009c912ece554ab9cc372c03221aae552a8dd07cc7bb03f630c78b5880f\"" Feb 13 19:51:09.852183 systemd[1]: Started cri-containerd-94afa009c912ece554ab9cc372c03221aae552a8dd07cc7bb03f630c78b5880f.scope - libcontainer container 94afa009c912ece554ab9cc372c03221aae552a8dd07cc7bb03f630c78b5880f. Feb 13 19:51:09.896572 containerd[1468]: time="2025-02-13T19:51:09.896081764Z" level=info msg="StartContainer for \"94afa009c912ece554ab9cc372c03221aae552a8dd07cc7bb03f630c78b5880f\" returns successfully" Feb 13 19:51:10.677981 kubelet[2575]: E0213 19:51:10.677898 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:10.719505 kubelet[2575]: I0213 19:51:10.718382 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-6b7676bc7f-2tjsj" podStartSLOduration=2.350631798 podStartE2EDuration="5.718359207s" podCreationTimestamp="2025-02-13 19:51:05 +0000 UTC" firstStartedPulling="2025-02-13 19:51:06.372847656 +0000 UTC m=+12.919164083" lastFinishedPulling="2025-02-13 19:51:09.740575066 +0000 UTC m=+16.286891492" observedRunningTime="2025-02-13 19:51:10.717833788 +0000 UTC m=+17.264150234" watchObservedRunningTime="2025-02-13 19:51:10.718359207 +0000 UTC m=+17.264675653" Feb 13 19:51:10.747954 kubelet[2575]: E0213 19:51:10.747911 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:10.747954 kubelet[2575]: W0213 19:51:10.747945 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:10.754297 kubelet[2575]: E0213 19:51:10.754237 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:10.754709 kubelet[2575]: E0213 19:51:10.754677 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:10.754709 kubelet[2575]: W0213 19:51:10.754698 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:10.754709 kubelet[2575]: E0213 19:51:10.754719 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:10.755103 kubelet[2575]: E0213 19:51:10.755070 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:10.755103 kubelet[2575]: W0213 19:51:10.755080 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:10.755103 kubelet[2575]: E0213 19:51:10.755091 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:10.755486 kubelet[2575]: E0213 19:51:10.755447 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:10.755544 kubelet[2575]: W0213 19:51:10.755483 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:10.755596 kubelet[2575]: E0213 19:51:10.755548 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:10.756160 kubelet[2575]: E0213 19:51:10.756126 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:10.756160 kubelet[2575]: W0213 19:51:10.756144 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:10.756160 kubelet[2575]: E0213 19:51:10.756156 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:10.756494 kubelet[2575]: E0213 19:51:10.756467 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:10.756494 kubelet[2575]: W0213 19:51:10.756485 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:10.756596 kubelet[2575]: E0213 19:51:10.756502 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:10.756790 kubelet[2575]: E0213 19:51:10.756771 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:10.756790 kubelet[2575]: W0213 19:51:10.756786 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:10.756905 kubelet[2575]: E0213 19:51:10.756799 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:10.757094 kubelet[2575]: E0213 19:51:10.757076 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:10.757094 kubelet[2575]: W0213 19:51:10.757090 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:10.757177 kubelet[2575]: E0213 19:51:10.757102 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:10.757370 kubelet[2575]: E0213 19:51:10.757340 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:10.757370 kubelet[2575]: W0213 19:51:10.757353 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:10.757370 kubelet[2575]: E0213 19:51:10.757366 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:10.757625 kubelet[2575]: E0213 19:51:10.757602 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:10.757672 kubelet[2575]: W0213 19:51:10.757631 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:10.757672 kubelet[2575]: E0213 19:51:10.757649 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:10.758026 kubelet[2575]: E0213 19:51:10.757990 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:10.758026 kubelet[2575]: W0213 19:51:10.758022 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:10.758150 kubelet[2575]: E0213 19:51:10.758057 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:10.758381 kubelet[2575]: E0213 19:51:10.758364 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:10.758381 kubelet[2575]: W0213 19:51:10.758378 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:10.758458 kubelet[2575]: E0213 19:51:10.758389 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:10.758687 kubelet[2575]: E0213 19:51:10.758667 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:10.758687 kubelet[2575]: W0213 19:51:10.758682 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:10.758761 kubelet[2575]: E0213 19:51:10.758694 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:10.759052 kubelet[2575]: E0213 19:51:10.759020 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:10.759052 kubelet[2575]: W0213 19:51:10.759035 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:10.759169 kubelet[2575]: E0213 19:51:10.759050 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:10.759339 kubelet[2575]: E0213 19:51:10.759317 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:10.759339 kubelet[2575]: W0213 19:51:10.759331 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:10.759339 kubelet[2575]: E0213 19:51:10.759344 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:10.761807 kubelet[2575]: E0213 19:51:10.761770 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:10.761807 kubelet[2575]: W0213 19:51:10.761803 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:10.761937 kubelet[2575]: E0213 19:51:10.761834 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:10.762332 kubelet[2575]: E0213 19:51:10.762312 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:10.762332 kubelet[2575]: W0213 19:51:10.762330 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:10.762411 kubelet[2575]: E0213 19:51:10.762352 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:10.762740 kubelet[2575]: E0213 19:51:10.762717 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:10.762796 kubelet[2575]: W0213 19:51:10.762739 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:10.762796 kubelet[2575]: E0213 19:51:10.762763 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:10.763171 kubelet[2575]: E0213 19:51:10.763149 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:10.763171 kubelet[2575]: W0213 19:51:10.763169 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:10.763240 kubelet[2575]: E0213 19:51:10.763191 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:10.763485 kubelet[2575]: E0213 19:51:10.763465 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:10.763485 kubelet[2575]: W0213 19:51:10.763482 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:10.763576 kubelet[2575]: E0213 19:51:10.763501 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:10.763763 kubelet[2575]: E0213 19:51:10.763750 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:10.763763 kubelet[2575]: W0213 19:51:10.763761 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:10.763831 kubelet[2575]: E0213 19:51:10.763779 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:10.764127 kubelet[2575]: E0213 19:51:10.764099 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:10.764127 kubelet[2575]: W0213 19:51:10.764112 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:10.764324 kubelet[2575]: E0213 19:51:10.764164 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:10.764363 kubelet[2575]: E0213 19:51:10.764327 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:10.764363 kubelet[2575]: W0213 19:51:10.764337 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:10.764507 kubelet[2575]: E0213 19:51:10.764442 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:10.764733 kubelet[2575]: E0213 19:51:10.764714 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:10.764733 kubelet[2575]: W0213 19:51:10.764728 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:10.764814 kubelet[2575]: E0213 19:51:10.764760 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:10.765067 kubelet[2575]: E0213 19:51:10.765048 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:10.765106 kubelet[2575]: W0213 19:51:10.765073 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:10.765152 kubelet[2575]: E0213 19:51:10.765097 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:10.765482 kubelet[2575]: E0213 19:51:10.765464 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:10.765522 kubelet[2575]: W0213 19:51:10.765481 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:10.765522 kubelet[2575]: E0213 19:51:10.765505 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:10.765935 kubelet[2575]: E0213 19:51:10.765845 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:10.765935 kubelet[2575]: W0213 19:51:10.765891 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:10.765935 kubelet[2575]: E0213 19:51:10.765916 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:10.766371 kubelet[2575]: E0213 19:51:10.766332 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:10.766428 kubelet[2575]: W0213 19:51:10.766369 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:10.766428 kubelet[2575]: E0213 19:51:10.766405 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:10.766682 kubelet[2575]: E0213 19:51:10.766655 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:10.766682 kubelet[2575]: W0213 19:51:10.766669 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:10.766765 kubelet[2575]: E0213 19:51:10.766695 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:10.767001 kubelet[2575]: E0213 19:51:10.766966 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:10.767001 kubelet[2575]: W0213 19:51:10.766980 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:10.767001 kubelet[2575]: E0213 19:51:10.767001 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:10.767352 kubelet[2575]: E0213 19:51:10.767265 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:10.767352 kubelet[2575]: W0213 19:51:10.767280 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:10.767352 kubelet[2575]: E0213 19:51:10.767325 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:10.767925 kubelet[2575]: E0213 19:51:10.767896 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:10.767925 kubelet[2575]: W0213 19:51:10.767912 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:10.767925 kubelet[2575]: E0213 19:51:10.767932 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:10.768227 kubelet[2575]: E0213 19:51:10.768210 2575 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:51:10.768271 kubelet[2575]: W0213 19:51:10.768225 2575 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:51:10.768271 kubelet[2575]: E0213 19:51:10.768240 2575 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:51:11.576939 containerd[1468]: time="2025-02-13T19:51:11.576830947Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:51:11.577622 containerd[1468]: time="2025-02-13T19:51:11.577576620Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1: active requests=0, bytes read=5362121" Feb 13 19:51:11.578829 containerd[1468]: time="2025-02-13T19:51:11.578786807Z" level=info msg="ImageCreate event name:\"sha256:2b7452b763ec8833ca0386ada5fd066e552a9b3b02b8538a5e34cc3d6d3840a6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:51:11.581215 containerd[1468]: time="2025-02-13T19:51:11.581168898Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:a63f8b4ff531912d12d143664eb263fdbc6cd7b3ff4aa777dfb6e318a090462c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:51:11.582023 containerd[1468]: time="2025-02-13T19:51:11.581964225Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\" with image id \"sha256:2b7452b763ec8833ca0386ada5fd066e552a9b3b02b8538a5e34cc3d6d3840a6\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:a63f8b4ff531912d12d143664eb263fdbc6cd7b3ff4aa777dfb6e318a090462c\", size \"6855165\" in 1.841049169s" Feb 13 19:51:11.582023 containerd[1468]: time="2025-02-13T19:51:11.582018918Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\" returns image reference \"sha256:2b7452b763ec8833ca0386ada5fd066e552a9b3b02b8538a5e34cc3d6d3840a6\"" Feb 13 19:51:11.588333 containerd[1468]: time="2025-02-13T19:51:11.587149602Z" level=info msg="CreateContainer within sandbox \"f9006b0c7095feabafe4193f0ce61148f5b8f4c8d7f93c8d390cefc8f6f6014f\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Feb 13 19:51:11.606005 containerd[1468]: time="2025-02-13T19:51:11.605958612Z" level=info msg="CreateContainer within sandbox \"f9006b0c7095feabafe4193f0ce61148f5b8f4c8d7f93c8d390cefc8f6f6014f\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"6ed372d4841e41adb56bf0da3dbea4ced9eac12fe0af4fcfbc0907cf380fc5f4\"" Feb 13 19:51:11.606431 containerd[1468]: time="2025-02-13T19:51:11.606400714Z" level=info msg="StartContainer for \"6ed372d4841e41adb56bf0da3dbea4ced9eac12fe0af4fcfbc0907cf380fc5f4\"" Feb 13 19:51:11.629654 kubelet[2575]: E0213 19:51:11.629600 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-sxwqd" podUID="bbd75eeb-e06b-4812-b2c6-522804c5ef2f" Feb 13 19:51:11.643244 systemd[1]: Started cri-containerd-6ed372d4841e41adb56bf0da3dbea4ced9eac12fe0af4fcfbc0907cf380fc5f4.scope - libcontainer container 6ed372d4841e41adb56bf0da3dbea4ced9eac12fe0af4fcfbc0907cf380fc5f4. Feb 13 19:51:11.681861 containerd[1468]: time="2025-02-13T19:51:11.681794894Z" level=info msg="StartContainer for \"6ed372d4841e41adb56bf0da3dbea4ced9eac12fe0af4fcfbc0907cf380fc5f4\" returns successfully" Feb 13 19:51:11.685423 kubelet[2575]: I0213 19:51:11.685383 2575 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 13 19:51:11.685844 kubelet[2575]: E0213 19:51:11.685801 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:11.695355 systemd[1]: cri-containerd-6ed372d4841e41adb56bf0da3dbea4ced9eac12fe0af4fcfbc0907cf380fc5f4.scope: Deactivated successfully. Feb 13 19:51:11.747325 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-6ed372d4841e41adb56bf0da3dbea4ced9eac12fe0af4fcfbc0907cf380fc5f4-rootfs.mount: Deactivated successfully. Feb 13 19:51:12.288939 containerd[1468]: time="2025-02-13T19:51:12.288844603Z" level=info msg="shim disconnected" id=6ed372d4841e41adb56bf0da3dbea4ced9eac12fe0af4fcfbc0907cf380fc5f4 namespace=k8s.io Feb 13 19:51:12.288939 containerd[1468]: time="2025-02-13T19:51:12.288932939Z" level=warning msg="cleaning up after shim disconnected" id=6ed372d4841e41adb56bf0da3dbea4ced9eac12fe0af4fcfbc0907cf380fc5f4 namespace=k8s.io Feb 13 19:51:12.288939 containerd[1468]: time="2025-02-13T19:51:12.288945343Z" level=info msg="cleaning up dead shim" namespace=k8s.io Feb 13 19:51:12.687603 kubelet[2575]: E0213 19:51:12.687432 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:12.688421 containerd[1468]: time="2025-02-13T19:51:12.688193923Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.1\"" Feb 13 19:51:12.993178 kubelet[2575]: I0213 19:51:12.993136 2575 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 13 19:51:12.993587 kubelet[2575]: E0213 19:51:12.993567 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:13.629041 kubelet[2575]: E0213 19:51:13.628955 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-sxwqd" podUID="bbd75eeb-e06b-4812-b2c6-522804c5ef2f" Feb 13 19:51:13.690041 kubelet[2575]: E0213 19:51:13.690001 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:15.628818 kubelet[2575]: E0213 19:51:15.628744 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-sxwqd" podUID="bbd75eeb-e06b-4812-b2c6-522804c5ef2f" Feb 13 19:51:17.628727 kubelet[2575]: E0213 19:51:17.628660 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-sxwqd" podUID="bbd75eeb-e06b-4812-b2c6-522804c5ef2f" Feb 13 19:51:18.453583 containerd[1468]: time="2025-02-13T19:51:18.453520835Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:51:18.454383 containerd[1468]: time="2025-02-13T19:51:18.454335806Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.29.1: active requests=0, bytes read=96154154" Feb 13 19:51:18.455734 containerd[1468]: time="2025-02-13T19:51:18.455686264Z" level=info msg="ImageCreate event name:\"sha256:7dd6ea186aba0d7a1791a79d426fe854527ca95192b26bbd19e8baf8373f7d0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:51:18.457853 containerd[1468]: time="2025-02-13T19:51:18.457824984Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:21e759d51c90dfb34fc1397dc180dd3a3fb564c2b0580d2f61ffe108f2a3c94b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:51:18.458469 containerd[1468]: time="2025-02-13T19:51:18.458425351Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.29.1\" with image id \"sha256:7dd6ea186aba0d7a1791a79d426fe854527ca95192b26bbd19e8baf8373f7d0e\", repo tag \"ghcr.io/flatcar/calico/cni:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:21e759d51c90dfb34fc1397dc180dd3a3fb564c2b0580d2f61ffe108f2a3c94b\", size \"97647238\" in 5.770190402s" Feb 13 19:51:18.458469 containerd[1468]: time="2025-02-13T19:51:18.458468532Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.1\" returns image reference \"sha256:7dd6ea186aba0d7a1791a79d426fe854527ca95192b26bbd19e8baf8373f7d0e\"" Feb 13 19:51:18.468538 containerd[1468]: time="2025-02-13T19:51:18.468504162Z" level=info msg="CreateContainer within sandbox \"f9006b0c7095feabafe4193f0ce61148f5b8f4c8d7f93c8d390cefc8f6f6014f\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Feb 13 19:51:18.486927 containerd[1468]: time="2025-02-13T19:51:18.486858672Z" level=info msg="CreateContainer within sandbox \"f9006b0c7095feabafe4193f0ce61148f5b8f4c8d7f93c8d390cefc8f6f6014f\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"427cc5ceb3960374e40a172dcf7e2280637717b6227d9697d447467419f5ef09\"" Feb 13 19:51:18.489796 containerd[1468]: time="2025-02-13T19:51:18.489763191Z" level=info msg="StartContainer for \"427cc5ceb3960374e40a172dcf7e2280637717b6227d9697d447467419f5ef09\"" Feb 13 19:51:18.521045 systemd[1]: Started cri-containerd-427cc5ceb3960374e40a172dcf7e2280637717b6227d9697d447467419f5ef09.scope - libcontainer container 427cc5ceb3960374e40a172dcf7e2280637717b6227d9697d447467419f5ef09. Feb 13 19:51:18.553070 containerd[1468]: time="2025-02-13T19:51:18.553027907Z" level=info msg="StartContainer for \"427cc5ceb3960374e40a172dcf7e2280637717b6227d9697d447467419f5ef09\" returns successfully" Feb 13 19:51:19.159319 kubelet[2575]: E0213 19:51:19.158786 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:19.628892 kubelet[2575]: E0213 19:51:19.628820 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-sxwqd" podUID="bbd75eeb-e06b-4812-b2c6-522804c5ef2f" Feb 13 19:51:19.724935 containerd[1468]: time="2025-02-13T19:51:19.724884099Z" level=error msg="failed to reload cni configuration after receiving fs change event(WRITE \"/etc/cni/net.d/calico-kubeconfig\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Feb 13 19:51:19.727788 systemd[1]: cri-containerd-427cc5ceb3960374e40a172dcf7e2280637717b6227d9697d447467419f5ef09.scope: Deactivated successfully. Feb 13 19:51:19.747839 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-427cc5ceb3960374e40a172dcf7e2280637717b6227d9697d447467419f5ef09-rootfs.mount: Deactivated successfully. Feb 13 19:51:19.815313 kubelet[2575]: I0213 19:51:19.815270 2575 kubelet_node_status.go:502] "Fast updating node status as it just became ready" Feb 13 19:51:20.021022 systemd[1]: Created slice kubepods-besteffort-poda5e91fd3_19fa_490e_be13_26d807029222.slice - libcontainer container kubepods-besteffort-poda5e91fd3_19fa_490e_be13_26d807029222.slice. Feb 13 19:51:20.026724 systemd[1]: Created slice kubepods-burstable-podbd8b42bf_05e5_489c_87c7_f5984836cd6a.slice - libcontainer container kubepods-burstable-podbd8b42bf_05e5_489c_87c7_f5984836cd6a.slice. Feb 13 19:51:20.034479 systemd[1]: Created slice kubepods-besteffort-poddca83c7e_397c_4cb7_b95d_96c1c013b67f.slice - libcontainer container kubepods-besteffort-poddca83c7e_397c_4cb7_b95d_96c1c013b67f.slice. Feb 13 19:51:20.046181 systemd[1]: Created slice kubepods-burstable-pod245734af_c06e_49f3_bfef_a966f57e2f98.slice - libcontainer container kubepods-burstable-pod245734af_c06e_49f3_bfef_a966f57e2f98.slice. Feb 13 19:51:20.052300 systemd[1]: Created slice kubepods-besteffort-pod77a9c264_ba41_4ec0_b704_771924c99471.slice - libcontainer container kubepods-besteffort-pod77a9c264_ba41_4ec0_b704_771924c99471.slice. Feb 13 19:51:20.062954 kubelet[2575]: I0213 19:51:20.062891 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/77a9c264-ba41-4ec0-b704-771924c99471-tigera-ca-bundle\") pod \"calico-kube-controllers-5bf7d587cb-s28ls\" (UID: \"77a9c264-ba41-4ec0-b704-771924c99471\") " pod="calico-system/calico-kube-controllers-5bf7d587cb-s28ls" Feb 13 19:51:20.062954 kubelet[2575]: I0213 19:51:20.062948 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjff9\" (UniqueName: \"kubernetes.io/projected/77a9c264-ba41-4ec0-b704-771924c99471-kube-api-access-bjff9\") pod \"calico-kube-controllers-5bf7d587cb-s28ls\" (UID: \"77a9c264-ba41-4ec0-b704-771924c99471\") " pod="calico-system/calico-kube-controllers-5bf7d587cb-s28ls" Feb 13 19:51:20.063150 kubelet[2575]: I0213 19:51:20.062984 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnfcv\" (UniqueName: \"kubernetes.io/projected/245734af-c06e-49f3-bfef-a966f57e2f98-kube-api-access-jnfcv\") pod \"coredns-668d6bf9bc-6d5rf\" (UID: \"245734af-c06e-49f3-bfef-a966f57e2f98\") " pod="kube-system/coredns-668d6bf9bc-6d5rf" Feb 13 19:51:20.063150 kubelet[2575]: I0213 19:51:20.063008 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/dca83c7e-397c-4cb7-b95d-96c1c013b67f-calico-apiserver-certs\") pod \"calico-apiserver-575f8f688d-rm27z\" (UID: \"dca83c7e-397c-4cb7-b95d-96c1c013b67f\") " pod="calico-apiserver/calico-apiserver-575f8f688d-rm27z" Feb 13 19:51:20.063150 kubelet[2575]: I0213 19:51:20.063031 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/a5e91fd3-19fa-490e-be13-26d807029222-calico-apiserver-certs\") pod \"calico-apiserver-575f8f688d-8855n\" (UID: \"a5e91fd3-19fa-490e-be13-26d807029222\") " pod="calico-apiserver/calico-apiserver-575f8f688d-8855n" Feb 13 19:51:20.063150 kubelet[2575]: I0213 19:51:20.063058 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54kl2\" (UniqueName: \"kubernetes.io/projected/a5e91fd3-19fa-490e-be13-26d807029222-kube-api-access-54kl2\") pod \"calico-apiserver-575f8f688d-8855n\" (UID: \"a5e91fd3-19fa-490e-be13-26d807029222\") " pod="calico-apiserver/calico-apiserver-575f8f688d-8855n" Feb 13 19:51:20.063245 kubelet[2575]: I0213 19:51:20.063149 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bd8b42bf-05e5-489c-87c7-f5984836cd6a-config-volume\") pod \"coredns-668d6bf9bc-s22xp\" (UID: \"bd8b42bf-05e5-489c-87c7-f5984836cd6a\") " pod="kube-system/coredns-668d6bf9bc-s22xp" Feb 13 19:51:20.063245 kubelet[2575]: I0213 19:51:20.063182 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qc5bn\" (UniqueName: \"kubernetes.io/projected/bd8b42bf-05e5-489c-87c7-f5984836cd6a-kube-api-access-qc5bn\") pod \"coredns-668d6bf9bc-s22xp\" (UID: \"bd8b42bf-05e5-489c-87c7-f5984836cd6a\") " pod="kube-system/coredns-668d6bf9bc-s22xp" Feb 13 19:51:20.063245 kubelet[2575]: I0213 19:51:20.063205 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzq49\" (UniqueName: \"kubernetes.io/projected/dca83c7e-397c-4cb7-b95d-96c1c013b67f-kube-api-access-dzq49\") pod \"calico-apiserver-575f8f688d-rm27z\" (UID: \"dca83c7e-397c-4cb7-b95d-96c1c013b67f\") " pod="calico-apiserver/calico-apiserver-575f8f688d-rm27z" Feb 13 19:51:20.063245 kubelet[2575]: I0213 19:51:20.063225 2575 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/245734af-c06e-49f3-bfef-a966f57e2f98-config-volume\") pod \"coredns-668d6bf9bc-6d5rf\" (UID: \"245734af-c06e-49f3-bfef-a966f57e2f98\") " pod="kube-system/coredns-668d6bf9bc-6d5rf" Feb 13 19:51:20.130094 containerd[1468]: time="2025-02-13T19:51:20.130020988Z" level=info msg="shim disconnected" id=427cc5ceb3960374e40a172dcf7e2280637717b6227d9697d447467419f5ef09 namespace=k8s.io Feb 13 19:51:20.130094 containerd[1468]: time="2025-02-13T19:51:20.130086341Z" level=warning msg="cleaning up after shim disconnected" id=427cc5ceb3960374e40a172dcf7e2280637717b6227d9697d447467419f5ef09 namespace=k8s.io Feb 13 19:51:20.130094 containerd[1468]: time="2025-02-13T19:51:20.130097932Z" level=info msg="cleaning up dead shim" namespace=k8s.io Feb 13 19:51:20.161328 kubelet[2575]: E0213 19:51:20.161294 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:20.161917 containerd[1468]: time="2025-02-13T19:51:20.161860284Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.1\"" Feb 13 19:51:20.325379 containerd[1468]: time="2025-02-13T19:51:20.325249189Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-575f8f688d-8855n,Uid:a5e91fd3-19fa-490e-be13-26d807029222,Namespace:calico-apiserver,Attempt:0,}" Feb 13 19:51:20.332822 kubelet[2575]: E0213 19:51:20.332769 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:20.333516 containerd[1468]: time="2025-02-13T19:51:20.333383611Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-s22xp,Uid:bd8b42bf-05e5-489c-87c7-f5984836cd6a,Namespace:kube-system,Attempt:0,}" Feb 13 19:51:20.341934 containerd[1468]: time="2025-02-13T19:51:20.341801455Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-575f8f688d-rm27z,Uid:dca83c7e-397c-4cb7-b95d-96c1c013b67f,Namespace:calico-apiserver,Attempt:0,}" Feb 13 19:51:20.350384 kubelet[2575]: E0213 19:51:20.350222 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:20.352105 containerd[1468]: time="2025-02-13T19:51:20.352053194Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-6d5rf,Uid:245734af-c06e-49f3-bfef-a966f57e2f98,Namespace:kube-system,Attempt:0,}" Feb 13 19:51:20.356756 containerd[1468]: time="2025-02-13T19:51:20.356702148Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5bf7d587cb-s28ls,Uid:77a9c264-ba41-4ec0-b704-771924c99471,Namespace:calico-system,Attempt:0,}" Feb 13 19:51:20.462171 containerd[1468]: time="2025-02-13T19:51:20.462110436Z" level=error msg="Failed to destroy network for sandbox \"517d17b8f015a9521e2907d73d8ffe6db808ece1034c7f320e60f66e57f87de3\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:20.463741 containerd[1468]: time="2025-02-13T19:51:20.463525274Z" level=error msg="encountered an error cleaning up failed sandbox \"517d17b8f015a9521e2907d73d8ffe6db808ece1034c7f320e60f66e57f87de3\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:20.463741 containerd[1468]: time="2025-02-13T19:51:20.463611456Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-575f8f688d-8855n,Uid:a5e91fd3-19fa-490e-be13-26d807029222,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"517d17b8f015a9521e2907d73d8ffe6db808ece1034c7f320e60f66e57f87de3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:20.464631 kubelet[2575]: E0213 19:51:20.464093 2575 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"517d17b8f015a9521e2907d73d8ffe6db808ece1034c7f320e60f66e57f87de3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:20.464631 kubelet[2575]: E0213 19:51:20.464192 2575 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"517d17b8f015a9521e2907d73d8ffe6db808ece1034c7f320e60f66e57f87de3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-575f8f688d-8855n" Feb 13 19:51:20.464631 kubelet[2575]: E0213 19:51:20.464220 2575 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"517d17b8f015a9521e2907d73d8ffe6db808ece1034c7f320e60f66e57f87de3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-575f8f688d-8855n" Feb 13 19:51:20.464956 kubelet[2575]: E0213 19:51:20.464286 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-575f8f688d-8855n_calico-apiserver(a5e91fd3-19fa-490e-be13-26d807029222)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-575f8f688d-8855n_calico-apiserver(a5e91fd3-19fa-490e-be13-26d807029222)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"517d17b8f015a9521e2907d73d8ffe6db808ece1034c7f320e60f66e57f87de3\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-575f8f688d-8855n" podUID="a5e91fd3-19fa-490e-be13-26d807029222" Feb 13 19:51:20.479580 containerd[1468]: time="2025-02-13T19:51:20.479476712Z" level=error msg="Failed to destroy network for sandbox \"4ff7a1ce33e60be86627d63cbaa39e854dfba3ccfcd25d3b18dc72196a2b96da\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:20.480079 containerd[1468]: time="2025-02-13T19:51:20.479927468Z" level=error msg="Failed to destroy network for sandbox \"aea5462f72694272b5bc284a0bd9fc173269129df512c011930e89667ad0f97d\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:20.480079 containerd[1468]: time="2025-02-13T19:51:20.480050871Z" level=error msg="encountered an error cleaning up failed sandbox \"4ff7a1ce33e60be86627d63cbaa39e854dfba3ccfcd25d3b18dc72196a2b96da\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:20.480262 containerd[1468]: time="2025-02-13T19:51:20.480124980Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-575f8f688d-rm27z,Uid:dca83c7e-397c-4cb7-b95d-96c1c013b67f,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"4ff7a1ce33e60be86627d63cbaa39e854dfba3ccfcd25d3b18dc72196a2b96da\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:20.480501 kubelet[2575]: E0213 19:51:20.480447 2575 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4ff7a1ce33e60be86627d63cbaa39e854dfba3ccfcd25d3b18dc72196a2b96da\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:20.480579 kubelet[2575]: E0213 19:51:20.480532 2575 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4ff7a1ce33e60be86627d63cbaa39e854dfba3ccfcd25d3b18dc72196a2b96da\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-575f8f688d-rm27z" Feb 13 19:51:20.480579 kubelet[2575]: E0213 19:51:20.480561 2575 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4ff7a1ce33e60be86627d63cbaa39e854dfba3ccfcd25d3b18dc72196a2b96da\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-575f8f688d-rm27z" Feb 13 19:51:20.480664 kubelet[2575]: E0213 19:51:20.480611 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-575f8f688d-rm27z_calico-apiserver(dca83c7e-397c-4cb7-b95d-96c1c013b67f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-575f8f688d-rm27z_calico-apiserver(dca83c7e-397c-4cb7-b95d-96c1c013b67f)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"4ff7a1ce33e60be86627d63cbaa39e854dfba3ccfcd25d3b18dc72196a2b96da\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-575f8f688d-rm27z" podUID="dca83c7e-397c-4cb7-b95d-96c1c013b67f" Feb 13 19:51:20.480764 containerd[1468]: time="2025-02-13T19:51:20.480612315Z" level=error msg="encountered an error cleaning up failed sandbox \"aea5462f72694272b5bc284a0bd9fc173269129df512c011930e89667ad0f97d\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:20.480764 containerd[1468]: time="2025-02-13T19:51:20.480677017Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-s22xp,Uid:bd8b42bf-05e5-489c-87c7-f5984836cd6a,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"aea5462f72694272b5bc284a0bd9fc173269129df512c011930e89667ad0f97d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:20.481057 kubelet[2575]: E0213 19:51:20.480919 2575 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"aea5462f72694272b5bc284a0bd9fc173269129df512c011930e89667ad0f97d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:20.481057 kubelet[2575]: E0213 19:51:20.480965 2575 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"aea5462f72694272b5bc284a0bd9fc173269129df512c011930e89667ad0f97d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-s22xp" Feb 13 19:51:20.481057 kubelet[2575]: E0213 19:51:20.480987 2575 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"aea5462f72694272b5bc284a0bd9fc173269129df512c011930e89667ad0f97d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-s22xp" Feb 13 19:51:20.481173 kubelet[2575]: E0213 19:51:20.481020 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-s22xp_kube-system(bd8b42bf-05e5-489c-87c7-f5984836cd6a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-s22xp_kube-system(bd8b42bf-05e5-489c-87c7-f5984836cd6a)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"aea5462f72694272b5bc284a0bd9fc173269129df512c011930e89667ad0f97d\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-s22xp" podUID="bd8b42bf-05e5-489c-87c7-f5984836cd6a" Feb 13 19:51:20.483070 containerd[1468]: time="2025-02-13T19:51:20.483022473Z" level=error msg="Failed to destroy network for sandbox \"a24dd5ac08e1be65e64f1e0e1bcc57fba990025660893d49024096981b2ba3aa\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:20.483564 containerd[1468]: time="2025-02-13T19:51:20.483471366Z" level=error msg="encountered an error cleaning up failed sandbox \"a24dd5ac08e1be65e64f1e0e1bcc57fba990025660893d49024096981b2ba3aa\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:20.483779 containerd[1468]: time="2025-02-13T19:51:20.483591442Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-6d5rf,Uid:245734af-c06e-49f3-bfef-a966f57e2f98,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"a24dd5ac08e1be65e64f1e0e1bcc57fba990025660893d49024096981b2ba3aa\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:20.483946 kubelet[2575]: E0213 19:51:20.483909 2575 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a24dd5ac08e1be65e64f1e0e1bcc57fba990025660893d49024096981b2ba3aa\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:20.483995 kubelet[2575]: E0213 19:51:20.483969 2575 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a24dd5ac08e1be65e64f1e0e1bcc57fba990025660893d49024096981b2ba3aa\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-6d5rf" Feb 13 19:51:20.484026 kubelet[2575]: E0213 19:51:20.483991 2575 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a24dd5ac08e1be65e64f1e0e1bcc57fba990025660893d49024096981b2ba3aa\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-6d5rf" Feb 13 19:51:20.484077 kubelet[2575]: E0213 19:51:20.484048 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-6d5rf_kube-system(245734af-c06e-49f3-bfef-a966f57e2f98)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-6d5rf_kube-system(245734af-c06e-49f3-bfef-a966f57e2f98)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"a24dd5ac08e1be65e64f1e0e1bcc57fba990025660893d49024096981b2ba3aa\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-6d5rf" podUID="245734af-c06e-49f3-bfef-a966f57e2f98" Feb 13 19:51:20.499951 containerd[1468]: time="2025-02-13T19:51:20.499830310Z" level=error msg="Failed to destroy network for sandbox \"c143929d3d37d0fd39eeb807647ccdb8bf35b3017d21a93a728ef6fa513987e4\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:20.500455 containerd[1468]: time="2025-02-13T19:51:20.500420759Z" level=error msg="encountered an error cleaning up failed sandbox \"c143929d3d37d0fd39eeb807647ccdb8bf35b3017d21a93a728ef6fa513987e4\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:20.500537 containerd[1468]: time="2025-02-13T19:51:20.500507361Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5bf7d587cb-s28ls,Uid:77a9c264-ba41-4ec0-b704-771924c99471,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"c143929d3d37d0fd39eeb807647ccdb8bf35b3017d21a93a728ef6fa513987e4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:20.500844 kubelet[2575]: E0213 19:51:20.500793 2575 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c143929d3d37d0fd39eeb807647ccdb8bf35b3017d21a93a728ef6fa513987e4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:20.500972 kubelet[2575]: E0213 19:51:20.500890 2575 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c143929d3d37d0fd39eeb807647ccdb8bf35b3017d21a93a728ef6fa513987e4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5bf7d587cb-s28ls" Feb 13 19:51:20.500972 kubelet[2575]: E0213 19:51:20.500929 2575 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c143929d3d37d0fd39eeb807647ccdb8bf35b3017d21a93a728ef6fa513987e4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5bf7d587cb-s28ls" Feb 13 19:51:20.501079 kubelet[2575]: E0213 19:51:20.500997 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-5bf7d587cb-s28ls_calico-system(77a9c264-ba41-4ec0-b704-771924c99471)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-5bf7d587cb-s28ls_calico-system(77a9c264-ba41-4ec0-b704-771924c99471)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"c143929d3d37d0fd39eeb807647ccdb8bf35b3017d21a93a728ef6fa513987e4\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-5bf7d587cb-s28ls" podUID="77a9c264-ba41-4ec0-b704-771924c99471" Feb 13 19:51:21.164287 kubelet[2575]: I0213 19:51:21.164239 2575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a24dd5ac08e1be65e64f1e0e1bcc57fba990025660893d49024096981b2ba3aa" Feb 13 19:51:21.165340 containerd[1468]: time="2025-02-13T19:51:21.164894422Z" level=info msg="StopPodSandbox for \"a24dd5ac08e1be65e64f1e0e1bcc57fba990025660893d49024096981b2ba3aa\"" Feb 13 19:51:21.165340 containerd[1468]: time="2025-02-13T19:51:21.165165111Z" level=info msg="Ensure that sandbox a24dd5ac08e1be65e64f1e0e1bcc57fba990025660893d49024096981b2ba3aa in task-service has been cleanup successfully" Feb 13 19:51:21.165896 kubelet[2575]: I0213 19:51:21.165152 2575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aea5462f72694272b5bc284a0bd9fc173269129df512c011930e89667ad0f97d" Feb 13 19:51:21.165951 containerd[1468]: time="2025-02-13T19:51:21.165495672Z" level=info msg="StopPodSandbox for \"aea5462f72694272b5bc284a0bd9fc173269129df512c011930e89667ad0f97d\"" Feb 13 19:51:21.165951 containerd[1468]: time="2025-02-13T19:51:21.165723280Z" level=info msg="Ensure that sandbox aea5462f72694272b5bc284a0bd9fc173269129df512c011930e89667ad0f97d in task-service has been cleanup successfully" Feb 13 19:51:21.166699 kubelet[2575]: I0213 19:51:21.166654 2575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="517d17b8f015a9521e2907d73d8ffe6db808ece1034c7f320e60f66e57f87de3" Feb 13 19:51:21.167188 containerd[1468]: time="2025-02-13T19:51:21.167167542Z" level=info msg="StopPodSandbox for \"517d17b8f015a9521e2907d73d8ffe6db808ece1034c7f320e60f66e57f87de3\"" Feb 13 19:51:21.167358 containerd[1468]: time="2025-02-13T19:51:21.167329396Z" level=info msg="Ensure that sandbox 517d17b8f015a9521e2907d73d8ffe6db808ece1034c7f320e60f66e57f87de3 in task-service has been cleanup successfully" Feb 13 19:51:21.168003 kubelet[2575]: I0213 19:51:21.167985 2575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c143929d3d37d0fd39eeb807647ccdb8bf35b3017d21a93a728ef6fa513987e4" Feb 13 19:51:21.168361 containerd[1468]: time="2025-02-13T19:51:21.168341488Z" level=info msg="StopPodSandbox for \"c143929d3d37d0fd39eeb807647ccdb8bf35b3017d21a93a728ef6fa513987e4\"" Feb 13 19:51:21.168509 containerd[1468]: time="2025-02-13T19:51:21.168493012Z" level=info msg="Ensure that sandbox c143929d3d37d0fd39eeb807647ccdb8bf35b3017d21a93a728ef6fa513987e4 in task-service has been cleanup successfully" Feb 13 19:51:21.168822 systemd[1]: run-netns-cni\x2d6768807d\x2d7813\x2dd23d\x2de284\x2deb832f363b2a.mount: Deactivated successfully. Feb 13 19:51:21.169238 containerd[1468]: time="2025-02-13T19:51:21.169209178Z" level=info msg="TearDown network for sandbox \"a24dd5ac08e1be65e64f1e0e1bcc57fba990025660893d49024096981b2ba3aa\" successfully" Feb 13 19:51:21.169238 containerd[1468]: time="2025-02-13T19:51:21.169230277Z" level=info msg="StopPodSandbox for \"a24dd5ac08e1be65e64f1e0e1bcc57fba990025660893d49024096981b2ba3aa\" returns successfully" Feb 13 19:51:21.169417 systemd[1]: run-netns-cni\x2db56a5a84\x2d86ec\x2d05ed\x2d9e1c\x2d76bd53150342.mount: Deactivated successfully. Feb 13 19:51:21.169514 kubelet[2575]: E0213 19:51:21.169427 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:21.169866 containerd[1468]: time="2025-02-13T19:51:21.169835824Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-6d5rf,Uid:245734af-c06e-49f3-bfef-a966f57e2f98,Namespace:kube-system,Attempt:1,}" Feb 13 19:51:21.170368 kubelet[2575]: I0213 19:51:21.170221 2575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ff7a1ce33e60be86627d63cbaa39e854dfba3ccfcd25d3b18dc72196a2b96da" Feb 13 19:51:21.170677 containerd[1468]: time="2025-02-13T19:51:21.170570624Z" level=info msg="TearDown network for sandbox \"aea5462f72694272b5bc284a0bd9fc173269129df512c011930e89667ad0f97d\" successfully" Feb 13 19:51:21.170677 containerd[1468]: time="2025-02-13T19:51:21.170594830Z" level=info msg="StopPodSandbox for \"aea5462f72694272b5bc284a0bd9fc173269129df512c011930e89667ad0f97d\" returns successfully" Feb 13 19:51:21.170677 containerd[1468]: time="2025-02-13T19:51:21.170602885Z" level=info msg="TearDown network for sandbox \"517d17b8f015a9521e2907d73d8ffe6db808ece1034c7f320e60f66e57f87de3\" successfully" Feb 13 19:51:21.170677 containerd[1468]: time="2025-02-13T19:51:21.170634976Z" level=info msg="StopPodSandbox for \"517d17b8f015a9521e2907d73d8ffe6db808ece1034c7f320e60f66e57f87de3\" returns successfully" Feb 13 19:51:21.170677 containerd[1468]: time="2025-02-13T19:51:21.170574632Z" level=info msg="TearDown network for sandbox \"c143929d3d37d0fd39eeb807647ccdb8bf35b3017d21a93a728ef6fa513987e4\" successfully" Feb 13 19:51:21.170677 containerd[1468]: time="2025-02-13T19:51:21.170654883Z" level=info msg="StopPodSandbox for \"c143929d3d37d0fd39eeb807647ccdb8bf35b3017d21a93a728ef6fa513987e4\" returns successfully" Feb 13 19:51:21.170930 kubelet[2575]: E0213 19:51:21.170912 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:21.171114 containerd[1468]: time="2025-02-13T19:51:21.171076054Z" level=info msg="StopPodSandbox for \"4ff7a1ce33e60be86627d63cbaa39e854dfba3ccfcd25d3b18dc72196a2b96da\"" Feb 13 19:51:21.171269 containerd[1468]: time="2025-02-13T19:51:21.171248799Z" level=info msg="Ensure that sandbox 4ff7a1ce33e60be86627d63cbaa39e854dfba3ccfcd25d3b18dc72196a2b96da in task-service has been cleanup successfully" Feb 13 19:51:21.171865 containerd[1468]: time="2025-02-13T19:51:21.171547820Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-575f8f688d-8855n,Uid:a5e91fd3-19fa-490e-be13-26d807029222,Namespace:calico-apiserver,Attempt:1,}" Feb 13 19:51:21.171865 containerd[1468]: time="2025-02-13T19:51:21.171578478Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5bf7d587cb-s28ls,Uid:77a9c264-ba41-4ec0-b704-771924c99471,Namespace:calico-system,Attempt:1,}" Feb 13 19:51:21.171865 containerd[1468]: time="2025-02-13T19:51:21.171752955Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-s22xp,Uid:bd8b42bf-05e5-489c-87c7-f5984836cd6a,Namespace:kube-system,Attempt:1,}" Feb 13 19:51:21.172048 containerd[1468]: time="2025-02-13T19:51:21.172023103Z" level=info msg="TearDown network for sandbox \"4ff7a1ce33e60be86627d63cbaa39e854dfba3ccfcd25d3b18dc72196a2b96da\" successfully" Feb 13 19:51:21.172048 containerd[1468]: time="2025-02-13T19:51:21.172039924Z" level=info msg="StopPodSandbox for \"4ff7a1ce33e60be86627d63cbaa39e854dfba3ccfcd25d3b18dc72196a2b96da\" returns successfully" Feb 13 19:51:21.172571 containerd[1468]: time="2025-02-13T19:51:21.172532420Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-575f8f688d-rm27z,Uid:dca83c7e-397c-4cb7-b95d-96c1c013b67f,Namespace:calico-apiserver,Attempt:1,}" Feb 13 19:51:21.172980 systemd[1]: run-netns-cni\x2da470442f\x2d0061\x2d6722\x2d2590\x2df78951a84d64.mount: Deactivated successfully. Feb 13 19:51:21.173109 systemd[1]: run-netns-cni\x2dd2c67f7a\x2d91ff\x2d266f\x2ddb63\x2dcab81e17017a.mount: Deactivated successfully. Feb 13 19:51:21.176462 systemd[1]: run-netns-cni\x2d0e683abf\x2d57a1\x2d352b\x2d97ab\x2d20699793b370.mount: Deactivated successfully. Feb 13 19:51:21.417315 containerd[1468]: time="2025-02-13T19:51:21.417149059Z" level=error msg="Failed to destroy network for sandbox \"899fcc18f7bfce71348bcd81eba1c4cca6504cfc06183532c64985ae4deedeb3\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:21.418653 containerd[1468]: time="2025-02-13T19:51:21.418587802Z" level=error msg="encountered an error cleaning up failed sandbox \"899fcc18f7bfce71348bcd81eba1c4cca6504cfc06183532c64985ae4deedeb3\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:21.418987 containerd[1468]: time="2025-02-13T19:51:21.418966012Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-6d5rf,Uid:245734af-c06e-49f3-bfef-a966f57e2f98,Namespace:kube-system,Attempt:1,} failed, error" error="failed to setup network for sandbox \"899fcc18f7bfce71348bcd81eba1c4cca6504cfc06183532c64985ae4deedeb3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:21.420023 kubelet[2575]: E0213 19:51:21.419564 2575 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"899fcc18f7bfce71348bcd81eba1c4cca6504cfc06183532c64985ae4deedeb3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:21.420023 kubelet[2575]: E0213 19:51:21.419656 2575 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"899fcc18f7bfce71348bcd81eba1c4cca6504cfc06183532c64985ae4deedeb3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-6d5rf" Feb 13 19:51:21.420023 kubelet[2575]: E0213 19:51:21.419701 2575 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"899fcc18f7bfce71348bcd81eba1c4cca6504cfc06183532c64985ae4deedeb3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-6d5rf" Feb 13 19:51:21.420217 kubelet[2575]: E0213 19:51:21.419809 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-6d5rf_kube-system(245734af-c06e-49f3-bfef-a966f57e2f98)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-6d5rf_kube-system(245734af-c06e-49f3-bfef-a966f57e2f98)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"899fcc18f7bfce71348bcd81eba1c4cca6504cfc06183532c64985ae4deedeb3\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-6d5rf" podUID="245734af-c06e-49f3-bfef-a966f57e2f98" Feb 13 19:51:21.420721 containerd[1468]: time="2025-02-13T19:51:21.420573391Z" level=error msg="Failed to destroy network for sandbox \"3af9c99589cb59a86666f237cc1b2ea19ad1577ee4287478e22ace117f62bf47\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:21.421094 containerd[1468]: time="2025-02-13T19:51:21.421069773Z" level=error msg="encountered an error cleaning up failed sandbox \"3af9c99589cb59a86666f237cc1b2ea19ad1577ee4287478e22ace117f62bf47\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:21.421217 containerd[1468]: time="2025-02-13T19:51:21.421195230Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-575f8f688d-rm27z,Uid:dca83c7e-397c-4cb7-b95d-96c1c013b67f,Namespace:calico-apiserver,Attempt:1,} failed, error" error="failed to setup network for sandbox \"3af9c99589cb59a86666f237cc1b2ea19ad1577ee4287478e22ace117f62bf47\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:21.421710 kubelet[2575]: E0213 19:51:21.421523 2575 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3af9c99589cb59a86666f237cc1b2ea19ad1577ee4287478e22ace117f62bf47\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:21.421710 kubelet[2575]: E0213 19:51:21.421553 2575 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3af9c99589cb59a86666f237cc1b2ea19ad1577ee4287478e22ace117f62bf47\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-575f8f688d-rm27z" Feb 13 19:51:21.421710 kubelet[2575]: E0213 19:51:21.421572 2575 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3af9c99589cb59a86666f237cc1b2ea19ad1577ee4287478e22ace117f62bf47\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-575f8f688d-rm27z" Feb 13 19:51:21.421803 kubelet[2575]: E0213 19:51:21.421682 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-575f8f688d-rm27z_calico-apiserver(dca83c7e-397c-4cb7-b95d-96c1c013b67f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-575f8f688d-rm27z_calico-apiserver(dca83c7e-397c-4cb7-b95d-96c1c013b67f)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"3af9c99589cb59a86666f237cc1b2ea19ad1577ee4287478e22ace117f62bf47\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-575f8f688d-rm27z" podUID="dca83c7e-397c-4cb7-b95d-96c1c013b67f" Feb 13 19:51:21.424940 containerd[1468]: time="2025-02-13T19:51:21.424282759Z" level=error msg="Failed to destroy network for sandbox \"cc0fab9693293af68d0473a6799ea1e188b038744dcf56927b1d4fffd60c2c96\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:21.426018 containerd[1468]: time="2025-02-13T19:51:21.425987922Z" level=error msg="encountered an error cleaning up failed sandbox \"cc0fab9693293af68d0473a6799ea1e188b038744dcf56927b1d4fffd60c2c96\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:21.426313 containerd[1468]: time="2025-02-13T19:51:21.426064415Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5bf7d587cb-s28ls,Uid:77a9c264-ba41-4ec0-b704-771924c99471,Namespace:calico-system,Attempt:1,} failed, error" error="failed to setup network for sandbox \"cc0fab9693293af68d0473a6799ea1e188b038744dcf56927b1d4fffd60c2c96\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:21.426453 containerd[1468]: time="2025-02-13T19:51:21.426255785Z" level=error msg="Failed to destroy network for sandbox \"38f72aa5f0a5062621013b365f7d00faaabff6dd65a3f268813b6c2c1534a463\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:21.426784 kubelet[2575]: E0213 19:51:21.426727 2575 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"cc0fab9693293af68d0473a6799ea1e188b038744dcf56927b1d4fffd60c2c96\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:21.426939 kubelet[2575]: E0213 19:51:21.426794 2575 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"cc0fab9693293af68d0473a6799ea1e188b038744dcf56927b1d4fffd60c2c96\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5bf7d587cb-s28ls" Feb 13 19:51:21.426939 kubelet[2575]: E0213 19:51:21.426812 2575 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"cc0fab9693293af68d0473a6799ea1e188b038744dcf56927b1d4fffd60c2c96\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5bf7d587cb-s28ls" Feb 13 19:51:21.427107 kubelet[2575]: E0213 19:51:21.426999 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-5bf7d587cb-s28ls_calico-system(77a9c264-ba41-4ec0-b704-771924c99471)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-5bf7d587cb-s28ls_calico-system(77a9c264-ba41-4ec0-b704-771924c99471)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"cc0fab9693293af68d0473a6799ea1e188b038744dcf56927b1d4fffd60c2c96\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-5bf7d587cb-s28ls" podUID="77a9c264-ba41-4ec0-b704-771924c99471" Feb 13 19:51:21.427475 containerd[1468]: time="2025-02-13T19:51:21.427433758Z" level=error msg="encountered an error cleaning up failed sandbox \"38f72aa5f0a5062621013b365f7d00faaabff6dd65a3f268813b6c2c1534a463\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:21.427676 containerd[1468]: time="2025-02-13T19:51:21.427497748Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-575f8f688d-8855n,Uid:a5e91fd3-19fa-490e-be13-26d807029222,Namespace:calico-apiserver,Attempt:1,} failed, error" error="failed to setup network for sandbox \"38f72aa5f0a5062621013b365f7d00faaabff6dd65a3f268813b6c2c1534a463\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:21.428023 kubelet[2575]: E0213 19:51:21.427986 2575 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"38f72aa5f0a5062621013b365f7d00faaabff6dd65a3f268813b6c2c1534a463\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:21.428085 kubelet[2575]: E0213 19:51:21.428052 2575 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"38f72aa5f0a5062621013b365f7d00faaabff6dd65a3f268813b6c2c1534a463\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-575f8f688d-8855n" Feb 13 19:51:21.428118 kubelet[2575]: E0213 19:51:21.428081 2575 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"38f72aa5f0a5062621013b365f7d00faaabff6dd65a3f268813b6c2c1534a463\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-575f8f688d-8855n" Feb 13 19:51:21.428162 kubelet[2575]: E0213 19:51:21.428134 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-575f8f688d-8855n_calico-apiserver(a5e91fd3-19fa-490e-be13-26d807029222)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-575f8f688d-8855n_calico-apiserver(a5e91fd3-19fa-490e-be13-26d807029222)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"38f72aa5f0a5062621013b365f7d00faaabff6dd65a3f268813b6c2c1534a463\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-575f8f688d-8855n" podUID="a5e91fd3-19fa-490e-be13-26d807029222" Feb 13 19:51:21.442518 containerd[1468]: time="2025-02-13T19:51:21.442453352Z" level=error msg="Failed to destroy network for sandbox \"d7392cd735f64bbb484b37d152068bd95b9e9585d02b78303932e81c6708295c\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:21.442960 containerd[1468]: time="2025-02-13T19:51:21.442929767Z" level=error msg="encountered an error cleaning up failed sandbox \"d7392cd735f64bbb484b37d152068bd95b9e9585d02b78303932e81c6708295c\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:21.443034 containerd[1468]: time="2025-02-13T19:51:21.443011310Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-s22xp,Uid:bd8b42bf-05e5-489c-87c7-f5984836cd6a,Namespace:kube-system,Attempt:1,} failed, error" error="failed to setup network for sandbox \"d7392cd735f64bbb484b37d152068bd95b9e9585d02b78303932e81c6708295c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:21.443333 kubelet[2575]: E0213 19:51:21.443282 2575 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d7392cd735f64bbb484b37d152068bd95b9e9585d02b78303932e81c6708295c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:21.443410 kubelet[2575]: E0213 19:51:21.443369 2575 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d7392cd735f64bbb484b37d152068bd95b9e9585d02b78303932e81c6708295c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-s22xp" Feb 13 19:51:21.443443 kubelet[2575]: E0213 19:51:21.443398 2575 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d7392cd735f64bbb484b37d152068bd95b9e9585d02b78303932e81c6708295c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-s22xp" Feb 13 19:51:21.443525 kubelet[2575]: E0213 19:51:21.443484 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-s22xp_kube-system(bd8b42bf-05e5-489c-87c7-f5984836cd6a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-s22xp_kube-system(bd8b42bf-05e5-489c-87c7-f5984836cd6a)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"d7392cd735f64bbb484b37d152068bd95b9e9585d02b78303932e81c6708295c\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-s22xp" podUID="bd8b42bf-05e5-489c-87c7-f5984836cd6a" Feb 13 19:51:21.635394 systemd[1]: Created slice kubepods-besteffort-podbbd75eeb_e06b_4812_b2c6_522804c5ef2f.slice - libcontainer container kubepods-besteffort-podbbd75eeb_e06b_4812_b2c6_522804c5ef2f.slice. Feb 13 19:51:21.637779 containerd[1468]: time="2025-02-13T19:51:21.637739856Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-sxwqd,Uid:bbd75eeb-e06b-4812-b2c6-522804c5ef2f,Namespace:calico-system,Attempt:0,}" Feb 13 19:51:21.700145 containerd[1468]: time="2025-02-13T19:51:21.700013494Z" level=error msg="Failed to destroy network for sandbox \"dee4103e3ed5a58df193e21973191414daa33c176df2ebb53f48f6427c812139\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:21.700482 containerd[1468]: time="2025-02-13T19:51:21.700445486Z" level=error msg="encountered an error cleaning up failed sandbox \"dee4103e3ed5a58df193e21973191414daa33c176df2ebb53f48f6427c812139\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:21.700528 containerd[1468]: time="2025-02-13T19:51:21.700510278Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-sxwqd,Uid:bbd75eeb-e06b-4812-b2c6-522804c5ef2f,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"dee4103e3ed5a58df193e21973191414daa33c176df2ebb53f48f6427c812139\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:21.700887 kubelet[2575]: E0213 19:51:21.700830 2575 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"dee4103e3ed5a58df193e21973191414daa33c176df2ebb53f48f6427c812139\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:21.700969 kubelet[2575]: E0213 19:51:21.700942 2575 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"dee4103e3ed5a58df193e21973191414daa33c176df2ebb53f48f6427c812139\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-sxwqd" Feb 13 19:51:21.700997 kubelet[2575]: E0213 19:51:21.700980 2575 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"dee4103e3ed5a58df193e21973191414daa33c176df2ebb53f48f6427c812139\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-sxwqd" Feb 13 19:51:21.701115 kubelet[2575]: E0213 19:51:21.701039 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-sxwqd_calico-system(bbd75eeb-e06b-4812-b2c6-522804c5ef2f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-sxwqd_calico-system(bbd75eeb-e06b-4812-b2c6-522804c5ef2f)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"dee4103e3ed5a58df193e21973191414daa33c176df2ebb53f48f6427c812139\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-sxwqd" podUID="bbd75eeb-e06b-4812-b2c6-522804c5ef2f" Feb 13 19:51:21.750628 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-899fcc18f7bfce71348bcd81eba1c4cca6504cfc06183532c64985ae4deedeb3-shm.mount: Deactivated successfully. Feb 13 19:51:22.173961 kubelet[2575]: I0213 19:51:22.173921 2575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc0fab9693293af68d0473a6799ea1e188b038744dcf56927b1d4fffd60c2c96" Feb 13 19:51:22.174519 containerd[1468]: time="2025-02-13T19:51:22.174418451Z" level=info msg="StopPodSandbox for \"cc0fab9693293af68d0473a6799ea1e188b038744dcf56927b1d4fffd60c2c96\"" Feb 13 19:51:22.174860 containerd[1468]: time="2025-02-13T19:51:22.174652301Z" level=info msg="Ensure that sandbox cc0fab9693293af68d0473a6799ea1e188b038744dcf56927b1d4fffd60c2c96 in task-service has been cleanup successfully" Feb 13 19:51:22.177440 systemd[1]: run-netns-cni\x2d3cdafbea\x2d437b\x2dd20c\x2dd133\x2db2ee1d0b542c.mount: Deactivated successfully. Feb 13 19:51:22.178482 containerd[1468]: time="2025-02-13T19:51:22.178216434Z" level=info msg="TearDown network for sandbox \"cc0fab9693293af68d0473a6799ea1e188b038744dcf56927b1d4fffd60c2c96\" successfully" Feb 13 19:51:22.178482 containerd[1468]: time="2025-02-13T19:51:22.178235530Z" level=info msg="StopPodSandbox for \"cc0fab9693293af68d0473a6799ea1e188b038744dcf56927b1d4fffd60c2c96\" returns successfully" Feb 13 19:51:22.178686 kubelet[2575]: I0213 19:51:22.178659 2575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3af9c99589cb59a86666f237cc1b2ea19ad1577ee4287478e22ace117f62bf47" Feb 13 19:51:22.179216 containerd[1468]: time="2025-02-13T19:51:22.179195563Z" level=info msg="StopPodSandbox for \"3af9c99589cb59a86666f237cc1b2ea19ad1577ee4287478e22ace117f62bf47\"" Feb 13 19:51:22.179433 containerd[1468]: time="2025-02-13T19:51:22.179385781Z" level=info msg="Ensure that sandbox 3af9c99589cb59a86666f237cc1b2ea19ad1577ee4287478e22ace117f62bf47 in task-service has been cleanup successfully" Feb 13 19:51:22.181998 containerd[1468]: time="2025-02-13T19:51:22.181971367Z" level=info msg="StopPodSandbox for \"c143929d3d37d0fd39eeb807647ccdb8bf35b3017d21a93a728ef6fa513987e4\"" Feb 13 19:51:22.182115 containerd[1468]: time="2025-02-13T19:51:22.182087526Z" level=info msg="TearDown network for sandbox \"c143929d3d37d0fd39eeb807647ccdb8bf35b3017d21a93a728ef6fa513987e4\" successfully" Feb 13 19:51:22.182115 containerd[1468]: time="2025-02-13T19:51:22.182108415Z" level=info msg="StopPodSandbox for \"c143929d3d37d0fd39eeb807647ccdb8bf35b3017d21a93a728ef6fa513987e4\" returns successfully" Feb 13 19:51:22.182469 containerd[1468]: time="2025-02-13T19:51:22.182380676Z" level=info msg="TearDown network for sandbox \"3af9c99589cb59a86666f237cc1b2ea19ad1577ee4287478e22ace117f62bf47\" successfully" Feb 13 19:51:22.182469 containerd[1468]: time="2025-02-13T19:51:22.182401315Z" level=info msg="StopPodSandbox for \"3af9c99589cb59a86666f237cc1b2ea19ad1577ee4287478e22ace117f62bf47\" returns successfully" Feb 13 19:51:22.182636 containerd[1468]: time="2025-02-13T19:51:22.182588155Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5bf7d587cb-s28ls,Uid:77a9c264-ba41-4ec0-b704-771924c99471,Namespace:calico-system,Attempt:2,}" Feb 13 19:51:22.182982 containerd[1468]: time="2025-02-13T19:51:22.182802819Z" level=info msg="StopPodSandbox for \"4ff7a1ce33e60be86627d63cbaa39e854dfba3ccfcd25d3b18dc72196a2b96da\"" Feb 13 19:51:22.182982 containerd[1468]: time="2025-02-13T19:51:22.182916642Z" level=info msg="TearDown network for sandbox \"4ff7a1ce33e60be86627d63cbaa39e854dfba3ccfcd25d3b18dc72196a2b96da\" successfully" Feb 13 19:51:22.182982 containerd[1468]: time="2025-02-13T19:51:22.182929236Z" level=info msg="StopPodSandbox for \"4ff7a1ce33e60be86627d63cbaa39e854dfba3ccfcd25d3b18dc72196a2b96da\" returns successfully" Feb 13 19:51:22.183241 systemd[1]: run-netns-cni\x2db164afc1\x2dc46e\x2dc18c\x2d7bd0\x2d11c2c5d1198f.mount: Deactivated successfully. Feb 13 19:51:22.183416 containerd[1468]: time="2025-02-13T19:51:22.183353574Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-575f8f688d-rm27z,Uid:dca83c7e-397c-4cb7-b95d-96c1c013b67f,Namespace:calico-apiserver,Attempt:2,}" Feb 13 19:51:22.193907 kubelet[2575]: I0213 19:51:22.193619 2575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38f72aa5f0a5062621013b365f7d00faaabff6dd65a3f268813b6c2c1534a463" Feb 13 19:51:22.194279 containerd[1468]: time="2025-02-13T19:51:22.194177143Z" level=info msg="StopPodSandbox for \"38f72aa5f0a5062621013b365f7d00faaabff6dd65a3f268813b6c2c1534a463\"" Feb 13 19:51:22.194498 containerd[1468]: time="2025-02-13T19:51:22.194404630Z" level=info msg="Ensure that sandbox 38f72aa5f0a5062621013b365f7d00faaabff6dd65a3f268813b6c2c1534a463 in task-service has been cleanup successfully" Feb 13 19:51:22.196755 containerd[1468]: time="2025-02-13T19:51:22.194638358Z" level=info msg="TearDown network for sandbox \"38f72aa5f0a5062621013b365f7d00faaabff6dd65a3f268813b6c2c1534a463\" successfully" Feb 13 19:51:22.196755 containerd[1468]: time="2025-02-13T19:51:22.194655851Z" level=info msg="StopPodSandbox for \"38f72aa5f0a5062621013b365f7d00faaabff6dd65a3f268813b6c2c1534a463\" returns successfully" Feb 13 19:51:22.197315 containerd[1468]: time="2025-02-13T19:51:22.197082429Z" level=info msg="StopPodSandbox for \"517d17b8f015a9521e2907d73d8ffe6db808ece1034c7f320e60f66e57f87de3\"" Feb 13 19:51:22.197315 containerd[1468]: time="2025-02-13T19:51:22.197200612Z" level=info msg="TearDown network for sandbox \"517d17b8f015a9521e2907d73d8ffe6db808ece1034c7f320e60f66e57f87de3\" successfully" Feb 13 19:51:22.197315 containerd[1468]: time="2025-02-13T19:51:22.197213676Z" level=info msg="StopPodSandbox for \"517d17b8f015a9521e2907d73d8ffe6db808ece1034c7f320e60f66e57f87de3\" returns successfully" Feb 13 19:51:22.197454 systemd[1]: run-netns-cni\x2d0f3a33f9\x2dbe8d\x2d4275\x2d4bcb\x2d6f054ef63c3f.mount: Deactivated successfully. Feb 13 19:51:22.199375 containerd[1468]: time="2025-02-13T19:51:22.199354497Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-575f8f688d-8855n,Uid:a5e91fd3-19fa-490e-be13-26d807029222,Namespace:calico-apiserver,Attempt:2,}" Feb 13 19:51:22.199727 kubelet[2575]: I0213 19:51:22.199700 2575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="899fcc18f7bfce71348bcd81eba1c4cca6504cfc06183532c64985ae4deedeb3" Feb 13 19:51:22.200767 containerd[1468]: time="2025-02-13T19:51:22.200137808Z" level=info msg="StopPodSandbox for \"899fcc18f7bfce71348bcd81eba1c4cca6504cfc06183532c64985ae4deedeb3\"" Feb 13 19:51:22.200767 containerd[1468]: time="2025-02-13T19:51:22.200349927Z" level=info msg="Ensure that sandbox 899fcc18f7bfce71348bcd81eba1c4cca6504cfc06183532c64985ae4deedeb3 in task-service has been cleanup successfully" Feb 13 19:51:22.202469 systemd[1]: run-netns-cni\x2d1a46e1fb\x2d9580\x2dffbe\x2dd60b\x2d883b8fd142a7.mount: Deactivated successfully. Feb 13 19:51:22.203316 containerd[1468]: time="2025-02-13T19:51:22.203288245Z" level=info msg="TearDown network for sandbox \"899fcc18f7bfce71348bcd81eba1c4cca6504cfc06183532c64985ae4deedeb3\" successfully" Feb 13 19:51:22.203316 containerd[1468]: time="2025-02-13T19:51:22.203310176Z" level=info msg="StopPodSandbox for \"899fcc18f7bfce71348bcd81eba1c4cca6504cfc06183532c64985ae4deedeb3\" returns successfully" Feb 13 19:51:22.203966 containerd[1468]: time="2025-02-13T19:51:22.203941934Z" level=info msg="StopPodSandbox for \"a24dd5ac08e1be65e64f1e0e1bcc57fba990025660893d49024096981b2ba3aa\"" Feb 13 19:51:22.204065 containerd[1468]: time="2025-02-13T19:51:22.204047381Z" level=info msg="TearDown network for sandbox \"a24dd5ac08e1be65e64f1e0e1bcc57fba990025660893d49024096981b2ba3aa\" successfully" Feb 13 19:51:22.204065 containerd[1468]: time="2025-02-13T19:51:22.204063121Z" level=info msg="StopPodSandbox for \"a24dd5ac08e1be65e64f1e0e1bcc57fba990025660893d49024096981b2ba3aa\" returns successfully" Feb 13 19:51:22.204577 kubelet[2575]: E0213 19:51:22.204556 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:22.204961 containerd[1468]: time="2025-02-13T19:51:22.204937563Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-6d5rf,Uid:245734af-c06e-49f3-bfef-a966f57e2f98,Namespace:kube-system,Attempt:2,}" Feb 13 19:51:22.205713 kubelet[2575]: I0213 19:51:22.205681 2575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7392cd735f64bbb484b37d152068bd95b9e9585d02b78303932e81c6708295c" Feb 13 19:51:22.206233 containerd[1468]: time="2025-02-13T19:51:22.206203080Z" level=info msg="StopPodSandbox for \"d7392cd735f64bbb484b37d152068bd95b9e9585d02b78303932e81c6708295c\"" Feb 13 19:51:22.206667 containerd[1468]: time="2025-02-13T19:51:22.206637906Z" level=info msg="Ensure that sandbox d7392cd735f64bbb484b37d152068bd95b9e9585d02b78303932e81c6708295c in task-service has been cleanup successfully" Feb 13 19:51:22.206880 containerd[1468]: time="2025-02-13T19:51:22.206846579Z" level=info msg="TearDown network for sandbox \"d7392cd735f64bbb484b37d152068bd95b9e9585d02b78303932e81c6708295c\" successfully" Feb 13 19:51:22.206918 containerd[1468]: time="2025-02-13T19:51:22.206863430Z" level=info msg="StopPodSandbox for \"d7392cd735f64bbb484b37d152068bd95b9e9585d02b78303932e81c6708295c\" returns successfully" Feb 13 19:51:22.207407 kubelet[2575]: I0213 19:51:22.207390 2575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dee4103e3ed5a58df193e21973191414daa33c176df2ebb53f48f6427c812139" Feb 13 19:51:22.207511 containerd[1468]: time="2025-02-13T19:51:22.207475580Z" level=info msg="StopPodSandbox for \"aea5462f72694272b5bc284a0bd9fc173269129df512c011930e89667ad0f97d\"" Feb 13 19:51:22.207580 containerd[1468]: time="2025-02-13T19:51:22.207562443Z" level=info msg="TearDown network for sandbox \"aea5462f72694272b5bc284a0bd9fc173269129df512c011930e89667ad0f97d\" successfully" Feb 13 19:51:22.207633 containerd[1468]: time="2025-02-13T19:51:22.207577942Z" level=info msg="StopPodSandbox for \"aea5462f72694272b5bc284a0bd9fc173269129df512c011930e89667ad0f97d\" returns successfully" Feb 13 19:51:22.207988 containerd[1468]: time="2025-02-13T19:51:22.207892643Z" level=info msg="StopPodSandbox for \"dee4103e3ed5a58df193e21973191414daa33c176df2ebb53f48f6427c812139\"" Feb 13 19:51:22.208078 containerd[1468]: time="2025-02-13T19:51:22.208041082Z" level=info msg="Ensure that sandbox dee4103e3ed5a58df193e21973191414daa33c176df2ebb53f48f6427c812139 in task-service has been cleanup successfully" Feb 13 19:51:22.208246 kubelet[2575]: E0213 19:51:22.208222 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:22.208579 containerd[1468]: time="2025-02-13T19:51:22.208472773Z" level=info msg="TearDown network for sandbox \"dee4103e3ed5a58df193e21973191414daa33c176df2ebb53f48f6427c812139\" successfully" Feb 13 19:51:22.208579 containerd[1468]: time="2025-02-13T19:51:22.208531323Z" level=info msg="StopPodSandbox for \"dee4103e3ed5a58df193e21973191414daa33c176df2ebb53f48f6427c812139\" returns successfully" Feb 13 19:51:22.208723 containerd[1468]: time="2025-02-13T19:51:22.208499062Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-s22xp,Uid:bd8b42bf-05e5-489c-87c7-f5984836cd6a,Namespace:kube-system,Attempt:2,}" Feb 13 19:51:22.210045 containerd[1468]: time="2025-02-13T19:51:22.210016723Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-sxwqd,Uid:bbd75eeb-e06b-4812-b2c6-522804c5ef2f,Namespace:calico-system,Attempt:1,}" Feb 13 19:51:22.650147 containerd[1468]: time="2025-02-13T19:51:22.650083410Z" level=error msg="Failed to destroy network for sandbox \"5f114ecd2418a98f886f02a3a5637918502bf71cd1bf3c74f6c1dd0e45034e21\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:22.650577 containerd[1468]: time="2025-02-13T19:51:22.650542933Z" level=error msg="encountered an error cleaning up failed sandbox \"5f114ecd2418a98f886f02a3a5637918502bf71cd1bf3c74f6c1dd0e45034e21\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:22.650669 containerd[1468]: time="2025-02-13T19:51:22.650626549Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-575f8f688d-rm27z,Uid:dca83c7e-397c-4cb7-b95d-96c1c013b67f,Namespace:calico-apiserver,Attempt:2,} failed, error" error="failed to setup network for sandbox \"5f114ecd2418a98f886f02a3a5637918502bf71cd1bf3c74f6c1dd0e45034e21\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:22.650993 kubelet[2575]: E0213 19:51:22.650936 2575 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5f114ecd2418a98f886f02a3a5637918502bf71cd1bf3c74f6c1dd0e45034e21\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:22.651068 kubelet[2575]: E0213 19:51:22.651025 2575 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5f114ecd2418a98f886f02a3a5637918502bf71cd1bf3c74f6c1dd0e45034e21\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-575f8f688d-rm27z" Feb 13 19:51:22.651068 kubelet[2575]: E0213 19:51:22.651058 2575 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5f114ecd2418a98f886f02a3a5637918502bf71cd1bf3c74f6c1dd0e45034e21\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-575f8f688d-rm27z" Feb 13 19:51:22.651177 kubelet[2575]: E0213 19:51:22.651139 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-575f8f688d-rm27z_calico-apiserver(dca83c7e-397c-4cb7-b95d-96c1c013b67f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-575f8f688d-rm27z_calico-apiserver(dca83c7e-397c-4cb7-b95d-96c1c013b67f)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"5f114ecd2418a98f886f02a3a5637918502bf71cd1bf3c74f6c1dd0e45034e21\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-575f8f688d-rm27z" podUID="dca83c7e-397c-4cb7-b95d-96c1c013b67f" Feb 13 19:51:22.653461 containerd[1468]: time="2025-02-13T19:51:22.653416200Z" level=error msg="Failed to destroy network for sandbox \"deadb290c2eabe3a6b5ed648ceb2f5121ab56b3728b197df70e6532dc828cd25\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:22.653995 containerd[1468]: time="2025-02-13T19:51:22.653965160Z" level=error msg="encountered an error cleaning up failed sandbox \"deadb290c2eabe3a6b5ed648ceb2f5121ab56b3728b197df70e6532dc828cd25\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:22.654066 containerd[1468]: time="2025-02-13T19:51:22.654037657Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5bf7d587cb-s28ls,Uid:77a9c264-ba41-4ec0-b704-771924c99471,Namespace:calico-system,Attempt:2,} failed, error" error="failed to setup network for sandbox \"deadb290c2eabe3a6b5ed648ceb2f5121ab56b3728b197df70e6532dc828cd25\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:22.654297 kubelet[2575]: E0213 19:51:22.654271 2575 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"deadb290c2eabe3a6b5ed648ceb2f5121ab56b3728b197df70e6532dc828cd25\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:22.654373 kubelet[2575]: E0213 19:51:22.654314 2575 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"deadb290c2eabe3a6b5ed648ceb2f5121ab56b3728b197df70e6532dc828cd25\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5bf7d587cb-s28ls" Feb 13 19:51:22.654373 kubelet[2575]: E0213 19:51:22.654337 2575 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"deadb290c2eabe3a6b5ed648ceb2f5121ab56b3728b197df70e6532dc828cd25\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5bf7d587cb-s28ls" Feb 13 19:51:22.654445 kubelet[2575]: E0213 19:51:22.654386 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-5bf7d587cb-s28ls_calico-system(77a9c264-ba41-4ec0-b704-771924c99471)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-5bf7d587cb-s28ls_calico-system(77a9c264-ba41-4ec0-b704-771924c99471)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"deadb290c2eabe3a6b5ed648ceb2f5121ab56b3728b197df70e6532dc828cd25\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-5bf7d587cb-s28ls" podUID="77a9c264-ba41-4ec0-b704-771924c99471" Feb 13 19:51:22.719610 containerd[1468]: time="2025-02-13T19:51:22.719533093Z" level=error msg="Failed to destroy network for sandbox \"13004b615f48f8fc71da84226292b2e4c272d980ea5cce23fb82164b712c2e30\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:22.723787 containerd[1468]: time="2025-02-13T19:51:22.720092232Z" level=error msg="encountered an error cleaning up failed sandbox \"13004b615f48f8fc71da84226292b2e4c272d980ea5cce23fb82164b712c2e30\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:22.723787 containerd[1468]: time="2025-02-13T19:51:22.720169939Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-6d5rf,Uid:245734af-c06e-49f3-bfef-a966f57e2f98,Namespace:kube-system,Attempt:2,} failed, error" error="failed to setup network for sandbox \"13004b615f48f8fc71da84226292b2e4c272d980ea5cce23fb82164b712c2e30\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:22.732553 kubelet[2575]: E0213 19:51:22.720464 2575 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"13004b615f48f8fc71da84226292b2e4c272d980ea5cce23fb82164b712c2e30\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:22.732553 kubelet[2575]: E0213 19:51:22.720604 2575 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"13004b615f48f8fc71da84226292b2e4c272d980ea5cce23fb82164b712c2e30\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-6d5rf" Feb 13 19:51:22.732553 kubelet[2575]: E0213 19:51:22.720642 2575 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"13004b615f48f8fc71da84226292b2e4c272d980ea5cce23fb82164b712c2e30\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-6d5rf" Feb 13 19:51:22.732729 kubelet[2575]: E0213 19:51:22.720699 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-6d5rf_kube-system(245734af-c06e-49f3-bfef-a966f57e2f98)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-6d5rf_kube-system(245734af-c06e-49f3-bfef-a966f57e2f98)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"13004b615f48f8fc71da84226292b2e4c272d980ea5cce23fb82164b712c2e30\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-6d5rf" podUID="245734af-c06e-49f3-bfef-a966f57e2f98" Feb 13 19:51:22.752768 systemd[1]: run-netns-cni\x2d3f1e8b74\x2df294\x2d8fad\x2d35fd\x2d1f45a9ad5fd9.mount: Deactivated successfully. Feb 13 19:51:22.752891 systemd[1]: run-netns-cni\x2d99f7ac3b\x2d6039\x2d96c2\x2d7879\x2d9a081828e10a.mount: Deactivated successfully. Feb 13 19:51:22.783729 containerd[1468]: time="2025-02-13T19:51:22.783675206Z" level=error msg="Failed to destroy network for sandbox \"86a985525093286f84ab532b093323eb20cae87c984957fc89def62353843bd5\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:22.784323 containerd[1468]: time="2025-02-13T19:51:22.784131493Z" level=error msg="encountered an error cleaning up failed sandbox \"86a985525093286f84ab532b093323eb20cae87c984957fc89def62353843bd5\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:22.784323 containerd[1468]: time="2025-02-13T19:51:22.784200974Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-s22xp,Uid:bd8b42bf-05e5-489c-87c7-f5984836cd6a,Namespace:kube-system,Attempt:2,} failed, error" error="failed to setup network for sandbox \"86a985525093286f84ab532b093323eb20cae87c984957fc89def62353843bd5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:22.784703 kubelet[2575]: E0213 19:51:22.784648 2575 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"86a985525093286f84ab532b093323eb20cae87c984957fc89def62353843bd5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:22.784759 kubelet[2575]: E0213 19:51:22.784725 2575 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"86a985525093286f84ab532b093323eb20cae87c984957fc89def62353843bd5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-s22xp" Feb 13 19:51:22.784759 kubelet[2575]: E0213 19:51:22.784750 2575 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"86a985525093286f84ab532b093323eb20cae87c984957fc89def62353843bd5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-s22xp" Feb 13 19:51:22.784821 kubelet[2575]: E0213 19:51:22.784803 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-s22xp_kube-system(bd8b42bf-05e5-489c-87c7-f5984836cd6a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-s22xp_kube-system(bd8b42bf-05e5-489c-87c7-f5984836cd6a)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"86a985525093286f84ab532b093323eb20cae87c984957fc89def62353843bd5\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-s22xp" podUID="bd8b42bf-05e5-489c-87c7-f5984836cd6a" Feb 13 19:51:22.787194 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-86a985525093286f84ab532b093323eb20cae87c984957fc89def62353843bd5-shm.mount: Deactivated successfully. Feb 13 19:51:22.967458 containerd[1468]: time="2025-02-13T19:51:22.967390887Z" level=error msg="Failed to destroy network for sandbox \"49835f65aea5736de7af75d5477e71209d3ab4edb28dc78ae778b60bcf5f2e76\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:22.969445 containerd[1468]: time="2025-02-13T19:51:22.969380805Z" level=error msg="encountered an error cleaning up failed sandbox \"49835f65aea5736de7af75d5477e71209d3ab4edb28dc78ae778b60bcf5f2e76\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:22.969560 containerd[1468]: time="2025-02-13T19:51:22.969468008Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-sxwqd,Uid:bbd75eeb-e06b-4812-b2c6-522804c5ef2f,Namespace:calico-system,Attempt:1,} failed, error" error="failed to setup network for sandbox \"49835f65aea5736de7af75d5477e71209d3ab4edb28dc78ae778b60bcf5f2e76\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:22.969847 kubelet[2575]: E0213 19:51:22.969802 2575 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"49835f65aea5736de7af75d5477e71209d3ab4edb28dc78ae778b60bcf5f2e76\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:22.969934 kubelet[2575]: E0213 19:51:22.969914 2575 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"49835f65aea5736de7af75d5477e71209d3ab4edb28dc78ae778b60bcf5f2e76\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-sxwqd" Feb 13 19:51:22.969971 kubelet[2575]: E0213 19:51:22.969942 2575 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"49835f65aea5736de7af75d5477e71209d3ab4edb28dc78ae778b60bcf5f2e76\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-sxwqd" Feb 13 19:51:22.970012 kubelet[2575]: E0213 19:51:22.969982 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-sxwqd_calico-system(bbd75eeb-e06b-4812-b2c6-522804c5ef2f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-sxwqd_calico-system(bbd75eeb-e06b-4812-b2c6-522804c5ef2f)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"49835f65aea5736de7af75d5477e71209d3ab4edb28dc78ae778b60bcf5f2e76\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-sxwqd" podUID="bbd75eeb-e06b-4812-b2c6-522804c5ef2f" Feb 13 19:51:22.970058 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-49835f65aea5736de7af75d5477e71209d3ab4edb28dc78ae778b60bcf5f2e76-shm.mount: Deactivated successfully. Feb 13 19:51:22.990783 containerd[1468]: time="2025-02-13T19:51:22.990721297Z" level=error msg="Failed to destroy network for sandbox \"269c22f856e2916019bfab8db86934d7f53640411cbfd7537abae3e996450f7b\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:22.991201 containerd[1468]: time="2025-02-13T19:51:22.991166643Z" level=error msg="encountered an error cleaning up failed sandbox \"269c22f856e2916019bfab8db86934d7f53640411cbfd7537abae3e996450f7b\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:22.991263 containerd[1468]: time="2025-02-13T19:51:22.991234771Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-575f8f688d-8855n,Uid:a5e91fd3-19fa-490e-be13-26d807029222,Namespace:calico-apiserver,Attempt:2,} failed, error" error="failed to setup network for sandbox \"269c22f856e2916019bfab8db86934d7f53640411cbfd7537abae3e996450f7b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:22.991586 kubelet[2575]: E0213 19:51:22.991531 2575 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"269c22f856e2916019bfab8db86934d7f53640411cbfd7537abae3e996450f7b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:22.991652 kubelet[2575]: E0213 19:51:22.991609 2575 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"269c22f856e2916019bfab8db86934d7f53640411cbfd7537abae3e996450f7b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-575f8f688d-8855n" Feb 13 19:51:22.991652 kubelet[2575]: E0213 19:51:22.991633 2575 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"269c22f856e2916019bfab8db86934d7f53640411cbfd7537abae3e996450f7b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-575f8f688d-8855n" Feb 13 19:51:22.991704 kubelet[2575]: E0213 19:51:22.991679 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-575f8f688d-8855n_calico-apiserver(a5e91fd3-19fa-490e-be13-26d807029222)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-575f8f688d-8855n_calico-apiserver(a5e91fd3-19fa-490e-be13-26d807029222)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"269c22f856e2916019bfab8db86934d7f53640411cbfd7537abae3e996450f7b\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-575f8f688d-8855n" podUID="a5e91fd3-19fa-490e-be13-26d807029222" Feb 13 19:51:23.210752 kubelet[2575]: I0213 19:51:23.210705 2575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="deadb290c2eabe3a6b5ed648ceb2f5121ab56b3728b197df70e6532dc828cd25" Feb 13 19:51:23.211329 containerd[1468]: time="2025-02-13T19:51:23.211287648Z" level=info msg="StopPodSandbox for \"deadb290c2eabe3a6b5ed648ceb2f5121ab56b3728b197df70e6532dc828cd25\"" Feb 13 19:51:23.211805 containerd[1468]: time="2025-02-13T19:51:23.211492303Z" level=info msg="Ensure that sandbox deadb290c2eabe3a6b5ed648ceb2f5121ab56b3728b197df70e6532dc828cd25 in task-service has been cleanup successfully" Feb 13 19:51:23.211805 containerd[1468]: time="2025-02-13T19:51:23.211686317Z" level=info msg="TearDown network for sandbox \"deadb290c2eabe3a6b5ed648ceb2f5121ab56b3728b197df70e6532dc828cd25\" successfully" Feb 13 19:51:23.211805 containerd[1468]: time="2025-02-13T19:51:23.211698149Z" level=info msg="StopPodSandbox for \"deadb290c2eabe3a6b5ed648ceb2f5121ab56b3728b197df70e6532dc828cd25\" returns successfully" Feb 13 19:51:23.212564 containerd[1468]: time="2025-02-13T19:51:23.212537746Z" level=info msg="StopPodSandbox for \"cc0fab9693293af68d0473a6799ea1e188b038744dcf56927b1d4fffd60c2c96\"" Feb 13 19:51:23.212672 containerd[1468]: time="2025-02-13T19:51:23.212648053Z" level=info msg="TearDown network for sandbox \"cc0fab9693293af68d0473a6799ea1e188b038744dcf56927b1d4fffd60c2c96\" successfully" Feb 13 19:51:23.212672 containerd[1468]: time="2025-02-13T19:51:23.212668241Z" level=info msg="StopPodSandbox for \"cc0fab9693293af68d0473a6799ea1e188b038744dcf56927b1d4fffd60c2c96\" returns successfully" Feb 13 19:51:23.212927 containerd[1468]: time="2025-02-13T19:51:23.212901319Z" level=info msg="StopPodSandbox for \"c143929d3d37d0fd39eeb807647ccdb8bf35b3017d21a93a728ef6fa513987e4\"" Feb 13 19:51:23.213648 kubelet[2575]: I0213 19:51:23.213628 2575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f114ecd2418a98f886f02a3a5637918502bf71cd1bf3c74f6c1dd0e45034e21" Feb 13 19:51:23.216019 kubelet[2575]: I0213 19:51:23.215224 2575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="269c22f856e2916019bfab8db86934d7f53640411cbfd7537abae3e996450f7b" Feb 13 19:51:23.216540 kubelet[2575]: I0213 19:51:23.216520 2575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13004b615f48f8fc71da84226292b2e4c272d980ea5cce23fb82164b712c2e30" Feb 13 19:51:23.218310 kubelet[2575]: I0213 19:51:23.218139 2575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="86a985525093286f84ab532b093323eb20cae87c984957fc89def62353843bd5" Feb 13 19:51:23.220024 kubelet[2575]: I0213 19:51:23.219907 2575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49835f65aea5736de7af75d5477e71209d3ab4edb28dc78ae778b60bcf5f2e76" Feb 13 19:51:23.225049 containerd[1468]: time="2025-02-13T19:51:23.212999393Z" level=info msg="TearDown network for sandbox \"c143929d3d37d0fd39eeb807647ccdb8bf35b3017d21a93a728ef6fa513987e4\" successfully" Feb 13 19:51:23.225143 containerd[1468]: time="2025-02-13T19:51:23.225044374Z" level=info msg="StopPodSandbox for \"c143929d3d37d0fd39eeb807647ccdb8bf35b3017d21a93a728ef6fa513987e4\" returns successfully" Feb 13 19:51:23.225143 containerd[1468]: time="2025-02-13T19:51:23.214067159Z" level=info msg="StopPodSandbox for \"5f114ecd2418a98f886f02a3a5637918502bf71cd1bf3c74f6c1dd0e45034e21\"" Feb 13 19:51:23.225211 containerd[1468]: time="2025-02-13T19:51:23.215669237Z" level=info msg="StopPodSandbox for \"269c22f856e2916019bfab8db86934d7f53640411cbfd7537abae3e996450f7b\"" Feb 13 19:51:23.225464 containerd[1468]: time="2025-02-13T19:51:23.216991620Z" level=info msg="StopPodSandbox for \"13004b615f48f8fc71da84226292b2e4c272d980ea5cce23fb82164b712c2e30\"" Feb 13 19:51:23.225464 containerd[1468]: time="2025-02-13T19:51:23.225319291Z" level=info msg="Ensure that sandbox 5f114ecd2418a98f886f02a3a5637918502bf71cd1bf3c74f6c1dd0e45034e21 in task-service has been cleanup successfully" Feb 13 19:51:23.225464 containerd[1468]: time="2025-02-13T19:51:23.225331524Z" level=info msg="Ensure that sandbox 269c22f856e2916019bfab8db86934d7f53640411cbfd7537abae3e996450f7b in task-service has been cleanup successfully" Feb 13 19:51:23.225590 containerd[1468]: time="2025-02-13T19:51:23.225502524Z" level=info msg="TearDown network for sandbox \"269c22f856e2916019bfab8db86934d7f53640411cbfd7537abae3e996450f7b\" successfully" Feb 13 19:51:23.225590 containerd[1468]: time="2025-02-13T19:51:23.225516270Z" level=info msg="StopPodSandbox for \"269c22f856e2916019bfab8db86934d7f53640411cbfd7537abae3e996450f7b\" returns successfully" Feb 13 19:51:23.225590 containerd[1468]: time="2025-02-13T19:51:23.225535827Z" level=info msg="TearDown network for sandbox \"5f114ecd2418a98f886f02a3a5637918502bf71cd1bf3c74f6c1dd0e45034e21\" successfully" Feb 13 19:51:23.225590 containerd[1468]: time="2025-02-13T19:51:23.225551767Z" level=info msg="StopPodSandbox for \"5f114ecd2418a98f886f02a3a5637918502bf71cd1bf3c74f6c1dd0e45034e21\" returns successfully" Feb 13 19:51:23.225756 containerd[1468]: time="2025-02-13T19:51:23.219199798Z" level=info msg="StopPodSandbox for \"86a985525093286f84ab532b093323eb20cae87c984957fc89def62353843bd5\"" Feb 13 19:51:23.225756 containerd[1468]: time="2025-02-13T19:51:23.225705446Z" level=info msg="StopPodSandbox for \"38f72aa5f0a5062621013b365f7d00faaabff6dd65a3f268813b6c2c1534a463\"" Feb 13 19:51:23.225756 containerd[1468]: time="2025-02-13T19:51:23.225745070Z" level=info msg="Ensure that sandbox 86a985525093286f84ab532b093323eb20cae87c984957fc89def62353843bd5 in task-service has been cleanup successfully" Feb 13 19:51:23.225882 containerd[1468]: time="2025-02-13T19:51:23.220260259Z" level=info msg="StopPodSandbox for \"49835f65aea5736de7af75d5477e71209d3ab4edb28dc78ae778b60bcf5f2e76\"" Feb 13 19:51:23.225921 containerd[1468]: time="2025-02-13T19:51:23.225888639Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5bf7d587cb-s28ls,Uid:77a9c264-ba41-4ec0-b704-771924c99471,Namespace:calico-system,Attempt:3,}" Feb 13 19:51:23.225921 containerd[1468]: time="2025-02-13T19:51:23.225906524Z" level=info msg="TearDown network for sandbox \"86a985525093286f84ab532b093323eb20cae87c984957fc89def62353843bd5\" successfully" Feb 13 19:51:23.225921 containerd[1468]: time="2025-02-13T19:51:23.225917945Z" level=info msg="StopPodSandbox for \"86a985525093286f84ab532b093323eb20cae87c984957fc89def62353843bd5\" returns successfully" Feb 13 19:51:23.226016 containerd[1468]: time="2025-02-13T19:51:23.226001041Z" level=info msg="Ensure that sandbox 49835f65aea5736de7af75d5477e71209d3ab4edb28dc78ae778b60bcf5f2e76 in task-service has been cleanup successfully" Feb 13 19:51:23.226051 containerd[1468]: time="2025-02-13T19:51:23.225795765Z" level=info msg="TearDown network for sandbox \"38f72aa5f0a5062621013b365f7d00faaabff6dd65a3f268813b6c2c1534a463\" successfully" Feb 13 19:51:23.226051 containerd[1468]: time="2025-02-13T19:51:23.226027551Z" level=info msg="StopPodSandbox for \"38f72aa5f0a5062621013b365f7d00faaabff6dd65a3f268813b6c2c1534a463\" returns successfully" Feb 13 19:51:23.226204 containerd[1468]: time="2025-02-13T19:51:23.226012362Z" level=info msg="StopPodSandbox for \"3af9c99589cb59a86666f237cc1b2ea19ad1577ee4287478e22ace117f62bf47\"" Feb 13 19:51:23.226289 containerd[1468]: time="2025-02-13T19:51:23.226272391Z" level=info msg="TearDown network for sandbox \"3af9c99589cb59a86666f237cc1b2ea19ad1577ee4287478e22ace117f62bf47\" successfully" Feb 13 19:51:23.226333 containerd[1468]: time="2025-02-13T19:51:23.226285205Z" level=info msg="StopPodSandbox for \"3af9c99589cb59a86666f237cc1b2ea19ad1577ee4287478e22ace117f62bf47\" returns successfully" Feb 13 19:51:23.226333 containerd[1468]: time="2025-02-13T19:51:23.226273403Z" level=info msg="TearDown network for sandbox \"49835f65aea5736de7af75d5477e71209d3ab4edb28dc78ae778b60bcf5f2e76\" successfully" Feb 13 19:51:23.226333 containerd[1468]: time="2025-02-13T19:51:23.226330650Z" level=info msg="StopPodSandbox for \"49835f65aea5736de7af75d5477e71209d3ab4edb28dc78ae778b60bcf5f2e76\" returns successfully" Feb 13 19:51:23.226522 containerd[1468]: time="2025-02-13T19:51:23.226482154Z" level=info msg="StopPodSandbox for \"d7392cd735f64bbb484b37d152068bd95b9e9585d02b78303932e81c6708295c\"" Feb 13 19:51:23.226552 containerd[1468]: time="2025-02-13T19:51:23.226540694Z" level=info msg="StopPodSandbox for \"dee4103e3ed5a58df193e21973191414daa33c176df2ebb53f48f6427c812139\"" Feb 13 19:51:23.226597 containerd[1468]: time="2025-02-13T19:51:23.226581311Z" level=info msg="TearDown network for sandbox \"d7392cd735f64bbb484b37d152068bd95b9e9585d02b78303932e81c6708295c\" successfully" Feb 13 19:51:23.226597 containerd[1468]: time="2025-02-13T19:51:23.226593544Z" level=info msg="StopPodSandbox for \"d7392cd735f64bbb484b37d152068bd95b9e9585d02b78303932e81c6708295c\" returns successfully" Feb 13 19:51:23.226751 containerd[1468]: time="2025-02-13T19:51:23.226639660Z" level=info msg="StopPodSandbox for \"517d17b8f015a9521e2907d73d8ffe6db808ece1034c7f320e60f66e57f87de3\"" Feb 13 19:51:23.226751 containerd[1468]: time="2025-02-13T19:51:23.226715072Z" level=info msg="TearDown network for sandbox \"517d17b8f015a9521e2907d73d8ffe6db808ece1034c7f320e60f66e57f87de3\" successfully" Feb 13 19:51:23.226751 containerd[1468]: time="2025-02-13T19:51:23.226726734Z" level=info msg="StopPodSandbox for \"517d17b8f015a9521e2907d73d8ffe6db808ece1034c7f320e60f66e57f87de3\" returns successfully" Feb 13 19:51:23.226842 containerd[1468]: time="2025-02-13T19:51:23.226804189Z" level=info msg="StopPodSandbox for \"4ff7a1ce33e60be86627d63cbaa39e854dfba3ccfcd25d3b18dc72196a2b96da\"" Feb 13 19:51:23.226926 containerd[1468]: time="2025-02-13T19:51:23.226901182Z" level=info msg="TearDown network for sandbox \"4ff7a1ce33e60be86627d63cbaa39e854dfba3ccfcd25d3b18dc72196a2b96da\" successfully" Feb 13 19:51:23.226926 containerd[1468]: time="2025-02-13T19:51:23.226919276Z" level=info msg="StopPodSandbox for \"4ff7a1ce33e60be86627d63cbaa39e854dfba3ccfcd25d3b18dc72196a2b96da\" returns successfully" Feb 13 19:51:23.227311 containerd[1468]: time="2025-02-13T19:51:23.227285784Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-575f8f688d-rm27z,Uid:dca83c7e-397c-4cb7-b95d-96c1c013b67f,Namespace:calico-apiserver,Attempt:3,}" Feb 13 19:51:23.227511 containerd[1468]: time="2025-02-13T19:51:23.227488284Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-575f8f688d-8855n,Uid:a5e91fd3-19fa-490e-be13-26d807029222,Namespace:calico-apiserver,Attempt:3,}" Feb 13 19:51:23.227688 containerd[1468]: time="2025-02-13T19:51:23.227665968Z" level=info msg="StopPodSandbox for \"aea5462f72694272b5bc284a0bd9fc173269129df512c011930e89667ad0f97d\"" Feb 13 19:51:23.227789 containerd[1468]: time="2025-02-13T19:51:23.227755696Z" level=info msg="TearDown network for sandbox \"aea5462f72694272b5bc284a0bd9fc173269129df512c011930e89667ad0f97d\" successfully" Feb 13 19:51:23.227920 containerd[1468]: time="2025-02-13T19:51:23.227787195Z" level=info msg="StopPodSandbox for \"aea5462f72694272b5bc284a0bd9fc173269129df512c011930e89667ad0f97d\" returns successfully" Feb 13 19:51:23.227970 kubelet[2575]: E0213 19:51:23.227953 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:23.228190 containerd[1468]: time="2025-02-13T19:51:23.228159073Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-s22xp,Uid:bd8b42bf-05e5-489c-87c7-f5984836cd6a,Namespace:kube-system,Attempt:3,}" Feb 13 19:51:23.230711 containerd[1468]: time="2025-02-13T19:51:23.230682122Z" level=info msg="Ensure that sandbox 13004b615f48f8fc71da84226292b2e4c272d980ea5cce23fb82164b712c2e30 in task-service has been cleanup successfully" Feb 13 19:51:23.230885 containerd[1468]: time="2025-02-13T19:51:23.230827434Z" level=info msg="TearDown network for sandbox \"13004b615f48f8fc71da84226292b2e4c272d980ea5cce23fb82164b712c2e30\" successfully" Feb 13 19:51:23.230885 containerd[1468]: time="2025-02-13T19:51:23.230847754Z" level=info msg="StopPodSandbox for \"13004b615f48f8fc71da84226292b2e4c272d980ea5cce23fb82164b712c2e30\" returns successfully" Feb 13 19:51:23.231171 containerd[1468]: time="2025-02-13T19:51:23.231143478Z" level=info msg="StopPodSandbox for \"899fcc18f7bfce71348bcd81eba1c4cca6504cfc06183532c64985ae4deedeb3\"" Feb 13 19:51:23.231251 containerd[1468]: time="2025-02-13T19:51:23.231218890Z" level=info msg="TearDown network for sandbox \"899fcc18f7bfce71348bcd81eba1c4cca6504cfc06183532c64985ae4deedeb3\" successfully" Feb 13 19:51:23.231251 containerd[1468]: time="2025-02-13T19:51:23.231231264Z" level=info msg="StopPodSandbox for \"899fcc18f7bfce71348bcd81eba1c4cca6504cfc06183532c64985ae4deedeb3\" returns successfully" Feb 13 19:51:23.231457 containerd[1468]: time="2025-02-13T19:51:23.231432080Z" level=info msg="StopPodSandbox for \"a24dd5ac08e1be65e64f1e0e1bcc57fba990025660893d49024096981b2ba3aa\"" Feb 13 19:51:23.231547 containerd[1468]: time="2025-02-13T19:51:23.231525546Z" level=info msg="TearDown network for sandbox \"a24dd5ac08e1be65e64f1e0e1bcc57fba990025660893d49024096981b2ba3aa\" successfully" Feb 13 19:51:23.231601 containerd[1468]: time="2025-02-13T19:51:23.231545504Z" level=info msg="StopPodSandbox for \"a24dd5ac08e1be65e64f1e0e1bcc57fba990025660893d49024096981b2ba3aa\" returns successfully" Feb 13 19:51:23.231766 kubelet[2575]: E0213 19:51:23.231749 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:23.232031 containerd[1468]: time="2025-02-13T19:51:23.232008002Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-6d5rf,Uid:245734af-c06e-49f3-bfef-a966f57e2f98,Namespace:kube-system,Attempt:3,}" Feb 13 19:51:23.267219 containerd[1468]: time="2025-02-13T19:51:23.267159177Z" level=info msg="TearDown network for sandbox \"dee4103e3ed5a58df193e21973191414daa33c176df2ebb53f48f6427c812139\" successfully" Feb 13 19:51:23.267219 containerd[1468]: time="2025-02-13T19:51:23.267192890Z" level=info msg="StopPodSandbox for \"dee4103e3ed5a58df193e21973191414daa33c176df2ebb53f48f6427c812139\" returns successfully" Feb 13 19:51:23.267913 containerd[1468]: time="2025-02-13T19:51:23.267865804Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-sxwqd,Uid:bbd75eeb-e06b-4812-b2c6-522804c5ef2f,Namespace:calico-system,Attempt:2,}" Feb 13 19:51:23.748459 systemd[1]: run-netns-cni\x2d19674e07\x2d0e2d\x2d5f0f\x2d342c\x2dcc0a5575cc09.mount: Deactivated successfully. Feb 13 19:51:23.748597 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-269c22f856e2916019bfab8db86934d7f53640411cbfd7537abae3e996450f7b-shm.mount: Deactivated successfully. Feb 13 19:51:23.748682 systemd[1]: run-netns-cni\x2dba49501e\x2dc186\x2d3760\x2d2550\x2dfaf2312ec306.mount: Deactivated successfully. Feb 13 19:51:23.748763 systemd[1]: run-netns-cni\x2d8d917bc2\x2d29c8\x2d42c3\x2dd143\x2d5e1cf397e2c7.mount: Deactivated successfully. Feb 13 19:51:23.748831 systemd[1]: run-netns-cni\x2dfe3997c6\x2dd26a\x2d09e5\x2d35dd\x2d55cdad81043c.mount: Deactivated successfully. Feb 13 19:51:23.748914 systemd[1]: run-netns-cni\x2df3a5140b\x2da56e\x2dd4a8\x2d166c\x2d527b72705868.mount: Deactivated successfully. Feb 13 19:51:23.748982 systemd[1]: run-netns-cni\x2d13603cb5\x2dfdab\x2db3a6\x2d9d4e\x2d22375a6332f8.mount: Deactivated successfully. Feb 13 19:51:25.230007 containerd[1468]: time="2025-02-13T19:51:25.229941878Z" level=error msg="Failed to destroy network for sandbox \"e4d2487cfe7f2e3ccbb95c87ade298917e21e16bc54e8c6d5da7846fe11b6c37\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:25.233619 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-e4d2487cfe7f2e3ccbb95c87ade298917e21e16bc54e8c6d5da7846fe11b6c37-shm.mount: Deactivated successfully. Feb 13 19:51:25.235697 containerd[1468]: time="2025-02-13T19:51:25.235504972Z" level=error msg="Failed to destroy network for sandbox \"3d79c24a7d960cef54edf513bffc4dd8a68cbc23ca3e3304eafaabcbc4d88e33\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:25.250015 containerd[1468]: time="2025-02-13T19:51:25.249965795Z" level=error msg="Failed to destroy network for sandbox \"20a630c64d17851512c12fc278a3718b46879b27d0ffedf611d9175c93614992\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:25.442684 containerd[1468]: time="2025-02-13T19:51:25.442586973Z" level=error msg="encountered an error cleaning up failed sandbox \"e4d2487cfe7f2e3ccbb95c87ade298917e21e16bc54e8c6d5da7846fe11b6c37\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:25.442915 containerd[1468]: time="2025-02-13T19:51:25.442720173Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-s22xp,Uid:bd8b42bf-05e5-489c-87c7-f5984836cd6a,Namespace:kube-system,Attempt:3,} failed, error" error="failed to setup network for sandbox \"e4d2487cfe7f2e3ccbb95c87ade298917e21e16bc54e8c6d5da7846fe11b6c37\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:25.442915 containerd[1468]: time="2025-02-13T19:51:25.442725012Z" level=error msg="encountered an error cleaning up failed sandbox \"3d79c24a7d960cef54edf513bffc4dd8a68cbc23ca3e3304eafaabcbc4d88e33\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:25.443585 kubelet[2575]: E0213 19:51:25.443439 2575 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e4d2487cfe7f2e3ccbb95c87ade298917e21e16bc54e8c6d5da7846fe11b6c37\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:25.443585 kubelet[2575]: E0213 19:51:25.443544 2575 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e4d2487cfe7f2e3ccbb95c87ade298917e21e16bc54e8c6d5da7846fe11b6c37\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-s22xp" Feb 13 19:51:25.443585 kubelet[2575]: E0213 19:51:25.443580 2575 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e4d2487cfe7f2e3ccbb95c87ade298917e21e16bc54e8c6d5da7846fe11b6c37\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-s22xp" Feb 13 19:51:25.444458 kubelet[2575]: E0213 19:51:25.443670 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-s22xp_kube-system(bd8b42bf-05e5-489c-87c7-f5984836cd6a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-s22xp_kube-system(bd8b42bf-05e5-489c-87c7-f5984836cd6a)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"e4d2487cfe7f2e3ccbb95c87ade298917e21e16bc54e8c6d5da7846fe11b6c37\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-s22xp" podUID="bd8b42bf-05e5-489c-87c7-f5984836cd6a" Feb 13 19:51:25.445785 containerd[1468]: time="2025-02-13T19:51:25.445528687Z" level=error msg="encountered an error cleaning up failed sandbox \"20a630c64d17851512c12fc278a3718b46879b27d0ffedf611d9175c93614992\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:25.445785 containerd[1468]: time="2025-02-13T19:51:25.445658500Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-sxwqd,Uid:bbd75eeb-e06b-4812-b2c6-522804c5ef2f,Namespace:calico-system,Attempt:2,} failed, error" error="failed to setup network for sandbox \"20a630c64d17851512c12fc278a3718b46879b27d0ffedf611d9175c93614992\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:25.446035 kubelet[2575]: E0213 19:51:25.445978 2575 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"20a630c64d17851512c12fc278a3718b46879b27d0ffedf611d9175c93614992\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:25.446101 kubelet[2575]: E0213 19:51:25.446056 2575 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"20a630c64d17851512c12fc278a3718b46879b27d0ffedf611d9175c93614992\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-sxwqd" Feb 13 19:51:25.446145 kubelet[2575]: E0213 19:51:25.446109 2575 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"20a630c64d17851512c12fc278a3718b46879b27d0ffedf611d9175c93614992\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-sxwqd" Feb 13 19:51:25.446186 kubelet[2575]: E0213 19:51:25.446162 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-sxwqd_calico-system(bbd75eeb-e06b-4812-b2c6-522804c5ef2f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-sxwqd_calico-system(bbd75eeb-e06b-4812-b2c6-522804c5ef2f)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"20a630c64d17851512c12fc278a3718b46879b27d0ffedf611d9175c93614992\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-sxwqd" podUID="bbd75eeb-e06b-4812-b2c6-522804c5ef2f" Feb 13 19:51:25.446738 containerd[1468]: time="2025-02-13T19:51:25.446673025Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-575f8f688d-rm27z,Uid:dca83c7e-397c-4cb7-b95d-96c1c013b67f,Namespace:calico-apiserver,Attempt:3,} failed, error" error="failed to setup network for sandbox \"3d79c24a7d960cef54edf513bffc4dd8a68cbc23ca3e3304eafaabcbc4d88e33\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:25.446904 kubelet[2575]: E0213 19:51:25.446879 2575 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3d79c24a7d960cef54edf513bffc4dd8a68cbc23ca3e3304eafaabcbc4d88e33\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:25.446962 kubelet[2575]: E0213 19:51:25.446914 2575 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3d79c24a7d960cef54edf513bffc4dd8a68cbc23ca3e3304eafaabcbc4d88e33\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-575f8f688d-rm27z" Feb 13 19:51:25.446962 kubelet[2575]: E0213 19:51:25.446936 2575 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3d79c24a7d960cef54edf513bffc4dd8a68cbc23ca3e3304eafaabcbc4d88e33\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-575f8f688d-rm27z" Feb 13 19:51:25.447030 kubelet[2575]: E0213 19:51:25.446969 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-575f8f688d-rm27z_calico-apiserver(dca83c7e-397c-4cb7-b95d-96c1c013b67f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-575f8f688d-rm27z_calico-apiserver(dca83c7e-397c-4cb7-b95d-96c1c013b67f)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"3d79c24a7d960cef54edf513bffc4dd8a68cbc23ca3e3304eafaabcbc4d88e33\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-575f8f688d-rm27z" podUID="dca83c7e-397c-4cb7-b95d-96c1c013b67f" Feb 13 19:51:25.462155 containerd[1468]: time="2025-02-13T19:51:25.462071048Z" level=error msg="Failed to destroy network for sandbox \"ebbbb0cee651304a47f8384965c605db95eabdf650756e32c6b70a0355fbec6d\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:25.462739 containerd[1468]: time="2025-02-13T19:51:25.462687885Z" level=error msg="encountered an error cleaning up failed sandbox \"ebbbb0cee651304a47f8384965c605db95eabdf650756e32c6b70a0355fbec6d\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:25.462950 containerd[1468]: time="2025-02-13T19:51:25.462808432Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-575f8f688d-8855n,Uid:a5e91fd3-19fa-490e-be13-26d807029222,Namespace:calico-apiserver,Attempt:3,} failed, error" error="failed to setup network for sandbox \"ebbbb0cee651304a47f8384965c605db95eabdf650756e32c6b70a0355fbec6d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:25.463224 kubelet[2575]: E0213 19:51:25.463152 2575 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ebbbb0cee651304a47f8384965c605db95eabdf650756e32c6b70a0355fbec6d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:25.463301 kubelet[2575]: E0213 19:51:25.463258 2575 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ebbbb0cee651304a47f8384965c605db95eabdf650756e32c6b70a0355fbec6d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-575f8f688d-8855n" Feb 13 19:51:25.463350 kubelet[2575]: E0213 19:51:25.463295 2575 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ebbbb0cee651304a47f8384965c605db95eabdf650756e32c6b70a0355fbec6d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-575f8f688d-8855n" Feb 13 19:51:25.463449 kubelet[2575]: E0213 19:51:25.463371 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-575f8f688d-8855n_calico-apiserver(a5e91fd3-19fa-490e-be13-26d807029222)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-575f8f688d-8855n_calico-apiserver(a5e91fd3-19fa-490e-be13-26d807029222)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"ebbbb0cee651304a47f8384965c605db95eabdf650756e32c6b70a0355fbec6d\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-575f8f688d-8855n" podUID="a5e91fd3-19fa-490e-be13-26d807029222" Feb 13 19:51:25.640690 containerd[1468]: time="2025-02-13T19:51:25.640562644Z" level=error msg="Failed to destroy network for sandbox \"22c8c985c21b7d2634c18c69f2f75ef4a82d17959955a493172581323c264379\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:25.641162 containerd[1468]: time="2025-02-13T19:51:25.641136482Z" level=error msg="encountered an error cleaning up failed sandbox \"22c8c985c21b7d2634c18c69f2f75ef4a82d17959955a493172581323c264379\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:25.641228 containerd[1468]: time="2025-02-13T19:51:25.641207355Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-6d5rf,Uid:245734af-c06e-49f3-bfef-a966f57e2f98,Namespace:kube-system,Attempt:3,} failed, error" error="failed to setup network for sandbox \"22c8c985c21b7d2634c18c69f2f75ef4a82d17959955a493172581323c264379\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:25.641563 kubelet[2575]: E0213 19:51:25.641504 2575 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"22c8c985c21b7d2634c18c69f2f75ef4a82d17959955a493172581323c264379\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:25.641629 kubelet[2575]: E0213 19:51:25.641602 2575 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"22c8c985c21b7d2634c18c69f2f75ef4a82d17959955a493172581323c264379\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-6d5rf" Feb 13 19:51:25.641663 kubelet[2575]: E0213 19:51:25.641636 2575 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"22c8c985c21b7d2634c18c69f2f75ef4a82d17959955a493172581323c264379\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-6d5rf" Feb 13 19:51:25.643195 kubelet[2575]: E0213 19:51:25.642821 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-6d5rf_kube-system(245734af-c06e-49f3-bfef-a966f57e2f98)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-6d5rf_kube-system(245734af-c06e-49f3-bfef-a966f57e2f98)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"22c8c985c21b7d2634c18c69f2f75ef4a82d17959955a493172581323c264379\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-6d5rf" podUID="245734af-c06e-49f3-bfef-a966f57e2f98" Feb 13 19:51:25.661758 containerd[1468]: time="2025-02-13T19:51:25.661685065Z" level=error msg="Failed to destroy network for sandbox \"339897f85c32bf6b58fdb7fa27b7f977f70adf895ce4769419e9e19adb77c464\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:25.662921 containerd[1468]: time="2025-02-13T19:51:25.662830996Z" level=error msg="encountered an error cleaning up failed sandbox \"339897f85c32bf6b58fdb7fa27b7f977f70adf895ce4769419e9e19adb77c464\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:25.663093 containerd[1468]: time="2025-02-13T19:51:25.662962053Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5bf7d587cb-s28ls,Uid:77a9c264-ba41-4ec0-b704-771924c99471,Namespace:calico-system,Attempt:3,} failed, error" error="failed to setup network for sandbox \"339897f85c32bf6b58fdb7fa27b7f977f70adf895ce4769419e9e19adb77c464\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:25.663268 kubelet[2575]: E0213 19:51:25.663217 2575 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"339897f85c32bf6b58fdb7fa27b7f977f70adf895ce4769419e9e19adb77c464\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:25.663365 kubelet[2575]: E0213 19:51:25.663287 2575 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"339897f85c32bf6b58fdb7fa27b7f977f70adf895ce4769419e9e19adb77c464\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5bf7d587cb-s28ls" Feb 13 19:51:25.663365 kubelet[2575]: E0213 19:51:25.663308 2575 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"339897f85c32bf6b58fdb7fa27b7f977f70adf895ce4769419e9e19adb77c464\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5bf7d587cb-s28ls" Feb 13 19:51:25.663432 kubelet[2575]: E0213 19:51:25.663396 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-5bf7d587cb-s28ls_calico-system(77a9c264-ba41-4ec0-b704-771924c99471)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-5bf7d587cb-s28ls_calico-system(77a9c264-ba41-4ec0-b704-771924c99471)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"339897f85c32bf6b58fdb7fa27b7f977f70adf895ce4769419e9e19adb77c464\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-5bf7d587cb-s28ls" podUID="77a9c264-ba41-4ec0-b704-771924c99471" Feb 13 19:51:26.108348 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-339897f85c32bf6b58fdb7fa27b7f977f70adf895ce4769419e9e19adb77c464-shm.mount: Deactivated successfully. Feb 13 19:51:26.108859 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-20a630c64d17851512c12fc278a3718b46879b27d0ffedf611d9175c93614992-shm.mount: Deactivated successfully. Feb 13 19:51:26.109082 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-22c8c985c21b7d2634c18c69f2f75ef4a82d17959955a493172581323c264379-shm.mount: Deactivated successfully. Feb 13 19:51:26.109185 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-ebbbb0cee651304a47f8384965c605db95eabdf650756e32c6b70a0355fbec6d-shm.mount: Deactivated successfully. Feb 13 19:51:26.109279 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-3d79c24a7d960cef54edf513bffc4dd8a68cbc23ca3e3304eafaabcbc4d88e33-shm.mount: Deactivated successfully. Feb 13 19:51:26.247744 kubelet[2575]: I0213 19:51:26.247699 2575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d79c24a7d960cef54edf513bffc4dd8a68cbc23ca3e3304eafaabcbc4d88e33" Feb 13 19:51:26.249384 containerd[1468]: time="2025-02-13T19:51:26.249320765Z" level=info msg="StopPodSandbox for \"3d79c24a7d960cef54edf513bffc4dd8a68cbc23ca3e3304eafaabcbc4d88e33\"" Feb 13 19:51:26.250008 containerd[1468]: time="2025-02-13T19:51:26.249587726Z" level=info msg="Ensure that sandbox 3d79c24a7d960cef54edf513bffc4dd8a68cbc23ca3e3304eafaabcbc4d88e33 in task-service has been cleanup successfully" Feb 13 19:51:26.252050 containerd[1468]: time="2025-02-13T19:51:26.252016045Z" level=info msg="TearDown network for sandbox \"3d79c24a7d960cef54edf513bffc4dd8a68cbc23ca3e3304eafaabcbc4d88e33\" successfully" Feb 13 19:51:26.252050 containerd[1468]: time="2025-02-13T19:51:26.252042896Z" level=info msg="StopPodSandbox for \"3d79c24a7d960cef54edf513bffc4dd8a68cbc23ca3e3304eafaabcbc4d88e33\" returns successfully" Feb 13 19:51:26.252997 systemd[1]: run-netns-cni\x2d4b7335db\x2d66b2\x2d76cf\x2d87f5\x2dce94326c49c1.mount: Deactivated successfully. Feb 13 19:51:26.254504 containerd[1468]: time="2025-02-13T19:51:26.254423175Z" level=info msg="StopPodSandbox for \"5f114ecd2418a98f886f02a3a5637918502bf71cd1bf3c74f6c1dd0e45034e21\"" Feb 13 19:51:26.254586 containerd[1468]: time="2025-02-13T19:51:26.254565552Z" level=info msg="TearDown network for sandbox \"5f114ecd2418a98f886f02a3a5637918502bf71cd1bf3c74f6c1dd0e45034e21\" successfully" Feb 13 19:51:26.254586 containerd[1468]: time="2025-02-13T19:51:26.254578686Z" level=info msg="StopPodSandbox for \"5f114ecd2418a98f886f02a3a5637918502bf71cd1bf3c74f6c1dd0e45034e21\" returns successfully" Feb 13 19:51:26.255193 containerd[1468]: time="2025-02-13T19:51:26.255169907Z" level=info msg="StopPodSandbox for \"3af9c99589cb59a86666f237cc1b2ea19ad1577ee4287478e22ace117f62bf47\"" Feb 13 19:51:26.255289 containerd[1468]: time="2025-02-13T19:51:26.255268632Z" level=info msg="TearDown network for sandbox \"3af9c99589cb59a86666f237cc1b2ea19ad1577ee4287478e22ace117f62bf47\" successfully" Feb 13 19:51:26.255341 containerd[1468]: time="2025-02-13T19:51:26.255286165Z" level=info msg="StopPodSandbox for \"3af9c99589cb59a86666f237cc1b2ea19ad1577ee4287478e22ace117f62bf47\" returns successfully" Feb 13 19:51:26.256192 containerd[1468]: time="2025-02-13T19:51:26.255536275Z" level=info msg="StopPodSandbox for \"4ff7a1ce33e60be86627d63cbaa39e854dfba3ccfcd25d3b18dc72196a2b96da\"" Feb 13 19:51:26.256192 containerd[1468]: time="2025-02-13T19:51:26.255623599Z" level=info msg="TearDown network for sandbox \"4ff7a1ce33e60be86627d63cbaa39e854dfba3ccfcd25d3b18dc72196a2b96da\" successfully" Feb 13 19:51:26.256192 containerd[1468]: time="2025-02-13T19:51:26.255634890Z" level=info msg="StopPodSandbox for \"4ff7a1ce33e60be86627d63cbaa39e854dfba3ccfcd25d3b18dc72196a2b96da\" returns successfully" Feb 13 19:51:26.256303 kubelet[2575]: I0213 19:51:26.255730 2575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ebbbb0cee651304a47f8384965c605db95eabdf650756e32c6b70a0355fbec6d" Feb 13 19:51:26.256864 containerd[1468]: time="2025-02-13T19:51:26.256823802Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-575f8f688d-rm27z,Uid:dca83c7e-397c-4cb7-b95d-96c1c013b67f,Namespace:calico-apiserver,Attempt:4,}" Feb 13 19:51:26.257106 containerd[1468]: time="2025-02-13T19:51:26.257075144Z" level=info msg="StopPodSandbox for \"ebbbb0cee651304a47f8384965c605db95eabdf650756e32c6b70a0355fbec6d\"" Feb 13 19:51:26.257309 containerd[1468]: time="2025-02-13T19:51:26.257286761Z" level=info msg="Ensure that sandbox ebbbb0cee651304a47f8384965c605db95eabdf650756e32c6b70a0355fbec6d in task-service has been cleanup successfully" Feb 13 19:51:26.257615 containerd[1468]: time="2025-02-13T19:51:26.257583789Z" level=info msg="TearDown network for sandbox \"ebbbb0cee651304a47f8384965c605db95eabdf650756e32c6b70a0355fbec6d\" successfully" Feb 13 19:51:26.257615 containerd[1468]: time="2025-02-13T19:51:26.257606982Z" level=info msg="StopPodSandbox for \"ebbbb0cee651304a47f8384965c605db95eabdf650756e32c6b70a0355fbec6d\" returns successfully" Feb 13 19:51:26.261398 systemd[1]: run-netns-cni\x2db1ab855c\x2dd01a\x2dabd7\x2dbf6b\x2d49b7f49a6e41.mount: Deactivated successfully. Feb 13 19:51:26.264650 kubelet[2575]: I0213 19:51:26.264566 2575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="339897f85c32bf6b58fdb7fa27b7f977f70adf895ce4769419e9e19adb77c464" Feb 13 19:51:26.265190 containerd[1468]: time="2025-02-13T19:51:26.265144394Z" level=info msg="StopPodSandbox for \"339897f85c32bf6b58fdb7fa27b7f977f70adf895ce4769419e9e19adb77c464\"" Feb 13 19:51:26.265467 containerd[1468]: time="2025-02-13T19:51:26.265435069Z" level=info msg="Ensure that sandbox 339897f85c32bf6b58fdb7fa27b7f977f70adf895ce4769419e9e19adb77c464 in task-service has been cleanup successfully" Feb 13 19:51:26.265728 containerd[1468]: time="2025-02-13T19:51:26.265692282Z" level=info msg="TearDown network for sandbox \"339897f85c32bf6b58fdb7fa27b7f977f70adf895ce4769419e9e19adb77c464\" successfully" Feb 13 19:51:26.265728 containerd[1468]: time="2025-02-13T19:51:26.265718171Z" level=info msg="StopPodSandbox for \"339897f85c32bf6b58fdb7fa27b7f977f70adf895ce4769419e9e19adb77c464\" returns successfully" Feb 13 19:51:26.266125 containerd[1468]: time="2025-02-13T19:51:26.266089087Z" level=info msg="StopPodSandbox for \"deadb290c2eabe3a6b5ed648ceb2f5121ab56b3728b197df70e6532dc828cd25\"" Feb 13 19:51:26.266207 containerd[1468]: time="2025-02-13T19:51:26.266182683Z" level=info msg="TearDown network for sandbox \"deadb290c2eabe3a6b5ed648ceb2f5121ab56b3728b197df70e6532dc828cd25\" successfully" Feb 13 19:51:26.266207 containerd[1468]: time="2025-02-13T19:51:26.266201468Z" level=info msg="StopPodSandbox for \"deadb290c2eabe3a6b5ed648ceb2f5121ab56b3728b197df70e6532dc828cd25\" returns successfully" Feb 13 19:51:26.267072 containerd[1468]: time="2025-02-13T19:51:26.266603763Z" level=info msg="StopPodSandbox for \"cc0fab9693293af68d0473a6799ea1e188b038744dcf56927b1d4fffd60c2c96\"" Feb 13 19:51:26.267072 containerd[1468]: time="2025-02-13T19:51:26.266673545Z" level=info msg="TearDown network for sandbox \"cc0fab9693293af68d0473a6799ea1e188b038744dcf56927b1d4fffd60c2c96\" successfully" Feb 13 19:51:26.267072 containerd[1468]: time="2025-02-13T19:51:26.266683604Z" level=info msg="StopPodSandbox for \"cc0fab9693293af68d0473a6799ea1e188b038744dcf56927b1d4fffd60c2c96\" returns successfully" Feb 13 19:51:26.267072 containerd[1468]: time="2025-02-13T19:51:26.266989688Z" level=info msg="StopPodSandbox for \"c143929d3d37d0fd39eeb807647ccdb8bf35b3017d21a93a728ef6fa513987e4\"" Feb 13 19:51:26.267072 containerd[1468]: time="2025-02-13T19:51:26.267062786Z" level=info msg="TearDown network for sandbox \"c143929d3d37d0fd39eeb807647ccdb8bf35b3017d21a93a728ef6fa513987e4\" successfully" Feb 13 19:51:26.267072 containerd[1468]: time="2025-02-13T19:51:26.267074027Z" level=info msg="StopPodSandbox for \"c143929d3d37d0fd39eeb807647ccdb8bf35b3017d21a93a728ef6fa513987e4\" returns successfully" Feb 13 19:51:26.267747 kubelet[2575]: I0213 19:51:26.267445 2575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="20a630c64d17851512c12fc278a3718b46879b27d0ffedf611d9175c93614992" Feb 13 19:51:26.268013 containerd[1468]: time="2025-02-13T19:51:26.267983053Z" level=info msg="StopPodSandbox for \"20a630c64d17851512c12fc278a3718b46879b27d0ffedf611d9175c93614992\"" Feb 13 19:51:26.268248 containerd[1468]: time="2025-02-13T19:51:26.268174823Z" level=info msg="Ensure that sandbox 20a630c64d17851512c12fc278a3718b46879b27d0ffedf611d9175c93614992 in task-service has been cleanup successfully" Feb 13 19:51:26.268538 containerd[1468]: time="2025-02-13T19:51:26.268402982Z" level=info msg="TearDown network for sandbox \"20a630c64d17851512c12fc278a3718b46879b27d0ffedf611d9175c93614992\" successfully" Feb 13 19:51:26.268538 containerd[1468]: time="2025-02-13T19:51:26.268420505Z" level=info msg="StopPodSandbox for \"20a630c64d17851512c12fc278a3718b46879b27d0ffedf611d9175c93614992\" returns successfully" Feb 13 19:51:26.268600 systemd[1]: run-netns-cni\x2d320ca717\x2df3d7\x2d84ef\x2df4c4\x2d6e6d5f3d0b15.mount: Deactivated successfully. Feb 13 19:51:26.269190 containerd[1468]: time="2025-02-13T19:51:26.269154533Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5bf7d587cb-s28ls,Uid:77a9c264-ba41-4ec0-b704-771924c99471,Namespace:calico-system,Attempt:4,}" Feb 13 19:51:26.269365 containerd[1468]: time="2025-02-13T19:51:26.269333749Z" level=info msg="StopPodSandbox for \"49835f65aea5736de7af75d5477e71209d3ab4edb28dc78ae778b60bcf5f2e76\"" Feb 13 19:51:26.269746 containerd[1468]: time="2025-02-13T19:51:26.269647759Z" level=info msg="TearDown network for sandbox \"49835f65aea5736de7af75d5477e71209d3ab4edb28dc78ae778b60bcf5f2e76\" successfully" Feb 13 19:51:26.269746 containerd[1468]: time="2025-02-13T19:51:26.269665242Z" level=info msg="StopPodSandbox for \"49835f65aea5736de7af75d5477e71209d3ab4edb28dc78ae778b60bcf5f2e76\" returns successfully" Feb 13 19:51:26.270978 containerd[1468]: time="2025-02-13T19:51:26.270525246Z" level=info msg="StopPodSandbox for \"dee4103e3ed5a58df193e21973191414daa33c176df2ebb53f48f6427c812139\"" Feb 13 19:51:26.270978 containerd[1468]: time="2025-02-13T19:51:26.270645151Z" level=info msg="TearDown network for sandbox \"dee4103e3ed5a58df193e21973191414daa33c176df2ebb53f48f6427c812139\" successfully" Feb 13 19:51:26.270978 containerd[1468]: time="2025-02-13T19:51:26.270684325Z" level=info msg="StopPodSandbox for \"dee4103e3ed5a58df193e21973191414daa33c176df2ebb53f48f6427c812139\" returns successfully" Feb 13 19:51:26.272853 containerd[1468]: time="2025-02-13T19:51:26.272808463Z" level=info msg="StopPodSandbox for \"269c22f856e2916019bfab8db86934d7f53640411cbfd7537abae3e996450f7b\"" Feb 13 19:51:26.272964 containerd[1468]: time="2025-02-13T19:51:26.272926384Z" level=info msg="TearDown network for sandbox \"269c22f856e2916019bfab8db86934d7f53640411cbfd7537abae3e996450f7b\" successfully" Feb 13 19:51:26.272964 containerd[1468]: time="2025-02-13T19:51:26.272939338Z" level=info msg="StopPodSandbox for \"269c22f856e2916019bfab8db86934d7f53640411cbfd7537abae3e996450f7b\" returns successfully" Feb 13 19:51:26.273330 containerd[1468]: time="2025-02-13T19:51:26.273141078Z" level=info msg="StopPodSandbox for \"38f72aa5f0a5062621013b365f7d00faaabff6dd65a3f268813b6c2c1534a463\"" Feb 13 19:51:26.273330 containerd[1468]: time="2025-02-13T19:51:26.273214796Z" level=info msg="TearDown network for sandbox \"38f72aa5f0a5062621013b365f7d00faaabff6dd65a3f268813b6c2c1534a463\" successfully" Feb 13 19:51:26.273330 containerd[1468]: time="2025-02-13T19:51:26.273225025Z" level=info msg="StopPodSandbox for \"38f72aa5f0a5062621013b365f7d00faaabff6dd65a3f268813b6c2c1534a463\" returns successfully" Feb 13 19:51:26.273423 kubelet[2575]: I0213 19:51:26.273155 2575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22c8c985c21b7d2634c18c69f2f75ef4a82d17959955a493172581323c264379" Feb 13 19:51:26.273583 systemd[1]: run-netns-cni\x2dd9a05dc5\x2dc4c9\x2d4f1f\x2d4e4c\x2dd33654f8e904.mount: Deactivated successfully. Feb 13 19:51:26.273701 containerd[1468]: time="2025-02-13T19:51:26.273608465Z" level=info msg="StopPodSandbox for \"22c8c985c21b7d2634c18c69f2f75ef4a82d17959955a493172581323c264379\"" Feb 13 19:51:26.273701 containerd[1468]: time="2025-02-13T19:51:26.273683506Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-sxwqd,Uid:bbd75eeb-e06b-4812-b2c6-522804c5ef2f,Namespace:calico-system,Attempt:3,}" Feb 13 19:51:26.274246 containerd[1468]: time="2025-02-13T19:51:26.274044684Z" level=info msg="Ensure that sandbox 22c8c985c21b7d2634c18c69f2f75ef4a82d17959955a493172581323c264379 in task-service has been cleanup successfully" Feb 13 19:51:26.274246 containerd[1468]: time="2025-02-13T19:51:26.274151544Z" level=info msg="StopPodSandbox for \"517d17b8f015a9521e2907d73d8ffe6db808ece1034c7f320e60f66e57f87de3\"" Feb 13 19:51:26.274522 containerd[1468]: time="2025-02-13T19:51:26.274266971Z" level=info msg="TearDown network for sandbox \"517d17b8f015a9521e2907d73d8ffe6db808ece1034c7f320e60f66e57f87de3\" successfully" Feb 13 19:51:26.274522 containerd[1468]: time="2025-02-13T19:51:26.274280357Z" level=info msg="StopPodSandbox for \"517d17b8f015a9521e2907d73d8ffe6db808ece1034c7f320e60f66e57f87de3\" returns successfully" Feb 13 19:51:26.274744 containerd[1468]: time="2025-02-13T19:51:26.274699594Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-575f8f688d-8855n,Uid:a5e91fd3-19fa-490e-be13-26d807029222,Namespace:calico-apiserver,Attempt:4,}" Feb 13 19:51:26.275021 containerd[1468]: time="2025-02-13T19:51:26.274815311Z" level=info msg="TearDown network for sandbox \"22c8c985c21b7d2634c18c69f2f75ef4a82d17959955a493172581323c264379\" successfully" Feb 13 19:51:26.275021 containerd[1468]: time="2025-02-13T19:51:26.274834377Z" level=info msg="StopPodSandbox for \"22c8c985c21b7d2634c18c69f2f75ef4a82d17959955a493172581323c264379\" returns successfully" Feb 13 19:51:26.275213 containerd[1468]: time="2025-02-13T19:51:26.275193892Z" level=info msg="StopPodSandbox for \"13004b615f48f8fc71da84226292b2e4c272d980ea5cce23fb82164b712c2e30\"" Feb 13 19:51:26.275397 containerd[1468]: time="2025-02-13T19:51:26.275271708Z" level=info msg="TearDown network for sandbox \"13004b615f48f8fc71da84226292b2e4c272d980ea5cce23fb82164b712c2e30\" successfully" Feb 13 19:51:26.275397 containerd[1468]: time="2025-02-13T19:51:26.275285874Z" level=info msg="StopPodSandbox for \"13004b615f48f8fc71da84226292b2e4c272d980ea5cce23fb82164b712c2e30\" returns successfully" Feb 13 19:51:26.275779 containerd[1468]: time="2025-02-13T19:51:26.275745958Z" level=info msg="StopPodSandbox for \"899fcc18f7bfce71348bcd81eba1c4cca6504cfc06183532c64985ae4deedeb3\"" Feb 13 19:51:26.275888 containerd[1468]: time="2025-02-13T19:51:26.275839443Z" level=info msg="TearDown network for sandbox \"899fcc18f7bfce71348bcd81eba1c4cca6504cfc06183532c64985ae4deedeb3\" successfully" Feb 13 19:51:26.275888 containerd[1468]: time="2025-02-13T19:51:26.275853469Z" level=info msg="StopPodSandbox for \"899fcc18f7bfce71348bcd81eba1c4cca6504cfc06183532c64985ae4deedeb3\" returns successfully" Feb 13 19:51:26.276237 containerd[1468]: time="2025-02-13T19:51:26.276181806Z" level=info msg="StopPodSandbox for \"a24dd5ac08e1be65e64f1e0e1bcc57fba990025660893d49024096981b2ba3aa\"" Feb 13 19:51:26.276283 containerd[1468]: time="2025-02-13T19:51:26.276254583Z" level=info msg="TearDown network for sandbox \"a24dd5ac08e1be65e64f1e0e1bcc57fba990025660893d49024096981b2ba3aa\" successfully" Feb 13 19:51:26.276283 containerd[1468]: time="2025-02-13T19:51:26.276266475Z" level=info msg="StopPodSandbox for \"a24dd5ac08e1be65e64f1e0e1bcc57fba990025660893d49024096981b2ba3aa\" returns successfully" Feb 13 19:51:26.276588 kubelet[2575]: E0213 19:51:26.276566 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:26.278562 kubelet[2575]: I0213 19:51:26.276689 2575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4d2487cfe7f2e3ccbb95c87ade298917e21e16bc54e8c6d5da7846fe11b6c37" Feb 13 19:51:26.278884 containerd[1468]: time="2025-02-13T19:51:26.278839176Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-6d5rf,Uid:245734af-c06e-49f3-bfef-a966f57e2f98,Namespace:kube-system,Attempt:4,}" Feb 13 19:51:26.279572 containerd[1468]: time="2025-02-13T19:51:26.279541434Z" level=info msg="StopPodSandbox for \"e4d2487cfe7f2e3ccbb95c87ade298917e21e16bc54e8c6d5da7846fe11b6c37\"" Feb 13 19:51:26.279839 containerd[1468]: time="2025-02-13T19:51:26.279765915Z" level=info msg="Ensure that sandbox e4d2487cfe7f2e3ccbb95c87ade298917e21e16bc54e8c6d5da7846fe11b6c37 in task-service has been cleanup successfully" Feb 13 19:51:26.280063 containerd[1468]: time="2025-02-13T19:51:26.280034219Z" level=info msg="TearDown network for sandbox \"e4d2487cfe7f2e3ccbb95c87ade298917e21e16bc54e8c6d5da7846fe11b6c37\" successfully" Feb 13 19:51:26.280063 containerd[1468]: time="2025-02-13T19:51:26.280057062Z" level=info msg="StopPodSandbox for \"e4d2487cfe7f2e3ccbb95c87ade298917e21e16bc54e8c6d5da7846fe11b6c37\" returns successfully" Feb 13 19:51:26.280912 containerd[1468]: time="2025-02-13T19:51:26.280652860Z" level=info msg="StopPodSandbox for \"86a985525093286f84ab532b093323eb20cae87c984957fc89def62353843bd5\"" Feb 13 19:51:26.280912 containerd[1468]: time="2025-02-13T19:51:26.280778938Z" level=info msg="TearDown network for sandbox \"86a985525093286f84ab532b093323eb20cae87c984957fc89def62353843bd5\" successfully" Feb 13 19:51:26.280912 containerd[1468]: time="2025-02-13T19:51:26.280789878Z" level=info msg="StopPodSandbox for \"86a985525093286f84ab532b093323eb20cae87c984957fc89def62353843bd5\" returns successfully" Feb 13 19:51:26.281316 containerd[1468]: time="2025-02-13T19:51:26.281283254Z" level=info msg="StopPodSandbox for \"d7392cd735f64bbb484b37d152068bd95b9e9585d02b78303932e81c6708295c\"" Feb 13 19:51:26.281488 containerd[1468]: time="2025-02-13T19:51:26.281469124Z" level=info msg="TearDown network for sandbox \"d7392cd735f64bbb484b37d152068bd95b9e9585d02b78303932e81c6708295c\" successfully" Feb 13 19:51:26.281488 containerd[1468]: time="2025-02-13T19:51:26.281483871Z" level=info msg="StopPodSandbox for \"d7392cd735f64bbb484b37d152068bd95b9e9585d02b78303932e81c6708295c\" returns successfully" Feb 13 19:51:26.282210 containerd[1468]: time="2025-02-13T19:51:26.282147016Z" level=info msg="StopPodSandbox for \"aea5462f72694272b5bc284a0bd9fc173269129df512c011930e89667ad0f97d\"" Feb 13 19:51:26.282359 containerd[1468]: time="2025-02-13T19:51:26.282337394Z" level=info msg="TearDown network for sandbox \"aea5462f72694272b5bc284a0bd9fc173269129df512c011930e89667ad0f97d\" successfully" Feb 13 19:51:26.282420 containerd[1468]: time="2025-02-13T19:51:26.282355748Z" level=info msg="StopPodSandbox for \"aea5462f72694272b5bc284a0bd9fc173269129df512c011930e89667ad0f97d\" returns successfully" Feb 13 19:51:26.282849 kubelet[2575]: E0213 19:51:26.282703 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:26.283370 containerd[1468]: time="2025-02-13T19:51:26.283154267Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-s22xp,Uid:bd8b42bf-05e5-489c-87c7-f5984836cd6a,Namespace:kube-system,Attempt:4,}" Feb 13 19:51:27.121648 systemd[1]: run-netns-cni\x2da69bbdc1\x2dfe41\x2d008c\x2d14ad\x2d6d9cd333c719.mount: Deactivated successfully. Feb 13 19:51:27.121793 systemd[1]: run-netns-cni\x2dd56097a6\x2d79f0\x2dcbf8\x2d0582\x2da26b9151a65c.mount: Deactivated successfully. Feb 13 19:51:27.194192 containerd[1468]: time="2025-02-13T19:51:27.193978606Z" level=error msg="Failed to destroy network for sandbox \"ed33748ae2f78ecf5ee2c805ac620f4f19dbb1b895af40e625d52cc88bec3832\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.197491 containerd[1468]: time="2025-02-13T19:51:27.197102000Z" level=error msg="encountered an error cleaning up failed sandbox \"ed33748ae2f78ecf5ee2c805ac620f4f19dbb1b895af40e625d52cc88bec3832\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.197491 containerd[1468]: time="2025-02-13T19:51:27.197191177Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5bf7d587cb-s28ls,Uid:77a9c264-ba41-4ec0-b704-771924c99471,Namespace:calico-system,Attempt:4,} failed, error" error="failed to setup network for sandbox \"ed33748ae2f78ecf5ee2c805ac620f4f19dbb1b895af40e625d52cc88bec3832\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.197760 kubelet[2575]: E0213 19:51:27.197502 2575 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ed33748ae2f78ecf5ee2c805ac620f4f19dbb1b895af40e625d52cc88bec3832\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.197760 kubelet[2575]: E0213 19:51:27.197598 2575 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ed33748ae2f78ecf5ee2c805ac620f4f19dbb1b895af40e625d52cc88bec3832\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5bf7d587cb-s28ls" Feb 13 19:51:27.197760 kubelet[2575]: E0213 19:51:27.197631 2575 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ed33748ae2f78ecf5ee2c805ac620f4f19dbb1b895af40e625d52cc88bec3832\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5bf7d587cb-s28ls" Feb 13 19:51:27.197559 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-ed33748ae2f78ecf5ee2c805ac620f4f19dbb1b895af40e625d52cc88bec3832-shm.mount: Deactivated successfully. Feb 13 19:51:27.198770 kubelet[2575]: E0213 19:51:27.197691 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-5bf7d587cb-s28ls_calico-system(77a9c264-ba41-4ec0-b704-771924c99471)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-5bf7d587cb-s28ls_calico-system(77a9c264-ba41-4ec0-b704-771924c99471)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"ed33748ae2f78ecf5ee2c805ac620f4f19dbb1b895af40e625d52cc88bec3832\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-5bf7d587cb-s28ls" podUID="77a9c264-ba41-4ec0-b704-771924c99471" Feb 13 19:51:27.221427 containerd[1468]: time="2025-02-13T19:51:27.221341836Z" level=error msg="Failed to destroy network for sandbox \"d42fa1fe42814b711ad3cb3d471c4742bf3051f8d634e2db762b56c01dc00d5a\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.228806 containerd[1468]: time="2025-02-13T19:51:27.227002203Z" level=error msg="encountered an error cleaning up failed sandbox \"d42fa1fe42814b711ad3cb3d471c4742bf3051f8d634e2db762b56c01dc00d5a\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.228806 containerd[1468]: time="2025-02-13T19:51:27.227120665Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-575f8f688d-rm27z,Uid:dca83c7e-397c-4cb7-b95d-96c1c013b67f,Namespace:calico-apiserver,Attempt:4,} failed, error" error="failed to setup network for sandbox \"d42fa1fe42814b711ad3cb3d471c4742bf3051f8d634e2db762b56c01dc00d5a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.229633 kubelet[2575]: E0213 19:51:27.227464 2575 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d42fa1fe42814b711ad3cb3d471c4742bf3051f8d634e2db762b56c01dc00d5a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.229633 kubelet[2575]: E0213 19:51:27.227851 2575 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d42fa1fe42814b711ad3cb3d471c4742bf3051f8d634e2db762b56c01dc00d5a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-575f8f688d-rm27z" Feb 13 19:51:27.229633 kubelet[2575]: E0213 19:51:27.227915 2575 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d42fa1fe42814b711ad3cb3d471c4742bf3051f8d634e2db762b56c01dc00d5a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-575f8f688d-rm27z" Feb 13 19:51:27.230835 containerd[1468]: time="2025-02-13T19:51:27.229141850Z" level=error msg="Failed to destroy network for sandbox \"785b364a151170b1b8e2dc79f0469cae0c56fbd88d84bde56a96da9c32fc64d7\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.230270 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-d42fa1fe42814b711ad3cb3d471c4742bf3051f8d634e2db762b56c01dc00d5a-shm.mount: Deactivated successfully. Feb 13 19:51:27.231076 kubelet[2575]: E0213 19:51:27.227985 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-575f8f688d-rm27z_calico-apiserver(dca83c7e-397c-4cb7-b95d-96c1c013b67f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-575f8f688d-rm27z_calico-apiserver(dca83c7e-397c-4cb7-b95d-96c1c013b67f)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"d42fa1fe42814b711ad3cb3d471c4742bf3051f8d634e2db762b56c01dc00d5a\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-575f8f688d-rm27z" podUID="dca83c7e-397c-4cb7-b95d-96c1c013b67f" Feb 13 19:51:27.235077 containerd[1468]: time="2025-02-13T19:51:27.231148165Z" level=error msg="encountered an error cleaning up failed sandbox \"785b364a151170b1b8e2dc79f0469cae0c56fbd88d84bde56a96da9c32fc64d7\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.235077 containerd[1468]: time="2025-02-13T19:51:27.231237022Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-6d5rf,Uid:245734af-c06e-49f3-bfef-a966f57e2f98,Namespace:kube-system,Attempt:4,} failed, error" error="failed to setup network for sandbox \"785b364a151170b1b8e2dc79f0469cae0c56fbd88d84bde56a96da9c32fc64d7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.235077 containerd[1468]: time="2025-02-13T19:51:27.231417501Z" level=error msg="Failed to destroy network for sandbox \"39ec85c7a0b06b4fc75b83977eee38fcc3ba3782a02a024285436451c00bfc91\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.235077 containerd[1468]: time="2025-02-13T19:51:27.233361360Z" level=error msg="encountered an error cleaning up failed sandbox \"39ec85c7a0b06b4fc75b83977eee38fcc3ba3782a02a024285436451c00bfc91\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.235077 containerd[1468]: time="2025-02-13T19:51:27.233400895Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-575f8f688d-8855n,Uid:a5e91fd3-19fa-490e-be13-26d807029222,Namespace:calico-apiserver,Attempt:4,} failed, error" error="failed to setup network for sandbox \"39ec85c7a0b06b4fc75b83977eee38fcc3ba3782a02a024285436451c00bfc91\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.235276 kubelet[2575]: E0213 19:51:27.232235 2575 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"785b364a151170b1b8e2dc79f0469cae0c56fbd88d84bde56a96da9c32fc64d7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.235276 kubelet[2575]: E0213 19:51:27.232484 2575 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"785b364a151170b1b8e2dc79f0469cae0c56fbd88d84bde56a96da9c32fc64d7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-6d5rf" Feb 13 19:51:27.235276 kubelet[2575]: E0213 19:51:27.232528 2575 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"785b364a151170b1b8e2dc79f0469cae0c56fbd88d84bde56a96da9c32fc64d7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-6d5rf" Feb 13 19:51:27.235359 kubelet[2575]: E0213 19:51:27.232713 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-6d5rf_kube-system(245734af-c06e-49f3-bfef-a966f57e2f98)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-6d5rf_kube-system(245734af-c06e-49f3-bfef-a966f57e2f98)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"785b364a151170b1b8e2dc79f0469cae0c56fbd88d84bde56a96da9c32fc64d7\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-6d5rf" podUID="245734af-c06e-49f3-bfef-a966f57e2f98" Feb 13 19:51:27.235359 kubelet[2575]: E0213 19:51:27.234513 2575 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"39ec85c7a0b06b4fc75b83977eee38fcc3ba3782a02a024285436451c00bfc91\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.235359 kubelet[2575]: E0213 19:51:27.234597 2575 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"39ec85c7a0b06b4fc75b83977eee38fcc3ba3782a02a024285436451c00bfc91\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-575f8f688d-8855n" Feb 13 19:51:27.235465 kubelet[2575]: E0213 19:51:27.234626 2575 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"39ec85c7a0b06b4fc75b83977eee38fcc3ba3782a02a024285436451c00bfc91\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-575f8f688d-8855n" Feb 13 19:51:27.236988 kubelet[2575]: E0213 19:51:27.234674 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-575f8f688d-8855n_calico-apiserver(a5e91fd3-19fa-490e-be13-26d807029222)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-575f8f688d-8855n_calico-apiserver(a5e91fd3-19fa-490e-be13-26d807029222)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"39ec85c7a0b06b4fc75b83977eee38fcc3ba3782a02a024285436451c00bfc91\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-575f8f688d-8855n" podUID="a5e91fd3-19fa-490e-be13-26d807029222" Feb 13 19:51:27.237621 containerd[1468]: time="2025-02-13T19:51:27.237578858Z" level=error msg="Failed to destroy network for sandbox \"6e045cd079efa77c85217df21eca6d5e6fd9d9297b9e159cbd34c6c2ec690015\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.237783 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-785b364a151170b1b8e2dc79f0469cae0c56fbd88d84bde56a96da9c32fc64d7-shm.mount: Deactivated successfully. Feb 13 19:51:27.238111 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-39ec85c7a0b06b4fc75b83977eee38fcc3ba3782a02a024285436451c00bfc91-shm.mount: Deactivated successfully. Feb 13 19:51:27.238599 containerd[1468]: time="2025-02-13T19:51:27.238518793Z" level=error msg="encountered an error cleaning up failed sandbox \"6e045cd079efa77c85217df21eca6d5e6fd9d9297b9e159cbd34c6c2ec690015\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.239991 containerd[1468]: time="2025-02-13T19:51:27.239820447Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-s22xp,Uid:bd8b42bf-05e5-489c-87c7-f5984836cd6a,Namespace:kube-system,Attempt:4,} failed, error" error="failed to setup network for sandbox \"6e045cd079efa77c85217df21eca6d5e6fd9d9297b9e159cbd34c6c2ec690015\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.241598 kubelet[2575]: E0213 19:51:27.241494 2575 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6e045cd079efa77c85217df21eca6d5e6fd9d9297b9e159cbd34c6c2ec690015\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.242401 kubelet[2575]: E0213 19:51:27.241945 2575 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6e045cd079efa77c85217df21eca6d5e6fd9d9297b9e159cbd34c6c2ec690015\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-s22xp" Feb 13 19:51:27.242401 kubelet[2575]: E0213 19:51:27.242100 2575 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6e045cd079efa77c85217df21eca6d5e6fd9d9297b9e159cbd34c6c2ec690015\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-s22xp" Feb 13 19:51:27.243076 kubelet[2575]: E0213 19:51:27.242959 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-s22xp_kube-system(bd8b42bf-05e5-489c-87c7-f5984836cd6a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-s22xp_kube-system(bd8b42bf-05e5-489c-87c7-f5984836cd6a)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"6e045cd079efa77c85217df21eca6d5e6fd9d9297b9e159cbd34c6c2ec690015\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-s22xp" podUID="bd8b42bf-05e5-489c-87c7-f5984836cd6a" Feb 13 19:51:27.243635 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-6e045cd079efa77c85217df21eca6d5e6fd9d9297b9e159cbd34c6c2ec690015-shm.mount: Deactivated successfully. Feb 13 19:51:27.245306 containerd[1468]: time="2025-02-13T19:51:27.244671864Z" level=error msg="Failed to destroy network for sandbox \"d52193e0d0af1f395378fbf458da41b44ce26d784835ca82240d3687ed53ca6f\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.245306 containerd[1468]: time="2025-02-13T19:51:27.245154270Z" level=error msg="encountered an error cleaning up failed sandbox \"d52193e0d0af1f395378fbf458da41b44ce26d784835ca82240d3687ed53ca6f\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.245306 containerd[1468]: time="2025-02-13T19:51:27.245220173Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-sxwqd,Uid:bbd75eeb-e06b-4812-b2c6-522804c5ef2f,Namespace:calico-system,Attempt:3,} failed, error" error="failed to setup network for sandbox \"d52193e0d0af1f395378fbf458da41b44ce26d784835ca82240d3687ed53ca6f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.245499 kubelet[2575]: E0213 19:51:27.245414 2575 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d52193e0d0af1f395378fbf458da41b44ce26d784835ca82240d3687ed53ca6f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.245499 kubelet[2575]: E0213 19:51:27.245481 2575 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d52193e0d0af1f395378fbf458da41b44ce26d784835ca82240d3687ed53ca6f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-sxwqd" Feb 13 19:51:27.245636 kubelet[2575]: E0213 19:51:27.245502 2575 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d52193e0d0af1f395378fbf458da41b44ce26d784835ca82240d3687ed53ca6f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-sxwqd" Feb 13 19:51:27.245636 kubelet[2575]: E0213 19:51:27.245541 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-sxwqd_calico-system(bbd75eeb-e06b-4812-b2c6-522804c5ef2f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-sxwqd_calico-system(bbd75eeb-e06b-4812-b2c6-522804c5ef2f)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"d52193e0d0af1f395378fbf458da41b44ce26d784835ca82240d3687ed53ca6f\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-sxwqd" podUID="bbd75eeb-e06b-4812-b2c6-522804c5ef2f" Feb 13 19:51:27.282047 kubelet[2575]: I0213 19:51:27.282012 2575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed33748ae2f78ecf5ee2c805ac620f4f19dbb1b895af40e625d52cc88bec3832" Feb 13 19:51:27.283354 containerd[1468]: time="2025-02-13T19:51:27.283285806Z" level=info msg="StopPodSandbox for \"ed33748ae2f78ecf5ee2c805ac620f4f19dbb1b895af40e625d52cc88bec3832\"" Feb 13 19:51:27.284385 containerd[1468]: time="2025-02-13T19:51:27.284181478Z" level=info msg="Ensure that sandbox ed33748ae2f78ecf5ee2c805ac620f4f19dbb1b895af40e625d52cc88bec3832 in task-service has been cleanup successfully" Feb 13 19:51:27.284797 containerd[1468]: time="2025-02-13T19:51:27.284780221Z" level=info msg="TearDown network for sandbox \"ed33748ae2f78ecf5ee2c805ac620f4f19dbb1b895af40e625d52cc88bec3832\" successfully" Feb 13 19:51:27.284938 containerd[1468]: time="2025-02-13T19:51:27.284896149Z" level=info msg="StopPodSandbox for \"ed33748ae2f78ecf5ee2c805ac620f4f19dbb1b895af40e625d52cc88bec3832\" returns successfully" Feb 13 19:51:27.285818 kubelet[2575]: I0213 19:51:27.285311 2575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d42fa1fe42814b711ad3cb3d471c4742bf3051f8d634e2db762b56c01dc00d5a" Feb 13 19:51:27.286090 containerd[1468]: time="2025-02-13T19:51:27.286068490Z" level=info msg="StopPodSandbox for \"339897f85c32bf6b58fdb7fa27b7f977f70adf895ce4769419e9e19adb77c464\"" Feb 13 19:51:27.286241 containerd[1468]: time="2025-02-13T19:51:27.286222238Z" level=info msg="TearDown network for sandbox \"339897f85c32bf6b58fdb7fa27b7f977f70adf895ce4769419e9e19adb77c464\" successfully" Feb 13 19:51:27.286327 containerd[1468]: time="2025-02-13T19:51:27.286307569Z" level=info msg="StopPodSandbox for \"339897f85c32bf6b58fdb7fa27b7f977f70adf895ce4769419e9e19adb77c464\" returns successfully" Feb 13 19:51:27.286651 containerd[1468]: time="2025-02-13T19:51:27.286623732Z" level=info msg="StopPodSandbox for \"d42fa1fe42814b711ad3cb3d471c4742bf3051f8d634e2db762b56c01dc00d5a\"" Feb 13 19:51:27.286958 containerd[1468]: time="2025-02-13T19:51:27.286936018Z" level=info msg="Ensure that sandbox d42fa1fe42814b711ad3cb3d471c4742bf3051f8d634e2db762b56c01dc00d5a in task-service has been cleanup successfully" Feb 13 19:51:27.287254 containerd[1468]: time="2025-02-13T19:51:27.287233197Z" level=info msg="TearDown network for sandbox \"d42fa1fe42814b711ad3cb3d471c4742bf3051f8d634e2db762b56c01dc00d5a\" successfully" Feb 13 19:51:27.287339 containerd[1468]: time="2025-02-13T19:51:27.287321402Z" level=info msg="StopPodSandbox for \"d42fa1fe42814b711ad3cb3d471c4742bf3051f8d634e2db762b56c01dc00d5a\" returns successfully" Feb 13 19:51:27.288370 containerd[1468]: time="2025-02-13T19:51:27.288314746Z" level=info msg="StopPodSandbox for \"3d79c24a7d960cef54edf513bffc4dd8a68cbc23ca3e3304eafaabcbc4d88e33\"" Feb 13 19:51:27.288514 containerd[1468]: time="2025-02-13T19:51:27.288483855Z" level=info msg="TearDown network for sandbox \"3d79c24a7d960cef54edf513bffc4dd8a68cbc23ca3e3304eafaabcbc4d88e33\" successfully" Feb 13 19:51:27.288514 containerd[1468]: time="2025-02-13T19:51:27.288512117Z" level=info msg="StopPodSandbox for \"3d79c24a7d960cef54edf513bffc4dd8a68cbc23ca3e3304eafaabcbc4d88e33\" returns successfully" Feb 13 19:51:27.288590 containerd[1468]: time="2025-02-13T19:51:27.288567321Z" level=info msg="StopPodSandbox for \"deadb290c2eabe3a6b5ed648ceb2f5121ab56b3728b197df70e6532dc828cd25\"" Feb 13 19:51:27.288684 containerd[1468]: time="2025-02-13T19:51:27.288655096Z" level=info msg="TearDown network for sandbox \"deadb290c2eabe3a6b5ed648ceb2f5121ab56b3728b197df70e6532dc828cd25\" successfully" Feb 13 19:51:27.288719 containerd[1468]: time="2025-02-13T19:51:27.288680684Z" level=info msg="StopPodSandbox for \"deadb290c2eabe3a6b5ed648ceb2f5121ab56b3728b197df70e6532dc828cd25\" returns successfully" Feb 13 19:51:27.289308 containerd[1468]: time="2025-02-13T19:51:27.289270872Z" level=info msg="StopPodSandbox for \"5f114ecd2418a98f886f02a3a5637918502bf71cd1bf3c74f6c1dd0e45034e21\"" Feb 13 19:51:27.289415 containerd[1468]: time="2025-02-13T19:51:27.289284898Z" level=info msg="StopPodSandbox for \"cc0fab9693293af68d0473a6799ea1e188b038744dcf56927b1d4fffd60c2c96\"" Feb 13 19:51:27.289535 containerd[1468]: time="2025-02-13T19:51:27.289389034Z" level=info msg="TearDown network for sandbox \"5f114ecd2418a98f886f02a3a5637918502bf71cd1bf3c74f6c1dd0e45034e21\" successfully" Feb 13 19:51:27.289535 containerd[1468]: time="2025-02-13T19:51:27.289492828Z" level=info msg="StopPodSandbox for \"5f114ecd2418a98f886f02a3a5637918502bf71cd1bf3c74f6c1dd0e45034e21\" returns successfully" Feb 13 19:51:27.289535 containerd[1468]: time="2025-02-13T19:51:27.289505933Z" level=info msg="TearDown network for sandbox \"cc0fab9693293af68d0473a6799ea1e188b038744dcf56927b1d4fffd60c2c96\" successfully" Feb 13 19:51:27.289535 containerd[1468]: time="2025-02-13T19:51:27.289522905Z" level=info msg="StopPodSandbox for \"cc0fab9693293af68d0473a6799ea1e188b038744dcf56927b1d4fffd60c2c96\" returns successfully" Feb 13 19:51:27.290912 containerd[1468]: time="2025-02-13T19:51:27.290840558Z" level=info msg="StopPodSandbox for \"3af9c99589cb59a86666f237cc1b2ea19ad1577ee4287478e22ace117f62bf47\"" Feb 13 19:51:27.291088 containerd[1468]: time="2025-02-13T19:51:27.291065371Z" level=info msg="StopPodSandbox for \"c143929d3d37d0fd39eeb807647ccdb8bf35b3017d21a93a728ef6fa513987e4\"" Feb 13 19:51:27.291332 containerd[1468]: time="2025-02-13T19:51:27.291166230Z" level=info msg="TearDown network for sandbox \"c143929d3d37d0fd39eeb807647ccdb8bf35b3017d21a93a728ef6fa513987e4\" successfully" Feb 13 19:51:27.291332 containerd[1468]: time="2025-02-13T19:51:27.291193641Z" level=info msg="StopPodSandbox for \"c143929d3d37d0fd39eeb807647ccdb8bf35b3017d21a93a728ef6fa513987e4\" returns successfully" Feb 13 19:51:27.292996 containerd[1468]: time="2025-02-13T19:51:27.292789698Z" level=info msg="TearDown network for sandbox \"3af9c99589cb59a86666f237cc1b2ea19ad1577ee4287478e22ace117f62bf47\" successfully" Feb 13 19:51:27.292996 containerd[1468]: time="2025-02-13T19:51:27.292833851Z" level=info msg="StopPodSandbox for \"3af9c99589cb59a86666f237cc1b2ea19ad1577ee4287478e22ace117f62bf47\" returns successfully" Feb 13 19:51:27.294908 containerd[1468]: time="2025-02-13T19:51:27.294403237Z" level=info msg="StopPodSandbox for \"4ff7a1ce33e60be86627d63cbaa39e854dfba3ccfcd25d3b18dc72196a2b96da\"" Feb 13 19:51:27.294908 containerd[1468]: time="2025-02-13T19:51:27.294582834Z" level=info msg="TearDown network for sandbox \"4ff7a1ce33e60be86627d63cbaa39e854dfba3ccfcd25d3b18dc72196a2b96da\" successfully" Feb 13 19:51:27.294908 containerd[1468]: time="2025-02-13T19:51:27.294595307Z" level=info msg="StopPodSandbox for \"4ff7a1ce33e60be86627d63cbaa39e854dfba3ccfcd25d3b18dc72196a2b96da\" returns successfully" Feb 13 19:51:27.294908 containerd[1468]: time="2025-02-13T19:51:27.294728387Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5bf7d587cb-s28ls,Uid:77a9c264-ba41-4ec0-b704-771924c99471,Namespace:calico-system,Attempt:5,}" Feb 13 19:51:27.296096 containerd[1468]: time="2025-02-13T19:51:27.296037424Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-575f8f688d-rm27z,Uid:dca83c7e-397c-4cb7-b95d-96c1c013b67f,Namespace:calico-apiserver,Attempt:5,}" Feb 13 19:51:27.296717 kubelet[2575]: I0213 19:51:27.296672 2575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39ec85c7a0b06b4fc75b83977eee38fcc3ba3782a02a024285436451c00bfc91" Feb 13 19:51:27.297366 containerd[1468]: time="2025-02-13T19:51:27.297236926Z" level=info msg="StopPodSandbox for \"39ec85c7a0b06b4fc75b83977eee38fcc3ba3782a02a024285436451c00bfc91\"" Feb 13 19:51:27.297448 containerd[1468]: time="2025-02-13T19:51:27.297423587Z" level=info msg="Ensure that sandbox 39ec85c7a0b06b4fc75b83977eee38fcc3ba3782a02a024285436451c00bfc91 in task-service has been cleanup successfully" Feb 13 19:51:27.297772 containerd[1468]: time="2025-02-13T19:51:27.297715615Z" level=info msg="TearDown network for sandbox \"39ec85c7a0b06b4fc75b83977eee38fcc3ba3782a02a024285436451c00bfc91\" successfully" Feb 13 19:51:27.297772 containerd[1468]: time="2025-02-13T19:51:27.297772882Z" level=info msg="StopPodSandbox for \"39ec85c7a0b06b4fc75b83977eee38fcc3ba3782a02a024285436451c00bfc91\" returns successfully" Feb 13 19:51:27.298370 containerd[1468]: time="2025-02-13T19:51:27.298333455Z" level=info msg="StopPodSandbox for \"ebbbb0cee651304a47f8384965c605db95eabdf650756e32c6b70a0355fbec6d\"" Feb 13 19:51:27.298459 containerd[1468]: time="2025-02-13T19:51:27.298429486Z" level=info msg="TearDown network for sandbox \"ebbbb0cee651304a47f8384965c605db95eabdf650756e32c6b70a0355fbec6d\" successfully" Feb 13 19:51:27.298459 containerd[1468]: time="2025-02-13T19:51:27.298450174Z" level=info msg="StopPodSandbox for \"ebbbb0cee651304a47f8384965c605db95eabdf650756e32c6b70a0355fbec6d\" returns successfully" Feb 13 19:51:27.299135 containerd[1468]: time="2025-02-13T19:51:27.299082632Z" level=info msg="StopPodSandbox for \"269c22f856e2916019bfab8db86934d7f53640411cbfd7537abae3e996450f7b\"" Feb 13 19:51:27.299196 containerd[1468]: time="2025-02-13T19:51:27.299178813Z" level=info msg="TearDown network for sandbox \"269c22f856e2916019bfab8db86934d7f53640411cbfd7537abae3e996450f7b\" successfully" Feb 13 19:51:27.299238 containerd[1468]: time="2025-02-13T19:51:27.299193139Z" level=info msg="StopPodSandbox for \"269c22f856e2916019bfab8db86934d7f53640411cbfd7537abae3e996450f7b\" returns successfully" Feb 13 19:51:27.299822 containerd[1468]: time="2025-02-13T19:51:27.299612326Z" level=info msg="StopPodSandbox for \"38f72aa5f0a5062621013b365f7d00faaabff6dd65a3f268813b6c2c1534a463\"" Feb 13 19:51:27.299822 containerd[1468]: time="2025-02-13T19:51:27.299739134Z" level=info msg="TearDown network for sandbox \"38f72aa5f0a5062621013b365f7d00faaabff6dd65a3f268813b6c2c1534a463\" successfully" Feb 13 19:51:27.299822 containerd[1468]: time="2025-02-13T19:51:27.299755355Z" level=info msg="StopPodSandbox for \"38f72aa5f0a5062621013b365f7d00faaabff6dd65a3f268813b6c2c1534a463\" returns successfully" Feb 13 19:51:27.300645 containerd[1468]: time="2025-02-13T19:51:27.300326176Z" level=info msg="StopPodSandbox for \"517d17b8f015a9521e2907d73d8ffe6db808ece1034c7f320e60f66e57f87de3\"" Feb 13 19:51:27.300645 containerd[1468]: time="2025-02-13T19:51:27.300501095Z" level=info msg="TearDown network for sandbox \"517d17b8f015a9521e2907d73d8ffe6db808ece1034c7f320e60f66e57f87de3\" successfully" Feb 13 19:51:27.300645 containerd[1468]: time="2025-02-13T19:51:27.300517265Z" level=info msg="StopPodSandbox for \"517d17b8f015a9521e2907d73d8ffe6db808ece1034c7f320e60f66e57f87de3\" returns successfully" Feb 13 19:51:27.301425 containerd[1468]: time="2025-02-13T19:51:27.301404491Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-575f8f688d-8855n,Uid:a5e91fd3-19fa-490e-be13-26d807029222,Namespace:calico-apiserver,Attempt:5,}" Feb 13 19:51:27.302570 kubelet[2575]: I0213 19:51:27.302522 2575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="785b364a151170b1b8e2dc79f0469cae0c56fbd88d84bde56a96da9c32fc64d7" Feb 13 19:51:27.303065 containerd[1468]: time="2025-02-13T19:51:27.303016307Z" level=info msg="StopPodSandbox for \"785b364a151170b1b8e2dc79f0469cae0c56fbd88d84bde56a96da9c32fc64d7\"" Feb 13 19:51:27.303245 containerd[1468]: time="2025-02-13T19:51:27.303224207Z" level=info msg="Ensure that sandbox 785b364a151170b1b8e2dc79f0469cae0c56fbd88d84bde56a96da9c32fc64d7 in task-service has been cleanup successfully" Feb 13 19:51:27.303697 containerd[1468]: time="2025-02-13T19:51:27.303675544Z" level=info msg="TearDown network for sandbox \"785b364a151170b1b8e2dc79f0469cae0c56fbd88d84bde56a96da9c32fc64d7\" successfully" Feb 13 19:51:27.303697 containerd[1468]: time="2025-02-13T19:51:27.303694590Z" level=info msg="StopPodSandbox for \"785b364a151170b1b8e2dc79f0469cae0c56fbd88d84bde56a96da9c32fc64d7\" returns successfully" Feb 13 19:51:27.303988 containerd[1468]: time="2025-02-13T19:51:27.303953626Z" level=info msg="StopPodSandbox for \"22c8c985c21b7d2634c18c69f2f75ef4a82d17959955a493172581323c264379\"" Feb 13 19:51:27.305154 containerd[1468]: time="2025-02-13T19:51:27.305038021Z" level=info msg="TearDown network for sandbox \"22c8c985c21b7d2634c18c69f2f75ef4a82d17959955a493172581323c264379\" successfully" Feb 13 19:51:27.305154 containerd[1468]: time="2025-02-13T19:51:27.305067317Z" level=info msg="StopPodSandbox for \"22c8c985c21b7d2634c18c69f2f75ef4a82d17959955a493172581323c264379\" returns successfully" Feb 13 19:51:27.305321 containerd[1468]: time="2025-02-13T19:51:27.305297870Z" level=info msg="StopPodSandbox for \"13004b615f48f8fc71da84226292b2e4c272d980ea5cce23fb82164b712c2e30\"" Feb 13 19:51:27.305489 containerd[1468]: time="2025-02-13T19:51:27.305373552Z" level=info msg="TearDown network for sandbox \"13004b615f48f8fc71da84226292b2e4c272d980ea5cce23fb82164b712c2e30\" successfully" Feb 13 19:51:27.305489 containerd[1468]: time="2025-02-13T19:51:27.305393349Z" level=info msg="StopPodSandbox for \"13004b615f48f8fc71da84226292b2e4c272d980ea5cce23fb82164b712c2e30\" returns successfully" Feb 13 19:51:27.305764 containerd[1468]: time="2025-02-13T19:51:27.305734860Z" level=info msg="StopPodSandbox for \"899fcc18f7bfce71348bcd81eba1c4cca6504cfc06183532c64985ae4deedeb3\"" Feb 13 19:51:27.305862 containerd[1468]: time="2025-02-13T19:51:27.305815782Z" level=info msg="TearDown network for sandbox \"899fcc18f7bfce71348bcd81eba1c4cca6504cfc06183532c64985ae4deedeb3\" successfully" Feb 13 19:51:27.305862 containerd[1468]: time="2025-02-13T19:51:27.305829808Z" level=info msg="StopPodSandbox for \"899fcc18f7bfce71348bcd81eba1c4cca6504cfc06183532c64985ae4deedeb3\" returns successfully" Feb 13 19:51:27.306433 containerd[1468]: time="2025-02-13T19:51:27.306302866Z" level=info msg="StopPodSandbox for \"a24dd5ac08e1be65e64f1e0e1bcc57fba990025660893d49024096981b2ba3aa\"" Feb 13 19:51:27.306720 containerd[1468]: time="2025-02-13T19:51:27.306668853Z" level=info msg="TearDown network for sandbox \"a24dd5ac08e1be65e64f1e0e1bcc57fba990025660893d49024096981b2ba3aa\" successfully" Feb 13 19:51:27.306720 containerd[1468]: time="2025-02-13T19:51:27.306685254Z" level=info msg="StopPodSandbox for \"a24dd5ac08e1be65e64f1e0e1bcc57fba990025660893d49024096981b2ba3aa\" returns successfully" Feb 13 19:51:27.306937 kubelet[2575]: I0213 19:51:27.306728 2575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e045cd079efa77c85217df21eca6d5e6fd9d9297b9e159cbd34c6c2ec690015" Feb 13 19:51:27.306937 kubelet[2575]: E0213 19:51:27.306844 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:27.307210 containerd[1468]: time="2025-02-13T19:51:27.307183410Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-6d5rf,Uid:245734af-c06e-49f3-bfef-a966f57e2f98,Namespace:kube-system,Attempt:5,}" Feb 13 19:51:27.307611 containerd[1468]: time="2025-02-13T19:51:27.307407771Z" level=info msg="StopPodSandbox for \"6e045cd079efa77c85217df21eca6d5e6fd9d9297b9e159cbd34c6c2ec690015\"" Feb 13 19:51:27.311739 kubelet[2575]: I0213 19:51:27.311157 2575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d52193e0d0af1f395378fbf458da41b44ce26d784835ca82240d3687ed53ca6f" Feb 13 19:51:27.311890 containerd[1468]: time="2025-02-13T19:51:27.311382853Z" level=info msg="Ensure that sandbox 6e045cd079efa77c85217df21eca6d5e6fd9d9297b9e159cbd34c6c2ec690015 in task-service has been cleanup successfully" Feb 13 19:51:27.311890 containerd[1468]: time="2025-02-13T19:51:27.311589772Z" level=info msg="TearDown network for sandbox \"6e045cd079efa77c85217df21eca6d5e6fd9d9297b9e159cbd34c6c2ec690015\" successfully" Feb 13 19:51:27.311890 containerd[1468]: time="2025-02-13T19:51:27.311601053Z" level=info msg="StopPodSandbox for \"6e045cd079efa77c85217df21eca6d5e6fd9d9297b9e159cbd34c6c2ec690015\" returns successfully" Feb 13 19:51:27.311890 containerd[1468]: time="2025-02-13T19:51:27.311772525Z" level=info msg="StopPodSandbox for \"d52193e0d0af1f395378fbf458da41b44ce26d784835ca82240d3687ed53ca6f\"" Feb 13 19:51:27.312031 containerd[1468]: time="2025-02-13T19:51:27.311952193Z" level=info msg="Ensure that sandbox d52193e0d0af1f395378fbf458da41b44ce26d784835ca82240d3687ed53ca6f in task-service has been cleanup successfully" Feb 13 19:51:27.312193 containerd[1468]: time="2025-02-13T19:51:27.312125988Z" level=info msg="TearDown network for sandbox \"d52193e0d0af1f395378fbf458da41b44ce26d784835ca82240d3687ed53ca6f\" successfully" Feb 13 19:51:27.312193 containerd[1468]: time="2025-02-13T19:51:27.312142259Z" level=info msg="StopPodSandbox for \"d52193e0d0af1f395378fbf458da41b44ce26d784835ca82240d3687ed53ca6f\" returns successfully" Feb 13 19:51:27.312664 containerd[1468]: time="2025-02-13T19:51:27.312620096Z" level=info msg="StopPodSandbox for \"e4d2487cfe7f2e3ccbb95c87ade298917e21e16bc54e8c6d5da7846fe11b6c37\"" Feb 13 19:51:27.312781 containerd[1468]: time="2025-02-13T19:51:27.312755681Z" level=info msg="TearDown network for sandbox \"e4d2487cfe7f2e3ccbb95c87ade298917e21e16bc54e8c6d5da7846fe11b6c37\" successfully" Feb 13 19:51:27.312822 containerd[1468]: time="2025-02-13T19:51:27.312778464Z" level=info msg="StopPodSandbox for \"e4d2487cfe7f2e3ccbb95c87ade298917e21e16bc54e8c6d5da7846fe11b6c37\" returns successfully" Feb 13 19:51:27.312864 containerd[1468]: time="2025-02-13T19:51:27.312842654Z" level=info msg="StopPodSandbox for \"20a630c64d17851512c12fc278a3718b46879b27d0ffedf611d9175c93614992\"" Feb 13 19:51:27.313179 containerd[1468]: time="2025-02-13T19:51:27.313146315Z" level=info msg="TearDown network for sandbox \"20a630c64d17851512c12fc278a3718b46879b27d0ffedf611d9175c93614992\" successfully" Feb 13 19:51:27.313179 containerd[1468]: time="2025-02-13T19:51:27.313168967Z" level=info msg="StopPodSandbox for \"20a630c64d17851512c12fc278a3718b46879b27d0ffedf611d9175c93614992\" returns successfully" Feb 13 19:51:27.313342 containerd[1468]: time="2025-02-13T19:51:27.313223730Z" level=info msg="StopPodSandbox for \"86a985525093286f84ab532b093323eb20cae87c984957fc89def62353843bd5\"" Feb 13 19:51:27.313432 containerd[1468]: time="2025-02-13T19:51:27.313411872Z" level=info msg="TearDown network for sandbox \"86a985525093286f84ab532b093323eb20cae87c984957fc89def62353843bd5\" successfully" Feb 13 19:51:27.313472 containerd[1468]: time="2025-02-13T19:51:27.313430497Z" level=info msg="StopPodSandbox for \"86a985525093286f84ab532b093323eb20cae87c984957fc89def62353843bd5\" returns successfully" Feb 13 19:51:27.313734 containerd[1468]: time="2025-02-13T19:51:27.313706025Z" level=info msg="StopPodSandbox for \"d7392cd735f64bbb484b37d152068bd95b9e9585d02b78303932e81c6708295c\"" Feb 13 19:51:27.313800 containerd[1468]: time="2025-02-13T19:51:27.313784061Z" level=info msg="TearDown network for sandbox \"d7392cd735f64bbb484b37d152068bd95b9e9585d02b78303932e81c6708295c\" successfully" Feb 13 19:51:27.313832 containerd[1468]: time="2025-02-13T19:51:27.313798458Z" level=info msg="StopPodSandbox for \"d7392cd735f64bbb484b37d152068bd95b9e9585d02b78303932e81c6708295c\" returns successfully" Feb 13 19:51:27.313857 containerd[1468]: time="2025-02-13T19:51:27.313834867Z" level=info msg="StopPodSandbox for \"49835f65aea5736de7af75d5477e71209d3ab4edb28dc78ae778b60bcf5f2e76\"" Feb 13 19:51:27.313935 containerd[1468]: time="2025-02-13T19:51:27.313919365Z" level=info msg="TearDown network for sandbox \"49835f65aea5736de7af75d5477e71209d3ab4edb28dc78ae778b60bcf5f2e76\" successfully" Feb 13 19:51:27.313959 containerd[1468]: time="2025-02-13T19:51:27.313932870Z" level=info msg="StopPodSandbox for \"49835f65aea5736de7af75d5477e71209d3ab4edb28dc78ae778b60bcf5f2e76\" returns successfully" Feb 13 19:51:27.314401 containerd[1468]: time="2025-02-13T19:51:27.314347008Z" level=info msg="StopPodSandbox for \"dee4103e3ed5a58df193e21973191414daa33c176df2ebb53f48f6427c812139\"" Feb 13 19:51:27.314589 containerd[1468]: time="2025-02-13T19:51:27.314347879Z" level=info msg="StopPodSandbox for \"aea5462f72694272b5bc284a0bd9fc173269129df512c011930e89667ad0f97d\"" Feb 13 19:51:27.314694 containerd[1468]: time="2025-02-13T19:51:27.314672840Z" level=info msg="TearDown network for sandbox \"aea5462f72694272b5bc284a0bd9fc173269129df512c011930e89667ad0f97d\" successfully" Feb 13 19:51:27.314717 containerd[1468]: time="2025-02-13T19:51:27.314692777Z" level=info msg="StopPodSandbox for \"aea5462f72694272b5bc284a0bd9fc173269129df512c011930e89667ad0f97d\" returns successfully" Feb 13 19:51:27.314974 containerd[1468]: time="2025-02-13T19:51:27.314912981Z" level=info msg="TearDown network for sandbox \"dee4103e3ed5a58df193e21973191414daa33c176df2ebb53f48f6427c812139\" successfully" Feb 13 19:51:27.314974 containerd[1468]: time="2025-02-13T19:51:27.314940713Z" level=info msg="StopPodSandbox for \"dee4103e3ed5a58df193e21973191414daa33c176df2ebb53f48f6427c812139\" returns successfully" Feb 13 19:51:27.315705 kubelet[2575]: E0213 19:51:27.314960 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:27.317077 containerd[1468]: time="2025-02-13T19:51:27.316474863Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-sxwqd,Uid:bbd75eeb-e06b-4812-b2c6-522804c5ef2f,Namespace:calico-system,Attempt:4,}" Feb 13 19:51:27.317077 containerd[1468]: time="2025-02-13T19:51:27.316814721Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-s22xp,Uid:bd8b42bf-05e5-489c-87c7-f5984836cd6a,Namespace:kube-system,Attempt:5,}" Feb 13 19:51:27.511618 systemd[1]: Started sshd@8-10.0.0.104:22-10.0.0.1:57054.service - OpenSSH per-connection server daemon (10.0.0.1:57054). Feb 13 19:51:27.593318 sshd[4441]: Accepted publickey for core from 10.0.0.1 port 57054 ssh2: RSA SHA256:8WP2kqV5KzwZsuVRMXRFZkAHZWbkdD5kizbT2H+wOcw Feb 13 19:51:27.595287 sshd-session[4441]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:51:27.600928 systemd-logind[1447]: New session 8 of user core. Feb 13 19:51:27.607086 systemd[1]: Started session-8.scope - Session 8 of User core. Feb 13 19:51:27.783424 containerd[1468]: time="2025-02-13T19:51:27.781532625Z" level=error msg="Failed to destroy network for sandbox \"1d976eb55d0e17adfbd3d8051a2fdf3026e7061589ae2a3f34816e887a92e46f\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.783424 containerd[1468]: time="2025-02-13T19:51:27.782096202Z" level=error msg="encountered an error cleaning up failed sandbox \"1d976eb55d0e17adfbd3d8051a2fdf3026e7061589ae2a3f34816e887a92e46f\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.783424 containerd[1468]: time="2025-02-13T19:51:27.782172045Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-sxwqd,Uid:bbd75eeb-e06b-4812-b2c6-522804c5ef2f,Namespace:calico-system,Attempt:4,} failed, error" error="failed to setup network for sandbox \"1d976eb55d0e17adfbd3d8051a2fdf3026e7061589ae2a3f34816e887a92e46f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.783858 kubelet[2575]: E0213 19:51:27.782473 2575 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1d976eb55d0e17adfbd3d8051a2fdf3026e7061589ae2a3f34816e887a92e46f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.783858 kubelet[2575]: E0213 19:51:27.782560 2575 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1d976eb55d0e17adfbd3d8051a2fdf3026e7061589ae2a3f34816e887a92e46f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-sxwqd" Feb 13 19:51:27.783858 kubelet[2575]: E0213 19:51:27.782596 2575 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1d976eb55d0e17adfbd3d8051a2fdf3026e7061589ae2a3f34816e887a92e46f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-sxwqd" Feb 13 19:51:27.784184 kubelet[2575]: E0213 19:51:27.782670 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-sxwqd_calico-system(bbd75eeb-e06b-4812-b2c6-522804c5ef2f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-sxwqd_calico-system(bbd75eeb-e06b-4812-b2c6-522804c5ef2f)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"1d976eb55d0e17adfbd3d8051a2fdf3026e7061589ae2a3f34816e887a92e46f\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-sxwqd" podUID="bbd75eeb-e06b-4812-b2c6-522804c5ef2f" Feb 13 19:51:27.785702 containerd[1468]: time="2025-02-13T19:51:27.785631489Z" level=error msg="Failed to destroy network for sandbox \"729c6e6d3b4530ba0b68738b226a62687ecd65bb8522ed1b880437d5541be3c3\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.786524 containerd[1468]: time="2025-02-13T19:51:27.786483639Z" level=error msg="encountered an error cleaning up failed sandbox \"729c6e6d3b4530ba0b68738b226a62687ecd65bb8522ed1b880437d5541be3c3\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.787072 containerd[1468]: time="2025-02-13T19:51:27.787035114Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-6d5rf,Uid:245734af-c06e-49f3-bfef-a966f57e2f98,Namespace:kube-system,Attempt:5,} failed, error" error="failed to setup network for sandbox \"729c6e6d3b4530ba0b68738b226a62687ecd65bb8522ed1b880437d5541be3c3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.788975 kubelet[2575]: E0213 19:51:27.788403 2575 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"729c6e6d3b4530ba0b68738b226a62687ecd65bb8522ed1b880437d5541be3c3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.788975 kubelet[2575]: E0213 19:51:27.788744 2575 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"729c6e6d3b4530ba0b68738b226a62687ecd65bb8522ed1b880437d5541be3c3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-6d5rf" Feb 13 19:51:27.788975 kubelet[2575]: E0213 19:51:27.788837 2575 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"729c6e6d3b4530ba0b68738b226a62687ecd65bb8522ed1b880437d5541be3c3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-6d5rf" Feb 13 19:51:27.789119 kubelet[2575]: E0213 19:51:27.788910 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-6d5rf_kube-system(245734af-c06e-49f3-bfef-a966f57e2f98)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-6d5rf_kube-system(245734af-c06e-49f3-bfef-a966f57e2f98)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"729c6e6d3b4530ba0b68738b226a62687ecd65bb8522ed1b880437d5541be3c3\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-6d5rf" podUID="245734af-c06e-49f3-bfef-a966f57e2f98" Feb 13 19:51:27.792950 containerd[1468]: time="2025-02-13T19:51:27.792897750Z" level=error msg="Failed to destroy network for sandbox \"5dd1dc9316e1d4cb2cc982113433992b2f865de8097e02c3d7f196ab0750361e\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.800846 containerd[1468]: time="2025-02-13T19:51:27.800779588Z" level=error msg="encountered an error cleaning up failed sandbox \"5dd1dc9316e1d4cb2cc982113433992b2f865de8097e02c3d7f196ab0750361e\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.801366 containerd[1468]: time="2025-02-13T19:51:27.801328297Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-575f8f688d-rm27z,Uid:dca83c7e-397c-4cb7-b95d-96c1c013b67f,Namespace:calico-apiserver,Attempt:5,} failed, error" error="failed to setup network for sandbox \"5dd1dc9316e1d4cb2cc982113433992b2f865de8097e02c3d7f196ab0750361e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.802384 kubelet[2575]: E0213 19:51:27.802340 2575 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5dd1dc9316e1d4cb2cc982113433992b2f865de8097e02c3d7f196ab0750361e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.802562 kubelet[2575]: E0213 19:51:27.802540 2575 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5dd1dc9316e1d4cb2cc982113433992b2f865de8097e02c3d7f196ab0750361e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-575f8f688d-rm27z" Feb 13 19:51:27.802649 kubelet[2575]: E0213 19:51:27.802631 2575 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5dd1dc9316e1d4cb2cc982113433992b2f865de8097e02c3d7f196ab0750361e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-575f8f688d-rm27z" Feb 13 19:51:27.802832 kubelet[2575]: E0213 19:51:27.802758 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-575f8f688d-rm27z_calico-apiserver(dca83c7e-397c-4cb7-b95d-96c1c013b67f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-575f8f688d-rm27z_calico-apiserver(dca83c7e-397c-4cb7-b95d-96c1c013b67f)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"5dd1dc9316e1d4cb2cc982113433992b2f865de8097e02c3d7f196ab0750361e\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-575f8f688d-rm27z" podUID="dca83c7e-397c-4cb7-b95d-96c1c013b67f" Feb 13 19:51:27.833393 sshd[4443]: Connection closed by 10.0.0.1 port 57054 Feb 13 19:51:27.834176 sshd-session[4441]: pam_unix(sshd:session): session closed for user core Feb 13 19:51:27.834935 containerd[1468]: time="2025-02-13T19:51:27.834748989Z" level=error msg="Failed to destroy network for sandbox \"725f264f85d2c0cd16ccc9f133fde5df46c364a1fdc97ca60f100ccf9eecc7c3\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.836417 containerd[1468]: time="2025-02-13T19:51:27.836360755Z" level=error msg="encountered an error cleaning up failed sandbox \"725f264f85d2c0cd16ccc9f133fde5df46c364a1fdc97ca60f100ccf9eecc7c3\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.839950 containerd[1468]: time="2025-02-13T19:51:27.836485068Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-s22xp,Uid:bd8b42bf-05e5-489c-87c7-f5984836cd6a,Namespace:kube-system,Attempt:5,} failed, error" error="failed to setup network for sandbox \"725f264f85d2c0cd16ccc9f133fde5df46c364a1fdc97ca60f100ccf9eecc7c3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.840012 kubelet[2575]: E0213 19:51:27.839088 2575 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"725f264f85d2c0cd16ccc9f133fde5df46c364a1fdc97ca60f100ccf9eecc7c3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.840012 kubelet[2575]: E0213 19:51:27.839180 2575 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"725f264f85d2c0cd16ccc9f133fde5df46c364a1fdc97ca60f100ccf9eecc7c3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-s22xp" Feb 13 19:51:27.840012 kubelet[2575]: E0213 19:51:27.839210 2575 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"725f264f85d2c0cd16ccc9f133fde5df46c364a1fdc97ca60f100ccf9eecc7c3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-s22xp" Feb 13 19:51:27.840125 kubelet[2575]: E0213 19:51:27.839266 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-s22xp_kube-system(bd8b42bf-05e5-489c-87c7-f5984836cd6a)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-s22xp_kube-system(bd8b42bf-05e5-489c-87c7-f5984836cd6a)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"725f264f85d2c0cd16ccc9f133fde5df46c364a1fdc97ca60f100ccf9eecc7c3\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-s22xp" podUID="bd8b42bf-05e5-489c-87c7-f5984836cd6a" Feb 13 19:51:27.841722 systemd[1]: sshd@8-10.0.0.104:22-10.0.0.1:57054.service: Deactivated successfully. Feb 13 19:51:27.846599 containerd[1468]: time="2025-02-13T19:51:27.845894763Z" level=error msg="Failed to destroy network for sandbox \"07cbe351afb62e0b27b5d8a863b7989470635218b3f4519289a79adcd80bb66e\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.846599 containerd[1468]: time="2025-02-13T19:51:27.846576193Z" level=error msg="encountered an error cleaning up failed sandbox \"07cbe351afb62e0b27b5d8a863b7989470635218b3f4519289a79adcd80bb66e\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.846776 containerd[1468]: time="2025-02-13T19:51:27.846652085Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5bf7d587cb-s28ls,Uid:77a9c264-ba41-4ec0-b704-771924c99471,Namespace:calico-system,Attempt:5,} failed, error" error="failed to setup network for sandbox \"07cbe351afb62e0b27b5d8a863b7989470635218b3f4519289a79adcd80bb66e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.847373 kubelet[2575]: E0213 19:51:27.846957 2575 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"07cbe351afb62e0b27b5d8a863b7989470635218b3f4519289a79adcd80bb66e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.847373 kubelet[2575]: E0213 19:51:27.847049 2575 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"07cbe351afb62e0b27b5d8a863b7989470635218b3f4519289a79adcd80bb66e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5bf7d587cb-s28ls" Feb 13 19:51:27.847373 kubelet[2575]: E0213 19:51:27.847081 2575 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"07cbe351afb62e0b27b5d8a863b7989470635218b3f4519289a79adcd80bb66e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5bf7d587cb-s28ls" Feb 13 19:51:27.847137 systemd[1]: session-8.scope: Deactivated successfully. Feb 13 19:51:27.847983 kubelet[2575]: E0213 19:51:27.847149 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-5bf7d587cb-s28ls_calico-system(77a9c264-ba41-4ec0-b704-771924c99471)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-5bf7d587cb-s28ls_calico-system(77a9c264-ba41-4ec0-b704-771924c99471)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"07cbe351afb62e0b27b5d8a863b7989470635218b3f4519289a79adcd80bb66e\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-5bf7d587cb-s28ls" podUID="77a9c264-ba41-4ec0-b704-771924c99471" Feb 13 19:51:27.850490 systemd-logind[1447]: Session 8 logged out. Waiting for processes to exit. Feb 13 19:51:27.852511 systemd-logind[1447]: Removed session 8. Feb 13 19:51:27.856043 containerd[1468]: time="2025-02-13T19:51:27.855852799Z" level=error msg="Failed to destroy network for sandbox \"b65733053a47537c9f659bcb260db972084c3f0ccd2ff597a9805d0ba86debdc\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.856599 containerd[1468]: time="2025-02-13T19:51:27.856538475Z" level=error msg="encountered an error cleaning up failed sandbox \"b65733053a47537c9f659bcb260db972084c3f0ccd2ff597a9805d0ba86debdc\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.856669 containerd[1468]: time="2025-02-13T19:51:27.856612675Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-575f8f688d-8855n,Uid:a5e91fd3-19fa-490e-be13-26d807029222,Namespace:calico-apiserver,Attempt:5,} failed, error" error="failed to setup network for sandbox \"b65733053a47537c9f659bcb260db972084c3f0ccd2ff597a9805d0ba86debdc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.856942 kubelet[2575]: E0213 19:51:27.856857 2575 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b65733053a47537c9f659bcb260db972084c3f0ccd2ff597a9805d0ba86debdc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:51:27.857028 kubelet[2575]: E0213 19:51:27.856960 2575 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b65733053a47537c9f659bcb260db972084c3f0ccd2ff597a9805d0ba86debdc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-575f8f688d-8855n" Feb 13 19:51:27.857028 kubelet[2575]: E0213 19:51:27.856992 2575 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b65733053a47537c9f659bcb260db972084c3f0ccd2ff597a9805d0ba86debdc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-575f8f688d-8855n" Feb 13 19:51:27.857095 kubelet[2575]: E0213 19:51:27.857046 2575 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-575f8f688d-8855n_calico-apiserver(a5e91fd3-19fa-490e-be13-26d807029222)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-575f8f688d-8855n_calico-apiserver(a5e91fd3-19fa-490e-be13-26d807029222)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"b65733053a47537c9f659bcb260db972084c3f0ccd2ff597a9805d0ba86debdc\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-575f8f688d-8855n" podUID="a5e91fd3-19fa-490e-be13-26d807029222" Feb 13 19:51:27.933978 containerd[1468]: time="2025-02-13T19:51:27.933841640Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:51:27.935519 containerd[1468]: time="2025-02-13T19:51:27.935375950Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.29.1: active requests=0, bytes read=142742010" Feb 13 19:51:27.936858 containerd[1468]: time="2025-02-13T19:51:27.936793061Z" level=info msg="ImageCreate event name:\"sha256:feb26d4585d68e875d9bd9bd6c27ea9f2d5c9ed9ef70f8b8cb0ebb0559a1d664\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:51:27.940532 containerd[1468]: time="2025-02-13T19:51:27.940430550Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:99c3917516efe1f807a0cfdf2d14b628b7c5cc6bd8a9ee5a253154f31756bea1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:51:27.941291 containerd[1468]: time="2025-02-13T19:51:27.941242193Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.29.1\" with image id \"sha256:feb26d4585d68e875d9bd9bd6c27ea9f2d5c9ed9ef70f8b8cb0ebb0559a1d664\", repo tag \"ghcr.io/flatcar/calico/node:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/node@sha256:99c3917516efe1f807a0cfdf2d14b628b7c5cc6bd8a9ee5a253154f31756bea1\", size \"142741872\" in 7.779327167s" Feb 13 19:51:27.941291 containerd[1468]: time="2025-02-13T19:51:27.941281567Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.1\" returns image reference \"sha256:feb26d4585d68e875d9bd9bd6c27ea9f2d5c9ed9ef70f8b8cb0ebb0559a1d664\"" Feb 13 19:51:27.950918 containerd[1468]: time="2025-02-13T19:51:27.950842696Z" level=info msg="CreateContainer within sandbox \"f9006b0c7095feabafe4193f0ce61148f5b8f4c8d7f93c8d390cefc8f6f6014f\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Feb 13 19:51:27.982941 containerd[1468]: time="2025-02-13T19:51:27.982840727Z" level=info msg="CreateContainer within sandbox \"f9006b0c7095feabafe4193f0ce61148f5b8f4c8d7f93c8d390cefc8f6f6014f\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"44b54026f546b44a6a62c8d474424c2abb287d0b0bdefe45730881d00c21e537\"" Feb 13 19:51:27.984079 containerd[1468]: time="2025-02-13T19:51:27.983704679Z" level=info msg="StartContainer for \"44b54026f546b44a6a62c8d474424c2abb287d0b0bdefe45730881d00c21e537\"" Feb 13 19:51:28.077254 systemd[1]: Started cri-containerd-44b54026f546b44a6a62c8d474424c2abb287d0b0bdefe45730881d00c21e537.scope - libcontainer container 44b54026f546b44a6a62c8d474424c2abb287d0b0bdefe45730881d00c21e537. Feb 13 19:51:28.116880 systemd[1]: run-netns-cni\x2d11095693\x2d1489\x2dce37\x2d6bf1\x2d06e5cd7935e5.mount: Deactivated successfully. Feb 13 19:51:28.117054 systemd[1]: run-netns-cni\x2dc435e0c2\x2dca91\x2d1571\x2dab71\x2d6f736c6c8fd0.mount: Deactivated successfully. Feb 13 19:51:28.117141 systemd[1]: run-netns-cni\x2d53889f61\x2d7582\x2d2ab4\x2dba4e\x2dc03f5271d48b.mount: Deactivated successfully. Feb 13 19:51:28.117224 systemd[1]: run-netns-cni\x2df9008414\x2dc39b\x2d4a9e\x2dcfe0\x2db4f03e3f8557.mount: Deactivated successfully. Feb 13 19:51:28.117310 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-d52193e0d0af1f395378fbf458da41b44ce26d784835ca82240d3687ed53ca6f-shm.mount: Deactivated successfully. Feb 13 19:51:28.117481 systemd[1]: run-netns-cni\x2dc3702f63\x2dff8d\x2d9fba\x2ddde0\x2d9afd5e7dc7ca.mount: Deactivated successfully. Feb 13 19:51:28.117644 systemd[1]: run-netns-cni\x2d1b2f9ff5\x2d17e8\x2dc4dd\x2dd30d\x2db8d0dd954570.mount: Deactivated successfully. Feb 13 19:51:28.117795 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2308271405.mount: Deactivated successfully. Feb 13 19:51:28.137291 containerd[1468]: time="2025-02-13T19:51:28.137123900Z" level=info msg="StartContainer for \"44b54026f546b44a6a62c8d474424c2abb287d0b0bdefe45730881d00c21e537\" returns successfully" Feb 13 19:51:28.242919 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Feb 13 19:51:28.243184 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Feb 13 19:51:28.323172 kubelet[2575]: I0213 19:51:28.323104 2575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5dd1dc9316e1d4cb2cc982113433992b2f865de8097e02c3d7f196ab0750361e" Feb 13 19:51:28.325088 containerd[1468]: time="2025-02-13T19:51:28.325034344Z" level=info msg="StopPodSandbox for \"5dd1dc9316e1d4cb2cc982113433992b2f865de8097e02c3d7f196ab0750361e\"" Feb 13 19:51:28.325379 containerd[1468]: time="2025-02-13T19:51:28.325353634Z" level=info msg="Ensure that sandbox 5dd1dc9316e1d4cb2cc982113433992b2f865de8097e02c3d7f196ab0750361e in task-service has been cleanup successfully" Feb 13 19:51:28.329179 containerd[1468]: time="2025-02-13T19:51:28.329023232Z" level=info msg="TearDown network for sandbox \"5dd1dc9316e1d4cb2cc982113433992b2f865de8097e02c3d7f196ab0750361e\" successfully" Feb 13 19:51:28.329179 containerd[1468]: time="2025-02-13T19:51:28.329072695Z" level=info msg="StopPodSandbox for \"5dd1dc9316e1d4cb2cc982113433992b2f865de8097e02c3d7f196ab0750361e\" returns successfully" Feb 13 19:51:28.330454 containerd[1468]: time="2025-02-13T19:51:28.330390148Z" level=info msg="StopPodSandbox for \"d42fa1fe42814b711ad3cb3d471c4742bf3051f8d634e2db762b56c01dc00d5a\"" Feb 13 19:51:28.330571 containerd[1468]: time="2025-02-13T19:51:28.330545369Z" level=info msg="TearDown network for sandbox \"d42fa1fe42814b711ad3cb3d471c4742bf3051f8d634e2db762b56c01dc00d5a\" successfully" Feb 13 19:51:28.330571 containerd[1468]: time="2025-02-13T19:51:28.330564866Z" level=info msg="StopPodSandbox for \"d42fa1fe42814b711ad3cb3d471c4742bf3051f8d634e2db762b56c01dc00d5a\" returns successfully" Feb 13 19:51:28.335202 containerd[1468]: time="2025-02-13T19:51:28.335145094Z" level=info msg="StopPodSandbox for \"3d79c24a7d960cef54edf513bffc4dd8a68cbc23ca3e3304eafaabcbc4d88e33\"" Feb 13 19:51:28.335385 containerd[1468]: time="2025-02-13T19:51:28.335296498Z" level=info msg="TearDown network for sandbox \"3d79c24a7d960cef54edf513bffc4dd8a68cbc23ca3e3304eafaabcbc4d88e33\" successfully" Feb 13 19:51:28.335385 containerd[1468]: time="2025-02-13T19:51:28.335307619Z" level=info msg="StopPodSandbox for \"3d79c24a7d960cef54edf513bffc4dd8a68cbc23ca3e3304eafaabcbc4d88e33\" returns successfully" Feb 13 19:51:28.336036 containerd[1468]: time="2025-02-13T19:51:28.335979079Z" level=info msg="StopPodSandbox for \"5f114ecd2418a98f886f02a3a5637918502bf71cd1bf3c74f6c1dd0e45034e21\"" Feb 13 19:51:28.336324 systemd[1]: run-netns-cni\x2dbf2d3b26\x2d6d24\x2dff4e\x2d3868\x2d906703b9ed09.mount: Deactivated successfully. Feb 13 19:51:28.337344 containerd[1468]: time="2025-02-13T19:51:28.336334336Z" level=info msg="TearDown network for sandbox \"5f114ecd2418a98f886f02a3a5637918502bf71cd1bf3c74f6c1dd0e45034e21\" successfully" Feb 13 19:51:28.337344 containerd[1468]: time="2025-02-13T19:51:28.336350707Z" level=info msg="StopPodSandbox for \"5f114ecd2418a98f886f02a3a5637918502bf71cd1bf3c74f6c1dd0e45034e21\" returns successfully" Feb 13 19:51:28.340035 containerd[1468]: time="2025-02-13T19:51:28.339968819Z" level=info msg="StopPodSandbox for \"3af9c99589cb59a86666f237cc1b2ea19ad1577ee4287478e22ace117f62bf47\"" Feb 13 19:51:28.340271 containerd[1468]: time="2025-02-13T19:51:28.340134600Z" level=info msg="TearDown network for sandbox \"3af9c99589cb59a86666f237cc1b2ea19ad1577ee4287478e22ace117f62bf47\" successfully" Feb 13 19:51:28.340271 containerd[1468]: time="2025-02-13T19:51:28.340159647Z" level=info msg="StopPodSandbox for \"3af9c99589cb59a86666f237cc1b2ea19ad1577ee4287478e22ace117f62bf47\" returns successfully" Feb 13 19:51:28.340678 containerd[1468]: time="2025-02-13T19:51:28.340654395Z" level=info msg="StopPodSandbox for \"4ff7a1ce33e60be86627d63cbaa39e854dfba3ccfcd25d3b18dc72196a2b96da\"" Feb 13 19:51:28.342894 containerd[1468]: time="2025-02-13T19:51:28.340972162Z" level=info msg="TearDown network for sandbox \"4ff7a1ce33e60be86627d63cbaa39e854dfba3ccfcd25d3b18dc72196a2b96da\" successfully" Feb 13 19:51:28.342894 containerd[1468]: time="2025-02-13T19:51:28.340994093Z" level=info msg="StopPodSandbox for \"4ff7a1ce33e60be86627d63cbaa39e854dfba3ccfcd25d3b18dc72196a2b96da\" returns successfully" Feb 13 19:51:28.344034 containerd[1468]: time="2025-02-13T19:51:28.344006479Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-575f8f688d-rm27z,Uid:dca83c7e-397c-4cb7-b95d-96c1c013b67f,Namespace:calico-apiserver,Attempt:6,}" Feb 13 19:51:28.345591 kubelet[2575]: I0213 19:51:28.345454 2575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b65733053a47537c9f659bcb260db972084c3f0ccd2ff597a9805d0ba86debdc" Feb 13 19:51:28.351821 containerd[1468]: time="2025-02-13T19:51:28.351267389Z" level=info msg="StopPodSandbox for \"b65733053a47537c9f659bcb260db972084c3f0ccd2ff597a9805d0ba86debdc\"" Feb 13 19:51:28.351821 containerd[1468]: time="2025-02-13T19:51:28.351582009Z" level=info msg="Ensure that sandbox b65733053a47537c9f659bcb260db972084c3f0ccd2ff597a9805d0ba86debdc in task-service has been cleanup successfully" Feb 13 19:51:28.355210 containerd[1468]: time="2025-02-13T19:51:28.355180814Z" level=info msg="TearDown network for sandbox \"b65733053a47537c9f659bcb260db972084c3f0ccd2ff597a9805d0ba86debdc\" successfully" Feb 13 19:51:28.355314 containerd[1468]: time="2025-02-13T19:51:28.355301782Z" level=info msg="StopPodSandbox for \"b65733053a47537c9f659bcb260db972084c3f0ccd2ff597a9805d0ba86debdc\" returns successfully" Feb 13 19:51:28.356252 containerd[1468]: time="2025-02-13T19:51:28.356234904Z" level=info msg="StopPodSandbox for \"39ec85c7a0b06b4fc75b83977eee38fcc3ba3782a02a024285436451c00bfc91\"" Feb 13 19:51:28.356401 containerd[1468]: time="2025-02-13T19:51:28.356386658Z" level=info msg="TearDown network for sandbox \"39ec85c7a0b06b4fc75b83977eee38fcc3ba3782a02a024285436451c00bfc91\" successfully" Feb 13 19:51:28.356450 systemd[1]: run-netns-cni\x2d731ad7b7\x2df217\x2d91dc\x2d2681\x2d3824810c250d.mount: Deactivated successfully. Feb 13 19:51:28.357102 containerd[1468]: time="2025-02-13T19:51:28.357085671Z" level=info msg="StopPodSandbox for \"39ec85c7a0b06b4fc75b83977eee38fcc3ba3782a02a024285436451c00bfc91\" returns successfully" Feb 13 19:51:28.357864 kubelet[2575]: E0213 19:51:28.357838 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:28.360612 containerd[1468]: time="2025-02-13T19:51:28.358035925Z" level=info msg="StopPodSandbox for \"ebbbb0cee651304a47f8384965c605db95eabdf650756e32c6b70a0355fbec6d\"" Feb 13 19:51:28.360612 containerd[1468]: time="2025-02-13T19:51:28.358124962Z" level=info msg="TearDown network for sandbox \"ebbbb0cee651304a47f8384965c605db95eabdf650756e32c6b70a0355fbec6d\" successfully" Feb 13 19:51:28.360612 containerd[1468]: time="2025-02-13T19:51:28.358135802Z" level=info msg="StopPodSandbox for \"ebbbb0cee651304a47f8384965c605db95eabdf650756e32c6b70a0355fbec6d\" returns successfully" Feb 13 19:51:28.362305 containerd[1468]: time="2025-02-13T19:51:28.361625343Z" level=info msg="StopPodSandbox for \"269c22f856e2916019bfab8db86934d7f53640411cbfd7537abae3e996450f7b\"" Feb 13 19:51:28.362305 containerd[1468]: time="2025-02-13T19:51:28.361780303Z" level=info msg="TearDown network for sandbox \"269c22f856e2916019bfab8db86934d7f53640411cbfd7537abae3e996450f7b\" successfully" Feb 13 19:51:28.362305 containerd[1468]: time="2025-02-13T19:51:28.361796033Z" level=info msg="StopPodSandbox for \"269c22f856e2916019bfab8db86934d7f53640411cbfd7537abae3e996450f7b\" returns successfully" Feb 13 19:51:28.363324 containerd[1468]: time="2025-02-13T19:51:28.363300777Z" level=info msg="StopPodSandbox for \"38f72aa5f0a5062621013b365f7d00faaabff6dd65a3f268813b6c2c1534a463\"" Feb 13 19:51:28.363543 containerd[1468]: time="2025-02-13T19:51:28.363520791Z" level=info msg="TearDown network for sandbox \"38f72aa5f0a5062621013b365f7d00faaabff6dd65a3f268813b6c2c1534a463\" successfully" Feb 13 19:51:28.363765 containerd[1468]: time="2025-02-13T19:51:28.363678296Z" level=info msg="StopPodSandbox for \"38f72aa5f0a5062621013b365f7d00faaabff6dd65a3f268813b6c2c1534a463\" returns successfully" Feb 13 19:51:28.364315 containerd[1468]: time="2025-02-13T19:51:28.364222708Z" level=info msg="StopPodSandbox for \"517d17b8f015a9521e2907d73d8ffe6db808ece1034c7f320e60f66e57f87de3\"" Feb 13 19:51:28.366608 containerd[1468]: time="2025-02-13T19:51:28.364605897Z" level=info msg="TearDown network for sandbox \"517d17b8f015a9521e2907d73d8ffe6db808ece1034c7f320e60f66e57f87de3\" successfully" Feb 13 19:51:28.366905 containerd[1468]: time="2025-02-13T19:51:28.366645116Z" level=info msg="StopPodSandbox for \"517d17b8f015a9521e2907d73d8ffe6db808ece1034c7f320e60f66e57f87de3\" returns successfully" Feb 13 19:51:28.368570 containerd[1468]: time="2025-02-13T19:51:28.368273232Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-575f8f688d-8855n,Uid:a5e91fd3-19fa-490e-be13-26d807029222,Namespace:calico-apiserver,Attempt:6,}" Feb 13 19:51:28.369348 kubelet[2575]: I0213 19:51:28.369311 2575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07cbe351afb62e0b27b5d8a863b7989470635218b3f4519289a79adcd80bb66e" Feb 13 19:51:28.370216 containerd[1468]: time="2025-02-13T19:51:28.370173388Z" level=info msg="StopPodSandbox for \"07cbe351afb62e0b27b5d8a863b7989470635218b3f4519289a79adcd80bb66e\"" Feb 13 19:51:28.370504 containerd[1468]: time="2025-02-13T19:51:28.370477610Z" level=info msg="Ensure that sandbox 07cbe351afb62e0b27b5d8a863b7989470635218b3f4519289a79adcd80bb66e in task-service has been cleanup successfully" Feb 13 19:51:28.370839 containerd[1468]: time="2025-02-13T19:51:28.370791810Z" level=info msg="TearDown network for sandbox \"07cbe351afb62e0b27b5d8a863b7989470635218b3f4519289a79adcd80bb66e\" successfully" Feb 13 19:51:28.370839 containerd[1468]: time="2025-02-13T19:51:28.370819241Z" level=info msg="StopPodSandbox for \"07cbe351afb62e0b27b5d8a863b7989470635218b3f4519289a79adcd80bb66e\" returns successfully" Feb 13 19:51:28.372435 containerd[1468]: time="2025-02-13T19:51:28.371961044Z" level=info msg="StopPodSandbox for \"ed33748ae2f78ecf5ee2c805ac620f4f19dbb1b895af40e625d52cc88bec3832\"" Feb 13 19:51:28.372435 containerd[1468]: time="2025-02-13T19:51:28.372117960Z" level=info msg="TearDown network for sandbox \"ed33748ae2f78ecf5ee2c805ac620f4f19dbb1b895af40e625d52cc88bec3832\" successfully" Feb 13 19:51:28.372435 containerd[1468]: time="2025-02-13T19:51:28.372134220Z" level=info msg="StopPodSandbox for \"ed33748ae2f78ecf5ee2c805ac620f4f19dbb1b895af40e625d52cc88bec3832\" returns successfully" Feb 13 19:51:28.373259 containerd[1468]: time="2025-02-13T19:51:28.373161608Z" level=info msg="StopPodSandbox for \"339897f85c32bf6b58fdb7fa27b7f977f70adf895ce4769419e9e19adb77c464\"" Feb 13 19:51:28.373360 containerd[1468]: time="2025-02-13T19:51:28.373307612Z" level=info msg="TearDown network for sandbox \"339897f85c32bf6b58fdb7fa27b7f977f70adf895ce4769419e9e19adb77c464\" successfully" Feb 13 19:51:28.373360 containerd[1468]: time="2025-02-13T19:51:28.373340163Z" level=info msg="StopPodSandbox for \"339897f85c32bf6b58fdb7fa27b7f977f70adf895ce4769419e9e19adb77c464\" returns successfully" Feb 13 19:51:28.374181 containerd[1468]: time="2025-02-13T19:51:28.373782824Z" level=info msg="StopPodSandbox for \"deadb290c2eabe3a6b5ed648ceb2f5121ab56b3728b197df70e6532dc828cd25\"" Feb 13 19:51:28.374181 containerd[1468]: time="2025-02-13T19:51:28.373924841Z" level=info msg="TearDown network for sandbox \"deadb290c2eabe3a6b5ed648ceb2f5121ab56b3728b197df70e6532dc828cd25\" successfully" Feb 13 19:51:28.374181 containerd[1468]: time="2025-02-13T19:51:28.373937144Z" level=info msg="StopPodSandbox for \"deadb290c2eabe3a6b5ed648ceb2f5121ab56b3728b197df70e6532dc828cd25\" returns successfully" Feb 13 19:51:28.374912 containerd[1468]: time="2025-02-13T19:51:28.374890234Z" level=info msg="StopPodSandbox for \"cc0fab9693293af68d0473a6799ea1e188b038744dcf56927b1d4fffd60c2c96\"" Feb 13 19:51:28.375151 containerd[1468]: time="2025-02-13T19:51:28.375109114Z" level=info msg="TearDown network for sandbox \"cc0fab9693293af68d0473a6799ea1e188b038744dcf56927b1d4fffd60c2c96\" successfully" Feb 13 19:51:28.375459 containerd[1468]: time="2025-02-13T19:51:28.375387397Z" level=info msg="StopPodSandbox for \"cc0fab9693293af68d0473a6799ea1e188b038744dcf56927b1d4fffd60c2c96\" returns successfully" Feb 13 19:51:28.375701 systemd[1]: run-netns-cni\x2d9bc456bb\x2d050a\x2dd3b8\x2da17c\x2d73a37865440c.mount: Deactivated successfully. Feb 13 19:51:28.377073 containerd[1468]: time="2025-02-13T19:51:28.376768800Z" level=info msg="StopPodSandbox for \"c143929d3d37d0fd39eeb807647ccdb8bf35b3017d21a93a728ef6fa513987e4\"" Feb 13 19:51:28.377073 containerd[1468]: time="2025-02-13T19:51:28.376996988Z" level=info msg="TearDown network for sandbox \"c143929d3d37d0fd39eeb807647ccdb8bf35b3017d21a93a728ef6fa513987e4\" successfully" Feb 13 19:51:28.377073 containerd[1468]: time="2025-02-13T19:51:28.377012697Z" level=info msg="StopPodSandbox for \"c143929d3d37d0fd39eeb807647ccdb8bf35b3017d21a93a728ef6fa513987e4\" returns successfully" Feb 13 19:51:28.379711 containerd[1468]: time="2025-02-13T19:51:28.379672701Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5bf7d587cb-s28ls,Uid:77a9c264-ba41-4ec0-b704-771924c99471,Namespace:calico-system,Attempt:6,}" Feb 13 19:51:28.381909 kubelet[2575]: I0213 19:51:28.381841 2575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d976eb55d0e17adfbd3d8051a2fdf3026e7061589ae2a3f34816e887a92e46f" Feb 13 19:51:28.388297 containerd[1468]: time="2025-02-13T19:51:28.385775648Z" level=info msg="StopPodSandbox for \"1d976eb55d0e17adfbd3d8051a2fdf3026e7061589ae2a3f34816e887a92e46f\"" Feb 13 19:51:28.388297 containerd[1468]: time="2025-02-13T19:51:28.386117959Z" level=info msg="Ensure that sandbox 1d976eb55d0e17adfbd3d8051a2fdf3026e7061589ae2a3f34816e887a92e46f in task-service has been cleanup successfully" Feb 13 19:51:28.389749 containerd[1468]: time="2025-02-13T19:51:28.389693331Z" level=info msg="TearDown network for sandbox \"1d976eb55d0e17adfbd3d8051a2fdf3026e7061589ae2a3f34816e887a92e46f\" successfully" Feb 13 19:51:28.389955 containerd[1468]: time="2025-02-13T19:51:28.389934464Z" level=info msg="StopPodSandbox for \"1d976eb55d0e17adfbd3d8051a2fdf3026e7061589ae2a3f34816e887a92e46f\" returns successfully" Feb 13 19:51:28.390046 systemd[1]: run-netns-cni\x2d5e5babc0\x2df1d7\x2de4f0\x2d6aca\x2d32b6b3f4309f.mount: Deactivated successfully. Feb 13 19:51:28.390962 containerd[1468]: time="2025-02-13T19:51:28.390463187Z" level=info msg="StopPodSandbox for \"d52193e0d0af1f395378fbf458da41b44ce26d784835ca82240d3687ed53ca6f\"" Feb 13 19:51:28.390962 containerd[1468]: time="2025-02-13T19:51:28.390680715Z" level=info msg="TearDown network for sandbox \"d52193e0d0af1f395378fbf458da41b44ce26d784835ca82240d3687ed53ca6f\" successfully" Feb 13 19:51:28.390962 containerd[1468]: time="2025-02-13T19:51:28.390695783Z" level=info msg="StopPodSandbox for \"d52193e0d0af1f395378fbf458da41b44ce26d784835ca82240d3687ed53ca6f\" returns successfully" Feb 13 19:51:28.392562 kubelet[2575]: I0213 19:51:28.392216 2575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="729c6e6d3b4530ba0b68738b226a62687ecd65bb8522ed1b880437d5541be3c3" Feb 13 19:51:28.394012 containerd[1468]: time="2025-02-13T19:51:28.393959120Z" level=info msg="StopPodSandbox for \"729c6e6d3b4530ba0b68738b226a62687ecd65bb8522ed1b880437d5541be3c3\"" Feb 13 19:51:28.394309 containerd[1468]: time="2025-02-13T19:51:28.394273669Z" level=info msg="Ensure that sandbox 729c6e6d3b4530ba0b68738b226a62687ecd65bb8522ed1b880437d5541be3c3 in task-service has been cleanup successfully" Feb 13 19:51:28.394950 containerd[1468]: time="2025-02-13T19:51:28.394910666Z" level=info msg="StopPodSandbox for \"20a630c64d17851512c12fc278a3718b46879b27d0ffedf611d9175c93614992\"" Feb 13 19:51:28.395058 containerd[1468]: time="2025-02-13T19:51:28.395029549Z" level=info msg="TearDown network for sandbox \"20a630c64d17851512c12fc278a3718b46879b27d0ffedf611d9175c93614992\" successfully" Feb 13 19:51:28.395058 containerd[1468]: time="2025-02-13T19:51:28.395051119Z" level=info msg="StopPodSandbox for \"20a630c64d17851512c12fc278a3718b46879b27d0ffedf611d9175c93614992\" returns successfully" Feb 13 19:51:28.396332 containerd[1468]: time="2025-02-13T19:51:28.396051347Z" level=info msg="StopPodSandbox for \"49835f65aea5736de7af75d5477e71209d3ab4edb28dc78ae778b60bcf5f2e76\"" Feb 13 19:51:28.396332 containerd[1468]: time="2025-02-13T19:51:28.396140834Z" level=info msg="TearDown network for sandbox \"49835f65aea5736de7af75d5477e71209d3ab4edb28dc78ae778b60bcf5f2e76\" successfully" Feb 13 19:51:28.396332 containerd[1468]: time="2025-02-13T19:51:28.396154670Z" level=info msg="StopPodSandbox for \"49835f65aea5736de7af75d5477e71209d3ab4edb28dc78ae778b60bcf5f2e76\" returns successfully" Feb 13 19:51:28.397233 containerd[1468]: time="2025-02-13T19:51:28.397175768Z" level=info msg="TearDown network for sandbox \"729c6e6d3b4530ba0b68738b226a62687ecd65bb8522ed1b880437d5541be3c3\" successfully" Feb 13 19:51:28.397409 containerd[1468]: time="2025-02-13T19:51:28.397393877Z" level=info msg="StopPodSandbox for \"729c6e6d3b4530ba0b68738b226a62687ecd65bb8522ed1b880437d5541be3c3\" returns successfully" Feb 13 19:51:28.399057 containerd[1468]: time="2025-02-13T19:51:28.398099071Z" level=info msg="StopPodSandbox for \"dee4103e3ed5a58df193e21973191414daa33c176df2ebb53f48f6427c812139\"" Feb 13 19:51:28.399450 containerd[1468]: time="2025-02-13T19:51:28.399415903Z" level=info msg="TearDown network for sandbox \"dee4103e3ed5a58df193e21973191414daa33c176df2ebb53f48f6427c812139\" successfully" Feb 13 19:51:28.399517 containerd[1468]: time="2025-02-13T19:51:28.399505130Z" level=info msg="StopPodSandbox for \"dee4103e3ed5a58df193e21973191414daa33c176df2ebb53f48f6427c812139\" returns successfully" Feb 13 19:51:28.400315 containerd[1468]: time="2025-02-13T19:51:28.399112343Z" level=info msg="StopPodSandbox for \"785b364a151170b1b8e2dc79f0469cae0c56fbd88d84bde56a96da9c32fc64d7\"" Feb 13 19:51:28.400315 containerd[1468]: time="2025-02-13T19:51:28.399815081Z" level=info msg="TearDown network for sandbox \"785b364a151170b1b8e2dc79f0469cae0c56fbd88d84bde56a96da9c32fc64d7\" successfully" Feb 13 19:51:28.400315 containerd[1468]: time="2025-02-13T19:51:28.399826112Z" level=info msg="StopPodSandbox for \"785b364a151170b1b8e2dc79f0469cae0c56fbd88d84bde56a96da9c32fc64d7\" returns successfully" Feb 13 19:51:28.400604 containerd[1468]: time="2025-02-13T19:51:28.400338204Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-sxwqd,Uid:bbd75eeb-e06b-4812-b2c6-522804c5ef2f,Namespace:calico-system,Attempt:5,}" Feb 13 19:51:28.401116 containerd[1468]: time="2025-02-13T19:51:28.401089254Z" level=info msg="StopPodSandbox for \"22c8c985c21b7d2634c18c69f2f75ef4a82d17959955a493172581323c264379\"" Feb 13 19:51:28.401203 containerd[1468]: time="2025-02-13T19:51:28.401177850Z" level=info msg="TearDown network for sandbox \"22c8c985c21b7d2634c18c69f2f75ef4a82d17959955a493172581323c264379\" successfully" Feb 13 19:51:28.401203 containerd[1468]: time="2025-02-13T19:51:28.401197436Z" level=info msg="StopPodSandbox for \"22c8c985c21b7d2634c18c69f2f75ef4a82d17959955a493172581323c264379\" returns successfully" Feb 13 19:51:28.401592 containerd[1468]: time="2025-02-13T19:51:28.401571198Z" level=info msg="StopPodSandbox for \"13004b615f48f8fc71da84226292b2e4c272d980ea5cce23fb82164b712c2e30\"" Feb 13 19:51:28.401778 containerd[1468]: time="2025-02-13T19:51:28.401717784Z" level=info msg="TearDown network for sandbox \"13004b615f48f8fc71da84226292b2e4c272d980ea5cce23fb82164b712c2e30\" successfully" Feb 13 19:51:28.401778 containerd[1468]: time="2025-02-13T19:51:28.401731319Z" level=info msg="StopPodSandbox for \"13004b615f48f8fc71da84226292b2e4c272d980ea5cce23fb82164b712c2e30\" returns successfully" Feb 13 19:51:28.402406 containerd[1468]: time="2025-02-13T19:51:28.402008319Z" level=info msg="StopPodSandbox for \"899fcc18f7bfce71348bcd81eba1c4cca6504cfc06183532c64985ae4deedeb3\"" Feb 13 19:51:28.402406 containerd[1468]: time="2025-02-13T19:51:28.402084822Z" level=info msg="TearDown network for sandbox \"899fcc18f7bfce71348bcd81eba1c4cca6504cfc06183532c64985ae4deedeb3\" successfully" Feb 13 19:51:28.402406 containerd[1468]: time="2025-02-13T19:51:28.402096123Z" level=info msg="StopPodSandbox for \"899fcc18f7bfce71348bcd81eba1c4cca6504cfc06183532c64985ae4deedeb3\" returns successfully" Feb 13 19:51:28.402406 containerd[1468]: time="2025-02-13T19:51:28.402264881Z" level=info msg="StopPodSandbox for \"a24dd5ac08e1be65e64f1e0e1bcc57fba990025660893d49024096981b2ba3aa\"" Feb 13 19:51:28.402406 containerd[1468]: time="2025-02-13T19:51:28.402351613Z" level=info msg="TearDown network for sandbox \"a24dd5ac08e1be65e64f1e0e1bcc57fba990025660893d49024096981b2ba3aa\" successfully" Feb 13 19:51:28.402406 containerd[1468]: time="2025-02-13T19:51:28.402361662Z" level=info msg="StopPodSandbox for \"a24dd5ac08e1be65e64f1e0e1bcc57fba990025660893d49024096981b2ba3aa\" returns successfully" Feb 13 19:51:28.402578 kubelet[2575]: I0213 19:51:28.402332 2575 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="725f264f85d2c0cd16ccc9f133fde5df46c364a1fdc97ca60f100ccf9eecc7c3" Feb 13 19:51:28.402627 kubelet[2575]: E0213 19:51:28.402601 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:28.402904 containerd[1468]: time="2025-02-13T19:51:28.402859266Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-6d5rf,Uid:245734af-c06e-49f3-bfef-a966f57e2f98,Namespace:kube-system,Attempt:6,}" Feb 13 19:51:28.403220 containerd[1468]: time="2025-02-13T19:51:28.403037672Z" level=info msg="StopPodSandbox for \"725f264f85d2c0cd16ccc9f133fde5df46c364a1fdc97ca60f100ccf9eecc7c3\"" Feb 13 19:51:28.403514 containerd[1468]: time="2025-02-13T19:51:28.403494169Z" level=info msg="Ensure that sandbox 725f264f85d2c0cd16ccc9f133fde5df46c364a1fdc97ca60f100ccf9eecc7c3 in task-service has been cleanup successfully" Feb 13 19:51:28.403724 containerd[1468]: time="2025-02-13T19:51:28.403710274Z" level=info msg="TearDown network for sandbox \"725f264f85d2c0cd16ccc9f133fde5df46c364a1fdc97ca60f100ccf9eecc7c3\" successfully" Feb 13 19:51:28.403778 containerd[1468]: time="2025-02-13T19:51:28.403766319Z" level=info msg="StopPodSandbox for \"725f264f85d2c0cd16ccc9f133fde5df46c364a1fdc97ca60f100ccf9eecc7c3\" returns successfully" Feb 13 19:51:28.404271 containerd[1468]: time="2025-02-13T19:51:28.404230560Z" level=info msg="StopPodSandbox for \"6e045cd079efa77c85217df21eca6d5e6fd9d9297b9e159cbd34c6c2ec690015\"" Feb 13 19:51:28.404378 containerd[1468]: time="2025-02-13T19:51:28.404357138Z" level=info msg="TearDown network for sandbox \"6e045cd079efa77c85217df21eca6d5e6fd9d9297b9e159cbd34c6c2ec690015\" successfully" Feb 13 19:51:28.404401 containerd[1468]: time="2025-02-13T19:51:28.404377366Z" level=info msg="StopPodSandbox for \"6e045cd079efa77c85217df21eca6d5e6fd9d9297b9e159cbd34c6c2ec690015\" returns successfully" Feb 13 19:51:28.404772 containerd[1468]: time="2025-02-13T19:51:28.404741920Z" level=info msg="StopPodSandbox for \"e4d2487cfe7f2e3ccbb95c87ade298917e21e16bc54e8c6d5da7846fe11b6c37\"" Feb 13 19:51:28.404937 containerd[1468]: time="2025-02-13T19:51:28.404912120Z" level=info msg="TearDown network for sandbox \"e4d2487cfe7f2e3ccbb95c87ade298917e21e16bc54e8c6d5da7846fe11b6c37\" successfully" Feb 13 19:51:28.404937 containerd[1468]: time="2025-02-13T19:51:28.404933661Z" level=info msg="StopPodSandbox for \"e4d2487cfe7f2e3ccbb95c87ade298917e21e16bc54e8c6d5da7846fe11b6c37\" returns successfully" Feb 13 19:51:28.405237 containerd[1468]: time="2025-02-13T19:51:28.405216892Z" level=info msg="StopPodSandbox for \"86a985525093286f84ab532b093323eb20cae87c984957fc89def62353843bd5\"" Feb 13 19:51:28.405428 containerd[1468]: time="2025-02-13T19:51:28.405349982Z" level=info msg="TearDown network for sandbox \"86a985525093286f84ab532b093323eb20cae87c984957fc89def62353843bd5\" successfully" Feb 13 19:51:28.405428 containerd[1468]: time="2025-02-13T19:51:28.405363818Z" level=info msg="StopPodSandbox for \"86a985525093286f84ab532b093323eb20cae87c984957fc89def62353843bd5\" returns successfully" Feb 13 19:51:28.406148 containerd[1468]: time="2025-02-13T19:51:28.406009650Z" level=info msg="StopPodSandbox for \"d7392cd735f64bbb484b37d152068bd95b9e9585d02b78303932e81c6708295c\"" Feb 13 19:51:28.406148 containerd[1468]: time="2025-02-13T19:51:28.406097755Z" level=info msg="TearDown network for sandbox \"d7392cd735f64bbb484b37d152068bd95b9e9585d02b78303932e81c6708295c\" successfully" Feb 13 19:51:28.406148 containerd[1468]: time="2025-02-13T19:51:28.406108245Z" level=info msg="StopPodSandbox for \"d7392cd735f64bbb484b37d152068bd95b9e9585d02b78303932e81c6708295c\" returns successfully" Feb 13 19:51:28.406414 containerd[1468]: time="2025-02-13T19:51:28.406395565Z" level=info msg="StopPodSandbox for \"aea5462f72694272b5bc284a0bd9fc173269129df512c011930e89667ad0f97d\"" Feb 13 19:51:28.407167 containerd[1468]: time="2025-02-13T19:51:28.406629053Z" level=info msg="TearDown network for sandbox \"aea5462f72694272b5bc284a0bd9fc173269129df512c011930e89667ad0f97d\" successfully" Feb 13 19:51:28.407167 containerd[1468]: time="2025-02-13T19:51:28.406645023Z" level=info msg="StopPodSandbox for \"aea5462f72694272b5bc284a0bd9fc173269129df512c011930e89667ad0f97d\" returns successfully" Feb 13 19:51:28.407255 kubelet[2575]: E0213 19:51:28.406843 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:28.407768 containerd[1468]: time="2025-02-13T19:51:28.407465723Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-s22xp,Uid:bd8b42bf-05e5-489c-87c7-f5984836cd6a,Namespace:kube-system,Attempt:6,}" Feb 13 19:51:28.417235 kubelet[2575]: I0213 19:51:28.417133 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-x7j5b" podStartSLOduration=1.883146929 podStartE2EDuration="23.417104517s" podCreationTimestamp="2025-02-13 19:51:05 +0000 UTC" firstStartedPulling="2025-02-13 19:51:06.408269945 +0000 UTC m=+12.954586381" lastFinishedPulling="2025-02-13 19:51:27.942227543 +0000 UTC m=+34.488543969" observedRunningTime="2025-02-13 19:51:28.414131877 +0000 UTC m=+34.960448324" watchObservedRunningTime="2025-02-13 19:51:28.417104517 +0000 UTC m=+34.963420943" Feb 13 19:51:29.099969 systemd-networkd[1392]: cali32ebcf50b73: Link UP Feb 13 19:51:29.100295 systemd-networkd[1392]: cali32ebcf50b73: Gained carrier Feb 13 19:51:29.126028 systemd[1]: run-netns-cni\x2d09904cff\x2d7f02\x2d99ab\x2df175\x2d129e65b331b4.mount: Deactivated successfully. Feb 13 19:51:29.126745 containerd[1468]: 2025-02-13 19:51:28.795 [INFO][4809] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Feb 13 19:51:29.126745 containerd[1468]: 2025-02-13 19:51:28.817 [INFO][4809] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--668d6bf9bc--6d5rf-eth0 coredns-668d6bf9bc- kube-system 245734af-c06e-49f3-bfef-a966f57e2f98 755 0 2025-02-13 19:51:00 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-668d6bf9bc-6d5rf eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali32ebcf50b73 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="3d203a0b6be532f34ff1b77b79cfa9e0cc3f4be37034a55c05cda2cdbc8535c7" Namespace="kube-system" Pod="coredns-668d6bf9bc-6d5rf" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--6d5rf-" Feb 13 19:51:29.126745 containerd[1468]: 2025-02-13 19:51:28.817 [INFO][4809] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="3d203a0b6be532f34ff1b77b79cfa9e0cc3f4be37034a55c05cda2cdbc8535c7" Namespace="kube-system" Pod="coredns-668d6bf9bc-6d5rf" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--6d5rf-eth0" Feb 13 19:51:29.126745 containerd[1468]: 2025-02-13 19:51:28.985 [INFO][4827] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="3d203a0b6be532f34ff1b77b79cfa9e0cc3f4be37034a55c05cda2cdbc8535c7" HandleID="k8s-pod-network.3d203a0b6be532f34ff1b77b79cfa9e0cc3f4be37034a55c05cda2cdbc8535c7" Workload="localhost-k8s-coredns--668d6bf9bc--6d5rf-eth0" Feb 13 19:51:29.126745 containerd[1468]: 2025-02-13 19:51:29.035 [INFO][4827] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="3d203a0b6be532f34ff1b77b79cfa9e0cc3f4be37034a55c05cda2cdbc8535c7" HandleID="k8s-pod-network.3d203a0b6be532f34ff1b77b79cfa9e0cc3f4be37034a55c05cda2cdbc8535c7" Workload="localhost-k8s-coredns--668d6bf9bc--6d5rf-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0003acb30), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-668d6bf9bc-6d5rf", "timestamp":"2025-02-13 19:51:28.985921167 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Feb 13 19:51:29.126745 containerd[1468]: 2025-02-13 19:51:29.035 [INFO][4827] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Feb 13 19:51:29.126745 containerd[1468]: 2025-02-13 19:51:29.035 [INFO][4827] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Feb 13 19:51:29.126745 containerd[1468]: 2025-02-13 19:51:29.035 [INFO][4827] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Feb 13 19:51:29.126745 containerd[1468]: 2025-02-13 19:51:29.039 [INFO][4827] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.3d203a0b6be532f34ff1b77b79cfa9e0cc3f4be37034a55c05cda2cdbc8535c7" host="localhost" Feb 13 19:51:29.126745 containerd[1468]: 2025-02-13 19:51:29.051 [INFO][4827] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" Feb 13 19:51:29.126745 containerd[1468]: 2025-02-13 19:51:29.058 [INFO][4827] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Feb 13 19:51:29.126745 containerd[1468]: 2025-02-13 19:51:29.061 [INFO][4827] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Feb 13 19:51:29.126745 containerd[1468]: 2025-02-13 19:51:29.065 [INFO][4827] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Feb 13 19:51:29.126745 containerd[1468]: 2025-02-13 19:51:29.065 [INFO][4827] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.3d203a0b6be532f34ff1b77b79cfa9e0cc3f4be37034a55c05cda2cdbc8535c7" host="localhost" Feb 13 19:51:29.126745 containerd[1468]: 2025-02-13 19:51:29.067 [INFO][4827] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.3d203a0b6be532f34ff1b77b79cfa9e0cc3f4be37034a55c05cda2cdbc8535c7 Feb 13 19:51:29.126745 containerd[1468]: 2025-02-13 19:51:29.074 [INFO][4827] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.3d203a0b6be532f34ff1b77b79cfa9e0cc3f4be37034a55c05cda2cdbc8535c7" host="localhost" Feb 13 19:51:29.126745 containerd[1468]: 2025-02-13 19:51:29.085 [INFO][4827] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.129/26] block=192.168.88.128/26 handle="k8s-pod-network.3d203a0b6be532f34ff1b77b79cfa9e0cc3f4be37034a55c05cda2cdbc8535c7" host="localhost" Feb 13 19:51:29.126745 containerd[1468]: 2025-02-13 19:51:29.085 [INFO][4827] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.129/26] handle="k8s-pod-network.3d203a0b6be532f34ff1b77b79cfa9e0cc3f4be37034a55c05cda2cdbc8535c7" host="localhost" Feb 13 19:51:29.126745 containerd[1468]: 2025-02-13 19:51:29.085 [INFO][4827] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Feb 13 19:51:29.126745 containerd[1468]: 2025-02-13 19:51:29.085 [INFO][4827] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.129/26] IPv6=[] ContainerID="3d203a0b6be532f34ff1b77b79cfa9e0cc3f4be37034a55c05cda2cdbc8535c7" HandleID="k8s-pod-network.3d203a0b6be532f34ff1b77b79cfa9e0cc3f4be37034a55c05cda2cdbc8535c7" Workload="localhost-k8s-coredns--668d6bf9bc--6d5rf-eth0" Feb 13 19:51:29.127734 containerd[1468]: 2025-02-13 19:51:29.089 [INFO][4809] cni-plugin/k8s.go 386: Populated endpoint ContainerID="3d203a0b6be532f34ff1b77b79cfa9e0cc3f4be37034a55c05cda2cdbc8535c7" Namespace="kube-system" Pod="coredns-668d6bf9bc-6d5rf" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--6d5rf-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--668d6bf9bc--6d5rf-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"245734af-c06e-49f3-bfef-a966f57e2f98", ResourceVersion:"755", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 19, 51, 0, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-668d6bf9bc-6d5rf", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali32ebcf50b73", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 19:51:29.127734 containerd[1468]: 2025-02-13 19:51:29.089 [INFO][4809] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.129/32] ContainerID="3d203a0b6be532f34ff1b77b79cfa9e0cc3f4be37034a55c05cda2cdbc8535c7" Namespace="kube-system" Pod="coredns-668d6bf9bc-6d5rf" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--6d5rf-eth0" Feb 13 19:51:29.127734 containerd[1468]: 2025-02-13 19:51:29.089 [INFO][4809] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali32ebcf50b73 ContainerID="3d203a0b6be532f34ff1b77b79cfa9e0cc3f4be37034a55c05cda2cdbc8535c7" Namespace="kube-system" Pod="coredns-668d6bf9bc-6d5rf" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--6d5rf-eth0" Feb 13 19:51:29.127734 containerd[1468]: 2025-02-13 19:51:29.100 [INFO][4809] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="3d203a0b6be532f34ff1b77b79cfa9e0cc3f4be37034a55c05cda2cdbc8535c7" Namespace="kube-system" Pod="coredns-668d6bf9bc-6d5rf" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--6d5rf-eth0" Feb 13 19:51:29.127734 containerd[1468]: 2025-02-13 19:51:29.100 [INFO][4809] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="3d203a0b6be532f34ff1b77b79cfa9e0cc3f4be37034a55c05cda2cdbc8535c7" Namespace="kube-system" Pod="coredns-668d6bf9bc-6d5rf" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--6d5rf-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--668d6bf9bc--6d5rf-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"245734af-c06e-49f3-bfef-a966f57e2f98", ResourceVersion:"755", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 19, 51, 0, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"3d203a0b6be532f34ff1b77b79cfa9e0cc3f4be37034a55c05cda2cdbc8535c7", Pod:"coredns-668d6bf9bc-6d5rf", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali32ebcf50b73", MAC:"6e:b5:5c:a1:36:95", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 19:51:29.127734 containerd[1468]: 2025-02-13 19:51:29.121 [INFO][4809] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="3d203a0b6be532f34ff1b77b79cfa9e0cc3f4be37034a55c05cda2cdbc8535c7" Namespace="kube-system" Pod="coredns-668d6bf9bc-6d5rf" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--6d5rf-eth0" Feb 13 19:51:29.128185 systemd[1]: run-netns-cni\x2dec3d194c\x2d1928\x2dfaf9\x2d2909\x2d057e99fb47d3.mount: Deactivated successfully. Feb 13 19:51:29.197949 containerd[1468]: time="2025-02-13T19:51:29.197481359Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:51:29.197949 containerd[1468]: time="2025-02-13T19:51:29.197579093Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:51:29.197949 containerd[1468]: time="2025-02-13T19:51:29.197603879Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:51:29.197949 containerd[1468]: time="2025-02-13T19:51:29.197748992Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:51:29.206698 systemd-networkd[1392]: cali0237eeff969: Link UP Feb 13 19:51:29.207462 systemd-networkd[1392]: cali0237eeff969: Gained carrier Feb 13 19:51:29.224058 containerd[1468]: 2025-02-13 19:51:28.685 [INFO][4775] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Feb 13 19:51:29.224058 containerd[1468]: 2025-02-13 19:51:28.779 [INFO][4775] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-csi--node--driver--sxwqd-eth0 csi-node-driver- calico-system bbd75eeb-e06b-4812-b2c6-522804c5ef2f 646 0 2025-02-13 19:51:06 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:84cddb44f k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s localhost csi-node-driver-sxwqd eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali0237eeff969 [] []}} ContainerID="76eb5868109750c25808ea1c71219c263c1d063e208322be3fc3c91f65657bfe" Namespace="calico-system" Pod="csi-node-driver-sxwqd" WorkloadEndpoint="localhost-k8s-csi--node--driver--sxwqd-" Feb 13 19:51:29.224058 containerd[1468]: 2025-02-13 19:51:28.779 [INFO][4775] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="76eb5868109750c25808ea1c71219c263c1d063e208322be3fc3c91f65657bfe" Namespace="calico-system" Pod="csi-node-driver-sxwqd" WorkloadEndpoint="localhost-k8s-csi--node--driver--sxwqd-eth0" Feb 13 19:51:29.224058 containerd[1468]: 2025-02-13 19:51:28.995 [INFO][4822] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="76eb5868109750c25808ea1c71219c263c1d063e208322be3fc3c91f65657bfe" HandleID="k8s-pod-network.76eb5868109750c25808ea1c71219c263c1d063e208322be3fc3c91f65657bfe" Workload="localhost-k8s-csi--node--driver--sxwqd-eth0" Feb 13 19:51:29.224058 containerd[1468]: 2025-02-13 19:51:29.035 [INFO][4822] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="76eb5868109750c25808ea1c71219c263c1d063e208322be3fc3c91f65657bfe" HandleID="k8s-pod-network.76eb5868109750c25808ea1c71219c263c1d063e208322be3fc3c91f65657bfe" Workload="localhost-k8s-csi--node--driver--sxwqd-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0003aad50), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"csi-node-driver-sxwqd", "timestamp":"2025-02-13 19:51:28.995893536 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Feb 13 19:51:29.224058 containerd[1468]: 2025-02-13 19:51:29.035 [INFO][4822] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Feb 13 19:51:29.224058 containerd[1468]: 2025-02-13 19:51:29.085 [INFO][4822] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Feb 13 19:51:29.224058 containerd[1468]: 2025-02-13 19:51:29.085 [INFO][4822] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Feb 13 19:51:29.224058 containerd[1468]: 2025-02-13 19:51:29.141 [INFO][4822] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.76eb5868109750c25808ea1c71219c263c1d063e208322be3fc3c91f65657bfe" host="localhost" Feb 13 19:51:29.224058 containerd[1468]: 2025-02-13 19:51:29.151 [INFO][4822] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" Feb 13 19:51:29.224058 containerd[1468]: 2025-02-13 19:51:29.160 [INFO][4822] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Feb 13 19:51:29.224058 containerd[1468]: 2025-02-13 19:51:29.164 [INFO][4822] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Feb 13 19:51:29.224058 containerd[1468]: 2025-02-13 19:51:29.168 [INFO][4822] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Feb 13 19:51:29.224058 containerd[1468]: 2025-02-13 19:51:29.168 [INFO][4822] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.76eb5868109750c25808ea1c71219c263c1d063e208322be3fc3c91f65657bfe" host="localhost" Feb 13 19:51:29.224058 containerd[1468]: 2025-02-13 19:51:29.172 [INFO][4822] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.76eb5868109750c25808ea1c71219c263c1d063e208322be3fc3c91f65657bfe Feb 13 19:51:29.224058 containerd[1468]: 2025-02-13 19:51:29.178 [INFO][4822] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.76eb5868109750c25808ea1c71219c263c1d063e208322be3fc3c91f65657bfe" host="localhost" Feb 13 19:51:29.224058 containerd[1468]: 2025-02-13 19:51:29.194 [INFO][4822] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.130/26] block=192.168.88.128/26 handle="k8s-pod-network.76eb5868109750c25808ea1c71219c263c1d063e208322be3fc3c91f65657bfe" host="localhost" Feb 13 19:51:29.224058 containerd[1468]: 2025-02-13 19:51:29.194 [INFO][4822] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.130/26] handle="k8s-pod-network.76eb5868109750c25808ea1c71219c263c1d063e208322be3fc3c91f65657bfe" host="localhost" Feb 13 19:51:29.224058 containerd[1468]: 2025-02-13 19:51:29.194 [INFO][4822] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Feb 13 19:51:29.224058 containerd[1468]: 2025-02-13 19:51:29.194 [INFO][4822] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.130/26] IPv6=[] ContainerID="76eb5868109750c25808ea1c71219c263c1d063e208322be3fc3c91f65657bfe" HandleID="k8s-pod-network.76eb5868109750c25808ea1c71219c263c1d063e208322be3fc3c91f65657bfe" Workload="localhost-k8s-csi--node--driver--sxwqd-eth0" Feb 13 19:51:29.228203 containerd[1468]: 2025-02-13 19:51:29.198 [INFO][4775] cni-plugin/k8s.go 386: Populated endpoint ContainerID="76eb5868109750c25808ea1c71219c263c1d063e208322be3fc3c91f65657bfe" Namespace="calico-system" Pod="csi-node-driver-sxwqd" WorkloadEndpoint="localhost-k8s-csi--node--driver--sxwqd-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--sxwqd-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"bbd75eeb-e06b-4812-b2c6-522804c5ef2f", ResourceVersion:"646", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 19, 51, 6, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"84cddb44f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"csi-node-driver-sxwqd", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali0237eeff969", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 19:51:29.228203 containerd[1468]: 2025-02-13 19:51:29.199 [INFO][4775] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.130/32] ContainerID="76eb5868109750c25808ea1c71219c263c1d063e208322be3fc3c91f65657bfe" Namespace="calico-system" Pod="csi-node-driver-sxwqd" WorkloadEndpoint="localhost-k8s-csi--node--driver--sxwqd-eth0" Feb 13 19:51:29.228203 containerd[1468]: 2025-02-13 19:51:29.199 [INFO][4775] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali0237eeff969 ContainerID="76eb5868109750c25808ea1c71219c263c1d063e208322be3fc3c91f65657bfe" Namespace="calico-system" Pod="csi-node-driver-sxwqd" WorkloadEndpoint="localhost-k8s-csi--node--driver--sxwqd-eth0" Feb 13 19:51:29.228203 containerd[1468]: 2025-02-13 19:51:29.205 [INFO][4775] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="76eb5868109750c25808ea1c71219c263c1d063e208322be3fc3c91f65657bfe" Namespace="calico-system" Pod="csi-node-driver-sxwqd" WorkloadEndpoint="localhost-k8s-csi--node--driver--sxwqd-eth0" Feb 13 19:51:29.228203 containerd[1468]: 2025-02-13 19:51:29.205 [INFO][4775] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="76eb5868109750c25808ea1c71219c263c1d063e208322be3fc3c91f65657bfe" Namespace="calico-system" Pod="csi-node-driver-sxwqd" WorkloadEndpoint="localhost-k8s-csi--node--driver--sxwqd-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--sxwqd-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"bbd75eeb-e06b-4812-b2c6-522804c5ef2f", ResourceVersion:"646", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 19, 51, 6, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"84cddb44f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"76eb5868109750c25808ea1c71219c263c1d063e208322be3fc3c91f65657bfe", Pod:"csi-node-driver-sxwqd", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali0237eeff969", MAC:"8a:7f:e6:2f:05:a7", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 19:51:29.228203 containerd[1468]: 2025-02-13 19:51:29.220 [INFO][4775] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="76eb5868109750c25808ea1c71219c263c1d063e208322be3fc3c91f65657bfe" Namespace="calico-system" Pod="csi-node-driver-sxwqd" WorkloadEndpoint="localhost-k8s-csi--node--driver--sxwqd-eth0" Feb 13 19:51:29.237631 systemd[1]: Started cri-containerd-3d203a0b6be532f34ff1b77b79cfa9e0cc3f4be37034a55c05cda2cdbc8535c7.scope - libcontainer container 3d203a0b6be532f34ff1b77b79cfa9e0cc3f4be37034a55c05cda2cdbc8535c7. Feb 13 19:51:29.257718 containerd[1468]: time="2025-02-13T19:51:29.257167313Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:51:29.257718 containerd[1468]: time="2025-02-13T19:51:29.257240821Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:51:29.257718 containerd[1468]: time="2025-02-13T19:51:29.257255529Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:51:29.257718 containerd[1468]: time="2025-02-13T19:51:29.257343914Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:51:29.258801 systemd-resolved[1331]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Feb 13 19:51:29.287295 systemd[1]: Started cri-containerd-76eb5868109750c25808ea1c71219c263c1d063e208322be3fc3c91f65657bfe.scope - libcontainer container 76eb5868109750c25808ea1c71219c263c1d063e208322be3fc3c91f65657bfe. Feb 13 19:51:29.306610 systemd-resolved[1331]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Feb 13 19:51:29.307943 containerd[1468]: time="2025-02-13T19:51:29.307861048Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-6d5rf,Uid:245734af-c06e-49f3-bfef-a966f57e2f98,Namespace:kube-system,Attempt:6,} returns sandbox id \"3d203a0b6be532f34ff1b77b79cfa9e0cc3f4be37034a55c05cda2cdbc8535c7\"" Feb 13 19:51:29.308836 kubelet[2575]: E0213 19:51:29.308807 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:29.318692 containerd[1468]: time="2025-02-13T19:51:29.318639520Z" level=info msg="CreateContainer within sandbox \"3d203a0b6be532f34ff1b77b79cfa9e0cc3f4be37034a55c05cda2cdbc8535c7\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Feb 13 19:51:29.325322 systemd-networkd[1392]: calif7d6af010e9: Link UP Feb 13 19:51:29.327210 systemd-networkd[1392]: calif7d6af010e9: Gained carrier Feb 13 19:51:29.328064 containerd[1468]: time="2025-02-13T19:51:29.327608886Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-sxwqd,Uid:bbd75eeb-e06b-4812-b2c6-522804c5ef2f,Namespace:calico-system,Attempt:5,} returns sandbox id \"76eb5868109750c25808ea1c71219c263c1d063e208322be3fc3c91f65657bfe\"" Feb 13 19:51:29.337780 containerd[1468]: time="2025-02-13T19:51:29.337214697Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.1\"" Feb 13 19:51:29.351824 containerd[1468]: 2025-02-13 19:51:28.648 [INFO][4750] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Feb 13 19:51:29.351824 containerd[1468]: 2025-02-13 19:51:28.677 [INFO][4750] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--575f8f688d--8855n-eth0 calico-apiserver-575f8f688d- calico-apiserver a5e91fd3-19fa-490e-be13-26d807029222 748 0 2025-02-13 19:51:05 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:575f8f688d projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-575f8f688d-8855n eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] calif7d6af010e9 [] []}} ContainerID="3262af959167e8c4b3c4b8c815b5f5f89da8209224bb65693a300d9ec019a4ff" Namespace="calico-apiserver" Pod="calico-apiserver-575f8f688d-8855n" WorkloadEndpoint="localhost-k8s-calico--apiserver--575f8f688d--8855n-" Feb 13 19:51:29.351824 containerd[1468]: 2025-02-13 19:51:28.678 [INFO][4750] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="3262af959167e8c4b3c4b8c815b5f5f89da8209224bb65693a300d9ec019a4ff" Namespace="calico-apiserver" Pod="calico-apiserver-575f8f688d-8855n" WorkloadEndpoint="localhost-k8s-calico--apiserver--575f8f688d--8855n-eth0" Feb 13 19:51:29.351824 containerd[1468]: 2025-02-13 19:51:28.986 [INFO][4798] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="3262af959167e8c4b3c4b8c815b5f5f89da8209224bb65693a300d9ec019a4ff" HandleID="k8s-pod-network.3262af959167e8c4b3c4b8c815b5f5f89da8209224bb65693a300d9ec019a4ff" Workload="localhost-k8s-calico--apiserver--575f8f688d--8855n-eth0" Feb 13 19:51:29.351824 containerd[1468]: 2025-02-13 19:51:29.035 [INFO][4798] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="3262af959167e8c4b3c4b8c815b5f5f89da8209224bb65693a300d9ec019a4ff" HandleID="k8s-pod-network.3262af959167e8c4b3c4b8c815b5f5f89da8209224bb65693a300d9ec019a4ff" Workload="localhost-k8s-calico--apiserver--575f8f688d--8855n-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000296150), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-575f8f688d-8855n", "timestamp":"2025-02-13 19:51:28.986135689 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Feb 13 19:51:29.351824 containerd[1468]: 2025-02-13 19:51:29.035 [INFO][4798] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Feb 13 19:51:29.351824 containerd[1468]: 2025-02-13 19:51:29.194 [INFO][4798] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Feb 13 19:51:29.351824 containerd[1468]: 2025-02-13 19:51:29.194 [INFO][4798] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Feb 13 19:51:29.351824 containerd[1468]: 2025-02-13 19:51:29.242 [INFO][4798] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.3262af959167e8c4b3c4b8c815b5f5f89da8209224bb65693a300d9ec019a4ff" host="localhost" Feb 13 19:51:29.351824 containerd[1468]: 2025-02-13 19:51:29.253 [INFO][4798] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" Feb 13 19:51:29.351824 containerd[1468]: 2025-02-13 19:51:29.263 [INFO][4798] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Feb 13 19:51:29.351824 containerd[1468]: 2025-02-13 19:51:29.269 [INFO][4798] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Feb 13 19:51:29.351824 containerd[1468]: 2025-02-13 19:51:29.273 [INFO][4798] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Feb 13 19:51:29.351824 containerd[1468]: 2025-02-13 19:51:29.274 [INFO][4798] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.3262af959167e8c4b3c4b8c815b5f5f89da8209224bb65693a300d9ec019a4ff" host="localhost" Feb 13 19:51:29.351824 containerd[1468]: 2025-02-13 19:51:29.277 [INFO][4798] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.3262af959167e8c4b3c4b8c815b5f5f89da8209224bb65693a300d9ec019a4ff Feb 13 19:51:29.351824 containerd[1468]: 2025-02-13 19:51:29.284 [INFO][4798] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.3262af959167e8c4b3c4b8c815b5f5f89da8209224bb65693a300d9ec019a4ff" host="localhost" Feb 13 19:51:29.351824 containerd[1468]: 2025-02-13 19:51:29.304 [INFO][4798] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.131/26] block=192.168.88.128/26 handle="k8s-pod-network.3262af959167e8c4b3c4b8c815b5f5f89da8209224bb65693a300d9ec019a4ff" host="localhost" Feb 13 19:51:29.351824 containerd[1468]: 2025-02-13 19:51:29.304 [INFO][4798] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.131/26] handle="k8s-pod-network.3262af959167e8c4b3c4b8c815b5f5f89da8209224bb65693a300d9ec019a4ff" host="localhost" Feb 13 19:51:29.351824 containerd[1468]: 2025-02-13 19:51:29.304 [INFO][4798] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Feb 13 19:51:29.351824 containerd[1468]: 2025-02-13 19:51:29.304 [INFO][4798] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.131/26] IPv6=[] ContainerID="3262af959167e8c4b3c4b8c815b5f5f89da8209224bb65693a300d9ec019a4ff" HandleID="k8s-pod-network.3262af959167e8c4b3c4b8c815b5f5f89da8209224bb65693a300d9ec019a4ff" Workload="localhost-k8s-calico--apiserver--575f8f688d--8855n-eth0" Feb 13 19:51:29.352678 containerd[1468]: 2025-02-13 19:51:29.319 [INFO][4750] cni-plugin/k8s.go 386: Populated endpoint ContainerID="3262af959167e8c4b3c4b8c815b5f5f89da8209224bb65693a300d9ec019a4ff" Namespace="calico-apiserver" Pod="calico-apiserver-575f8f688d-8855n" WorkloadEndpoint="localhost-k8s-calico--apiserver--575f8f688d--8855n-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--575f8f688d--8855n-eth0", GenerateName:"calico-apiserver-575f8f688d-", Namespace:"calico-apiserver", SelfLink:"", UID:"a5e91fd3-19fa-490e-be13-26d807029222", ResourceVersion:"748", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 19, 51, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"575f8f688d", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-575f8f688d-8855n", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calif7d6af010e9", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 19:51:29.352678 containerd[1468]: 2025-02-13 19:51:29.319 [INFO][4750] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.131/32] ContainerID="3262af959167e8c4b3c4b8c815b5f5f89da8209224bb65693a300d9ec019a4ff" Namespace="calico-apiserver" Pod="calico-apiserver-575f8f688d-8855n" WorkloadEndpoint="localhost-k8s-calico--apiserver--575f8f688d--8855n-eth0" Feb 13 19:51:29.352678 containerd[1468]: 2025-02-13 19:51:29.319 [INFO][4750] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calif7d6af010e9 ContainerID="3262af959167e8c4b3c4b8c815b5f5f89da8209224bb65693a300d9ec019a4ff" Namespace="calico-apiserver" Pod="calico-apiserver-575f8f688d-8855n" WorkloadEndpoint="localhost-k8s-calico--apiserver--575f8f688d--8855n-eth0" Feb 13 19:51:29.352678 containerd[1468]: 2025-02-13 19:51:29.328 [INFO][4750] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="3262af959167e8c4b3c4b8c815b5f5f89da8209224bb65693a300d9ec019a4ff" Namespace="calico-apiserver" Pod="calico-apiserver-575f8f688d-8855n" WorkloadEndpoint="localhost-k8s-calico--apiserver--575f8f688d--8855n-eth0" Feb 13 19:51:29.352678 containerd[1468]: 2025-02-13 19:51:29.329 [INFO][4750] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="3262af959167e8c4b3c4b8c815b5f5f89da8209224bb65693a300d9ec019a4ff" Namespace="calico-apiserver" Pod="calico-apiserver-575f8f688d-8855n" WorkloadEndpoint="localhost-k8s-calico--apiserver--575f8f688d--8855n-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--575f8f688d--8855n-eth0", GenerateName:"calico-apiserver-575f8f688d-", Namespace:"calico-apiserver", SelfLink:"", UID:"a5e91fd3-19fa-490e-be13-26d807029222", ResourceVersion:"748", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 19, 51, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"575f8f688d", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"3262af959167e8c4b3c4b8c815b5f5f89da8209224bb65693a300d9ec019a4ff", Pod:"calico-apiserver-575f8f688d-8855n", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calif7d6af010e9", MAC:"a2:16:12:79:52:ea", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 19:51:29.352678 containerd[1468]: 2025-02-13 19:51:29.348 [INFO][4750] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="3262af959167e8c4b3c4b8c815b5f5f89da8209224bb65693a300d9ec019a4ff" Namespace="calico-apiserver" Pod="calico-apiserver-575f8f688d-8855n" WorkloadEndpoint="localhost-k8s-calico--apiserver--575f8f688d--8855n-eth0" Feb 13 19:51:29.367445 containerd[1468]: time="2025-02-13T19:51:29.367236419Z" level=info msg="CreateContainer within sandbox \"3d203a0b6be532f34ff1b77b79cfa9e0cc3f4be37034a55c05cda2cdbc8535c7\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"02a1b7e3862aeb5f1281236e15f5c74d18bab8442f4e84358e3a2b6b09e7bd42\"" Feb 13 19:51:29.369328 containerd[1468]: time="2025-02-13T19:51:29.368101512Z" level=info msg="StartContainer for \"02a1b7e3862aeb5f1281236e15f5c74d18bab8442f4e84358e3a2b6b09e7bd42\"" Feb 13 19:51:29.406121 systemd[1]: Started cri-containerd-02a1b7e3862aeb5f1281236e15f5c74d18bab8442f4e84358e3a2b6b09e7bd42.scope - libcontainer container 02a1b7e3862aeb5f1281236e15f5c74d18bab8442f4e84358e3a2b6b09e7bd42. Feb 13 19:51:29.408835 systemd-networkd[1392]: cali316c6680fd5: Link UP Feb 13 19:51:29.412106 systemd-networkd[1392]: cali316c6680fd5: Gained carrier Feb 13 19:51:29.422182 containerd[1468]: time="2025-02-13T19:51:29.421565228Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:51:29.422182 containerd[1468]: time="2025-02-13T19:51:29.422112645Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:51:29.422613 containerd[1468]: time="2025-02-13T19:51:29.422243260Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:51:29.422613 containerd[1468]: time="2025-02-13T19:51:29.422356232Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:51:29.433913 kubelet[2575]: I0213 19:51:29.432889 2575 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 13 19:51:29.433913 kubelet[2575]: E0213 19:51:29.433435 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:29.437322 containerd[1468]: 2025-02-13 19:51:28.583 [INFO][4735] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Feb 13 19:51:29.437322 containerd[1468]: 2025-02-13 19:51:28.607 [INFO][4735] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--575f8f688d--rm27z-eth0 calico-apiserver-575f8f688d- calico-apiserver dca83c7e-397c-4cb7-b95d-96c1c013b67f 751 0 2025-02-13 19:51:05 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:575f8f688d projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-575f8f688d-rm27z eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali316c6680fd5 [] []}} ContainerID="951ca50beaa6c68b43833e5784b46e75de16e07e3b775ed28b40c57afe780d60" Namespace="calico-apiserver" Pod="calico-apiserver-575f8f688d-rm27z" WorkloadEndpoint="localhost-k8s-calico--apiserver--575f8f688d--rm27z-" Feb 13 19:51:29.437322 containerd[1468]: 2025-02-13 19:51:28.608 [INFO][4735] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="951ca50beaa6c68b43833e5784b46e75de16e07e3b775ed28b40c57afe780d60" Namespace="calico-apiserver" Pod="calico-apiserver-575f8f688d-rm27z" WorkloadEndpoint="localhost-k8s-calico--apiserver--575f8f688d--rm27z-eth0" Feb 13 19:51:29.437322 containerd[1468]: 2025-02-13 19:51:28.987 [INFO][4773] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="951ca50beaa6c68b43833e5784b46e75de16e07e3b775ed28b40c57afe780d60" HandleID="k8s-pod-network.951ca50beaa6c68b43833e5784b46e75de16e07e3b775ed28b40c57afe780d60" Workload="localhost-k8s-calico--apiserver--575f8f688d--rm27z-eth0" Feb 13 19:51:29.437322 containerd[1468]: 2025-02-13 19:51:29.035 [INFO][4773] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="951ca50beaa6c68b43833e5784b46e75de16e07e3b775ed28b40c57afe780d60" HandleID="k8s-pod-network.951ca50beaa6c68b43833e5784b46e75de16e07e3b775ed28b40c57afe780d60" Workload="localhost-k8s-calico--apiserver--575f8f688d--rm27z-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000051a90), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-575f8f688d-rm27z", "timestamp":"2025-02-13 19:51:28.986416096 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Feb 13 19:51:29.437322 containerd[1468]: 2025-02-13 19:51:29.035 [INFO][4773] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Feb 13 19:51:29.437322 containerd[1468]: 2025-02-13 19:51:29.304 [INFO][4773] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Feb 13 19:51:29.437322 containerd[1468]: 2025-02-13 19:51:29.304 [INFO][4773] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Feb 13 19:51:29.437322 containerd[1468]: 2025-02-13 19:51:29.342 [INFO][4773] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.951ca50beaa6c68b43833e5784b46e75de16e07e3b775ed28b40c57afe780d60" host="localhost" Feb 13 19:51:29.437322 containerd[1468]: 2025-02-13 19:51:29.352 [INFO][4773] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" Feb 13 19:51:29.437322 containerd[1468]: 2025-02-13 19:51:29.364 [INFO][4773] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Feb 13 19:51:29.437322 containerd[1468]: 2025-02-13 19:51:29.368 [INFO][4773] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Feb 13 19:51:29.437322 containerd[1468]: 2025-02-13 19:51:29.374 [INFO][4773] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Feb 13 19:51:29.437322 containerd[1468]: 2025-02-13 19:51:29.374 [INFO][4773] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.951ca50beaa6c68b43833e5784b46e75de16e07e3b775ed28b40c57afe780d60" host="localhost" Feb 13 19:51:29.437322 containerd[1468]: 2025-02-13 19:51:29.377 [INFO][4773] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.951ca50beaa6c68b43833e5784b46e75de16e07e3b775ed28b40c57afe780d60 Feb 13 19:51:29.437322 containerd[1468]: 2025-02-13 19:51:29.385 [INFO][4773] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.951ca50beaa6c68b43833e5784b46e75de16e07e3b775ed28b40c57afe780d60" host="localhost" Feb 13 19:51:29.437322 containerd[1468]: 2025-02-13 19:51:29.397 [INFO][4773] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.132/26] block=192.168.88.128/26 handle="k8s-pod-network.951ca50beaa6c68b43833e5784b46e75de16e07e3b775ed28b40c57afe780d60" host="localhost" Feb 13 19:51:29.437322 containerd[1468]: 2025-02-13 19:51:29.397 [INFO][4773] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.132/26] handle="k8s-pod-network.951ca50beaa6c68b43833e5784b46e75de16e07e3b775ed28b40c57afe780d60" host="localhost" Feb 13 19:51:29.437322 containerd[1468]: 2025-02-13 19:51:29.397 [INFO][4773] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Feb 13 19:51:29.437322 containerd[1468]: 2025-02-13 19:51:29.397 [INFO][4773] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.132/26] IPv6=[] ContainerID="951ca50beaa6c68b43833e5784b46e75de16e07e3b775ed28b40c57afe780d60" HandleID="k8s-pod-network.951ca50beaa6c68b43833e5784b46e75de16e07e3b775ed28b40c57afe780d60" Workload="localhost-k8s-calico--apiserver--575f8f688d--rm27z-eth0" Feb 13 19:51:29.438184 containerd[1468]: 2025-02-13 19:51:29.404 [INFO][4735] cni-plugin/k8s.go 386: Populated endpoint ContainerID="951ca50beaa6c68b43833e5784b46e75de16e07e3b775ed28b40c57afe780d60" Namespace="calico-apiserver" Pod="calico-apiserver-575f8f688d-rm27z" WorkloadEndpoint="localhost-k8s-calico--apiserver--575f8f688d--rm27z-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--575f8f688d--rm27z-eth0", GenerateName:"calico-apiserver-575f8f688d-", Namespace:"calico-apiserver", SelfLink:"", UID:"dca83c7e-397c-4cb7-b95d-96c1c013b67f", ResourceVersion:"751", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 19, 51, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"575f8f688d", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-575f8f688d-rm27z", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali316c6680fd5", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 19:51:29.438184 containerd[1468]: 2025-02-13 19:51:29.404 [INFO][4735] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.132/32] ContainerID="951ca50beaa6c68b43833e5784b46e75de16e07e3b775ed28b40c57afe780d60" Namespace="calico-apiserver" Pod="calico-apiserver-575f8f688d-rm27z" WorkloadEndpoint="localhost-k8s-calico--apiserver--575f8f688d--rm27z-eth0" Feb 13 19:51:29.438184 containerd[1468]: 2025-02-13 19:51:29.404 [INFO][4735] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali316c6680fd5 ContainerID="951ca50beaa6c68b43833e5784b46e75de16e07e3b775ed28b40c57afe780d60" Namespace="calico-apiserver" Pod="calico-apiserver-575f8f688d-rm27z" WorkloadEndpoint="localhost-k8s-calico--apiserver--575f8f688d--rm27z-eth0" Feb 13 19:51:29.438184 containerd[1468]: 2025-02-13 19:51:29.408 [INFO][4735] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="951ca50beaa6c68b43833e5784b46e75de16e07e3b775ed28b40c57afe780d60" Namespace="calico-apiserver" Pod="calico-apiserver-575f8f688d-rm27z" WorkloadEndpoint="localhost-k8s-calico--apiserver--575f8f688d--rm27z-eth0" Feb 13 19:51:29.438184 containerd[1468]: 2025-02-13 19:51:29.409 [INFO][4735] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="951ca50beaa6c68b43833e5784b46e75de16e07e3b775ed28b40c57afe780d60" Namespace="calico-apiserver" Pod="calico-apiserver-575f8f688d-rm27z" WorkloadEndpoint="localhost-k8s-calico--apiserver--575f8f688d--rm27z-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--575f8f688d--rm27z-eth0", GenerateName:"calico-apiserver-575f8f688d-", Namespace:"calico-apiserver", SelfLink:"", UID:"dca83c7e-397c-4cb7-b95d-96c1c013b67f", ResourceVersion:"751", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 19, 51, 5, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"575f8f688d", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"951ca50beaa6c68b43833e5784b46e75de16e07e3b775ed28b40c57afe780d60", Pod:"calico-apiserver-575f8f688d-rm27z", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali316c6680fd5", MAC:"26:5f:bc:f8:ec:db", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 19:51:29.438184 containerd[1468]: 2025-02-13 19:51:29.430 [INFO][4735] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="951ca50beaa6c68b43833e5784b46e75de16e07e3b775ed28b40c57afe780d60" Namespace="calico-apiserver" Pod="calico-apiserver-575f8f688d-rm27z" WorkloadEndpoint="localhost-k8s-calico--apiserver--575f8f688d--rm27z-eth0" Feb 13 19:51:29.452805 systemd[1]: Started cri-containerd-3262af959167e8c4b3c4b8c815b5f5f89da8209224bb65693a300d9ec019a4ff.scope - libcontainer container 3262af959167e8c4b3c4b8c815b5f5f89da8209224bb65693a300d9ec019a4ff. Feb 13 19:51:29.476020 systemd-resolved[1331]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Feb 13 19:51:29.476295 containerd[1468]: time="2025-02-13T19:51:29.476005335Z" level=info msg="StartContainer for \"02a1b7e3862aeb5f1281236e15f5c74d18bab8442f4e84358e3a2b6b09e7bd42\" returns successfully" Feb 13 19:51:29.482504 containerd[1468]: time="2025-02-13T19:51:29.482248293Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:51:29.482695 containerd[1468]: time="2025-02-13T19:51:29.482467765Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:51:29.482695 containerd[1468]: time="2025-02-13T19:51:29.482491900Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:51:29.484453 containerd[1468]: time="2025-02-13T19:51:29.483939327Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:51:29.516211 systemd[1]: Started cri-containerd-951ca50beaa6c68b43833e5784b46e75de16e07e3b775ed28b40c57afe780d60.scope - libcontainer container 951ca50beaa6c68b43833e5784b46e75de16e07e3b775ed28b40c57afe780d60. Feb 13 19:51:29.525904 systemd-networkd[1392]: cali879ea2817f7: Link UP Feb 13 19:51:29.526734 systemd-networkd[1392]: cali879ea2817f7: Gained carrier Feb 13 19:51:29.529181 containerd[1468]: time="2025-02-13T19:51:29.529133519Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-575f8f688d-8855n,Uid:a5e91fd3-19fa-490e-be13-26d807029222,Namespace:calico-apiserver,Attempt:6,} returns sandbox id \"3262af959167e8c4b3c4b8c815b5f5f89da8209224bb65693a300d9ec019a4ff\"" Feb 13 19:51:29.552111 containerd[1468]: 2025-02-13 19:51:28.648 [INFO][4761] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Feb 13 19:51:29.552111 containerd[1468]: 2025-02-13 19:51:28.677 [INFO][4761] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--kube--controllers--5bf7d587cb--s28ls-eth0 calico-kube-controllers-5bf7d587cb- calico-system 77a9c264-ba41-4ec0-b704-771924c99471 753 0 2025-02-13 19:51:06 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:5bf7d587cb projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s localhost calico-kube-controllers-5bf7d587cb-s28ls eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali879ea2817f7 [] []}} ContainerID="4e303ccae692d7180aaa799bf448dec3718270dd1fd1a4774323275681d6f11a" Namespace="calico-system" Pod="calico-kube-controllers-5bf7d587cb-s28ls" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5bf7d587cb--s28ls-" Feb 13 19:51:29.552111 containerd[1468]: 2025-02-13 19:51:28.677 [INFO][4761] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="4e303ccae692d7180aaa799bf448dec3718270dd1fd1a4774323275681d6f11a" Namespace="calico-system" Pod="calico-kube-controllers-5bf7d587cb-s28ls" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5bf7d587cb--s28ls-eth0" Feb 13 19:51:29.552111 containerd[1468]: 2025-02-13 19:51:28.990 [INFO][4803] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="4e303ccae692d7180aaa799bf448dec3718270dd1fd1a4774323275681d6f11a" HandleID="k8s-pod-network.4e303ccae692d7180aaa799bf448dec3718270dd1fd1a4774323275681d6f11a" Workload="localhost-k8s-calico--kube--controllers--5bf7d587cb--s28ls-eth0" Feb 13 19:51:29.552111 containerd[1468]: 2025-02-13 19:51:29.036 [INFO][4803] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="4e303ccae692d7180aaa799bf448dec3718270dd1fd1a4774323275681d6f11a" HandleID="k8s-pod-network.4e303ccae692d7180aaa799bf448dec3718270dd1fd1a4774323275681d6f11a" Workload="localhost-k8s-calico--kube--controllers--5bf7d587cb--s28ls-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000415f70), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"calico-kube-controllers-5bf7d587cb-s28ls", "timestamp":"2025-02-13 19:51:28.990347315 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Feb 13 19:51:29.552111 containerd[1468]: 2025-02-13 19:51:29.036 [INFO][4803] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Feb 13 19:51:29.552111 containerd[1468]: 2025-02-13 19:51:29.400 [INFO][4803] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Feb 13 19:51:29.552111 containerd[1468]: 2025-02-13 19:51:29.400 [INFO][4803] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Feb 13 19:51:29.552111 containerd[1468]: 2025-02-13 19:51:29.444 [INFO][4803] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.4e303ccae692d7180aaa799bf448dec3718270dd1fd1a4774323275681d6f11a" host="localhost" Feb 13 19:51:29.552111 containerd[1468]: 2025-02-13 19:51:29.459 [INFO][4803] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" Feb 13 19:51:29.552111 containerd[1468]: 2025-02-13 19:51:29.471 [INFO][4803] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Feb 13 19:51:29.552111 containerd[1468]: 2025-02-13 19:51:29.477 [INFO][4803] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Feb 13 19:51:29.552111 containerd[1468]: 2025-02-13 19:51:29.483 [INFO][4803] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Feb 13 19:51:29.552111 containerd[1468]: 2025-02-13 19:51:29.483 [INFO][4803] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.4e303ccae692d7180aaa799bf448dec3718270dd1fd1a4774323275681d6f11a" host="localhost" Feb 13 19:51:29.552111 containerd[1468]: 2025-02-13 19:51:29.489 [INFO][4803] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.4e303ccae692d7180aaa799bf448dec3718270dd1fd1a4774323275681d6f11a Feb 13 19:51:29.552111 containerd[1468]: 2025-02-13 19:51:29.499 [INFO][4803] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.4e303ccae692d7180aaa799bf448dec3718270dd1fd1a4774323275681d6f11a" host="localhost" Feb 13 19:51:29.552111 containerd[1468]: 2025-02-13 19:51:29.517 [INFO][4803] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.133/26] block=192.168.88.128/26 handle="k8s-pod-network.4e303ccae692d7180aaa799bf448dec3718270dd1fd1a4774323275681d6f11a" host="localhost" Feb 13 19:51:29.552111 containerd[1468]: 2025-02-13 19:51:29.517 [INFO][4803] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.133/26] handle="k8s-pod-network.4e303ccae692d7180aaa799bf448dec3718270dd1fd1a4774323275681d6f11a" host="localhost" Feb 13 19:51:29.552111 containerd[1468]: 2025-02-13 19:51:29.518 [INFO][4803] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Feb 13 19:51:29.552111 containerd[1468]: 2025-02-13 19:51:29.518 [INFO][4803] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.133/26] IPv6=[] ContainerID="4e303ccae692d7180aaa799bf448dec3718270dd1fd1a4774323275681d6f11a" HandleID="k8s-pod-network.4e303ccae692d7180aaa799bf448dec3718270dd1fd1a4774323275681d6f11a" Workload="localhost-k8s-calico--kube--controllers--5bf7d587cb--s28ls-eth0" Feb 13 19:51:29.552919 systemd-resolved[1331]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Feb 13 19:51:29.553567 containerd[1468]: 2025-02-13 19:51:29.523 [INFO][4761] cni-plugin/k8s.go 386: Populated endpoint ContainerID="4e303ccae692d7180aaa799bf448dec3718270dd1fd1a4774323275681d6f11a" Namespace="calico-system" Pod="calico-kube-controllers-5bf7d587cb-s28ls" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5bf7d587cb--s28ls-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--5bf7d587cb--s28ls-eth0", GenerateName:"calico-kube-controllers-5bf7d587cb-", Namespace:"calico-system", SelfLink:"", UID:"77a9c264-ba41-4ec0-b704-771924c99471", ResourceVersion:"753", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 19, 51, 6, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"5bf7d587cb", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-kube-controllers-5bf7d587cb-s28ls", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali879ea2817f7", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 19:51:29.553567 containerd[1468]: 2025-02-13 19:51:29.523 [INFO][4761] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.133/32] ContainerID="4e303ccae692d7180aaa799bf448dec3718270dd1fd1a4774323275681d6f11a" Namespace="calico-system" Pod="calico-kube-controllers-5bf7d587cb-s28ls" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5bf7d587cb--s28ls-eth0" Feb 13 19:51:29.553567 containerd[1468]: 2025-02-13 19:51:29.523 [INFO][4761] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali879ea2817f7 ContainerID="4e303ccae692d7180aaa799bf448dec3718270dd1fd1a4774323275681d6f11a" Namespace="calico-system" Pod="calico-kube-controllers-5bf7d587cb-s28ls" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5bf7d587cb--s28ls-eth0" Feb 13 19:51:29.553567 containerd[1468]: 2025-02-13 19:51:29.526 [INFO][4761] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="4e303ccae692d7180aaa799bf448dec3718270dd1fd1a4774323275681d6f11a" Namespace="calico-system" Pod="calico-kube-controllers-5bf7d587cb-s28ls" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5bf7d587cb--s28ls-eth0" Feb 13 19:51:29.553567 containerd[1468]: 2025-02-13 19:51:29.526 [INFO][4761] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="4e303ccae692d7180aaa799bf448dec3718270dd1fd1a4774323275681d6f11a" Namespace="calico-system" Pod="calico-kube-controllers-5bf7d587cb-s28ls" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5bf7d587cb--s28ls-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--5bf7d587cb--s28ls-eth0", GenerateName:"calico-kube-controllers-5bf7d587cb-", Namespace:"calico-system", SelfLink:"", UID:"77a9c264-ba41-4ec0-b704-771924c99471", ResourceVersion:"753", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 19, 51, 6, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"5bf7d587cb", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"4e303ccae692d7180aaa799bf448dec3718270dd1fd1a4774323275681d6f11a", Pod:"calico-kube-controllers-5bf7d587cb-s28ls", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali879ea2817f7", MAC:"ee:ca:e6:56:7f:bd", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 19:51:29.553567 containerd[1468]: 2025-02-13 19:51:29.548 [INFO][4761] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="4e303ccae692d7180aaa799bf448dec3718270dd1fd1a4774323275681d6f11a" Namespace="calico-system" Pod="calico-kube-controllers-5bf7d587cb-s28ls" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5bf7d587cb--s28ls-eth0" Feb 13 19:51:29.589337 containerd[1468]: time="2025-02-13T19:51:29.589147659Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:51:29.590175 containerd[1468]: time="2025-02-13T19:51:29.589476326Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:51:29.590175 containerd[1468]: time="2025-02-13T19:51:29.589542891Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:51:29.590175 containerd[1468]: time="2025-02-13T19:51:29.589710616Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:51:29.594978 containerd[1468]: time="2025-02-13T19:51:29.594549068Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-575f8f688d-rm27z,Uid:dca83c7e-397c-4cb7-b95d-96c1c013b67f,Namespace:calico-apiserver,Attempt:6,} returns sandbox id \"951ca50beaa6c68b43833e5784b46e75de16e07e3b775ed28b40c57afe780d60\"" Feb 13 19:51:29.613975 systemd-networkd[1392]: cali4d56333c87a: Link UP Feb 13 19:51:29.614256 systemd-networkd[1392]: cali4d56333c87a: Gained carrier Feb 13 19:51:29.616817 systemd[1]: Started cri-containerd-4e303ccae692d7180aaa799bf448dec3718270dd1fd1a4774323275681d6f11a.scope - libcontainer container 4e303ccae692d7180aaa799bf448dec3718270dd1fd1a4774323275681d6f11a. Feb 13 19:51:29.638136 containerd[1468]: 2025-02-13 19:51:28.951 [INFO][4838] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Feb 13 19:51:29.638136 containerd[1468]: 2025-02-13 19:51:28.970 [INFO][4838] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--668d6bf9bc--s22xp-eth0 coredns-668d6bf9bc- kube-system bd8b42bf-05e5-489c-87c7-f5984836cd6a 754 0 2025-02-13 19:51:00 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-668d6bf9bc-s22xp eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali4d56333c87a [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="230edc07b37def1ff6a53304b624572a10d0b1c7b6e38d3627bda8b16a94cacc" Namespace="kube-system" Pod="coredns-668d6bf9bc-s22xp" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--s22xp-" Feb 13 19:51:29.638136 containerd[1468]: 2025-02-13 19:51:28.970 [INFO][4838] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="230edc07b37def1ff6a53304b624572a10d0b1c7b6e38d3627bda8b16a94cacc" Namespace="kube-system" Pod="coredns-668d6bf9bc-s22xp" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--s22xp-eth0" Feb 13 19:51:29.638136 containerd[1468]: 2025-02-13 19:51:29.021 [INFO][4867] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="230edc07b37def1ff6a53304b624572a10d0b1c7b6e38d3627bda8b16a94cacc" HandleID="k8s-pod-network.230edc07b37def1ff6a53304b624572a10d0b1c7b6e38d3627bda8b16a94cacc" Workload="localhost-k8s-coredns--668d6bf9bc--s22xp-eth0" Feb 13 19:51:29.638136 containerd[1468]: 2025-02-13 19:51:29.036 [INFO][4867] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="230edc07b37def1ff6a53304b624572a10d0b1c7b6e38d3627bda8b16a94cacc" HandleID="k8s-pod-network.230edc07b37def1ff6a53304b624572a10d0b1c7b6e38d3627bda8b16a94cacc" Workload="localhost-k8s-coredns--668d6bf9bc--s22xp-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0000517e0), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-668d6bf9bc-s22xp", "timestamp":"2025-02-13 19:51:29.021278506 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Feb 13 19:51:29.638136 containerd[1468]: 2025-02-13 19:51:29.036 [INFO][4867] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Feb 13 19:51:29.638136 containerd[1468]: 2025-02-13 19:51:29.518 [INFO][4867] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Feb 13 19:51:29.638136 containerd[1468]: 2025-02-13 19:51:29.518 [INFO][4867] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Feb 13 19:51:29.638136 containerd[1468]: 2025-02-13 19:51:29.543 [INFO][4867] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.230edc07b37def1ff6a53304b624572a10d0b1c7b6e38d3627bda8b16a94cacc" host="localhost" Feb 13 19:51:29.638136 containerd[1468]: 2025-02-13 19:51:29.555 [INFO][4867] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" Feb 13 19:51:29.638136 containerd[1468]: 2025-02-13 19:51:29.571 [INFO][4867] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Feb 13 19:51:29.638136 containerd[1468]: 2025-02-13 19:51:29.575 [INFO][4867] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Feb 13 19:51:29.638136 containerd[1468]: 2025-02-13 19:51:29.580 [INFO][4867] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Feb 13 19:51:29.638136 containerd[1468]: 2025-02-13 19:51:29.580 [INFO][4867] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.230edc07b37def1ff6a53304b624572a10d0b1c7b6e38d3627bda8b16a94cacc" host="localhost" Feb 13 19:51:29.638136 containerd[1468]: 2025-02-13 19:51:29.583 [INFO][4867] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.230edc07b37def1ff6a53304b624572a10d0b1c7b6e38d3627bda8b16a94cacc Feb 13 19:51:29.638136 containerd[1468]: 2025-02-13 19:51:29.592 [INFO][4867] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.230edc07b37def1ff6a53304b624572a10d0b1c7b6e38d3627bda8b16a94cacc" host="localhost" Feb 13 19:51:29.638136 containerd[1468]: 2025-02-13 19:51:29.604 [INFO][4867] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.134/26] block=192.168.88.128/26 handle="k8s-pod-network.230edc07b37def1ff6a53304b624572a10d0b1c7b6e38d3627bda8b16a94cacc" host="localhost" Feb 13 19:51:29.638136 containerd[1468]: 2025-02-13 19:51:29.605 [INFO][4867] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.134/26] handle="k8s-pod-network.230edc07b37def1ff6a53304b624572a10d0b1c7b6e38d3627bda8b16a94cacc" host="localhost" Feb 13 19:51:29.638136 containerd[1468]: 2025-02-13 19:51:29.605 [INFO][4867] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Feb 13 19:51:29.638136 containerd[1468]: 2025-02-13 19:51:29.605 [INFO][4867] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.134/26] IPv6=[] ContainerID="230edc07b37def1ff6a53304b624572a10d0b1c7b6e38d3627bda8b16a94cacc" HandleID="k8s-pod-network.230edc07b37def1ff6a53304b624572a10d0b1c7b6e38d3627bda8b16a94cacc" Workload="localhost-k8s-coredns--668d6bf9bc--s22xp-eth0" Feb 13 19:51:29.638766 containerd[1468]: 2025-02-13 19:51:29.609 [INFO][4838] cni-plugin/k8s.go 386: Populated endpoint ContainerID="230edc07b37def1ff6a53304b624572a10d0b1c7b6e38d3627bda8b16a94cacc" Namespace="kube-system" Pod="coredns-668d6bf9bc-s22xp" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--s22xp-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--668d6bf9bc--s22xp-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"bd8b42bf-05e5-489c-87c7-f5984836cd6a", ResourceVersion:"754", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 19, 51, 0, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-668d6bf9bc-s22xp", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali4d56333c87a", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 19:51:29.638766 containerd[1468]: 2025-02-13 19:51:29.609 [INFO][4838] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.134/32] ContainerID="230edc07b37def1ff6a53304b624572a10d0b1c7b6e38d3627bda8b16a94cacc" Namespace="kube-system" Pod="coredns-668d6bf9bc-s22xp" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--s22xp-eth0" Feb 13 19:51:29.638766 containerd[1468]: 2025-02-13 19:51:29.609 [INFO][4838] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali4d56333c87a ContainerID="230edc07b37def1ff6a53304b624572a10d0b1c7b6e38d3627bda8b16a94cacc" Namespace="kube-system" Pod="coredns-668d6bf9bc-s22xp" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--s22xp-eth0" Feb 13 19:51:29.638766 containerd[1468]: 2025-02-13 19:51:29.614 [INFO][4838] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="230edc07b37def1ff6a53304b624572a10d0b1c7b6e38d3627bda8b16a94cacc" Namespace="kube-system" Pod="coredns-668d6bf9bc-s22xp" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--s22xp-eth0" Feb 13 19:51:29.638766 containerd[1468]: 2025-02-13 19:51:29.614 [INFO][4838] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="230edc07b37def1ff6a53304b624572a10d0b1c7b6e38d3627bda8b16a94cacc" Namespace="kube-system" Pod="coredns-668d6bf9bc-s22xp" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--s22xp-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--668d6bf9bc--s22xp-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"bd8b42bf-05e5-489c-87c7-f5984836cd6a", ResourceVersion:"754", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 19, 51, 0, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"230edc07b37def1ff6a53304b624572a10d0b1c7b6e38d3627bda8b16a94cacc", Pod:"coredns-668d6bf9bc-s22xp", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali4d56333c87a", MAC:"62:a7:f5:5b:04:1a", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 19:51:29.638766 containerd[1468]: 2025-02-13 19:51:29.631 [INFO][4838] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="230edc07b37def1ff6a53304b624572a10d0b1c7b6e38d3627bda8b16a94cacc" Namespace="kube-system" Pod="coredns-668d6bf9bc-s22xp" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--s22xp-eth0" Feb 13 19:51:29.642826 systemd-resolved[1331]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Feb 13 19:51:29.667968 containerd[1468]: time="2025-02-13T19:51:29.667734862Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:51:29.667968 containerd[1468]: time="2025-02-13T19:51:29.667794874Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:51:29.667968 containerd[1468]: time="2025-02-13T19:51:29.667809341Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:51:29.668838 containerd[1468]: time="2025-02-13T19:51:29.668737995Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:51:29.679800 containerd[1468]: time="2025-02-13T19:51:29.679628987Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5bf7d587cb-s28ls,Uid:77a9c264-ba41-4ec0-b704-771924c99471,Namespace:calico-system,Attempt:6,} returns sandbox id \"4e303ccae692d7180aaa799bf448dec3718270dd1fd1a4774323275681d6f11a\"" Feb 13 19:51:29.691172 systemd[1]: Started cri-containerd-230edc07b37def1ff6a53304b624572a10d0b1c7b6e38d3627bda8b16a94cacc.scope - libcontainer container 230edc07b37def1ff6a53304b624572a10d0b1c7b6e38d3627bda8b16a94cacc. Feb 13 19:51:29.708218 systemd-resolved[1331]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Feb 13 19:51:29.739188 containerd[1468]: time="2025-02-13T19:51:29.739139773Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-s22xp,Uid:bd8b42bf-05e5-489c-87c7-f5984836cd6a,Namespace:kube-system,Attempt:6,} returns sandbox id \"230edc07b37def1ff6a53304b624572a10d0b1c7b6e38d3627bda8b16a94cacc\"" Feb 13 19:51:29.741139 kubelet[2575]: E0213 19:51:29.740924 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:29.744459 containerd[1468]: time="2025-02-13T19:51:29.744097468Z" level=info msg="CreateContainer within sandbox \"230edc07b37def1ff6a53304b624572a10d0b1c7b6e38d3627bda8b16a94cacc\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Feb 13 19:51:29.789284 containerd[1468]: time="2025-02-13T19:51:29.789192614Z" level=info msg="CreateContainer within sandbox \"230edc07b37def1ff6a53304b624572a10d0b1c7b6e38d3627bda8b16a94cacc\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"891a1e7d3dce21096d0ef3436597af36370596fc25d0f5d00bddf72a0cadf596\"" Feb 13 19:51:29.793338 containerd[1468]: time="2025-02-13T19:51:29.793274045Z" level=info msg="StartContainer for \"891a1e7d3dce21096d0ef3436597af36370596fc25d0f5d00bddf72a0cadf596\"" Feb 13 19:51:29.834249 systemd[1]: Started cri-containerd-891a1e7d3dce21096d0ef3436597af36370596fc25d0f5d00bddf72a0cadf596.scope - libcontainer container 891a1e7d3dce21096d0ef3436597af36370596fc25d0f5d00bddf72a0cadf596. Feb 13 19:51:29.874752 containerd[1468]: time="2025-02-13T19:51:29.874272814Z" level=info msg="StartContainer for \"891a1e7d3dce21096d0ef3436597af36370596fc25d0f5d00bddf72a0cadf596\" returns successfully" Feb 13 19:51:30.211048 kernel: bpftool[5398]: memfd_create() called without MFD_EXEC or MFD_NOEXEC_SEAL set Feb 13 19:51:30.460701 kubelet[2575]: E0213 19:51:30.460621 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:30.476209 kubelet[2575]: I0213 19:51:30.476073 2575 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 13 19:51:30.476566 kubelet[2575]: E0213 19:51:30.476530 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:30.478583 kubelet[2575]: E0213 19:51:30.477722 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:30.496514 systemd-networkd[1392]: vxlan.calico: Link UP Feb 13 19:51:30.496526 systemd-networkd[1392]: vxlan.calico: Gained carrier Feb 13 19:51:30.507563 systemd-networkd[1392]: calif7d6af010e9: Gained IPv6LL Feb 13 19:51:30.735484 kubelet[2575]: I0213 19:51:30.733474 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-6d5rf" podStartSLOduration=30.733449684 podStartE2EDuration="30.733449684s" podCreationTimestamp="2025-02-13 19:51:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 19:51:30.497443327 +0000 UTC m=+37.043759773" watchObservedRunningTime="2025-02-13 19:51:30.733449684 +0000 UTC m=+37.279766110" Feb 13 19:51:30.769652 systemd-networkd[1392]: cali316c6680fd5: Gained IPv6LL Feb 13 19:51:30.775337 kubelet[2575]: I0213 19:51:30.775265 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-s22xp" podStartSLOduration=30.775230995 podStartE2EDuration="30.775230995s" podCreationTimestamp="2025-02-13 19:51:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 19:51:30.770217785 +0000 UTC m=+37.316534211" watchObservedRunningTime="2025-02-13 19:51:30.775230995 +0000 UTC m=+37.321547421" Feb 13 19:51:30.891185 systemd-networkd[1392]: cali0237eeff969: Gained IPv6LL Feb 13 19:51:31.083063 systemd-networkd[1392]: cali879ea2817f7: Gained IPv6LL Feb 13 19:51:31.083704 systemd-networkd[1392]: cali4d56333c87a: Gained IPv6LL Feb 13 19:51:31.147314 systemd-networkd[1392]: cali32ebcf50b73: Gained IPv6LL Feb 13 19:51:31.459474 containerd[1468]: time="2025-02-13T19:51:31.459284088Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:51:31.477966 kubelet[2575]: E0213 19:51:31.477907 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:31.477966 kubelet[2575]: E0213 19:51:31.477965 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:31.510666 containerd[1468]: time="2025-02-13T19:51:31.510568959Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.29.1: active requests=0, bytes read=7902632" Feb 13 19:51:31.537493 containerd[1468]: time="2025-02-13T19:51:31.537428980Z" level=info msg="ImageCreate event name:\"sha256:bda8c42e04758c4f061339e213f50ccdc7502c4176fbf631aa12357e62b63540\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:51:31.630301 containerd[1468]: time="2025-02-13T19:51:31.630239618Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:eaa7e01fb16b603c155a67b81f16992281db7f831684c7b2081d3434587a7ff3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:51:31.631258 containerd[1468]: time="2025-02-13T19:51:31.631207284Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.29.1\" with image id \"sha256:bda8c42e04758c4f061339e213f50ccdc7502c4176fbf631aa12357e62b63540\", repo tag \"ghcr.io/flatcar/calico/csi:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:eaa7e01fb16b603c155a67b81f16992281db7f831684c7b2081d3434587a7ff3\", size \"9395716\" in 2.293936271s" Feb 13 19:51:31.631258 containerd[1468]: time="2025-02-13T19:51:31.631254372Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.1\" returns image reference \"sha256:bda8c42e04758c4f061339e213f50ccdc7502c4176fbf631aa12357e62b63540\"" Feb 13 19:51:31.632560 containerd[1468]: time="2025-02-13T19:51:31.632514327Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.1\"" Feb 13 19:51:31.633594 containerd[1468]: time="2025-02-13T19:51:31.633567363Z" level=info msg="CreateContainer within sandbox \"76eb5868109750c25808ea1c71219c263c1d063e208322be3fc3c91f65657bfe\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Feb 13 19:51:31.659041 systemd-networkd[1392]: vxlan.calico: Gained IPv6LL Feb 13 19:51:31.756659 containerd[1468]: time="2025-02-13T19:51:31.756593570Z" level=info msg="CreateContainer within sandbox \"76eb5868109750c25808ea1c71219c263c1d063e208322be3fc3c91f65657bfe\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"b2dbdf355d0a77b670cfd328cf66e6302c6ddd8090a04de3bba9d56c76c5f313\"" Feb 13 19:51:31.757240 containerd[1468]: time="2025-02-13T19:51:31.757204837Z" level=info msg="StartContainer for \"b2dbdf355d0a77b670cfd328cf66e6302c6ddd8090a04de3bba9d56c76c5f313\"" Feb 13 19:51:31.792322 systemd[1]: Started cri-containerd-b2dbdf355d0a77b670cfd328cf66e6302c6ddd8090a04de3bba9d56c76c5f313.scope - libcontainer container b2dbdf355d0a77b670cfd328cf66e6302c6ddd8090a04de3bba9d56c76c5f313. Feb 13 19:51:31.838524 containerd[1468]: time="2025-02-13T19:51:31.838459125Z" level=info msg="StartContainer for \"b2dbdf355d0a77b670cfd328cf66e6302c6ddd8090a04de3bba9d56c76c5f313\" returns successfully" Feb 13 19:51:32.483402 kubelet[2575]: E0213 19:51:32.483360 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:32.483913 kubelet[2575]: E0213 19:51:32.483364 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:32.849644 systemd[1]: Started sshd@9-10.0.0.104:22-10.0.0.1:58324.service - OpenSSH per-connection server daemon (10.0.0.1:58324). Feb 13 19:51:32.912116 sshd[5571]: Accepted publickey for core from 10.0.0.1 port 58324 ssh2: RSA SHA256:8WP2kqV5KzwZsuVRMXRFZkAHZWbkdD5kizbT2H+wOcw Feb 13 19:51:32.914506 sshd-session[5571]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:51:32.920305 systemd-logind[1447]: New session 9 of user core. Feb 13 19:51:32.928157 systemd[1]: Started session-9.scope - Session 9 of User core. Feb 13 19:51:33.075495 sshd[5573]: Connection closed by 10.0.0.1 port 58324 Feb 13 19:51:33.075961 sshd-session[5571]: pam_unix(sshd:session): session closed for user core Feb 13 19:51:33.080631 systemd[1]: sshd@9-10.0.0.104:22-10.0.0.1:58324.service: Deactivated successfully. Feb 13 19:51:33.082983 systemd[1]: session-9.scope: Deactivated successfully. Feb 13 19:51:33.083908 systemd-logind[1447]: Session 9 logged out. Waiting for processes to exit. Feb 13 19:51:33.085702 systemd-logind[1447]: Removed session 9. Feb 13 19:51:33.486055 kubelet[2575]: E0213 19:51:33.486013 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:51:34.205076 containerd[1468]: time="2025-02-13T19:51:34.204936792Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:51:34.205808 containerd[1468]: time="2025-02-13T19:51:34.205759626Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.29.1: active requests=0, bytes read=42001404" Feb 13 19:51:34.206991 containerd[1468]: time="2025-02-13T19:51:34.206950922Z" level=info msg="ImageCreate event name:\"sha256:421726ace5ed13894f7edf594dd3a462947aedc13d0f69d08525d7369477fb70\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:51:34.209487 containerd[1468]: time="2025-02-13T19:51:34.209456904Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:b8c43e264fe52e0c327b0bf3ac882a0224b33bdd7f4ff58a74242da7d9b00486\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:51:34.210073 containerd[1468]: time="2025-02-13T19:51:34.210036952Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.29.1\" with image id \"sha256:421726ace5ed13894f7edf594dd3a462947aedc13d0f69d08525d7369477fb70\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:b8c43e264fe52e0c327b0bf3ac882a0224b33bdd7f4ff58a74242da7d9b00486\", size \"43494504\" in 2.577489172s" Feb 13 19:51:34.210073 containerd[1468]: time="2025-02-13T19:51:34.210066748Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.1\" returns image reference \"sha256:421726ace5ed13894f7edf594dd3a462947aedc13d0f69d08525d7369477fb70\"" Feb 13 19:51:34.211827 containerd[1468]: time="2025-02-13T19:51:34.211320140Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.1\"" Feb 13 19:51:34.212018 containerd[1468]: time="2025-02-13T19:51:34.211908604Z" level=info msg="CreateContainer within sandbox \"3262af959167e8c4b3c4b8c815b5f5f89da8209224bb65693a300d9ec019a4ff\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Feb 13 19:51:34.231880 containerd[1468]: time="2025-02-13T19:51:34.231805561Z" level=info msg="CreateContainer within sandbox \"3262af959167e8c4b3c4b8c815b5f5f89da8209224bb65693a300d9ec019a4ff\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"759c5200a451670ab11066c9c096fabb1b74235533c2ed8f33553133dc076891\"" Feb 13 19:51:34.232417 containerd[1468]: time="2025-02-13T19:51:34.232387794Z" level=info msg="StartContainer for \"759c5200a451670ab11066c9c096fabb1b74235533c2ed8f33553133dc076891\"" Feb 13 19:51:34.269062 systemd[1]: Started cri-containerd-759c5200a451670ab11066c9c096fabb1b74235533c2ed8f33553133dc076891.scope - libcontainer container 759c5200a451670ab11066c9c096fabb1b74235533c2ed8f33553133dc076891. Feb 13 19:51:34.312552 containerd[1468]: time="2025-02-13T19:51:34.312496957Z" level=info msg="StartContainer for \"759c5200a451670ab11066c9c096fabb1b74235533c2ed8f33553133dc076891\" returns successfully" Feb 13 19:51:34.673604 containerd[1468]: time="2025-02-13T19:51:34.673551352Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:51:34.674421 containerd[1468]: time="2025-02-13T19:51:34.674319473Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.29.1: active requests=0, bytes read=77" Feb 13 19:51:34.676371 containerd[1468]: time="2025-02-13T19:51:34.676333122Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.29.1\" with image id \"sha256:421726ace5ed13894f7edf594dd3a462947aedc13d0f69d08525d7369477fb70\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:b8c43e264fe52e0c327b0bf3ac882a0224b33bdd7f4ff58a74242da7d9b00486\", size \"43494504\" in 464.974119ms" Feb 13 19:51:34.676444 containerd[1468]: time="2025-02-13T19:51:34.676372906Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.1\" returns image reference \"sha256:421726ace5ed13894f7edf594dd3a462947aedc13d0f69d08525d7369477fb70\"" Feb 13 19:51:34.677688 containerd[1468]: time="2025-02-13T19:51:34.677637720Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.29.1\"" Feb 13 19:51:34.678860 containerd[1468]: time="2025-02-13T19:51:34.678815039Z" level=info msg="CreateContainer within sandbox \"951ca50beaa6c68b43833e5784b46e75de16e07e3b775ed28b40c57afe780d60\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Feb 13 19:51:34.698709 containerd[1468]: time="2025-02-13T19:51:34.698567755Z" level=info msg="CreateContainer within sandbox \"951ca50beaa6c68b43833e5784b46e75de16e07e3b775ed28b40c57afe780d60\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"a0ba01aefe26558f7c33a7afeac6ece3b8d61569a092a830c80cddb991bd44a1\"" Feb 13 19:51:34.699170 containerd[1468]: time="2025-02-13T19:51:34.699151470Z" level=info msg="StartContainer for \"a0ba01aefe26558f7c33a7afeac6ece3b8d61569a092a830c80cddb991bd44a1\"" Feb 13 19:51:34.733582 systemd[1]: Started cri-containerd-a0ba01aefe26558f7c33a7afeac6ece3b8d61569a092a830c80cddb991bd44a1.scope - libcontainer container a0ba01aefe26558f7c33a7afeac6ece3b8d61569a092a830c80cddb991bd44a1. Feb 13 19:51:34.794924 containerd[1468]: time="2025-02-13T19:51:34.794738549Z" level=info msg="StartContainer for \"a0ba01aefe26558f7c33a7afeac6ece3b8d61569a092a830c80cddb991bd44a1\" returns successfully" Feb 13 19:51:35.500809 kubelet[2575]: I0213 19:51:35.500771 2575 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 13 19:51:35.518397 kubelet[2575]: I0213 19:51:35.518284 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-575f8f688d-8855n" podStartSLOduration=25.841574964 podStartE2EDuration="30.518259289s" podCreationTimestamp="2025-02-13 19:51:05 +0000 UTC" firstStartedPulling="2025-02-13 19:51:29.5340731 +0000 UTC m=+36.080389526" lastFinishedPulling="2025-02-13 19:51:34.210757415 +0000 UTC m=+40.757073851" observedRunningTime="2025-02-13 19:51:34.502434764 +0000 UTC m=+41.048751190" watchObservedRunningTime="2025-02-13 19:51:35.518259289 +0000 UTC m=+42.064575715" Feb 13 19:51:36.502859 kubelet[2575]: I0213 19:51:36.502813 2575 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 13 19:51:37.277232 containerd[1468]: time="2025-02-13T19:51:37.277140202Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:51:37.279148 containerd[1468]: time="2025-02-13T19:51:37.279001794Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.29.1: active requests=0, bytes read=34141192" Feb 13 19:51:37.281078 containerd[1468]: time="2025-02-13T19:51:37.281011826Z" level=info msg="ImageCreate event name:\"sha256:6331715a2ae96b18a770a395cac108321d108e445e08b616e5bc9fbd1f9c21da\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:51:37.284410 containerd[1468]: time="2025-02-13T19:51:37.284336904Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:1072d6a98167a14ca361e9ce757733f9bae36d1f1c6a9621ea10934b6b1e10d9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:51:37.285141 containerd[1468]: time="2025-02-13T19:51:37.285104505Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.29.1\" with image id \"sha256:6331715a2ae96b18a770a395cac108321d108e445e08b616e5bc9fbd1f9c21da\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:1072d6a98167a14ca361e9ce757733f9bae36d1f1c6a9621ea10934b6b1e10d9\", size \"35634244\" in 2.607436047s" Feb 13 19:51:37.285221 containerd[1468]: time="2025-02-13T19:51:37.285142887Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.29.1\" returns image reference \"sha256:6331715a2ae96b18a770a395cac108321d108e445e08b616e5bc9fbd1f9c21da\"" Feb 13 19:51:37.286398 containerd[1468]: time="2025-02-13T19:51:37.286351163Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\"" Feb 13 19:51:37.301367 containerd[1468]: time="2025-02-13T19:51:37.301311251Z" level=info msg="CreateContainer within sandbox \"4e303ccae692d7180aaa799bf448dec3718270dd1fd1a4774323275681d6f11a\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Feb 13 19:51:37.320686 containerd[1468]: time="2025-02-13T19:51:37.320603458Z" level=info msg="CreateContainer within sandbox \"4e303ccae692d7180aaa799bf448dec3718270dd1fd1a4774323275681d6f11a\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"8d0161f152d1e4d24f945c90939e2ed9109ac08eb4bff552dc0e550ff7bc5a46\"" Feb 13 19:51:37.321438 containerd[1468]: time="2025-02-13T19:51:37.321335131Z" level=info msg="StartContainer for \"8d0161f152d1e4d24f945c90939e2ed9109ac08eb4bff552dc0e550ff7bc5a46\"" Feb 13 19:51:37.362281 systemd[1]: Started cri-containerd-8d0161f152d1e4d24f945c90939e2ed9109ac08eb4bff552dc0e550ff7bc5a46.scope - libcontainer container 8d0161f152d1e4d24f945c90939e2ed9109ac08eb4bff552dc0e550ff7bc5a46. Feb 13 19:51:37.413184 containerd[1468]: time="2025-02-13T19:51:37.413124619Z" level=info msg="StartContainer for \"8d0161f152d1e4d24f945c90939e2ed9109ac08eb4bff552dc0e550ff7bc5a46\" returns successfully" Feb 13 19:51:37.523242 kubelet[2575]: I0213 19:51:37.523153 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-5bf7d587cb-s28ls" podStartSLOduration=23.918536513 podStartE2EDuration="31.523132158s" podCreationTimestamp="2025-02-13 19:51:06 +0000 UTC" firstStartedPulling="2025-02-13 19:51:29.681628089 +0000 UTC m=+36.227944515" lastFinishedPulling="2025-02-13 19:51:37.286223734 +0000 UTC m=+43.832540160" observedRunningTime="2025-02-13 19:51:37.52117178 +0000 UTC m=+44.067488206" watchObservedRunningTime="2025-02-13 19:51:37.523132158 +0000 UTC m=+44.069448584" Feb 13 19:51:37.523833 kubelet[2575]: I0213 19:51:37.523503 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-575f8f688d-rm27z" podStartSLOduration=27.444617752 podStartE2EDuration="32.523495711s" podCreationTimestamp="2025-02-13 19:51:05 +0000 UTC" firstStartedPulling="2025-02-13 19:51:29.598384958 +0000 UTC m=+36.144701384" lastFinishedPulling="2025-02-13 19:51:34.677262927 +0000 UTC m=+41.223579343" observedRunningTime="2025-02-13 19:51:35.518999027 +0000 UTC m=+42.065315453" watchObservedRunningTime="2025-02-13 19:51:37.523495711 +0000 UTC m=+44.069812147" Feb 13 19:51:38.095623 systemd[1]: Started sshd@10-10.0.0.104:22-10.0.0.1:58328.service - OpenSSH per-connection server daemon (10.0.0.1:58328). Feb 13 19:51:38.176797 sshd[5752]: Accepted publickey for core from 10.0.0.1 port 58328 ssh2: RSA SHA256:8WP2kqV5KzwZsuVRMXRFZkAHZWbkdD5kizbT2H+wOcw Feb 13 19:51:38.178598 sshd-session[5752]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:51:38.183125 systemd-logind[1447]: New session 10 of user core. Feb 13 19:51:38.191032 systemd[1]: Started session-10.scope - Session 10 of User core. Feb 13 19:51:38.337346 sshd[5754]: Connection closed by 10.0.0.1 port 58328 Feb 13 19:51:38.337742 sshd-session[5752]: pam_unix(sshd:session): session closed for user core Feb 13 19:51:38.343426 systemd[1]: sshd@10-10.0.0.104:22-10.0.0.1:58328.service: Deactivated successfully. Feb 13 19:51:38.345810 systemd[1]: session-10.scope: Deactivated successfully. Feb 13 19:51:38.346923 systemd-logind[1447]: Session 10 logged out. Waiting for processes to exit. Feb 13 19:51:38.349576 systemd-logind[1447]: Removed session 10. Feb 13 19:51:39.444228 containerd[1468]: time="2025-02-13T19:51:39.444160351Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:51:39.444979 containerd[1468]: time="2025-02-13T19:51:39.444929333Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1: active requests=0, bytes read=10501081" Feb 13 19:51:39.446201 containerd[1468]: time="2025-02-13T19:51:39.446173057Z" level=info msg="ImageCreate event name:\"sha256:8b7d18f262d5cf6a6343578ad0db68a140c4c9989d9e02c58c27cb5d2c70320f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:51:39.449181 containerd[1468]: time="2025-02-13T19:51:39.449127770Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:a338da9488cbaa83c78457c3d7354d84149969c0480e88dd768e036632ff5b76\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:51:39.449773 containerd[1468]: time="2025-02-13T19:51:39.449745660Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\" with image id \"sha256:8b7d18f262d5cf6a6343578ad0db68a140c4c9989d9e02c58c27cb5d2c70320f\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:a338da9488cbaa83c78457c3d7354d84149969c0480e88dd768e036632ff5b76\", size \"11994117\" in 2.163357256s" Feb 13 19:51:39.449822 containerd[1468]: time="2025-02-13T19:51:39.449775606Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\" returns image reference \"sha256:8b7d18f262d5cf6a6343578ad0db68a140c4c9989d9e02c58c27cb5d2c70320f\"" Feb 13 19:51:39.452137 containerd[1468]: time="2025-02-13T19:51:39.452090879Z" level=info msg="CreateContainer within sandbox \"76eb5868109750c25808ea1c71219c263c1d063e208322be3fc3c91f65657bfe\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Feb 13 19:51:39.469713 containerd[1468]: time="2025-02-13T19:51:39.469653508Z" level=info msg="CreateContainer within sandbox \"76eb5868109750c25808ea1c71219c263c1d063e208322be3fc3c91f65657bfe\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"9c1ed1202ba72daaf2057aabde2f6adf40939b0a6c61481bf317eb0b62b8521c\"" Feb 13 19:51:39.470403 containerd[1468]: time="2025-02-13T19:51:39.470378759Z" level=info msg="StartContainer for \"9c1ed1202ba72daaf2057aabde2f6adf40939b0a6c61481bf317eb0b62b8521c\"" Feb 13 19:51:39.505064 systemd[1]: Started cri-containerd-9c1ed1202ba72daaf2057aabde2f6adf40939b0a6c61481bf317eb0b62b8521c.scope - libcontainer container 9c1ed1202ba72daaf2057aabde2f6adf40939b0a6c61481bf317eb0b62b8521c. Feb 13 19:51:39.680806 containerd[1468]: time="2025-02-13T19:51:39.680738070Z" level=info msg="StartContainer for \"9c1ed1202ba72daaf2057aabde2f6adf40939b0a6c61481bf317eb0b62b8521c\" returns successfully" Feb 13 19:51:39.692832 kubelet[2575]: I0213 19:51:39.692766 2575 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Feb 13 19:51:39.692832 kubelet[2575]: I0213 19:51:39.692836 2575 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Feb 13 19:51:40.543396 kubelet[2575]: I0213 19:51:40.541498 2575 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-sxwqd" podStartSLOduration=24.424278863 podStartE2EDuration="34.541477397s" podCreationTimestamp="2025-02-13 19:51:06 +0000 UTC" firstStartedPulling="2025-02-13 19:51:29.333633484 +0000 UTC m=+35.879949910" lastFinishedPulling="2025-02-13 19:51:39.450832018 +0000 UTC m=+45.997148444" observedRunningTime="2025-02-13 19:51:40.541470124 +0000 UTC m=+47.087786570" watchObservedRunningTime="2025-02-13 19:51:40.541477397 +0000 UTC m=+47.087793823" Feb 13 19:51:43.348736 systemd[1]: Started sshd@11-10.0.0.104:22-10.0.0.1:48746.service - OpenSSH per-connection server daemon (10.0.0.1:48746). Feb 13 19:51:43.397275 sshd[5816]: Accepted publickey for core from 10.0.0.1 port 48746 ssh2: RSA SHA256:8WP2kqV5KzwZsuVRMXRFZkAHZWbkdD5kizbT2H+wOcw Feb 13 19:51:43.399061 sshd-session[5816]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:51:43.403731 systemd-logind[1447]: New session 11 of user core. Feb 13 19:51:43.413150 systemd[1]: Started session-11.scope - Session 11 of User core. Feb 13 19:51:43.556973 sshd[5818]: Connection closed by 10.0.0.1 port 48746 Feb 13 19:51:43.557358 sshd-session[5816]: pam_unix(sshd:session): session closed for user core Feb 13 19:51:43.567108 systemd[1]: sshd@11-10.0.0.104:22-10.0.0.1:48746.service: Deactivated successfully. Feb 13 19:51:43.569073 systemd[1]: session-11.scope: Deactivated successfully. Feb 13 19:51:43.570659 systemd-logind[1447]: Session 11 logged out. Waiting for processes to exit. Feb 13 19:51:43.578386 systemd[1]: Started sshd@12-10.0.0.104:22-10.0.0.1:48760.service - OpenSSH per-connection server daemon (10.0.0.1:48760). Feb 13 19:51:43.579550 systemd-logind[1447]: Removed session 11. Feb 13 19:51:43.616744 sshd[5831]: Accepted publickey for core from 10.0.0.1 port 48760 ssh2: RSA SHA256:8WP2kqV5KzwZsuVRMXRFZkAHZWbkdD5kizbT2H+wOcw Feb 13 19:51:43.618373 sshd-session[5831]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:51:43.622572 systemd-logind[1447]: New session 12 of user core. Feb 13 19:51:43.631038 systemd[1]: Started session-12.scope - Session 12 of User core. Feb 13 19:51:43.835034 sshd[5833]: Connection closed by 10.0.0.1 port 48760 Feb 13 19:51:43.836784 sshd-session[5831]: pam_unix(sshd:session): session closed for user core Feb 13 19:51:43.848127 systemd[1]: sshd@12-10.0.0.104:22-10.0.0.1:48760.service: Deactivated successfully. Feb 13 19:51:43.850132 systemd[1]: session-12.scope: Deactivated successfully. Feb 13 19:51:43.852801 systemd-logind[1447]: Session 12 logged out. Waiting for processes to exit. Feb 13 19:51:43.858170 systemd[1]: Started sshd@13-10.0.0.104:22-10.0.0.1:48776.service - OpenSSH per-connection server daemon (10.0.0.1:48776). Feb 13 19:51:43.859328 systemd-logind[1447]: Removed session 12. Feb 13 19:51:43.903679 sshd[5845]: Accepted publickey for core from 10.0.0.1 port 48776 ssh2: RSA SHA256:8WP2kqV5KzwZsuVRMXRFZkAHZWbkdD5kizbT2H+wOcw Feb 13 19:51:43.906169 sshd-session[5845]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:51:43.910774 systemd-logind[1447]: New session 13 of user core. Feb 13 19:51:43.921190 systemd[1]: Started session-13.scope - Session 13 of User core. Feb 13 19:51:44.113369 sshd[5847]: Connection closed by 10.0.0.1 port 48776 Feb 13 19:51:44.113780 sshd-session[5845]: pam_unix(sshd:session): session closed for user core Feb 13 19:51:44.118051 systemd[1]: sshd@13-10.0.0.104:22-10.0.0.1:48776.service: Deactivated successfully. Feb 13 19:51:44.120007 systemd[1]: session-13.scope: Deactivated successfully. Feb 13 19:51:44.120651 systemd-logind[1447]: Session 13 logged out. Waiting for processes to exit. Feb 13 19:51:44.121695 systemd-logind[1447]: Removed session 13. Feb 13 19:51:49.135764 systemd[1]: Started sshd@14-10.0.0.104:22-10.0.0.1:48784.service - OpenSSH per-connection server daemon (10.0.0.1:48784). Feb 13 19:51:49.188755 sshd[5863]: Accepted publickey for core from 10.0.0.1 port 48784 ssh2: RSA SHA256:8WP2kqV5KzwZsuVRMXRFZkAHZWbkdD5kizbT2H+wOcw Feb 13 19:51:49.190524 sshd-session[5863]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:51:49.195154 systemd-logind[1447]: New session 14 of user core. Feb 13 19:51:49.205007 systemd[1]: Started session-14.scope - Session 14 of User core. Feb 13 19:51:49.332323 sshd[5865]: Connection closed by 10.0.0.1 port 48784 Feb 13 19:51:49.332684 sshd-session[5863]: pam_unix(sshd:session): session closed for user core Feb 13 19:51:49.336834 systemd[1]: sshd@14-10.0.0.104:22-10.0.0.1:48784.service: Deactivated successfully. Feb 13 19:51:49.339003 systemd[1]: session-14.scope: Deactivated successfully. Feb 13 19:51:49.339758 systemd-logind[1447]: Session 14 logged out. Waiting for processes to exit. Feb 13 19:51:49.340771 systemd-logind[1447]: Removed session 14. Feb 13 19:51:53.616504 containerd[1468]: time="2025-02-13T19:51:53.616460122Z" level=info msg="StopPodSandbox for \"c143929d3d37d0fd39eeb807647ccdb8bf35b3017d21a93a728ef6fa513987e4\"" Feb 13 19:51:53.617864 containerd[1468]: time="2025-02-13T19:51:53.616597736Z" level=info msg="TearDown network for sandbox \"c143929d3d37d0fd39eeb807647ccdb8bf35b3017d21a93a728ef6fa513987e4\" successfully" Feb 13 19:51:53.617864 containerd[1468]: time="2025-02-13T19:51:53.616610761Z" level=info msg="StopPodSandbox for \"c143929d3d37d0fd39eeb807647ccdb8bf35b3017d21a93a728ef6fa513987e4\" returns successfully" Feb 13 19:51:53.617864 containerd[1468]: time="2025-02-13T19:51:53.617593481Z" level=info msg="RemovePodSandbox for \"c143929d3d37d0fd39eeb807647ccdb8bf35b3017d21a93a728ef6fa513987e4\"" Feb 13 19:51:53.629561 containerd[1468]: time="2025-02-13T19:51:53.629521725Z" level=info msg="Forcibly stopping sandbox \"c143929d3d37d0fd39eeb807647ccdb8bf35b3017d21a93a728ef6fa513987e4\"" Feb 13 19:51:53.629670 containerd[1468]: time="2025-02-13T19:51:53.629620195Z" level=info msg="TearDown network for sandbox \"c143929d3d37d0fd39eeb807647ccdb8bf35b3017d21a93a728ef6fa513987e4\" successfully" Feb 13 19:51:53.676942 containerd[1468]: time="2025-02-13T19:51:53.676857429Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"c143929d3d37d0fd39eeb807647ccdb8bf35b3017d21a93a728ef6fa513987e4\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:51:53.677193 containerd[1468]: time="2025-02-13T19:51:53.676967100Z" level=info msg="RemovePodSandbox \"c143929d3d37d0fd39eeb807647ccdb8bf35b3017d21a93a728ef6fa513987e4\" returns successfully" Feb 13 19:51:53.677522 containerd[1468]: time="2025-02-13T19:51:53.677474456Z" level=info msg="StopPodSandbox for \"cc0fab9693293af68d0473a6799ea1e188b038744dcf56927b1d4fffd60c2c96\"" Feb 13 19:51:53.677656 containerd[1468]: time="2025-02-13T19:51:53.677612731Z" level=info msg="TearDown network for sandbox \"cc0fab9693293af68d0473a6799ea1e188b038744dcf56927b1d4fffd60c2c96\" successfully" Feb 13 19:51:53.677656 containerd[1468]: time="2025-02-13T19:51:53.677624545Z" level=info msg="StopPodSandbox for \"cc0fab9693293af68d0473a6799ea1e188b038744dcf56927b1d4fffd60c2c96\" returns successfully" Feb 13 19:51:53.677909 containerd[1468]: time="2025-02-13T19:51:53.677883823Z" level=info msg="RemovePodSandbox for \"cc0fab9693293af68d0473a6799ea1e188b038744dcf56927b1d4fffd60c2c96\"" Feb 13 19:51:53.677966 containerd[1468]: time="2025-02-13T19:51:53.677910495Z" level=info msg="Forcibly stopping sandbox \"cc0fab9693293af68d0473a6799ea1e188b038744dcf56927b1d4fffd60c2c96\"" Feb 13 19:51:53.678048 containerd[1468]: time="2025-02-13T19:51:53.677993914Z" level=info msg="TearDown network for sandbox \"cc0fab9693293af68d0473a6799ea1e188b038744dcf56927b1d4fffd60c2c96\" successfully" Feb 13 19:51:53.713996 containerd[1468]: time="2025-02-13T19:51:53.713938938Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"cc0fab9693293af68d0473a6799ea1e188b038744dcf56927b1d4fffd60c2c96\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:51:53.714150 containerd[1468]: time="2025-02-13T19:51:53.714033369Z" level=info msg="RemovePodSandbox \"cc0fab9693293af68d0473a6799ea1e188b038744dcf56927b1d4fffd60c2c96\" returns successfully" Feb 13 19:51:53.714530 containerd[1468]: time="2025-02-13T19:51:53.714494706Z" level=info msg="StopPodSandbox for \"deadb290c2eabe3a6b5ed648ceb2f5121ab56b3728b197df70e6532dc828cd25\"" Feb 13 19:51:53.714694 containerd[1468]: time="2025-02-13T19:51:53.714612513Z" level=info msg="TearDown network for sandbox \"deadb290c2eabe3a6b5ed648ceb2f5121ab56b3728b197df70e6532dc828cd25\" successfully" Feb 13 19:51:53.714694 containerd[1468]: time="2025-02-13T19:51:53.714626269Z" level=info msg="StopPodSandbox for \"deadb290c2eabe3a6b5ed648ceb2f5121ab56b3728b197df70e6532dc828cd25\" returns successfully" Feb 13 19:51:53.715905 containerd[1468]: time="2025-02-13T19:51:53.714965692Z" level=info msg="RemovePodSandbox for \"deadb290c2eabe3a6b5ed648ceb2f5121ab56b3728b197df70e6532dc828cd25\"" Feb 13 19:51:53.715905 containerd[1468]: time="2025-02-13T19:51:53.714996471Z" level=info msg="Forcibly stopping sandbox \"deadb290c2eabe3a6b5ed648ceb2f5121ab56b3728b197df70e6532dc828cd25\"" Feb 13 19:51:53.715905 containerd[1468]: time="2025-02-13T19:51:53.715075352Z" level=info msg="TearDown network for sandbox \"deadb290c2eabe3a6b5ed648ceb2f5121ab56b3728b197df70e6532dc828cd25\" successfully" Feb 13 19:51:53.798422 containerd[1468]: time="2025-02-13T19:51:53.798336522Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"deadb290c2eabe3a6b5ed648ceb2f5121ab56b3728b197df70e6532dc828cd25\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:51:53.798422 containerd[1468]: time="2025-02-13T19:51:53.798430593Z" level=info msg="RemovePodSandbox \"deadb290c2eabe3a6b5ed648ceb2f5121ab56b3728b197df70e6532dc828cd25\" returns successfully" Feb 13 19:51:53.799114 containerd[1468]: time="2025-02-13T19:51:53.799061807Z" level=info msg="StopPodSandbox for \"339897f85c32bf6b58fdb7fa27b7f977f70adf895ce4769419e9e19adb77c464\"" Feb 13 19:51:53.799237 containerd[1468]: time="2025-02-13T19:51:53.799216724Z" level=info msg="TearDown network for sandbox \"339897f85c32bf6b58fdb7fa27b7f977f70adf895ce4769419e9e19adb77c464\" successfully" Feb 13 19:51:53.799281 containerd[1468]: time="2025-02-13T19:51:53.799235691Z" level=info msg="StopPodSandbox for \"339897f85c32bf6b58fdb7fa27b7f977f70adf895ce4769419e9e19adb77c464\" returns successfully" Feb 13 19:51:53.799719 containerd[1468]: time="2025-02-13T19:51:53.799659415Z" level=info msg="RemovePodSandbox for \"339897f85c32bf6b58fdb7fa27b7f977f70adf895ce4769419e9e19adb77c464\"" Feb 13 19:51:53.799719 containerd[1468]: time="2025-02-13T19:51:53.799688401Z" level=info msg="Forcibly stopping sandbox \"339897f85c32bf6b58fdb7fa27b7f977f70adf895ce4769419e9e19adb77c464\"" Feb 13 19:51:53.799837 containerd[1468]: time="2025-02-13T19:51:53.799763325Z" level=info msg="TearDown network for sandbox \"339897f85c32bf6b58fdb7fa27b7f977f70adf895ce4769419e9e19adb77c464\" successfully" Feb 13 19:51:53.864927 containerd[1468]: time="2025-02-13T19:51:53.864856492Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"339897f85c32bf6b58fdb7fa27b7f977f70adf895ce4769419e9e19adb77c464\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:51:53.865094 containerd[1468]: time="2025-02-13T19:51:53.864950413Z" level=info msg="RemovePodSandbox \"339897f85c32bf6b58fdb7fa27b7f977f70adf895ce4769419e9e19adb77c464\" returns successfully" Feb 13 19:51:53.865351 containerd[1468]: time="2025-02-13T19:51:53.865323671Z" level=info msg="StopPodSandbox for \"ed33748ae2f78ecf5ee2c805ac620f4f19dbb1b895af40e625d52cc88bec3832\"" Feb 13 19:51:53.865484 containerd[1468]: time="2025-02-13T19:51:53.865456425Z" level=info msg="TearDown network for sandbox \"ed33748ae2f78ecf5ee2c805ac620f4f19dbb1b895af40e625d52cc88bec3832\" successfully" Feb 13 19:51:53.865484 containerd[1468]: time="2025-02-13T19:51:53.865477326Z" level=info msg="StopPodSandbox for \"ed33748ae2f78ecf5ee2c805ac620f4f19dbb1b895af40e625d52cc88bec3832\" returns successfully" Feb 13 19:51:53.865763 containerd[1468]: time="2025-02-13T19:51:53.865741734Z" level=info msg="RemovePodSandbox for \"ed33748ae2f78ecf5ee2c805ac620f4f19dbb1b895af40e625d52cc88bec3832\"" Feb 13 19:51:53.865830 containerd[1468]: time="2025-02-13T19:51:53.865766532Z" level=info msg="Forcibly stopping sandbox \"ed33748ae2f78ecf5ee2c805ac620f4f19dbb1b895af40e625d52cc88bec3832\"" Feb 13 19:51:53.865936 containerd[1468]: time="2025-02-13T19:51:53.865885660Z" level=info msg="TearDown network for sandbox \"ed33748ae2f78ecf5ee2c805ac620f4f19dbb1b895af40e625d52cc88bec3832\" successfully" Feb 13 19:51:53.902835 containerd[1468]: time="2025-02-13T19:51:53.902719532Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"ed33748ae2f78ecf5ee2c805ac620f4f19dbb1b895af40e625d52cc88bec3832\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:51:53.902835 containerd[1468]: time="2025-02-13T19:51:53.902766002Z" level=info msg="RemovePodSandbox \"ed33748ae2f78ecf5ee2c805ac620f4f19dbb1b895af40e625d52cc88bec3832\" returns successfully" Feb 13 19:51:53.903319 containerd[1468]: time="2025-02-13T19:51:53.903288015Z" level=info msg="StopPodSandbox for \"07cbe351afb62e0b27b5d8a863b7989470635218b3f4519289a79adcd80bb66e\"" Feb 13 19:51:53.903420 containerd[1468]: time="2025-02-13T19:51:53.903386064Z" level=info msg="TearDown network for sandbox \"07cbe351afb62e0b27b5d8a863b7989470635218b3f4519289a79adcd80bb66e\" successfully" Feb 13 19:51:53.903420 containerd[1468]: time="2025-02-13T19:51:53.903413426Z" level=info msg="StopPodSandbox for \"07cbe351afb62e0b27b5d8a863b7989470635218b3f4519289a79adcd80bb66e\" returns successfully" Feb 13 19:51:53.903858 containerd[1468]: time="2025-02-13T19:51:53.903834657Z" level=info msg="RemovePodSandbox for \"07cbe351afb62e0b27b5d8a863b7989470635218b3f4519289a79adcd80bb66e\"" Feb 13 19:51:53.904007 containerd[1468]: time="2025-02-13T19:51:53.903972031Z" level=info msg="Forcibly stopping sandbox \"07cbe351afb62e0b27b5d8a863b7989470635218b3f4519289a79adcd80bb66e\"" Feb 13 19:51:53.904129 containerd[1468]: time="2025-02-13T19:51:53.904061763Z" level=info msg="TearDown network for sandbox \"07cbe351afb62e0b27b5d8a863b7989470635218b3f4519289a79adcd80bb66e\" successfully" Feb 13 19:51:53.965383 containerd[1468]: time="2025-02-13T19:51:53.965328923Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"07cbe351afb62e0b27b5d8a863b7989470635218b3f4519289a79adcd80bb66e\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:51:53.965383 containerd[1468]: time="2025-02-13T19:51:53.965420270Z" level=info msg="RemovePodSandbox \"07cbe351afb62e0b27b5d8a863b7989470635218b3f4519289a79adcd80bb66e\" returns successfully" Feb 13 19:51:53.966493 containerd[1468]: time="2025-02-13T19:51:53.965912836Z" level=info msg="StopPodSandbox for \"a24dd5ac08e1be65e64f1e0e1bcc57fba990025660893d49024096981b2ba3aa\"" Feb 13 19:51:53.966493 containerd[1468]: time="2025-02-13T19:51:53.966034360Z" level=info msg="TearDown network for sandbox \"a24dd5ac08e1be65e64f1e0e1bcc57fba990025660893d49024096981b2ba3aa\" successfully" Feb 13 19:51:53.966493 containerd[1468]: time="2025-02-13T19:51:53.966050291Z" level=info msg="StopPodSandbox for \"a24dd5ac08e1be65e64f1e0e1bcc57fba990025660893d49024096981b2ba3aa\" returns successfully" Feb 13 19:51:53.966493 containerd[1468]: time="2025-02-13T19:51:53.966370697Z" level=info msg="RemovePodSandbox for \"a24dd5ac08e1be65e64f1e0e1bcc57fba990025660893d49024096981b2ba3aa\"" Feb 13 19:51:53.966493 containerd[1468]: time="2025-02-13T19:51:53.966394182Z" level=info msg="Forcibly stopping sandbox \"a24dd5ac08e1be65e64f1e0e1bcc57fba990025660893d49024096981b2ba3aa\"" Feb 13 19:51:53.966715 containerd[1468]: time="2025-02-13T19:51:53.966645715Z" level=info msg="TearDown network for sandbox \"a24dd5ac08e1be65e64f1e0e1bcc57fba990025660893d49024096981b2ba3aa\" successfully" Feb 13 19:51:53.984130 containerd[1468]: time="2025-02-13T19:51:53.984051773Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"a24dd5ac08e1be65e64f1e0e1bcc57fba990025660893d49024096981b2ba3aa\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:51:53.984287 containerd[1468]: time="2025-02-13T19:51:53.984142928Z" level=info msg="RemovePodSandbox \"a24dd5ac08e1be65e64f1e0e1bcc57fba990025660893d49024096981b2ba3aa\" returns successfully" Feb 13 19:51:53.984778 containerd[1468]: time="2025-02-13T19:51:53.984743523Z" level=info msg="StopPodSandbox for \"899fcc18f7bfce71348bcd81eba1c4cca6504cfc06183532c64985ae4deedeb3\"" Feb 13 19:51:53.984916 containerd[1468]: time="2025-02-13T19:51:53.984892590Z" level=info msg="TearDown network for sandbox \"899fcc18f7bfce71348bcd81eba1c4cca6504cfc06183532c64985ae4deedeb3\" successfully" Feb 13 19:51:53.984916 containerd[1468]: time="2025-02-13T19:51:53.984908480Z" level=info msg="StopPodSandbox for \"899fcc18f7bfce71348bcd81eba1c4cca6504cfc06183532c64985ae4deedeb3\" returns successfully" Feb 13 19:51:53.985407 containerd[1468]: time="2025-02-13T19:51:53.985352514Z" level=info msg="RemovePodSandbox for \"899fcc18f7bfce71348bcd81eba1c4cca6504cfc06183532c64985ae4deedeb3\"" Feb 13 19:51:53.985407 containerd[1468]: time="2025-02-13T19:51:53.985385157Z" level=info msg="Forcibly stopping sandbox \"899fcc18f7bfce71348bcd81eba1c4cca6504cfc06183532c64985ae4deedeb3\"" Feb 13 19:51:53.985594 containerd[1468]: time="2025-02-13T19:51:53.985465751Z" level=info msg="TearDown network for sandbox \"899fcc18f7bfce71348bcd81eba1c4cca6504cfc06183532c64985ae4deedeb3\" successfully" Feb 13 19:51:53.995581 containerd[1468]: time="2025-02-13T19:51:53.995282286Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"899fcc18f7bfce71348bcd81eba1c4cca6504cfc06183532c64985ae4deedeb3\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:51:53.995581 containerd[1468]: time="2025-02-13T19:51:53.995381527Z" level=info msg="RemovePodSandbox \"899fcc18f7bfce71348bcd81eba1c4cca6504cfc06183532c64985ae4deedeb3\" returns successfully" Feb 13 19:51:53.996201 containerd[1468]: time="2025-02-13T19:51:53.996170373Z" level=info msg="StopPodSandbox for \"13004b615f48f8fc71da84226292b2e4c272d980ea5cce23fb82164b712c2e30\"" Feb 13 19:51:53.996329 containerd[1468]: time="2025-02-13T19:51:53.996296325Z" level=info msg="TearDown network for sandbox \"13004b615f48f8fc71da84226292b2e4c272d980ea5cce23fb82164b712c2e30\" successfully" Feb 13 19:51:53.996329 containerd[1468]: time="2025-02-13T19:51:53.996314871Z" level=info msg="StopPodSandbox for \"13004b615f48f8fc71da84226292b2e4c272d980ea5cce23fb82164b712c2e30\" returns successfully" Feb 13 19:51:53.996720 containerd[1468]: time="2025-02-13T19:51:53.996680505Z" level=info msg="RemovePodSandbox for \"13004b615f48f8fc71da84226292b2e4c272d980ea5cce23fb82164b712c2e30\"" Feb 13 19:51:53.996782 containerd[1468]: time="2025-02-13T19:51:53.996711133Z" level=info msg="Forcibly stopping sandbox \"13004b615f48f8fc71da84226292b2e4c272d980ea5cce23fb82164b712c2e30\"" Feb 13 19:51:53.996850 containerd[1468]: time="2025-02-13T19:51:53.996831425Z" level=info msg="TearDown network for sandbox \"13004b615f48f8fc71da84226292b2e4c272d980ea5cce23fb82164b712c2e30\" successfully" Feb 13 19:51:54.003979 containerd[1468]: time="2025-02-13T19:51:54.003896175Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"13004b615f48f8fc71da84226292b2e4c272d980ea5cce23fb82164b712c2e30\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:51:54.004181 containerd[1468]: time="2025-02-13T19:51:54.004001146Z" level=info msg="RemovePodSandbox \"13004b615f48f8fc71da84226292b2e4c272d980ea5cce23fb82164b712c2e30\" returns successfully" Feb 13 19:51:54.004863 containerd[1468]: time="2025-02-13T19:51:54.004632118Z" level=info msg="StopPodSandbox for \"22c8c985c21b7d2634c18c69f2f75ef4a82d17959955a493172581323c264379\"" Feb 13 19:51:54.004863 containerd[1468]: time="2025-02-13T19:51:54.004765014Z" level=info msg="TearDown network for sandbox \"22c8c985c21b7d2634c18c69f2f75ef4a82d17959955a493172581323c264379\" successfully" Feb 13 19:51:54.004863 containerd[1468]: time="2025-02-13T19:51:54.004782337Z" level=info msg="StopPodSandbox for \"22c8c985c21b7d2634c18c69f2f75ef4a82d17959955a493172581323c264379\" returns successfully" Feb 13 19:51:54.005410 containerd[1468]: time="2025-02-13T19:51:54.005359997Z" level=info msg="RemovePodSandbox for \"22c8c985c21b7d2634c18c69f2f75ef4a82d17959955a493172581323c264379\"" Feb 13 19:51:54.005483 containerd[1468]: time="2025-02-13T19:51:54.005421876Z" level=info msg="Forcibly stopping sandbox \"22c8c985c21b7d2634c18c69f2f75ef4a82d17959955a493172581323c264379\"" Feb 13 19:51:54.005598 containerd[1468]: time="2025-02-13T19:51:54.005541396Z" level=info msg="TearDown network for sandbox \"22c8c985c21b7d2634c18c69f2f75ef4a82d17959955a493172581323c264379\" successfully" Feb 13 19:51:54.010106 containerd[1468]: time="2025-02-13T19:51:54.010077221Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"22c8c985c21b7d2634c18c69f2f75ef4a82d17959955a493172581323c264379\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:51:54.010192 containerd[1468]: time="2025-02-13T19:51:54.010129962Z" level=info msg="RemovePodSandbox \"22c8c985c21b7d2634c18c69f2f75ef4a82d17959955a493172581323c264379\" returns successfully" Feb 13 19:51:54.010652 containerd[1468]: time="2025-02-13T19:51:54.010611097Z" level=info msg="StopPodSandbox for \"785b364a151170b1b8e2dc79f0469cae0c56fbd88d84bde56a96da9c32fc64d7\"" Feb 13 19:51:54.010763 containerd[1468]: time="2025-02-13T19:51:54.010736968Z" level=info msg="TearDown network for sandbox \"785b364a151170b1b8e2dc79f0469cae0c56fbd88d84bde56a96da9c32fc64d7\" successfully" Feb 13 19:51:54.010794 containerd[1468]: time="2025-02-13T19:51:54.010758319Z" level=info msg="StopPodSandbox for \"785b364a151170b1b8e2dc79f0469cae0c56fbd88d84bde56a96da9c32fc64d7\" returns successfully" Feb 13 19:51:54.011102 containerd[1468]: time="2025-02-13T19:51:54.011071041Z" level=info msg="RemovePodSandbox for \"785b364a151170b1b8e2dc79f0469cae0c56fbd88d84bde56a96da9c32fc64d7\"" Feb 13 19:51:54.011102 containerd[1468]: time="2025-02-13T19:51:54.011094085Z" level=info msg="Forcibly stopping sandbox \"785b364a151170b1b8e2dc79f0469cae0c56fbd88d84bde56a96da9c32fc64d7\"" Feb 13 19:51:54.011248 containerd[1468]: time="2025-02-13T19:51:54.011161655Z" level=info msg="TearDown network for sandbox \"785b364a151170b1b8e2dc79f0469cae0c56fbd88d84bde56a96da9c32fc64d7\" successfully" Feb 13 19:51:54.015531 containerd[1468]: time="2025-02-13T19:51:54.015484130Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"785b364a151170b1b8e2dc79f0469cae0c56fbd88d84bde56a96da9c32fc64d7\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:51:54.015599 containerd[1468]: time="2025-02-13T19:51:54.015533315Z" level=info msg="RemovePodSandbox \"785b364a151170b1b8e2dc79f0469cae0c56fbd88d84bde56a96da9c32fc64d7\" returns successfully" Feb 13 19:51:54.015920 containerd[1468]: time="2025-02-13T19:51:54.015884980Z" level=info msg="StopPodSandbox for \"729c6e6d3b4530ba0b68738b226a62687ecd65bb8522ed1b880437d5541be3c3\"" Feb 13 19:51:54.016081 containerd[1468]: time="2025-02-13T19:51:54.015963241Z" level=info msg="TearDown network for sandbox \"729c6e6d3b4530ba0b68738b226a62687ecd65bb8522ed1b880437d5541be3c3\" successfully" Feb 13 19:51:54.016081 containerd[1468]: time="2025-02-13T19:51:54.015973160Z" level=info msg="StopPodSandbox for \"729c6e6d3b4530ba0b68738b226a62687ecd65bb8522ed1b880437d5541be3c3\" returns successfully" Feb 13 19:51:54.016323 containerd[1468]: time="2025-02-13T19:51:54.016294768Z" level=info msg="RemovePodSandbox for \"729c6e6d3b4530ba0b68738b226a62687ecd65bb8522ed1b880437d5541be3c3\"" Feb 13 19:51:54.016387 containerd[1468]: time="2025-02-13T19:51:54.016326700Z" level=info msg="Forcibly stopping sandbox \"729c6e6d3b4530ba0b68738b226a62687ecd65bb8522ed1b880437d5541be3c3\"" Feb 13 19:51:54.016470 containerd[1468]: time="2025-02-13T19:51:54.016421531Z" level=info msg="TearDown network for sandbox \"729c6e6d3b4530ba0b68738b226a62687ecd65bb8522ed1b880437d5541be3c3\" successfully" Feb 13 19:51:54.020444 containerd[1468]: time="2025-02-13T19:51:54.020386621Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"729c6e6d3b4530ba0b68738b226a62687ecd65bb8522ed1b880437d5541be3c3\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:51:54.020556 containerd[1468]: time="2025-02-13T19:51:54.020452868Z" level=info msg="RemovePodSandbox \"729c6e6d3b4530ba0b68738b226a62687ecd65bb8522ed1b880437d5541be3c3\" returns successfully" Feb 13 19:51:54.020951 containerd[1468]: time="2025-02-13T19:51:54.020925256Z" level=info msg="StopPodSandbox for \"aea5462f72694272b5bc284a0bd9fc173269129df512c011930e89667ad0f97d\"" Feb 13 19:51:54.021052 containerd[1468]: time="2025-02-13T19:51:54.021013355Z" level=info msg="TearDown network for sandbox \"aea5462f72694272b5bc284a0bd9fc173269129df512c011930e89667ad0f97d\" successfully" Feb 13 19:51:54.021052 containerd[1468]: time="2025-02-13T19:51:54.021027873Z" level=info msg="StopPodSandbox for \"aea5462f72694272b5bc284a0bd9fc173269129df512c011930e89667ad0f97d\" returns successfully" Feb 13 19:51:54.021456 containerd[1468]: time="2025-02-13T19:51:54.021407562Z" level=info msg="RemovePodSandbox for \"aea5462f72694272b5bc284a0bd9fc173269129df512c011930e89667ad0f97d\"" Feb 13 19:51:54.021456 containerd[1468]: time="2025-02-13T19:51:54.021451326Z" level=info msg="Forcibly stopping sandbox \"aea5462f72694272b5bc284a0bd9fc173269129df512c011930e89667ad0f97d\"" Feb 13 19:51:54.021614 containerd[1468]: time="2025-02-13T19:51:54.021548703Z" level=info msg="TearDown network for sandbox \"aea5462f72694272b5bc284a0bd9fc173269129df512c011930e89667ad0f97d\" successfully" Feb 13 19:51:54.026176 containerd[1468]: time="2025-02-13T19:51:54.026119376Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"aea5462f72694272b5bc284a0bd9fc173269129df512c011930e89667ad0f97d\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:51:54.026243 containerd[1468]: time="2025-02-13T19:51:54.026186996Z" level=info msg="RemovePodSandbox \"aea5462f72694272b5bc284a0bd9fc173269129df512c011930e89667ad0f97d\" returns successfully" Feb 13 19:51:54.026608 containerd[1468]: time="2025-02-13T19:51:54.026583758Z" level=info msg="StopPodSandbox for \"d7392cd735f64bbb484b37d152068bd95b9e9585d02b78303932e81c6708295c\"" Feb 13 19:51:54.026697 containerd[1468]: time="2025-02-13T19:51:54.026677869Z" level=info msg="TearDown network for sandbox \"d7392cd735f64bbb484b37d152068bd95b9e9585d02b78303932e81c6708295c\" successfully" Feb 13 19:51:54.026697 containerd[1468]: time="2025-02-13T19:51:54.026691344Z" level=info msg="StopPodSandbox for \"d7392cd735f64bbb484b37d152068bd95b9e9585d02b78303932e81c6708295c\" returns successfully" Feb 13 19:51:54.027115 containerd[1468]: time="2025-02-13T19:51:54.027080292Z" level=info msg="RemovePodSandbox for \"d7392cd735f64bbb484b37d152068bd95b9e9585d02b78303932e81c6708295c\"" Feb 13 19:51:54.027151 containerd[1468]: time="2025-02-13T19:51:54.027125809Z" level=info msg="Forcibly stopping sandbox \"d7392cd735f64bbb484b37d152068bd95b9e9585d02b78303932e81c6708295c\"" Feb 13 19:51:54.027275 containerd[1468]: time="2025-02-13T19:51:54.027228977Z" level=info msg="TearDown network for sandbox \"d7392cd735f64bbb484b37d152068bd95b9e9585d02b78303932e81c6708295c\" successfully" Feb 13 19:51:54.031547 containerd[1468]: time="2025-02-13T19:51:54.031508862Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"d7392cd735f64bbb484b37d152068bd95b9e9585d02b78303932e81c6708295c\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:51:54.031620 containerd[1468]: time="2025-02-13T19:51:54.031575269Z" level=info msg="RemovePodSandbox \"d7392cd735f64bbb484b37d152068bd95b9e9585d02b78303932e81c6708295c\" returns successfully" Feb 13 19:51:54.032009 containerd[1468]: time="2025-02-13T19:51:54.031967383Z" level=info msg="StopPodSandbox for \"86a985525093286f84ab532b093323eb20cae87c984957fc89def62353843bd5\"" Feb 13 19:51:54.032191 containerd[1468]: time="2025-02-13T19:51:54.032066062Z" level=info msg="TearDown network for sandbox \"86a985525093286f84ab532b093323eb20cae87c984957fc89def62353843bd5\" successfully" Feb 13 19:51:54.032191 containerd[1468]: time="2025-02-13T19:51:54.032076983Z" level=info msg="StopPodSandbox for \"86a985525093286f84ab532b093323eb20cae87c984957fc89def62353843bd5\" returns successfully" Feb 13 19:51:54.032351 containerd[1468]: time="2025-02-13T19:51:54.032310943Z" level=info msg="RemovePodSandbox for \"86a985525093286f84ab532b093323eb20cae87c984957fc89def62353843bd5\"" Feb 13 19:51:54.032351 containerd[1468]: time="2025-02-13T19:51:54.032345539Z" level=info msg="Forcibly stopping sandbox \"86a985525093286f84ab532b093323eb20cae87c984957fc89def62353843bd5\"" Feb 13 19:51:54.032498 containerd[1468]: time="2025-02-13T19:51:54.032441183Z" level=info msg="TearDown network for sandbox \"86a985525093286f84ab532b093323eb20cae87c984957fc89def62353843bd5\" successfully" Feb 13 19:51:54.037094 containerd[1468]: time="2025-02-13T19:51:54.037038016Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"86a985525093286f84ab532b093323eb20cae87c984957fc89def62353843bd5\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:51:54.037239 containerd[1468]: time="2025-02-13T19:51:54.037129261Z" level=info msg="RemovePodSandbox \"86a985525093286f84ab532b093323eb20cae87c984957fc89def62353843bd5\" returns successfully" Feb 13 19:51:54.037612 containerd[1468]: time="2025-02-13T19:51:54.037581059Z" level=info msg="StopPodSandbox for \"e4d2487cfe7f2e3ccbb95c87ade298917e21e16bc54e8c6d5da7846fe11b6c37\"" Feb 13 19:51:54.037729 containerd[1468]: time="2025-02-13T19:51:54.037705228Z" level=info msg="TearDown network for sandbox \"e4d2487cfe7f2e3ccbb95c87ade298917e21e16bc54e8c6d5da7846fe11b6c37\" successfully" Feb 13 19:51:54.037729 containerd[1468]: time="2025-02-13T19:51:54.037724475Z" level=info msg="StopPodSandbox for \"e4d2487cfe7f2e3ccbb95c87ade298917e21e16bc54e8c6d5da7846fe11b6c37\" returns successfully" Feb 13 19:51:54.038297 containerd[1468]: time="2025-02-13T19:51:54.038255926Z" level=info msg="RemovePodSandbox for \"e4d2487cfe7f2e3ccbb95c87ade298917e21e16bc54e8c6d5da7846fe11b6c37\"" Feb 13 19:51:54.038341 containerd[1468]: time="2025-02-13T19:51:54.038305671Z" level=info msg="Forcibly stopping sandbox \"e4d2487cfe7f2e3ccbb95c87ade298917e21e16bc54e8c6d5da7846fe11b6c37\"" Feb 13 19:51:54.038517 containerd[1468]: time="2025-02-13T19:51:54.038457373Z" level=info msg="TearDown network for sandbox \"e4d2487cfe7f2e3ccbb95c87ade298917e21e16bc54e8c6d5da7846fe11b6c37\" successfully" Feb 13 19:51:54.044164 containerd[1468]: time="2025-02-13T19:51:54.044106157Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"e4d2487cfe7f2e3ccbb95c87ade298917e21e16bc54e8c6d5da7846fe11b6c37\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:51:54.044164 containerd[1468]: time="2025-02-13T19:51:54.044166142Z" level=info msg="RemovePodSandbox \"e4d2487cfe7f2e3ccbb95c87ade298917e21e16bc54e8c6d5da7846fe11b6c37\" returns successfully" Feb 13 19:51:54.044521 containerd[1468]: time="2025-02-13T19:51:54.044493431Z" level=info msg="StopPodSandbox for \"6e045cd079efa77c85217df21eca6d5e6fd9d9297b9e159cbd34c6c2ec690015\"" Feb 13 19:51:54.044597 containerd[1468]: time="2025-02-13T19:51:54.044578434Z" level=info msg="TearDown network for sandbox \"6e045cd079efa77c85217df21eca6d5e6fd9d9297b9e159cbd34c6c2ec690015\" successfully" Feb 13 19:51:54.044597 containerd[1468]: time="2025-02-13T19:51:54.044592421Z" level=info msg="StopPodSandbox for \"6e045cd079efa77c85217df21eca6d5e6fd9d9297b9e159cbd34c6c2ec690015\" returns successfully" Feb 13 19:51:54.045012 containerd[1468]: time="2025-02-13T19:51:54.044966860Z" level=info msg="RemovePodSandbox for \"6e045cd079efa77c85217df21eca6d5e6fd9d9297b9e159cbd34c6c2ec690015\"" Feb 13 19:51:54.045012 containerd[1468]: time="2025-02-13T19:51:54.044992790Z" level=info msg="Forcibly stopping sandbox \"6e045cd079efa77c85217df21eca6d5e6fd9d9297b9e159cbd34c6c2ec690015\"" Feb 13 19:51:54.045157 containerd[1468]: time="2025-02-13T19:51:54.045061342Z" level=info msg="TearDown network for sandbox \"6e045cd079efa77c85217df21eca6d5e6fd9d9297b9e159cbd34c6c2ec690015\" successfully" Feb 13 19:51:54.049104 containerd[1468]: time="2025-02-13T19:51:54.049051158Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"6e045cd079efa77c85217df21eca6d5e6fd9d9297b9e159cbd34c6c2ec690015\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:51:54.049104 containerd[1468]: time="2025-02-13T19:51:54.049093930Z" level=info msg="RemovePodSandbox \"6e045cd079efa77c85217df21eca6d5e6fd9d9297b9e159cbd34c6c2ec690015\" returns successfully" Feb 13 19:51:54.049428 containerd[1468]: time="2025-02-13T19:51:54.049390090Z" level=info msg="StopPodSandbox for \"725f264f85d2c0cd16ccc9f133fde5df46c364a1fdc97ca60f100ccf9eecc7c3\"" Feb 13 19:51:54.049511 containerd[1468]: time="2025-02-13T19:51:54.049489681Z" level=info msg="TearDown network for sandbox \"725f264f85d2c0cd16ccc9f133fde5df46c364a1fdc97ca60f100ccf9eecc7c3\" successfully" Feb 13 19:51:54.049511 containerd[1468]: time="2025-02-13T19:51:54.049506182Z" level=info msg="StopPodSandbox for \"725f264f85d2c0cd16ccc9f133fde5df46c364a1fdc97ca60f100ccf9eecc7c3\" returns successfully" Feb 13 19:51:54.049805 containerd[1468]: time="2025-02-13T19:51:54.049779979Z" level=info msg="RemovePodSandbox for \"725f264f85d2c0cd16ccc9f133fde5df46c364a1fdc97ca60f100ccf9eecc7c3\"" Feb 13 19:51:54.049805 containerd[1468]: time="2025-02-13T19:51:54.049805277Z" level=info msg="Forcibly stopping sandbox \"725f264f85d2c0cd16ccc9f133fde5df46c364a1fdc97ca60f100ccf9eecc7c3\"" Feb 13 19:51:54.049938 containerd[1468]: time="2025-02-13T19:51:54.049898527Z" level=info msg="TearDown network for sandbox \"725f264f85d2c0cd16ccc9f133fde5df46c364a1fdc97ca60f100ccf9eecc7c3\" successfully" Feb 13 19:51:54.054021 containerd[1468]: time="2025-02-13T19:51:54.053983245Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"725f264f85d2c0cd16ccc9f133fde5df46c364a1fdc97ca60f100ccf9eecc7c3\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:51:54.054021 containerd[1468]: time="2025-02-13T19:51:54.054021559Z" level=info msg="RemovePodSandbox \"725f264f85d2c0cd16ccc9f133fde5df46c364a1fdc97ca60f100ccf9eecc7c3\" returns successfully" Feb 13 19:51:54.054391 containerd[1468]: time="2025-02-13T19:51:54.054356482Z" level=info msg="StopPodSandbox for \"dee4103e3ed5a58df193e21973191414daa33c176df2ebb53f48f6427c812139\"" Feb 13 19:51:54.054468 containerd[1468]: time="2025-02-13T19:51:54.054446876Z" level=info msg="TearDown network for sandbox \"dee4103e3ed5a58df193e21973191414daa33c176df2ebb53f48f6427c812139\" successfully" Feb 13 19:51:54.054468 containerd[1468]: time="2025-02-13T19:51:54.054460132Z" level=info msg="StopPodSandbox for \"dee4103e3ed5a58df193e21973191414daa33c176df2ebb53f48f6427c812139\" returns successfully" Feb 13 19:51:54.054762 containerd[1468]: time="2025-02-13T19:51:54.054717175Z" level=info msg="RemovePodSandbox for \"dee4103e3ed5a58df193e21973191414daa33c176df2ebb53f48f6427c812139\"" Feb 13 19:51:54.054762 containerd[1468]: time="2025-02-13T19:51:54.054743015Z" level=info msg="Forcibly stopping sandbox \"dee4103e3ed5a58df193e21973191414daa33c176df2ebb53f48f6427c812139\"" Feb 13 19:51:54.054853 containerd[1468]: time="2025-02-13T19:51:54.054808300Z" level=info msg="TearDown network for sandbox \"dee4103e3ed5a58df193e21973191414daa33c176df2ebb53f48f6427c812139\" successfully" Feb 13 19:51:54.058552 containerd[1468]: time="2025-02-13T19:51:54.058511466Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"dee4103e3ed5a58df193e21973191414daa33c176df2ebb53f48f6427c812139\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:51:54.058645 containerd[1468]: time="2025-02-13T19:51:54.058559358Z" level=info msg="RemovePodSandbox \"dee4103e3ed5a58df193e21973191414daa33c176df2ebb53f48f6427c812139\" returns successfully" Feb 13 19:51:54.058899 containerd[1468]: time="2025-02-13T19:51:54.058860677Z" level=info msg="StopPodSandbox for \"49835f65aea5736de7af75d5477e71209d3ab4edb28dc78ae778b60bcf5f2e76\"" Feb 13 19:51:54.058991 containerd[1468]: time="2025-02-13T19:51:54.058973634Z" level=info msg="TearDown network for sandbox \"49835f65aea5736de7af75d5477e71209d3ab4edb28dc78ae778b60bcf5f2e76\" successfully" Feb 13 19:51:54.059036 containerd[1468]: time="2025-02-13T19:51:54.058990917Z" level=info msg="StopPodSandbox for \"49835f65aea5736de7af75d5477e71209d3ab4edb28dc78ae778b60bcf5f2e76\" returns successfully" Feb 13 19:51:54.059365 containerd[1468]: time="2025-02-13T19:51:54.059340499Z" level=info msg="RemovePodSandbox for \"49835f65aea5736de7af75d5477e71209d3ab4edb28dc78ae778b60bcf5f2e76\"" Feb 13 19:51:54.059407 containerd[1468]: time="2025-02-13T19:51:54.059368313Z" level=info msg="Forcibly stopping sandbox \"49835f65aea5736de7af75d5477e71209d3ab4edb28dc78ae778b60bcf5f2e76\"" Feb 13 19:51:54.059491 containerd[1468]: time="2025-02-13T19:51:54.059455711Z" level=info msg="TearDown network for sandbox \"49835f65aea5736de7af75d5477e71209d3ab4edb28dc78ae778b60bcf5f2e76\" successfully" Feb 13 19:51:54.064285 containerd[1468]: time="2025-02-13T19:51:54.064224504Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"49835f65aea5736de7af75d5477e71209d3ab4edb28dc78ae778b60bcf5f2e76\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:51:54.064360 containerd[1468]: time="2025-02-13T19:51:54.064341037Z" level=info msg="RemovePodSandbox \"49835f65aea5736de7af75d5477e71209d3ab4edb28dc78ae778b60bcf5f2e76\" returns successfully" Feb 13 19:51:54.064847 containerd[1468]: time="2025-02-13T19:51:54.064822432Z" level=info msg="StopPodSandbox for \"20a630c64d17851512c12fc278a3718b46879b27d0ffedf611d9175c93614992\"" Feb 13 19:51:54.064970 containerd[1468]: time="2025-02-13T19:51:54.064952071Z" level=info msg="TearDown network for sandbox \"20a630c64d17851512c12fc278a3718b46879b27d0ffedf611d9175c93614992\" successfully" Feb 13 19:51:54.064970 containerd[1468]: time="2025-02-13T19:51:54.064966379Z" level=info msg="StopPodSandbox for \"20a630c64d17851512c12fc278a3718b46879b27d0ffedf611d9175c93614992\" returns successfully" Feb 13 19:51:54.065231 containerd[1468]: time="2025-02-13T19:51:54.065205408Z" level=info msg="RemovePodSandbox for \"20a630c64d17851512c12fc278a3718b46879b27d0ffedf611d9175c93614992\"" Feb 13 19:51:54.065291 containerd[1468]: time="2025-02-13T19:51:54.065235146Z" level=info msg="Forcibly stopping sandbox \"20a630c64d17851512c12fc278a3718b46879b27d0ffedf611d9175c93614992\"" Feb 13 19:51:54.065346 containerd[1468]: time="2025-02-13T19:51:54.065316111Z" level=info msg="TearDown network for sandbox \"20a630c64d17851512c12fc278a3718b46879b27d0ffedf611d9175c93614992\" successfully" Feb 13 19:51:54.069576 containerd[1468]: time="2025-02-13T19:51:54.069548984Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"20a630c64d17851512c12fc278a3718b46879b27d0ffedf611d9175c93614992\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:51:54.069651 containerd[1468]: time="2025-02-13T19:51:54.069594192Z" level=info msg="RemovePodSandbox \"20a630c64d17851512c12fc278a3718b46879b27d0ffedf611d9175c93614992\" returns successfully" Feb 13 19:51:54.070007 containerd[1468]: time="2025-02-13T19:51:54.069968470Z" level=info msg="StopPodSandbox for \"d52193e0d0af1f395378fbf458da41b44ce26d784835ca82240d3687ed53ca6f\"" Feb 13 19:51:54.070169 containerd[1468]: time="2025-02-13T19:51:54.070117276Z" level=info msg="TearDown network for sandbox \"d52193e0d0af1f395378fbf458da41b44ce26d784835ca82240d3687ed53ca6f\" successfully" Feb 13 19:51:54.070169 containerd[1468]: time="2025-02-13T19:51:54.070136413Z" level=info msg="StopPodSandbox for \"d52193e0d0af1f395378fbf458da41b44ce26d784835ca82240d3687ed53ca6f\" returns successfully" Feb 13 19:51:54.070796 containerd[1468]: time="2025-02-13T19:51:54.070755373Z" level=info msg="RemovePodSandbox for \"d52193e0d0af1f395378fbf458da41b44ce26d784835ca82240d3687ed53ca6f\"" Feb 13 19:51:54.070864 containerd[1468]: time="2025-02-13T19:51:54.070803214Z" level=info msg="Forcibly stopping sandbox \"d52193e0d0af1f395378fbf458da41b44ce26d784835ca82240d3687ed53ca6f\"" Feb 13 19:51:54.071036 containerd[1468]: time="2025-02-13T19:51:54.070977950Z" level=info msg="TearDown network for sandbox \"d52193e0d0af1f395378fbf458da41b44ce26d784835ca82240d3687ed53ca6f\" successfully" Feb 13 19:51:54.075498 containerd[1468]: time="2025-02-13T19:51:54.075454631Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"d52193e0d0af1f395378fbf458da41b44ce26d784835ca82240d3687ed53ca6f\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:51:54.075556 containerd[1468]: time="2025-02-13T19:51:54.075509407Z" level=info msg="RemovePodSandbox \"d52193e0d0af1f395378fbf458da41b44ce26d784835ca82240d3687ed53ca6f\" returns successfully" Feb 13 19:51:54.075939 containerd[1468]: time="2025-02-13T19:51:54.075907291Z" level=info msg="StopPodSandbox for \"1d976eb55d0e17adfbd3d8051a2fdf3026e7061589ae2a3f34816e887a92e46f\"" Feb 13 19:51:54.076071 containerd[1468]: time="2025-02-13T19:51:54.076013605Z" level=info msg="TearDown network for sandbox \"1d976eb55d0e17adfbd3d8051a2fdf3026e7061589ae2a3f34816e887a92e46f\" successfully" Feb 13 19:51:54.076071 containerd[1468]: time="2025-02-13T19:51:54.076068541Z" level=info msg="StopPodSandbox for \"1d976eb55d0e17adfbd3d8051a2fdf3026e7061589ae2a3f34816e887a92e46f\" returns successfully" Feb 13 19:51:54.076432 containerd[1468]: time="2025-02-13T19:51:54.076381794Z" level=info msg="RemovePodSandbox for \"1d976eb55d0e17adfbd3d8051a2fdf3026e7061589ae2a3f34816e887a92e46f\"" Feb 13 19:51:54.076432 containerd[1468]: time="2025-02-13T19:51:54.076420177Z" level=info msg="Forcibly stopping sandbox \"1d976eb55d0e17adfbd3d8051a2fdf3026e7061589ae2a3f34816e887a92e46f\"" Feb 13 19:51:54.076621 containerd[1468]: time="2025-02-13T19:51:54.076494500Z" level=info msg="TearDown network for sandbox \"1d976eb55d0e17adfbd3d8051a2fdf3026e7061589ae2a3f34816e887a92e46f\" successfully" Feb 13 19:51:54.081152 containerd[1468]: time="2025-02-13T19:51:54.081092094Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"1d976eb55d0e17adfbd3d8051a2fdf3026e7061589ae2a3f34816e887a92e46f\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:51:54.081222 containerd[1468]: time="2025-02-13T19:51:54.081159994Z" level=info msg="RemovePodSandbox \"1d976eb55d0e17adfbd3d8051a2fdf3026e7061589ae2a3f34816e887a92e46f\" returns successfully" Feb 13 19:51:54.081574 containerd[1468]: time="2025-02-13T19:51:54.081534905Z" level=info msg="StopPodSandbox for \"4ff7a1ce33e60be86627d63cbaa39e854dfba3ccfcd25d3b18dc72196a2b96da\"" Feb 13 19:51:54.081742 containerd[1468]: time="2025-02-13T19:51:54.081664595Z" level=info msg="TearDown network for sandbox \"4ff7a1ce33e60be86627d63cbaa39e854dfba3ccfcd25d3b18dc72196a2b96da\" successfully" Feb 13 19:51:54.081742 containerd[1468]: time="2025-02-13T19:51:54.081677268Z" level=info msg="StopPodSandbox for \"4ff7a1ce33e60be86627d63cbaa39e854dfba3ccfcd25d3b18dc72196a2b96da\" returns successfully" Feb 13 19:51:54.081943 containerd[1468]: time="2025-02-13T19:51:54.081915566Z" level=info msg="RemovePodSandbox for \"4ff7a1ce33e60be86627d63cbaa39e854dfba3ccfcd25d3b18dc72196a2b96da\"" Feb 13 19:51:54.081943 containerd[1468]: time="2025-02-13T19:51:54.081939963Z" level=info msg="Forcibly stopping sandbox \"4ff7a1ce33e60be86627d63cbaa39e854dfba3ccfcd25d3b18dc72196a2b96da\"" Feb 13 19:51:54.082069 containerd[1468]: time="2025-02-13T19:51:54.082023503Z" level=info msg="TearDown network for sandbox \"4ff7a1ce33e60be86627d63cbaa39e854dfba3ccfcd25d3b18dc72196a2b96da\" successfully" Feb 13 19:51:54.086286 containerd[1468]: time="2025-02-13T19:51:54.086224055Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"4ff7a1ce33e60be86627d63cbaa39e854dfba3ccfcd25d3b18dc72196a2b96da\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:51:54.086389 containerd[1468]: time="2025-02-13T19:51:54.086296103Z" level=info msg="RemovePodSandbox \"4ff7a1ce33e60be86627d63cbaa39e854dfba3ccfcd25d3b18dc72196a2b96da\" returns successfully" Feb 13 19:51:54.086788 containerd[1468]: time="2025-02-13T19:51:54.086709287Z" level=info msg="StopPodSandbox for \"3af9c99589cb59a86666f237cc1b2ea19ad1577ee4287478e22ace117f62bf47\"" Feb 13 19:51:54.086848 containerd[1468]: time="2025-02-13T19:51:54.086809079Z" level=info msg="TearDown network for sandbox \"3af9c99589cb59a86666f237cc1b2ea19ad1577ee4287478e22ace117f62bf47\" successfully" Feb 13 19:51:54.086848 containerd[1468]: time="2025-02-13T19:51:54.086819679Z" level=info msg="StopPodSandbox for \"3af9c99589cb59a86666f237cc1b2ea19ad1577ee4287478e22ace117f62bf47\" returns successfully" Feb 13 19:51:54.087171 containerd[1468]: time="2025-02-13T19:51:54.087138121Z" level=info msg="RemovePodSandbox for \"3af9c99589cb59a86666f237cc1b2ea19ad1577ee4287478e22ace117f62bf47\"" Feb 13 19:51:54.087171 containerd[1468]: time="2025-02-13T19:51:54.087163901Z" level=info msg="Forcibly stopping sandbox \"3af9c99589cb59a86666f237cc1b2ea19ad1577ee4287478e22ace117f62bf47\"" Feb 13 19:51:54.087319 containerd[1468]: time="2025-02-13T19:51:54.087226150Z" level=info msg="TearDown network for sandbox \"3af9c99589cb59a86666f237cc1b2ea19ad1577ee4287478e22ace117f62bf47\" successfully" Feb 13 19:51:54.091338 containerd[1468]: time="2025-02-13T19:51:54.091288166Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"3af9c99589cb59a86666f237cc1b2ea19ad1577ee4287478e22ace117f62bf47\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:51:54.091436 containerd[1468]: time="2025-02-13T19:51:54.091385042Z" level=info msg="RemovePodSandbox \"3af9c99589cb59a86666f237cc1b2ea19ad1577ee4287478e22ace117f62bf47\" returns successfully" Feb 13 19:51:54.093025 containerd[1468]: time="2025-02-13T19:51:54.092975688Z" level=info msg="StopPodSandbox for \"5f114ecd2418a98f886f02a3a5637918502bf71cd1bf3c74f6c1dd0e45034e21\"" Feb 13 19:51:54.093180 containerd[1468]: time="2025-02-13T19:51:54.093154231Z" level=info msg="TearDown network for sandbox \"5f114ecd2418a98f886f02a3a5637918502bf71cd1bf3c74f6c1dd0e45034e21\" successfully" Feb 13 19:51:54.093180 containerd[1468]: time="2025-02-13T19:51:54.093170502Z" level=info msg="StopPodSandbox for \"5f114ecd2418a98f886f02a3a5637918502bf71cd1bf3c74f6c1dd0e45034e21\" returns successfully" Feb 13 19:51:54.093978 containerd[1468]: time="2025-02-13T19:51:54.093938949Z" level=info msg="RemovePodSandbox for \"5f114ecd2418a98f886f02a3a5637918502bf71cd1bf3c74f6c1dd0e45034e21\"" Feb 13 19:51:54.093978 containerd[1468]: time="2025-02-13T19:51:54.093972404Z" level=info msg="Forcibly stopping sandbox \"5f114ecd2418a98f886f02a3a5637918502bf71cd1bf3c74f6c1dd0e45034e21\"" Feb 13 19:51:54.094149 containerd[1468]: time="2025-02-13T19:51:54.094047497Z" level=info msg="TearDown network for sandbox \"5f114ecd2418a98f886f02a3a5637918502bf71cd1bf3c74f6c1dd0e45034e21\" successfully" Feb 13 19:51:54.103495 containerd[1468]: time="2025-02-13T19:51:54.103411019Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"5f114ecd2418a98f886f02a3a5637918502bf71cd1bf3c74f6c1dd0e45034e21\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:51:54.103618 containerd[1468]: time="2025-02-13T19:51:54.103544205Z" level=info msg="RemovePodSandbox \"5f114ecd2418a98f886f02a3a5637918502bf71cd1bf3c74f6c1dd0e45034e21\" returns successfully" Feb 13 19:51:54.104043 containerd[1468]: time="2025-02-13T19:51:54.104005512Z" level=info msg="StopPodSandbox for \"3d79c24a7d960cef54edf513bffc4dd8a68cbc23ca3e3304eafaabcbc4d88e33\"" Feb 13 19:51:54.104181 containerd[1468]: time="2025-02-13T19:51:54.104124820Z" level=info msg="TearDown network for sandbox \"3d79c24a7d960cef54edf513bffc4dd8a68cbc23ca3e3304eafaabcbc4d88e33\" successfully" Feb 13 19:51:54.104181 containerd[1468]: time="2025-02-13T19:51:54.104176500Z" level=info msg="StopPodSandbox for \"3d79c24a7d960cef54edf513bffc4dd8a68cbc23ca3e3304eafaabcbc4d88e33\" returns successfully" Feb 13 19:51:54.104454 containerd[1468]: time="2025-02-13T19:51:54.104428474Z" level=info msg="RemovePodSandbox for \"3d79c24a7d960cef54edf513bffc4dd8a68cbc23ca3e3304eafaabcbc4d88e33\"" Feb 13 19:51:54.104489 containerd[1468]: time="2025-02-13T19:51:54.104456649Z" level=info msg="Forcibly stopping sandbox \"3d79c24a7d960cef54edf513bffc4dd8a68cbc23ca3e3304eafaabcbc4d88e33\"" Feb 13 19:51:54.104596 containerd[1468]: time="2025-02-13T19:51:54.104545389Z" level=info msg="TearDown network for sandbox \"3d79c24a7d960cef54edf513bffc4dd8a68cbc23ca3e3304eafaabcbc4d88e33\" successfully" Feb 13 19:51:54.352548 systemd[1]: Started sshd@15-10.0.0.104:22-10.0.0.1:46548.service - OpenSSH per-connection server daemon (10.0.0.1:46548). Feb 13 19:51:54.395015 sshd[5887]: Accepted publickey for core from 10.0.0.1 port 46548 ssh2: RSA SHA256:8WP2kqV5KzwZsuVRMXRFZkAHZWbkdD5kizbT2H+wOcw Feb 13 19:51:54.396520 sshd-session[5887]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:51:54.401074 systemd-logind[1447]: New session 15 of user core. Feb 13 19:51:54.414056 systemd[1]: Started session-15.scope - Session 15 of User core. Feb 13 19:51:54.556782 sshd[5889]: Connection closed by 10.0.0.1 port 46548 Feb 13 19:51:54.557246 sshd-session[5887]: pam_unix(sshd:session): session closed for user core Feb 13 19:51:54.567394 systemd[1]: sshd@15-10.0.0.104:22-10.0.0.1:46548.service: Deactivated successfully. Feb 13 19:51:54.569210 systemd[1]: session-15.scope: Deactivated successfully. Feb 13 19:51:54.572980 systemd-logind[1447]: Session 15 logged out. Waiting for processes to exit. Feb 13 19:51:54.583187 systemd[1]: Started sshd@16-10.0.0.104:22-10.0.0.1:46552.service - OpenSSH per-connection server daemon (10.0.0.1:46552). Feb 13 19:51:54.584531 systemd-logind[1447]: Removed session 15. Feb 13 19:51:54.625331 sshd[5901]: Accepted publickey for core from 10.0.0.1 port 46552 ssh2: RSA SHA256:8WP2kqV5KzwZsuVRMXRFZkAHZWbkdD5kizbT2H+wOcw Feb 13 19:51:54.627329 sshd-session[5901]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:51:54.631695 systemd-logind[1447]: New session 16 of user core. Feb 13 19:51:54.644995 systemd[1]: Started session-16.scope - Session 16 of User core. Feb 13 19:51:54.879555 containerd[1468]: time="2025-02-13T19:51:54.879358708Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"3d79c24a7d960cef54edf513bffc4dd8a68cbc23ca3e3304eafaabcbc4d88e33\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:51:54.879555 containerd[1468]: time="2025-02-13T19:51:54.879454231Z" level=info msg="RemovePodSandbox \"3d79c24a7d960cef54edf513bffc4dd8a68cbc23ca3e3304eafaabcbc4d88e33\" returns successfully" Feb 13 19:51:54.880411 containerd[1468]: time="2025-02-13T19:51:54.880044366Z" level=info msg="StopPodSandbox for \"d42fa1fe42814b711ad3cb3d471c4742bf3051f8d634e2db762b56c01dc00d5a\"" Feb 13 19:51:54.880411 containerd[1468]: time="2025-02-13T19:51:54.880275710Z" level=info msg="TearDown network for sandbox \"d42fa1fe42814b711ad3cb3d471c4742bf3051f8d634e2db762b56c01dc00d5a\" successfully" Feb 13 19:51:54.880411 containerd[1468]: time="2025-02-13T19:51:54.880290418Z" level=info msg="StopPodSandbox for \"d42fa1fe42814b711ad3cb3d471c4742bf3051f8d634e2db762b56c01dc00d5a\" returns successfully" Feb 13 19:51:54.880695 containerd[1468]: time="2025-02-13T19:51:54.880651282Z" level=info msg="RemovePodSandbox for \"d42fa1fe42814b711ad3cb3d471c4742bf3051f8d634e2db762b56c01dc00d5a\"" Feb 13 19:51:54.880695 containerd[1468]: time="2025-02-13T19:51:54.880674296Z" level=info msg="Forcibly stopping sandbox \"d42fa1fe42814b711ad3cb3d471c4742bf3051f8d634e2db762b56c01dc00d5a\"" Feb 13 19:51:54.880860 containerd[1468]: time="2025-02-13T19:51:54.880759930Z" level=info msg="TearDown network for sandbox \"d42fa1fe42814b711ad3cb3d471c4742bf3051f8d634e2db762b56c01dc00d5a\" successfully" Feb 13 19:51:54.887065 containerd[1468]: time="2025-02-13T19:51:54.887023817Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"d42fa1fe42814b711ad3cb3d471c4742bf3051f8d634e2db762b56c01dc00d5a\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:51:54.887065 containerd[1468]: time="2025-02-13T19:51:54.887077960Z" level=info msg="RemovePodSandbox \"d42fa1fe42814b711ad3cb3d471c4742bf3051f8d634e2db762b56c01dc00d5a\" returns successfully" Feb 13 19:51:54.887469 containerd[1468]: time="2025-02-13T19:51:54.887431169Z" level=info msg="StopPodSandbox for \"5dd1dc9316e1d4cb2cc982113433992b2f865de8097e02c3d7f196ab0750361e\"" Feb 13 19:51:54.887543 containerd[1468]: time="2025-02-13T19:51:54.887523797Z" level=info msg="TearDown network for sandbox \"5dd1dc9316e1d4cb2cc982113433992b2f865de8097e02c3d7f196ab0750361e\" successfully" Feb 13 19:51:54.887543 containerd[1468]: time="2025-02-13T19:51:54.887538626Z" level=info msg="StopPodSandbox for \"5dd1dc9316e1d4cb2cc982113433992b2f865de8097e02c3d7f196ab0750361e\" returns successfully" Feb 13 19:51:54.887820 containerd[1468]: time="2025-02-13T19:51:54.887784848Z" level=info msg="RemovePodSandbox for \"5dd1dc9316e1d4cb2cc982113433992b2f865de8097e02c3d7f196ab0750361e\"" Feb 13 19:51:54.887820 containerd[1468]: time="2025-02-13T19:51:54.887805899Z" level=info msg="Forcibly stopping sandbox \"5dd1dc9316e1d4cb2cc982113433992b2f865de8097e02c3d7f196ab0750361e\"" Feb 13 19:51:54.887923 containerd[1468]: time="2025-02-13T19:51:54.887865504Z" level=info msg="TearDown network for sandbox \"5dd1dc9316e1d4cb2cc982113433992b2f865de8097e02c3d7f196ab0750361e\" successfully" Feb 13 19:51:54.895437 containerd[1468]: time="2025-02-13T19:51:54.895379953Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"5dd1dc9316e1d4cb2cc982113433992b2f865de8097e02c3d7f196ab0750361e\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:51:54.895437 containerd[1468]: time="2025-02-13T19:51:54.895436190Z" level=info msg="RemovePodSandbox \"5dd1dc9316e1d4cb2cc982113433992b2f865de8097e02c3d7f196ab0750361e\" returns successfully" Feb 13 19:51:54.895831 containerd[1468]: time="2025-02-13T19:51:54.895785482Z" level=info msg="StopPodSandbox for \"517d17b8f015a9521e2907d73d8ffe6db808ece1034c7f320e60f66e57f87de3\"" Feb 13 19:51:54.895939 containerd[1468]: time="2025-02-13T19:51:54.895914239Z" level=info msg="TearDown network for sandbox \"517d17b8f015a9521e2907d73d8ffe6db808ece1034c7f320e60f66e57f87de3\" successfully" Feb 13 19:51:54.895939 containerd[1468]: time="2025-02-13T19:51:54.895933557Z" level=info msg="StopPodSandbox for \"517d17b8f015a9521e2907d73d8ffe6db808ece1034c7f320e60f66e57f87de3\" returns successfully" Feb 13 19:51:54.896226 containerd[1468]: time="2025-02-13T19:51:54.896183907Z" level=info msg="RemovePodSandbox for \"517d17b8f015a9521e2907d73d8ffe6db808ece1034c7f320e60f66e57f87de3\"" Feb 13 19:51:54.896226 containerd[1468]: time="2025-02-13T19:51:54.896208935Z" level=info msg="Forcibly stopping sandbox \"517d17b8f015a9521e2907d73d8ffe6db808ece1034c7f320e60f66e57f87de3\"" Feb 13 19:51:54.896311 containerd[1468]: time="2025-02-13T19:51:54.896278499Z" level=info msg="TearDown network for sandbox \"517d17b8f015a9521e2907d73d8ffe6db808ece1034c7f320e60f66e57f87de3\" successfully" Feb 13 19:51:54.921682 containerd[1468]: time="2025-02-13T19:51:54.921616104Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"517d17b8f015a9521e2907d73d8ffe6db808ece1034c7f320e60f66e57f87de3\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:51:54.922005 containerd[1468]: time="2025-02-13T19:51:54.921702089Z" level=info msg="RemovePodSandbox \"517d17b8f015a9521e2907d73d8ffe6db808ece1034c7f320e60f66e57f87de3\" returns successfully" Feb 13 19:51:54.922293 containerd[1468]: time="2025-02-13T19:51:54.922250293Z" level=info msg="StopPodSandbox for \"38f72aa5f0a5062621013b365f7d00faaabff6dd65a3f268813b6c2c1534a463\"" Feb 13 19:51:54.922411 containerd[1468]: time="2025-02-13T19:51:54.922385733Z" level=info msg="TearDown network for sandbox \"38f72aa5f0a5062621013b365f7d00faaabff6dd65a3f268813b6c2c1534a463\" successfully" Feb 13 19:51:54.922411 containerd[1468]: time="2025-02-13T19:51:54.922404900Z" level=info msg="StopPodSandbox for \"38f72aa5f0a5062621013b365f7d00faaabff6dd65a3f268813b6c2c1534a463\" returns successfully" Feb 13 19:51:54.922706 containerd[1468]: time="2025-02-13T19:51:54.922673286Z" level=info msg="RemovePodSandbox for \"38f72aa5f0a5062621013b365f7d00faaabff6dd65a3f268813b6c2c1534a463\"" Feb 13 19:51:54.922706 containerd[1468]: time="2025-02-13T19:51:54.922694747Z" level=info msg="Forcibly stopping sandbox \"38f72aa5f0a5062621013b365f7d00faaabff6dd65a3f268813b6c2c1534a463\"" Feb 13 19:51:54.922806 containerd[1468]: time="2025-02-13T19:51:54.922761205Z" level=info msg="TearDown network for sandbox \"38f72aa5f0a5062621013b365f7d00faaabff6dd65a3f268813b6c2c1534a463\" successfully" Feb 13 19:51:54.929845 containerd[1468]: time="2025-02-13T19:51:54.929777756Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"38f72aa5f0a5062621013b365f7d00faaabff6dd65a3f268813b6c2c1534a463\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:51:54.929952 containerd[1468]: time="2025-02-13T19:51:54.929895844Z" level=info msg="RemovePodSandbox \"38f72aa5f0a5062621013b365f7d00faaabff6dd65a3f268813b6c2c1534a463\" returns successfully" Feb 13 19:51:54.930478 containerd[1468]: time="2025-02-13T19:51:54.930447102Z" level=info msg="StopPodSandbox for \"269c22f856e2916019bfab8db86934d7f53640411cbfd7537abae3e996450f7b\"" Feb 13 19:51:54.930591 containerd[1468]: time="2025-02-13T19:51:54.930568466Z" level=info msg="TearDown network for sandbox \"269c22f856e2916019bfab8db86934d7f53640411cbfd7537abae3e996450f7b\" successfully" Feb 13 19:51:54.930620 containerd[1468]: time="2025-02-13T19:51:54.930587362Z" level=info msg="StopPodSandbox for \"269c22f856e2916019bfab8db86934d7f53640411cbfd7537abae3e996450f7b\" returns successfully" Feb 13 19:51:54.931682 containerd[1468]: time="2025-02-13T19:51:54.930958325Z" level=info msg="RemovePodSandbox for \"269c22f856e2916019bfab8db86934d7f53640411cbfd7537abae3e996450f7b\"" Feb 13 19:51:54.931682 containerd[1468]: time="2025-02-13T19:51:54.930981880Z" level=info msg="Forcibly stopping sandbox \"269c22f856e2916019bfab8db86934d7f53640411cbfd7537abae3e996450f7b\"" Feb 13 19:51:54.931682 containerd[1468]: time="2025-02-13T19:51:54.931057656Z" level=info msg="TearDown network for sandbox \"269c22f856e2916019bfab8db86934d7f53640411cbfd7537abae3e996450f7b\" successfully" Feb 13 19:51:54.944439 containerd[1468]: time="2025-02-13T19:51:54.944362811Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"269c22f856e2916019bfab8db86934d7f53640411cbfd7537abae3e996450f7b\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:51:54.944586 containerd[1468]: time="2025-02-13T19:51:54.944455991Z" level=info msg="RemovePodSandbox \"269c22f856e2916019bfab8db86934d7f53640411cbfd7537abae3e996450f7b\" returns successfully" Feb 13 19:51:54.944985 containerd[1468]: time="2025-02-13T19:51:54.944951132Z" level=info msg="StopPodSandbox for \"ebbbb0cee651304a47f8384965c605db95eabdf650756e32c6b70a0355fbec6d\"" Feb 13 19:51:54.945154 containerd[1468]: time="2025-02-13T19:51:54.945066784Z" level=info msg="TearDown network for sandbox \"ebbbb0cee651304a47f8384965c605db95eabdf650756e32c6b70a0355fbec6d\" successfully" Feb 13 19:51:54.945154 containerd[1468]: time="2025-02-13T19:51:54.945131558Z" level=info msg="StopPodSandbox for \"ebbbb0cee651304a47f8384965c605db95eabdf650756e32c6b70a0355fbec6d\" returns successfully" Feb 13 19:51:54.945459 containerd[1468]: time="2025-02-13T19:51:54.945430533Z" level=info msg="RemovePodSandbox for \"ebbbb0cee651304a47f8384965c605db95eabdf650756e32c6b70a0355fbec6d\"" Feb 13 19:51:54.945509 containerd[1468]: time="2025-02-13T19:51:54.945458076Z" level=info msg="Forcibly stopping sandbox \"ebbbb0cee651304a47f8384965c605db95eabdf650756e32c6b70a0355fbec6d\"" Feb 13 19:51:54.945549 containerd[1468]: time="2025-02-13T19:51:54.945537148Z" level=info msg="TearDown network for sandbox \"ebbbb0cee651304a47f8384965c605db95eabdf650756e32c6b70a0355fbec6d\" successfully" Feb 13 19:51:54.952656 containerd[1468]: time="2025-02-13T19:51:54.952581443Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"ebbbb0cee651304a47f8384965c605db95eabdf650756e32c6b70a0355fbec6d\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:51:54.952852 containerd[1468]: time="2025-02-13T19:51:54.952677498Z" level=info msg="RemovePodSandbox \"ebbbb0cee651304a47f8384965c605db95eabdf650756e32c6b70a0355fbec6d\" returns successfully" Feb 13 19:51:54.953291 containerd[1468]: time="2025-02-13T19:51:54.953229629Z" level=info msg="StopPodSandbox for \"39ec85c7a0b06b4fc75b83977eee38fcc3ba3782a02a024285436451c00bfc91\"" Feb 13 19:51:54.953385 containerd[1468]: time="2025-02-13T19:51:54.953361632Z" level=info msg="TearDown network for sandbox \"39ec85c7a0b06b4fc75b83977eee38fcc3ba3782a02a024285436451c00bfc91\" successfully" Feb 13 19:51:54.953429 containerd[1468]: time="2025-02-13T19:51:54.953385037Z" level=info msg="StopPodSandbox for \"39ec85c7a0b06b4fc75b83977eee38fcc3ba3782a02a024285436451c00bfc91\" returns successfully" Feb 13 19:51:54.953835 containerd[1468]: time="2025-02-13T19:51:54.953804353Z" level=info msg="RemovePodSandbox for \"39ec85c7a0b06b4fc75b83977eee38fcc3ba3782a02a024285436451c00bfc91\"" Feb 13 19:51:54.953835 containerd[1468]: time="2025-02-13T19:51:54.953826195Z" level=info msg="Forcibly stopping sandbox \"39ec85c7a0b06b4fc75b83977eee38fcc3ba3782a02a024285436451c00bfc91\"" Feb 13 19:51:54.954058 containerd[1468]: time="2025-02-13T19:51:54.953905698Z" level=info msg="TearDown network for sandbox \"39ec85c7a0b06b4fc75b83977eee38fcc3ba3782a02a024285436451c00bfc91\" successfully" Feb 13 19:51:54.960290 containerd[1468]: time="2025-02-13T19:51:54.960225441Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"39ec85c7a0b06b4fc75b83977eee38fcc3ba3782a02a024285436451c00bfc91\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:51:54.960429 containerd[1468]: time="2025-02-13T19:51:54.960318279Z" level=info msg="RemovePodSandbox \"39ec85c7a0b06b4fc75b83977eee38fcc3ba3782a02a024285436451c00bfc91\" returns successfully" Feb 13 19:51:54.960899 containerd[1468]: time="2025-02-13T19:51:54.960862195Z" level=info msg="StopPodSandbox for \"b65733053a47537c9f659bcb260db972084c3f0ccd2ff597a9805d0ba86debdc\"" Feb 13 19:51:54.961023 containerd[1468]: time="2025-02-13T19:51:54.961002965Z" level=info msg="TearDown network for sandbox \"b65733053a47537c9f659bcb260db972084c3f0ccd2ff597a9805d0ba86debdc\" successfully" Feb 13 19:51:54.961023 containerd[1468]: time="2025-02-13T19:51:54.961021711Z" level=info msg="StopPodSandbox for \"b65733053a47537c9f659bcb260db972084c3f0ccd2ff597a9805d0ba86debdc\" returns successfully" Feb 13 19:51:54.962615 containerd[1468]: time="2025-02-13T19:51:54.961392904Z" level=info msg="RemovePodSandbox for \"b65733053a47537c9f659bcb260db972084c3f0ccd2ff597a9805d0ba86debdc\"" Feb 13 19:51:54.962615 containerd[1468]: time="2025-02-13T19:51:54.961425377Z" level=info msg="Forcibly stopping sandbox \"b65733053a47537c9f659bcb260db972084c3f0ccd2ff597a9805d0ba86debdc\"" Feb 13 19:51:54.962615 containerd[1468]: time="2025-02-13T19:51:54.961492606Z" level=info msg="TearDown network for sandbox \"b65733053a47537c9f659bcb260db972084c3f0ccd2ff597a9805d0ba86debdc\" successfully" Feb 13 19:51:54.972026 containerd[1468]: time="2025-02-13T19:51:54.971986700Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"b65733053a47537c9f659bcb260db972084c3f0ccd2ff597a9805d0ba86debdc\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:51:54.972119 containerd[1468]: time="2025-02-13T19:51:54.972042096Z" level=info msg="RemovePodSandbox \"b65733053a47537c9f659bcb260db972084c3f0ccd2ff597a9805d0ba86debdc\" returns successfully" Feb 13 19:51:55.109525 sshd[5903]: Connection closed by 10.0.0.1 port 46552 Feb 13 19:51:55.110757 sshd-session[5901]: pam_unix(sshd:session): session closed for user core Feb 13 19:51:55.121497 systemd[1]: sshd@16-10.0.0.104:22-10.0.0.1:46552.service: Deactivated successfully. Feb 13 19:51:55.123851 systemd[1]: session-16.scope: Deactivated successfully. Feb 13 19:51:55.126046 systemd-logind[1447]: Session 16 logged out. Waiting for processes to exit. Feb 13 19:51:55.137236 systemd[1]: Started sshd@17-10.0.0.104:22-10.0.0.1:46568.service - OpenSSH per-connection server daemon (10.0.0.1:46568). Feb 13 19:51:55.138371 systemd-logind[1447]: Removed session 16. Feb 13 19:51:55.178948 sshd[5914]: Accepted publickey for core from 10.0.0.1 port 46568 ssh2: RSA SHA256:8WP2kqV5KzwZsuVRMXRFZkAHZWbkdD5kizbT2H+wOcw Feb 13 19:51:55.181020 sshd-session[5914]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:51:55.185682 systemd-logind[1447]: New session 17 of user core. Feb 13 19:51:55.192132 systemd[1]: Started session-17.scope - Session 17 of User core. Feb 13 19:51:55.974961 sshd[5917]: Connection closed by 10.0.0.1 port 46568 Feb 13 19:51:55.977192 sshd-session[5914]: pam_unix(sshd:session): session closed for user core Feb 13 19:51:55.985690 systemd[1]: sshd@17-10.0.0.104:22-10.0.0.1:46568.service: Deactivated successfully. Feb 13 19:51:55.987723 systemd[1]: session-17.scope: Deactivated successfully. Feb 13 19:51:55.990803 systemd-logind[1447]: Session 17 logged out. Waiting for processes to exit. Feb 13 19:51:55.999525 systemd[1]: Started sshd@18-10.0.0.104:22-10.0.0.1:46576.service - OpenSSH per-connection server daemon (10.0.0.1:46576). Feb 13 19:51:56.001297 systemd-logind[1447]: Removed session 17. Feb 13 19:51:56.048570 sshd[5938]: Accepted publickey for core from 10.0.0.1 port 46576 ssh2: RSA SHA256:8WP2kqV5KzwZsuVRMXRFZkAHZWbkdD5kizbT2H+wOcw Feb 13 19:51:56.050259 sshd-session[5938]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:51:56.055416 systemd-logind[1447]: New session 18 of user core. Feb 13 19:51:56.062113 systemd[1]: Started session-18.scope - Session 18 of User core. Feb 13 19:51:56.423429 sshd[5940]: Connection closed by 10.0.0.1 port 46576 Feb 13 19:51:56.423732 sshd-session[5938]: pam_unix(sshd:session): session closed for user core Feb 13 19:51:56.436251 systemd[1]: sshd@18-10.0.0.104:22-10.0.0.1:46576.service: Deactivated successfully. Feb 13 19:51:56.438426 systemd[1]: session-18.scope: Deactivated successfully. Feb 13 19:51:56.440311 systemd-logind[1447]: Session 18 logged out. Waiting for processes to exit. Feb 13 19:51:56.448511 systemd[1]: Started sshd@19-10.0.0.104:22-10.0.0.1:46588.service - OpenSSH per-connection server daemon (10.0.0.1:46588). Feb 13 19:51:56.449486 systemd-logind[1447]: Removed session 18. Feb 13 19:51:56.487247 sshd[5950]: Accepted publickey for core from 10.0.0.1 port 46588 ssh2: RSA SHA256:8WP2kqV5KzwZsuVRMXRFZkAHZWbkdD5kizbT2H+wOcw Feb 13 19:51:56.488858 sshd-session[5950]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:51:56.493170 systemd-logind[1447]: New session 19 of user core. Feb 13 19:51:56.503192 systemd[1]: Started session-19.scope - Session 19 of User core. Feb 13 19:51:56.633690 sshd[5952]: Connection closed by 10.0.0.1 port 46588 Feb 13 19:51:56.634134 sshd-session[5950]: pam_unix(sshd:session): session closed for user core Feb 13 19:51:56.638173 systemd[1]: sshd@19-10.0.0.104:22-10.0.0.1:46588.service: Deactivated successfully. Feb 13 19:51:56.640607 systemd[1]: session-19.scope: Deactivated successfully. Feb 13 19:51:56.641363 systemd-logind[1447]: Session 19 logged out. Waiting for processes to exit. Feb 13 19:51:56.642301 systemd-logind[1447]: Removed session 19. Feb 13 19:52:00.740517 kubelet[2575]: E0213 19:52:00.740386 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:52:01.647941 systemd[1]: Started sshd@20-10.0.0.104:22-10.0.0.1:41368.service - OpenSSH per-connection server daemon (10.0.0.1:41368). Feb 13 19:52:01.702384 sshd[5989]: Accepted publickey for core from 10.0.0.1 port 41368 ssh2: RSA SHA256:8WP2kqV5KzwZsuVRMXRFZkAHZWbkdD5kizbT2H+wOcw Feb 13 19:52:01.704162 sshd-session[5989]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:52:01.709078 systemd-logind[1447]: New session 20 of user core. Feb 13 19:52:01.718092 systemd[1]: Started session-20.scope - Session 20 of User core. Feb 13 19:52:01.844253 sshd[5991]: Connection closed by 10.0.0.1 port 41368 Feb 13 19:52:01.844716 sshd-session[5989]: pam_unix(sshd:session): session closed for user core Feb 13 19:52:01.849777 systemd[1]: sshd@20-10.0.0.104:22-10.0.0.1:41368.service: Deactivated successfully. Feb 13 19:52:01.852435 systemd[1]: session-20.scope: Deactivated successfully. Feb 13 19:52:01.853267 systemd-logind[1447]: Session 20 logged out. Waiting for processes to exit. Feb 13 19:52:01.854558 systemd-logind[1447]: Removed session 20. Feb 13 19:52:06.306560 kubelet[2575]: I0213 19:52:06.306031 2575 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 13 19:52:06.629170 kubelet[2575]: E0213 19:52:06.628995 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:52:06.858928 systemd[1]: Started sshd@21-10.0.0.104:22-10.0.0.1:41378.service - OpenSSH per-connection server daemon (10.0.0.1:41378). Feb 13 19:52:06.905613 sshd[6007]: Accepted publickey for core from 10.0.0.1 port 41378 ssh2: RSA SHA256:8WP2kqV5KzwZsuVRMXRFZkAHZWbkdD5kizbT2H+wOcw Feb 13 19:52:06.907331 sshd-session[6007]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:52:06.911367 systemd-logind[1447]: New session 21 of user core. Feb 13 19:52:06.918007 systemd[1]: Started session-21.scope - Session 21 of User core. Feb 13 19:52:07.038180 sshd[6009]: Connection closed by 10.0.0.1 port 41378 Feb 13 19:52:07.038525 sshd-session[6007]: pam_unix(sshd:session): session closed for user core Feb 13 19:52:07.042759 systemd[1]: sshd@21-10.0.0.104:22-10.0.0.1:41378.service: Deactivated successfully. Feb 13 19:52:07.045593 systemd[1]: session-21.scope: Deactivated successfully. Feb 13 19:52:07.046376 systemd-logind[1447]: Session 21 logged out. Waiting for processes to exit. Feb 13 19:52:07.047383 systemd-logind[1447]: Removed session 21. Feb 13 19:52:07.487124 kubelet[2575]: I0213 19:52:07.487053 2575 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 13 19:52:12.056470 systemd[1]: Started sshd@22-10.0.0.104:22-10.0.0.1:39342.service - OpenSSH per-connection server daemon (10.0.0.1:39342). Feb 13 19:52:12.108790 sshd[6053]: Accepted publickey for core from 10.0.0.1 port 39342 ssh2: RSA SHA256:8WP2kqV5KzwZsuVRMXRFZkAHZWbkdD5kizbT2H+wOcw Feb 13 19:52:12.111057 sshd-session[6053]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:52:12.116086 systemd-logind[1447]: New session 22 of user core. Feb 13 19:52:12.124088 systemd[1]: Started session-22.scope - Session 22 of User core. Feb 13 19:52:12.267197 sshd[6055]: Connection closed by 10.0.0.1 port 39342 Feb 13 19:52:12.267858 sshd-session[6053]: pam_unix(sshd:session): session closed for user core Feb 13 19:52:12.272428 systemd[1]: sshd@22-10.0.0.104:22-10.0.0.1:39342.service: Deactivated successfully. Feb 13 19:52:12.275034 systemd[1]: session-22.scope: Deactivated successfully. Feb 13 19:52:12.276414 systemd-logind[1447]: Session 22 logged out. Waiting for processes to exit. Feb 13 19:52:12.277556 systemd-logind[1447]: Removed session 22. Feb 13 19:52:17.278668 systemd[1]: Started sshd@23-10.0.0.104:22-10.0.0.1:39344.service - OpenSSH per-connection server daemon (10.0.0.1:39344). Feb 13 19:52:17.328130 sshd[6067]: Accepted publickey for core from 10.0.0.1 port 39344 ssh2: RSA SHA256:8WP2kqV5KzwZsuVRMXRFZkAHZWbkdD5kizbT2H+wOcw Feb 13 19:52:17.329887 sshd-session[6067]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:52:17.333988 systemd-logind[1447]: New session 23 of user core. Feb 13 19:52:17.347041 systemd[1]: Started session-23.scope - Session 23 of User core. Feb 13 19:52:17.497989 sshd[6069]: Connection closed by 10.0.0.1 port 39344 Feb 13 19:52:17.498364 sshd-session[6067]: pam_unix(sshd:session): session closed for user core Feb 13 19:52:17.501340 systemd[1]: sshd@23-10.0.0.104:22-10.0.0.1:39344.service: Deactivated successfully. Feb 13 19:52:17.503361 systemd[1]: session-23.scope: Deactivated successfully. Feb 13 19:52:17.505271 systemd-logind[1447]: Session 23 logged out. Waiting for processes to exit. Feb 13 19:52:17.506235 systemd-logind[1447]: Removed session 23. Feb 13 19:52:17.629424 kubelet[2575]: E0213 19:52:17.629256 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:52:19.629264 kubelet[2575]: E0213 19:52:19.629233 2575 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:52:22.513930 systemd[1]: Started sshd@24-10.0.0.104:22-10.0.0.1:39262.service - OpenSSH per-connection server daemon (10.0.0.1:39262). Feb 13 19:52:22.582631 sshd[6081]: Accepted publickey for core from 10.0.0.1 port 39262 ssh2: RSA SHA256:8WP2kqV5KzwZsuVRMXRFZkAHZWbkdD5kizbT2H+wOcw Feb 13 19:52:22.584628 sshd-session[6081]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:52:22.589335 systemd-logind[1447]: New session 24 of user core. Feb 13 19:52:22.596048 systemd[1]: Started session-24.scope - Session 24 of User core. Feb 13 19:52:22.738556 sshd[6083]: Connection closed by 10.0.0.1 port 39262 Feb 13 19:52:22.740089 sshd-session[6081]: pam_unix(sshd:session): session closed for user core Feb 13 19:52:22.745184 systemd[1]: sshd@24-10.0.0.104:22-10.0.0.1:39262.service: Deactivated successfully. Feb 13 19:52:22.747942 systemd[1]: session-24.scope: Deactivated successfully. Feb 13 19:52:22.748955 systemd-logind[1447]: Session 24 logged out. Waiting for processes to exit. Feb 13 19:52:22.750520 systemd-logind[1447]: Removed session 24.