Feb 13 19:18:23.904012 kernel: Linux version 6.6.74-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.2.1_p20241221 p7) 14.2.1 20241221, GNU ld (Gentoo 2.43 p3) 2.43.1) #1 SMP PREEMPT_DYNAMIC Thu Feb 13 17:40:15 -00 2025 Feb 13 19:18:23.904034 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=f28373bbaddf11103b551b595069cf5faacb27d62f1aab4f9911393ba418b416 Feb 13 19:18:23.904045 kernel: BIOS-provided physical RAM map: Feb 13 19:18:23.904052 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable Feb 13 19:18:23.904058 kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000007fffff] usable Feb 13 19:18:23.904064 kernel: BIOS-e820: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS Feb 13 19:18:23.904072 kernel: BIOS-e820: [mem 0x0000000000808000-0x000000000080afff] usable Feb 13 19:18:23.904078 kernel: BIOS-e820: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS Feb 13 19:18:23.904085 kernel: BIOS-e820: [mem 0x000000000080c000-0x0000000000810fff] usable Feb 13 19:18:23.904091 kernel: BIOS-e820: [mem 0x0000000000811000-0x00000000008fffff] ACPI NVS Feb 13 19:18:23.904098 kernel: BIOS-e820: [mem 0x0000000000900000-0x000000009bd3efff] usable Feb 13 19:18:23.904106 kernel: BIOS-e820: [mem 0x000000009bd3f000-0x000000009bdfffff] reserved Feb 13 19:18:23.904113 kernel: BIOS-e820: [mem 0x000000009be00000-0x000000009c8ecfff] usable Feb 13 19:18:23.904119 kernel: BIOS-e820: [mem 0x000000009c8ed000-0x000000009cb6cfff] reserved Feb 13 19:18:23.904127 kernel: BIOS-e820: [mem 0x000000009cb6d000-0x000000009cb7efff] ACPI data Feb 13 19:18:23.904134 kernel: BIOS-e820: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS Feb 13 19:18:23.904143 kernel: BIOS-e820: [mem 0x000000009cbff000-0x000000009ce91fff] usable Feb 13 19:18:23.904150 kernel: BIOS-e820: [mem 0x000000009ce92000-0x000000009ce95fff] reserved Feb 13 19:18:23.904157 kernel: BIOS-e820: [mem 0x000000009ce96000-0x000000009ce97fff] ACPI NVS Feb 13 19:18:23.904164 kernel: BIOS-e820: [mem 0x000000009ce98000-0x000000009cedbfff] usable Feb 13 19:18:23.904171 kernel: BIOS-e820: [mem 0x000000009cedc000-0x000000009cf5ffff] reserved Feb 13 19:18:23.904178 kernel: BIOS-e820: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS Feb 13 19:18:23.904185 kernel: BIOS-e820: [mem 0x00000000e0000000-0x00000000efffffff] reserved Feb 13 19:18:23.904192 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Feb 13 19:18:23.904199 kernel: BIOS-e820: [mem 0x00000000ffc00000-0x00000000ffffffff] reserved Feb 13 19:18:23.904206 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Feb 13 19:18:23.904213 kernel: NX (Execute Disable) protection: active Feb 13 19:18:23.904222 kernel: APIC: Static calls initialized Feb 13 19:18:23.904229 kernel: e820: update [mem 0x9b351018-0x9b35ac57] usable ==> usable Feb 13 19:18:23.904236 kernel: e820: update [mem 0x9b351018-0x9b35ac57] usable ==> usable Feb 13 19:18:23.904243 kernel: e820: update [mem 0x9b314018-0x9b350e57] usable ==> usable Feb 13 19:18:23.904250 kernel: e820: update [mem 0x9b314018-0x9b350e57] usable ==> usable Feb 13 19:18:23.904256 kernel: extended physical RAM map: Feb 13 19:18:23.904263 kernel: reserve setup_data: [mem 0x0000000000000000-0x000000000009ffff] usable Feb 13 19:18:23.904270 kernel: reserve setup_data: [mem 0x0000000000100000-0x00000000007fffff] usable Feb 13 19:18:23.904277 kernel: reserve setup_data: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS Feb 13 19:18:23.904284 kernel: reserve setup_data: [mem 0x0000000000808000-0x000000000080afff] usable Feb 13 19:18:23.904291 kernel: reserve setup_data: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS Feb 13 19:18:23.904302 kernel: reserve setup_data: [mem 0x000000000080c000-0x0000000000810fff] usable Feb 13 19:18:23.904318 kernel: reserve setup_data: [mem 0x0000000000811000-0x00000000008fffff] ACPI NVS Feb 13 19:18:23.904341 kernel: reserve setup_data: [mem 0x0000000000900000-0x000000009b314017] usable Feb 13 19:18:23.904354 kernel: reserve setup_data: [mem 0x000000009b314018-0x000000009b350e57] usable Feb 13 19:18:23.904369 kernel: reserve setup_data: [mem 0x000000009b350e58-0x000000009b351017] usable Feb 13 19:18:23.904388 kernel: reserve setup_data: [mem 0x000000009b351018-0x000000009b35ac57] usable Feb 13 19:18:23.904402 kernel: reserve setup_data: [mem 0x000000009b35ac58-0x000000009bd3efff] usable Feb 13 19:18:23.904421 kernel: reserve setup_data: [mem 0x000000009bd3f000-0x000000009bdfffff] reserved Feb 13 19:18:23.904438 kernel: reserve setup_data: [mem 0x000000009be00000-0x000000009c8ecfff] usable Feb 13 19:18:23.904451 kernel: reserve setup_data: [mem 0x000000009c8ed000-0x000000009cb6cfff] reserved Feb 13 19:18:23.904464 kernel: reserve setup_data: [mem 0x000000009cb6d000-0x000000009cb7efff] ACPI data Feb 13 19:18:23.904478 kernel: reserve setup_data: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS Feb 13 19:18:23.904494 kernel: reserve setup_data: [mem 0x000000009cbff000-0x000000009ce91fff] usable Feb 13 19:18:23.904507 kernel: reserve setup_data: [mem 0x000000009ce92000-0x000000009ce95fff] reserved Feb 13 19:18:23.904521 kernel: reserve setup_data: [mem 0x000000009ce96000-0x000000009ce97fff] ACPI NVS Feb 13 19:18:23.904535 kernel: reserve setup_data: [mem 0x000000009ce98000-0x000000009cedbfff] usable Feb 13 19:18:23.904553 kernel: reserve setup_data: [mem 0x000000009cedc000-0x000000009cf5ffff] reserved Feb 13 19:18:23.904570 kernel: reserve setup_data: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS Feb 13 19:18:23.904583 kernel: reserve setup_data: [mem 0x00000000e0000000-0x00000000efffffff] reserved Feb 13 19:18:23.904596 kernel: reserve setup_data: [mem 0x00000000feffc000-0x00000000feffffff] reserved Feb 13 19:18:23.904609 kernel: reserve setup_data: [mem 0x00000000ffc00000-0x00000000ffffffff] reserved Feb 13 19:18:23.904622 kernel: reserve setup_data: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Feb 13 19:18:23.904638 kernel: efi: EFI v2.7 by EDK II Feb 13 19:18:23.904652 kernel: efi: SMBIOS=0x9c988000 ACPI=0x9cb7e000 ACPI 2.0=0x9cb7e014 MEMATTR=0x9ba0d198 RNG=0x9cb73018 Feb 13 19:18:23.904665 kernel: random: crng init done Feb 13 19:18:23.904682 kernel: efi: Remove mem142: MMIO range=[0xffc00000-0xffffffff] (4MB) from e820 map Feb 13 19:18:23.904695 kernel: e820: remove [mem 0xffc00000-0xffffffff] reserved Feb 13 19:18:23.904709 kernel: secureboot: Secure boot disabled Feb 13 19:18:23.904727 kernel: SMBIOS 2.8 present. Feb 13 19:18:23.904741 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS unknown 02/02/2022 Feb 13 19:18:23.904754 kernel: Hypervisor detected: KVM Feb 13 19:18:23.904767 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Feb 13 19:18:23.904784 kernel: kvm-clock: using sched offset of 2568459983 cycles Feb 13 19:18:23.904798 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Feb 13 19:18:23.904811 kernel: tsc: Detected 2794.748 MHz processor Feb 13 19:18:23.904829 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Feb 13 19:18:23.904842 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Feb 13 19:18:23.904870 kernel: last_pfn = 0x9cedc max_arch_pfn = 0x400000000 Feb 13 19:18:23.904889 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs Feb 13 19:18:23.904903 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Feb 13 19:18:23.904917 kernel: Using GB pages for direct mapping Feb 13 19:18:23.904933 kernel: ACPI: Early table checksum verification disabled Feb 13 19:18:23.904947 kernel: ACPI: RSDP 0x000000009CB7E014 000024 (v02 BOCHS ) Feb 13 19:18:23.904961 kernel: ACPI: XSDT 0x000000009CB7D0E8 000054 (v01 BOCHS BXPC 00000001 01000013) Feb 13 19:18:23.904978 kernel: ACPI: FACP 0x000000009CB79000 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Feb 13 19:18:23.904991 kernel: ACPI: DSDT 0x000000009CB7A000 0021A8 (v01 BOCHS BXPC 00000001 BXPC 00000001) Feb 13 19:18:23.905004 kernel: ACPI: FACS 0x000000009CBDD000 000040 Feb 13 19:18:23.905024 kernel: ACPI: APIC 0x000000009CB78000 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) Feb 13 19:18:23.905038 kernel: ACPI: HPET 0x000000009CB77000 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Feb 13 19:18:23.905055 kernel: ACPI: MCFG 0x000000009CB76000 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Feb 13 19:18:23.905068 kernel: ACPI: WAET 0x000000009CB75000 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Feb 13 19:18:23.905082 kernel: ACPI: BGRT 0x000000009CB74000 000038 (v01 INTEL EDK2 00000002 01000013) Feb 13 19:18:23.905099 kernel: ACPI: Reserving FACP table memory at [mem 0x9cb79000-0x9cb790f3] Feb 13 19:18:23.905113 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cb7a000-0x9cb7c1a7] Feb 13 19:18:23.905127 kernel: ACPI: Reserving FACS table memory at [mem 0x9cbdd000-0x9cbdd03f] Feb 13 19:18:23.905141 kernel: ACPI: Reserving APIC table memory at [mem 0x9cb78000-0x9cb7808f] Feb 13 19:18:23.905160 kernel: ACPI: Reserving HPET table memory at [mem 0x9cb77000-0x9cb77037] Feb 13 19:18:23.905176 kernel: ACPI: Reserving MCFG table memory at [mem 0x9cb76000-0x9cb7603b] Feb 13 19:18:23.905184 kernel: ACPI: Reserving WAET table memory at [mem 0x9cb75000-0x9cb75027] Feb 13 19:18:23.905191 kernel: ACPI: Reserving BGRT table memory at [mem 0x9cb74000-0x9cb74037] Feb 13 19:18:23.905198 kernel: No NUMA configuration found Feb 13 19:18:23.905206 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cedbfff] Feb 13 19:18:23.905213 kernel: NODE_DATA(0) allocated [mem 0x9ce3a000-0x9ce3ffff] Feb 13 19:18:23.905220 kernel: Zone ranges: Feb 13 19:18:23.905228 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Feb 13 19:18:23.905237 kernel: DMA32 [mem 0x0000000001000000-0x000000009cedbfff] Feb 13 19:18:23.905245 kernel: Normal empty Feb 13 19:18:23.905252 kernel: Movable zone start for each node Feb 13 19:18:23.905259 kernel: Early memory node ranges Feb 13 19:18:23.905266 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] Feb 13 19:18:23.905274 kernel: node 0: [mem 0x0000000000100000-0x00000000007fffff] Feb 13 19:18:23.905281 kernel: node 0: [mem 0x0000000000808000-0x000000000080afff] Feb 13 19:18:23.905288 kernel: node 0: [mem 0x000000000080c000-0x0000000000810fff] Feb 13 19:18:23.905296 kernel: node 0: [mem 0x0000000000900000-0x000000009bd3efff] Feb 13 19:18:23.905303 kernel: node 0: [mem 0x000000009be00000-0x000000009c8ecfff] Feb 13 19:18:23.905312 kernel: node 0: [mem 0x000000009cbff000-0x000000009ce91fff] Feb 13 19:18:23.905320 kernel: node 0: [mem 0x000000009ce98000-0x000000009cedbfff] Feb 13 19:18:23.905327 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cedbfff] Feb 13 19:18:23.905335 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Feb 13 19:18:23.905342 kernel: On node 0, zone DMA: 96 pages in unavailable ranges Feb 13 19:18:23.905357 kernel: On node 0, zone DMA: 8 pages in unavailable ranges Feb 13 19:18:23.905367 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Feb 13 19:18:23.905382 kernel: On node 0, zone DMA: 239 pages in unavailable ranges Feb 13 19:18:23.905390 kernel: On node 0, zone DMA32: 193 pages in unavailable ranges Feb 13 19:18:23.905398 kernel: On node 0, zone DMA32: 786 pages in unavailable ranges Feb 13 19:18:23.905405 kernel: On node 0, zone DMA32: 6 pages in unavailable ranges Feb 13 19:18:23.905413 kernel: On node 0, zone DMA32: 12580 pages in unavailable ranges Feb 13 19:18:23.905423 kernel: ACPI: PM-Timer IO Port: 0x608 Feb 13 19:18:23.905431 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Feb 13 19:18:23.905438 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Feb 13 19:18:23.905446 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Feb 13 19:18:23.905454 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Feb 13 19:18:23.905463 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Feb 13 19:18:23.905471 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Feb 13 19:18:23.905479 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Feb 13 19:18:23.905486 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Feb 13 19:18:23.905494 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Feb 13 19:18:23.905501 kernel: TSC deadline timer available Feb 13 19:18:23.905509 kernel: smpboot: Allowing 4 CPUs, 0 hotplug CPUs Feb 13 19:18:23.905517 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Feb 13 19:18:23.905524 kernel: kvm-guest: KVM setup pv remote TLB flush Feb 13 19:18:23.905534 kernel: kvm-guest: setup PV sched yield Feb 13 19:18:23.905542 kernel: [mem 0x9d000000-0xdfffffff] available for PCI devices Feb 13 19:18:23.905550 kernel: Booting paravirtualized kernel on KVM Feb 13 19:18:23.905557 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Feb 13 19:18:23.905565 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:4 nr_cpu_ids:4 nr_node_ids:1 Feb 13 19:18:23.905573 kernel: percpu: Embedded 58 pages/cpu s197032 r8192 d32344 u524288 Feb 13 19:18:23.905581 kernel: pcpu-alloc: s197032 r8192 d32344 u524288 alloc=1*2097152 Feb 13 19:18:23.905588 kernel: pcpu-alloc: [0] 0 1 2 3 Feb 13 19:18:23.905596 kernel: kvm-guest: PV spinlocks enabled Feb 13 19:18:23.905606 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Feb 13 19:18:23.905615 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=f28373bbaddf11103b551b595069cf5faacb27d62f1aab4f9911393ba418b416 Feb 13 19:18:23.905623 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Feb 13 19:18:23.905630 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Feb 13 19:18:23.905638 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Feb 13 19:18:23.905646 kernel: Fallback order for Node 0: 0 Feb 13 19:18:23.905654 kernel: Built 1 zonelists, mobility grouping on. Total pages: 629460 Feb 13 19:18:23.905661 kernel: Policy zone: DMA32 Feb 13 19:18:23.905671 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Feb 13 19:18:23.905679 kernel: Memory: 2387720K/2565800K available (14336K kernel code, 2301K rwdata, 22852K rodata, 43476K init, 1596K bss, 177824K reserved, 0K cma-reserved) Feb 13 19:18:23.905687 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Feb 13 19:18:23.905695 kernel: ftrace: allocating 37893 entries in 149 pages Feb 13 19:18:23.905703 kernel: ftrace: allocated 149 pages with 4 groups Feb 13 19:18:23.905710 kernel: Dynamic Preempt: voluntary Feb 13 19:18:23.905718 kernel: rcu: Preemptible hierarchical RCU implementation. Feb 13 19:18:23.905726 kernel: rcu: RCU event tracing is enabled. Feb 13 19:18:23.905734 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Feb 13 19:18:23.905745 kernel: Trampoline variant of Tasks RCU enabled. Feb 13 19:18:23.905752 kernel: Rude variant of Tasks RCU enabled. Feb 13 19:18:23.905760 kernel: Tracing variant of Tasks RCU enabled. Feb 13 19:18:23.905768 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Feb 13 19:18:23.905775 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Feb 13 19:18:23.905783 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 Feb 13 19:18:23.905791 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Feb 13 19:18:23.905798 kernel: Console: colour dummy device 80x25 Feb 13 19:18:23.905806 kernel: printk: console [ttyS0] enabled Feb 13 19:18:23.905815 kernel: ACPI: Core revision 20230628 Feb 13 19:18:23.905823 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Feb 13 19:18:23.905831 kernel: APIC: Switch to symmetric I/O mode setup Feb 13 19:18:23.905839 kernel: x2apic enabled Feb 13 19:18:23.905846 kernel: APIC: Switched APIC routing to: physical x2apic Feb 13 19:18:23.905872 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Feb 13 19:18:23.905881 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Feb 13 19:18:23.905888 kernel: kvm-guest: setup PV IPIs Feb 13 19:18:23.905896 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Feb 13 19:18:23.905907 kernel: tsc: Marking TSC unstable due to TSCs unsynchronized Feb 13 19:18:23.905915 kernel: Calibrating delay loop (skipped) preset value.. 5589.49 BogoMIPS (lpj=2794748) Feb 13 19:18:23.905923 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Feb 13 19:18:23.905930 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Feb 13 19:18:23.905938 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Feb 13 19:18:23.905946 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Feb 13 19:18:23.905953 kernel: Spectre V2 : Mitigation: Retpolines Feb 13 19:18:23.905961 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch Feb 13 19:18:23.905968 kernel: Spectre V2 : Spectre v2 / SpectreRSB : Filling RSB on VMEXIT Feb 13 19:18:23.905978 kernel: Spectre V2 : Enabling Speculation Barrier for firmware calls Feb 13 19:18:23.905986 kernel: RETBleed: Mitigation: untrained return thunk Feb 13 19:18:23.905994 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Feb 13 19:18:23.906002 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Feb 13 19:18:23.906009 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! Feb 13 19:18:23.906018 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. Feb 13 19:18:23.906025 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode Feb 13 19:18:23.906033 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Feb 13 19:18:23.906043 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Feb 13 19:18:23.906051 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Feb 13 19:18:23.906058 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Feb 13 19:18:23.906066 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Feb 13 19:18:23.906074 kernel: Freeing SMP alternatives memory: 32K Feb 13 19:18:23.906081 kernel: pid_max: default: 32768 minimum: 301 Feb 13 19:18:23.906089 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Feb 13 19:18:23.906096 kernel: landlock: Up and running. Feb 13 19:18:23.906104 kernel: SELinux: Initializing. Feb 13 19:18:23.906112 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Feb 13 19:18:23.906122 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Feb 13 19:18:23.906130 kernel: smpboot: CPU0: AMD EPYC 7402P 24-Core Processor (family: 0x17, model: 0x31, stepping: 0x0) Feb 13 19:18:23.906140 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Feb 13 19:18:23.906148 kernel: RCU Tasks Rude: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Feb 13 19:18:23.906158 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Feb 13 19:18:23.906166 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. Feb 13 19:18:23.906173 kernel: ... version: 0 Feb 13 19:18:23.906181 kernel: ... bit width: 48 Feb 13 19:18:23.906191 kernel: ... generic registers: 6 Feb 13 19:18:23.906198 kernel: ... value mask: 0000ffffffffffff Feb 13 19:18:23.906206 kernel: ... max period: 00007fffffffffff Feb 13 19:18:23.906214 kernel: ... fixed-purpose events: 0 Feb 13 19:18:23.906222 kernel: ... event mask: 000000000000003f Feb 13 19:18:23.906229 kernel: signal: max sigframe size: 1776 Feb 13 19:18:23.906237 kernel: rcu: Hierarchical SRCU implementation. Feb 13 19:18:23.906244 kernel: rcu: Max phase no-delay instances is 400. Feb 13 19:18:23.906252 kernel: smp: Bringing up secondary CPUs ... Feb 13 19:18:23.906262 kernel: smpboot: x86: Booting SMP configuration: Feb 13 19:18:23.906270 kernel: .... node #0, CPUs: #1 #2 #3 Feb 13 19:18:23.906277 kernel: smp: Brought up 1 node, 4 CPUs Feb 13 19:18:23.906285 kernel: smpboot: Max logical packages: 1 Feb 13 19:18:23.906292 kernel: smpboot: Total of 4 processors activated (22357.98 BogoMIPS) Feb 13 19:18:23.906300 kernel: devtmpfs: initialized Feb 13 19:18:23.906308 kernel: x86/mm: Memory block size: 128MB Feb 13 19:18:23.906315 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00800000-0x00807fff] (32768 bytes) Feb 13 19:18:23.906323 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x0080b000-0x0080bfff] (4096 bytes) Feb 13 19:18:23.906331 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00811000-0x008fffff] (978944 bytes) Feb 13 19:18:23.906341 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cb7f000-0x9cbfefff] (524288 bytes) Feb 13 19:18:23.906349 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9ce96000-0x9ce97fff] (8192 bytes) Feb 13 19:18:23.906357 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cf60000-0x9cffffff] (655360 bytes) Feb 13 19:18:23.906364 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Feb 13 19:18:23.906380 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Feb 13 19:18:23.906388 kernel: pinctrl core: initialized pinctrl subsystem Feb 13 19:18:23.906396 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Feb 13 19:18:23.906404 kernel: audit: initializing netlink subsys (disabled) Feb 13 19:18:23.906414 kernel: audit: type=2000 audit(1739474303.849:1): state=initialized audit_enabled=0 res=1 Feb 13 19:18:23.906421 kernel: thermal_sys: Registered thermal governor 'step_wise' Feb 13 19:18:23.906429 kernel: thermal_sys: Registered thermal governor 'user_space' Feb 13 19:18:23.906437 kernel: cpuidle: using governor menu Feb 13 19:18:23.906444 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Feb 13 19:18:23.906452 kernel: dca service started, version 1.12.1 Feb 13 19:18:23.906459 kernel: PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xe0000000-0xefffffff] (base 0xe0000000) Feb 13 19:18:23.906467 kernel: PCI: Using configuration type 1 for base access Feb 13 19:18:23.906475 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Feb 13 19:18:23.906485 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Feb 13 19:18:23.906493 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Feb 13 19:18:23.906501 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Feb 13 19:18:23.906508 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Feb 13 19:18:23.906516 kernel: ACPI: Added _OSI(Module Device) Feb 13 19:18:23.906523 kernel: ACPI: Added _OSI(Processor Device) Feb 13 19:18:23.906531 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Feb 13 19:18:23.906539 kernel: ACPI: Added _OSI(Processor Aggregator Device) Feb 13 19:18:23.906546 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Feb 13 19:18:23.906556 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Feb 13 19:18:23.906564 kernel: ACPI: Interpreter enabled Feb 13 19:18:23.906571 kernel: ACPI: PM: (supports S0 S3 S5) Feb 13 19:18:23.906579 kernel: ACPI: Using IOAPIC for interrupt routing Feb 13 19:18:23.906587 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Feb 13 19:18:23.906594 kernel: PCI: Using E820 reservations for host bridge windows Feb 13 19:18:23.906602 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Feb 13 19:18:23.906609 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Feb 13 19:18:23.906793 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Feb 13 19:18:23.906956 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Feb 13 19:18:23.907083 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Feb 13 19:18:23.907093 kernel: PCI host bridge to bus 0000:00 Feb 13 19:18:23.907220 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Feb 13 19:18:23.907337 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Feb 13 19:18:23.907460 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Feb 13 19:18:23.907579 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xdfffffff window] Feb 13 19:18:23.907695 kernel: pci_bus 0000:00: root bus resource [mem 0xf0000000-0xfebfffff window] Feb 13 19:18:23.907807 kernel: pci_bus 0000:00: root bus resource [mem 0x380000000000-0x3807ffffffff window] Feb 13 19:18:23.907939 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Feb 13 19:18:23.909201 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 Feb 13 19:18:23.909350 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 Feb 13 19:18:23.909487 kernel: pci 0000:00:01.0: reg 0x10: [mem 0xc0000000-0xc0ffffff pref] Feb 13 19:18:23.909618 kernel: pci 0000:00:01.0: reg 0x18: [mem 0xc1044000-0xc1044fff] Feb 13 19:18:23.909742 kernel: pci 0000:00:01.0: reg 0x30: [mem 0xffff0000-0xffffffff pref] Feb 13 19:18:23.909886 kernel: pci 0000:00:01.0: BAR 0: assigned to efifb Feb 13 19:18:23.910011 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Feb 13 19:18:23.910144 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 Feb 13 19:18:23.910268 kernel: pci 0000:00:02.0: reg 0x10: [io 0x6100-0x611f] Feb 13 19:18:23.910407 kernel: pci 0000:00:02.0: reg 0x14: [mem 0xc1043000-0xc1043fff] Feb 13 19:18:23.910532 kernel: pci 0000:00:02.0: reg 0x20: [mem 0x380000000000-0x380000003fff 64bit pref] Feb 13 19:18:23.912392 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 Feb 13 19:18:23.912526 kernel: pci 0000:00:03.0: reg 0x10: [io 0x6000-0x607f] Feb 13 19:18:23.912650 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xc1042000-0xc1042fff] Feb 13 19:18:23.912776 kernel: pci 0000:00:03.0: reg 0x20: [mem 0x380000004000-0x380000007fff 64bit pref] Feb 13 19:18:23.912924 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 Feb 13 19:18:23.913056 kernel: pci 0000:00:04.0: reg 0x10: [io 0x60e0-0x60ff] Feb 13 19:18:23.913180 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xc1041000-0xc1041fff] Feb 13 19:18:23.913305 kernel: pci 0000:00:04.0: reg 0x20: [mem 0x380000008000-0x38000000bfff 64bit pref] Feb 13 19:18:23.913439 kernel: pci 0000:00:04.0: reg 0x30: [mem 0xfffc0000-0xffffffff pref] Feb 13 19:18:23.913569 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 Feb 13 19:18:23.913693 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Feb 13 19:18:23.913823 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 Feb 13 19:18:23.913983 kernel: pci 0000:00:1f.2: reg 0x20: [io 0x60c0-0x60df] Feb 13 19:18:23.914109 kernel: pci 0000:00:1f.2: reg 0x24: [mem 0xc1040000-0xc1040fff] Feb 13 19:18:23.914252 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 Feb 13 19:18:23.915429 kernel: pci 0000:00:1f.3: reg 0x20: [io 0x6080-0x60bf] Feb 13 19:18:23.915442 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Feb 13 19:18:23.915450 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Feb 13 19:18:23.915458 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Feb 13 19:18:23.915470 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Feb 13 19:18:23.915478 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Feb 13 19:18:23.915486 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Feb 13 19:18:23.915493 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Feb 13 19:18:23.915501 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Feb 13 19:18:23.915509 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Feb 13 19:18:23.915517 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Feb 13 19:18:23.915524 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Feb 13 19:18:23.915532 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Feb 13 19:18:23.915542 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Feb 13 19:18:23.915550 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Feb 13 19:18:23.915557 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Feb 13 19:18:23.915565 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Feb 13 19:18:23.915573 kernel: iommu: Default domain type: Translated Feb 13 19:18:23.915581 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Feb 13 19:18:23.915588 kernel: efivars: Registered efivars operations Feb 13 19:18:23.915596 kernel: PCI: Using ACPI for IRQ routing Feb 13 19:18:23.915604 kernel: PCI: pci_cache_line_size set to 64 bytes Feb 13 19:18:23.915612 kernel: e820: reserve RAM buffer [mem 0x0080b000-0x008fffff] Feb 13 19:18:23.915622 kernel: e820: reserve RAM buffer [mem 0x00811000-0x008fffff] Feb 13 19:18:23.915629 kernel: e820: reserve RAM buffer [mem 0x9b314018-0x9bffffff] Feb 13 19:18:23.915637 kernel: e820: reserve RAM buffer [mem 0x9b351018-0x9bffffff] Feb 13 19:18:23.915645 kernel: e820: reserve RAM buffer [mem 0x9bd3f000-0x9bffffff] Feb 13 19:18:23.915652 kernel: e820: reserve RAM buffer [mem 0x9c8ed000-0x9fffffff] Feb 13 19:18:23.915660 kernel: e820: reserve RAM buffer [mem 0x9ce92000-0x9fffffff] Feb 13 19:18:23.915667 kernel: e820: reserve RAM buffer [mem 0x9cedc000-0x9fffffff] Feb 13 19:18:23.915795 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Feb 13 19:18:23.915997 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Feb 13 19:18:23.916119 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Feb 13 19:18:23.916129 kernel: vgaarb: loaded Feb 13 19:18:23.916137 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Feb 13 19:18:23.916145 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Feb 13 19:18:23.916153 kernel: clocksource: Switched to clocksource kvm-clock Feb 13 19:18:23.916161 kernel: VFS: Disk quotas dquot_6.6.0 Feb 13 19:18:23.916169 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Feb 13 19:18:23.916177 kernel: pnp: PnP ACPI init Feb 13 19:18:23.916319 kernel: system 00:05: [mem 0xe0000000-0xefffffff window] has been reserved Feb 13 19:18:23.916331 kernel: pnp: PnP ACPI: found 6 devices Feb 13 19:18:23.916339 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Feb 13 19:18:23.916347 kernel: NET: Registered PF_INET protocol family Feb 13 19:18:23.916410 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Feb 13 19:18:23.916421 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Feb 13 19:18:23.916431 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Feb 13 19:18:23.916439 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Feb 13 19:18:23.916450 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Feb 13 19:18:23.916458 kernel: TCP: Hash tables configured (established 32768 bind 32768) Feb 13 19:18:23.916466 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Feb 13 19:18:23.916474 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Feb 13 19:18:23.916482 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Feb 13 19:18:23.917467 kernel: NET: Registered PF_XDP protocol family Feb 13 19:18:23.917603 kernel: pci 0000:00:04.0: can't claim BAR 6 [mem 0xfffc0000-0xffffffff pref]: no compatible bridge window Feb 13 19:18:23.917728 kernel: pci 0000:00:04.0: BAR 6: assigned [mem 0x9d000000-0x9d03ffff pref] Feb 13 19:18:23.917848 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Feb 13 19:18:23.917995 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Feb 13 19:18:23.918110 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Feb 13 19:18:23.918222 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xdfffffff window] Feb 13 19:18:23.918333 kernel: pci_bus 0000:00: resource 8 [mem 0xf0000000-0xfebfffff window] Feb 13 19:18:23.918454 kernel: pci_bus 0000:00: resource 9 [mem 0x380000000000-0x3807ffffffff window] Feb 13 19:18:23.918465 kernel: PCI: CLS 0 bytes, default 64 Feb 13 19:18:23.918473 kernel: Initialise system trusted keyrings Feb 13 19:18:23.918486 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Feb 13 19:18:23.918494 kernel: Key type asymmetric registered Feb 13 19:18:23.918502 kernel: Asymmetric key parser 'x509' registered Feb 13 19:18:23.918510 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Feb 13 19:18:23.918518 kernel: io scheduler mq-deadline registered Feb 13 19:18:23.918526 kernel: io scheduler kyber registered Feb 13 19:18:23.918534 kernel: io scheduler bfq registered Feb 13 19:18:23.918542 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Feb 13 19:18:23.918551 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Feb 13 19:18:23.918562 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Feb 13 19:18:23.918572 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Feb 13 19:18:23.918580 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Feb 13 19:18:23.918589 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Feb 13 19:18:23.918597 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Feb 13 19:18:23.918605 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Feb 13 19:18:23.918616 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Feb 13 19:18:23.918624 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Feb 13 19:18:23.918752 kernel: rtc_cmos 00:04: RTC can wake from S4 Feb 13 19:18:23.918885 kernel: rtc_cmos 00:04: registered as rtc0 Feb 13 19:18:23.919003 kernel: rtc_cmos 00:04: setting system clock to 2025-02-13T19:18:23 UTC (1739474303) Feb 13 19:18:23.919122 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram Feb 13 19:18:23.919133 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Feb 13 19:18:23.919142 kernel: efifb: probing for efifb Feb 13 19:18:23.919154 kernel: efifb: framebuffer at 0xc0000000, using 4000k, total 4000k Feb 13 19:18:23.919164 kernel: efifb: mode is 1280x800x32, linelength=5120, pages=1 Feb 13 19:18:23.919173 kernel: efifb: scrolling: redraw Feb 13 19:18:23.919183 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Feb 13 19:18:23.919191 kernel: Console: switching to colour frame buffer device 160x50 Feb 13 19:18:23.919200 kernel: fb0: EFI VGA frame buffer device Feb 13 19:18:23.919208 kernel: pstore: Using crash dump compression: deflate Feb 13 19:18:23.919216 kernel: pstore: Registered efi_pstore as persistent store backend Feb 13 19:18:23.919224 kernel: NET: Registered PF_INET6 protocol family Feb 13 19:18:23.919234 kernel: Segment Routing with IPv6 Feb 13 19:18:23.919242 kernel: In-situ OAM (IOAM) with IPv6 Feb 13 19:18:23.919250 kernel: NET: Registered PF_PACKET protocol family Feb 13 19:18:23.919258 kernel: Key type dns_resolver registered Feb 13 19:18:23.919266 kernel: IPI shorthand broadcast: enabled Feb 13 19:18:23.919274 kernel: sched_clock: Marking stable (602002436, 153429658)->(768903649, -13471555) Feb 13 19:18:23.919282 kernel: registered taskstats version 1 Feb 13 19:18:23.919290 kernel: Loading compiled-in X.509 certificates Feb 13 19:18:23.919298 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.74-flatcar: 6c364ddae48101e091a28279a8d953535f596d53' Feb 13 19:18:23.919311 kernel: Key type .fscrypt registered Feb 13 19:18:23.919318 kernel: Key type fscrypt-provisioning registered Feb 13 19:18:23.919327 kernel: ima: No TPM chip found, activating TPM-bypass! Feb 13 19:18:23.919335 kernel: ima: Allocated hash algorithm: sha1 Feb 13 19:18:23.919343 kernel: ima: No architecture policies found Feb 13 19:18:23.919351 kernel: clk: Disabling unused clocks Feb 13 19:18:23.919359 kernel: Freeing unused kernel image (initmem) memory: 43476K Feb 13 19:18:23.919367 kernel: Write protecting the kernel read-only data: 38912k Feb 13 19:18:23.919383 kernel: Freeing unused kernel image (rodata/data gap) memory: 1724K Feb 13 19:18:23.919394 kernel: Run /init as init process Feb 13 19:18:23.919402 kernel: with arguments: Feb 13 19:18:23.919410 kernel: /init Feb 13 19:18:23.919418 kernel: with environment: Feb 13 19:18:23.919425 kernel: HOME=/ Feb 13 19:18:23.919433 kernel: TERM=linux Feb 13 19:18:23.919441 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Feb 13 19:18:23.919450 systemd[1]: Successfully made /usr/ read-only. Feb 13 19:18:23.919464 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Feb 13 19:18:23.919473 systemd[1]: Detected virtualization kvm. Feb 13 19:18:23.919482 systemd[1]: Detected architecture x86-64. Feb 13 19:18:23.919490 systemd[1]: Running in initrd. Feb 13 19:18:23.919499 systemd[1]: No hostname configured, using default hostname. Feb 13 19:18:23.919508 systemd[1]: Hostname set to . Feb 13 19:18:23.919516 systemd[1]: Initializing machine ID from VM UUID. Feb 13 19:18:23.919524 systemd[1]: Queued start job for default target initrd.target. Feb 13 19:18:23.919536 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Feb 13 19:18:23.919544 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Feb 13 19:18:23.919553 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Feb 13 19:18:23.919562 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Feb 13 19:18:23.919571 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Feb 13 19:18:23.919580 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Feb 13 19:18:23.919590 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Feb 13 19:18:23.919601 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Feb 13 19:18:23.919610 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Feb 13 19:18:23.919618 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Feb 13 19:18:23.919627 systemd[1]: Reached target paths.target - Path Units. Feb 13 19:18:23.919635 systemd[1]: Reached target slices.target - Slice Units. Feb 13 19:18:23.919643 systemd[1]: Reached target swap.target - Swaps. Feb 13 19:18:23.919652 systemd[1]: Reached target timers.target - Timer Units. Feb 13 19:18:23.919670 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Feb 13 19:18:23.919681 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Feb 13 19:18:23.919690 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Feb 13 19:18:23.919698 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Feb 13 19:18:23.919707 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Feb 13 19:18:23.919715 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Feb 13 19:18:23.919724 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Feb 13 19:18:23.919732 systemd[1]: Reached target sockets.target - Socket Units. Feb 13 19:18:23.919741 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Feb 13 19:18:23.919750 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Feb 13 19:18:23.919761 systemd[1]: Finished network-cleanup.service - Network Cleanup. Feb 13 19:18:23.919769 systemd[1]: Starting systemd-fsck-usr.service... Feb 13 19:18:23.919778 systemd[1]: Starting systemd-journald.service - Journal Service... Feb 13 19:18:23.919786 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Feb 13 19:18:23.919795 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 19:18:23.919804 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Feb 13 19:18:23.919812 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Feb 13 19:18:23.919824 systemd[1]: Finished systemd-fsck-usr.service. Feb 13 19:18:23.919832 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Feb 13 19:18:23.919841 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Feb 13 19:18:23.919888 systemd-journald[194]: Collecting audit messages is disabled. Feb 13 19:18:23.919914 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Feb 13 19:18:23.919923 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 19:18:23.919932 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Feb 13 19:18:23.919941 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Feb 13 19:18:23.919950 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Feb 13 19:18:23.919961 kernel: Bridge firewalling registered Feb 13 19:18:23.919969 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Feb 13 19:18:23.919978 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Feb 13 19:18:23.919987 systemd-journald[194]: Journal started Feb 13 19:18:23.920006 systemd-journald[194]: Runtime Journal (/run/log/journal/2193d0da5b2849829be986e4c92b6580) is 6M, max 48.2M, 42.2M free. Feb 13 19:18:23.880093 systemd-modules-load[195]: Inserted module 'overlay' Feb 13 19:18:23.922741 systemd[1]: Started systemd-journald.service - Journal Service. Feb 13 19:18:23.912316 systemd-modules-load[195]: Inserted module 'br_netfilter' Feb 13 19:18:23.924434 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Feb 13 19:18:23.930626 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Feb 13 19:18:23.933101 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 19:18:23.945076 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Feb 13 19:18:23.947279 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Feb 13 19:18:23.952058 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Feb 13 19:18:23.958130 dracut-cmdline[229]: dracut-dracut-053 Feb 13 19:18:23.961257 dracut-cmdline[229]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=f28373bbaddf11103b551b595069cf5faacb27d62f1aab4f9911393ba418b416 Feb 13 19:18:23.993440 systemd-resolved[238]: Positive Trust Anchors: Feb 13 19:18:23.993455 systemd-resolved[238]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Feb 13 19:18:23.993486 systemd-resolved[238]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Feb 13 19:18:23.995923 systemd-resolved[238]: Defaulting to hostname 'linux'. Feb 13 19:18:23.996974 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Feb 13 19:18:24.002399 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Feb 13 19:18:24.059896 kernel: SCSI subsystem initialized Feb 13 19:18:24.068875 kernel: Loading iSCSI transport class v2.0-870. Feb 13 19:18:24.079897 kernel: iscsi: registered transport (tcp) Feb 13 19:18:24.103898 kernel: iscsi: registered transport (qla4xxx) Feb 13 19:18:24.103933 kernel: QLogic iSCSI HBA Driver Feb 13 19:18:24.156539 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Feb 13 19:18:24.170092 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Feb 13 19:18:24.196774 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Feb 13 19:18:24.196835 kernel: device-mapper: uevent: version 1.0.3 Feb 13 19:18:24.196868 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Feb 13 19:18:24.238890 kernel: raid6: avx2x4 gen() 28105 MB/s Feb 13 19:18:24.255880 kernel: raid6: avx2x2 gen() 30191 MB/s Feb 13 19:18:24.272952 kernel: raid6: avx2x1 gen() 25779 MB/s Feb 13 19:18:24.272975 kernel: raid6: using algorithm avx2x2 gen() 30191 MB/s Feb 13 19:18:24.290959 kernel: raid6: .... xor() 19829 MB/s, rmw enabled Feb 13 19:18:24.291002 kernel: raid6: using avx2x2 recovery algorithm Feb 13 19:18:24.310892 kernel: xor: automatically using best checksumming function avx Feb 13 19:18:24.457901 kernel: Btrfs loaded, zoned=no, fsverity=no Feb 13 19:18:24.471906 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Feb 13 19:18:24.484029 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Feb 13 19:18:24.499141 systemd-udevd[417]: Using default interface naming scheme 'v255'. Feb 13 19:18:24.504506 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Feb 13 19:18:24.514020 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Feb 13 19:18:24.530131 dracut-pre-trigger[424]: rd.md=0: removing MD RAID activation Feb 13 19:18:24.570335 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Feb 13 19:18:24.583043 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Feb 13 19:18:24.652895 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Feb 13 19:18:24.661518 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Feb 13 19:18:24.674203 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Feb 13 19:18:24.677706 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Feb 13 19:18:24.680145 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Feb 13 19:18:24.682731 systemd[1]: Reached target remote-fs.target - Remote File Systems. Feb 13 19:18:24.692056 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Feb 13 19:18:24.701833 kernel: virtio_blk virtio1: 4/0/0 default/read/poll queues Feb 13 19:18:24.723223 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Feb 13 19:18:24.723477 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Feb 13 19:18:24.723494 kernel: GPT:9289727 != 19775487 Feb 13 19:18:24.723508 kernel: GPT:Alternate GPT header not at the end of the disk. Feb 13 19:18:24.723521 kernel: GPT:9289727 != 19775487 Feb 13 19:18:24.723534 kernel: GPT: Use GNU Parted to correct GPT errors. Feb 13 19:18:24.723548 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Feb 13 19:18:24.723579 kernel: libata version 3.00 loaded. Feb 13 19:18:24.723598 kernel: cryptd: max_cpu_qlen set to 1000 Feb 13 19:18:24.710193 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Feb 13 19:18:24.728879 kernel: ahci 0000:00:1f.2: version 3.0 Feb 13 19:18:24.747476 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Feb 13 19:18:24.747492 kernel: ahci 0000:00:1f.2: AHCI 0001.0000 32 slots 6 ports 1.5 Gbps 0x3f impl SATA mode Feb 13 19:18:24.747648 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Feb 13 19:18:24.747801 kernel: AVX2 version of gcm_enc/dec engaged. Feb 13 19:18:24.747812 kernel: AES CTR mode by8 optimization enabled Feb 13 19:18:24.747823 kernel: scsi host0: ahci Feb 13 19:18:24.748418 kernel: scsi host1: ahci Feb 13 19:18:24.748606 kernel: scsi host2: ahci Feb 13 19:18:24.748762 kernel: scsi host3: ahci Feb 13 19:18:24.749270 kernel: scsi host4: ahci Feb 13 19:18:24.749441 kernel: scsi host5: ahci Feb 13 19:18:24.749590 kernel: ata1: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040100 irq 34 Feb 13 19:18:24.749602 kernel: ata2: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040180 irq 34 Feb 13 19:18:24.749612 kernel: ata3: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040200 irq 34 Feb 13 19:18:24.749623 kernel: ata4: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040280 irq 34 Feb 13 19:18:24.749633 kernel: ata5: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040300 irq 34 Feb 13 19:18:24.749643 kernel: ata6: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040380 irq 34 Feb 13 19:18:24.729700 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Feb 13 19:18:24.730031 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 19:18:24.760647 kernel: BTRFS: device fsid 60f89c25-9096-4268-99ca-ef7992742f2b devid 1 transid 39 /dev/vda3 scanned by (udev-worker) (465) Feb 13 19:18:24.732541 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Feb 13 19:18:24.733848 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Feb 13 19:18:24.734449 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 19:18:24.739967 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 19:18:24.772007 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/vda6 scanned by (udev-worker) (479) Feb 13 19:18:24.750747 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 19:18:24.779508 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Feb 13 19:18:24.800101 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Feb 13 19:18:24.813027 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Feb 13 19:18:24.813304 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Feb 13 19:18:24.824798 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Feb 13 19:18:24.835972 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Feb 13 19:18:24.836229 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Feb 13 19:18:24.836282 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 19:18:24.840295 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 19:18:24.841512 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 19:18:24.854080 disk-uuid[559]: Primary Header is updated. Feb 13 19:18:24.854080 disk-uuid[559]: Secondary Entries is updated. Feb 13 19:18:24.854080 disk-uuid[559]: Secondary Header is updated. Feb 13 19:18:24.857939 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Feb 13 19:18:24.859537 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 19:18:24.866008 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Feb 13 19:18:24.892099 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 19:18:25.058921 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Feb 13 19:18:25.059004 kernel: ata1: SATA link down (SStatus 0 SControl 300) Feb 13 19:18:25.059019 kernel: ata4: SATA link down (SStatus 0 SControl 300) Feb 13 19:18:25.059029 kernel: ata5: SATA link down (SStatus 0 SControl 300) Feb 13 19:18:25.059881 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Feb 13 19:18:25.060880 kernel: ata3.00: applying bridge limits Feb 13 19:18:25.060895 kernel: ata6: SATA link down (SStatus 0 SControl 300) Feb 13 19:18:25.061886 kernel: ata2: SATA link down (SStatus 0 SControl 300) Feb 13 19:18:25.062878 kernel: ata3.00: configured for UDMA/100 Feb 13 19:18:25.064886 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Feb 13 19:18:25.111416 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Feb 13 19:18:25.123525 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Feb 13 19:18:25.123542 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Feb 13 19:18:25.866915 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Feb 13 19:18:25.867730 disk-uuid[561]: The operation has completed successfully. Feb 13 19:18:25.899706 systemd[1]: disk-uuid.service: Deactivated successfully. Feb 13 19:18:25.899845 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Feb 13 19:18:25.948106 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Feb 13 19:18:25.953699 sh[600]: Success Feb 13 19:18:25.965886 kernel: device-mapper: verity: sha256 using implementation "sha256-ni" Feb 13 19:18:26.001513 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Feb 13 19:18:26.021466 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Feb 13 19:18:26.026208 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Feb 13 19:18:26.037698 kernel: BTRFS info (device dm-0): first mount of filesystem 60f89c25-9096-4268-99ca-ef7992742f2b Feb 13 19:18:26.037729 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Feb 13 19:18:26.037741 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Feb 13 19:18:26.038716 kernel: BTRFS info (device dm-0): disabling log replay at mount time Feb 13 19:18:26.039445 kernel: BTRFS info (device dm-0): using free space tree Feb 13 19:18:26.044219 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Feb 13 19:18:26.045240 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Feb 13 19:18:26.060011 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Feb 13 19:18:26.060937 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Feb 13 19:18:26.072421 kernel: BTRFS info (device vda6): first mount of filesystem 9d862461-eab1-477f-8790-b61f63b2958e Feb 13 19:18:26.072452 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Feb 13 19:18:26.072469 kernel: BTRFS info (device vda6): using free space tree Feb 13 19:18:26.075894 kernel: BTRFS info (device vda6): auto enabling async discard Feb 13 19:18:26.084699 systemd[1]: mnt-oem.mount: Deactivated successfully. Feb 13 19:18:26.086411 kernel: BTRFS info (device vda6): last unmount of filesystem 9d862461-eab1-477f-8790-b61f63b2958e Feb 13 19:18:26.095819 systemd[1]: Finished ignition-setup.service - Ignition (setup). Feb 13 19:18:26.103020 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Feb 13 19:18:26.155353 ignition[697]: Ignition 2.20.0 Feb 13 19:18:26.155367 ignition[697]: Stage: fetch-offline Feb 13 19:18:26.155412 ignition[697]: no configs at "/usr/lib/ignition/base.d" Feb 13 19:18:26.155422 ignition[697]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Feb 13 19:18:26.155519 ignition[697]: parsed url from cmdline: "" Feb 13 19:18:26.155523 ignition[697]: no config URL provided Feb 13 19:18:26.155528 ignition[697]: reading system config file "/usr/lib/ignition/user.ign" Feb 13 19:18:26.155538 ignition[697]: no config at "/usr/lib/ignition/user.ign" Feb 13 19:18:26.155563 ignition[697]: op(1): [started] loading QEMU firmware config module Feb 13 19:18:26.155570 ignition[697]: op(1): executing: "modprobe" "qemu_fw_cfg" Feb 13 19:18:26.163543 ignition[697]: op(1): [finished] loading QEMU firmware config module Feb 13 19:18:26.163569 ignition[697]: QEMU firmware config was not found. Ignoring... Feb 13 19:18:26.186348 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Feb 13 19:18:26.198009 systemd[1]: Starting systemd-networkd.service - Network Configuration... Feb 13 19:18:26.214135 ignition[697]: parsing config with SHA512: 4d9c2b3ec34b01b771602240d7c8b9cce68329bbfdd6e859f7713955f7f768f7b955efa80faf659f8ab4e1d3d208edf14a8ab2b8e031913bfa7083a6b71615c0 Feb 13 19:18:26.219117 unknown[697]: fetched base config from "system" Feb 13 19:18:26.219128 unknown[697]: fetched user config from "qemu" Feb 13 19:18:26.219466 ignition[697]: fetch-offline: fetch-offline passed Feb 13 19:18:26.222048 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Feb 13 19:18:26.219531 ignition[697]: Ignition finished successfully Feb 13 19:18:26.225570 systemd-networkd[790]: lo: Link UP Feb 13 19:18:26.225573 systemd-networkd[790]: lo: Gained carrier Feb 13 19:18:26.227216 systemd-networkd[790]: Enumeration completed Feb 13 19:18:26.227292 systemd[1]: Started systemd-networkd.service - Network Configuration. Feb 13 19:18:26.227563 systemd-networkd[790]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 19:18:26.227567 systemd-networkd[790]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Feb 13 19:18:26.228613 systemd-networkd[790]: eth0: Link UP Feb 13 19:18:26.228617 systemd-networkd[790]: eth0: Gained carrier Feb 13 19:18:26.228623 systemd-networkd[790]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 19:18:26.229110 systemd[1]: Reached target network.target - Network. Feb 13 19:18:26.230648 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Feb 13 19:18:26.239997 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Feb 13 19:18:26.242063 systemd-networkd[790]: eth0: DHCPv4 address 10.0.0.46/16, gateway 10.0.0.1 acquired from 10.0.0.1 Feb 13 19:18:26.259904 ignition[794]: Ignition 2.20.0 Feb 13 19:18:26.259914 ignition[794]: Stage: kargs Feb 13 19:18:26.260072 ignition[794]: no configs at "/usr/lib/ignition/base.d" Feb 13 19:18:26.260084 ignition[794]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Feb 13 19:18:26.260884 ignition[794]: kargs: kargs passed Feb 13 19:18:26.260925 ignition[794]: Ignition finished successfully Feb 13 19:18:26.264705 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Feb 13 19:18:26.273055 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Feb 13 19:18:26.284803 ignition[803]: Ignition 2.20.0 Feb 13 19:18:26.284814 ignition[803]: Stage: disks Feb 13 19:18:26.284990 ignition[803]: no configs at "/usr/lib/ignition/base.d" Feb 13 19:18:26.285001 ignition[803]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Feb 13 19:18:26.285762 ignition[803]: disks: disks passed Feb 13 19:18:26.288107 systemd[1]: Finished ignition-disks.service - Ignition (disks). Feb 13 19:18:26.285807 ignition[803]: Ignition finished successfully Feb 13 19:18:26.289320 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Feb 13 19:18:26.290752 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Feb 13 19:18:26.292850 systemd[1]: Reached target local-fs.target - Local File Systems. Feb 13 19:18:26.293882 systemd[1]: Reached target sysinit.target - System Initialization. Feb 13 19:18:26.295531 systemd[1]: Reached target basic.target - Basic System. Feb 13 19:18:26.303022 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Feb 13 19:18:26.315356 systemd-fsck[813]: ROOT: clean, 14/553520 files, 52654/553472 blocks Feb 13 19:18:26.321265 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Feb 13 19:18:27.044963 systemd[1]: Mounting sysroot.mount - /sysroot... Feb 13 19:18:27.126885 kernel: EXT4-fs (vda9): mounted filesystem 157595f2-1515-4117-a2d1-73fe2ed647fc r/w with ordered data mode. Quota mode: none. Feb 13 19:18:27.127537 systemd[1]: Mounted sysroot.mount - /sysroot. Feb 13 19:18:27.128771 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Feb 13 19:18:27.149951 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Feb 13 19:18:27.151766 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Feb 13 19:18:27.153298 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Feb 13 19:18:27.158275 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/vda6 scanned by mount (821) Feb 13 19:18:27.158301 kernel: BTRFS info (device vda6): first mount of filesystem 9d862461-eab1-477f-8790-b61f63b2958e Feb 13 19:18:27.153354 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Feb 13 19:18:27.164384 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Feb 13 19:18:27.164399 kernel: BTRFS info (device vda6): using free space tree Feb 13 19:18:27.164410 kernel: BTRFS info (device vda6): auto enabling async discard Feb 13 19:18:27.153387 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Feb 13 19:18:27.160892 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Feb 13 19:18:27.166902 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Feb 13 19:18:27.169522 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Feb 13 19:18:27.204668 initrd-setup-root[845]: cut: /sysroot/etc/passwd: No such file or directory Feb 13 19:18:27.209713 initrd-setup-root[852]: cut: /sysroot/etc/group: No such file or directory Feb 13 19:18:27.214375 initrd-setup-root[859]: cut: /sysroot/etc/shadow: No such file or directory Feb 13 19:18:27.218178 initrd-setup-root[866]: cut: /sysroot/etc/gshadow: No such file or directory Feb 13 19:18:27.307652 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Feb 13 19:18:27.319940 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Feb 13 19:18:27.321063 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Feb 13 19:18:27.327878 kernel: BTRFS info (device vda6): last unmount of filesystem 9d862461-eab1-477f-8790-b61f63b2958e Feb 13 19:18:27.345200 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Feb 13 19:18:27.347161 ignition[934]: INFO : Ignition 2.20.0 Feb 13 19:18:27.347161 ignition[934]: INFO : Stage: mount Feb 13 19:18:27.347161 ignition[934]: INFO : no configs at "/usr/lib/ignition/base.d" Feb 13 19:18:27.347161 ignition[934]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Feb 13 19:18:27.347161 ignition[934]: INFO : mount: mount passed Feb 13 19:18:27.347161 ignition[934]: INFO : Ignition finished successfully Feb 13 19:18:27.352526 systemd[1]: Finished ignition-mount.service - Ignition (mount). Feb 13 19:18:27.363955 systemd[1]: Starting ignition-files.service - Ignition (files)... Feb 13 19:18:27.530989 systemd-networkd[790]: eth0: Gained IPv6LL Feb 13 19:18:28.037197 systemd[1]: sysroot-oem.mount: Deactivated successfully. Feb 13 19:18:28.052001 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Feb 13 19:18:28.058882 kernel: BTRFS: device label OEM devid 1 transid 16 /dev/vda6 scanned by mount (947) Feb 13 19:18:28.062345 kernel: BTRFS info (device vda6): first mount of filesystem 9d862461-eab1-477f-8790-b61f63b2958e Feb 13 19:18:28.062398 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Feb 13 19:18:28.062410 kernel: BTRFS info (device vda6): using free space tree Feb 13 19:18:28.065882 kernel: BTRFS info (device vda6): auto enabling async discard Feb 13 19:18:28.066788 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Feb 13 19:18:28.084631 ignition[964]: INFO : Ignition 2.20.0 Feb 13 19:18:28.084631 ignition[964]: INFO : Stage: files Feb 13 19:18:28.086493 ignition[964]: INFO : no configs at "/usr/lib/ignition/base.d" Feb 13 19:18:28.086493 ignition[964]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Feb 13 19:18:28.086493 ignition[964]: DEBUG : files: compiled without relabeling support, skipping Feb 13 19:18:28.086493 ignition[964]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Feb 13 19:18:28.086493 ignition[964]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Feb 13 19:18:28.093096 ignition[964]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Feb 13 19:18:28.093096 ignition[964]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Feb 13 19:18:28.093096 ignition[964]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Feb 13 19:18:28.093096 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Feb 13 19:18:28.093096 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-amd64.tar.gz: attempt #1 Feb 13 19:18:28.088796 unknown[964]: wrote ssh authorized keys file for user: core Feb 13 19:18:28.132672 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Feb 13 19:18:28.485739 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Feb 13 19:18:28.487744 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Feb 13 19:18:28.487744 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Feb 13 19:18:28.487744 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Feb 13 19:18:28.487744 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Feb 13 19:18:28.487744 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Feb 13 19:18:28.487744 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Feb 13 19:18:28.487744 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Feb 13 19:18:28.487744 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Feb 13 19:18:28.487744 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Feb 13 19:18:28.487744 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Feb 13 19:18:28.487744 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.0-x86-64.raw" Feb 13 19:18:28.487744 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.0-x86-64.raw" Feb 13 19:18:28.487744 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.0-x86-64.raw" Feb 13 19:18:28.487744 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.31.0-x86-64.raw: attempt #1 Feb 13 19:18:29.041077 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Feb 13 19:18:29.385218 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.0-x86-64.raw" Feb 13 19:18:29.385218 ignition[964]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Feb 13 19:18:29.389182 ignition[964]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Feb 13 19:18:29.389182 ignition[964]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Feb 13 19:18:29.389182 ignition[964]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Feb 13 19:18:29.389182 ignition[964]: INFO : files: op(d): [started] processing unit "coreos-metadata.service" Feb 13 19:18:29.389182 ignition[964]: INFO : files: op(d): op(e): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Feb 13 19:18:29.389182 ignition[964]: INFO : files: op(d): op(e): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Feb 13 19:18:29.389182 ignition[964]: INFO : files: op(d): [finished] processing unit "coreos-metadata.service" Feb 13 19:18:29.389182 ignition[964]: INFO : files: op(f): [started] setting preset to disabled for "coreos-metadata.service" Feb 13 19:18:29.408315 ignition[964]: INFO : files: op(f): op(10): [started] removing enablement symlink(s) for "coreos-metadata.service" Feb 13 19:18:29.412214 ignition[964]: INFO : files: op(f): op(10): [finished] removing enablement symlink(s) for "coreos-metadata.service" Feb 13 19:18:29.413762 ignition[964]: INFO : files: op(f): [finished] setting preset to disabled for "coreos-metadata.service" Feb 13 19:18:29.413762 ignition[964]: INFO : files: op(11): [started] setting preset to enabled for "prepare-helm.service" Feb 13 19:18:29.413762 ignition[964]: INFO : files: op(11): [finished] setting preset to enabled for "prepare-helm.service" Feb 13 19:18:29.413762 ignition[964]: INFO : files: createResultFile: createFiles: op(12): [started] writing file "/sysroot/etc/.ignition-result.json" Feb 13 19:18:29.413762 ignition[964]: INFO : files: createResultFile: createFiles: op(12): [finished] writing file "/sysroot/etc/.ignition-result.json" Feb 13 19:18:29.413762 ignition[964]: INFO : files: files passed Feb 13 19:18:29.413762 ignition[964]: INFO : Ignition finished successfully Feb 13 19:18:29.415028 systemd[1]: Finished ignition-files.service - Ignition (files). Feb 13 19:18:29.427964 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Feb 13 19:18:29.429685 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Feb 13 19:18:29.431434 systemd[1]: ignition-quench.service: Deactivated successfully. Feb 13 19:18:29.431537 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Feb 13 19:18:29.439064 initrd-setup-root-after-ignition[993]: grep: /sysroot/oem/oem-release: No such file or directory Feb 13 19:18:29.441417 initrd-setup-root-after-ignition[995]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Feb 13 19:18:29.443030 initrd-setup-root-after-ignition[995]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Feb 13 19:18:29.445697 initrd-setup-root-after-ignition[999]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Feb 13 19:18:29.444281 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Feb 13 19:18:29.445985 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Feb 13 19:18:29.460974 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Feb 13 19:18:29.483032 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Feb 13 19:18:29.483148 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Feb 13 19:18:29.485343 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Feb 13 19:18:29.487352 systemd[1]: Reached target initrd.target - Initrd Default Target. Feb 13 19:18:29.489313 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Feb 13 19:18:29.498968 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Feb 13 19:18:29.511731 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Feb 13 19:18:29.528080 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Feb 13 19:18:29.537548 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Feb 13 19:18:29.538786 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Feb 13 19:18:29.540930 systemd[1]: Stopped target timers.target - Timer Units. Feb 13 19:18:29.542874 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Feb 13 19:18:29.542985 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Feb 13 19:18:29.545212 systemd[1]: Stopped target initrd.target - Initrd Default Target. Feb 13 19:18:29.546698 systemd[1]: Stopped target basic.target - Basic System. Feb 13 19:18:29.548752 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Feb 13 19:18:29.550844 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Feb 13 19:18:29.552785 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Feb 13 19:18:29.555006 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Feb 13 19:18:29.557051 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Feb 13 19:18:29.559326 systemd[1]: Stopped target sysinit.target - System Initialization. Feb 13 19:18:29.561311 systemd[1]: Stopped target local-fs.target - Local File Systems. Feb 13 19:18:29.563468 systemd[1]: Stopped target swap.target - Swaps. Feb 13 19:18:29.565237 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Feb 13 19:18:29.565394 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Feb 13 19:18:29.567458 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Feb 13 19:18:29.569091 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Feb 13 19:18:29.571146 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Feb 13 19:18:29.571267 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Feb 13 19:18:29.573548 systemd[1]: dracut-initqueue.service: Deactivated successfully. Feb 13 19:18:29.573690 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Feb 13 19:18:29.575812 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Feb 13 19:18:29.575945 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Feb 13 19:18:29.577907 systemd[1]: Stopped target paths.target - Path Units. Feb 13 19:18:29.579604 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Feb 13 19:18:29.582911 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Feb 13 19:18:29.585106 systemd[1]: Stopped target slices.target - Slice Units. Feb 13 19:18:29.587085 systemd[1]: Stopped target sockets.target - Socket Units. Feb 13 19:18:29.588939 systemd[1]: iscsid.socket: Deactivated successfully. Feb 13 19:18:29.589030 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Feb 13 19:18:29.591113 systemd[1]: iscsiuio.socket: Deactivated successfully. Feb 13 19:18:29.591195 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Feb 13 19:18:29.593552 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Feb 13 19:18:29.593661 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Feb 13 19:18:29.595715 systemd[1]: ignition-files.service: Deactivated successfully. Feb 13 19:18:29.595818 systemd[1]: Stopped ignition-files.service - Ignition (files). Feb 13 19:18:29.609002 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Feb 13 19:18:29.610670 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Feb 13 19:18:29.611785 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Feb 13 19:18:29.611913 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Feb 13 19:18:29.614075 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Feb 13 19:18:29.614290 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Feb 13 19:18:29.620314 systemd[1]: initrd-cleanup.service: Deactivated successfully. Feb 13 19:18:29.621493 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Feb 13 19:18:29.624116 ignition[1020]: INFO : Ignition 2.20.0 Feb 13 19:18:29.625192 ignition[1020]: INFO : Stage: umount Feb 13 19:18:29.625192 ignition[1020]: INFO : no configs at "/usr/lib/ignition/base.d" Feb 13 19:18:29.625192 ignition[1020]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Feb 13 19:18:29.628161 ignition[1020]: INFO : umount: umount passed Feb 13 19:18:29.628161 ignition[1020]: INFO : Ignition finished successfully Feb 13 19:18:29.628678 systemd[1]: ignition-mount.service: Deactivated successfully. Feb 13 19:18:29.628791 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Feb 13 19:18:29.630127 systemd[1]: Stopped target network.target - Network. Feb 13 19:18:29.631686 systemd[1]: ignition-disks.service: Deactivated successfully. Feb 13 19:18:29.631739 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Feb 13 19:18:29.633720 systemd[1]: ignition-kargs.service: Deactivated successfully. Feb 13 19:18:29.633766 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Feb 13 19:18:29.635634 systemd[1]: ignition-setup.service: Deactivated successfully. Feb 13 19:18:29.635678 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Feb 13 19:18:29.637443 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Feb 13 19:18:29.637488 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Feb 13 19:18:29.639643 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Feb 13 19:18:29.641590 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Feb 13 19:18:29.644524 systemd[1]: sysroot-boot.mount: Deactivated successfully. Feb 13 19:18:29.649597 systemd[1]: systemd-resolved.service: Deactivated successfully. Feb 13 19:18:29.649726 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Feb 13 19:18:29.654733 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Feb 13 19:18:29.655049 systemd[1]: systemd-networkd.service: Deactivated successfully. Feb 13 19:18:29.655167 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Feb 13 19:18:29.658894 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Feb 13 19:18:29.660014 systemd[1]: systemd-networkd.socket: Deactivated successfully. Feb 13 19:18:29.660063 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Feb 13 19:18:29.676996 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Feb 13 19:18:29.677922 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Feb 13 19:18:29.678001 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Feb 13 19:18:29.680176 systemd[1]: systemd-sysctl.service: Deactivated successfully. Feb 13 19:18:29.680239 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Feb 13 19:18:29.682451 systemd[1]: systemd-modules-load.service: Deactivated successfully. Feb 13 19:18:29.682500 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Feb 13 19:18:29.684678 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Feb 13 19:18:29.684726 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Feb 13 19:18:29.686815 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Feb 13 19:18:29.695692 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Feb 13 19:18:29.696921 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Feb 13 19:18:29.712114 systemd[1]: systemd-udevd.service: Deactivated successfully. Feb 13 19:18:29.722568 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Feb 13 19:18:29.725390 systemd[1]: network-cleanup.service: Deactivated successfully. Feb 13 19:18:29.726364 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Feb 13 19:18:29.729507 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Feb 13 19:18:29.730637 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Feb 13 19:18:29.732786 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Feb 13 19:18:29.732830 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Feb 13 19:18:29.735795 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Feb 13 19:18:29.736704 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Feb 13 19:18:29.738780 systemd[1]: dracut-cmdline.service: Deactivated successfully. Feb 13 19:18:29.739705 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Feb 13 19:18:29.741716 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Feb 13 19:18:29.742704 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Feb 13 19:18:29.759992 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Feb 13 19:18:29.760427 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Feb 13 19:18:29.760488 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Feb 13 19:18:29.764467 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Feb 13 19:18:29.764527 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Feb 13 19:18:29.765328 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Feb 13 19:18:29.765382 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Feb 13 19:18:29.765642 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Feb 13 19:18:29.765690 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 19:18:29.773178 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Feb 13 19:18:29.773271 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Feb 13 19:18:29.784343 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Feb 13 19:18:29.784483 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Feb 13 19:18:29.854506 systemd[1]: sysroot-boot.service: Deactivated successfully. Feb 13 19:18:29.854637 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Feb 13 19:18:29.857026 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Feb 13 19:18:29.857845 systemd[1]: initrd-setup-root.service: Deactivated successfully. Feb 13 19:18:29.857943 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Feb 13 19:18:29.876014 systemd[1]: Starting initrd-switch-root.service - Switch Root... Feb 13 19:18:29.882505 systemd[1]: Switching root. Feb 13 19:18:29.915183 systemd-journald[194]: Journal stopped Feb 13 19:18:31.075885 systemd-journald[194]: Received SIGTERM from PID 1 (systemd). Feb 13 19:18:31.075946 kernel: SELinux: policy capability network_peer_controls=1 Feb 13 19:18:31.075961 kernel: SELinux: policy capability open_perms=1 Feb 13 19:18:31.075972 kernel: SELinux: policy capability extended_socket_class=1 Feb 13 19:18:31.075983 kernel: SELinux: policy capability always_check_network=0 Feb 13 19:18:31.075994 kernel: SELinux: policy capability cgroup_seclabel=1 Feb 13 19:18:31.076012 kernel: SELinux: policy capability nnp_nosuid_transition=1 Feb 13 19:18:31.076030 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Feb 13 19:18:31.076053 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Feb 13 19:18:31.076065 kernel: audit: type=1403 audit(1739474310.295:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Feb 13 19:18:31.076077 systemd[1]: Successfully loaded SELinux policy in 38.675ms. Feb 13 19:18:31.076092 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 13.175ms. Feb 13 19:18:31.076106 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Feb 13 19:18:31.076128 systemd[1]: Detected virtualization kvm. Feb 13 19:18:31.076144 systemd[1]: Detected architecture x86-64. Feb 13 19:18:31.076158 systemd[1]: Detected first boot. Feb 13 19:18:31.076172 systemd[1]: Initializing machine ID from VM UUID. Feb 13 19:18:31.076185 zram_generator::config[1066]: No configuration found. Feb 13 19:18:31.076205 kernel: Guest personality initialized and is inactive Feb 13 19:18:31.076217 kernel: VMCI host device registered (name=vmci, major=10, minor=125) Feb 13 19:18:31.076228 kernel: Initialized host personality Feb 13 19:18:31.076245 kernel: NET: Registered PF_VSOCK protocol family Feb 13 19:18:31.076257 systemd[1]: Populated /etc with preset unit settings. Feb 13 19:18:31.076270 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Feb 13 19:18:31.076284 systemd[1]: initrd-switch-root.service: Deactivated successfully. Feb 13 19:18:31.076295 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Feb 13 19:18:31.076308 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Feb 13 19:18:31.076320 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Feb 13 19:18:31.076332 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Feb 13 19:18:31.076344 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Feb 13 19:18:31.076355 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Feb 13 19:18:31.076367 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Feb 13 19:18:31.076388 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Feb 13 19:18:31.076402 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Feb 13 19:18:31.076414 systemd[1]: Created slice user.slice - User and Session Slice. Feb 13 19:18:31.076426 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Feb 13 19:18:31.076440 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Feb 13 19:18:31.076459 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Feb 13 19:18:31.076471 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Feb 13 19:18:31.076483 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Feb 13 19:18:31.076496 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Feb 13 19:18:31.076510 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Feb 13 19:18:31.076522 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Feb 13 19:18:31.076534 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Feb 13 19:18:31.076546 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Feb 13 19:18:31.076558 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Feb 13 19:18:31.076571 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Feb 13 19:18:31.076583 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Feb 13 19:18:31.076595 systemd[1]: Reached target remote-fs.target - Remote File Systems. Feb 13 19:18:31.076609 systemd[1]: Reached target slices.target - Slice Units. Feb 13 19:18:31.076621 systemd[1]: Reached target swap.target - Swaps. Feb 13 19:18:31.076634 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Feb 13 19:18:31.076645 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Feb 13 19:18:31.076658 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Feb 13 19:18:31.076670 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Feb 13 19:18:31.076682 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Feb 13 19:18:31.076694 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Feb 13 19:18:31.076707 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Feb 13 19:18:31.076719 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Feb 13 19:18:31.076733 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Feb 13 19:18:31.076745 systemd[1]: Mounting media.mount - External Media Directory... Feb 13 19:18:31.076758 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 13 19:18:31.076769 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Feb 13 19:18:31.076781 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Feb 13 19:18:31.076793 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Feb 13 19:18:31.076805 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Feb 13 19:18:31.076818 systemd[1]: Reached target machines.target - Containers. Feb 13 19:18:31.076878 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Feb 13 19:18:31.076891 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Feb 13 19:18:31.076903 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Feb 13 19:18:31.076916 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Feb 13 19:18:31.076928 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Feb 13 19:18:31.076940 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Feb 13 19:18:31.076952 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Feb 13 19:18:31.076964 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Feb 13 19:18:31.076979 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Feb 13 19:18:31.076992 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Feb 13 19:18:31.077004 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Feb 13 19:18:31.077018 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Feb 13 19:18:31.077029 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Feb 13 19:18:31.077042 systemd[1]: Stopped systemd-fsck-usr.service. Feb 13 19:18:31.077054 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Feb 13 19:18:31.077066 kernel: fuse: init (API version 7.39) Feb 13 19:18:31.077078 kernel: loop: module loaded Feb 13 19:18:31.077092 systemd[1]: Starting systemd-journald.service - Journal Service... Feb 13 19:18:31.077105 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Feb 13 19:18:31.077117 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Feb 13 19:18:31.077128 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Feb 13 19:18:31.077141 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Feb 13 19:18:31.077155 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Feb 13 19:18:31.077167 systemd[1]: verity-setup.service: Deactivated successfully. Feb 13 19:18:31.077179 systemd[1]: Stopped verity-setup.service. Feb 13 19:18:31.077200 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 13 19:18:31.077264 systemd-journald[1141]: Collecting audit messages is disabled. Feb 13 19:18:31.077287 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Feb 13 19:18:31.077299 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Feb 13 19:18:31.077314 kernel: ACPI: bus type drm_connector registered Feb 13 19:18:31.077328 systemd-journald[1141]: Journal started Feb 13 19:18:31.077350 systemd-journald[1141]: Runtime Journal (/run/log/journal/2193d0da5b2849829be986e4c92b6580) is 6M, max 48.2M, 42.2M free. Feb 13 19:18:30.846115 systemd[1]: Queued start job for default target multi-user.target. Feb 13 19:18:30.863741 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Feb 13 19:18:30.864263 systemd[1]: systemd-journald.service: Deactivated successfully. Feb 13 19:18:31.079884 systemd[1]: Started systemd-journald.service - Journal Service. Feb 13 19:18:31.081397 systemd[1]: Mounted media.mount - External Media Directory. Feb 13 19:18:31.082504 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Feb 13 19:18:31.083705 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Feb 13 19:18:31.084941 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Feb 13 19:18:31.086272 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Feb 13 19:18:31.087795 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Feb 13 19:18:31.089367 systemd[1]: modprobe@configfs.service: Deactivated successfully. Feb 13 19:18:31.089582 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Feb 13 19:18:31.091074 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Feb 13 19:18:31.091290 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Feb 13 19:18:31.092712 systemd[1]: modprobe@drm.service: Deactivated successfully. Feb 13 19:18:31.092939 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Feb 13 19:18:31.094336 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Feb 13 19:18:31.094550 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Feb 13 19:18:31.096162 systemd[1]: modprobe@fuse.service: Deactivated successfully. Feb 13 19:18:31.096418 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Feb 13 19:18:31.097808 systemd[1]: modprobe@loop.service: Deactivated successfully. Feb 13 19:18:31.098079 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Feb 13 19:18:31.099511 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Feb 13 19:18:31.101097 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Feb 13 19:18:31.102671 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Feb 13 19:18:31.104264 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Feb 13 19:18:31.120450 systemd[1]: Reached target network-pre.target - Preparation for Network. Feb 13 19:18:31.131970 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Feb 13 19:18:31.134457 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Feb 13 19:18:31.135740 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Feb 13 19:18:31.135781 systemd[1]: Reached target local-fs.target - Local File Systems. Feb 13 19:18:31.138137 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Feb 13 19:18:31.140720 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Feb 13 19:18:31.144314 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Feb 13 19:18:31.145626 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Feb 13 19:18:31.148355 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Feb 13 19:18:31.151697 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Feb 13 19:18:31.152256 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Feb 13 19:18:31.157137 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Feb 13 19:18:31.158693 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Feb 13 19:18:31.160449 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Feb 13 19:18:31.161468 systemd-journald[1141]: Time spent on flushing to /var/log/journal/2193d0da5b2849829be986e4c92b6580 is 15.513ms for 1057 entries. Feb 13 19:18:31.161468 systemd-journald[1141]: System Journal (/var/log/journal/2193d0da5b2849829be986e4c92b6580) is 8M, max 195.6M, 187.6M free. Feb 13 19:18:31.202012 systemd-journald[1141]: Received client request to flush runtime journal. Feb 13 19:18:31.202069 kernel: loop0: detected capacity change from 0 to 138176 Feb 13 19:18:31.166974 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Feb 13 19:18:31.170871 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Feb 13 19:18:31.174770 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Feb 13 19:18:31.177175 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Feb 13 19:18:31.179138 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Feb 13 19:18:31.183797 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Feb 13 19:18:31.192795 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Feb 13 19:18:31.196013 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Feb 13 19:18:31.204758 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Feb 13 19:18:31.211264 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Feb 13 19:18:31.223073 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Feb 13 19:18:31.224755 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Feb 13 19:18:31.233075 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Feb 13 19:18:31.234796 systemd-tmpfiles[1187]: ACLs are not supported, ignoring. Feb 13 19:18:31.234810 systemd-tmpfiles[1187]: ACLs are not supported, ignoring. Feb 13 19:18:31.240016 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Feb 13 19:18:31.242343 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Feb 13 19:18:31.256921 kernel: loop1: detected capacity change from 0 to 147912 Feb 13 19:18:31.258146 systemd[1]: Starting systemd-sysusers.service - Create System Users... Feb 13 19:18:31.259568 udevadm[1202]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Feb 13 19:18:31.283014 systemd[1]: Finished systemd-sysusers.service - Create System Users. Feb 13 19:18:31.290254 kernel: loop2: detected capacity change from 0 to 205544 Feb 13 19:18:31.295141 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Feb 13 19:18:31.315178 systemd-tmpfiles[1209]: ACLs are not supported, ignoring. Feb 13 19:18:31.315205 systemd-tmpfiles[1209]: ACLs are not supported, ignoring. Feb 13 19:18:31.320866 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Feb 13 19:18:31.324888 kernel: loop3: detected capacity change from 0 to 138176 Feb 13 19:18:31.337877 kernel: loop4: detected capacity change from 0 to 147912 Feb 13 19:18:31.347889 kernel: loop5: detected capacity change from 0 to 205544 Feb 13 19:18:31.355307 (sd-merge)[1213]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. Feb 13 19:18:31.355934 (sd-merge)[1213]: Merged extensions into '/usr'. Feb 13 19:18:31.360338 systemd[1]: Reload requested from client PID 1186 ('systemd-sysext') (unit systemd-sysext.service)... Feb 13 19:18:31.360354 systemd[1]: Reloading... Feb 13 19:18:31.424880 zram_generator::config[1240]: No configuration found. Feb 13 19:18:31.479483 ldconfig[1181]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Feb 13 19:18:31.548165 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 19:18:31.615827 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Feb 13 19:18:31.616164 systemd[1]: Reloading finished in 255 ms. Feb 13 19:18:31.634311 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Feb 13 19:18:31.635970 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Feb 13 19:18:31.657652 systemd[1]: Starting ensure-sysext.service... Feb 13 19:18:31.659742 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Feb 13 19:18:31.680149 systemd-tmpfiles[1279]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Feb 13 19:18:31.680437 systemd-tmpfiles[1279]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Feb 13 19:18:31.681363 systemd-tmpfiles[1279]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Feb 13 19:18:31.681639 systemd-tmpfiles[1279]: ACLs are not supported, ignoring. Feb 13 19:18:31.681718 systemd-tmpfiles[1279]: ACLs are not supported, ignoring. Feb 13 19:18:31.682613 systemd[1]: Reload requested from client PID 1278 ('systemctl') (unit ensure-sysext.service)... Feb 13 19:18:31.682629 systemd[1]: Reloading... Feb 13 19:18:31.685532 systemd-tmpfiles[1279]: Detected autofs mount point /boot during canonicalization of boot. Feb 13 19:18:31.685543 systemd-tmpfiles[1279]: Skipping /boot Feb 13 19:18:31.700154 systemd-tmpfiles[1279]: Detected autofs mount point /boot during canonicalization of boot. Feb 13 19:18:31.700166 systemd-tmpfiles[1279]: Skipping /boot Feb 13 19:18:31.735043 zram_generator::config[1308]: No configuration found. Feb 13 19:18:31.850636 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 19:18:31.916233 systemd[1]: Reloading finished in 233 ms. Feb 13 19:18:31.930687 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Feb 13 19:18:31.949453 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Feb 13 19:18:31.958960 systemd[1]: Starting audit-rules.service - Load Audit Rules... Feb 13 19:18:31.961557 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Feb 13 19:18:31.964050 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Feb 13 19:18:31.967813 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Feb 13 19:18:31.972404 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Feb 13 19:18:31.978405 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Feb 13 19:18:31.983157 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 13 19:18:31.983342 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Feb 13 19:18:31.989303 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Feb 13 19:18:31.992390 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Feb 13 19:18:32.002189 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Feb 13 19:18:32.003328 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Feb 13 19:18:32.003432 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Feb 13 19:18:32.005329 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Feb 13 19:18:32.006574 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 13 19:18:32.008500 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Feb 13 19:18:32.009583 systemd-udevd[1351]: Using default interface naming scheme 'v255'. Feb 13 19:18:32.010359 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Feb 13 19:18:32.010592 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Feb 13 19:18:32.016221 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Feb 13 19:18:32.016427 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Feb 13 19:18:32.018292 systemd[1]: modprobe@loop.service: Deactivated successfully. Feb 13 19:18:32.018762 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Feb 13 19:18:32.025544 augenrules[1376]: No rules Feb 13 19:18:32.026660 systemd[1]: audit-rules.service: Deactivated successfully. Feb 13 19:18:32.027557 systemd[1]: Finished audit-rules.service - Load Audit Rules. Feb 13 19:18:32.034876 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Feb 13 19:18:32.036527 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Feb 13 19:18:32.045811 systemd[1]: Finished ensure-sysext.service. Feb 13 19:18:32.052360 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 13 19:18:32.061054 systemd[1]: Starting audit-rules.service - Load Audit Rules... Feb 13 19:18:32.062140 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Feb 13 19:18:32.064933 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Feb 13 19:18:32.070497 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Feb 13 19:18:32.073621 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Feb 13 19:18:32.079792 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Feb 13 19:18:32.081111 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Feb 13 19:18:32.081170 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Feb 13 19:18:32.087731 systemd[1]: Starting systemd-networkd.service - Network Configuration... Feb 13 19:18:32.092205 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Feb 13 19:18:32.095296 systemd[1]: Starting systemd-update-done.service - Update is Completed... Feb 13 19:18:32.095743 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 13 19:18:32.096265 systemd[1]: Started systemd-userdbd.service - User Database Manager. Feb 13 19:18:32.099676 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Feb 13 19:18:32.101336 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Feb 13 19:18:32.101950 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Feb 13 19:18:32.110127 augenrules[1403]: /sbin/augenrules: No change Feb 13 19:18:32.111978 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Feb 13 19:18:32.122254 systemd[1]: modprobe@loop.service: Deactivated successfully. Feb 13 19:18:32.122488 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Feb 13 19:18:32.126100 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Feb 13 19:18:32.126331 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Feb 13 19:18:32.128007 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 39 scanned by (udev-worker) (1400) Feb 13 19:18:32.131425 systemd[1]: modprobe@drm.service: Deactivated successfully. Feb 13 19:18:32.131654 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Feb 13 19:18:32.133504 augenrules[1443]: No rules Feb 13 19:18:32.157353 systemd[1]: audit-rules.service: Deactivated successfully. Feb 13 19:18:32.157659 systemd[1]: Finished audit-rules.service - Load Audit Rules. Feb 13 19:18:32.159243 systemd[1]: Finished systemd-update-done.service - Update is Completed. Feb 13 19:18:32.173633 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Feb 13 19:18:32.179379 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Feb 13 19:18:32.179457 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Feb 13 19:18:32.194876 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input2 Feb 13 19:18:32.205571 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Feb 13 19:18:32.215130 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Feb 13 19:18:32.217682 kernel: ACPI: button: Power Button [PWRF] Feb 13 19:18:32.220882 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input3 Feb 13 19:18:32.225764 kernel: i801_smbus 0000:00:1f.3: Enabling SMBus device Feb 13 19:18:32.229974 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Feb 13 19:18:32.230142 kernel: i2c i2c-0: 1/1 memory slots populated (from DMI) Feb 13 19:18:32.241619 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Feb 13 19:18:32.227242 systemd-networkd[1422]: lo: Link UP Feb 13 19:18:32.227248 systemd-networkd[1422]: lo: Gained carrier Feb 13 19:18:32.231382 systemd-networkd[1422]: Enumeration completed Feb 13 19:18:32.231473 systemd[1]: Started systemd-networkd.service - Network Configuration. Feb 13 19:18:32.234420 systemd-networkd[1422]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 19:18:32.234425 systemd-networkd[1422]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Feb 13 19:18:32.236179 systemd-networkd[1422]: eth0: Link UP Feb 13 19:18:32.236183 systemd-networkd[1422]: eth0: Gained carrier Feb 13 19:18:32.236196 systemd-networkd[1422]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Feb 13 19:18:32.242687 systemd-resolved[1350]: Positive Trust Anchors: Feb 13 19:18:32.242697 systemd-resolved[1350]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Feb 13 19:18:32.242728 systemd-resolved[1350]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Feb 13 19:18:32.246949 systemd-networkd[1422]: eth0: DHCPv4 address 10.0.0.46/16, gateway 10.0.0.1 acquired from 10.0.0.1 Feb 13 19:18:32.249575 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Feb 13 19:18:32.250660 systemd-timesyncd[1424]: Network configuration changed, trying to establish connection. Feb 13 19:18:33.232398 systemd-timesyncd[1424]: Contacted time server 10.0.0.1:123 (10.0.0.1). Feb 13 19:18:33.232453 systemd-timesyncd[1424]: Initial clock synchronization to Thu 2025-02-13 19:18:33.232310 UTC. Feb 13 19:18:33.235552 systemd-resolved[1350]: Defaulting to hostname 'linux'. Feb 13 19:18:33.238394 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Feb 13 19:18:33.239807 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Feb 13 19:18:33.242660 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Feb 13 19:18:33.244045 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Feb 13 19:18:33.248261 systemd[1]: Reached target network.target - Network. Feb 13 19:18:33.249216 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Feb 13 19:18:33.250388 systemd[1]: Reached target time-set.target - System Time Set. Feb 13 19:18:33.266647 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Feb 13 19:18:33.298311 kernel: mousedev: PS/2 mouse device common for all mice Feb 13 19:18:33.311728 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 19:18:33.322162 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Feb 13 19:18:33.322577 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 19:18:33.325258 kernel: kvm_amd: TSC scaling supported Feb 13 19:18:33.325289 kernel: kvm_amd: Nested Virtualization enabled Feb 13 19:18:33.325302 kernel: kvm_amd: Nested Paging enabled Feb 13 19:18:33.325314 kernel: kvm_amd: LBR virtualization supported Feb 13 19:18:33.326320 kernel: kvm_amd: Virtual VMLOAD VMSAVE supported Feb 13 19:18:33.326359 kernel: kvm_amd: Virtual GIF supported Feb 13 19:18:33.327446 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Feb 13 19:18:33.344819 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Feb 13 19:18:33.349211 kernel: EDAC MC: Ver: 3.0.0 Feb 13 19:18:33.384862 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Feb 13 19:18:33.392463 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Feb 13 19:18:33.394103 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Feb 13 19:18:33.401728 lvm[1479]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Feb 13 19:18:33.434470 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Feb 13 19:18:33.436032 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Feb 13 19:18:33.437155 systemd[1]: Reached target sysinit.target - System Initialization. Feb 13 19:18:33.438344 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Feb 13 19:18:33.439604 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Feb 13 19:18:33.441034 systemd[1]: Started logrotate.timer - Daily rotation of log files. Feb 13 19:18:33.442232 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Feb 13 19:18:33.443474 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Feb 13 19:18:33.444718 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Feb 13 19:18:33.444747 systemd[1]: Reached target paths.target - Path Units. Feb 13 19:18:33.445652 systemd[1]: Reached target timers.target - Timer Units. Feb 13 19:18:33.447556 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Feb 13 19:18:33.450333 systemd[1]: Starting docker.socket - Docker Socket for the API... Feb 13 19:18:33.453897 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Feb 13 19:18:33.455323 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Feb 13 19:18:33.456587 systemd[1]: Reached target ssh-access.target - SSH Access Available. Feb 13 19:18:33.464776 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Feb 13 19:18:33.466318 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Feb 13 19:18:33.468793 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Feb 13 19:18:33.470439 systemd[1]: Listening on docker.socket - Docker Socket for the API. Feb 13 19:18:33.471589 systemd[1]: Reached target sockets.target - Socket Units. Feb 13 19:18:33.472557 systemd[1]: Reached target basic.target - Basic System. Feb 13 19:18:33.472844 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Feb 13 19:18:33.472876 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Feb 13 19:18:33.473800 systemd[1]: Starting containerd.service - containerd container runtime... Feb 13 19:18:33.475813 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Feb 13 19:18:33.479205 lvm[1484]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Feb 13 19:18:33.479273 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Feb 13 19:18:33.482339 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Feb 13 19:18:33.483435 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Feb 13 19:18:33.487345 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Feb 13 19:18:33.488252 jq[1487]: false Feb 13 19:18:33.490310 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Feb 13 19:18:33.495004 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Feb 13 19:18:33.498015 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Feb 13 19:18:33.504356 systemd[1]: Starting systemd-logind.service - User Login Management... Feb 13 19:18:33.506821 extend-filesystems[1488]: Found loop3 Feb 13 19:18:33.506821 extend-filesystems[1488]: Found loop4 Feb 13 19:18:33.506821 extend-filesystems[1488]: Found loop5 Feb 13 19:18:33.506821 extend-filesystems[1488]: Found sr0 Feb 13 19:18:33.506821 extend-filesystems[1488]: Found vda Feb 13 19:18:33.506821 extend-filesystems[1488]: Found vda1 Feb 13 19:18:33.506821 extend-filesystems[1488]: Found vda2 Feb 13 19:18:33.506821 extend-filesystems[1488]: Found vda3 Feb 13 19:18:33.506821 extend-filesystems[1488]: Found usr Feb 13 19:18:33.506821 extend-filesystems[1488]: Found vda4 Feb 13 19:18:33.506821 extend-filesystems[1488]: Found vda6 Feb 13 19:18:33.506821 extend-filesystems[1488]: Found vda7 Feb 13 19:18:33.506821 extend-filesystems[1488]: Found vda9 Feb 13 19:18:33.506821 extend-filesystems[1488]: Checking size of /dev/vda9 Feb 13 19:18:33.507778 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Feb 13 19:18:33.515473 dbus-daemon[1486]: [system] SELinux support is enabled Feb 13 19:18:33.508299 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Feb 13 19:18:33.517415 systemd[1]: Starting update-engine.service - Update Engine... Feb 13 19:18:33.521294 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Feb 13 19:18:33.524786 systemd[1]: Started dbus.service - D-Bus System Message Bus. Feb 13 19:18:33.527693 extend-filesystems[1488]: Resized partition /dev/vda9 Feb 13 19:18:33.528368 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Feb 13 19:18:33.533230 extend-filesystems[1509]: resize2fs 1.47.1 (20-May-2024) Feb 13 19:18:33.545099 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Feb 13 19:18:33.536829 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Feb 13 19:18:33.545392 update_engine[1501]: I20250213 19:18:33.536366 1501 main.cc:92] Flatcar Update Engine starting Feb 13 19:18:33.545392 update_engine[1501]: I20250213 19:18:33.537578 1501 update_check_scheduler.cc:74] Next update check in 4m54s Feb 13 19:18:33.546112 jq[1506]: true Feb 13 19:18:33.537273 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Feb 13 19:18:33.537614 systemd[1]: motdgen.service: Deactivated successfully. Feb 13 19:18:33.538247 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Feb 13 19:18:33.540174 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Feb 13 19:18:33.540446 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Feb 13 19:18:33.550847 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 39 scanned by (udev-worker) (1394) Feb 13 19:18:33.559586 (ntainerd)[1515]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Feb 13 19:18:33.567825 jq[1514]: true Feb 13 19:18:33.580744 systemd[1]: Started update-engine.service - Update Engine. Feb 13 19:18:33.582499 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Feb 13 19:18:33.582523 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Feb 13 19:18:33.584795 tar[1511]: linux-amd64/helm Feb 13 19:18:33.584317 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Feb 13 19:18:33.629634 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Feb 13 19:18:33.637030 extend-filesystems[1509]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Feb 13 19:18:33.637030 extend-filesystems[1509]: old_desc_blocks = 1, new_desc_blocks = 1 Feb 13 19:18:33.637030 extend-filesystems[1509]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Feb 13 19:18:33.584333 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Feb 13 19:18:33.642591 extend-filesystems[1488]: Resized filesystem in /dev/vda9 Feb 13 19:18:33.599359 systemd[1]: Started locksmithd.service - Cluster reboot manager. Feb 13 19:18:33.626990 systemd-logind[1495]: Watching system buttons on /dev/input/event1 (Power Button) Feb 13 19:18:33.627016 systemd-logind[1495]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Feb 13 19:18:33.628928 systemd-logind[1495]: New seat seat0. Feb 13 19:18:33.630672 systemd[1]: extend-filesystems.service: Deactivated successfully. Feb 13 19:18:33.630930 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Feb 13 19:18:33.633322 systemd[1]: Started systemd-logind.service - User Login Management. Feb 13 19:18:33.647218 bash[1540]: Updated "/home/core/.ssh/authorized_keys" Feb 13 19:18:33.648069 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Feb 13 19:18:33.651555 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Feb 13 19:18:33.652358 sshd_keygen[1505]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Feb 13 19:18:33.653857 locksmithd[1526]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Feb 13 19:18:33.679566 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Feb 13 19:18:33.693086 systemd[1]: Starting issuegen.service - Generate /run/issue... Feb 13 19:18:33.700373 systemd[1]: issuegen.service: Deactivated successfully. Feb 13 19:18:33.700644 systemd[1]: Finished issuegen.service - Generate /run/issue. Feb 13 19:18:33.703410 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Feb 13 19:18:33.718246 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Feb 13 19:18:33.725610 systemd[1]: Started getty@tty1.service - Getty on tty1. Feb 13 19:18:33.728606 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Feb 13 19:18:33.729964 systemd[1]: Reached target getty.target - Login Prompts. Feb 13 19:18:33.776131 containerd[1515]: time="2025-02-13T19:18:33.776037467Z" level=info msg="starting containerd" revision=9b2ad7760328148397346d10c7b2004271249db4 version=v1.7.23 Feb 13 19:18:33.802767 containerd[1515]: time="2025-02-13T19:18:33.802699067Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Feb 13 19:18:33.804557 containerd[1515]: time="2025-02-13T19:18:33.804524992Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.74-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Feb 13 19:18:33.804557 containerd[1515]: time="2025-02-13T19:18:33.804552443Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Feb 13 19:18:33.804609 containerd[1515]: time="2025-02-13T19:18:33.804569956Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Feb 13 19:18:33.804788 containerd[1515]: time="2025-02-13T19:18:33.804767576Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Feb 13 19:18:33.804817 containerd[1515]: time="2025-02-13T19:18:33.804789327Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Feb 13 19:18:33.804884 containerd[1515]: time="2025-02-13T19:18:33.804864077Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Feb 13 19:18:33.804884 containerd[1515]: time="2025-02-13T19:18:33.804879977Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Feb 13 19:18:33.805181 containerd[1515]: time="2025-02-13T19:18:33.805151406Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Feb 13 19:18:33.805181 containerd[1515]: time="2025-02-13T19:18:33.805171203Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Feb 13 19:18:33.805233 containerd[1515]: time="2025-02-13T19:18:33.805205518Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Feb 13 19:18:33.805233 containerd[1515]: time="2025-02-13T19:18:33.805217340Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Feb 13 19:18:33.805348 containerd[1515]: time="2025-02-13T19:18:33.805324491Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Feb 13 19:18:33.805617 containerd[1515]: time="2025-02-13T19:18:33.805586392Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Feb 13 19:18:33.805794 containerd[1515]: time="2025-02-13T19:18:33.805766430Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Feb 13 19:18:33.805794 containerd[1515]: time="2025-02-13T19:18:33.805784303Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Feb 13 19:18:33.805901 containerd[1515]: time="2025-02-13T19:18:33.805880874Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Feb 13 19:18:33.805963 containerd[1515]: time="2025-02-13T19:18:33.805945295Z" level=info msg="metadata content store policy set" policy=shared Feb 13 19:18:33.811954 containerd[1515]: time="2025-02-13T19:18:33.811925560Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Feb 13 19:18:33.811990 containerd[1515]: time="2025-02-13T19:18:33.811972037Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Feb 13 19:18:33.812011 containerd[1515]: time="2025-02-13T19:18:33.811995251Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Feb 13 19:18:33.812031 containerd[1515]: time="2025-02-13T19:18:33.812013084Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Feb 13 19:18:33.812031 containerd[1515]: time="2025-02-13T19:18:33.812028614Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Feb 13 19:18:33.812202 containerd[1515]: time="2025-02-13T19:18:33.812173505Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Feb 13 19:18:33.812455 containerd[1515]: time="2025-02-13T19:18:33.812417433Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Feb 13 19:18:33.812556 containerd[1515]: time="2025-02-13T19:18:33.812540073Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Feb 13 19:18:33.812579 containerd[1515]: time="2025-02-13T19:18:33.812556864Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Feb 13 19:18:33.812579 containerd[1515]: time="2025-02-13T19:18:33.812571201Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Feb 13 19:18:33.812617 containerd[1515]: time="2025-02-13T19:18:33.812583935Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Feb 13 19:18:33.812617 containerd[1515]: time="2025-02-13T19:18:33.812596719Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Feb 13 19:18:33.812617 containerd[1515]: time="2025-02-13T19:18:33.812609012Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Feb 13 19:18:33.812689 containerd[1515]: time="2025-02-13T19:18:33.812623038Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Feb 13 19:18:33.812689 containerd[1515]: time="2025-02-13T19:18:33.812637375Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Feb 13 19:18:33.812689 containerd[1515]: time="2025-02-13T19:18:33.812649919Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Feb 13 19:18:33.812689 containerd[1515]: time="2025-02-13T19:18:33.812662913Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Feb 13 19:18:33.812689 containerd[1515]: time="2025-02-13T19:18:33.812674134Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Feb 13 19:18:33.812786 containerd[1515]: time="2025-02-13T19:18:33.812695063Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Feb 13 19:18:33.812786 containerd[1515]: time="2025-02-13T19:18:33.812708999Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Feb 13 19:18:33.812786 containerd[1515]: time="2025-02-13T19:18:33.812723206Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Feb 13 19:18:33.812786 containerd[1515]: time="2025-02-13T19:18:33.812740749Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Feb 13 19:18:33.812786 containerd[1515]: time="2025-02-13T19:18:33.812754324Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Feb 13 19:18:33.812786 containerd[1515]: time="2025-02-13T19:18:33.812779492Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Feb 13 19:18:33.812903 containerd[1515]: time="2025-02-13T19:18:33.812792296Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Feb 13 19:18:33.812903 containerd[1515]: time="2025-02-13T19:18:33.812806172Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Feb 13 19:18:33.812903 containerd[1515]: time="2025-02-13T19:18:33.812820218Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Feb 13 19:18:33.812903 containerd[1515]: time="2025-02-13T19:18:33.812834605Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Feb 13 19:18:33.812903 containerd[1515]: time="2025-02-13T19:18:33.812846507Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Feb 13 19:18:33.812903 containerd[1515]: time="2025-02-13T19:18:33.812858430Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Feb 13 19:18:33.812903 containerd[1515]: time="2025-02-13T19:18:33.812870062Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Feb 13 19:18:33.812903 containerd[1515]: time="2025-02-13T19:18:33.812883837Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Feb 13 19:18:33.812903 containerd[1515]: time="2025-02-13T19:18:33.812902613Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Feb 13 19:18:33.813064 containerd[1515]: time="2025-02-13T19:18:33.812916348Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Feb 13 19:18:33.813064 containerd[1515]: time="2025-02-13T19:18:33.812929162Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Feb 13 19:18:33.813064 containerd[1515]: time="2025-02-13T19:18:33.812971963Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Feb 13 19:18:33.813064 containerd[1515]: time="2025-02-13T19:18:33.812988544Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Feb 13 19:18:33.813064 containerd[1515]: time="2025-02-13T19:18:33.812999414Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Feb 13 19:18:33.813064 containerd[1515]: time="2025-02-13T19:18:33.813013400Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Feb 13 19:18:33.813064 containerd[1515]: time="2025-02-13T19:18:33.813024191Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Feb 13 19:18:33.813064 containerd[1515]: time="2025-02-13T19:18:33.813037145Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Feb 13 19:18:33.813064 containerd[1515]: time="2025-02-13T19:18:33.813048937Z" level=info msg="NRI interface is disabled by configuration." Feb 13 19:18:33.813064 containerd[1515]: time="2025-02-13T19:18:33.813064687Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Feb 13 19:18:33.813406 containerd[1515]: time="2025-02-13T19:18:33.813353638Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Feb 13 19:18:33.813406 containerd[1515]: time="2025-02-13T19:18:33.813406898Z" level=info msg="Connect containerd service" Feb 13 19:18:33.813563 containerd[1515]: time="2025-02-13T19:18:33.813442846Z" level=info msg="using legacy CRI server" Feb 13 19:18:33.813563 containerd[1515]: time="2025-02-13T19:18:33.813450199Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Feb 13 19:18:33.813601 containerd[1515]: time="2025-02-13T19:18:33.813564654Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Feb 13 19:18:33.814165 containerd[1515]: time="2025-02-13T19:18:33.814133962Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Feb 13 19:18:33.814388 containerd[1515]: time="2025-02-13T19:18:33.814350829Z" level=info msg="Start subscribing containerd event" Feb 13 19:18:33.814427 containerd[1515]: time="2025-02-13T19:18:33.814393899Z" level=info msg="Start recovering state" Feb 13 19:18:33.814464 containerd[1515]: time="2025-02-13T19:18:33.814449443Z" level=info msg="Start event monitor" Feb 13 19:18:33.814495 containerd[1515]: time="2025-02-13T19:18:33.814482105Z" level=info msg="Start snapshots syncer" Feb 13 19:18:33.814495 containerd[1515]: time="2025-02-13T19:18:33.814492444Z" level=info msg="Start cni network conf syncer for default" Feb 13 19:18:33.814532 containerd[1515]: time="2025-02-13T19:18:33.814502944Z" level=info msg="Start streaming server" Feb 13 19:18:33.814860 containerd[1515]: time="2025-02-13T19:18:33.814830027Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Feb 13 19:18:33.814932 containerd[1515]: time="2025-02-13T19:18:33.814911951Z" level=info msg=serving... address=/run/containerd/containerd.sock Feb 13 19:18:33.815090 systemd[1]: Started containerd.service - containerd container runtime. Feb 13 19:18:33.815873 containerd[1515]: time="2025-02-13T19:18:33.815606443Z" level=info msg="containerd successfully booted in 0.040726s" Feb 13 19:18:33.981838 tar[1511]: linux-amd64/LICENSE Feb 13 19:18:33.981838 tar[1511]: linux-amd64/README.md Feb 13 19:18:33.996337 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Feb 13 19:18:34.249308 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Feb 13 19:18:34.263387 systemd[1]: Started sshd@0-10.0.0.46:22-10.0.0.1:54668.service - OpenSSH per-connection server daemon (10.0.0.1:54668). Feb 13 19:18:34.313335 sshd[1577]: Accepted publickey for core from 10.0.0.1 port 54668 ssh2: RSA SHA256:xgLbxCKtIvCmXzj7C6d4ih050Hrbkh61XCRduaX62E8 Feb 13 19:18:34.315379 sshd-session[1577]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:18:34.325719 systemd-logind[1495]: New session 1 of user core. Feb 13 19:18:34.326932 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Feb 13 19:18:34.344383 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Feb 13 19:18:34.356535 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Feb 13 19:18:34.367407 systemd[1]: Starting user@500.service - User Manager for UID 500... Feb 13 19:18:34.371507 (systemd)[1581]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Feb 13 19:18:34.374034 systemd-logind[1495]: New session c1 of user core. Feb 13 19:18:34.519076 systemd[1581]: Queued start job for default target default.target. Feb 13 19:18:34.525305 systemd-networkd[1422]: eth0: Gained IPv6LL Feb 13 19:18:34.528214 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Feb 13 19:18:34.528792 systemd[1581]: Created slice app.slice - User Application Slice. Feb 13 19:18:34.528827 systemd[1581]: Reached target paths.target - Paths. Feb 13 19:18:34.528866 systemd[1581]: Reached target timers.target - Timers. Feb 13 19:18:34.529973 systemd[1]: Reached target network-online.target - Network is Online. Feb 13 19:18:34.530484 systemd[1581]: Starting dbus.socket - D-Bus User Message Bus Socket... Feb 13 19:18:34.532616 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Feb 13 19:18:34.535339 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:18:34.538539 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Feb 13 19:18:34.543099 systemd[1581]: Listening on dbus.socket - D-Bus User Message Bus Socket. Feb 13 19:18:34.543250 systemd[1581]: Reached target sockets.target - Sockets. Feb 13 19:18:34.543289 systemd[1581]: Reached target basic.target - Basic System. Feb 13 19:18:34.543332 systemd[1581]: Reached target default.target - Main User Target. Feb 13 19:18:34.543366 systemd[1581]: Startup finished in 161ms. Feb 13 19:18:34.544287 systemd[1]: Started user@500.service - User Manager for UID 500. Feb 13 19:18:34.551216 systemd[1]: Started session-1.scope - Session 1 of User core. Feb 13 19:18:34.562810 systemd[1]: coreos-metadata.service: Deactivated successfully. Feb 13 19:18:34.563111 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Feb 13 19:18:34.564662 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Feb 13 19:18:34.565136 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Feb 13 19:18:34.613231 systemd[1]: Started sshd@1-10.0.0.46:22-10.0.0.1:54676.service - OpenSSH per-connection server daemon (10.0.0.1:54676). Feb 13 19:18:34.655437 sshd[1609]: Accepted publickey for core from 10.0.0.1 port 54676 ssh2: RSA SHA256:xgLbxCKtIvCmXzj7C6d4ih050Hrbkh61XCRduaX62E8 Feb 13 19:18:34.657214 sshd-session[1609]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:18:34.661976 systemd-logind[1495]: New session 2 of user core. Feb 13 19:18:34.668320 systemd[1]: Started session-2.scope - Session 2 of User core. Feb 13 19:18:34.722936 sshd[1611]: Connection closed by 10.0.0.1 port 54676 Feb 13 19:18:34.723301 sshd-session[1609]: pam_unix(sshd:session): session closed for user core Feb 13 19:18:34.734010 systemd[1]: sshd@1-10.0.0.46:22-10.0.0.1:54676.service: Deactivated successfully. Feb 13 19:18:34.736004 systemd[1]: session-2.scope: Deactivated successfully. Feb 13 19:18:34.737601 systemd-logind[1495]: Session 2 logged out. Waiting for processes to exit. Feb 13 19:18:34.744652 systemd[1]: Started sshd@2-10.0.0.46:22-10.0.0.1:54678.service - OpenSSH per-connection server daemon (10.0.0.1:54678). Feb 13 19:18:34.747270 systemd-logind[1495]: Removed session 2. Feb 13 19:18:34.781508 sshd[1616]: Accepted publickey for core from 10.0.0.1 port 54678 ssh2: RSA SHA256:xgLbxCKtIvCmXzj7C6d4ih050Hrbkh61XCRduaX62E8 Feb 13 19:18:34.783065 sshd-session[1616]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:18:34.787494 systemd-logind[1495]: New session 3 of user core. Feb 13 19:18:34.803335 systemd[1]: Started session-3.scope - Session 3 of User core. Feb 13 19:18:34.857662 sshd[1619]: Connection closed by 10.0.0.1 port 54678 Feb 13 19:18:34.858168 sshd-session[1616]: pam_unix(sshd:session): session closed for user core Feb 13 19:18:34.862465 systemd[1]: sshd@2-10.0.0.46:22-10.0.0.1:54678.service: Deactivated successfully. Feb 13 19:18:34.864387 systemd[1]: session-3.scope: Deactivated successfully. Feb 13 19:18:34.865014 systemd-logind[1495]: Session 3 logged out. Waiting for processes to exit. Feb 13 19:18:34.865800 systemd-logind[1495]: Removed session 3. Feb 13 19:18:35.157707 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:18:35.159373 systemd[1]: Reached target multi-user.target - Multi-User System. Feb 13 19:18:35.163042 (kubelet)[1629]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 19:18:35.163282 systemd[1]: Startup finished in 737ms (kernel) + 6.591s (initrd) + 3.926s (userspace) = 11.256s. Feb 13 19:18:35.548667 kubelet[1629]: E0213 19:18:35.548477 1629 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 19:18:35.552747 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 19:18:35.552956 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 19:18:35.553362 systemd[1]: kubelet.service: Consumed 912ms CPU time, 237.6M memory peak. Feb 13 19:18:44.871292 systemd[1]: Started sshd@3-10.0.0.46:22-10.0.0.1:34346.service - OpenSSH per-connection server daemon (10.0.0.1:34346). Feb 13 19:18:44.913597 sshd[1643]: Accepted publickey for core from 10.0.0.1 port 34346 ssh2: RSA SHA256:xgLbxCKtIvCmXzj7C6d4ih050Hrbkh61XCRduaX62E8 Feb 13 19:18:44.915384 sshd-session[1643]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:18:44.919548 systemd-logind[1495]: New session 4 of user core. Feb 13 19:18:44.930442 systemd[1]: Started session-4.scope - Session 4 of User core. Feb 13 19:18:44.984610 sshd[1645]: Connection closed by 10.0.0.1 port 34346 Feb 13 19:18:44.984981 sshd-session[1643]: pam_unix(sshd:session): session closed for user core Feb 13 19:18:45.001244 systemd[1]: sshd@3-10.0.0.46:22-10.0.0.1:34346.service: Deactivated successfully. Feb 13 19:18:45.003140 systemd[1]: session-4.scope: Deactivated successfully. Feb 13 19:18:45.004741 systemd-logind[1495]: Session 4 logged out. Waiting for processes to exit. Feb 13 19:18:45.014605 systemd[1]: Started sshd@4-10.0.0.46:22-10.0.0.1:34350.service - OpenSSH per-connection server daemon (10.0.0.1:34350). Feb 13 19:18:45.015635 systemd-logind[1495]: Removed session 4. Feb 13 19:18:45.053036 sshd[1650]: Accepted publickey for core from 10.0.0.1 port 34350 ssh2: RSA SHA256:xgLbxCKtIvCmXzj7C6d4ih050Hrbkh61XCRduaX62E8 Feb 13 19:18:45.054839 sshd-session[1650]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:18:45.059005 systemd-logind[1495]: New session 5 of user core. Feb 13 19:18:45.068328 systemd[1]: Started session-5.scope - Session 5 of User core. Feb 13 19:18:45.116110 sshd[1653]: Connection closed by 10.0.0.1 port 34350 Feb 13 19:18:45.116422 sshd-session[1650]: pam_unix(sshd:session): session closed for user core Feb 13 19:18:45.131794 systemd[1]: sshd@4-10.0.0.46:22-10.0.0.1:34350.service: Deactivated successfully. Feb 13 19:18:45.134081 systemd[1]: session-5.scope: Deactivated successfully. Feb 13 19:18:45.135505 systemd-logind[1495]: Session 5 logged out. Waiting for processes to exit. Feb 13 19:18:45.142482 systemd[1]: Started sshd@5-10.0.0.46:22-10.0.0.1:34366.service - OpenSSH per-connection server daemon (10.0.0.1:34366). Feb 13 19:18:45.143293 systemd-logind[1495]: Removed session 5. Feb 13 19:18:45.180116 sshd[1658]: Accepted publickey for core from 10.0.0.1 port 34366 ssh2: RSA SHA256:xgLbxCKtIvCmXzj7C6d4ih050Hrbkh61XCRduaX62E8 Feb 13 19:18:45.181673 sshd-session[1658]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:18:45.185606 systemd-logind[1495]: New session 6 of user core. Feb 13 19:18:45.202316 systemd[1]: Started session-6.scope - Session 6 of User core. Feb 13 19:18:45.254031 sshd[1661]: Connection closed by 10.0.0.1 port 34366 Feb 13 19:18:45.254396 sshd-session[1658]: pam_unix(sshd:session): session closed for user core Feb 13 19:18:45.265624 systemd[1]: sshd@5-10.0.0.46:22-10.0.0.1:34366.service: Deactivated successfully. Feb 13 19:18:45.267254 systemd[1]: session-6.scope: Deactivated successfully. Feb 13 19:18:45.268555 systemd-logind[1495]: Session 6 logged out. Waiting for processes to exit. Feb 13 19:18:45.269748 systemd[1]: Started sshd@6-10.0.0.46:22-10.0.0.1:34382.service - OpenSSH per-connection server daemon (10.0.0.1:34382). Feb 13 19:18:45.270415 systemd-logind[1495]: Removed session 6. Feb 13 19:18:45.309654 sshd[1666]: Accepted publickey for core from 10.0.0.1 port 34382 ssh2: RSA SHA256:xgLbxCKtIvCmXzj7C6d4ih050Hrbkh61XCRduaX62E8 Feb 13 19:18:45.310984 sshd-session[1666]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:18:45.315170 systemd-logind[1495]: New session 7 of user core. Feb 13 19:18:45.326312 systemd[1]: Started session-7.scope - Session 7 of User core. Feb 13 19:18:45.382508 sudo[1670]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Feb 13 19:18:45.382847 sudo[1670]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Feb 13 19:18:45.395874 sudo[1670]: pam_unix(sudo:session): session closed for user root Feb 13 19:18:45.397090 sshd[1669]: Connection closed by 10.0.0.1 port 34382 Feb 13 19:18:45.397538 sshd-session[1666]: pam_unix(sshd:session): session closed for user core Feb 13 19:18:45.408950 systemd[1]: sshd@6-10.0.0.46:22-10.0.0.1:34382.service: Deactivated successfully. Feb 13 19:18:45.410802 systemd[1]: session-7.scope: Deactivated successfully. Feb 13 19:18:45.412515 systemd-logind[1495]: Session 7 logged out. Waiting for processes to exit. Feb 13 19:18:45.421406 systemd[1]: Started sshd@7-10.0.0.46:22-10.0.0.1:34392.service - OpenSSH per-connection server daemon (10.0.0.1:34392). Feb 13 19:18:45.422263 systemd-logind[1495]: Removed session 7. Feb 13 19:18:45.458659 sshd[1675]: Accepted publickey for core from 10.0.0.1 port 34392 ssh2: RSA SHA256:xgLbxCKtIvCmXzj7C6d4ih050Hrbkh61XCRduaX62E8 Feb 13 19:18:45.460024 sshd-session[1675]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:18:45.464091 systemd-logind[1495]: New session 8 of user core. Feb 13 19:18:45.473308 systemd[1]: Started session-8.scope - Session 8 of User core. Feb 13 19:18:45.525941 sudo[1680]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Feb 13 19:18:45.526290 sudo[1680]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Feb 13 19:18:45.529987 sudo[1680]: pam_unix(sudo:session): session closed for user root Feb 13 19:18:45.536338 sudo[1679]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Feb 13 19:18:45.536651 sudo[1679]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Feb 13 19:18:45.554448 systemd[1]: Starting audit-rules.service - Load Audit Rules... Feb 13 19:18:45.555397 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Feb 13 19:18:45.556847 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:18:45.584924 augenrules[1705]: No rules Feb 13 19:18:45.586838 systemd[1]: audit-rules.service: Deactivated successfully. Feb 13 19:18:45.587119 systemd[1]: Finished audit-rules.service - Load Audit Rules. Feb 13 19:18:45.588427 sudo[1679]: pam_unix(sudo:session): session closed for user root Feb 13 19:18:45.590093 sshd[1678]: Connection closed by 10.0.0.1 port 34392 Feb 13 19:18:45.591912 sshd-session[1675]: pam_unix(sshd:session): session closed for user core Feb 13 19:18:45.601456 systemd[1]: sshd@7-10.0.0.46:22-10.0.0.1:34392.service: Deactivated successfully. Feb 13 19:18:45.603517 systemd[1]: session-8.scope: Deactivated successfully. Feb 13 19:18:45.604246 systemd-logind[1495]: Session 8 logged out. Waiting for processes to exit. Feb 13 19:18:45.612584 systemd[1]: Started sshd@8-10.0.0.46:22-10.0.0.1:34398.service - OpenSSH per-connection server daemon (10.0.0.1:34398). Feb 13 19:18:45.613608 systemd-logind[1495]: Removed session 8. Feb 13 19:18:45.655955 sshd[1713]: Accepted publickey for core from 10.0.0.1 port 34398 ssh2: RSA SHA256:xgLbxCKtIvCmXzj7C6d4ih050Hrbkh61XCRduaX62E8 Feb 13 19:18:45.657518 sshd-session[1713]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:18:45.662424 systemd-logind[1495]: New session 9 of user core. Feb 13 19:18:45.672319 systemd[1]: Started session-9.scope - Session 9 of User core. Feb 13 19:18:45.714503 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:18:45.718555 (kubelet)[1722]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 19:18:45.725497 sudo[1723]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Feb 13 19:18:45.725818 sudo[1723]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Feb 13 19:18:45.753087 kubelet[1722]: E0213 19:18:45.753040 1722 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 19:18:45.759705 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 19:18:45.759914 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 19:18:45.760471 systemd[1]: kubelet.service: Consumed 188ms CPU time, 97.8M memory peak. Feb 13 19:18:46.006398 systemd[1]: Starting docker.service - Docker Application Container Engine... Feb 13 19:18:46.006586 (dockerd)[1750]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Feb 13 19:18:46.283343 dockerd[1750]: time="2025-02-13T19:18:46.283173612Z" level=info msg="Starting up" Feb 13 19:18:46.678436 dockerd[1750]: time="2025-02-13T19:18:46.678285057Z" level=info msg="Loading containers: start." Feb 13 19:18:46.913227 kernel: Initializing XFRM netlink socket Feb 13 19:18:47.001657 systemd-networkd[1422]: docker0: Link UP Feb 13 19:18:47.300167 dockerd[1750]: time="2025-02-13T19:18:47.300030502Z" level=info msg="Loading containers: done." Feb 13 19:18:47.313465 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck4045910941-merged.mount: Deactivated successfully. Feb 13 19:18:47.315537 dockerd[1750]: time="2025-02-13T19:18:47.315502237Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Feb 13 19:18:47.315601 dockerd[1750]: time="2025-02-13T19:18:47.315571206Z" level=info msg="Docker daemon" commit=41ca978a0a5400cc24b274137efa9f25517fcc0b containerd-snapshotter=false storage-driver=overlay2 version=27.3.1 Feb 13 19:18:47.315703 dockerd[1750]: time="2025-02-13T19:18:47.315678227Z" level=info msg="Daemon has completed initialization" Feb 13 19:18:47.351063 dockerd[1750]: time="2025-02-13T19:18:47.351008063Z" level=info msg="API listen on /run/docker.sock" Feb 13 19:18:47.351210 systemd[1]: Started docker.service - Docker Application Container Engine. Feb 13 19:18:47.984604 containerd[1515]: time="2025-02-13T19:18:47.984565549Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.6\"" Feb 13 19:18:50.205890 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3121897669.mount: Deactivated successfully. Feb 13 19:18:52.248829 containerd[1515]: time="2025-02-13T19:18:52.248776697Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.31.6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:18:52.249407 containerd[1515]: time="2025-02-13T19:18:52.249367585Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.31.6: active requests=0, bytes read=27976588" Feb 13 19:18:52.250537 containerd[1515]: time="2025-02-13T19:18:52.250512212Z" level=info msg="ImageCreate event name:\"sha256:1372127edc9da70a68712c470a11f621ed256e8be0dfec4c4d58ca09109352a3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:18:52.253200 containerd[1515]: time="2025-02-13T19:18:52.253151582Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:be0a2d815793b0408d921a50b82759e654cf1bba718cac480498391926902905\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:18:52.254064 containerd[1515]: time="2025-02-13T19:18:52.254016193Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.31.6\" with image id \"sha256:1372127edc9da70a68712c470a11f621ed256e8be0dfec4c4d58ca09109352a3\", repo tag \"registry.k8s.io/kube-apiserver:v1.31.6\", repo digest \"registry.k8s.io/kube-apiserver@sha256:be0a2d815793b0408d921a50b82759e654cf1bba718cac480498391926902905\", size \"27973388\" in 4.269402284s" Feb 13 19:18:52.254106 containerd[1515]: time="2025-02-13T19:18:52.254066237Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.6\" returns image reference \"sha256:1372127edc9da70a68712c470a11f621ed256e8be0dfec4c4d58ca09109352a3\"" Feb 13 19:18:52.255437 containerd[1515]: time="2025-02-13T19:18:52.255400559Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.6\"" Feb 13 19:18:53.400333 containerd[1515]: time="2025-02-13T19:18:53.400271541Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.31.6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:18:53.401072 containerd[1515]: time="2025-02-13T19:18:53.401004686Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.31.6: active requests=0, bytes read=24708193" Feb 13 19:18:53.402308 containerd[1515]: time="2025-02-13T19:18:53.402267354Z" level=info msg="ImageCreate event name:\"sha256:5f23cb154eea1f587685082e456e95e5480c1d459849b1c634119d7de897e34e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:18:53.405015 containerd[1515]: time="2025-02-13T19:18:53.404967668Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:63166e537a82855ac9b54ffa8b510429fe799ed9b062bf6b788b74e1d5995d12\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:18:53.407457 containerd[1515]: time="2025-02-13T19:18:53.407135033Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.31.6\" with image id \"sha256:5f23cb154eea1f587685082e456e95e5480c1d459849b1c634119d7de897e34e\", repo tag \"registry.k8s.io/kube-controller-manager:v1.31.6\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:63166e537a82855ac9b54ffa8b510429fe799ed9b062bf6b788b74e1d5995d12\", size \"26154739\" in 1.151689379s" Feb 13 19:18:53.407457 containerd[1515]: time="2025-02-13T19:18:53.407166161Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.6\" returns image reference \"sha256:5f23cb154eea1f587685082e456e95e5480c1d459849b1c634119d7de897e34e\"" Feb 13 19:18:53.408589 containerd[1515]: time="2025-02-13T19:18:53.408561458Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.6\"" Feb 13 19:18:54.620871 containerd[1515]: time="2025-02-13T19:18:54.620805929Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.31.6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:18:54.621598 containerd[1515]: time="2025-02-13T19:18:54.621522252Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.31.6: active requests=0, bytes read=18652425" Feb 13 19:18:54.622637 containerd[1515]: time="2025-02-13T19:18:54.622603229Z" level=info msg="ImageCreate event name:\"sha256:9195ad415d31e3c2df6dddf4603bc56915b71486f514455bc3b5389b9b0ed9c1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:18:54.625590 containerd[1515]: time="2025-02-13T19:18:54.625558652Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:8a64af33c57346355dc3cc6f9225dbe771da30e2f427e802ce2340ec3b5dd9b5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:18:54.626675 containerd[1515]: time="2025-02-13T19:18:54.626635601Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.31.6\" with image id \"sha256:9195ad415d31e3c2df6dddf4603bc56915b71486f514455bc3b5389b9b0ed9c1\", repo tag \"registry.k8s.io/kube-scheduler:v1.31.6\", repo digest \"registry.k8s.io/kube-scheduler@sha256:8a64af33c57346355dc3cc6f9225dbe771da30e2f427e802ce2340ec3b5dd9b5\", size \"20098989\" in 1.217935023s" Feb 13 19:18:54.626713 containerd[1515]: time="2025-02-13T19:18:54.626674274Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.6\" returns image reference \"sha256:9195ad415d31e3c2df6dddf4603bc56915b71486f514455bc3b5389b9b0ed9c1\"" Feb 13 19:18:54.627239 containerd[1515]: time="2025-02-13T19:18:54.627211772Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.6\"" Feb 13 19:18:55.555381 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3519853884.mount: Deactivated successfully. Feb 13 19:18:55.971639 containerd[1515]: time="2025-02-13T19:18:55.971488601Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.31.6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:18:55.972347 containerd[1515]: time="2025-02-13T19:18:55.972311935Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.31.6: active requests=0, bytes read=30229108" Feb 13 19:18:55.973484 containerd[1515]: time="2025-02-13T19:18:55.973456952Z" level=info msg="ImageCreate event name:\"sha256:d2448f015605e48efb6b06ceaba0cb6d48bfd82e5d30ba357a9bd78c8566348a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:18:55.975383 containerd[1515]: time="2025-02-13T19:18:55.975335696Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:e72a4bc769f10b56ffdfe2cdb21d84d49d9bc194b3658648207998a5bd924b72\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:18:55.975849 containerd[1515]: time="2025-02-13T19:18:55.975816046Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.31.6\" with image id \"sha256:d2448f015605e48efb6b06ceaba0cb6d48bfd82e5d30ba357a9bd78c8566348a\", repo tag \"registry.k8s.io/kube-proxy:v1.31.6\", repo digest \"registry.k8s.io/kube-proxy@sha256:e72a4bc769f10b56ffdfe2cdb21d84d49d9bc194b3658648207998a5bd924b72\", size \"30228127\" in 1.348568908s" Feb 13 19:18:55.975875 containerd[1515]: time="2025-02-13T19:18:55.975850621Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.6\" returns image reference \"sha256:d2448f015605e48efb6b06ceaba0cb6d48bfd82e5d30ba357a9bd78c8566348a\"" Feb 13 19:18:55.976344 containerd[1515]: time="2025-02-13T19:18:55.976322937Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\"" Feb 13 19:18:56.010324 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Feb 13 19:18:56.019412 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:18:56.171487 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:18:56.176330 (kubelet)[2027]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Feb 13 19:18:56.332209 kubelet[2027]: E0213 19:18:56.331992 2027 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Feb 13 19:18:56.336135 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 13 19:18:56.336399 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 13 19:18:56.336833 systemd[1]: kubelet.service: Consumed 191ms CPU time, 97.6M memory peak. Feb 13 19:18:56.869103 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2898625412.mount: Deactivated successfully. Feb 13 19:18:57.792817 containerd[1515]: time="2025-02-13T19:18:57.792751118Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:18:57.793799 containerd[1515]: time="2025-02-13T19:18:57.793715006Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.1: active requests=0, bytes read=18185761" Feb 13 19:18:57.795452 containerd[1515]: time="2025-02-13T19:18:57.795412369Z" level=info msg="ImageCreate event name:\"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:18:57.799234 containerd[1515]: time="2025-02-13T19:18:57.799206875Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:18:57.800249 containerd[1515]: time="2025-02-13T19:18:57.800212721Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.1\" with image id \"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\", size \"18182961\" in 1.823862012s" Feb 13 19:18:57.800249 containerd[1515]: time="2025-02-13T19:18:57.800243098Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\" returns image reference \"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\"" Feb 13 19:18:57.800867 containerd[1515]: time="2025-02-13T19:18:57.800703171Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Feb 13 19:18:58.310676 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2827219528.mount: Deactivated successfully. Feb 13 19:18:58.316689 containerd[1515]: time="2025-02-13T19:18:58.316633845Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:18:58.317338 containerd[1515]: time="2025-02-13T19:18:58.317278374Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Feb 13 19:18:58.318323 containerd[1515]: time="2025-02-13T19:18:58.318289901Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:18:58.320425 containerd[1515]: time="2025-02-13T19:18:58.320392694Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:18:58.320977 containerd[1515]: time="2025-02-13T19:18:58.320939049Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 520.208366ms" Feb 13 19:18:58.320977 containerd[1515]: time="2025-02-13T19:18:58.320970799Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Feb 13 19:18:58.321456 containerd[1515]: time="2025-02-13T19:18:58.321430391Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\"" Feb 13 19:18:59.090047 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1833442496.mount: Deactivated successfully. Feb 13 19:19:01.176241 containerd[1515]: time="2025-02-13T19:19:01.176170748Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.15-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:19:01.176990 containerd[1515]: time="2025-02-13T19:19:01.176946122Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.15-0: active requests=0, bytes read=56779973" Feb 13 19:19:01.178536 containerd[1515]: time="2025-02-13T19:19:01.178483034Z" level=info msg="ImageCreate event name:\"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:19:01.183429 containerd[1515]: time="2025-02-13T19:19:01.183375499Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:19:01.184743 containerd[1515]: time="2025-02-13T19:19:01.184682821Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.15-0\" with image id \"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\", repo tag \"registry.k8s.io/etcd:3.5.15-0\", repo digest \"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\", size \"56909194\" in 2.863228235s" Feb 13 19:19:01.184809 containerd[1515]: time="2025-02-13T19:19:01.184745629Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\" returns image reference \"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\"" Feb 13 19:19:03.436053 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:19:03.436375 systemd[1]: kubelet.service: Consumed 191ms CPU time, 97.6M memory peak. Feb 13 19:19:03.448402 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:19:03.476030 systemd[1]: Reload requested from client PID 2166 ('systemctl') (unit session-9.scope)... Feb 13 19:19:03.476049 systemd[1]: Reloading... Feb 13 19:19:03.556217 zram_generator::config[2213]: No configuration found. Feb 13 19:19:03.773312 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 19:19:03.874484 systemd[1]: Reloading finished in 398 ms. Feb 13 19:19:03.917817 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:19:03.921102 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:19:03.922300 systemd[1]: kubelet.service: Deactivated successfully. Feb 13 19:19:03.922544 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:19:03.922580 systemd[1]: kubelet.service: Consumed 129ms CPU time, 83.6M memory peak. Feb 13 19:19:03.924265 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:19:04.063457 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:19:04.068007 (kubelet)[2260]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Feb 13 19:19:04.098395 kubelet[2260]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 13 19:19:04.098395 kubelet[2260]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Feb 13 19:19:04.098395 kubelet[2260]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 13 19:19:04.098853 kubelet[2260]: I0213 19:19:04.098439 2260 server.go:206] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 13 19:19:04.270249 kubelet[2260]: I0213 19:19:04.270213 2260 server.go:486] "Kubelet version" kubeletVersion="v1.31.0" Feb 13 19:19:04.270249 kubelet[2260]: I0213 19:19:04.270237 2260 server.go:488] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 13 19:19:04.270444 kubelet[2260]: I0213 19:19:04.270428 2260 server.go:929] "Client rotation is on, will bootstrap in background" Feb 13 19:19:04.291055 kubelet[2260]: E0213 19:19:04.291015 2260 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.0.0.46:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.46:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:19:04.293504 kubelet[2260]: I0213 19:19:04.293473 2260 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Feb 13 19:19:04.298595 kubelet[2260]: E0213 19:19:04.298567 2260 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Feb 13 19:19:04.298595 kubelet[2260]: I0213 19:19:04.298591 2260 server.go:1403] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Feb 13 19:19:04.304253 kubelet[2260]: I0213 19:19:04.304222 2260 server.go:744] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Feb 13 19:19:04.305120 kubelet[2260]: I0213 19:19:04.305094 2260 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Feb 13 19:19:04.305292 kubelet[2260]: I0213 19:19:04.305255 2260 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 13 19:19:04.305419 kubelet[2260]: I0213 19:19:04.305280 2260 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 13 19:19:04.305514 kubelet[2260]: I0213 19:19:04.305422 2260 topology_manager.go:138] "Creating topology manager with none policy" Feb 13 19:19:04.305514 kubelet[2260]: I0213 19:19:04.305430 2260 container_manager_linux.go:300] "Creating device plugin manager" Feb 13 19:19:04.305559 kubelet[2260]: I0213 19:19:04.305530 2260 state_mem.go:36] "Initialized new in-memory state store" Feb 13 19:19:04.306897 kubelet[2260]: I0213 19:19:04.306869 2260 kubelet.go:408] "Attempting to sync node with API server" Feb 13 19:19:04.306897 kubelet[2260]: I0213 19:19:04.306891 2260 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 13 19:19:04.306987 kubelet[2260]: I0213 19:19:04.306934 2260 kubelet.go:314] "Adding apiserver pod source" Feb 13 19:19:04.306987 kubelet[2260]: I0213 19:19:04.306952 2260 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 13 19:19:04.309636 kubelet[2260]: W0213 19:19:04.309579 2260 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.46:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.0.0.46:6443: connect: connection refused Feb 13 19:19:04.309703 kubelet[2260]: E0213 19:19:04.309628 2260 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.0.0.46:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.46:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:19:04.310057 kubelet[2260]: W0213 19:19:04.310015 2260 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.46:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.46:6443: connect: connection refused Feb 13 19:19:04.310100 kubelet[2260]: E0213 19:19:04.310055 2260 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.0.0.46:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.46:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:19:04.311529 kubelet[2260]: I0213 19:19:04.311499 2260 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v1.7.23" apiVersion="v1" Feb 13 19:19:04.313143 kubelet[2260]: I0213 19:19:04.313124 2260 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 13 19:19:04.313683 kubelet[2260]: W0213 19:19:04.313618 2260 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Feb 13 19:19:04.317142 kubelet[2260]: I0213 19:19:04.316945 2260 server.go:1269] "Started kubelet" Feb 13 19:19:04.318538 kubelet[2260]: I0213 19:19:04.318397 2260 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 13 19:19:04.318702 kubelet[2260]: I0213 19:19:04.318645 2260 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Feb 13 19:19:04.320312 kubelet[2260]: I0213 19:19:04.319966 2260 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 13 19:19:04.321419 kubelet[2260]: I0213 19:19:04.321400 2260 server.go:460] "Adding debug handlers to kubelet server" Feb 13 19:19:04.323017 kubelet[2260]: I0213 19:19:04.322587 2260 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 13 19:19:04.323121 kubelet[2260]: I0213 19:19:04.323102 2260 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Feb 13 19:19:04.323243 kubelet[2260]: I0213 19:19:04.323222 2260 volume_manager.go:289] "Starting Kubelet Volume Manager" Feb 13 19:19:04.323350 kubelet[2260]: I0213 19:19:04.323327 2260 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Feb 13 19:19:04.323412 kubelet[2260]: I0213 19:19:04.323396 2260 reconciler.go:26] "Reconciler: start to sync state" Feb 13 19:19:04.323749 kubelet[2260]: W0213 19:19:04.323705 2260 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.46:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.46:6443: connect: connection refused Feb 13 19:19:04.323809 kubelet[2260]: E0213 19:19:04.323758 2260 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.0.0.46:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.46:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:19:04.323958 kubelet[2260]: E0213 19:19:04.323934 2260 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:19:04.324137 kubelet[2260]: E0213 19:19:04.324003 2260 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.46:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.46:6443: connect: connection refused" interval="200ms" Feb 13 19:19:04.324744 kubelet[2260]: E0213 19:19:04.322331 2260 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.46:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.46:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.1823dab334ea7994 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-02-13 19:19:04.316926356 +0000 UTC m=+0.245058974,LastTimestamp:2025-02-13 19:19:04.316926356 +0000 UTC m=+0.245058974,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Feb 13 19:19:04.324744 kubelet[2260]: I0213 19:19:04.324458 2260 factory.go:221] Registration of the systemd container factory successfully Feb 13 19:19:04.324744 kubelet[2260]: I0213 19:19:04.324520 2260 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Feb 13 19:19:04.324744 kubelet[2260]: E0213 19:19:04.324610 2260 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Feb 13 19:19:04.325597 kubelet[2260]: I0213 19:19:04.325580 2260 factory.go:221] Registration of the containerd container factory successfully Feb 13 19:19:04.339684 kubelet[2260]: I0213 19:19:04.339563 2260 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 13 19:19:04.340511 kubelet[2260]: I0213 19:19:04.340376 2260 cpu_manager.go:214] "Starting CPU manager" policy="none" Feb 13 19:19:04.340556 kubelet[2260]: I0213 19:19:04.340544 2260 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Feb 13 19:19:04.340604 kubelet[2260]: I0213 19:19:04.340563 2260 state_mem.go:36] "Initialized new in-memory state store" Feb 13 19:19:04.341602 kubelet[2260]: I0213 19:19:04.341572 2260 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 13 19:19:04.341647 kubelet[2260]: I0213 19:19:04.341620 2260 status_manager.go:217] "Starting to sync pod status with apiserver" Feb 13 19:19:04.341647 kubelet[2260]: I0213 19:19:04.341637 2260 kubelet.go:2321] "Starting kubelet main sync loop" Feb 13 19:19:04.341741 kubelet[2260]: E0213 19:19:04.341679 2260 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 13 19:19:04.349025 kubelet[2260]: W0213 19:19:04.348968 2260 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.46:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.46:6443: connect: connection refused Feb 13 19:19:04.349080 kubelet[2260]: E0213 19:19:04.349032 2260 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.0.0.46:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.46:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:19:04.424347 kubelet[2260]: E0213 19:19:04.424313 2260 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:19:04.442577 kubelet[2260]: E0213 19:19:04.442552 2260 kubelet.go:2345] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Feb 13 19:19:04.524916 kubelet[2260]: E0213 19:19:04.524884 2260 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:19:04.525057 kubelet[2260]: E0213 19:19:04.525024 2260 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.46:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.46:6443: connect: connection refused" interval="400ms" Feb 13 19:19:04.625914 kubelet[2260]: E0213 19:19:04.625813 2260 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Feb 13 19:19:04.643027 kubelet[2260]: E0213 19:19:04.642989 2260 kubelet.go:2345] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Feb 13 19:19:04.680076 kubelet[2260]: I0213 19:19:04.680040 2260 policy_none.go:49] "None policy: Start" Feb 13 19:19:04.680827 kubelet[2260]: I0213 19:19:04.680807 2260 memory_manager.go:170] "Starting memorymanager" policy="None" Feb 13 19:19:04.680872 kubelet[2260]: I0213 19:19:04.680833 2260 state_mem.go:35] "Initializing new in-memory state store" Feb 13 19:19:04.688812 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Feb 13 19:19:04.700337 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Feb 13 19:19:04.703422 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Feb 13 19:19:04.718218 kubelet[2260]: I0213 19:19:04.718170 2260 manager.go:510] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 13 19:19:04.718441 kubelet[2260]: I0213 19:19:04.718424 2260 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 13 19:19:04.718500 kubelet[2260]: I0213 19:19:04.718442 2260 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 13 19:19:04.718666 kubelet[2260]: I0213 19:19:04.718650 2260 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 13 19:19:04.720372 kubelet[2260]: E0213 19:19:04.720351 2260 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Feb 13 19:19:04.819741 kubelet[2260]: I0213 19:19:04.819695 2260 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Feb 13 19:19:04.820084 kubelet[2260]: E0213 19:19:04.820046 2260 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.0.0.46:6443/api/v1/nodes\": dial tcp 10.0.0.46:6443: connect: connection refused" node="localhost" Feb 13 19:19:04.926178 kubelet[2260]: E0213 19:19:04.926052 2260 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.46:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.46:6443: connect: connection refused" interval="800ms" Feb 13 19:19:05.021491 kubelet[2260]: I0213 19:19:05.021450 2260 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Feb 13 19:19:05.021803 kubelet[2260]: E0213 19:19:05.021777 2260 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.0.0.46:6443/api/v1/nodes\": dial tcp 10.0.0.46:6443: connect: connection refused" node="localhost" Feb 13 19:19:05.049818 systemd[1]: Created slice kubepods-burstable-podcd2dab34f219cb156e1e5a1181eb00ad.slice - libcontainer container kubepods-burstable-podcd2dab34f219cb156e1e5a1181eb00ad.slice. Feb 13 19:19:05.063206 systemd[1]: Created slice kubepods-burstable-pod98eb2295280bc6da80e83f7636be329c.slice - libcontainer container kubepods-burstable-pod98eb2295280bc6da80e83f7636be329c.slice. Feb 13 19:19:05.067295 systemd[1]: Created slice kubepods-burstable-pod04cca2c455deeb5da380812dcab224d8.slice - libcontainer container kubepods-burstable-pod04cca2c455deeb5da380812dcab224d8.slice. Feb 13 19:19:05.129914 kubelet[2260]: I0213 19:19:05.129863 2260 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/cd2dab34f219cb156e1e5a1181eb00ad-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"cd2dab34f219cb156e1e5a1181eb00ad\") " pod="kube-system/kube-apiserver-localhost" Feb 13 19:19:05.129914 kubelet[2260]: I0213 19:19:05.129904 2260 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/98eb2295280bc6da80e83f7636be329c-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"98eb2295280bc6da80e83f7636be329c\") " pod="kube-system/kube-controller-manager-localhost" Feb 13 19:19:05.130387 kubelet[2260]: I0213 19:19:05.129927 2260 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/98eb2295280bc6da80e83f7636be329c-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"98eb2295280bc6da80e83f7636be329c\") " pod="kube-system/kube-controller-manager-localhost" Feb 13 19:19:05.130387 kubelet[2260]: I0213 19:19:05.129946 2260 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/98eb2295280bc6da80e83f7636be329c-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"98eb2295280bc6da80e83f7636be329c\") " pod="kube-system/kube-controller-manager-localhost" Feb 13 19:19:05.130387 kubelet[2260]: I0213 19:19:05.129964 2260 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/98eb2295280bc6da80e83f7636be329c-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"98eb2295280bc6da80e83f7636be329c\") " pod="kube-system/kube-controller-manager-localhost" Feb 13 19:19:05.130387 kubelet[2260]: I0213 19:19:05.129982 2260 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/cd2dab34f219cb156e1e5a1181eb00ad-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"cd2dab34f219cb156e1e5a1181eb00ad\") " pod="kube-system/kube-apiserver-localhost" Feb 13 19:19:05.130387 kubelet[2260]: I0213 19:19:05.129999 2260 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/cd2dab34f219cb156e1e5a1181eb00ad-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"cd2dab34f219cb156e1e5a1181eb00ad\") " pod="kube-system/kube-apiserver-localhost" Feb 13 19:19:05.130500 kubelet[2260]: I0213 19:19:05.130016 2260 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/98eb2295280bc6da80e83f7636be329c-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"98eb2295280bc6da80e83f7636be329c\") " pod="kube-system/kube-controller-manager-localhost" Feb 13 19:19:05.130500 kubelet[2260]: I0213 19:19:05.130033 2260 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/04cca2c455deeb5da380812dcab224d8-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"04cca2c455deeb5da380812dcab224d8\") " pod="kube-system/kube-scheduler-localhost" Feb 13 19:19:05.339099 kubelet[2260]: W0213 19:19:05.338956 2260 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.46:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.46:6443: connect: connection refused Feb 13 19:19:05.339099 kubelet[2260]: E0213 19:19:05.339031 2260 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.0.0.46:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.46:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:19:05.361522 kubelet[2260]: E0213 19:19:05.361482 2260 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:05.362073 containerd[1515]: time="2025-02-13T19:19:05.362031010Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:cd2dab34f219cb156e1e5a1181eb00ad,Namespace:kube-system,Attempt:0,}" Feb 13 19:19:05.366229 kubelet[2260]: E0213 19:19:05.366208 2260 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:05.366545 containerd[1515]: time="2025-02-13T19:19:05.366516281Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:98eb2295280bc6da80e83f7636be329c,Namespace:kube-system,Attempt:0,}" Feb 13 19:19:05.371800 kubelet[2260]: E0213 19:19:05.371775 2260 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:05.372068 containerd[1515]: time="2025-02-13T19:19:05.372044128Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:04cca2c455deeb5da380812dcab224d8,Namespace:kube-system,Attempt:0,}" Feb 13 19:19:05.423210 kubelet[2260]: I0213 19:19:05.423168 2260 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Feb 13 19:19:05.423508 kubelet[2260]: E0213 19:19:05.423473 2260 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.0.0.46:6443/api/v1/nodes\": dial tcp 10.0.0.46:6443: connect: connection refused" node="localhost" Feb 13 19:19:05.559575 kubelet[2260]: W0213 19:19:05.559516 2260 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.46:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.0.0.46:6443: connect: connection refused Feb 13 19:19:05.559710 kubelet[2260]: E0213 19:19:05.559579 2260 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.0.0.46:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.46:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:19:05.630388 kubelet[2260]: W0213 19:19:05.630275 2260 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.46:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.46:6443: connect: connection refused Feb 13 19:19:05.630388 kubelet[2260]: E0213 19:19:05.630328 2260 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.0.0.46:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.46:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:19:05.727164 kubelet[2260]: E0213 19:19:05.727122 2260 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.46:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.46:6443: connect: connection refused" interval="1.6s" Feb 13 19:19:05.909644 kubelet[2260]: W0213 19:19:05.909535 2260 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.46:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.46:6443: connect: connection refused Feb 13 19:19:05.909644 kubelet[2260]: E0213 19:19:05.909586 2260 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.0.0.46:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.46:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:19:06.224935 kubelet[2260]: I0213 19:19:06.224853 2260 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Feb 13 19:19:06.225291 kubelet[2260]: E0213 19:19:06.225157 2260 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.0.0.46:6443/api/v1/nodes\": dial tcp 10.0.0.46:6443: connect: connection refused" node="localhost" Feb 13 19:19:06.328136 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1670747245.mount: Deactivated successfully. Feb 13 19:19:06.333332 containerd[1515]: time="2025-02-13T19:19:06.333288787Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 19:19:06.335141 containerd[1515]: time="2025-02-13T19:19:06.335054225Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312056" Feb 13 19:19:06.338651 containerd[1515]: time="2025-02-13T19:19:06.338603817Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 19:19:06.339965 containerd[1515]: time="2025-02-13T19:19:06.339910321Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 19:19:06.340575 containerd[1515]: time="2025-02-13T19:19:06.340536757Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 19:19:06.341216 containerd[1515]: time="2025-02-13T19:19:06.341106894Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Feb 13 19:19:06.342325 containerd[1515]: time="2025-02-13T19:19:06.342291104Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Feb 13 19:19:06.343183 containerd[1515]: time="2025-02-13T19:19:06.343144266Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Feb 13 19:19:06.343889 containerd[1515]: time="2025-02-13T19:19:06.343867839Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 971.762104ms" Feb 13 19:19:06.349670 containerd[1515]: time="2025-02-13T19:19:06.349627985Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 987.494011ms" Feb 13 19:19:06.350746 containerd[1515]: time="2025-02-13T19:19:06.350704537Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 984.123243ms" Feb 13 19:19:06.364036 kubelet[2260]: E0213 19:19:06.363982 2260 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.0.0.46:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.46:6443: connect: connection refused" logger="UnhandledError" Feb 13 19:19:06.468848 containerd[1515]: time="2025-02-13T19:19:06.468580011Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:19:06.468848 containerd[1515]: time="2025-02-13T19:19:06.468634024Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:19:06.468848 containerd[1515]: time="2025-02-13T19:19:06.468654834Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:19:06.468848 containerd[1515]: time="2025-02-13T19:19:06.468734648Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:19:06.469446 containerd[1515]: time="2025-02-13T19:19:06.469352798Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:19:06.469517 containerd[1515]: time="2025-02-13T19:19:06.469483179Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:19:06.469707 containerd[1515]: time="2025-02-13T19:19:06.469515090Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:19:06.469707 containerd[1515]: time="2025-02-13T19:19:06.469652705Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:19:06.470620 containerd[1515]: time="2025-02-13T19:19:06.470423008Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:19:06.470620 containerd[1515]: time="2025-02-13T19:19:06.470464638Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:19:06.470620 containerd[1515]: time="2025-02-13T19:19:06.470479507Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:19:06.470620 containerd[1515]: time="2025-02-13T19:19:06.470547838Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:19:06.488354 systemd[1]: Started cri-containerd-45efa36cd1efaa0462052b08996350ce6504a9133bf1c707f36e123cb71e8784.scope - libcontainer container 45efa36cd1efaa0462052b08996350ce6504a9133bf1c707f36e123cb71e8784. Feb 13 19:19:06.492455 systemd[1]: Started cri-containerd-b6be9fadf5815ca2770be1334a50d38c9bc17bdfbd5f6fbcf99b22dd5bf617b7.scope - libcontainer container b6be9fadf5815ca2770be1334a50d38c9bc17bdfbd5f6fbcf99b22dd5bf617b7. Feb 13 19:19:06.494166 systemd[1]: Started cri-containerd-ca3c86c1f02c8ad2765344e47bfbe53b9b72b95d70319bc18c67517034eff842.scope - libcontainer container ca3c86c1f02c8ad2765344e47bfbe53b9b72b95d70319bc18c67517034eff842. Feb 13 19:19:06.525984 containerd[1515]: time="2025-02-13T19:19:06.525924551Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:98eb2295280bc6da80e83f7636be329c,Namespace:kube-system,Attempt:0,} returns sandbox id \"45efa36cd1efaa0462052b08996350ce6504a9133bf1c707f36e123cb71e8784\"" Feb 13 19:19:06.527499 kubelet[2260]: E0213 19:19:06.527424 2260 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:06.530747 containerd[1515]: time="2025-02-13T19:19:06.530720492Z" level=info msg="CreateContainer within sandbox \"45efa36cd1efaa0462052b08996350ce6504a9133bf1c707f36e123cb71e8784\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Feb 13 19:19:06.531516 containerd[1515]: time="2025-02-13T19:19:06.531472068Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:04cca2c455deeb5da380812dcab224d8,Namespace:kube-system,Attempt:0,} returns sandbox id \"b6be9fadf5815ca2770be1334a50d38c9bc17bdfbd5f6fbcf99b22dd5bf617b7\"" Feb 13 19:19:06.531922 kubelet[2260]: E0213 19:19:06.531901 2260 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:06.533618 containerd[1515]: time="2025-02-13T19:19:06.533426670Z" level=info msg="CreateContainer within sandbox \"b6be9fadf5815ca2770be1334a50d38c9bc17bdfbd5f6fbcf99b22dd5bf617b7\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Feb 13 19:19:06.536394 containerd[1515]: time="2025-02-13T19:19:06.536359285Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:cd2dab34f219cb156e1e5a1181eb00ad,Namespace:kube-system,Attempt:0,} returns sandbox id \"ca3c86c1f02c8ad2765344e47bfbe53b9b72b95d70319bc18c67517034eff842\"" Feb 13 19:19:06.537002 kubelet[2260]: E0213 19:19:06.536978 2260 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:06.538555 containerd[1515]: time="2025-02-13T19:19:06.538525143Z" level=info msg="CreateContainer within sandbox \"ca3c86c1f02c8ad2765344e47bfbe53b9b72b95d70319bc18c67517034eff842\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Feb 13 19:19:06.551450 containerd[1515]: time="2025-02-13T19:19:06.551408761Z" level=info msg="CreateContainer within sandbox \"45efa36cd1efaa0462052b08996350ce6504a9133bf1c707f36e123cb71e8784\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"e153b1ed2508685195958e0832b089bbf739eff164e29c49acf0502d8b5f90ad\"" Feb 13 19:19:06.551880 containerd[1515]: time="2025-02-13T19:19:06.551849969Z" level=info msg="StartContainer for \"e153b1ed2508685195958e0832b089bbf739eff164e29c49acf0502d8b5f90ad\"" Feb 13 19:19:06.557393 containerd[1515]: time="2025-02-13T19:19:06.557362339Z" level=info msg="CreateContainer within sandbox \"b6be9fadf5815ca2770be1334a50d38c9bc17bdfbd5f6fbcf99b22dd5bf617b7\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"7a068ffa4ba019e158585d959dc599904f5d1cbffe4641a299d3b6634a1ff75e\"" Feb 13 19:19:06.557876 containerd[1515]: time="2025-02-13T19:19:06.557854436Z" level=info msg="StartContainer for \"7a068ffa4ba019e158585d959dc599904f5d1cbffe4641a299d3b6634a1ff75e\"" Feb 13 19:19:06.565641 containerd[1515]: time="2025-02-13T19:19:06.565612628Z" level=info msg="CreateContainer within sandbox \"ca3c86c1f02c8ad2765344e47bfbe53b9b72b95d70319bc18c67517034eff842\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"183a1eb9401f180735757516d11a0a94495b81800c8eb967758670923148a1ea\"" Feb 13 19:19:06.566861 containerd[1515]: time="2025-02-13T19:19:06.565969174Z" level=info msg="StartContainer for \"183a1eb9401f180735757516d11a0a94495b81800c8eb967758670923148a1ea\"" Feb 13 19:19:06.579355 systemd[1]: Started cri-containerd-e153b1ed2508685195958e0832b089bbf739eff164e29c49acf0502d8b5f90ad.scope - libcontainer container e153b1ed2508685195958e0832b089bbf739eff164e29c49acf0502d8b5f90ad. Feb 13 19:19:06.583139 systemd[1]: Started cri-containerd-7a068ffa4ba019e158585d959dc599904f5d1cbffe4641a299d3b6634a1ff75e.scope - libcontainer container 7a068ffa4ba019e158585d959dc599904f5d1cbffe4641a299d3b6634a1ff75e. Feb 13 19:19:06.589230 systemd[1]: Started cri-containerd-183a1eb9401f180735757516d11a0a94495b81800c8eb967758670923148a1ea.scope - libcontainer container 183a1eb9401f180735757516d11a0a94495b81800c8eb967758670923148a1ea. Feb 13 19:19:06.620809 containerd[1515]: time="2025-02-13T19:19:06.620777403Z" level=info msg="StartContainer for \"e153b1ed2508685195958e0832b089bbf739eff164e29c49acf0502d8b5f90ad\" returns successfully" Feb 13 19:19:06.633099 containerd[1515]: time="2025-02-13T19:19:06.633050385Z" level=info msg="StartContainer for \"7a068ffa4ba019e158585d959dc599904f5d1cbffe4641a299d3b6634a1ff75e\" returns successfully" Feb 13 19:19:06.635957 containerd[1515]: time="2025-02-13T19:19:06.635870002Z" level=info msg="StartContainer for \"183a1eb9401f180735757516d11a0a94495b81800c8eb967758670923148a1ea\" returns successfully" Feb 13 19:19:07.348606 kubelet[2260]: E0213 19:19:07.348506 2260 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:07.350232 kubelet[2260]: E0213 19:19:07.350216 2260 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:07.351363 kubelet[2260]: E0213 19:19:07.351348 2260 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:07.584088 kubelet[2260]: E0213 19:19:07.583930 2260 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Feb 13 19:19:07.826758 kubelet[2260]: I0213 19:19:07.826619 2260 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Feb 13 19:19:07.832417 kubelet[2260]: I0213 19:19:07.832366 2260 kubelet_node_status.go:75] "Successfully registered node" node="localhost" Feb 13 19:19:07.832417 kubelet[2260]: E0213 19:19:07.832411 2260 kubelet_node_status.go:535] "Error updating node status, will retry" err="error getting node \"localhost\": node \"localhost\" not found" Feb 13 19:19:08.311748 kubelet[2260]: I0213 19:19:08.311609 2260 apiserver.go:52] "Watching apiserver" Feb 13 19:19:08.324066 kubelet[2260]: I0213 19:19:08.324024 2260 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Feb 13 19:19:08.357113 kubelet[2260]: E0213 19:19:08.357072 2260 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-localhost" Feb 13 19:19:08.357501 kubelet[2260]: E0213 19:19:08.357226 2260 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:09.437238 kubelet[2260]: E0213 19:19:09.437210 2260 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:10.355338 kubelet[2260]: E0213 19:19:10.355304 2260 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:11.005960 systemd[1]: Reload requested from client PID 2541 ('systemctl') (unit session-9.scope)... Feb 13 19:19:11.005976 systemd[1]: Reloading... Feb 13 19:19:11.100870 zram_generator::config[2586]: No configuration found. Feb 13 19:19:11.233215 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 13 19:19:11.347419 systemd[1]: Reloading finished in 341 ms. Feb 13 19:19:11.374657 kubelet[2260]: I0213 19:19:11.374615 2260 dynamic_cafile_content.go:174] "Shutting down controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Feb 13 19:19:11.374717 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:19:11.391585 systemd[1]: kubelet.service: Deactivated successfully. Feb 13 19:19:11.391881 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:19:11.391929 systemd[1]: kubelet.service: Consumed 718ms CPU time, 119.6M memory peak. Feb 13 19:19:11.403557 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Feb 13 19:19:11.569278 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Feb 13 19:19:11.573104 (kubelet)[2630]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Feb 13 19:19:11.608578 kubelet[2630]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 13 19:19:11.608578 kubelet[2630]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Feb 13 19:19:11.608578 kubelet[2630]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 13 19:19:11.608578 kubelet[2630]: I0213 19:19:11.608547 2630 server.go:206] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 13 19:19:11.614455 kubelet[2630]: I0213 19:19:11.614414 2630 server.go:486] "Kubelet version" kubeletVersion="v1.31.0" Feb 13 19:19:11.614455 kubelet[2630]: I0213 19:19:11.614442 2630 server.go:488] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 13 19:19:11.614774 kubelet[2630]: I0213 19:19:11.614751 2630 server.go:929] "Client rotation is on, will bootstrap in background" Feb 13 19:19:11.617274 kubelet[2630]: I0213 19:19:11.617228 2630 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Feb 13 19:19:11.619532 kubelet[2630]: I0213 19:19:11.619394 2630 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Feb 13 19:19:11.623855 kubelet[2630]: E0213 19:19:11.623825 2630 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Feb 13 19:19:11.623855 kubelet[2630]: I0213 19:19:11.623856 2630 server.go:1403] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Feb 13 19:19:11.629313 kubelet[2630]: I0213 19:19:11.629280 2630 server.go:744] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Feb 13 19:19:11.629423 kubelet[2630]: I0213 19:19:11.629406 2630 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Feb 13 19:19:11.629569 kubelet[2630]: I0213 19:19:11.629543 2630 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 13 19:19:11.629841 kubelet[2630]: I0213 19:19:11.629567 2630 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 13 19:19:11.629921 kubelet[2630]: I0213 19:19:11.629848 2630 topology_manager.go:138] "Creating topology manager with none policy" Feb 13 19:19:11.629921 kubelet[2630]: I0213 19:19:11.629858 2630 container_manager_linux.go:300] "Creating device plugin manager" Feb 13 19:19:11.629921 kubelet[2630]: I0213 19:19:11.629889 2630 state_mem.go:36] "Initialized new in-memory state store" Feb 13 19:19:11.630004 kubelet[2630]: I0213 19:19:11.629991 2630 kubelet.go:408] "Attempting to sync node with API server" Feb 13 19:19:11.630032 kubelet[2630]: I0213 19:19:11.630006 2630 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 13 19:19:11.630056 kubelet[2630]: I0213 19:19:11.630036 2630 kubelet.go:314] "Adding apiserver pod source" Feb 13 19:19:11.630056 kubelet[2630]: I0213 19:19:11.630050 2630 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 13 19:19:11.632613 kubelet[2630]: I0213 19:19:11.632584 2630 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v1.7.23" apiVersion="v1" Feb 13 19:19:11.634561 kubelet[2630]: I0213 19:19:11.633048 2630 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 13 19:19:11.634561 kubelet[2630]: I0213 19:19:11.633477 2630 server.go:1269] "Started kubelet" Feb 13 19:19:11.634561 kubelet[2630]: I0213 19:19:11.634474 2630 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 13 19:19:11.638568 kubelet[2630]: I0213 19:19:11.638215 2630 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Feb 13 19:19:11.640855 kubelet[2630]: I0213 19:19:11.640697 2630 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 13 19:19:11.640855 kubelet[2630]: I0213 19:19:11.640718 2630 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 13 19:19:11.641241 kubelet[2630]: I0213 19:19:11.641222 2630 server.go:460] "Adding debug handlers to kubelet server" Feb 13 19:19:11.643432 kubelet[2630]: I0213 19:19:11.643369 2630 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Feb 13 19:19:11.646226 kubelet[2630]: I0213 19:19:11.645691 2630 volume_manager.go:289] "Starting Kubelet Volume Manager" Feb 13 19:19:11.646226 kubelet[2630]: I0213 19:19:11.646035 2630 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Feb 13 19:19:11.646226 kubelet[2630]: I0213 19:19:11.646176 2630 reconciler.go:26] "Reconciler: start to sync state" Feb 13 19:19:11.650558 kubelet[2630]: I0213 19:19:11.650490 2630 factory.go:221] Registration of the systemd container factory successfully Feb 13 19:19:11.651385 kubelet[2630]: E0213 19:19:11.651358 2630 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Feb 13 19:19:11.652050 kubelet[2630]: I0213 19:19:11.651529 2630 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Feb 13 19:19:11.652762 kubelet[2630]: I0213 19:19:11.652733 2630 factory.go:221] Registration of the containerd container factory successfully Feb 13 19:19:11.655749 kubelet[2630]: I0213 19:19:11.655700 2630 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 13 19:19:11.656900 kubelet[2630]: I0213 19:19:11.656874 2630 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 13 19:19:11.656938 kubelet[2630]: I0213 19:19:11.656909 2630 status_manager.go:217] "Starting to sync pod status with apiserver" Feb 13 19:19:11.656938 kubelet[2630]: I0213 19:19:11.656928 2630 kubelet.go:2321] "Starting kubelet main sync loop" Feb 13 19:19:11.657030 kubelet[2630]: E0213 19:19:11.656988 2630 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 13 19:19:11.691944 kubelet[2630]: I0213 19:19:11.691906 2630 cpu_manager.go:214] "Starting CPU manager" policy="none" Feb 13 19:19:11.691944 kubelet[2630]: I0213 19:19:11.691925 2630 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Feb 13 19:19:11.691944 kubelet[2630]: I0213 19:19:11.691941 2630 state_mem.go:36] "Initialized new in-memory state store" Feb 13 19:19:11.692118 kubelet[2630]: I0213 19:19:11.692076 2630 state_mem.go:88] "Updated default CPUSet" cpuSet="" Feb 13 19:19:11.692118 kubelet[2630]: I0213 19:19:11.692085 2630 state_mem.go:96] "Updated CPUSet assignments" assignments={} Feb 13 19:19:11.692118 kubelet[2630]: I0213 19:19:11.692102 2630 policy_none.go:49] "None policy: Start" Feb 13 19:19:11.692586 kubelet[2630]: I0213 19:19:11.692568 2630 memory_manager.go:170] "Starting memorymanager" policy="None" Feb 13 19:19:11.692586 kubelet[2630]: I0213 19:19:11.692586 2630 state_mem.go:35] "Initializing new in-memory state store" Feb 13 19:19:11.692717 kubelet[2630]: I0213 19:19:11.692700 2630 state_mem.go:75] "Updated machine memory state" Feb 13 19:19:11.697737 kubelet[2630]: I0213 19:19:11.697708 2630 manager.go:510] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 13 19:19:11.698030 kubelet[2630]: I0213 19:19:11.697877 2630 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 13 19:19:11.698030 kubelet[2630]: I0213 19:19:11.697893 2630 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 13 19:19:11.698100 kubelet[2630]: I0213 19:19:11.698036 2630 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 13 19:19:12.035385 kubelet[2630]: E0213 19:19:12.035232 2630 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Feb 13 19:19:12.071722 kubelet[2630]: I0213 19:19:12.071674 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/98eb2295280bc6da80e83f7636be329c-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"98eb2295280bc6da80e83f7636be329c\") " pod="kube-system/kube-controller-manager-localhost" Feb 13 19:19:12.071722 kubelet[2630]: I0213 19:19:12.071714 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/98eb2295280bc6da80e83f7636be329c-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"98eb2295280bc6da80e83f7636be329c\") " pod="kube-system/kube-controller-manager-localhost" Feb 13 19:19:12.071722 kubelet[2630]: I0213 19:19:12.071734 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/98eb2295280bc6da80e83f7636be329c-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"98eb2295280bc6da80e83f7636be329c\") " pod="kube-system/kube-controller-manager-localhost" Feb 13 19:19:12.071955 kubelet[2630]: I0213 19:19:12.071760 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/cd2dab34f219cb156e1e5a1181eb00ad-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"cd2dab34f219cb156e1e5a1181eb00ad\") " pod="kube-system/kube-apiserver-localhost" Feb 13 19:19:12.071955 kubelet[2630]: I0213 19:19:12.071776 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/cd2dab34f219cb156e1e5a1181eb00ad-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"cd2dab34f219cb156e1e5a1181eb00ad\") " pod="kube-system/kube-apiserver-localhost" Feb 13 19:19:12.071955 kubelet[2630]: I0213 19:19:12.071868 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/98eb2295280bc6da80e83f7636be329c-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"98eb2295280bc6da80e83f7636be329c\") " pod="kube-system/kube-controller-manager-localhost" Feb 13 19:19:12.071955 kubelet[2630]: I0213 19:19:12.071907 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/04cca2c455deeb5da380812dcab224d8-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"04cca2c455deeb5da380812dcab224d8\") " pod="kube-system/kube-scheduler-localhost" Feb 13 19:19:12.071955 kubelet[2630]: I0213 19:19:12.071927 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/cd2dab34f219cb156e1e5a1181eb00ad-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"cd2dab34f219cb156e1e5a1181eb00ad\") " pod="kube-system/kube-apiserver-localhost" Feb 13 19:19:12.072076 kubelet[2630]: I0213 19:19:12.071941 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/98eb2295280bc6da80e83f7636be329c-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"98eb2295280bc6da80e83f7636be329c\") " pod="kube-system/kube-controller-manager-localhost" Feb 13 19:19:12.077981 kubelet[2630]: I0213 19:19:12.077962 2630 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Feb 13 19:19:12.137298 kubelet[2630]: I0213 19:19:12.137255 2630 kubelet_node_status.go:111] "Node was previously registered" node="localhost" Feb 13 19:19:12.137471 kubelet[2630]: I0213 19:19:12.137335 2630 kubelet_node_status.go:75] "Successfully registered node" node="localhost" Feb 13 19:19:12.326733 kubelet[2630]: E0213 19:19:12.326617 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:12.327281 kubelet[2630]: E0213 19:19:12.327203 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:12.335761 kubelet[2630]: E0213 19:19:12.335711 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:12.630451 kubelet[2630]: I0213 19:19:12.630351 2630 apiserver.go:52] "Watching apiserver" Feb 13 19:19:12.646375 kubelet[2630]: I0213 19:19:12.646323 2630 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Feb 13 19:19:12.744215 kubelet[2630]: E0213 19:19:12.741336 2630 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-localhost\" already exists" pod="kube-system/kube-controller-manager-localhost" Feb 13 19:19:12.744215 kubelet[2630]: E0213 19:19:12.741538 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:12.744215 kubelet[2630]: E0213 19:19:12.742478 2630 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Feb 13 19:19:12.744215 kubelet[2630]: E0213 19:19:12.742680 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:12.744215 kubelet[2630]: E0213 19:19:12.742733 2630 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost" Feb 13 19:19:12.744215 kubelet[2630]: E0213 19:19:12.744000 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:12.761741 kubelet[2630]: I0213 19:19:12.761542 2630 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.761523454 podStartE2EDuration="1.761523454s" podCreationTimestamp="2025-02-13 19:19:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 19:19:12.76145298 +0000 UTC m=+1.183306080" watchObservedRunningTime="2025-02-13 19:19:12.761523454 +0000 UTC m=+1.183376544" Feb 13 19:19:12.777218 kubelet[2630]: I0213 19:19:12.774642 2630 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=3.77462858 podStartE2EDuration="3.77462858s" podCreationTimestamp="2025-02-13 19:19:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 19:19:12.774492059 +0000 UTC m=+1.196345149" watchObservedRunningTime="2025-02-13 19:19:12.77462858 +0000 UTC m=+1.196481670" Feb 13 19:19:12.807878 kubelet[2630]: I0213 19:19:12.807820 2630 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.807802634 podStartE2EDuration="1.807802634s" podCreationTimestamp="2025-02-13 19:19:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 19:19:12.78950671 +0000 UTC m=+1.211359800" watchObservedRunningTime="2025-02-13 19:19:12.807802634 +0000 UTC m=+1.229655714" Feb 13 19:19:13.675064 kubelet[2630]: E0213 19:19:13.674372 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:13.675064 kubelet[2630]: E0213 19:19:13.674799 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:13.675064 kubelet[2630]: E0213 19:19:13.674974 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:16.070856 kubelet[2630]: I0213 19:19:16.070820 2630 kuberuntime_manager.go:1633] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Feb 13 19:19:16.071569 kubelet[2630]: I0213 19:19:16.071414 2630 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Feb 13 19:19:16.071608 containerd[1515]: time="2025-02-13T19:19:16.071255986Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Feb 13 19:19:16.818085 systemd[1]: Created slice kubepods-besteffort-pode1988f8f_32c4_4d13_ad84_210cd55d68a0.slice - libcontainer container kubepods-besteffort-pode1988f8f_32c4_4d13_ad84_210cd55d68a0.slice. Feb 13 19:19:16.867497 sudo[1723]: pam_unix(sudo:session): session closed for user root Feb 13 19:19:16.869174 sshd[1716]: Connection closed by 10.0.0.1 port 34398 Feb 13 19:19:16.869676 sshd-session[1713]: pam_unix(sshd:session): session closed for user core Feb 13 19:19:16.874233 systemd[1]: sshd@8-10.0.0.46:22-10.0.0.1:34398.service: Deactivated successfully. Feb 13 19:19:16.876374 systemd[1]: session-9.scope: Deactivated successfully. Feb 13 19:19:16.876588 systemd[1]: session-9.scope: Consumed 4.292s CPU time, 214M memory peak. Feb 13 19:19:16.877777 systemd-logind[1495]: Session 9 logged out. Waiting for processes to exit. Feb 13 19:19:16.878642 systemd-logind[1495]: Removed session 9. Feb 13 19:19:16.883147 kubelet[2630]: E0213 19:19:16.883114 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:16.936772 kubelet[2630]: I0213 19:19:16.936694 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e1988f8f-32c4-4d13-ad84-210cd55d68a0-lib-modules\") pod \"kube-proxy-bq9xd\" (UID: \"e1988f8f-32c4-4d13-ad84-210cd55d68a0\") " pod="kube-system/kube-proxy-bq9xd" Feb 13 19:19:16.936772 kubelet[2630]: I0213 19:19:16.936761 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mbn7\" (UniqueName: \"kubernetes.io/projected/e1988f8f-32c4-4d13-ad84-210cd55d68a0-kube-api-access-7mbn7\") pod \"kube-proxy-bq9xd\" (UID: \"e1988f8f-32c4-4d13-ad84-210cd55d68a0\") " pod="kube-system/kube-proxy-bq9xd" Feb 13 19:19:16.936772 kubelet[2630]: I0213 19:19:16.936793 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/e1988f8f-32c4-4d13-ad84-210cd55d68a0-kube-proxy\") pod \"kube-proxy-bq9xd\" (UID: \"e1988f8f-32c4-4d13-ad84-210cd55d68a0\") " pod="kube-system/kube-proxy-bq9xd" Feb 13 19:19:16.936989 kubelet[2630]: I0213 19:19:16.936808 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/e1988f8f-32c4-4d13-ad84-210cd55d68a0-xtables-lock\") pod \"kube-proxy-bq9xd\" (UID: \"e1988f8f-32c4-4d13-ad84-210cd55d68a0\") " pod="kube-system/kube-proxy-bq9xd" Feb 13 19:19:17.136758 kubelet[2630]: E0213 19:19:17.136613 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:17.137513 containerd[1515]: time="2025-02-13T19:19:17.137274311Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-bq9xd,Uid:e1988f8f-32c4-4d13-ad84-210cd55d68a0,Namespace:kube-system,Attempt:0,}" Feb 13 19:19:17.169220 containerd[1515]: time="2025-02-13T19:19:17.168667256Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:19:17.169220 containerd[1515]: time="2025-02-13T19:19:17.168742569Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:19:17.169220 containerd[1515]: time="2025-02-13T19:19:17.168762055Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:19:17.169220 containerd[1515]: time="2025-02-13T19:19:17.168996461Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:19:17.187449 systemd[1]: Created slice kubepods-besteffort-pod1a93465c_981c_4768_beb7_190f707eb4f8.slice - libcontainer container kubepods-besteffort-pod1a93465c_981c_4768_beb7_190f707eb4f8.slice. Feb 13 19:19:17.209343 systemd[1]: Started cri-containerd-15f6932c4a364a0c4fa77f4da1320158306f9cc027f9051adf5e7dbb4d8d2348.scope - libcontainer container 15f6932c4a364a0c4fa77f4da1320158306f9cc027f9051adf5e7dbb4d8d2348. Feb 13 19:19:17.229737 containerd[1515]: time="2025-02-13T19:19:17.229637045Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-bq9xd,Uid:e1988f8f-32c4-4d13-ad84-210cd55d68a0,Namespace:kube-system,Attempt:0,} returns sandbox id \"15f6932c4a364a0c4fa77f4da1320158306f9cc027f9051adf5e7dbb4d8d2348\"" Feb 13 19:19:17.230341 kubelet[2630]: E0213 19:19:17.230317 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:17.232915 containerd[1515]: time="2025-02-13T19:19:17.232846641Z" level=info msg="CreateContainer within sandbox \"15f6932c4a364a0c4fa77f4da1320158306f9cc027f9051adf5e7dbb4d8d2348\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Feb 13 19:19:17.250026 containerd[1515]: time="2025-02-13T19:19:17.249979167Z" level=info msg="CreateContainer within sandbox \"15f6932c4a364a0c4fa77f4da1320158306f9cc027f9051adf5e7dbb4d8d2348\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"4dcc41010075bc72f05de97511dec2c1bd095259457a9422ded63af9714f1f55\"" Feb 13 19:19:17.250597 containerd[1515]: time="2025-02-13T19:19:17.250565130Z" level=info msg="StartContainer for \"4dcc41010075bc72f05de97511dec2c1bd095259457a9422ded63af9714f1f55\"" Feb 13 19:19:17.278416 systemd[1]: Started cri-containerd-4dcc41010075bc72f05de97511dec2c1bd095259457a9422ded63af9714f1f55.scope - libcontainer container 4dcc41010075bc72f05de97511dec2c1bd095259457a9422ded63af9714f1f55. Feb 13 19:19:17.311787 containerd[1515]: time="2025-02-13T19:19:17.311690114Z" level=info msg="StartContainer for \"4dcc41010075bc72f05de97511dec2c1bd095259457a9422ded63af9714f1f55\" returns successfully" Feb 13 19:19:17.338357 kubelet[2630]: I0213 19:19:17.338294 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/1a93465c-981c-4768-beb7-190f707eb4f8-var-lib-calico\") pod \"tigera-operator-76c4976dd7-mgpk6\" (UID: \"1a93465c-981c-4768-beb7-190f707eb4f8\") " pod="tigera-operator/tigera-operator-76c4976dd7-mgpk6" Feb 13 19:19:17.338577 kubelet[2630]: I0213 19:19:17.338531 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ln7cc\" (UniqueName: \"kubernetes.io/projected/1a93465c-981c-4768-beb7-190f707eb4f8-kube-api-access-ln7cc\") pod \"tigera-operator-76c4976dd7-mgpk6\" (UID: \"1a93465c-981c-4768-beb7-190f707eb4f8\") " pod="tigera-operator/tigera-operator-76c4976dd7-mgpk6" Feb 13 19:19:17.492798 containerd[1515]: time="2025-02-13T19:19:17.492742623Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-76c4976dd7-mgpk6,Uid:1a93465c-981c-4768-beb7-190f707eb4f8,Namespace:tigera-operator,Attempt:0,}" Feb 13 19:19:17.520044 containerd[1515]: time="2025-02-13T19:19:17.519946631Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:19:17.520044 containerd[1515]: time="2025-02-13T19:19:17.520006654Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:19:17.520044 containerd[1515]: time="2025-02-13T19:19:17.520018807Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:19:17.520296 containerd[1515]: time="2025-02-13T19:19:17.520096184Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:19:17.540336 systemd[1]: Started cri-containerd-a0074be09150660c12027fccee5ecec70184b4dc858cf609107c4cfe21e8eb8c.scope - libcontainer container a0074be09150660c12027fccee5ecec70184b4dc858cf609107c4cfe21e8eb8c. Feb 13 19:19:17.578360 containerd[1515]: time="2025-02-13T19:19:17.578315128Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-76c4976dd7-mgpk6,Uid:1a93465c-981c-4768-beb7-190f707eb4f8,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"a0074be09150660c12027fccee5ecec70184b4dc858cf609107c4cfe21e8eb8c\"" Feb 13 19:19:17.580061 containerd[1515]: time="2025-02-13T19:19:17.580013784Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.36.2\"" Feb 13 19:19:17.681741 kubelet[2630]: E0213 19:19:17.681444 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:17.682002 kubelet[2630]: E0213 19:19:17.681915 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:17.697805 kubelet[2630]: I0213 19:19:17.697710 2630 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-bq9xd" podStartSLOduration=1.697693551 podStartE2EDuration="1.697693551s" podCreationTimestamp="2025-02-13 19:19:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 19:19:17.697477721 +0000 UTC m=+6.119330811" watchObservedRunningTime="2025-02-13 19:19:17.697693551 +0000 UTC m=+6.119546641" Feb 13 19:19:18.310677 kubelet[2630]: E0213 19:19:18.310621 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:18.682708 kubelet[2630]: E0213 19:19:18.682685 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:18.908768 update_engine[1501]: I20250213 19:19:18.908699 1501 update_attempter.cc:509] Updating boot flags... Feb 13 19:19:18.932231 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 39 scanned by (udev-worker) (2974) Feb 13 19:19:18.973330 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 39 scanned by (udev-worker) (2975) Feb 13 19:19:19.269814 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2029325810.mount: Deactivated successfully. Feb 13 19:19:19.358215 kubelet[2630]: E0213 19:19:19.358169 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:19.561715 containerd[1515]: time="2025-02-13T19:19:19.561600342Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.36.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:19:19.562420 containerd[1515]: time="2025-02-13T19:19:19.562385802Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.36.2: active requests=0, bytes read=21762497" Feb 13 19:19:19.563538 containerd[1515]: time="2025-02-13T19:19:19.563512389Z" level=info msg="ImageCreate event name:\"sha256:3045aa4a360d468ed15090f280e94c54bf4678269a6e863a9ebcf5b31534a346\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:19:19.565822 containerd[1515]: time="2025-02-13T19:19:19.565745465Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:fc9ea45f2475fd99db1b36d2ff180a50017b1a5ea0e82a171c6b439b3a620764\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:19:19.566310 containerd[1515]: time="2025-02-13T19:19:19.566282593Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.36.2\" with image id \"sha256:3045aa4a360d468ed15090f280e94c54bf4678269a6e863a9ebcf5b31534a346\", repo tag \"quay.io/tigera/operator:v1.36.2\", repo digest \"quay.io/tigera/operator@sha256:fc9ea45f2475fd99db1b36d2ff180a50017b1a5ea0e82a171c6b439b3a620764\", size \"21758492\" in 1.986241326s" Feb 13 19:19:19.566374 containerd[1515]: time="2025-02-13T19:19:19.566311348Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.36.2\" returns image reference \"sha256:3045aa4a360d468ed15090f280e94c54bf4678269a6e863a9ebcf5b31534a346\"" Feb 13 19:19:19.567899 containerd[1515]: time="2025-02-13T19:19:19.567856609Z" level=info msg="CreateContainer within sandbox \"a0074be09150660c12027fccee5ecec70184b4dc858cf609107c4cfe21e8eb8c\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Feb 13 19:19:19.580216 containerd[1515]: time="2025-02-13T19:19:19.580085723Z" level=info msg="CreateContainer within sandbox \"a0074be09150660c12027fccee5ecec70184b4dc858cf609107c4cfe21e8eb8c\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"8ea22b54b68f56d1d9b83b8a31c9f5a8a914d276d6fdf52d2ba333112dfd85d7\"" Feb 13 19:19:19.580661 containerd[1515]: time="2025-02-13T19:19:19.580623153Z" level=info msg="StartContainer for \"8ea22b54b68f56d1d9b83b8a31c9f5a8a914d276d6fdf52d2ba333112dfd85d7\"" Feb 13 19:19:19.604350 systemd[1]: Started cri-containerd-8ea22b54b68f56d1d9b83b8a31c9f5a8a914d276d6fdf52d2ba333112dfd85d7.scope - libcontainer container 8ea22b54b68f56d1d9b83b8a31c9f5a8a914d276d6fdf52d2ba333112dfd85d7. Feb 13 19:19:19.664869 containerd[1515]: time="2025-02-13T19:19:19.664808716Z" level=info msg="StartContainer for \"8ea22b54b68f56d1d9b83b8a31c9f5a8a914d276d6fdf52d2ba333112dfd85d7\" returns successfully" Feb 13 19:19:19.685444 kubelet[2630]: E0213 19:19:19.685389 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:19.699624 kubelet[2630]: I0213 19:19:19.699580 2630 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-76c4976dd7-mgpk6" podStartSLOduration=0.712204385 podStartE2EDuration="2.699567531s" podCreationTimestamp="2025-02-13 19:19:17 +0000 UTC" firstStartedPulling="2025-02-13 19:19:17.579546828 +0000 UTC m=+6.001399918" lastFinishedPulling="2025-02-13 19:19:19.566909973 +0000 UTC m=+7.988763064" observedRunningTime="2025-02-13 19:19:19.699433467 +0000 UTC m=+8.121286557" watchObservedRunningTime="2025-02-13 19:19:19.699567531 +0000 UTC m=+8.121420621" Feb 13 19:19:22.557283 systemd[1]: Created slice kubepods-besteffort-pod931387ed_771a_4e79_8c00_0ba6bf4617bf.slice - libcontainer container kubepods-besteffort-pod931387ed_771a_4e79_8c00_0ba6bf4617bf.slice. Feb 13 19:19:22.565084 systemd[1]: Created slice kubepods-besteffort-pod34f2a7db_9549_41d9_8605_83d375f71336.slice - libcontainer container kubepods-besteffort-pod34f2a7db_9549_41d9_8605_83d375f71336.slice. Feb 13 19:19:22.610447 kubelet[2630]: E0213 19:19:22.610386 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-drlnv" podUID="84096616-cf79-421c-9e94-8cbcf0a204a5" Feb 13 19:19:22.672109 kubelet[2630]: I0213 19:19:22.672050 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/34f2a7db-9549-41d9-8605-83d375f71336-tigera-ca-bundle\") pod \"calico-node-4n8v4\" (UID: \"34f2a7db-9549-41d9-8605-83d375f71336\") " pod="calico-system/calico-node-4n8v4" Feb 13 19:19:22.672109 kubelet[2630]: I0213 19:19:22.672095 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/34f2a7db-9549-41d9-8605-83d375f71336-lib-modules\") pod \"calico-node-4n8v4\" (UID: \"34f2a7db-9549-41d9-8605-83d375f71336\") " pod="calico-system/calico-node-4n8v4" Feb 13 19:19:22.672109 kubelet[2630]: I0213 19:19:22.672111 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/34f2a7db-9549-41d9-8605-83d375f71336-node-certs\") pod \"calico-node-4n8v4\" (UID: \"34f2a7db-9549-41d9-8605-83d375f71336\") " pod="calico-system/calico-node-4n8v4" Feb 13 19:19:22.672318 kubelet[2630]: I0213 19:19:22.672128 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/34f2a7db-9549-41d9-8605-83d375f71336-var-lib-calico\") pod \"calico-node-4n8v4\" (UID: \"34f2a7db-9549-41d9-8605-83d375f71336\") " pod="calico-system/calico-node-4n8v4" Feb 13 19:19:22.672318 kubelet[2630]: I0213 19:19:22.672145 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/34f2a7db-9549-41d9-8605-83d375f71336-cni-net-dir\") pod \"calico-node-4n8v4\" (UID: \"34f2a7db-9549-41d9-8605-83d375f71336\") " pod="calico-system/calico-node-4n8v4" Feb 13 19:19:22.672318 kubelet[2630]: I0213 19:19:22.672159 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/34f2a7db-9549-41d9-8605-83d375f71336-cni-log-dir\") pod \"calico-node-4n8v4\" (UID: \"34f2a7db-9549-41d9-8605-83d375f71336\") " pod="calico-system/calico-node-4n8v4" Feb 13 19:19:22.672318 kubelet[2630]: I0213 19:19:22.672304 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/34f2a7db-9549-41d9-8605-83d375f71336-var-run-calico\") pod \"calico-node-4n8v4\" (UID: \"34f2a7db-9549-41d9-8605-83d375f71336\") " pod="calico-system/calico-node-4n8v4" Feb 13 19:19:22.672422 kubelet[2630]: I0213 19:19:22.672333 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glp2p\" (UniqueName: \"kubernetes.io/projected/931387ed-771a-4e79-8c00-0ba6bf4617bf-kube-api-access-glp2p\") pod \"calico-typha-65d7d8c4ff-kk8zb\" (UID: \"931387ed-771a-4e79-8c00-0ba6bf4617bf\") " pod="calico-system/calico-typha-65d7d8c4ff-kk8zb" Feb 13 19:19:22.672422 kubelet[2630]: I0213 19:19:22.672388 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/34f2a7db-9549-41d9-8605-83d375f71336-flexvol-driver-host\") pod \"calico-node-4n8v4\" (UID: \"34f2a7db-9549-41d9-8605-83d375f71336\") " pod="calico-system/calico-node-4n8v4" Feb 13 19:19:22.672422 kubelet[2630]: I0213 19:19:22.672410 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/34f2a7db-9549-41d9-8605-83d375f71336-xtables-lock\") pod \"calico-node-4n8v4\" (UID: \"34f2a7db-9549-41d9-8605-83d375f71336\") " pod="calico-system/calico-node-4n8v4" Feb 13 19:19:22.672497 kubelet[2630]: I0213 19:19:22.672432 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chj4q\" (UniqueName: \"kubernetes.io/projected/34f2a7db-9549-41d9-8605-83d375f71336-kube-api-access-chj4q\") pod \"calico-node-4n8v4\" (UID: \"34f2a7db-9549-41d9-8605-83d375f71336\") " pod="calico-system/calico-node-4n8v4" Feb 13 19:19:22.672497 kubelet[2630]: I0213 19:19:22.672453 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/34f2a7db-9549-41d9-8605-83d375f71336-policysync\") pod \"calico-node-4n8v4\" (UID: \"34f2a7db-9549-41d9-8605-83d375f71336\") " pod="calico-system/calico-node-4n8v4" Feb 13 19:19:22.672497 kubelet[2630]: I0213 19:19:22.672473 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/34f2a7db-9549-41d9-8605-83d375f71336-cni-bin-dir\") pod \"calico-node-4n8v4\" (UID: \"34f2a7db-9549-41d9-8605-83d375f71336\") " pod="calico-system/calico-node-4n8v4" Feb 13 19:19:22.672579 kubelet[2630]: I0213 19:19:22.672547 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/931387ed-771a-4e79-8c00-0ba6bf4617bf-tigera-ca-bundle\") pod \"calico-typha-65d7d8c4ff-kk8zb\" (UID: \"931387ed-771a-4e79-8c00-0ba6bf4617bf\") " pod="calico-system/calico-typha-65d7d8c4ff-kk8zb" Feb 13 19:19:22.672579 kubelet[2630]: I0213 19:19:22.672573 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/931387ed-771a-4e79-8c00-0ba6bf4617bf-typha-certs\") pod \"calico-typha-65d7d8c4ff-kk8zb\" (UID: \"931387ed-771a-4e79-8c00-0ba6bf4617bf\") " pod="calico-system/calico-typha-65d7d8c4ff-kk8zb" Feb 13 19:19:22.773487 kubelet[2630]: I0213 19:19:22.773227 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/84096616-cf79-421c-9e94-8cbcf0a204a5-varrun\") pod \"csi-node-driver-drlnv\" (UID: \"84096616-cf79-421c-9e94-8cbcf0a204a5\") " pod="calico-system/csi-node-driver-drlnv" Feb 13 19:19:22.773487 kubelet[2630]: I0213 19:19:22.773264 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/84096616-cf79-421c-9e94-8cbcf0a204a5-registration-dir\") pod \"csi-node-driver-drlnv\" (UID: \"84096616-cf79-421c-9e94-8cbcf0a204a5\") " pod="calico-system/csi-node-driver-drlnv" Feb 13 19:19:22.773487 kubelet[2630]: I0213 19:19:22.773321 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdw8x\" (UniqueName: \"kubernetes.io/projected/84096616-cf79-421c-9e94-8cbcf0a204a5-kube-api-access-vdw8x\") pod \"csi-node-driver-drlnv\" (UID: \"84096616-cf79-421c-9e94-8cbcf0a204a5\") " pod="calico-system/csi-node-driver-drlnv" Feb 13 19:19:22.773487 kubelet[2630]: I0213 19:19:22.773346 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/84096616-cf79-421c-9e94-8cbcf0a204a5-socket-dir\") pod \"csi-node-driver-drlnv\" (UID: \"84096616-cf79-421c-9e94-8cbcf0a204a5\") " pod="calico-system/csi-node-driver-drlnv" Feb 13 19:19:22.773487 kubelet[2630]: I0213 19:19:22.773377 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/84096616-cf79-421c-9e94-8cbcf0a204a5-kubelet-dir\") pod \"csi-node-driver-drlnv\" (UID: \"84096616-cf79-421c-9e94-8cbcf0a204a5\") " pod="calico-system/csi-node-driver-drlnv" Feb 13 19:19:22.775265 kubelet[2630]: E0213 19:19:22.774991 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.775265 kubelet[2630]: W0213 19:19:22.775005 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.775265 kubelet[2630]: E0213 19:19:22.775020 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.775536 kubelet[2630]: E0213 19:19:22.775497 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.775536 kubelet[2630]: W0213 19:19:22.775508 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.775815 kubelet[2630]: E0213 19:19:22.775802 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.775939 kubelet[2630]: W0213 19:19:22.775861 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.775939 kubelet[2630]: E0213 19:19:22.775875 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.776016 kubelet[2630]: E0213 19:19:22.776004 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.777113 kubelet[2630]: E0213 19:19:22.777102 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.777305 kubelet[2630]: W0213 19:19:22.777219 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.777305 kubelet[2630]: E0213 19:19:22.777232 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.777499 kubelet[2630]: E0213 19:19:22.777488 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.777554 kubelet[2630]: W0213 19:19:22.777544 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.777605 kubelet[2630]: E0213 19:19:22.777595 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.777849 kubelet[2630]: E0213 19:19:22.777839 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.777970 kubelet[2630]: W0213 19:19:22.777896 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.777970 kubelet[2630]: E0213 19:19:22.777908 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.778180 kubelet[2630]: E0213 19:19:22.778169 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.778247 kubelet[2630]: W0213 19:19:22.778237 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.778311 kubelet[2630]: E0213 19:19:22.778290 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.779125 kubelet[2630]: E0213 19:19:22.778927 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.779125 kubelet[2630]: W0213 19:19:22.778957 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.779125 kubelet[2630]: E0213 19:19:22.778982 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.781521 kubelet[2630]: E0213 19:19:22.781474 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.781521 kubelet[2630]: W0213 19:19:22.781512 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.781624 kubelet[2630]: E0213 19:19:22.781534 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.784994 kubelet[2630]: E0213 19:19:22.784967 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.784994 kubelet[2630]: W0213 19:19:22.784988 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.784994 kubelet[2630]: E0213 19:19:22.785011 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.785327 kubelet[2630]: E0213 19:19:22.785308 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.785327 kubelet[2630]: W0213 19:19:22.785325 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.785398 kubelet[2630]: E0213 19:19:22.785337 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.863493 kubelet[2630]: E0213 19:19:22.863376 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:22.866259 containerd[1515]: time="2025-02-13T19:19:22.865888308Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-65d7d8c4ff-kk8zb,Uid:931387ed-771a-4e79-8c00-0ba6bf4617bf,Namespace:calico-system,Attempt:0,}" Feb 13 19:19:22.868305 kubelet[2630]: E0213 19:19:22.868245 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:22.868751 containerd[1515]: time="2025-02-13T19:19:22.868693988Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-4n8v4,Uid:34f2a7db-9549-41d9-8605-83d375f71336,Namespace:calico-system,Attempt:0,}" Feb 13 19:19:22.873934 kubelet[2630]: E0213 19:19:22.873901 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.873934 kubelet[2630]: W0213 19:19:22.873921 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.874020 kubelet[2630]: E0213 19:19:22.873941 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.874268 kubelet[2630]: E0213 19:19:22.874234 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.874268 kubelet[2630]: W0213 19:19:22.874258 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.874334 kubelet[2630]: E0213 19:19:22.874288 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.874598 kubelet[2630]: E0213 19:19:22.874584 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.874631 kubelet[2630]: W0213 19:19:22.874599 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.874631 kubelet[2630]: E0213 19:19:22.874619 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.874867 kubelet[2630]: E0213 19:19:22.874849 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.874867 kubelet[2630]: W0213 19:19:22.874861 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.874960 kubelet[2630]: E0213 19:19:22.874878 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.875055 kubelet[2630]: E0213 19:19:22.875040 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.875055 kubelet[2630]: W0213 19:19:22.875049 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.875132 kubelet[2630]: E0213 19:19:22.875062 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.875277 kubelet[2630]: E0213 19:19:22.875261 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.875277 kubelet[2630]: W0213 19:19:22.875271 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.875368 kubelet[2630]: E0213 19:19:22.875285 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.875506 kubelet[2630]: E0213 19:19:22.875496 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.875506 kubelet[2630]: W0213 19:19:22.875504 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.875553 kubelet[2630]: E0213 19:19:22.875517 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.875781 kubelet[2630]: E0213 19:19:22.875771 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.875781 kubelet[2630]: W0213 19:19:22.875780 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.875845 kubelet[2630]: E0213 19:19:22.875825 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.876085 kubelet[2630]: E0213 19:19:22.876061 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.876085 kubelet[2630]: W0213 19:19:22.876072 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.876303 kubelet[2630]: E0213 19:19:22.876102 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.876303 kubelet[2630]: E0213 19:19:22.876267 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.876303 kubelet[2630]: W0213 19:19:22.876274 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.876303 kubelet[2630]: E0213 19:19:22.876295 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.876494 kubelet[2630]: E0213 19:19:22.876479 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.876494 kubelet[2630]: W0213 19:19:22.876489 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.876576 kubelet[2630]: E0213 19:19:22.876501 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.876722 kubelet[2630]: E0213 19:19:22.876703 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.876722 kubelet[2630]: W0213 19:19:22.876715 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.876821 kubelet[2630]: E0213 19:19:22.876730 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.876953 kubelet[2630]: E0213 19:19:22.876936 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.876953 kubelet[2630]: W0213 19:19:22.876947 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.877037 kubelet[2630]: E0213 19:19:22.876970 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.877253 kubelet[2630]: E0213 19:19:22.877131 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.877253 kubelet[2630]: W0213 19:19:22.877143 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.877253 kubelet[2630]: E0213 19:19:22.877181 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.877408 kubelet[2630]: E0213 19:19:22.877371 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.877408 kubelet[2630]: W0213 19:19:22.877388 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.877611 kubelet[2630]: E0213 19:19:22.877414 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.877611 kubelet[2630]: E0213 19:19:22.877601 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.877611 kubelet[2630]: W0213 19:19:22.877608 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.877706 kubelet[2630]: E0213 19:19:22.877637 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.877850 kubelet[2630]: E0213 19:19:22.877834 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.877850 kubelet[2630]: W0213 19:19:22.877845 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.878030 kubelet[2630]: E0213 19:19:22.877858 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.878066 kubelet[2630]: E0213 19:19:22.878044 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.878066 kubelet[2630]: W0213 19:19:22.878051 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.878066 kubelet[2630]: E0213 19:19:22.878062 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.878265 kubelet[2630]: E0213 19:19:22.878246 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.878265 kubelet[2630]: W0213 19:19:22.878259 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.878370 kubelet[2630]: E0213 19:19:22.878283 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.878588 kubelet[2630]: E0213 19:19:22.878559 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.878588 kubelet[2630]: W0213 19:19:22.878572 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.878588 kubelet[2630]: E0213 19:19:22.878586 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.878953 kubelet[2630]: E0213 19:19:22.878914 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.878953 kubelet[2630]: W0213 19:19:22.878950 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.879048 kubelet[2630]: E0213 19:19:22.878970 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.879241 kubelet[2630]: E0213 19:19:22.879226 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.879241 kubelet[2630]: W0213 19:19:22.879238 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.879298 kubelet[2630]: E0213 19:19:22.879247 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.879577 kubelet[2630]: E0213 19:19:22.879559 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.879617 kubelet[2630]: W0213 19:19:22.879577 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.879617 kubelet[2630]: E0213 19:19:22.879598 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.879902 kubelet[2630]: E0213 19:19:22.879885 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.879902 kubelet[2630]: W0213 19:19:22.879900 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.879960 kubelet[2630]: E0213 19:19:22.879919 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.880212 kubelet[2630]: E0213 19:19:22.880177 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.880212 kubelet[2630]: W0213 19:19:22.880210 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.880278 kubelet[2630]: E0213 19:19:22.880222 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.884891 kubelet[2630]: E0213 19:19:22.884870 2630 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Feb 13 19:19:22.884891 kubelet[2630]: W0213 19:19:22.884882 2630 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Feb 13 19:19:22.884891 kubelet[2630]: E0213 19:19:22.884892 2630 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Feb 13 19:19:22.960628 containerd[1515]: time="2025-02-13T19:19:22.960527522Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:19:22.960628 containerd[1515]: time="2025-02-13T19:19:22.960583668Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:19:22.960628 containerd[1515]: time="2025-02-13T19:19:22.960594039Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:19:22.960779 containerd[1515]: time="2025-02-13T19:19:22.960670012Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:19:22.969874 containerd[1515]: time="2025-02-13T19:19:22.969534151Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:19:22.969874 containerd[1515]: time="2025-02-13T19:19:22.969586400Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:19:22.969874 containerd[1515]: time="2025-02-13T19:19:22.969599475Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:19:22.969874 containerd[1515]: time="2025-02-13T19:19:22.969750150Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:19:22.981360 systemd[1]: Started cri-containerd-fbab82d383b2625dee20792f00483d9ca41f94ed2a69567af3212ec83f8165aa.scope - libcontainer container fbab82d383b2625dee20792f00483d9ca41f94ed2a69567af3212ec83f8165aa. Feb 13 19:19:22.985293 systemd[1]: Started cri-containerd-de4c4714d324a399cc66586668f0f0dccf310109ef8a4b507b491a5bbf423afa.scope - libcontainer container de4c4714d324a399cc66586668f0f0dccf310109ef8a4b507b491a5bbf423afa. Feb 13 19:19:23.008647 containerd[1515]: time="2025-02-13T19:19:23.008603197Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-4n8v4,Uid:34f2a7db-9549-41d9-8605-83d375f71336,Namespace:calico-system,Attempt:0,} returns sandbox id \"de4c4714d324a399cc66586668f0f0dccf310109ef8a4b507b491a5bbf423afa\"" Feb 13 19:19:23.009304 kubelet[2630]: E0213 19:19:23.009265 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:23.010385 containerd[1515]: time="2025-02-13T19:19:23.010359881Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\"" Feb 13 19:19:23.026570 containerd[1515]: time="2025-02-13T19:19:23.025632764Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-65d7d8c4ff-kk8zb,Uid:931387ed-771a-4e79-8c00-0ba6bf4617bf,Namespace:calico-system,Attempt:0,} returns sandbox id \"fbab82d383b2625dee20792f00483d9ca41f94ed2a69567af3212ec83f8165aa\"" Feb 13 19:19:23.027937 kubelet[2630]: E0213 19:19:23.027913 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:24.642451 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount234710855.mount: Deactivated successfully. Feb 13 19:19:24.658112 kubelet[2630]: E0213 19:19:24.658073 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-drlnv" podUID="84096616-cf79-421c-9e94-8cbcf0a204a5" Feb 13 19:19:24.803794 containerd[1515]: time="2025-02-13T19:19:24.803743053Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:19:24.804448 containerd[1515]: time="2025-02-13T19:19:24.804415926Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1: active requests=0, bytes read=6855343" Feb 13 19:19:24.805481 containerd[1515]: time="2025-02-13T19:19:24.805430173Z" level=info msg="ImageCreate event name:\"sha256:2b7452b763ec8833ca0386ada5fd066e552a9b3b02b8538a5e34cc3d6d3840a6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:19:24.807579 containerd[1515]: time="2025-02-13T19:19:24.807546235Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:a63f8b4ff531912d12d143664eb263fdbc6cd7b3ff4aa777dfb6e318a090462c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:19:24.808071 containerd[1515]: time="2025-02-13T19:19:24.808040580Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\" with image id \"sha256:2b7452b763ec8833ca0386ada5fd066e552a9b3b02b8538a5e34cc3d6d3840a6\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:a63f8b4ff531912d12d143664eb263fdbc6cd7b3ff4aa777dfb6e318a090462c\", size \"6855165\" in 1.797656182s" Feb 13 19:19:24.808071 containerd[1515]: time="2025-02-13T19:19:24.808065977Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\" returns image reference \"sha256:2b7452b763ec8833ca0386ada5fd066e552a9b3b02b8538a5e34cc3d6d3840a6\"" Feb 13 19:19:24.808799 containerd[1515]: time="2025-02-13T19:19:24.808737417Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.29.1\"" Feb 13 19:19:24.809549 containerd[1515]: time="2025-02-13T19:19:24.809527551Z" level=info msg="CreateContainer within sandbox \"de4c4714d324a399cc66586668f0f0dccf310109ef8a4b507b491a5bbf423afa\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Feb 13 19:19:24.827018 containerd[1515]: time="2025-02-13T19:19:24.826982826Z" level=info msg="CreateContainer within sandbox \"de4c4714d324a399cc66586668f0f0dccf310109ef8a4b507b491a5bbf423afa\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"a0dd21b49e34ef1bfc9a9efb0e48383bc26a70984c471217b6d5d37657875e76\"" Feb 13 19:19:24.827454 containerd[1515]: time="2025-02-13T19:19:24.827404352Z" level=info msg="StartContainer for \"a0dd21b49e34ef1bfc9a9efb0e48383bc26a70984c471217b6d5d37657875e76\"" Feb 13 19:19:24.854326 systemd[1]: Started cri-containerd-a0dd21b49e34ef1bfc9a9efb0e48383bc26a70984c471217b6d5d37657875e76.scope - libcontainer container a0dd21b49e34ef1bfc9a9efb0e48383bc26a70984c471217b6d5d37657875e76. Feb 13 19:19:24.881855 containerd[1515]: time="2025-02-13T19:19:24.881814341Z" level=info msg="StartContainer for \"a0dd21b49e34ef1bfc9a9efb0e48383bc26a70984c471217b6d5d37657875e76\" returns successfully" Feb 13 19:19:24.900759 systemd[1]: cri-containerd-a0dd21b49e34ef1bfc9a9efb0e48383bc26a70984c471217b6d5d37657875e76.scope: Deactivated successfully. Feb 13 19:19:24.901359 systemd[1]: cri-containerd-a0dd21b49e34ef1bfc9a9efb0e48383bc26a70984c471217b6d5d37657875e76.scope: Consumed 32ms CPU time, 8.1M memory peak, 6.3M written to disk. Feb 13 19:19:24.922636 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a0dd21b49e34ef1bfc9a9efb0e48383bc26a70984c471217b6d5d37657875e76-rootfs.mount: Deactivated successfully. Feb 13 19:19:24.949229 containerd[1515]: time="2025-02-13T19:19:24.949149818Z" level=info msg="shim disconnected" id=a0dd21b49e34ef1bfc9a9efb0e48383bc26a70984c471217b6d5d37657875e76 namespace=k8s.io Feb 13 19:19:24.949229 containerd[1515]: time="2025-02-13T19:19:24.949224911Z" level=warning msg="cleaning up after shim disconnected" id=a0dd21b49e34ef1bfc9a9efb0e48383bc26a70984c471217b6d5d37657875e76 namespace=k8s.io Feb 13 19:19:24.949229 containerd[1515]: time="2025-02-13T19:19:24.949233687Z" level=info msg="cleaning up dead shim" namespace=k8s.io Feb 13 19:19:25.696309 kubelet[2630]: E0213 19:19:25.696270 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:26.658020 kubelet[2630]: E0213 19:19:26.657969 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-drlnv" podUID="84096616-cf79-421c-9e94-8cbcf0a204a5" Feb 13 19:19:27.040581 containerd[1515]: time="2025-02-13T19:19:27.040534751Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:19:27.042865 containerd[1515]: time="2025-02-13T19:19:27.042774649Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.29.1: active requests=0, bytes read=29850141" Feb 13 19:19:27.043885 containerd[1515]: time="2025-02-13T19:19:27.043858216Z" level=info msg="ImageCreate event name:\"sha256:4cb3738506f5a9c530033d1e24fd6b9ec618518a2ec8b012ded33572be06ab44\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:19:27.046543 containerd[1515]: time="2025-02-13T19:19:27.046227409Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:768a194e1115c73bcbf35edb7afd18a63e16e08d940c79993565b6a3cca2da7c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:19:27.046773 containerd[1515]: time="2025-02-13T19:19:27.046734266Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.29.1\" with image id \"sha256:4cb3738506f5a9c530033d1e24fd6b9ec618518a2ec8b012ded33572be06ab44\", repo tag \"ghcr.io/flatcar/calico/typha:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:768a194e1115c73bcbf35edb7afd18a63e16e08d940c79993565b6a3cca2da7c\", size \"31343217\" in 2.237975208s" Feb 13 19:19:27.046845 containerd[1515]: time="2025-02-13T19:19:27.046772327Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.29.1\" returns image reference \"sha256:4cb3738506f5a9c530033d1e24fd6b9ec618518a2ec8b012ded33572be06ab44\"" Feb 13 19:19:27.048058 containerd[1515]: time="2025-02-13T19:19:27.048018540Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.1\"" Feb 13 19:19:27.058386 containerd[1515]: time="2025-02-13T19:19:27.058316153Z" level=info msg="CreateContainer within sandbox \"fbab82d383b2625dee20792f00483d9ca41f94ed2a69567af3212ec83f8165aa\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Feb 13 19:19:27.072849 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3813244597.mount: Deactivated successfully. Feb 13 19:19:27.077921 containerd[1515]: time="2025-02-13T19:19:27.077873610Z" level=info msg="CreateContainer within sandbox \"fbab82d383b2625dee20792f00483d9ca41f94ed2a69567af3212ec83f8165aa\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"718310b51f0a91c729a9ac72b4ad08c9a9a8cc1677b946cffe03606d069cdabd\"" Feb 13 19:19:27.078778 containerd[1515]: time="2025-02-13T19:19:27.078682346Z" level=info msg="StartContainer for \"718310b51f0a91c729a9ac72b4ad08c9a9a8cc1677b946cffe03606d069cdabd\"" Feb 13 19:19:27.111414 systemd[1]: Started cri-containerd-718310b51f0a91c729a9ac72b4ad08c9a9a8cc1677b946cffe03606d069cdabd.scope - libcontainer container 718310b51f0a91c729a9ac72b4ad08c9a9a8cc1677b946cffe03606d069cdabd. Feb 13 19:19:27.159504 containerd[1515]: time="2025-02-13T19:19:27.159462455Z" level=info msg="StartContainer for \"718310b51f0a91c729a9ac72b4ad08c9a9a8cc1677b946cffe03606d069cdabd\" returns successfully" Feb 13 19:19:27.700806 kubelet[2630]: E0213 19:19:27.700772 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:28.658233 kubelet[2630]: E0213 19:19:28.658152 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-drlnv" podUID="84096616-cf79-421c-9e94-8cbcf0a204a5" Feb 13 19:19:28.702880 kubelet[2630]: I0213 19:19:28.702844 2630 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 13 19:19:28.704453 kubelet[2630]: E0213 19:19:28.704392 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:30.657303 kubelet[2630]: E0213 19:19:30.657252 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-drlnv" podUID="84096616-cf79-421c-9e94-8cbcf0a204a5" Feb 13 19:19:31.902923 containerd[1515]: time="2025-02-13T19:19:31.902869056Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:19:31.904059 containerd[1515]: time="2025-02-13T19:19:31.903805411Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.29.1: active requests=0, bytes read=96154154" Feb 13 19:19:31.905031 containerd[1515]: time="2025-02-13T19:19:31.905006786Z" level=info msg="ImageCreate event name:\"sha256:7dd6ea186aba0d7a1791a79d426fe854527ca95192b26bbd19e8baf8373f7d0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:19:31.907269 containerd[1515]: time="2025-02-13T19:19:31.907231480Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:21e759d51c90dfb34fc1397dc180dd3a3fb564c2b0580d2f61ffe108f2a3c94b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:19:31.908098 containerd[1515]: time="2025-02-13T19:19:31.908034534Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.29.1\" with image id \"sha256:7dd6ea186aba0d7a1791a79d426fe854527ca95192b26bbd19e8baf8373f7d0e\", repo tag \"ghcr.io/flatcar/calico/cni:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:21e759d51c90dfb34fc1397dc180dd3a3fb564c2b0580d2f61ffe108f2a3c94b\", size \"97647238\" in 4.859984334s" Feb 13 19:19:31.908098 containerd[1515]: time="2025-02-13T19:19:31.908065001Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.1\" returns image reference \"sha256:7dd6ea186aba0d7a1791a79d426fe854527ca95192b26bbd19e8baf8373f7d0e\"" Feb 13 19:19:31.910677 containerd[1515]: time="2025-02-13T19:19:31.910631460Z" level=info msg="CreateContainer within sandbox \"de4c4714d324a399cc66586668f0f0dccf310109ef8a4b507b491a5bbf423afa\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Feb 13 19:19:31.927806 containerd[1515]: time="2025-02-13T19:19:31.927745472Z" level=info msg="CreateContainer within sandbox \"de4c4714d324a399cc66586668f0f0dccf310109ef8a4b507b491a5bbf423afa\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"57e48a0007d012bb3be6a4b8025af7f2b12964e6e40e2c4bc13a7c7075366299\"" Feb 13 19:19:31.928464 containerd[1515]: time="2025-02-13T19:19:31.928401649Z" level=info msg="StartContainer for \"57e48a0007d012bb3be6a4b8025af7f2b12964e6e40e2c4bc13a7c7075366299\"" Feb 13 19:19:31.963427 systemd[1]: Started cri-containerd-57e48a0007d012bb3be6a4b8025af7f2b12964e6e40e2c4bc13a7c7075366299.scope - libcontainer container 57e48a0007d012bb3be6a4b8025af7f2b12964e6e40e2c4bc13a7c7075366299. Feb 13 19:19:32.003009 containerd[1515]: time="2025-02-13T19:19:32.002963351Z" level=info msg="StartContainer for \"57e48a0007d012bb3be6a4b8025af7f2b12964e6e40e2c4bc13a7c7075366299\" returns successfully" Feb 13 19:19:33.072541 kubelet[2630]: E0213 19:19:33.072490 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-drlnv" podUID="84096616-cf79-421c-9e94-8cbcf0a204a5" Feb 13 19:19:33.073941 kubelet[2630]: E0213 19:19:33.073922 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:33.144924 kubelet[2630]: I0213 19:19:33.144843 2630 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-65d7d8c4ff-kk8zb" podStartSLOduration=7.125698163 podStartE2EDuration="11.144824996s" podCreationTimestamp="2025-02-13 19:19:22 +0000 UTC" firstStartedPulling="2025-02-13 19:19:23.028460975 +0000 UTC m=+11.450314065" lastFinishedPulling="2025-02-13 19:19:27.047587808 +0000 UTC m=+15.469440898" observedRunningTime="2025-02-13 19:19:27.74316957 +0000 UTC m=+16.165022660" watchObservedRunningTime="2025-02-13 19:19:33.144824996 +0000 UTC m=+21.566678086" Feb 13 19:19:33.368152 systemd[1]: cri-containerd-57e48a0007d012bb3be6a4b8025af7f2b12964e6e40e2c4bc13a7c7075366299.scope: Deactivated successfully. Feb 13 19:19:33.368625 systemd[1]: cri-containerd-57e48a0007d012bb3be6a4b8025af7f2b12964e6e40e2c4bc13a7c7075366299.scope: Consumed 531ms CPU time, 157.9M memory peak, 8K read from disk, 151M written to disk. Feb 13 19:19:33.373795 kubelet[2630]: I0213 19:19:33.373771 2630 kubelet_node_status.go:488] "Fast updating node status as it just became ready" Feb 13 19:19:33.395390 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-57e48a0007d012bb3be6a4b8025af7f2b12964e6e40e2c4bc13a7c7075366299-rootfs.mount: Deactivated successfully. Feb 13 19:19:33.402248 containerd[1515]: time="2025-02-13T19:19:33.402163499Z" level=info msg="shim disconnected" id=57e48a0007d012bb3be6a4b8025af7f2b12964e6e40e2c4bc13a7c7075366299 namespace=k8s.io Feb 13 19:19:33.402248 containerd[1515]: time="2025-02-13T19:19:33.402244933Z" level=warning msg="cleaning up after shim disconnected" id=57e48a0007d012bb3be6a4b8025af7f2b12964e6e40e2c4bc13a7c7075366299 namespace=k8s.io Feb 13 19:19:33.402722 containerd[1515]: time="2025-02-13T19:19:33.402257126Z" level=info msg="cleaning up dead shim" namespace=k8s.io Feb 13 19:19:33.408490 systemd[1]: Created slice kubepods-burstable-podf48c8839_9a17_43c4_a0d5_f9e8d9767814.slice - libcontainer container kubepods-burstable-podf48c8839_9a17_43c4_a0d5_f9e8d9767814.slice. Feb 13 19:19:33.417724 systemd[1]: Created slice kubepods-burstable-podf2fdfd8a_780b_4320_8f24_ca003c5d0c77.slice - libcontainer container kubepods-burstable-podf2fdfd8a_780b_4320_8f24_ca003c5d0c77.slice. Feb 13 19:19:33.419084 containerd[1515]: time="2025-02-13T19:19:33.418604685Z" level=warning msg="cleanup warnings time=\"2025-02-13T19:19:33Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Feb 13 19:19:33.429641 systemd[1]: Created slice kubepods-besteffort-pod4e40d638_ead9_4cd4_859c_b3835db75e0c.slice - libcontainer container kubepods-besteffort-pod4e40d638_ead9_4cd4_859c_b3835db75e0c.slice. Feb 13 19:19:33.436034 systemd[1]: Created slice kubepods-besteffort-pod5a39efb7_aca1_4893_a36f_840c3c504ece.slice - libcontainer container kubepods-besteffort-pod5a39efb7_aca1_4893_a36f_840c3c504ece.slice. Feb 13 19:19:33.442727 systemd[1]: Created slice kubepods-besteffort-podf6e023ee_9ef0_491c_bd3c_e70ebf96e0c1.slice - libcontainer container kubepods-besteffort-podf6e023ee_9ef0_491c_bd3c_e70ebf96e0c1.slice. Feb 13 19:19:33.550009 kubelet[2630]: I0213 19:19:33.549948 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hf695\" (UniqueName: \"kubernetes.io/projected/f48c8839-9a17-43c4-a0d5-f9e8d9767814-kube-api-access-hf695\") pod \"coredns-6f6b679f8f-frtk4\" (UID: \"f48c8839-9a17-43c4-a0d5-f9e8d9767814\") " pod="kube-system/coredns-6f6b679f8f-frtk4" Feb 13 19:19:33.550009 kubelet[2630]: I0213 19:19:33.549994 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrjrc\" (UniqueName: \"kubernetes.io/projected/f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1-kube-api-access-hrjrc\") pod \"calico-apiserver-86475b859c-kwlfl\" (UID: \"f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1\") " pod="calico-apiserver/calico-apiserver-86475b859c-kwlfl" Feb 13 19:19:33.550009 kubelet[2630]: I0213 19:19:33.550017 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f48c8839-9a17-43c4-a0d5-f9e8d9767814-config-volume\") pod \"coredns-6f6b679f8f-frtk4\" (UID: \"f48c8839-9a17-43c4-a0d5-f9e8d9767814\") " pod="kube-system/coredns-6f6b679f8f-frtk4" Feb 13 19:19:33.550365 kubelet[2630]: I0213 19:19:33.550039 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1-calico-apiserver-certs\") pod \"calico-apiserver-86475b859c-kwlfl\" (UID: \"f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1\") " pod="calico-apiserver/calico-apiserver-86475b859c-kwlfl" Feb 13 19:19:33.550365 kubelet[2630]: I0213 19:19:33.550107 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2fdfd8a-780b-4320-8f24-ca003c5d0c77-config-volume\") pod \"coredns-6f6b679f8f-s8264\" (UID: \"f2fdfd8a-780b-4320-8f24-ca003c5d0c77\") " pod="kube-system/coredns-6f6b679f8f-s8264" Feb 13 19:19:33.550365 kubelet[2630]: I0213 19:19:33.550131 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhpvb\" (UniqueName: \"kubernetes.io/projected/f2fdfd8a-780b-4320-8f24-ca003c5d0c77-kube-api-access-lhpvb\") pod \"coredns-6f6b679f8f-s8264\" (UID: \"f2fdfd8a-780b-4320-8f24-ca003c5d0c77\") " pod="kube-system/coredns-6f6b679f8f-s8264" Feb 13 19:19:33.550365 kubelet[2630]: I0213 19:19:33.550148 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9fdl\" (UniqueName: \"kubernetes.io/projected/5a39efb7-aca1-4893-a36f-840c3c504ece-kube-api-access-t9fdl\") pod \"calico-apiserver-86475b859c-swqrn\" (UID: \"5a39efb7-aca1-4893-a36f-840c3c504ece\") " pod="calico-apiserver/calico-apiserver-86475b859c-swqrn" Feb 13 19:19:33.550365 kubelet[2630]: I0213 19:19:33.550223 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4e40d638-ead9-4cd4-859c-b3835db75e0c-tigera-ca-bundle\") pod \"calico-kube-controllers-664f6f5454-mnbf7\" (UID: \"4e40d638-ead9-4cd4-859c-b3835db75e0c\") " pod="calico-system/calico-kube-controllers-664f6f5454-mnbf7" Feb 13 19:19:33.550490 kubelet[2630]: I0213 19:19:33.550251 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/5a39efb7-aca1-4893-a36f-840c3c504ece-calico-apiserver-certs\") pod \"calico-apiserver-86475b859c-swqrn\" (UID: \"5a39efb7-aca1-4893-a36f-840c3c504ece\") " pod="calico-apiserver/calico-apiserver-86475b859c-swqrn" Feb 13 19:19:33.550490 kubelet[2630]: I0213 19:19:33.550272 2630 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75k6f\" (UniqueName: \"kubernetes.io/projected/4e40d638-ead9-4cd4-859c-b3835db75e0c-kube-api-access-75k6f\") pod \"calico-kube-controllers-664f6f5454-mnbf7\" (UID: \"4e40d638-ead9-4cd4-859c-b3835db75e0c\") " pod="calico-system/calico-kube-controllers-664f6f5454-mnbf7" Feb 13 19:19:33.714747 kubelet[2630]: E0213 19:19:33.714669 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:33.715555 containerd[1515]: time="2025-02-13T19:19:33.715510383Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-frtk4,Uid:f48c8839-9a17-43c4-a0d5-f9e8d9767814,Namespace:kube-system,Attempt:0,}" Feb 13 19:19:33.723614 kubelet[2630]: E0213 19:19:33.723575 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:33.724035 containerd[1515]: time="2025-02-13T19:19:33.723981703Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-s8264,Uid:f2fdfd8a-780b-4320-8f24-ca003c5d0c77,Namespace:kube-system,Attempt:0,}" Feb 13 19:19:33.733081 containerd[1515]: time="2025-02-13T19:19:33.732983342Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-664f6f5454-mnbf7,Uid:4e40d638-ead9-4cd4-859c-b3835db75e0c,Namespace:calico-system,Attempt:0,}" Feb 13 19:19:33.741880 containerd[1515]: time="2025-02-13T19:19:33.741827325Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86475b859c-swqrn,Uid:5a39efb7-aca1-4893-a36f-840c3c504ece,Namespace:calico-apiserver,Attempt:0,}" Feb 13 19:19:33.746522 containerd[1515]: time="2025-02-13T19:19:33.746487444Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86475b859c-kwlfl,Uid:f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1,Namespace:calico-apiserver,Attempt:0,}" Feb 13 19:19:33.811302 containerd[1515]: time="2025-02-13T19:19:33.811253153Z" level=error msg="Failed to destroy network for sandbox \"f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:33.811481 containerd[1515]: time="2025-02-13T19:19:33.811308977Z" level=error msg="Failed to destroy network for sandbox \"e2995529ca797998e1cf161d883bb684a3464a8a8d57602758f8705313f0d601\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:33.811669 containerd[1515]: time="2025-02-13T19:19:33.811643858Z" level=error msg="encountered an error cleaning up failed sandbox \"f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:33.811724 containerd[1515]: time="2025-02-13T19:19:33.811704903Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-frtk4,Uid:f48c8839-9a17-43c4-a0d5-f9e8d9767814,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:33.812048 kubelet[2630]: E0213 19:19:33.811894 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:33.812048 kubelet[2630]: E0213 19:19:33.811956 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-frtk4" Feb 13 19:19:33.812048 kubelet[2630]: E0213 19:19:33.811975 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-frtk4" Feb 13 19:19:33.812262 kubelet[2630]: E0213 19:19:33.812016 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-6f6b679f8f-frtk4_kube-system(f48c8839-9a17-43c4-a0d5-f9e8d9767814)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-6f6b679f8f-frtk4_kube-system(f48c8839-9a17-43c4-a0d5-f9e8d9767814)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-6f6b679f8f-frtk4" podUID="f48c8839-9a17-43c4-a0d5-f9e8d9767814" Feb 13 19:19:33.812941 containerd[1515]: time="2025-02-13T19:19:33.812916155Z" level=error msg="encountered an error cleaning up failed sandbox \"e2995529ca797998e1cf161d883bb684a3464a8a8d57602758f8705313f0d601\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:33.813114 containerd[1515]: time="2025-02-13T19:19:33.813027265Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-s8264,Uid:f2fdfd8a-780b-4320-8f24-ca003c5d0c77,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"e2995529ca797998e1cf161d883bb684a3464a8a8d57602758f8705313f0d601\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:33.813313 kubelet[2630]: E0213 19:19:33.813148 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e2995529ca797998e1cf161d883bb684a3464a8a8d57602758f8705313f0d601\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:33.813313 kubelet[2630]: E0213 19:19:33.813202 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e2995529ca797998e1cf161d883bb684a3464a8a8d57602758f8705313f0d601\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-s8264" Feb 13 19:19:33.813313 kubelet[2630]: E0213 19:19:33.813220 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e2995529ca797998e1cf161d883bb684a3464a8a8d57602758f8705313f0d601\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-s8264" Feb 13 19:19:33.813395 kubelet[2630]: E0213 19:19:33.813244 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-6f6b679f8f-s8264_kube-system(f2fdfd8a-780b-4320-8f24-ca003c5d0c77)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-6f6b679f8f-s8264_kube-system(f2fdfd8a-780b-4320-8f24-ca003c5d0c77)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"e2995529ca797998e1cf161d883bb684a3464a8a8d57602758f8705313f0d601\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-6f6b679f8f-s8264" podUID="f2fdfd8a-780b-4320-8f24-ca003c5d0c77" Feb 13 19:19:33.838925 containerd[1515]: time="2025-02-13T19:19:33.838703549Z" level=error msg="Failed to destroy network for sandbox \"f7214e7c2d9f81df76f610a79140dcb0085636bf81f33d1eb139eb52cb12a6af\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:33.839857 containerd[1515]: time="2025-02-13T19:19:33.839787570Z" level=error msg="encountered an error cleaning up failed sandbox \"f7214e7c2d9f81df76f610a79140dcb0085636bf81f33d1eb139eb52cb12a6af\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:33.839988 containerd[1515]: time="2025-02-13T19:19:33.839844288Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86475b859c-kwlfl,Uid:f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"f7214e7c2d9f81df76f610a79140dcb0085636bf81f33d1eb139eb52cb12a6af\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:33.840245 kubelet[2630]: E0213 19:19:33.840206 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f7214e7c2d9f81df76f610a79140dcb0085636bf81f33d1eb139eb52cb12a6af\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:33.840301 kubelet[2630]: E0213 19:19:33.840266 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f7214e7c2d9f81df76f610a79140dcb0085636bf81f33d1eb139eb52cb12a6af\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86475b859c-kwlfl" Feb 13 19:19:33.840301 kubelet[2630]: E0213 19:19:33.840284 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f7214e7c2d9f81df76f610a79140dcb0085636bf81f33d1eb139eb52cb12a6af\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86475b859c-kwlfl" Feb 13 19:19:33.840350 kubelet[2630]: E0213 19:19:33.840321 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-86475b859c-kwlfl_calico-apiserver(f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-86475b859c-kwlfl_calico-apiserver(f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"f7214e7c2d9f81df76f610a79140dcb0085636bf81f33d1eb139eb52cb12a6af\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-86475b859c-kwlfl" podUID="f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1" Feb 13 19:19:33.841876 containerd[1515]: time="2025-02-13T19:19:33.841835780Z" level=error msg="Failed to destroy network for sandbox \"aa373a8af75a289d061969e5a22e647e175a6a451ba7c421e5c44fa363ec90f8\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:33.842154 containerd[1515]: time="2025-02-13T19:19:33.842135014Z" level=error msg="encountered an error cleaning up failed sandbox \"aa373a8af75a289d061969e5a22e647e175a6a451ba7c421e5c44fa363ec90f8\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:33.842199 containerd[1515]: time="2025-02-13T19:19:33.842170691Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-664f6f5454-mnbf7,Uid:4e40d638-ead9-4cd4-859c-b3835db75e0c,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"aa373a8af75a289d061969e5a22e647e175a6a451ba7c421e5c44fa363ec90f8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:33.842476 kubelet[2630]: E0213 19:19:33.842299 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"aa373a8af75a289d061969e5a22e647e175a6a451ba7c421e5c44fa363ec90f8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:33.842476 kubelet[2630]: E0213 19:19:33.842331 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"aa373a8af75a289d061969e5a22e647e175a6a451ba7c421e5c44fa363ec90f8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-664f6f5454-mnbf7" Feb 13 19:19:33.842476 kubelet[2630]: E0213 19:19:33.842350 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"aa373a8af75a289d061969e5a22e647e175a6a451ba7c421e5c44fa363ec90f8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-664f6f5454-mnbf7" Feb 13 19:19:33.842566 kubelet[2630]: E0213 19:19:33.842382 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-664f6f5454-mnbf7_calico-system(4e40d638-ead9-4cd4-859c-b3835db75e0c)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-664f6f5454-mnbf7_calico-system(4e40d638-ead9-4cd4-859c-b3835db75e0c)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"aa373a8af75a289d061969e5a22e647e175a6a451ba7c421e5c44fa363ec90f8\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-664f6f5454-mnbf7" podUID="4e40d638-ead9-4cd4-859c-b3835db75e0c" Feb 13 19:19:33.864811 containerd[1515]: time="2025-02-13T19:19:33.864736274Z" level=error msg="Failed to destroy network for sandbox \"d195bcf8f9551fdb26f6b42c8b0e4b63d6953f1bbdc1c4b497262fc12be34248\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:33.865153 containerd[1515]: time="2025-02-13T19:19:33.865125017Z" level=error msg="encountered an error cleaning up failed sandbox \"d195bcf8f9551fdb26f6b42c8b0e4b63d6953f1bbdc1c4b497262fc12be34248\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:33.865226 containerd[1515]: time="2025-02-13T19:19:33.865178447Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86475b859c-swqrn,Uid:5a39efb7-aca1-4893-a36f-840c3c504ece,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"d195bcf8f9551fdb26f6b42c8b0e4b63d6953f1bbdc1c4b497262fc12be34248\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:33.865415 kubelet[2630]: E0213 19:19:33.865387 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d195bcf8f9551fdb26f6b42c8b0e4b63d6953f1bbdc1c4b497262fc12be34248\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:33.865484 kubelet[2630]: E0213 19:19:33.865432 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d195bcf8f9551fdb26f6b42c8b0e4b63d6953f1bbdc1c4b497262fc12be34248\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86475b859c-swqrn" Feb 13 19:19:33.865484 kubelet[2630]: E0213 19:19:33.865450 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d195bcf8f9551fdb26f6b42c8b0e4b63d6953f1bbdc1c4b497262fc12be34248\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86475b859c-swqrn" Feb 13 19:19:33.865532 kubelet[2630]: E0213 19:19:33.865486 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-86475b859c-swqrn_calico-apiserver(5a39efb7-aca1-4893-a36f-840c3c504ece)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-86475b859c-swqrn_calico-apiserver(5a39efb7-aca1-4893-a36f-840c3c504ece)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"d195bcf8f9551fdb26f6b42c8b0e4b63d6953f1bbdc1c4b497262fc12be34248\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-86475b859c-swqrn" podUID="5a39efb7-aca1-4893-a36f-840c3c504ece" Feb 13 19:19:34.076140 kubelet[2630]: I0213 19:19:34.076041 2630 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d195bcf8f9551fdb26f6b42c8b0e4b63d6953f1bbdc1c4b497262fc12be34248" Feb 13 19:19:34.076913 containerd[1515]: time="2025-02-13T19:19:34.076712906Z" level=info msg="StopPodSandbox for \"d195bcf8f9551fdb26f6b42c8b0e4b63d6953f1bbdc1c4b497262fc12be34248\"" Feb 13 19:19:34.077750 containerd[1515]: time="2025-02-13T19:19:34.077713751Z" level=info msg="Ensure that sandbox d195bcf8f9551fdb26f6b42c8b0e4b63d6953f1bbdc1c4b497262fc12be34248 in task-service has been cleanup successfully" Feb 13 19:19:34.077945 containerd[1515]: time="2025-02-13T19:19:34.077906334Z" level=info msg="TearDown network for sandbox \"d195bcf8f9551fdb26f6b42c8b0e4b63d6953f1bbdc1c4b497262fc12be34248\" successfully" Feb 13 19:19:34.079958 kubelet[2630]: I0213 19:19:34.077986 2630 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa373a8af75a289d061969e5a22e647e175a6a451ba7c421e5c44fa363ec90f8" Feb 13 19:19:34.080070 containerd[1515]: time="2025-02-13T19:19:34.078015921Z" level=info msg="StopPodSandbox for \"d195bcf8f9551fdb26f6b42c8b0e4b63d6953f1bbdc1c4b497262fc12be34248\" returns successfully" Feb 13 19:19:34.080070 containerd[1515]: time="2025-02-13T19:19:34.079617427Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86475b859c-swqrn,Uid:5a39efb7-aca1-4893-a36f-840c3c504ece,Namespace:calico-apiserver,Attempt:1,}" Feb 13 19:19:34.080070 containerd[1515]: time="2025-02-13T19:19:34.079865975Z" level=info msg="StopPodSandbox for \"aa373a8af75a289d061969e5a22e647e175a6a451ba7c421e5c44fa363ec90f8\"" Feb 13 19:19:34.080157 containerd[1515]: time="2025-02-13T19:19:34.080091400Z" level=info msg="Ensure that sandbox aa373a8af75a289d061969e5a22e647e175a6a451ba7c421e5c44fa363ec90f8 in task-service has been cleanup successfully" Feb 13 19:19:34.080619 containerd[1515]: time="2025-02-13T19:19:34.080297979Z" level=info msg="TearDown network for sandbox \"aa373a8af75a289d061969e5a22e647e175a6a451ba7c421e5c44fa363ec90f8\" successfully" Feb 13 19:19:34.080619 containerd[1515]: time="2025-02-13T19:19:34.080317605Z" level=info msg="StopPodSandbox for \"aa373a8af75a289d061969e5a22e647e175a6a451ba7c421e5c44fa363ec90f8\" returns successfully" Feb 13 19:19:34.081305 containerd[1515]: time="2025-02-13T19:19:34.081281992Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-664f6f5454-mnbf7,Uid:4e40d638-ead9-4cd4-859c-b3835db75e0c,Namespace:calico-system,Attempt:1,}" Feb 13 19:19:34.082391 kubelet[2630]: I0213 19:19:34.082373 2630 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7" Feb 13 19:19:34.082811 containerd[1515]: time="2025-02-13T19:19:34.082777158Z" level=info msg="StopPodSandbox for \"f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7\"" Feb 13 19:19:34.082973 containerd[1515]: time="2025-02-13T19:19:34.082957548Z" level=info msg="Ensure that sandbox f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7 in task-service has been cleanup successfully" Feb 13 19:19:34.083141 containerd[1515]: time="2025-02-13T19:19:34.083114634Z" level=info msg="TearDown network for sandbox \"f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7\" successfully" Feb 13 19:19:34.083141 containerd[1515]: time="2025-02-13T19:19:34.083128570Z" level=info msg="StopPodSandbox for \"f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7\" returns successfully" Feb 13 19:19:34.083320 kubelet[2630]: E0213 19:19:34.083297 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:34.083540 containerd[1515]: time="2025-02-13T19:19:34.083517282Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-frtk4,Uid:f48c8839-9a17-43c4-a0d5-f9e8d9767814,Namespace:kube-system,Attempt:1,}" Feb 13 19:19:34.084246 kubelet[2630]: E0213 19:19:34.084163 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:34.084748 kubelet[2630]: I0213 19:19:34.084651 2630 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7214e7c2d9f81df76f610a79140dcb0085636bf81f33d1eb139eb52cb12a6af" Feb 13 19:19:34.084794 containerd[1515]: time="2025-02-13T19:19:34.084656057Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.1\"" Feb 13 19:19:34.084979 containerd[1515]: time="2025-02-13T19:19:34.084960200Z" level=info msg="StopPodSandbox for \"f7214e7c2d9f81df76f610a79140dcb0085636bf81f33d1eb139eb52cb12a6af\"" Feb 13 19:19:34.085104 containerd[1515]: time="2025-02-13T19:19:34.085090114Z" level=info msg="Ensure that sandbox f7214e7c2d9f81df76f610a79140dcb0085636bf81f33d1eb139eb52cb12a6af in task-service has been cleanup successfully" Feb 13 19:19:34.085374 containerd[1515]: time="2025-02-13T19:19:34.085336979Z" level=info msg="TearDown network for sandbox \"f7214e7c2d9f81df76f610a79140dcb0085636bf81f33d1eb139eb52cb12a6af\" successfully" Feb 13 19:19:34.085374 containerd[1515]: time="2025-02-13T19:19:34.085356396Z" level=info msg="StopPodSandbox for \"f7214e7c2d9f81df76f610a79140dcb0085636bf81f33d1eb139eb52cb12a6af\" returns successfully" Feb 13 19:19:34.085727 containerd[1515]: time="2025-02-13T19:19:34.085686398Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86475b859c-kwlfl,Uid:f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1,Namespace:calico-apiserver,Attempt:1,}" Feb 13 19:19:34.085838 kubelet[2630]: I0213 19:19:34.085820 2630 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2995529ca797998e1cf161d883bb684a3464a8a8d57602758f8705313f0d601" Feb 13 19:19:34.086167 containerd[1515]: time="2025-02-13T19:19:34.086133440Z" level=info msg="StopPodSandbox for \"e2995529ca797998e1cf161d883bb684a3464a8a8d57602758f8705313f0d601\"" Feb 13 19:19:34.086318 containerd[1515]: time="2025-02-13T19:19:34.086277451Z" level=info msg="Ensure that sandbox e2995529ca797998e1cf161d883bb684a3464a8a8d57602758f8705313f0d601 in task-service has been cleanup successfully" Feb 13 19:19:34.086411 containerd[1515]: time="2025-02-13T19:19:34.086391797Z" level=info msg="TearDown network for sandbox \"e2995529ca797998e1cf161d883bb684a3464a8a8d57602758f8705313f0d601\" successfully" Feb 13 19:19:34.086411 containerd[1515]: time="2025-02-13T19:19:34.086407406Z" level=info msg="StopPodSandbox for \"e2995529ca797998e1cf161d883bb684a3464a8a8d57602758f8705313f0d601\" returns successfully" Feb 13 19:19:34.086565 kubelet[2630]: E0213 19:19:34.086547 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:34.086750 containerd[1515]: time="2025-02-13T19:19:34.086733670Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-s8264,Uid:f2fdfd8a-780b-4320-8f24-ca003c5d0c77,Namespace:kube-system,Attempt:1,}" Feb 13 19:19:34.374670 containerd[1515]: time="2025-02-13T19:19:34.374533678Z" level=error msg="Failed to destroy network for sandbox \"a20bc39d31c8f248a56fc8b89d2d3ce62836f574c54b75502431e23bb908b697\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:34.375085 containerd[1515]: time="2025-02-13T19:19:34.375024713Z" level=error msg="encountered an error cleaning up failed sandbox \"a20bc39d31c8f248a56fc8b89d2d3ce62836f574c54b75502431e23bb908b697\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:34.375127 containerd[1515]: time="2025-02-13T19:19:34.375088804Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-frtk4,Uid:f48c8839-9a17-43c4-a0d5-f9e8d9767814,Namespace:kube-system,Attempt:1,} failed, error" error="failed to setup network for sandbox \"a20bc39d31c8f248a56fc8b89d2d3ce62836f574c54b75502431e23bb908b697\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:34.375410 kubelet[2630]: E0213 19:19:34.375370 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a20bc39d31c8f248a56fc8b89d2d3ce62836f574c54b75502431e23bb908b697\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:34.375459 kubelet[2630]: E0213 19:19:34.375430 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a20bc39d31c8f248a56fc8b89d2d3ce62836f574c54b75502431e23bb908b697\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-frtk4" Feb 13 19:19:34.375459 kubelet[2630]: E0213 19:19:34.375449 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a20bc39d31c8f248a56fc8b89d2d3ce62836f574c54b75502431e23bb908b697\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-frtk4" Feb 13 19:19:34.375510 kubelet[2630]: E0213 19:19:34.375489 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-6f6b679f8f-frtk4_kube-system(f48c8839-9a17-43c4-a0d5-f9e8d9767814)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-6f6b679f8f-frtk4_kube-system(f48c8839-9a17-43c4-a0d5-f9e8d9767814)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"a20bc39d31c8f248a56fc8b89d2d3ce62836f574c54b75502431e23bb908b697\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-6f6b679f8f-frtk4" podUID="f48c8839-9a17-43c4-a0d5-f9e8d9767814" Feb 13 19:19:34.382268 containerd[1515]: time="2025-02-13T19:19:34.382129134Z" level=error msg="Failed to destroy network for sandbox \"7d827c446a1e808002e0f19943157ab97bcc163d150e79960814ed79c5b4fbf3\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:34.382771 containerd[1515]: time="2025-02-13T19:19:34.382739844Z" level=error msg="encountered an error cleaning up failed sandbox \"7d827c446a1e808002e0f19943157ab97bcc163d150e79960814ed79c5b4fbf3\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:34.382912 containerd[1515]: time="2025-02-13T19:19:34.382889997Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86475b859c-swqrn,Uid:5a39efb7-aca1-4893-a36f-840c3c504ece,Namespace:calico-apiserver,Attempt:1,} failed, error" error="failed to setup network for sandbox \"7d827c446a1e808002e0f19943157ab97bcc163d150e79960814ed79c5b4fbf3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:34.383269 kubelet[2630]: E0213 19:19:34.383239 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7d827c446a1e808002e0f19943157ab97bcc163d150e79960814ed79c5b4fbf3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:34.383319 kubelet[2630]: E0213 19:19:34.383292 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7d827c446a1e808002e0f19943157ab97bcc163d150e79960814ed79c5b4fbf3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86475b859c-swqrn" Feb 13 19:19:34.383370 kubelet[2630]: E0213 19:19:34.383314 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7d827c446a1e808002e0f19943157ab97bcc163d150e79960814ed79c5b4fbf3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86475b859c-swqrn" Feb 13 19:19:34.383484 kubelet[2630]: E0213 19:19:34.383445 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-86475b859c-swqrn_calico-apiserver(5a39efb7-aca1-4893-a36f-840c3c504ece)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-86475b859c-swqrn_calico-apiserver(5a39efb7-aca1-4893-a36f-840c3c504ece)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"7d827c446a1e808002e0f19943157ab97bcc163d150e79960814ed79c5b4fbf3\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-86475b859c-swqrn" podUID="5a39efb7-aca1-4893-a36f-840c3c504ece" Feb 13 19:19:34.385356 containerd[1515]: time="2025-02-13T19:19:34.385143110Z" level=error msg="Failed to destroy network for sandbox \"b4101238369440d1fc253b3e442bebf940a9f5fa609c91f1252b6dcea502c067\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:34.385698 containerd[1515]: time="2025-02-13T19:19:34.385675484Z" level=error msg="encountered an error cleaning up failed sandbox \"b4101238369440d1fc253b3e442bebf940a9f5fa609c91f1252b6dcea502c067\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:34.385759 containerd[1515]: time="2025-02-13T19:19:34.385723374Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86475b859c-kwlfl,Uid:f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1,Namespace:calico-apiserver,Attempt:1,} failed, error" error="failed to setup network for sandbox \"b4101238369440d1fc253b3e442bebf940a9f5fa609c91f1252b6dcea502c067\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:34.385952 kubelet[2630]: E0213 19:19:34.385918 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b4101238369440d1fc253b3e442bebf940a9f5fa609c91f1252b6dcea502c067\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:34.385990 kubelet[2630]: E0213 19:19:34.385973 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b4101238369440d1fc253b3e442bebf940a9f5fa609c91f1252b6dcea502c067\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86475b859c-kwlfl" Feb 13 19:19:34.386105 kubelet[2630]: E0213 19:19:34.386041 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b4101238369440d1fc253b3e442bebf940a9f5fa609c91f1252b6dcea502c067\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86475b859c-kwlfl" Feb 13 19:19:34.386230 kubelet[2630]: E0213 19:19:34.386128 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-86475b859c-kwlfl_calico-apiserver(f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-86475b859c-kwlfl_calico-apiserver(f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"b4101238369440d1fc253b3e442bebf940a9f5fa609c91f1252b6dcea502c067\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-86475b859c-kwlfl" podUID="f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1" Feb 13 19:19:34.391897 containerd[1515]: time="2025-02-13T19:19:34.391785621Z" level=error msg="Failed to destroy network for sandbox \"4304f46dfb22ea4aced73143fa3f59c4ed426a670211c8559cc8ed545f1b3804\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:34.392360 containerd[1515]: time="2025-02-13T19:19:34.392337201Z" level=error msg="encountered an error cleaning up failed sandbox \"4304f46dfb22ea4aced73143fa3f59c4ed426a670211c8559cc8ed545f1b3804\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:34.392474 containerd[1515]: time="2025-02-13T19:19:34.392455774Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-664f6f5454-mnbf7,Uid:4e40d638-ead9-4cd4-859c-b3835db75e0c,Namespace:calico-system,Attempt:1,} failed, error" error="failed to setup network for sandbox \"4304f46dfb22ea4aced73143fa3f59c4ed426a670211c8559cc8ed545f1b3804\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:34.392781 kubelet[2630]: E0213 19:19:34.392727 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4304f46dfb22ea4aced73143fa3f59c4ed426a670211c8559cc8ed545f1b3804\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:34.392841 kubelet[2630]: E0213 19:19:34.392791 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4304f46dfb22ea4aced73143fa3f59c4ed426a670211c8559cc8ed545f1b3804\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-664f6f5454-mnbf7" Feb 13 19:19:34.392841 kubelet[2630]: E0213 19:19:34.392810 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4304f46dfb22ea4aced73143fa3f59c4ed426a670211c8559cc8ed545f1b3804\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-664f6f5454-mnbf7" Feb 13 19:19:34.392893 kubelet[2630]: E0213 19:19:34.392849 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-664f6f5454-mnbf7_calico-system(4e40d638-ead9-4cd4-859c-b3835db75e0c)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-664f6f5454-mnbf7_calico-system(4e40d638-ead9-4cd4-859c-b3835db75e0c)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"4304f46dfb22ea4aced73143fa3f59c4ed426a670211c8559cc8ed545f1b3804\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-664f6f5454-mnbf7" podUID="4e40d638-ead9-4cd4-859c-b3835db75e0c" Feb 13 19:19:34.397834 systemd[1]: run-netns-cni\x2dd9e70649\x2d55f8\x2d86fa\x2ddbab\x2ddd09083a6669.mount: Deactivated successfully. Feb 13 19:19:34.397972 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7-shm.mount: Deactivated successfully. Feb 13 19:19:34.400272 containerd[1515]: time="2025-02-13T19:19:34.400024609Z" level=error msg="Failed to destroy network for sandbox \"706e09ab546bf3d89157cd9a2c32b16d198260f6686345817b1acf65f680fdd2\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:34.400573 containerd[1515]: time="2025-02-13T19:19:34.400552674Z" level=error msg="encountered an error cleaning up failed sandbox \"706e09ab546bf3d89157cd9a2c32b16d198260f6686345817b1acf65f680fdd2\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:34.400703 containerd[1515]: time="2025-02-13T19:19:34.400686045Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-s8264,Uid:f2fdfd8a-780b-4320-8f24-ca003c5d0c77,Namespace:kube-system,Attempt:1,} failed, error" error="failed to setup network for sandbox \"706e09ab546bf3d89157cd9a2c32b16d198260f6686345817b1acf65f680fdd2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:34.400957 kubelet[2630]: E0213 19:19:34.400920 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"706e09ab546bf3d89157cd9a2c32b16d198260f6686345817b1acf65f680fdd2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:34.401033 kubelet[2630]: E0213 19:19:34.400970 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"706e09ab546bf3d89157cd9a2c32b16d198260f6686345817b1acf65f680fdd2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-s8264" Feb 13 19:19:34.401033 kubelet[2630]: E0213 19:19:34.400989 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"706e09ab546bf3d89157cd9a2c32b16d198260f6686345817b1acf65f680fdd2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-s8264" Feb 13 19:19:34.401091 kubelet[2630]: E0213 19:19:34.401025 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-6f6b679f8f-s8264_kube-system(f2fdfd8a-780b-4320-8f24-ca003c5d0c77)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-6f6b679f8f-s8264_kube-system(f2fdfd8a-780b-4320-8f24-ca003c5d0c77)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"706e09ab546bf3d89157cd9a2c32b16d198260f6686345817b1acf65f680fdd2\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-6f6b679f8f-s8264" podUID="f2fdfd8a-780b-4320-8f24-ca003c5d0c77" Feb 13 19:19:34.403489 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-706e09ab546bf3d89157cd9a2c32b16d198260f6686345817b1acf65f680fdd2-shm.mount: Deactivated successfully. Feb 13 19:19:34.663289 systemd[1]: Created slice kubepods-besteffort-pod84096616_cf79_421c_9e94_8cbcf0a204a5.slice - libcontainer container kubepods-besteffort-pod84096616_cf79_421c_9e94_8cbcf0a204a5.slice. Feb 13 19:19:34.665590 containerd[1515]: time="2025-02-13T19:19:34.665546367Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-drlnv,Uid:84096616-cf79-421c-9e94-8cbcf0a204a5,Namespace:calico-system,Attempt:0,}" Feb 13 19:19:34.719717 containerd[1515]: time="2025-02-13T19:19:34.719653946Z" level=error msg="Failed to destroy network for sandbox \"fa92388357966f3e183c59506d6cfde3e38d36deede0f30821c07e22d4146a49\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:34.720074 containerd[1515]: time="2025-02-13T19:19:34.720043911Z" level=error msg="encountered an error cleaning up failed sandbox \"fa92388357966f3e183c59506d6cfde3e38d36deede0f30821c07e22d4146a49\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:34.720120 containerd[1515]: time="2025-02-13T19:19:34.720103514Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-drlnv,Uid:84096616-cf79-421c-9e94-8cbcf0a204a5,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"fa92388357966f3e183c59506d6cfde3e38d36deede0f30821c07e22d4146a49\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:34.720365 kubelet[2630]: E0213 19:19:34.720310 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fa92388357966f3e183c59506d6cfde3e38d36deede0f30821c07e22d4146a49\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:34.720408 kubelet[2630]: E0213 19:19:34.720385 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fa92388357966f3e183c59506d6cfde3e38d36deede0f30821c07e22d4146a49\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-drlnv" Feb 13 19:19:34.720431 kubelet[2630]: E0213 19:19:34.720404 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fa92388357966f3e183c59506d6cfde3e38d36deede0f30821c07e22d4146a49\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-drlnv" Feb 13 19:19:34.720467 kubelet[2630]: E0213 19:19:34.720445 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-drlnv_calico-system(84096616-cf79-421c-9e94-8cbcf0a204a5)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-drlnv_calico-system(84096616-cf79-421c-9e94-8cbcf0a204a5)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"fa92388357966f3e183c59506d6cfde3e38d36deede0f30821c07e22d4146a49\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-drlnv" podUID="84096616-cf79-421c-9e94-8cbcf0a204a5" Feb 13 19:19:35.088402 kubelet[2630]: I0213 19:19:35.088360 2630 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d827c446a1e808002e0f19943157ab97bcc163d150e79960814ed79c5b4fbf3" Feb 13 19:19:35.088875 containerd[1515]: time="2025-02-13T19:19:35.088841871Z" level=info msg="StopPodSandbox for \"7d827c446a1e808002e0f19943157ab97bcc163d150e79960814ed79c5b4fbf3\"" Feb 13 19:19:35.089082 containerd[1515]: time="2025-02-13T19:19:35.089054030Z" level=info msg="Ensure that sandbox 7d827c446a1e808002e0f19943157ab97bcc163d150e79960814ed79c5b4fbf3 in task-service has been cleanup successfully" Feb 13 19:19:35.089322 containerd[1515]: time="2025-02-13T19:19:35.089268364Z" level=info msg="TearDown network for sandbox \"7d827c446a1e808002e0f19943157ab97bcc163d150e79960814ed79c5b4fbf3\" successfully" Feb 13 19:19:35.089322 containerd[1515]: time="2025-02-13T19:19:35.089286899Z" level=info msg="StopPodSandbox for \"7d827c446a1e808002e0f19943157ab97bcc163d150e79960814ed79c5b4fbf3\" returns successfully" Feb 13 19:19:35.089772 containerd[1515]: time="2025-02-13T19:19:35.089539334Z" level=info msg="StopPodSandbox for \"d195bcf8f9551fdb26f6b42c8b0e4b63d6953f1bbdc1c4b497262fc12be34248\"" Feb 13 19:19:35.089772 containerd[1515]: time="2025-02-13T19:19:35.089634203Z" level=info msg="TearDown network for sandbox \"d195bcf8f9551fdb26f6b42c8b0e4b63d6953f1bbdc1c4b497262fc12be34248\" successfully" Feb 13 19:19:35.089772 containerd[1515]: time="2025-02-13T19:19:35.089646446Z" level=info msg="StopPodSandbox for \"d195bcf8f9551fdb26f6b42c8b0e4b63d6953f1bbdc1c4b497262fc12be34248\" returns successfully" Feb 13 19:19:35.089993 kubelet[2630]: I0213 19:19:35.089961 2630 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4304f46dfb22ea4aced73143fa3f59c4ed426a670211c8559cc8ed545f1b3804" Feb 13 19:19:35.090038 containerd[1515]: time="2025-02-13T19:19:35.089994111Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86475b859c-swqrn,Uid:5a39efb7-aca1-4893-a36f-840c3c504ece,Namespace:calico-apiserver,Attempt:2,}" Feb 13 19:19:35.090320 containerd[1515]: time="2025-02-13T19:19:35.090275901Z" level=info msg="StopPodSandbox for \"4304f46dfb22ea4aced73143fa3f59c4ed426a670211c8559cc8ed545f1b3804\"" Feb 13 19:19:35.090606 containerd[1515]: time="2025-02-13T19:19:35.090457062Z" level=info msg="Ensure that sandbox 4304f46dfb22ea4aced73143fa3f59c4ed426a670211c8559cc8ed545f1b3804 in task-service has been cleanup successfully" Feb 13 19:19:35.090741 containerd[1515]: time="2025-02-13T19:19:35.090717222Z" level=info msg="TearDown network for sandbox \"4304f46dfb22ea4aced73143fa3f59c4ed426a670211c8559cc8ed545f1b3804\" successfully" Feb 13 19:19:35.090741 containerd[1515]: time="2025-02-13T19:19:35.090736799Z" level=info msg="StopPodSandbox for \"4304f46dfb22ea4aced73143fa3f59c4ed426a670211c8559cc8ed545f1b3804\" returns successfully" Feb 13 19:19:35.091043 containerd[1515]: time="2025-02-13T19:19:35.091007228Z" level=info msg="StopPodSandbox for \"aa373a8af75a289d061969e5a22e647e175a6a451ba7c421e5c44fa363ec90f8\"" Feb 13 19:19:35.091112 containerd[1515]: time="2025-02-13T19:19:35.091094332Z" level=info msg="TearDown network for sandbox \"aa373a8af75a289d061969e5a22e647e175a6a451ba7c421e5c44fa363ec90f8\" successfully" Feb 13 19:19:35.091155 containerd[1515]: time="2025-02-13T19:19:35.091110292Z" level=info msg="StopPodSandbox for \"aa373a8af75a289d061969e5a22e647e175a6a451ba7c421e5c44fa363ec90f8\" returns successfully" Feb 13 19:19:35.091625 containerd[1515]: time="2025-02-13T19:19:35.091589314Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-664f6f5454-mnbf7,Uid:4e40d638-ead9-4cd4-859c-b3835db75e0c,Namespace:calico-system,Attempt:2,}" Feb 13 19:19:35.092145 kubelet[2630]: I0213 19:19:35.092112 2630 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a20bc39d31c8f248a56fc8b89d2d3ce62836f574c54b75502431e23bb908b697" Feb 13 19:19:35.092671 containerd[1515]: time="2025-02-13T19:19:35.092648258Z" level=info msg="StopPodSandbox for \"a20bc39d31c8f248a56fc8b89d2d3ce62836f574c54b75502431e23bb908b697\"" Feb 13 19:19:35.092861 containerd[1515]: time="2025-02-13T19:19:35.092828658Z" level=info msg="Ensure that sandbox a20bc39d31c8f248a56fc8b89d2d3ce62836f574c54b75502431e23bb908b697 in task-service has been cleanup successfully" Feb 13 19:19:35.093160 containerd[1515]: time="2025-02-13T19:19:35.093140105Z" level=info msg="TearDown network for sandbox \"a20bc39d31c8f248a56fc8b89d2d3ce62836f574c54b75502431e23bb908b697\" successfully" Feb 13 19:19:35.093218 containerd[1515]: time="2025-02-13T19:19:35.093159170Z" level=info msg="StopPodSandbox for \"a20bc39d31c8f248a56fc8b89d2d3ce62836f574c54b75502431e23bb908b697\" returns successfully" Feb 13 19:19:35.093465 containerd[1515]: time="2025-02-13T19:19:35.093368294Z" level=info msg="StopPodSandbox for \"f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7\"" Feb 13 19:19:35.093465 containerd[1515]: time="2025-02-13T19:19:35.093447684Z" level=info msg="TearDown network for sandbox \"f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7\" successfully" Feb 13 19:19:35.093465 containerd[1515]: time="2025-02-13T19:19:35.093457041Z" level=info msg="StopPodSandbox for \"f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7\" returns successfully" Feb 13 19:19:35.093661 kubelet[2630]: E0213 19:19:35.093614 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:35.093661 kubelet[2630]: I0213 19:19:35.093633 2630 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa92388357966f3e183c59506d6cfde3e38d36deede0f30821c07e22d4146a49" Feb 13 19:19:35.093838 containerd[1515]: time="2025-02-13T19:19:35.093811068Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-frtk4,Uid:f48c8839-9a17-43c4-a0d5-f9e8d9767814,Namespace:kube-system,Attempt:2,}" Feb 13 19:19:35.094055 containerd[1515]: time="2025-02-13T19:19:35.094038435Z" level=info msg="StopPodSandbox for \"fa92388357966f3e183c59506d6cfde3e38d36deede0f30821c07e22d4146a49\"" Feb 13 19:19:35.094231 containerd[1515]: time="2025-02-13T19:19:35.094211722Z" level=info msg="Ensure that sandbox fa92388357966f3e183c59506d6cfde3e38d36deede0f30821c07e22d4146a49 in task-service has been cleanup successfully" Feb 13 19:19:35.095323 containerd[1515]: time="2025-02-13T19:19:35.095262200Z" level=info msg="TearDown network for sandbox \"fa92388357966f3e183c59506d6cfde3e38d36deede0f30821c07e22d4146a49\" successfully" Feb 13 19:19:35.095323 containerd[1515]: time="2025-02-13T19:19:35.095279483Z" level=info msg="StopPodSandbox for \"fa92388357966f3e183c59506d6cfde3e38d36deede0f30821c07e22d4146a49\" returns successfully" Feb 13 19:19:35.095465 kubelet[2630]: I0213 19:19:35.095393 2630 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4101238369440d1fc253b3e442bebf940a9f5fa609c91f1252b6dcea502c067" Feb 13 19:19:35.096236 containerd[1515]: time="2025-02-13T19:19:35.095919017Z" level=info msg="StopPodSandbox for \"b4101238369440d1fc253b3e442bebf940a9f5fa609c91f1252b6dcea502c067\"" Feb 13 19:19:35.096236 containerd[1515]: time="2025-02-13T19:19:35.095958952Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-drlnv,Uid:84096616-cf79-421c-9e94-8cbcf0a204a5,Namespace:calico-system,Attempt:1,}" Feb 13 19:19:35.096236 containerd[1515]: time="2025-02-13T19:19:35.096098194Z" level=info msg="Ensure that sandbox b4101238369440d1fc253b3e442bebf940a9f5fa609c91f1252b6dcea502c067 in task-service has been cleanup successfully" Feb 13 19:19:35.096442 containerd[1515]: time="2025-02-13T19:19:35.096421883Z" level=info msg="TearDown network for sandbox \"b4101238369440d1fc253b3e442bebf940a9f5fa609c91f1252b6dcea502c067\" successfully" Feb 13 19:19:35.096502 containerd[1515]: time="2025-02-13T19:19:35.096487187Z" level=info msg="StopPodSandbox for \"b4101238369440d1fc253b3e442bebf940a9f5fa609c91f1252b6dcea502c067\" returns successfully" Feb 13 19:19:35.096883 containerd[1515]: time="2025-02-13T19:19:35.096861932Z" level=info msg="StopPodSandbox for \"f7214e7c2d9f81df76f610a79140dcb0085636bf81f33d1eb139eb52cb12a6af\"" Feb 13 19:19:35.097253 containerd[1515]: time="2025-02-13T19:19:35.096940269Z" level=info msg="TearDown network for sandbox \"f7214e7c2d9f81df76f610a79140dcb0085636bf81f33d1eb139eb52cb12a6af\" successfully" Feb 13 19:19:35.097253 containerd[1515]: time="2025-02-13T19:19:35.096950689Z" level=info msg="StopPodSandbox for \"f7214e7c2d9f81df76f610a79140dcb0085636bf81f33d1eb139eb52cb12a6af\" returns successfully" Feb 13 19:19:35.097314 kubelet[2630]: I0213 19:19:35.096962 2630 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="706e09ab546bf3d89157cd9a2c32b16d198260f6686345817b1acf65f680fdd2" Feb 13 19:19:35.097641 containerd[1515]: time="2025-02-13T19:19:35.097617085Z" level=info msg="StopPodSandbox for \"706e09ab546bf3d89157cd9a2c32b16d198260f6686345817b1acf65f680fdd2\"" Feb 13 19:19:35.097747 containerd[1515]: time="2025-02-13T19:19:35.097719607Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86475b859c-kwlfl,Uid:f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1,Namespace:calico-apiserver,Attempt:2,}" Feb 13 19:19:35.097810 containerd[1515]: time="2025-02-13T19:19:35.097789439Z" level=info msg="Ensure that sandbox 706e09ab546bf3d89157cd9a2c32b16d198260f6686345817b1acf65f680fdd2 in task-service has been cleanup successfully" Feb 13 19:19:35.098059 containerd[1515]: time="2025-02-13T19:19:35.098040652Z" level=info msg="TearDown network for sandbox \"706e09ab546bf3d89157cd9a2c32b16d198260f6686345817b1acf65f680fdd2\" successfully" Feb 13 19:19:35.098059 containerd[1515]: time="2025-02-13T19:19:35.098055510Z" level=info msg="StopPodSandbox for \"706e09ab546bf3d89157cd9a2c32b16d198260f6686345817b1acf65f680fdd2\" returns successfully" Feb 13 19:19:35.098524 containerd[1515]: time="2025-02-13T19:19:35.098387354Z" level=info msg="StopPodSandbox for \"e2995529ca797998e1cf161d883bb684a3464a8a8d57602758f8705313f0d601\"" Feb 13 19:19:35.098524 containerd[1515]: time="2025-02-13T19:19:35.098469449Z" level=info msg="TearDown network for sandbox \"e2995529ca797998e1cf161d883bb684a3464a8a8d57602758f8705313f0d601\" successfully" Feb 13 19:19:35.098524 containerd[1515]: time="2025-02-13T19:19:35.098478897Z" level=info msg="StopPodSandbox for \"e2995529ca797998e1cf161d883bb684a3464a8a8d57602758f8705313f0d601\" returns successfully" Feb 13 19:19:35.098645 kubelet[2630]: E0213 19:19:35.098625 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:35.098896 containerd[1515]: time="2025-02-13T19:19:35.098843874Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-s8264,Uid:f2fdfd8a-780b-4320-8f24-ca003c5d0c77,Namespace:kube-system,Attempt:2,}" Feb 13 19:19:35.204791 containerd[1515]: time="2025-02-13T19:19:35.204739770Z" level=error msg="Failed to destroy network for sandbox \"154203fe3c4258e99eeb24eefabf42a73b9226503cd839391e62422056e09db6\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:35.207239 containerd[1515]: time="2025-02-13T19:19:35.207180006Z" level=error msg="encountered an error cleaning up failed sandbox \"154203fe3c4258e99eeb24eefabf42a73b9226503cd839391e62422056e09db6\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:35.207709 containerd[1515]: time="2025-02-13T19:19:35.207441097Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86475b859c-swqrn,Uid:5a39efb7-aca1-4893-a36f-840c3c504ece,Namespace:calico-apiserver,Attempt:2,} failed, error" error="failed to setup network for sandbox \"154203fe3c4258e99eeb24eefabf42a73b9226503cd839391e62422056e09db6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:35.207797 kubelet[2630]: E0213 19:19:35.207665 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"154203fe3c4258e99eeb24eefabf42a73b9226503cd839391e62422056e09db6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:35.207797 kubelet[2630]: E0213 19:19:35.207728 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"154203fe3c4258e99eeb24eefabf42a73b9226503cd839391e62422056e09db6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86475b859c-swqrn" Feb 13 19:19:35.207797 kubelet[2630]: E0213 19:19:35.207750 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"154203fe3c4258e99eeb24eefabf42a73b9226503cd839391e62422056e09db6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86475b859c-swqrn" Feb 13 19:19:35.207937 kubelet[2630]: E0213 19:19:35.207791 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-86475b859c-swqrn_calico-apiserver(5a39efb7-aca1-4893-a36f-840c3c504ece)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-86475b859c-swqrn_calico-apiserver(5a39efb7-aca1-4893-a36f-840c3c504ece)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"154203fe3c4258e99eeb24eefabf42a73b9226503cd839391e62422056e09db6\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-86475b859c-swqrn" podUID="5a39efb7-aca1-4893-a36f-840c3c504ece" Feb 13 19:19:35.213819 containerd[1515]: time="2025-02-13T19:19:35.213719870Z" level=error msg="Failed to destroy network for sandbox \"e5e6531fa4d44fae3b14be640ca5608599c77b930300d189004cbcc636c08e1b\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:35.215939 containerd[1515]: time="2025-02-13T19:19:35.215916906Z" level=error msg="encountered an error cleaning up failed sandbox \"e5e6531fa4d44fae3b14be640ca5608599c77b930300d189004cbcc636c08e1b\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:35.216153 containerd[1515]: time="2025-02-13T19:19:35.216102997Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-664f6f5454-mnbf7,Uid:4e40d638-ead9-4cd4-859c-b3835db75e0c,Namespace:calico-system,Attempt:2,} failed, error" error="failed to setup network for sandbox \"e5e6531fa4d44fae3b14be640ca5608599c77b930300d189004cbcc636c08e1b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:35.216597 kubelet[2630]: E0213 19:19:35.216438 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e5e6531fa4d44fae3b14be640ca5608599c77b930300d189004cbcc636c08e1b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:35.216597 kubelet[2630]: E0213 19:19:35.216494 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e5e6531fa4d44fae3b14be640ca5608599c77b930300d189004cbcc636c08e1b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-664f6f5454-mnbf7" Feb 13 19:19:35.216597 kubelet[2630]: E0213 19:19:35.216516 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e5e6531fa4d44fae3b14be640ca5608599c77b930300d189004cbcc636c08e1b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-664f6f5454-mnbf7" Feb 13 19:19:35.216711 kubelet[2630]: E0213 19:19:35.216551 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-664f6f5454-mnbf7_calico-system(4e40d638-ead9-4cd4-859c-b3835db75e0c)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-664f6f5454-mnbf7_calico-system(4e40d638-ead9-4cd4-859c-b3835db75e0c)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"e5e6531fa4d44fae3b14be640ca5608599c77b930300d189004cbcc636c08e1b\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-664f6f5454-mnbf7" podUID="4e40d638-ead9-4cd4-859c-b3835db75e0c" Feb 13 19:19:35.227090 containerd[1515]: time="2025-02-13T19:19:35.227035683Z" level=error msg="Failed to destroy network for sandbox \"1d811fd694e600e47efc43338f2c97ee7c9f512c6106b43a1a049dbc81724231\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:35.227477 containerd[1515]: time="2025-02-13T19:19:35.227448290Z" level=error msg="encountered an error cleaning up failed sandbox \"1d811fd694e600e47efc43338f2c97ee7c9f512c6106b43a1a049dbc81724231\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:35.227549 containerd[1515]: time="2025-02-13T19:19:35.227512600Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-drlnv,Uid:84096616-cf79-421c-9e94-8cbcf0a204a5,Namespace:calico-system,Attempt:1,} failed, error" error="failed to setup network for sandbox \"1d811fd694e600e47efc43338f2c97ee7c9f512c6106b43a1a049dbc81724231\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:35.227815 kubelet[2630]: E0213 19:19:35.227763 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1d811fd694e600e47efc43338f2c97ee7c9f512c6106b43a1a049dbc81724231\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:35.227867 kubelet[2630]: E0213 19:19:35.227838 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1d811fd694e600e47efc43338f2c97ee7c9f512c6106b43a1a049dbc81724231\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-drlnv" Feb 13 19:19:35.227894 kubelet[2630]: E0213 19:19:35.227862 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1d811fd694e600e47efc43338f2c97ee7c9f512c6106b43a1a049dbc81724231\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-drlnv" Feb 13 19:19:35.227944 kubelet[2630]: E0213 19:19:35.227906 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-drlnv_calico-system(84096616-cf79-421c-9e94-8cbcf0a204a5)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-drlnv_calico-system(84096616-cf79-421c-9e94-8cbcf0a204a5)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"1d811fd694e600e47efc43338f2c97ee7c9f512c6106b43a1a049dbc81724231\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-drlnv" podUID="84096616-cf79-421c-9e94-8cbcf0a204a5" Feb 13 19:19:35.233203 containerd[1515]: time="2025-02-13T19:19:35.232068940Z" level=error msg="Failed to destroy network for sandbox \"28a752266e9ffa0b0f35649f5eba403b238d216dc8ce20a8fe3622cc28799217\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:35.233203 containerd[1515]: time="2025-02-13T19:19:35.232491235Z" level=error msg="encountered an error cleaning up failed sandbox \"28a752266e9ffa0b0f35649f5eba403b238d216dc8ce20a8fe3622cc28799217\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:35.233203 containerd[1515]: time="2025-02-13T19:19:35.232557009Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-frtk4,Uid:f48c8839-9a17-43c4-a0d5-f9e8d9767814,Namespace:kube-system,Attempt:2,} failed, error" error="failed to setup network for sandbox \"28a752266e9ffa0b0f35649f5eba403b238d216dc8ce20a8fe3622cc28799217\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:35.233352 kubelet[2630]: E0213 19:19:35.232835 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"28a752266e9ffa0b0f35649f5eba403b238d216dc8ce20a8fe3622cc28799217\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:35.233352 kubelet[2630]: E0213 19:19:35.232900 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"28a752266e9ffa0b0f35649f5eba403b238d216dc8ce20a8fe3622cc28799217\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-frtk4" Feb 13 19:19:35.233352 kubelet[2630]: E0213 19:19:35.232922 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"28a752266e9ffa0b0f35649f5eba403b238d216dc8ce20a8fe3622cc28799217\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-frtk4" Feb 13 19:19:35.233439 kubelet[2630]: E0213 19:19:35.232970 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-6f6b679f8f-frtk4_kube-system(f48c8839-9a17-43c4-a0d5-f9e8d9767814)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-6f6b679f8f-frtk4_kube-system(f48c8839-9a17-43c4-a0d5-f9e8d9767814)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"28a752266e9ffa0b0f35649f5eba403b238d216dc8ce20a8fe3622cc28799217\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-6f6b679f8f-frtk4" podUID="f48c8839-9a17-43c4-a0d5-f9e8d9767814" Feb 13 19:19:35.249985 containerd[1515]: time="2025-02-13T19:19:35.249927196Z" level=error msg="Failed to destroy network for sandbox \"787aa28bb52b7073a96ced7755cf59d24334348abd3a51ce97858110663e1200\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:35.252609 containerd[1515]: time="2025-02-13T19:19:35.252439276Z" level=error msg="encountered an error cleaning up failed sandbox \"787aa28bb52b7073a96ced7755cf59d24334348abd3a51ce97858110663e1200\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:35.252609 containerd[1515]: time="2025-02-13T19:19:35.252506864Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86475b859c-kwlfl,Uid:f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1,Namespace:calico-apiserver,Attempt:2,} failed, error" error="failed to setup network for sandbox \"787aa28bb52b7073a96ced7755cf59d24334348abd3a51ce97858110663e1200\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:35.252783 kubelet[2630]: E0213 19:19:35.252742 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"787aa28bb52b7073a96ced7755cf59d24334348abd3a51ce97858110663e1200\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:35.252834 kubelet[2630]: E0213 19:19:35.252807 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"787aa28bb52b7073a96ced7755cf59d24334348abd3a51ce97858110663e1200\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86475b859c-kwlfl" Feb 13 19:19:35.252863 kubelet[2630]: E0213 19:19:35.252838 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"787aa28bb52b7073a96ced7755cf59d24334348abd3a51ce97858110663e1200\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86475b859c-kwlfl" Feb 13 19:19:35.252913 kubelet[2630]: E0213 19:19:35.252884 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-86475b859c-kwlfl_calico-apiserver(f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-86475b859c-kwlfl_calico-apiserver(f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"787aa28bb52b7073a96ced7755cf59d24334348abd3a51ce97858110663e1200\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-86475b859c-kwlfl" podUID="f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1" Feb 13 19:19:35.266953 containerd[1515]: time="2025-02-13T19:19:35.266888794Z" level=error msg="Failed to destroy network for sandbox \"8509c778c3294467a0ad8b8059bdecd145729001ee04e3c09d32d7f0d60b9468\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:35.267413 containerd[1515]: time="2025-02-13T19:19:35.267381943Z" level=error msg="encountered an error cleaning up failed sandbox \"8509c778c3294467a0ad8b8059bdecd145729001ee04e3c09d32d7f0d60b9468\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:35.267476 containerd[1515]: time="2025-02-13T19:19:35.267447666Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-s8264,Uid:f2fdfd8a-780b-4320-8f24-ca003c5d0c77,Namespace:kube-system,Attempt:2,} failed, error" error="failed to setup network for sandbox \"8509c778c3294467a0ad8b8059bdecd145729001ee04e3c09d32d7f0d60b9468\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:35.267708 kubelet[2630]: E0213 19:19:35.267666 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8509c778c3294467a0ad8b8059bdecd145729001ee04e3c09d32d7f0d60b9468\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:35.267777 kubelet[2630]: E0213 19:19:35.267728 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8509c778c3294467a0ad8b8059bdecd145729001ee04e3c09d32d7f0d60b9468\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-s8264" Feb 13 19:19:35.267777 kubelet[2630]: E0213 19:19:35.267751 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8509c778c3294467a0ad8b8059bdecd145729001ee04e3c09d32d7f0d60b9468\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-s8264" Feb 13 19:19:35.267865 kubelet[2630]: E0213 19:19:35.267796 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-6f6b679f8f-s8264_kube-system(f2fdfd8a-780b-4320-8f24-ca003c5d0c77)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-6f6b679f8f-s8264_kube-system(f2fdfd8a-780b-4320-8f24-ca003c5d0c77)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"8509c778c3294467a0ad8b8059bdecd145729001ee04e3c09d32d7f0d60b9468\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-6f6b679f8f-s8264" podUID="f2fdfd8a-780b-4320-8f24-ca003c5d0c77" Feb 13 19:19:35.397089 systemd[1]: run-netns-cni\x2dd9132d0b\x2d5989\x2d77b9\x2df2d0\x2d67280498dc8f.mount: Deactivated successfully. Feb 13 19:19:35.397235 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-fa92388357966f3e183c59506d6cfde3e38d36deede0f30821c07e22d4146a49-shm.mount: Deactivated successfully. Feb 13 19:19:35.397344 systemd[1]: run-netns-cni\x2d181e931e\x2dc951\x2dd8c8\x2dfa2f\x2dc8752b0cdd69.mount: Deactivated successfully. Feb 13 19:19:35.397432 systemd[1]: run-netns-cni\x2da357a058\x2d03da\x2da946\x2da18c\x2d88cf20951171.mount: Deactivated successfully. Feb 13 19:19:35.397519 systemd[1]: run-netns-cni\x2d6e81f692\x2d352e\x2d6641\x2ddb40\x2da466e4cc7720.mount: Deactivated successfully. Feb 13 19:19:35.397618 systemd[1]: run-netns-cni\x2d951d9e20\x2da145\x2d061f\x2df247\x2d3409069116fc.mount: Deactivated successfully. Feb 13 19:19:35.397711 systemd[1]: run-netns-cni\x2d7d16551c\x2dcdaa\x2d3e08\x2db3b0\x2d46e5c00e21ca.mount: Deactivated successfully. Feb 13 19:19:36.103872 kubelet[2630]: I0213 19:19:36.103812 2630 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5e6531fa4d44fae3b14be640ca5608599c77b930300d189004cbcc636c08e1b" Feb 13 19:19:36.104476 containerd[1515]: time="2025-02-13T19:19:36.104441954Z" level=info msg="StopPodSandbox for \"e5e6531fa4d44fae3b14be640ca5608599c77b930300d189004cbcc636c08e1b\"" Feb 13 19:19:36.106542 containerd[1515]: time="2025-02-13T19:19:36.104727021Z" level=info msg="Ensure that sandbox e5e6531fa4d44fae3b14be640ca5608599c77b930300d189004cbcc636c08e1b in task-service has been cleanup successfully" Feb 13 19:19:36.107816 containerd[1515]: time="2025-02-13T19:19:36.107564492Z" level=info msg="TearDown network for sandbox \"e5e6531fa4d44fae3b14be640ca5608599c77b930300d189004cbcc636c08e1b\" successfully" Feb 13 19:19:36.107816 containerd[1515]: time="2025-02-13T19:19:36.107750352Z" level=info msg="StopPodSandbox for \"e5e6531fa4d44fae3b14be640ca5608599c77b930300d189004cbcc636c08e1b\" returns successfully" Feb 13 19:19:36.108265 containerd[1515]: time="2025-02-13T19:19:36.108235205Z" level=info msg="StopPodSandbox for \"4304f46dfb22ea4aced73143fa3f59c4ed426a670211c8559cc8ed545f1b3804\"" Feb 13 19:19:36.108416 containerd[1515]: time="2025-02-13T19:19:36.108337868Z" level=info msg="TearDown network for sandbox \"4304f46dfb22ea4aced73143fa3f59c4ed426a670211c8559cc8ed545f1b3804\" successfully" Feb 13 19:19:36.108416 containerd[1515]: time="2025-02-13T19:19:36.108353227Z" level=info msg="StopPodSandbox for \"4304f46dfb22ea4aced73143fa3f59c4ed426a670211c8559cc8ed545f1b3804\" returns successfully" Feb 13 19:19:36.108874 systemd[1]: run-netns-cni\x2dee629b20\x2da8e1\x2d95d8\x2d5d53\x2dbbd5551b1271.mount: Deactivated successfully. Feb 13 19:19:36.110498 kubelet[2630]: I0213 19:19:36.110097 2630 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8509c778c3294467a0ad8b8059bdecd145729001ee04e3c09d32d7f0d60b9468" Feb 13 19:19:36.110557 containerd[1515]: time="2025-02-13T19:19:36.110312796Z" level=info msg="StopPodSandbox for \"aa373a8af75a289d061969e5a22e647e175a6a451ba7c421e5c44fa363ec90f8\"" Feb 13 19:19:36.110557 containerd[1515]: time="2025-02-13T19:19:36.110387677Z" level=info msg="TearDown network for sandbox \"aa373a8af75a289d061969e5a22e647e175a6a451ba7c421e5c44fa363ec90f8\" successfully" Feb 13 19:19:36.110557 containerd[1515]: time="2025-02-13T19:19:36.110396553Z" level=info msg="StopPodSandbox for \"aa373a8af75a289d061969e5a22e647e175a6a451ba7c421e5c44fa363ec90f8\" returns successfully" Feb 13 19:19:36.111512 containerd[1515]: time="2025-02-13T19:19:36.110968069Z" level=info msg="StopPodSandbox for \"8509c778c3294467a0ad8b8059bdecd145729001ee04e3c09d32d7f0d60b9468\"" Feb 13 19:19:36.111512 containerd[1515]: time="2025-02-13T19:19:36.111211047Z" level=info msg="Ensure that sandbox 8509c778c3294467a0ad8b8059bdecd145729001ee04e3c09d32d7f0d60b9468 in task-service has been cleanup successfully" Feb 13 19:19:36.115307 containerd[1515]: time="2025-02-13T19:19:36.112253479Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-664f6f5454-mnbf7,Uid:4e40d638-ead9-4cd4-859c-b3835db75e0c,Namespace:calico-system,Attempt:3,}" Feb 13 19:19:36.115536 containerd[1515]: time="2025-02-13T19:19:36.115513226Z" level=info msg="TearDown network for sandbox \"8509c778c3294467a0ad8b8059bdecd145729001ee04e3c09d32d7f0d60b9468\" successfully" Feb 13 19:19:36.115785 containerd[1515]: time="2025-02-13T19:19:36.115697432Z" level=info msg="StopPodSandbox for \"8509c778c3294467a0ad8b8059bdecd145729001ee04e3c09d32d7f0d60b9468\" returns successfully" Feb 13 19:19:36.116802 containerd[1515]: time="2025-02-13T19:19:36.116783938Z" level=info msg="StopPodSandbox for \"706e09ab546bf3d89157cd9a2c32b16d198260f6686345817b1acf65f680fdd2\"" Feb 13 19:19:36.117097 containerd[1515]: time="2025-02-13T19:19:36.117037004Z" level=info msg="TearDown network for sandbox \"706e09ab546bf3d89157cd9a2c32b16d198260f6686345817b1acf65f680fdd2\" successfully" Feb 13 19:19:36.117097 containerd[1515]: time="2025-02-13T19:19:36.117050580Z" level=info msg="StopPodSandbox for \"706e09ab546bf3d89157cd9a2c32b16d198260f6686345817b1acf65f680fdd2\" returns successfully" Feb 13 19:19:36.118525 kubelet[2630]: I0213 19:19:36.117406 2630 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="28a752266e9ffa0b0f35649f5eba403b238d216dc8ce20a8fe3622cc28799217" Feb 13 19:19:36.117680 systemd[1]: run-netns-cni\x2da18dabf8\x2d320b\x2df139\x2d306a\x2da18c84434463.mount: Deactivated successfully. Feb 13 19:19:36.119252 containerd[1515]: time="2025-02-13T19:19:36.119217138Z" level=info msg="StopPodSandbox for \"e2995529ca797998e1cf161d883bb684a3464a8a8d57602758f8705313f0d601\"" Feb 13 19:19:36.119524 containerd[1515]: time="2025-02-13T19:19:36.119493268Z" level=info msg="TearDown network for sandbox \"e2995529ca797998e1cf161d883bb684a3464a8a8d57602758f8705313f0d601\" successfully" Feb 13 19:19:36.119812 containerd[1515]: time="2025-02-13T19:19:36.119773736Z" level=info msg="StopPodSandbox for \"e2995529ca797998e1cf161d883bb684a3464a8a8d57602758f8705313f0d601\" returns successfully" Feb 13 19:19:36.121147 containerd[1515]: time="2025-02-13T19:19:36.120702244Z" level=info msg="StopPodSandbox for \"28a752266e9ffa0b0f35649f5eba403b238d216dc8ce20a8fe3622cc28799217\"" Feb 13 19:19:36.121147 containerd[1515]: time="2025-02-13T19:19:36.120886941Z" level=info msg="Ensure that sandbox 28a752266e9ffa0b0f35649f5eba403b238d216dc8ce20a8fe3622cc28799217 in task-service has been cleanup successfully" Feb 13 19:19:36.123573 kubelet[2630]: E0213 19:19:36.123541 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:36.125643 containerd[1515]: time="2025-02-13T19:19:36.124591264Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-s8264,Uid:f2fdfd8a-780b-4320-8f24-ca003c5d0c77,Namespace:kube-system,Attempt:3,}" Feb 13 19:19:36.125708 kubelet[2630]: I0213 19:19:36.125457 2630 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d811fd694e600e47efc43338f2c97ee7c9f512c6106b43a1a049dbc81724231" Feb 13 19:19:36.126243 containerd[1515]: time="2025-02-13T19:19:36.126218758Z" level=info msg="StopPodSandbox for \"1d811fd694e600e47efc43338f2c97ee7c9f512c6106b43a1a049dbc81724231\"" Feb 13 19:19:36.126539 containerd[1515]: time="2025-02-13T19:19:36.126512391Z" level=info msg="Ensure that sandbox 1d811fd694e600e47efc43338f2c97ee7c9f512c6106b43a1a049dbc81724231 in task-service has been cleanup successfully" Feb 13 19:19:36.127626 containerd[1515]: time="2025-02-13T19:19:36.127606791Z" level=info msg="TearDown network for sandbox \"1d811fd694e600e47efc43338f2c97ee7c9f512c6106b43a1a049dbc81724231\" successfully" Feb 13 19:19:36.127696 containerd[1515]: time="2025-02-13T19:19:36.127683445Z" level=info msg="StopPodSandbox for \"1d811fd694e600e47efc43338f2c97ee7c9f512c6106b43a1a049dbc81724231\" returns successfully" Feb 13 19:19:36.129807 containerd[1515]: time="2025-02-13T19:19:36.129744505Z" level=info msg="StopPodSandbox for \"fa92388357966f3e183c59506d6cfde3e38d36deede0f30821c07e22d4146a49\"" Feb 13 19:19:36.129949 containerd[1515]: time="2025-02-13T19:19:36.129850776Z" level=info msg="TearDown network for sandbox \"fa92388357966f3e183c59506d6cfde3e38d36deede0f30821c07e22d4146a49\" successfully" Feb 13 19:19:36.129949 containerd[1515]: time="2025-02-13T19:19:36.129861115Z" level=info msg="StopPodSandbox for \"fa92388357966f3e183c59506d6cfde3e38d36deede0f30821c07e22d4146a49\" returns successfully" Feb 13 19:19:36.130352 containerd[1515]: time="2025-02-13T19:19:36.130315690Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-drlnv,Uid:84096616-cf79-421c-9e94-8cbcf0a204a5,Namespace:calico-system,Attempt:2,}" Feb 13 19:19:36.131222 kubelet[2630]: I0213 19:19:36.131143 2630 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="787aa28bb52b7073a96ced7755cf59d24334348abd3a51ce97858110663e1200" Feb 13 19:19:36.131597 systemd[1]: run-netns-cni\x2d151f026c\x2d3584\x2dba63\x2d5fdd\x2d4e53d97eb912.mount: Deactivated successfully. Feb 13 19:19:36.133097 containerd[1515]: time="2025-02-13T19:19:36.133060288Z" level=info msg="StopPodSandbox for \"787aa28bb52b7073a96ced7755cf59d24334348abd3a51ce97858110663e1200\"" Feb 13 19:19:36.133822 containerd[1515]: time="2025-02-13T19:19:36.133611565Z" level=info msg="Ensure that sandbox 787aa28bb52b7073a96ced7755cf59d24334348abd3a51ce97858110663e1200 in task-service has been cleanup successfully" Feb 13 19:19:36.133944 containerd[1515]: time="2025-02-13T19:19:36.133925897Z" level=info msg="TearDown network for sandbox \"787aa28bb52b7073a96ced7755cf59d24334348abd3a51ce97858110663e1200\" successfully" Feb 13 19:19:36.134012 containerd[1515]: time="2025-02-13T19:19:36.133997442Z" level=info msg="StopPodSandbox for \"787aa28bb52b7073a96ced7755cf59d24334348abd3a51ce97858110663e1200\" returns successfully" Feb 13 19:19:36.136682 containerd[1515]: time="2025-02-13T19:19:36.136648181Z" level=info msg="StopPodSandbox for \"b4101238369440d1fc253b3e442bebf940a9f5fa609c91f1252b6dcea502c067\"" Feb 13 19:19:36.136786 containerd[1515]: time="2025-02-13T19:19:36.136767446Z" level=info msg="TearDown network for sandbox \"b4101238369440d1fc253b3e442bebf940a9f5fa609c91f1252b6dcea502c067\" successfully" Feb 13 19:19:36.136831 containerd[1515]: time="2025-02-13T19:19:36.136784227Z" level=info msg="StopPodSandbox for \"b4101238369440d1fc253b3e442bebf940a9f5fa609c91f1252b6dcea502c067\" returns successfully" Feb 13 19:19:36.137039 systemd[1]: run-netns-cni\x2d0bb5fc93\x2df23f\x2d6c08\x2d72b2\x2d3449a47171e2.mount: Deactivated successfully. Feb 13 19:19:36.137309 kubelet[2630]: I0213 19:19:36.137052 2630 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="154203fe3c4258e99eeb24eefabf42a73b9226503cd839391e62422056e09db6" Feb 13 19:19:36.139868 containerd[1515]: time="2025-02-13T19:19:36.139509418Z" level=info msg="StopPodSandbox for \"154203fe3c4258e99eeb24eefabf42a73b9226503cd839391e62422056e09db6\"" Feb 13 19:19:36.139868 containerd[1515]: time="2025-02-13T19:19:36.139556666Z" level=info msg="StopPodSandbox for \"f7214e7c2d9f81df76f610a79140dcb0085636bf81f33d1eb139eb52cb12a6af\"" Feb 13 19:19:36.139868 containerd[1515]: time="2025-02-13T19:19:36.139652577Z" level=info msg="TearDown network for sandbox \"f7214e7c2d9f81df76f610a79140dcb0085636bf81f33d1eb139eb52cb12a6af\" successfully" Feb 13 19:19:36.139868 containerd[1515]: time="2025-02-13T19:19:36.139665842Z" level=info msg="StopPodSandbox for \"f7214e7c2d9f81df76f610a79140dcb0085636bf81f33d1eb139eb52cb12a6af\" returns successfully" Feb 13 19:19:36.139868 containerd[1515]: time="2025-02-13T19:19:36.139704024Z" level=info msg="Ensure that sandbox 154203fe3c4258e99eeb24eefabf42a73b9226503cd839391e62422056e09db6 in task-service has been cleanup successfully" Feb 13 19:19:36.141662 containerd[1515]: time="2025-02-13T19:19:36.141005383Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86475b859c-kwlfl,Uid:f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1,Namespace:calico-apiserver,Attempt:3,}" Feb 13 19:19:36.146876 containerd[1515]: time="2025-02-13T19:19:36.146269113Z" level=info msg="TearDown network for sandbox \"154203fe3c4258e99eeb24eefabf42a73b9226503cd839391e62422056e09db6\" successfully" Feb 13 19:19:36.146876 containerd[1515]: time="2025-02-13T19:19:36.146297726Z" level=info msg="StopPodSandbox for \"154203fe3c4258e99eeb24eefabf42a73b9226503cd839391e62422056e09db6\" returns successfully" Feb 13 19:19:36.146876 containerd[1515]: time="2025-02-13T19:19:36.146657875Z" level=info msg="StopPodSandbox for \"7d827c446a1e808002e0f19943157ab97bcc163d150e79960814ed79c5b4fbf3\"" Feb 13 19:19:36.147346 containerd[1515]: time="2025-02-13T19:19:36.146763763Z" level=info msg="TearDown network for sandbox \"7d827c446a1e808002e0f19943157ab97bcc163d150e79960814ed79c5b4fbf3\" successfully" Feb 13 19:19:36.147494 containerd[1515]: time="2025-02-13T19:19:36.147409399Z" level=info msg="StopPodSandbox for \"7d827c446a1e808002e0f19943157ab97bcc163d150e79960814ed79c5b4fbf3\" returns successfully" Feb 13 19:19:36.147856 containerd[1515]: time="2025-02-13T19:19:36.147832315Z" level=info msg="StopPodSandbox for \"d195bcf8f9551fdb26f6b42c8b0e4b63d6953f1bbdc1c4b497262fc12be34248\"" Feb 13 19:19:36.147973 containerd[1515]: time="2025-02-13T19:19:36.147946370Z" level=info msg="TearDown network for sandbox \"d195bcf8f9551fdb26f6b42c8b0e4b63d6953f1bbdc1c4b497262fc12be34248\" successfully" Feb 13 19:19:36.147973 containerd[1515]: time="2025-02-13T19:19:36.147966528Z" level=info msg="StopPodSandbox for \"d195bcf8f9551fdb26f6b42c8b0e4b63d6953f1bbdc1c4b497262fc12be34248\" returns successfully" Feb 13 19:19:36.148434 containerd[1515]: time="2025-02-13T19:19:36.148410273Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86475b859c-swqrn,Uid:5a39efb7-aca1-4893-a36f-840c3c504ece,Namespace:calico-apiserver,Attempt:3,}" Feb 13 19:19:36.154441 containerd[1515]: time="2025-02-13T19:19:36.154403295Z" level=info msg="TearDown network for sandbox \"28a752266e9ffa0b0f35649f5eba403b238d216dc8ce20a8fe3622cc28799217\" successfully" Feb 13 19:19:36.154441 containerd[1515]: time="2025-02-13T19:19:36.154422892Z" level=info msg="StopPodSandbox for \"28a752266e9ffa0b0f35649f5eba403b238d216dc8ce20a8fe3622cc28799217\" returns successfully" Feb 13 19:19:36.154872 containerd[1515]: time="2025-02-13T19:19:36.154688091Z" level=info msg="StopPodSandbox for \"a20bc39d31c8f248a56fc8b89d2d3ce62836f574c54b75502431e23bb908b697\"" Feb 13 19:19:36.154872 containerd[1515]: time="2025-02-13T19:19:36.154761971Z" level=info msg="TearDown network for sandbox \"a20bc39d31c8f248a56fc8b89d2d3ce62836f574c54b75502431e23bb908b697\" successfully" Feb 13 19:19:36.154872 containerd[1515]: time="2025-02-13T19:19:36.154770116Z" level=info msg="StopPodSandbox for \"a20bc39d31c8f248a56fc8b89d2d3ce62836f574c54b75502431e23bb908b697\" returns successfully" Feb 13 19:19:36.155477 containerd[1515]: time="2025-02-13T19:19:36.154961816Z" level=info msg="StopPodSandbox for \"f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7\"" Feb 13 19:19:36.155477 containerd[1515]: time="2025-02-13T19:19:36.155032290Z" level=info msg="TearDown network for sandbox \"f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7\" successfully" Feb 13 19:19:36.155477 containerd[1515]: time="2025-02-13T19:19:36.155040595Z" level=info msg="StopPodSandbox for \"f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7\" returns successfully" Feb 13 19:19:36.155611 kubelet[2630]: E0213 19:19:36.155280 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:36.156307 containerd[1515]: time="2025-02-13T19:19:36.155958042Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-frtk4,Uid:f48c8839-9a17-43c4-a0d5-f9e8d9767814,Namespace:kube-system,Attempt:3,}" Feb 13 19:19:36.394052 systemd[1]: run-netns-cni\x2de69d7695\x2d4c7d\x2df142\x2dd560\x2d4d5f14897bb9.mount: Deactivated successfully. Feb 13 19:19:36.394168 systemd[1]: run-netns-cni\x2d666b613e\x2d1aff\x2d13bf\x2d5af2\x2d34aa42488a47.mount: Deactivated successfully. Feb 13 19:19:36.801725 containerd[1515]: time="2025-02-13T19:19:36.801682070Z" level=error msg="Failed to destroy network for sandbox \"c03a59deff35e415f0920de8fd1c0462db72392ea452d0eafd388e65b284eb39\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:36.804104 containerd[1515]: time="2025-02-13T19:19:36.804035801Z" level=error msg="encountered an error cleaning up failed sandbox \"c03a59deff35e415f0920de8fd1c0462db72392ea452d0eafd388e65b284eb39\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:36.804690 containerd[1515]: time="2025-02-13T19:19:36.804668633Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86475b859c-kwlfl,Uid:f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1,Namespace:calico-apiserver,Attempt:3,} failed, error" error="failed to setup network for sandbox \"c03a59deff35e415f0920de8fd1c0462db72392ea452d0eafd388e65b284eb39\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:36.806298 kubelet[2630]: E0213 19:19:36.805354 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c03a59deff35e415f0920de8fd1c0462db72392ea452d0eafd388e65b284eb39\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:36.806298 kubelet[2630]: E0213 19:19:36.805415 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c03a59deff35e415f0920de8fd1c0462db72392ea452d0eafd388e65b284eb39\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86475b859c-kwlfl" Feb 13 19:19:36.806298 kubelet[2630]: E0213 19:19:36.805437 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c03a59deff35e415f0920de8fd1c0462db72392ea452d0eafd388e65b284eb39\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86475b859c-kwlfl" Feb 13 19:19:36.806410 kubelet[2630]: E0213 19:19:36.805476 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-86475b859c-kwlfl_calico-apiserver(f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-86475b859c-kwlfl_calico-apiserver(f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"c03a59deff35e415f0920de8fd1c0462db72392ea452d0eafd388e65b284eb39\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-86475b859c-kwlfl" podUID="f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1" Feb 13 19:19:36.818530 containerd[1515]: time="2025-02-13T19:19:36.818468591Z" level=error msg="Failed to destroy network for sandbox \"216f8481f6c55f7a146dde6ed21a0f30b75cfe7059570683b2fb520eadf70207\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:36.819099 containerd[1515]: time="2025-02-13T19:19:36.819076344Z" level=error msg="encountered an error cleaning up failed sandbox \"216f8481f6c55f7a146dde6ed21a0f30b75cfe7059570683b2fb520eadf70207\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:36.819253 containerd[1515]: time="2025-02-13T19:19:36.819233260Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-drlnv,Uid:84096616-cf79-421c-9e94-8cbcf0a204a5,Namespace:calico-system,Attempt:2,} failed, error" error="failed to setup network for sandbox \"216f8481f6c55f7a146dde6ed21a0f30b75cfe7059570683b2fb520eadf70207\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:36.819896 kubelet[2630]: E0213 19:19:36.819526 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"216f8481f6c55f7a146dde6ed21a0f30b75cfe7059570683b2fb520eadf70207\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:36.819896 kubelet[2630]: E0213 19:19:36.819596 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"216f8481f6c55f7a146dde6ed21a0f30b75cfe7059570683b2fb520eadf70207\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-drlnv" Feb 13 19:19:36.819896 kubelet[2630]: E0213 19:19:36.819616 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"216f8481f6c55f7a146dde6ed21a0f30b75cfe7059570683b2fb520eadf70207\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-drlnv" Feb 13 19:19:36.820006 kubelet[2630]: E0213 19:19:36.819662 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-drlnv_calico-system(84096616-cf79-421c-9e94-8cbcf0a204a5)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-drlnv_calico-system(84096616-cf79-421c-9e94-8cbcf0a204a5)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"216f8481f6c55f7a146dde6ed21a0f30b75cfe7059570683b2fb520eadf70207\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-drlnv" podUID="84096616-cf79-421c-9e94-8cbcf0a204a5" Feb 13 19:19:36.833702 containerd[1515]: time="2025-02-13T19:19:36.833634119Z" level=error msg="Failed to destroy network for sandbox \"334cbee9ef51499256254b9201c8014a8e1e8be612d9ae143735bc19f2a5f6a4\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:36.834288 containerd[1515]: time="2025-02-13T19:19:36.834268003Z" level=error msg="encountered an error cleaning up failed sandbox \"334cbee9ef51499256254b9201c8014a8e1e8be612d9ae143735bc19f2a5f6a4\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:36.834412 containerd[1515]: time="2025-02-13T19:19:36.834394321Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-664f6f5454-mnbf7,Uid:4e40d638-ead9-4cd4-859c-b3835db75e0c,Namespace:calico-system,Attempt:3,} failed, error" error="failed to setup network for sandbox \"334cbee9ef51499256254b9201c8014a8e1e8be612d9ae143735bc19f2a5f6a4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:36.834751 kubelet[2630]: E0213 19:19:36.834716 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"334cbee9ef51499256254b9201c8014a8e1e8be612d9ae143735bc19f2a5f6a4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:36.834962 kubelet[2630]: E0213 19:19:36.834856 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"334cbee9ef51499256254b9201c8014a8e1e8be612d9ae143735bc19f2a5f6a4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-664f6f5454-mnbf7" Feb 13 19:19:36.834962 kubelet[2630]: E0213 19:19:36.834883 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"334cbee9ef51499256254b9201c8014a8e1e8be612d9ae143735bc19f2a5f6a4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-664f6f5454-mnbf7" Feb 13 19:19:36.834962 kubelet[2630]: E0213 19:19:36.834926 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-664f6f5454-mnbf7_calico-system(4e40d638-ead9-4cd4-859c-b3835db75e0c)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-664f6f5454-mnbf7_calico-system(4e40d638-ead9-4cd4-859c-b3835db75e0c)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"334cbee9ef51499256254b9201c8014a8e1e8be612d9ae143735bc19f2a5f6a4\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-664f6f5454-mnbf7" podUID="4e40d638-ead9-4cd4-859c-b3835db75e0c" Feb 13 19:19:36.838572 containerd[1515]: time="2025-02-13T19:19:36.838501342Z" level=error msg="Failed to destroy network for sandbox \"ee0d44b3886e78bb5043872859e92bfd2620af442c433a402780bc0453826443\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:36.839217 containerd[1515]: time="2025-02-13T19:19:36.839144022Z" level=error msg="encountered an error cleaning up failed sandbox \"ee0d44b3886e78bb5043872859e92bfd2620af442c433a402780bc0453826443\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:36.839383 containerd[1515]: time="2025-02-13T19:19:36.839324361Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-frtk4,Uid:f48c8839-9a17-43c4-a0d5-f9e8d9767814,Namespace:kube-system,Attempt:3,} failed, error" error="failed to setup network for sandbox \"ee0d44b3886e78bb5043872859e92bfd2620af442c433a402780bc0453826443\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:36.839678 kubelet[2630]: E0213 19:19:36.839637 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ee0d44b3886e78bb5043872859e92bfd2620af442c433a402780bc0453826443\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:36.839731 kubelet[2630]: E0213 19:19:36.839694 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ee0d44b3886e78bb5043872859e92bfd2620af442c433a402780bc0453826443\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-frtk4" Feb 13 19:19:36.839731 kubelet[2630]: E0213 19:19:36.839712 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ee0d44b3886e78bb5043872859e92bfd2620af442c433a402780bc0453826443\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-frtk4" Feb 13 19:19:36.839784 kubelet[2630]: E0213 19:19:36.839764 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-6f6b679f8f-frtk4_kube-system(f48c8839-9a17-43c4-a0d5-f9e8d9767814)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-6f6b679f8f-frtk4_kube-system(f48c8839-9a17-43c4-a0d5-f9e8d9767814)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"ee0d44b3886e78bb5043872859e92bfd2620af442c433a402780bc0453826443\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-6f6b679f8f-frtk4" podUID="f48c8839-9a17-43c4-a0d5-f9e8d9767814" Feb 13 19:19:36.841097 containerd[1515]: time="2025-02-13T19:19:36.840749023Z" level=error msg="Failed to destroy network for sandbox \"fd762dfcc2c780c334b7852f3da6760c5d53d6cb6d243b5ae4cf1085971b8f57\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:36.841880 containerd[1515]: time="2025-02-13T19:19:36.841859824Z" level=error msg="encountered an error cleaning up failed sandbox \"fd762dfcc2c780c334b7852f3da6760c5d53d6cb6d243b5ae4cf1085971b8f57\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:36.842013 containerd[1515]: time="2025-02-13T19:19:36.841974160Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-s8264,Uid:f2fdfd8a-780b-4320-8f24-ca003c5d0c77,Namespace:kube-system,Attempt:3,} failed, error" error="failed to setup network for sandbox \"fd762dfcc2c780c334b7852f3da6760c5d53d6cb6d243b5ae4cf1085971b8f57\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:36.842811 kubelet[2630]: E0213 19:19:36.842781 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fd762dfcc2c780c334b7852f3da6760c5d53d6cb6d243b5ae4cf1085971b8f57\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:36.842867 kubelet[2630]: E0213 19:19:36.842818 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fd762dfcc2c780c334b7852f3da6760c5d53d6cb6d243b5ae4cf1085971b8f57\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-s8264" Feb 13 19:19:36.842867 kubelet[2630]: E0213 19:19:36.842835 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fd762dfcc2c780c334b7852f3da6760c5d53d6cb6d243b5ae4cf1085971b8f57\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-s8264" Feb 13 19:19:36.842913 kubelet[2630]: E0213 19:19:36.842863 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-6f6b679f8f-s8264_kube-system(f2fdfd8a-780b-4320-8f24-ca003c5d0c77)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-6f6b679f8f-s8264_kube-system(f2fdfd8a-780b-4320-8f24-ca003c5d0c77)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"fd762dfcc2c780c334b7852f3da6760c5d53d6cb6d243b5ae4cf1085971b8f57\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-6f6b679f8f-s8264" podUID="f2fdfd8a-780b-4320-8f24-ca003c5d0c77" Feb 13 19:19:36.843094 containerd[1515]: time="2025-02-13T19:19:36.843069541Z" level=error msg="Failed to destroy network for sandbox \"8a568a618a4c0eab0195822f2ccafce3787a78e075bf7c3f5b50c840e692188a\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:36.843626 containerd[1515]: time="2025-02-13T19:19:36.843584661Z" level=error msg="encountered an error cleaning up failed sandbox \"8a568a618a4c0eab0195822f2ccafce3787a78e075bf7c3f5b50c840e692188a\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:36.843626 containerd[1515]: time="2025-02-13T19:19:36.843621209Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86475b859c-swqrn,Uid:5a39efb7-aca1-4893-a36f-840c3c504ece,Namespace:calico-apiserver,Attempt:3,} failed, error" error="failed to setup network for sandbox \"8a568a618a4c0eab0195822f2ccafce3787a78e075bf7c3f5b50c840e692188a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:36.843803 kubelet[2630]: E0213 19:19:36.843766 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8a568a618a4c0eab0195822f2ccafce3787a78e075bf7c3f5b50c840e692188a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:36.843844 kubelet[2630]: E0213 19:19:36.843816 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8a568a618a4c0eab0195822f2ccafce3787a78e075bf7c3f5b50c840e692188a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86475b859c-swqrn" Feb 13 19:19:36.843844 kubelet[2630]: E0213 19:19:36.843836 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8a568a618a4c0eab0195822f2ccafce3787a78e075bf7c3f5b50c840e692188a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86475b859c-swqrn" Feb 13 19:19:36.843913 kubelet[2630]: E0213 19:19:36.843876 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-86475b859c-swqrn_calico-apiserver(5a39efb7-aca1-4893-a36f-840c3c504ece)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-86475b859c-swqrn_calico-apiserver(5a39efb7-aca1-4893-a36f-840c3c504ece)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"8a568a618a4c0eab0195822f2ccafce3787a78e075bf7c3f5b50c840e692188a\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-86475b859c-swqrn" podUID="5a39efb7-aca1-4893-a36f-840c3c504ece" Feb 13 19:19:37.142960 kubelet[2630]: I0213 19:19:37.141876 2630 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd762dfcc2c780c334b7852f3da6760c5d53d6cb6d243b5ae4cf1085971b8f57" Feb 13 19:19:37.143445 containerd[1515]: time="2025-02-13T19:19:37.142704674Z" level=info msg="StopPodSandbox for \"fd762dfcc2c780c334b7852f3da6760c5d53d6cb6d243b5ae4cf1085971b8f57\"" Feb 13 19:19:37.143445 containerd[1515]: time="2025-02-13T19:19:37.143207750Z" level=info msg="Ensure that sandbox fd762dfcc2c780c334b7852f3da6760c5d53d6cb6d243b5ae4cf1085971b8f57 in task-service has been cleanup successfully" Feb 13 19:19:37.143445 containerd[1515]: time="2025-02-13T19:19:37.143404952Z" level=info msg="TearDown network for sandbox \"fd762dfcc2c780c334b7852f3da6760c5d53d6cb6d243b5ae4cf1085971b8f57\" successfully" Feb 13 19:19:37.143445 containerd[1515]: time="2025-02-13T19:19:37.143416093Z" level=info msg="StopPodSandbox for \"fd762dfcc2c780c334b7852f3da6760c5d53d6cb6d243b5ae4cf1085971b8f57\" returns successfully" Feb 13 19:19:37.144428 containerd[1515]: time="2025-02-13T19:19:37.144404392Z" level=info msg="StopPodSandbox for \"8509c778c3294467a0ad8b8059bdecd145729001ee04e3c09d32d7f0d60b9468\"" Feb 13 19:19:37.144490 containerd[1515]: time="2025-02-13T19:19:37.144479374Z" level=info msg="TearDown network for sandbox \"8509c778c3294467a0ad8b8059bdecd145729001ee04e3c09d32d7f0d60b9468\" successfully" Feb 13 19:19:37.144517 containerd[1515]: time="2025-02-13T19:19:37.144489413Z" level=info msg="StopPodSandbox for \"8509c778c3294467a0ad8b8059bdecd145729001ee04e3c09d32d7f0d60b9468\" returns successfully" Feb 13 19:19:37.144762 kubelet[2630]: I0213 19:19:37.144715 2630 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c03a59deff35e415f0920de8fd1c0462db72392ea452d0eafd388e65b284eb39" Feb 13 19:19:37.144919 containerd[1515]: time="2025-02-13T19:19:37.144724384Z" level=info msg="StopPodSandbox for \"706e09ab546bf3d89157cd9a2c32b16d198260f6686345817b1acf65f680fdd2\"" Feb 13 19:19:37.144919 containerd[1515]: time="2025-02-13T19:19:37.144818813Z" level=info msg="TearDown network for sandbox \"706e09ab546bf3d89157cd9a2c32b16d198260f6686345817b1acf65f680fdd2\" successfully" Feb 13 19:19:37.144919 containerd[1515]: time="2025-02-13T19:19:37.144828871Z" level=info msg="StopPodSandbox for \"706e09ab546bf3d89157cd9a2c32b16d198260f6686345817b1acf65f680fdd2\" returns successfully" Feb 13 19:19:37.145584 containerd[1515]: time="2025-02-13T19:19:37.145554207Z" level=info msg="StopPodSandbox for \"e2995529ca797998e1cf161d883bb684a3464a8a8d57602758f8705313f0d601\"" Feb 13 19:19:37.146100 containerd[1515]: time="2025-02-13T19:19:37.146046443Z" level=info msg="TearDown network for sandbox \"e2995529ca797998e1cf161d883bb684a3464a8a8d57602758f8705313f0d601\" successfully" Feb 13 19:19:37.146172 containerd[1515]: time="2025-02-13T19:19:37.146159406Z" level=info msg="StopPodSandbox for \"e2995529ca797998e1cf161d883bb684a3464a8a8d57602758f8705313f0d601\" returns successfully" Feb 13 19:19:37.146332 containerd[1515]: time="2025-02-13T19:19:37.145675104Z" level=info msg="StopPodSandbox for \"c03a59deff35e415f0920de8fd1c0462db72392ea452d0eafd388e65b284eb39\"" Feb 13 19:19:37.146535 containerd[1515]: time="2025-02-13T19:19:37.146503723Z" level=info msg="Ensure that sandbox c03a59deff35e415f0920de8fd1c0462db72392ea452d0eafd388e65b284eb39 in task-service has been cleanup successfully" Feb 13 19:19:37.146621 kubelet[2630]: E0213 19:19:37.146503 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:37.146805 containerd[1515]: time="2025-02-13T19:19:37.146786335Z" level=info msg="TearDown network for sandbox \"c03a59deff35e415f0920de8fd1c0462db72392ea452d0eafd388e65b284eb39\" successfully" Feb 13 19:19:37.146914 containerd[1515]: time="2025-02-13T19:19:37.146802385Z" level=info msg="StopPodSandbox for \"c03a59deff35e415f0920de8fd1c0462db72392ea452d0eafd388e65b284eb39\" returns successfully" Feb 13 19:19:37.146966 containerd[1515]: time="2025-02-13T19:19:37.146934294Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-s8264,Uid:f2fdfd8a-780b-4320-8f24-ca003c5d0c77,Namespace:kube-system,Attempt:4,}" Feb 13 19:19:37.147157 containerd[1515]: time="2025-02-13T19:19:37.147125855Z" level=info msg="StopPodSandbox for \"787aa28bb52b7073a96ced7755cf59d24334348abd3a51ce97858110663e1200\"" Feb 13 19:19:37.147265 containerd[1515]: time="2025-02-13T19:19:37.147242734Z" level=info msg="TearDown network for sandbox \"787aa28bb52b7073a96ced7755cf59d24334348abd3a51ce97858110663e1200\" successfully" Feb 13 19:19:37.147265 containerd[1515]: time="2025-02-13T19:19:37.147260688Z" level=info msg="StopPodSandbox for \"787aa28bb52b7073a96ced7755cf59d24334348abd3a51ce97858110663e1200\" returns successfully" Feb 13 19:19:37.147816 kubelet[2630]: I0213 19:19:37.147771 2630 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a568a618a4c0eab0195822f2ccafce3787a78e075bf7c3f5b50c840e692188a" Feb 13 19:19:37.148210 containerd[1515]: time="2025-02-13T19:19:37.148125516Z" level=info msg="StopPodSandbox for \"8a568a618a4c0eab0195822f2ccafce3787a78e075bf7c3f5b50c840e692188a\"" Feb 13 19:19:37.148568 containerd[1515]: time="2025-02-13T19:19:37.148235002Z" level=info msg="StopPodSandbox for \"b4101238369440d1fc253b3e442bebf940a9f5fa609c91f1252b6dcea502c067\"" Feb 13 19:19:37.148654 containerd[1515]: time="2025-02-13T19:19:37.148632220Z" level=info msg="TearDown network for sandbox \"b4101238369440d1fc253b3e442bebf940a9f5fa609c91f1252b6dcea502c067\" successfully" Feb 13 19:19:37.148654 containerd[1515]: time="2025-02-13T19:19:37.148650394Z" level=info msg="StopPodSandbox for \"b4101238369440d1fc253b3e442bebf940a9f5fa609c91f1252b6dcea502c067\" returns successfully" Feb 13 19:19:37.148737 containerd[1515]: time="2025-02-13T19:19:37.148719363Z" level=info msg="Ensure that sandbox 8a568a618a4c0eab0195822f2ccafce3787a78e075bf7c3f5b50c840e692188a in task-service has been cleanup successfully" Feb 13 19:19:37.148884 containerd[1515]: time="2025-02-13T19:19:37.148831515Z" level=info msg="StopPodSandbox for \"f7214e7c2d9f81df76f610a79140dcb0085636bf81f33d1eb139eb52cb12a6af\"" Feb 13 19:19:37.148934 containerd[1515]: time="2025-02-13T19:19:37.148916234Z" level=info msg="TearDown network for sandbox \"f7214e7c2d9f81df76f610a79140dcb0085636bf81f33d1eb139eb52cb12a6af\" successfully" Feb 13 19:19:37.148934 containerd[1515]: time="2025-02-13T19:19:37.148932455Z" level=info msg="StopPodSandbox for \"f7214e7c2d9f81df76f610a79140dcb0085636bf81f33d1eb139eb52cb12a6af\" returns successfully" Feb 13 19:19:37.149086 containerd[1515]: time="2025-02-13T19:19:37.149035588Z" level=info msg="TearDown network for sandbox \"8a568a618a4c0eab0195822f2ccafce3787a78e075bf7c3f5b50c840e692188a\" successfully" Feb 13 19:19:37.149086 containerd[1515]: time="2025-02-13T19:19:37.149050476Z" level=info msg="StopPodSandbox for \"8a568a618a4c0eab0195822f2ccafce3787a78e075bf7c3f5b50c840e692188a\" returns successfully" Feb 13 19:19:37.149388 containerd[1515]: time="2025-02-13T19:19:37.149359127Z" level=info msg="StopPodSandbox for \"154203fe3c4258e99eeb24eefabf42a73b9226503cd839391e62422056e09db6\"" Feb 13 19:19:37.149519 containerd[1515]: time="2025-02-13T19:19:37.149447263Z" level=info msg="TearDown network for sandbox \"154203fe3c4258e99eeb24eefabf42a73b9226503cd839391e62422056e09db6\" successfully" Feb 13 19:19:37.149519 containerd[1515]: time="2025-02-13T19:19:37.149463674Z" level=info msg="StopPodSandbox for \"154203fe3c4258e99eeb24eefabf42a73b9226503cd839391e62422056e09db6\" returns successfully" Feb 13 19:19:37.149686 containerd[1515]: time="2025-02-13T19:19:37.149656387Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86475b859c-kwlfl,Uid:f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1,Namespace:calico-apiserver,Attempt:4,}" Feb 13 19:19:37.150018 containerd[1515]: time="2025-02-13T19:19:37.149967552Z" level=info msg="StopPodSandbox for \"7d827c446a1e808002e0f19943157ab97bcc163d150e79960814ed79c5b4fbf3\"" Feb 13 19:19:37.150091 containerd[1515]: time="2025-02-13T19:19:37.150068102Z" level=info msg="TearDown network for sandbox \"7d827c446a1e808002e0f19943157ab97bcc163d150e79960814ed79c5b4fbf3\" successfully" Feb 13 19:19:37.150091 containerd[1515]: time="2025-02-13T19:19:37.150083471Z" level=info msg="StopPodSandbox for \"7d827c446a1e808002e0f19943157ab97bcc163d150e79960814ed79c5b4fbf3\" returns successfully" Feb 13 19:19:37.150354 kubelet[2630]: I0213 19:19:37.150288 2630 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="334cbee9ef51499256254b9201c8014a8e1e8be612d9ae143735bc19f2a5f6a4" Feb 13 19:19:37.150734 containerd[1515]: time="2025-02-13T19:19:37.150715570Z" level=info msg="StopPodSandbox for \"334cbee9ef51499256254b9201c8014a8e1e8be612d9ae143735bc19f2a5f6a4\"" Feb 13 19:19:37.150933 containerd[1515]: time="2025-02-13T19:19:37.150723806Z" level=info msg="StopPodSandbox for \"d195bcf8f9551fdb26f6b42c8b0e4b63d6953f1bbdc1c4b497262fc12be34248\"" Feb 13 19:19:37.150933 containerd[1515]: time="2025-02-13T19:19:37.150871044Z" level=info msg="TearDown network for sandbox \"d195bcf8f9551fdb26f6b42c8b0e4b63d6953f1bbdc1c4b497262fc12be34248\" successfully" Feb 13 19:19:37.150933 containerd[1515]: time="2025-02-13T19:19:37.150880712Z" level=info msg="StopPodSandbox for \"d195bcf8f9551fdb26f6b42c8b0e4b63d6953f1bbdc1c4b497262fc12be34248\" returns successfully" Feb 13 19:19:37.154010 containerd[1515]: time="2025-02-13T19:19:37.153988540Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86475b859c-swqrn,Uid:5a39efb7-aca1-4893-a36f-840c3c504ece,Namespace:calico-apiserver,Attempt:4,}" Feb 13 19:19:37.155659 kubelet[2630]: I0213 19:19:37.155620 2630 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee0d44b3886e78bb5043872859e92bfd2620af442c433a402780bc0453826443" Feb 13 19:19:37.156138 containerd[1515]: time="2025-02-13T19:19:37.156084895Z" level=info msg="StopPodSandbox for \"ee0d44b3886e78bb5043872859e92bfd2620af442c433a402780bc0453826443\"" Feb 13 19:19:37.156292 containerd[1515]: time="2025-02-13T19:19:37.156260437Z" level=info msg="Ensure that sandbox ee0d44b3886e78bb5043872859e92bfd2620af442c433a402780bc0453826443 in task-service has been cleanup successfully" Feb 13 19:19:37.156640 containerd[1515]: time="2025-02-13T19:19:37.156562896Z" level=info msg="TearDown network for sandbox \"ee0d44b3886e78bb5043872859e92bfd2620af442c433a402780bc0453826443\" successfully" Feb 13 19:19:37.156640 containerd[1515]: time="2025-02-13T19:19:37.156579767Z" level=info msg="StopPodSandbox for \"ee0d44b3886e78bb5043872859e92bfd2620af442c433a402780bc0453826443\" returns successfully" Feb 13 19:19:37.156797 containerd[1515]: time="2025-02-13T19:19:37.156760878Z" level=info msg="StopPodSandbox for \"28a752266e9ffa0b0f35649f5eba403b238d216dc8ce20a8fe3622cc28799217\"" Feb 13 19:19:37.156861 containerd[1515]: time="2025-02-13T19:19:37.156843283Z" level=info msg="TearDown network for sandbox \"28a752266e9ffa0b0f35649f5eba403b238d216dc8ce20a8fe3622cc28799217\" successfully" Feb 13 19:19:37.156861 containerd[1515]: time="2025-02-13T19:19:37.156857009Z" level=info msg="StopPodSandbox for \"28a752266e9ffa0b0f35649f5eba403b238d216dc8ce20a8fe3622cc28799217\" returns successfully" Feb 13 19:19:37.157114 containerd[1515]: time="2025-02-13T19:19:37.157087422Z" level=info msg="StopPodSandbox for \"a20bc39d31c8f248a56fc8b89d2d3ce62836f574c54b75502431e23bb908b697\"" Feb 13 19:19:37.158178 kubelet[2630]: I0213 19:19:37.157858 2630 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="216f8481f6c55f7a146dde6ed21a0f30b75cfe7059570683b2fb520eadf70207" Feb 13 19:19:37.394908 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-334cbee9ef51499256254b9201c8014a8e1e8be612d9ae143735bc19f2a5f6a4-shm.mount: Deactivated successfully. Feb 13 19:19:37.395036 systemd[1]: run-netns-cni\x2d81b54b9a\x2d31a3\x2d07a9\x2dd50a\x2dcbacc059d7f7.mount: Deactivated successfully. Feb 13 19:19:37.395114 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-fd762dfcc2c780c334b7852f3da6760c5d53d6cb6d243b5ae4cf1085971b8f57-shm.mount: Deactivated successfully. Feb 13 19:19:37.395210 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-216f8481f6c55f7a146dde6ed21a0f30b75cfe7059570683b2fb520eadf70207-shm.mount: Deactivated successfully. Feb 13 19:19:37.722075 containerd[1515]: time="2025-02-13T19:19:37.157356860Z" level=info msg="TearDown network for sandbox \"a20bc39d31c8f248a56fc8b89d2d3ce62836f574c54b75502431e23bb908b697\" successfully" Feb 13 19:19:37.722211 containerd[1515]: time="2025-02-13T19:19:37.722074825Z" level=info msg="StopPodSandbox for \"a20bc39d31c8f248a56fc8b89d2d3ce62836f574c54b75502431e23bb908b697\" returns successfully" Feb 13 19:19:37.722211 containerd[1515]: time="2025-02-13T19:19:37.158169068Z" level=info msg="StopPodSandbox for \"216f8481f6c55f7a146dde6ed21a0f30b75cfe7059570683b2fb520eadf70207\"" Feb 13 19:19:37.722867 containerd[1515]: time="2025-02-13T19:19:37.722840546Z" level=info msg="Ensure that sandbox 216f8481f6c55f7a146dde6ed21a0f30b75cfe7059570683b2fb520eadf70207 in task-service has been cleanup successfully" Feb 13 19:19:37.723463 containerd[1515]: time="2025-02-13T19:19:37.723200023Z" level=info msg="StopPodSandbox for \"f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7\"" Feb 13 19:19:37.723463 containerd[1515]: time="2025-02-13T19:19:37.723378118Z" level=info msg="TearDown network for sandbox \"f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7\" successfully" Feb 13 19:19:37.723463 containerd[1515]: time="2025-02-13T19:19:37.723397825Z" level=info msg="StopPodSandbox for \"f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7\" returns successfully" Feb 13 19:19:37.727869 containerd[1515]: time="2025-02-13T19:19:37.722035021Z" level=info msg="Ensure that sandbox 334cbee9ef51499256254b9201c8014a8e1e8be612d9ae143735bc19f2a5f6a4 in task-service has been cleanup successfully" Feb 13 19:19:37.727932 kubelet[2630]: E0213 19:19:37.727488 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:37.725161 systemd[1]: run-netns-cni\x2dc67debcc\x2ddf62\x2de853\x2d6740\x2dd08bbff46377.mount: Deactivated successfully. Feb 13 19:19:37.730966 containerd[1515]: time="2025-02-13T19:19:37.728288399Z" level=info msg="TearDown network for sandbox \"334cbee9ef51499256254b9201c8014a8e1e8be612d9ae143735bc19f2a5f6a4\" successfully" Feb 13 19:19:37.730966 containerd[1515]: time="2025-02-13T19:19:37.728310170Z" level=info msg="StopPodSandbox for \"334cbee9ef51499256254b9201c8014a8e1e8be612d9ae143735bc19f2a5f6a4\" returns successfully" Feb 13 19:19:37.730966 containerd[1515]: time="2025-02-13T19:19:37.728879431Z" level=info msg="StopPodSandbox for \"e5e6531fa4d44fae3b14be640ca5608599c77b930300d189004cbcc636c08e1b\"" Feb 13 19:19:37.730966 containerd[1515]: time="2025-02-13T19:19:37.729155631Z" level=info msg="TearDown network for sandbox \"e5e6531fa4d44fae3b14be640ca5608599c77b930300d189004cbcc636c08e1b\" successfully" Feb 13 19:19:37.730966 containerd[1515]: time="2025-02-13T19:19:37.729166341Z" level=info msg="StopPodSandbox for \"e5e6531fa4d44fae3b14be640ca5608599c77b930300d189004cbcc636c08e1b\" returns successfully" Feb 13 19:19:37.730966 containerd[1515]: time="2025-02-13T19:19:37.729453402Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-frtk4,Uid:f48c8839-9a17-43c4-a0d5-f9e8d9767814,Namespace:kube-system,Attempt:4,}" Feb 13 19:19:37.730161 systemd[1]: run-netns-cni\x2d0acb25cd\x2d8969\x2dd9c1\x2d5805\x2d4d82482b3190.mount: Deactivated successfully. Feb 13 19:19:37.739990 containerd[1515]: time="2025-02-13T19:19:37.739943174Z" level=info msg="TearDown network for sandbox \"216f8481f6c55f7a146dde6ed21a0f30b75cfe7059570683b2fb520eadf70207\" successfully" Feb 13 19:19:37.739990 containerd[1515]: time="2025-02-13T19:19:37.739977378Z" level=info msg="StopPodSandbox for \"216f8481f6c55f7a146dde6ed21a0f30b75cfe7059570683b2fb520eadf70207\" returns successfully" Feb 13 19:19:37.754561 containerd[1515]: time="2025-02-13T19:19:37.754412345Z" level=info msg="StopPodSandbox for \"1d811fd694e600e47efc43338f2c97ee7c9f512c6106b43a1a049dbc81724231\"" Feb 13 19:19:37.766656 containerd[1515]: time="2025-02-13T19:19:37.766635730Z" level=info msg="TearDown network for sandbox \"1d811fd694e600e47efc43338f2c97ee7c9f512c6106b43a1a049dbc81724231\" successfully" Feb 13 19:19:37.773771 containerd[1515]: time="2025-02-13T19:19:37.766719688Z" level=info msg="StopPodSandbox for \"1d811fd694e600e47efc43338f2c97ee7c9f512c6106b43a1a049dbc81724231\" returns successfully" Feb 13 19:19:37.773899 containerd[1515]: time="2025-02-13T19:19:37.773851370Z" level=info msg="StopPodSandbox for \"4304f46dfb22ea4aced73143fa3f59c4ed426a670211c8559cc8ed545f1b3804\"" Feb 13 19:19:37.774084 containerd[1515]: time="2025-02-13T19:19:37.774042670Z" level=info msg="TearDown network for sandbox \"4304f46dfb22ea4aced73143fa3f59c4ed426a670211c8559cc8ed545f1b3804\" successfully" Feb 13 19:19:37.774259 containerd[1515]: time="2025-02-13T19:19:37.774233499Z" level=info msg="StopPodSandbox for \"4304f46dfb22ea4aced73143fa3f59c4ed426a670211c8559cc8ed545f1b3804\" returns successfully" Feb 13 19:19:37.775063 containerd[1515]: time="2025-02-13T19:19:37.775012335Z" level=info msg="StopPodSandbox for \"aa373a8af75a289d061969e5a22e647e175a6a451ba7c421e5c44fa363ec90f8\"" Feb 13 19:19:37.775220 containerd[1515]: time="2025-02-13T19:19:37.775145416Z" level=info msg="TearDown network for sandbox \"aa373a8af75a289d061969e5a22e647e175a6a451ba7c421e5c44fa363ec90f8\" successfully" Feb 13 19:19:37.775247 containerd[1515]: time="2025-02-13T19:19:37.775217651Z" level=info msg="StopPodSandbox for \"aa373a8af75a289d061969e5a22e647e175a6a451ba7c421e5c44fa363ec90f8\" returns successfully" Feb 13 19:19:37.775529 containerd[1515]: time="2025-02-13T19:19:37.775287283Z" level=info msg="StopPodSandbox for \"fa92388357966f3e183c59506d6cfde3e38d36deede0f30821c07e22d4146a49\"" Feb 13 19:19:37.775529 containerd[1515]: time="2025-02-13T19:19:37.775377773Z" level=info msg="TearDown network for sandbox \"fa92388357966f3e183c59506d6cfde3e38d36deede0f30821c07e22d4146a49\" successfully" Feb 13 19:19:37.775529 containerd[1515]: time="2025-02-13T19:19:37.775389956Z" level=info msg="StopPodSandbox for \"fa92388357966f3e183c59506d6cfde3e38d36deede0f30821c07e22d4146a49\" returns successfully" Feb 13 19:19:37.775834 containerd[1515]: time="2025-02-13T19:19:37.775796902Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-drlnv,Uid:84096616-cf79-421c-9e94-8cbcf0a204a5,Namespace:calico-system,Attempt:3,}" Feb 13 19:19:37.775877 containerd[1515]: time="2025-02-13T19:19:37.775857185Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-664f6f5454-mnbf7,Uid:4e40d638-ead9-4cd4-859c-b3835db75e0c,Namespace:calico-system,Attempt:4,}" Feb 13 19:19:38.163643 containerd[1515]: time="2025-02-13T19:19:38.162866109Z" level=error msg="Failed to destroy network for sandbox \"913a65e5860c4b13ea05638c4cdf916cfe4d5179d77f6f08a60f9aa7332fc559\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:38.163643 containerd[1515]: time="2025-02-13T19:19:38.163270811Z" level=error msg="encountered an error cleaning up failed sandbox \"913a65e5860c4b13ea05638c4cdf916cfe4d5179d77f6f08a60f9aa7332fc559\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:38.163643 containerd[1515]: time="2025-02-13T19:19:38.163320284Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-s8264,Uid:f2fdfd8a-780b-4320-8f24-ca003c5d0c77,Namespace:kube-system,Attempt:4,} failed, error" error="failed to setup network for sandbox \"913a65e5860c4b13ea05638c4cdf916cfe4d5179d77f6f08a60f9aa7332fc559\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:38.164039 kubelet[2630]: E0213 19:19:38.163482 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"913a65e5860c4b13ea05638c4cdf916cfe4d5179d77f6f08a60f9aa7332fc559\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:38.164039 kubelet[2630]: E0213 19:19:38.163532 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"913a65e5860c4b13ea05638c4cdf916cfe4d5179d77f6f08a60f9aa7332fc559\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-s8264" Feb 13 19:19:38.164039 kubelet[2630]: E0213 19:19:38.163553 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"913a65e5860c4b13ea05638c4cdf916cfe4d5179d77f6f08a60f9aa7332fc559\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-s8264" Feb 13 19:19:38.164341 kubelet[2630]: E0213 19:19:38.163592 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-6f6b679f8f-s8264_kube-system(f2fdfd8a-780b-4320-8f24-ca003c5d0c77)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-6f6b679f8f-s8264_kube-system(f2fdfd8a-780b-4320-8f24-ca003c5d0c77)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"913a65e5860c4b13ea05638c4cdf916cfe4d5179d77f6f08a60f9aa7332fc559\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-6f6b679f8f-s8264" podUID="f2fdfd8a-780b-4320-8f24-ca003c5d0c77" Feb 13 19:19:38.168049 containerd[1515]: time="2025-02-13T19:19:38.166249164Z" level=error msg="Failed to destroy network for sandbox \"b9281cb17530bfce23ceee39b94a1f286e88137aa4b1dc73b5faa8a4e7379e99\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:38.169481 containerd[1515]: time="2025-02-13T19:19:38.169437193Z" level=error msg="encountered an error cleaning up failed sandbox \"b9281cb17530bfce23ceee39b94a1f286e88137aa4b1dc73b5faa8a4e7379e99\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:38.169554 containerd[1515]: time="2025-02-13T19:19:38.169530579Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-frtk4,Uid:f48c8839-9a17-43c4-a0d5-f9e8d9767814,Namespace:kube-system,Attempt:4,} failed, error" error="failed to setup network for sandbox \"b9281cb17530bfce23ceee39b94a1f286e88137aa4b1dc73b5faa8a4e7379e99\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:38.169810 kubelet[2630]: E0213 19:19:38.169754 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b9281cb17530bfce23ceee39b94a1f286e88137aa4b1dc73b5faa8a4e7379e99\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:38.170399 kubelet[2630]: E0213 19:19:38.169894 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b9281cb17530bfce23ceee39b94a1f286e88137aa4b1dc73b5faa8a4e7379e99\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-frtk4" Feb 13 19:19:38.170399 kubelet[2630]: E0213 19:19:38.169917 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b9281cb17530bfce23ceee39b94a1f286e88137aa4b1dc73b5faa8a4e7379e99\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-frtk4" Feb 13 19:19:38.170399 kubelet[2630]: E0213 19:19:38.169999 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-6f6b679f8f-frtk4_kube-system(f48c8839-9a17-43c4-a0d5-f9e8d9767814)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-6f6b679f8f-frtk4_kube-system(f48c8839-9a17-43c4-a0d5-f9e8d9767814)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"b9281cb17530bfce23ceee39b94a1f286e88137aa4b1dc73b5faa8a4e7379e99\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-6f6b679f8f-frtk4" podUID="f48c8839-9a17-43c4-a0d5-f9e8d9767814" Feb 13 19:19:38.172330 containerd[1515]: time="2025-02-13T19:19:38.172274332Z" level=error msg="Failed to destroy network for sandbox \"70821f98f9d2702e4d63972ae34f265d6851151a70f0070211ef163eb0840795\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:38.172755 containerd[1515]: time="2025-02-13T19:19:38.172734187Z" level=error msg="Failed to destroy network for sandbox \"9cfc4e8026fe9d3540dd0b1d209b241ede8d2453bfec90606b125f81c566e640\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:38.172897 containerd[1515]: time="2025-02-13T19:19:38.172860665Z" level=error msg="encountered an error cleaning up failed sandbox \"70821f98f9d2702e4d63972ae34f265d6851151a70f0070211ef163eb0840795\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:38.172950 containerd[1515]: time="2025-02-13T19:19:38.172920087Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86475b859c-swqrn,Uid:5a39efb7-aca1-4893-a36f-840c3c504ece,Namespace:calico-apiserver,Attempt:4,} failed, error" error="failed to setup network for sandbox \"70821f98f9d2702e4d63972ae34f265d6851151a70f0070211ef163eb0840795\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:38.173076 kubelet[2630]: E0213 19:19:38.173047 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"70821f98f9d2702e4d63972ae34f265d6851151a70f0070211ef163eb0840795\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:38.173370 containerd[1515]: time="2025-02-13T19:19:38.173346319Z" level=error msg="encountered an error cleaning up failed sandbox \"9cfc4e8026fe9d3540dd0b1d209b241ede8d2453bfec90606b125f81c566e640\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:38.173505 containerd[1515]: time="2025-02-13T19:19:38.173460504Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-664f6f5454-mnbf7,Uid:4e40d638-ead9-4cd4-859c-b3835db75e0c,Namespace:calico-system,Attempt:4,} failed, error" error="failed to setup network for sandbox \"9cfc4e8026fe9d3540dd0b1d209b241ede8d2453bfec90606b125f81c566e640\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:38.173765 kubelet[2630]: E0213 19:19:38.173736 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9cfc4e8026fe9d3540dd0b1d209b241ede8d2453bfec90606b125f81c566e640\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:38.173940 kubelet[2630]: E0213 19:19:38.173884 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9cfc4e8026fe9d3540dd0b1d209b241ede8d2453bfec90606b125f81c566e640\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-664f6f5454-mnbf7" Feb 13 19:19:38.173940 kubelet[2630]: E0213 19:19:38.173892 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"70821f98f9d2702e4d63972ae34f265d6851151a70f0070211ef163eb0840795\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86475b859c-swqrn" Feb 13 19:19:38.173940 kubelet[2630]: E0213 19:19:38.173910 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9cfc4e8026fe9d3540dd0b1d209b241ede8d2453bfec90606b125f81c566e640\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-664f6f5454-mnbf7" Feb 13 19:19:38.173940 kubelet[2630]: E0213 19:19:38.173916 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"70821f98f9d2702e4d63972ae34f265d6851151a70f0070211ef163eb0840795\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86475b859c-swqrn" Feb 13 19:19:38.174355 kubelet[2630]: E0213 19:19:38.174269 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-86475b859c-swqrn_calico-apiserver(5a39efb7-aca1-4893-a36f-840c3c504ece)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-86475b859c-swqrn_calico-apiserver(5a39efb7-aca1-4893-a36f-840c3c504ece)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"70821f98f9d2702e4d63972ae34f265d6851151a70f0070211ef163eb0840795\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-86475b859c-swqrn" podUID="5a39efb7-aca1-4893-a36f-840c3c504ece" Feb 13 19:19:38.174355 kubelet[2630]: E0213 19:19:38.174334 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-664f6f5454-mnbf7_calico-system(4e40d638-ead9-4cd4-859c-b3835db75e0c)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-664f6f5454-mnbf7_calico-system(4e40d638-ead9-4cd4-859c-b3835db75e0c)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"9cfc4e8026fe9d3540dd0b1d209b241ede8d2453bfec90606b125f81c566e640\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-664f6f5454-mnbf7" podUID="4e40d638-ead9-4cd4-859c-b3835db75e0c" Feb 13 19:19:38.177754 containerd[1515]: time="2025-02-13T19:19:38.177346536Z" level=error msg="Failed to destroy network for sandbox \"52abc7c7229296fa61042f80a0bb0c1bc30c4e6c9818f7329496601bde28a193\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:38.178412 containerd[1515]: time="2025-02-13T19:19:38.178380521Z" level=error msg="encountered an error cleaning up failed sandbox \"52abc7c7229296fa61042f80a0bb0c1bc30c4e6c9818f7329496601bde28a193\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:38.178538 containerd[1515]: time="2025-02-13T19:19:38.178439082Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86475b859c-kwlfl,Uid:f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1,Namespace:calico-apiserver,Attempt:4,} failed, error" error="failed to setup network for sandbox \"52abc7c7229296fa61042f80a0bb0c1bc30c4e6c9818f7329496601bde28a193\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:38.179730 kubelet[2630]: E0213 19:19:38.179660 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"52abc7c7229296fa61042f80a0bb0c1bc30c4e6c9818f7329496601bde28a193\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:38.179730 kubelet[2630]: E0213 19:19:38.179712 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"52abc7c7229296fa61042f80a0bb0c1bc30c4e6c9818f7329496601bde28a193\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86475b859c-kwlfl" Feb 13 19:19:38.179880 kubelet[2630]: E0213 19:19:38.179727 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"52abc7c7229296fa61042f80a0bb0c1bc30c4e6c9818f7329496601bde28a193\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86475b859c-kwlfl" Feb 13 19:19:38.179880 kubelet[2630]: E0213 19:19:38.179773 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-86475b859c-kwlfl_calico-apiserver(f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-86475b859c-kwlfl_calico-apiserver(f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"52abc7c7229296fa61042f80a0bb0c1bc30c4e6c9818f7329496601bde28a193\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-86475b859c-kwlfl" podUID="f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1" Feb 13 19:19:38.189724 containerd[1515]: time="2025-02-13T19:19:38.189590514Z" level=error msg="Failed to destroy network for sandbox \"a072e0a626a78003d707d9120f0d4982782f1be11bc8ddc22d2f8a1dfb030680\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:38.190067 containerd[1515]: time="2025-02-13T19:19:38.190044759Z" level=error msg="encountered an error cleaning up failed sandbox \"a072e0a626a78003d707d9120f0d4982782f1be11bc8ddc22d2f8a1dfb030680\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:38.190165 containerd[1515]: time="2025-02-13T19:19:38.190147833Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-drlnv,Uid:84096616-cf79-421c-9e94-8cbcf0a204a5,Namespace:calico-system,Attempt:3,} failed, error" error="failed to setup network for sandbox \"a072e0a626a78003d707d9120f0d4982782f1be11bc8ddc22d2f8a1dfb030680\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:38.190682 kubelet[2630]: E0213 19:19:38.190381 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a072e0a626a78003d707d9120f0d4982782f1be11bc8ddc22d2f8a1dfb030680\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:38.190682 kubelet[2630]: E0213 19:19:38.190430 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a072e0a626a78003d707d9120f0d4982782f1be11bc8ddc22d2f8a1dfb030680\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-drlnv" Feb 13 19:19:38.190682 kubelet[2630]: E0213 19:19:38.190446 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a072e0a626a78003d707d9120f0d4982782f1be11bc8ddc22d2f8a1dfb030680\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-drlnv" Feb 13 19:19:38.190786 kubelet[2630]: E0213 19:19:38.190477 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-drlnv_calico-system(84096616-cf79-421c-9e94-8cbcf0a204a5)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-drlnv_calico-system(84096616-cf79-421c-9e94-8cbcf0a204a5)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"a072e0a626a78003d707d9120f0d4982782f1be11bc8ddc22d2f8a1dfb030680\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-drlnv" podUID="84096616-cf79-421c-9e94-8cbcf0a204a5" Feb 13 19:19:38.395966 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-9cfc4e8026fe9d3540dd0b1d209b241ede8d2453bfec90606b125f81c566e640-shm.mount: Deactivated successfully. Feb 13 19:19:39.116584 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1866762993.mount: Deactivated successfully. Feb 13 19:19:39.164630 kubelet[2630]: I0213 19:19:39.164595 2630 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="913a65e5860c4b13ea05638c4cdf916cfe4d5179d77f6f08a60f9aa7332fc559" Feb 13 19:19:39.165259 containerd[1515]: time="2025-02-13T19:19:39.165217571Z" level=info msg="StopPodSandbox for \"913a65e5860c4b13ea05638c4cdf916cfe4d5179d77f6f08a60f9aa7332fc559\"" Feb 13 19:19:39.165851 containerd[1515]: time="2025-02-13T19:19:39.165416415Z" level=info msg="Ensure that sandbox 913a65e5860c4b13ea05638c4cdf916cfe4d5179d77f6f08a60f9aa7332fc559 in task-service has been cleanup successfully" Feb 13 19:19:39.165851 containerd[1515]: time="2025-02-13T19:19:39.165622644Z" level=info msg="TearDown network for sandbox \"913a65e5860c4b13ea05638c4cdf916cfe4d5179d77f6f08a60f9aa7332fc559\" successfully" Feb 13 19:19:39.165851 containerd[1515]: time="2025-02-13T19:19:39.165633735Z" level=info msg="StopPodSandbox for \"913a65e5860c4b13ea05638c4cdf916cfe4d5179d77f6f08a60f9aa7332fc559\" returns successfully" Feb 13 19:19:39.167524 containerd[1515]: time="2025-02-13T19:19:39.167479596Z" level=info msg="StopPodSandbox for \"fd762dfcc2c780c334b7852f3da6760c5d53d6cb6d243b5ae4cf1085971b8f57\"" Feb 13 19:19:39.167635 containerd[1515]: time="2025-02-13T19:19:39.167582280Z" level=info msg="TearDown network for sandbox \"fd762dfcc2c780c334b7852f3da6760c5d53d6cb6d243b5ae4cf1085971b8f57\" successfully" Feb 13 19:19:39.167635 containerd[1515]: time="2025-02-13T19:19:39.167629218Z" level=info msg="StopPodSandbox for \"fd762dfcc2c780c334b7852f3da6760c5d53d6cb6d243b5ae4cf1085971b8f57\" returns successfully" Feb 13 19:19:39.167801 systemd[1]: run-netns-cni\x2d47ba5b6d\x2dc7c0\x2d4fcc\x2d5299\x2d2a660deaf1ed.mount: Deactivated successfully. Feb 13 19:19:39.167990 containerd[1515]: time="2025-02-13T19:19:39.167823142Z" level=info msg="StopPodSandbox for \"8509c778c3294467a0ad8b8059bdecd145729001ee04e3c09d32d7f0d60b9468\"" Feb 13 19:19:39.167990 containerd[1515]: time="2025-02-13T19:19:39.167908544Z" level=info msg="TearDown network for sandbox \"8509c778c3294467a0ad8b8059bdecd145729001ee04e3c09d32d7f0d60b9468\" successfully" Feb 13 19:19:39.167990 containerd[1515]: time="2025-02-13T19:19:39.167920857Z" level=info msg="StopPodSandbox for \"8509c778c3294467a0ad8b8059bdecd145729001ee04e3c09d32d7f0d60b9468\" returns successfully" Feb 13 19:19:39.168855 containerd[1515]: time="2025-02-13T19:19:39.168813606Z" level=info msg="StopPodSandbox for \"706e09ab546bf3d89157cd9a2c32b16d198260f6686345817b1acf65f680fdd2\"" Feb 13 19:19:39.169013 containerd[1515]: time="2025-02-13T19:19:39.168918884Z" level=info msg="TearDown network for sandbox \"706e09ab546bf3d89157cd9a2c32b16d198260f6686345817b1acf65f680fdd2\" successfully" Feb 13 19:19:39.169013 containerd[1515]: time="2025-02-13T19:19:39.168984688Z" level=info msg="StopPodSandbox for \"706e09ab546bf3d89157cd9a2c32b16d198260f6686345817b1acf65f680fdd2\" returns successfully" Feb 13 19:19:39.169218 containerd[1515]: time="2025-02-13T19:19:39.169162703Z" level=info msg="StopPodSandbox for \"e2995529ca797998e1cf161d883bb684a3464a8a8d57602758f8705313f0d601\"" Feb 13 19:19:39.169307 containerd[1515]: time="2025-02-13T19:19:39.169282719Z" level=info msg="TearDown network for sandbox \"e2995529ca797998e1cf161d883bb684a3464a8a8d57602758f8705313f0d601\" successfully" Feb 13 19:19:39.169307 containerd[1515]: time="2025-02-13T19:19:39.169299841Z" level=info msg="StopPodSandbox for \"e2995529ca797998e1cf161d883bb684a3464a8a8d57602758f8705313f0d601\" returns successfully" Feb 13 19:19:39.169602 kubelet[2630]: I0213 19:19:39.169537 2630 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70821f98f9d2702e4d63972ae34f265d6851151a70f0070211ef163eb0840795" Feb 13 19:19:39.169602 kubelet[2630]: E0213 19:19:39.169573 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:39.169796 containerd[1515]: time="2025-02-13T19:19:39.169767481Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-s8264,Uid:f2fdfd8a-780b-4320-8f24-ca003c5d0c77,Namespace:kube-system,Attempt:5,}" Feb 13 19:19:39.170139 containerd[1515]: time="2025-02-13T19:19:39.170118992Z" level=info msg="StopPodSandbox for \"70821f98f9d2702e4d63972ae34f265d6851151a70f0070211ef163eb0840795\"" Feb 13 19:19:39.170569 containerd[1515]: time="2025-02-13T19:19:39.170544502Z" level=info msg="Ensure that sandbox 70821f98f9d2702e4d63972ae34f265d6851151a70f0070211ef163eb0840795 in task-service has been cleanup successfully" Feb 13 19:19:39.170787 containerd[1515]: time="2025-02-13T19:19:39.170760719Z" level=info msg="TearDown network for sandbox \"70821f98f9d2702e4d63972ae34f265d6851151a70f0070211ef163eb0840795\" successfully" Feb 13 19:19:39.170787 containerd[1515]: time="2025-02-13T19:19:39.170783592Z" level=info msg="StopPodSandbox for \"70821f98f9d2702e4d63972ae34f265d6851151a70f0070211ef163eb0840795\" returns successfully" Feb 13 19:19:39.171134 containerd[1515]: time="2025-02-13T19:19:39.171102382Z" level=info msg="StopPodSandbox for \"8a568a618a4c0eab0195822f2ccafce3787a78e075bf7c3f5b50c840e692188a\"" Feb 13 19:19:39.171219 containerd[1515]: time="2025-02-13T19:19:39.171175099Z" level=info msg="TearDown network for sandbox \"8a568a618a4c0eab0195822f2ccafce3787a78e075bf7c3f5b50c840e692188a\" successfully" Feb 13 19:19:39.171219 containerd[1515]: time="2025-02-13T19:19:39.171200556Z" level=info msg="StopPodSandbox for \"8a568a618a4c0eab0195822f2ccafce3787a78e075bf7c3f5b50c840e692188a\" returns successfully" Feb 13 19:19:39.171528 containerd[1515]: time="2025-02-13T19:19:39.171506502Z" level=info msg="StopPodSandbox for \"154203fe3c4258e99eeb24eefabf42a73b9226503cd839391e62422056e09db6\"" Feb 13 19:19:39.171619 containerd[1515]: time="2025-02-13T19:19:39.171602864Z" level=info msg="TearDown network for sandbox \"154203fe3c4258e99eeb24eefabf42a73b9226503cd839391e62422056e09db6\" successfully" Feb 13 19:19:39.171655 containerd[1515]: time="2025-02-13T19:19:39.171618864Z" level=info msg="StopPodSandbox for \"154203fe3c4258e99eeb24eefabf42a73b9226503cd839391e62422056e09db6\" returns successfully" Feb 13 19:19:39.172072 kubelet[2630]: I0213 19:19:39.171767 2630 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9cfc4e8026fe9d3540dd0b1d209b241ede8d2453bfec90606b125f81c566e640" Feb 13 19:19:39.172269 containerd[1515]: time="2025-02-13T19:19:39.172008987Z" level=info msg="StopPodSandbox for \"7d827c446a1e808002e0f19943157ab97bcc163d150e79960814ed79c5b4fbf3\"" Feb 13 19:19:39.172269 containerd[1515]: time="2025-02-13T19:19:39.172094909Z" level=info msg="TearDown network for sandbox \"7d827c446a1e808002e0f19943157ab97bcc163d150e79960814ed79c5b4fbf3\" successfully" Feb 13 19:19:39.172269 containerd[1515]: time="2025-02-13T19:19:39.172108936Z" level=info msg="StopPodSandbox for \"7d827c446a1e808002e0f19943157ab97bcc163d150e79960814ed79c5b4fbf3\" returns successfully" Feb 13 19:19:39.172269 containerd[1515]: time="2025-02-13T19:19:39.172216988Z" level=info msg="StopPodSandbox for \"9cfc4e8026fe9d3540dd0b1d209b241ede8d2453bfec90606b125f81c566e640\"" Feb 13 19:19:39.172413 containerd[1515]: time="2025-02-13T19:19:39.172354207Z" level=info msg="Ensure that sandbox 9cfc4e8026fe9d3540dd0b1d209b241ede8d2453bfec90606b125f81c566e640 in task-service has been cleanup successfully" Feb 13 19:19:39.172664 containerd[1515]: time="2025-02-13T19:19:39.172502997Z" level=info msg="TearDown network for sandbox \"9cfc4e8026fe9d3540dd0b1d209b241ede8d2453bfec90606b125f81c566e640\" successfully" Feb 13 19:19:39.172664 containerd[1515]: time="2025-02-13T19:19:39.172515651Z" level=info msg="StopPodSandbox for \"9cfc4e8026fe9d3540dd0b1d209b241ede8d2453bfec90606b125f81c566e640\" returns successfully" Feb 13 19:19:39.172740 containerd[1515]: time="2025-02-13T19:19:39.172710427Z" level=info msg="StopPodSandbox for \"334cbee9ef51499256254b9201c8014a8e1e8be612d9ae143735bc19f2a5f6a4\"" Feb 13 19:19:39.172804 containerd[1515]: time="2025-02-13T19:19:39.172787663Z" level=info msg="TearDown network for sandbox \"334cbee9ef51499256254b9201c8014a8e1e8be612d9ae143735bc19f2a5f6a4\" successfully" Feb 13 19:19:39.172830 containerd[1515]: time="2025-02-13T19:19:39.172803262Z" level=info msg="StopPodSandbox for \"334cbee9ef51499256254b9201c8014a8e1e8be612d9ae143735bc19f2a5f6a4\" returns successfully" Feb 13 19:19:39.172915 containerd[1515]: time="2025-02-13T19:19:39.172861501Z" level=info msg="StopPodSandbox for \"d195bcf8f9551fdb26f6b42c8b0e4b63d6953f1bbdc1c4b497262fc12be34248\"" Feb 13 19:19:39.172941 containerd[1515]: time="2025-02-13T19:19:39.172931312Z" level=info msg="TearDown network for sandbox \"d195bcf8f9551fdb26f6b42c8b0e4b63d6953f1bbdc1c4b497262fc12be34248\" successfully" Feb 13 19:19:39.172930 systemd[1]: run-netns-cni\x2d91a3b4c8\x2dd8ad\x2d0738\x2d49a0\x2db808aca6ed22.mount: Deactivated successfully. Feb 13 19:19:39.173032 containerd[1515]: time="2025-02-13T19:19:39.172943255Z" level=info msg="StopPodSandbox for \"d195bcf8f9551fdb26f6b42c8b0e4b63d6953f1bbdc1c4b497262fc12be34248\" returns successfully" Feb 13 19:19:39.173753 containerd[1515]: time="2025-02-13T19:19:39.173358156Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86475b859c-swqrn,Uid:5a39efb7-aca1-4893-a36f-840c3c504ece,Namespace:calico-apiserver,Attempt:5,}" Feb 13 19:19:39.173753 containerd[1515]: time="2025-02-13T19:19:39.173609738Z" level=info msg="StopPodSandbox for \"e5e6531fa4d44fae3b14be640ca5608599c77b930300d189004cbcc636c08e1b\"" Feb 13 19:19:39.173753 containerd[1515]: time="2025-02-13T19:19:39.173675713Z" level=info msg="TearDown network for sandbox \"e5e6531fa4d44fae3b14be640ca5608599c77b930300d189004cbcc636c08e1b\" successfully" Feb 13 19:19:39.173753 containerd[1515]: time="2025-02-13T19:19:39.173684229Z" level=info msg="StopPodSandbox for \"e5e6531fa4d44fae3b14be640ca5608599c77b930300d189004cbcc636c08e1b\" returns successfully" Feb 13 19:19:39.174123 containerd[1515]: time="2025-02-13T19:19:39.174099590Z" level=info msg="StopPodSandbox for \"4304f46dfb22ea4aced73143fa3f59c4ed426a670211c8559cc8ed545f1b3804\"" Feb 13 19:19:39.174330 containerd[1515]: time="2025-02-13T19:19:39.174171275Z" level=info msg="TearDown network for sandbox \"4304f46dfb22ea4aced73143fa3f59c4ed426a670211c8559cc8ed545f1b3804\" successfully" Feb 13 19:19:39.174330 containerd[1515]: time="2025-02-13T19:19:39.174182717Z" level=info msg="StopPodSandbox for \"4304f46dfb22ea4aced73143fa3f59c4ed426a670211c8559cc8ed545f1b3804\" returns successfully" Feb 13 19:19:39.174390 containerd[1515]: time="2025-02-13T19:19:39.174337869Z" level=info msg="StopPodSandbox for \"aa373a8af75a289d061969e5a22e647e175a6a451ba7c421e5c44fa363ec90f8\"" Feb 13 19:19:39.174420 containerd[1515]: time="2025-02-13T19:19:39.174400777Z" level=info msg="TearDown network for sandbox \"aa373a8af75a289d061969e5a22e647e175a6a451ba7c421e5c44fa363ec90f8\" successfully" Feb 13 19:19:39.174420 containerd[1515]: time="2025-02-13T19:19:39.174409403Z" level=info msg="StopPodSandbox for \"aa373a8af75a289d061969e5a22e647e175a6a451ba7c421e5c44fa363ec90f8\" returns successfully" Feb 13 19:19:39.175662 systemd[1]: run-netns-cni\x2ddfa57b63\x2d43eb\x2dc14a\x2d3b4f\x2d080ad03ca076.mount: Deactivated successfully. Feb 13 19:19:39.176448 containerd[1515]: time="2025-02-13T19:19:39.176417040Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-664f6f5454-mnbf7,Uid:4e40d638-ead9-4cd4-859c-b3835db75e0c,Namespace:calico-system,Attempt:5,}" Feb 13 19:19:39.177383 kubelet[2630]: I0213 19:19:39.177357 2630 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9281cb17530bfce23ceee39b94a1f286e88137aa4b1dc73b5faa8a4e7379e99" Feb 13 19:19:39.178115 containerd[1515]: time="2025-02-13T19:19:39.177906522Z" level=info msg="StopPodSandbox for \"b9281cb17530bfce23ceee39b94a1f286e88137aa4b1dc73b5faa8a4e7379e99\"" Feb 13 19:19:39.178115 containerd[1515]: time="2025-02-13T19:19:39.178091550Z" level=info msg="Ensure that sandbox b9281cb17530bfce23ceee39b94a1f286e88137aa4b1dc73b5faa8a4e7379e99 in task-service has been cleanup successfully" Feb 13 19:19:39.178374 containerd[1515]: time="2025-02-13T19:19:39.178297167Z" level=info msg="TearDown network for sandbox \"b9281cb17530bfce23ceee39b94a1f286e88137aa4b1dc73b5faa8a4e7379e99\" successfully" Feb 13 19:19:39.178374 containerd[1515]: time="2025-02-13T19:19:39.178308698Z" level=info msg="StopPodSandbox for \"b9281cb17530bfce23ceee39b94a1f286e88137aa4b1dc73b5faa8a4e7379e99\" returns successfully" Feb 13 19:19:39.178659 containerd[1515]: time="2025-02-13T19:19:39.178625945Z" level=info msg="StopPodSandbox for \"ee0d44b3886e78bb5043872859e92bfd2620af442c433a402780bc0453826443\"" Feb 13 19:19:39.178763 containerd[1515]: time="2025-02-13T19:19:39.178721705Z" level=info msg="TearDown network for sandbox \"ee0d44b3886e78bb5043872859e92bfd2620af442c433a402780bc0453826443\" successfully" Feb 13 19:19:39.178763 containerd[1515]: time="2025-02-13T19:19:39.178733278Z" level=info msg="StopPodSandbox for \"ee0d44b3886e78bb5043872859e92bfd2620af442c433a402780bc0453826443\" returns successfully" Feb 13 19:19:39.179077 containerd[1515]: time="2025-02-13T19:19:39.179040906Z" level=info msg="StopPodSandbox for \"28a752266e9ffa0b0f35649f5eba403b238d216dc8ce20a8fe3622cc28799217\"" Feb 13 19:19:39.179280 containerd[1515]: time="2025-02-13T19:19:39.179119264Z" level=info msg="TearDown network for sandbox \"28a752266e9ffa0b0f35649f5eba403b238d216dc8ce20a8fe3622cc28799217\" successfully" Feb 13 19:19:39.179280 containerd[1515]: time="2025-02-13T19:19:39.179176381Z" level=info msg="StopPodSandbox for \"28a752266e9ffa0b0f35649f5eba403b238d216dc8ce20a8fe3622cc28799217\" returns successfully" Feb 13 19:19:39.179637 containerd[1515]: time="2025-02-13T19:19:39.179614044Z" level=info msg="StopPodSandbox for \"a20bc39d31c8f248a56fc8b89d2d3ce62836f574c54b75502431e23bb908b697\"" Feb 13 19:19:39.181053 containerd[1515]: time="2025-02-13T19:19:39.179702330Z" level=info msg="TearDown network for sandbox \"a20bc39d31c8f248a56fc8b89d2d3ce62836f574c54b75502431e23bb908b697\" successfully" Feb 13 19:19:39.181053 containerd[1515]: time="2025-02-13T19:19:39.179713531Z" level=info msg="StopPodSandbox for \"a20bc39d31c8f248a56fc8b89d2d3ce62836f574c54b75502431e23bb908b697\" returns successfully" Feb 13 19:19:39.181053 containerd[1515]: time="2025-02-13T19:19:39.180129684Z" level=info msg="StopPodSandbox for \"a072e0a626a78003d707d9120f0d4982782f1be11bc8ddc22d2f8a1dfb030680\"" Feb 13 19:19:39.181053 containerd[1515]: time="2025-02-13T19:19:39.180281250Z" level=info msg="StopPodSandbox for \"f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7\"" Feb 13 19:19:39.181053 containerd[1515]: time="2025-02-13T19:19:39.180334299Z" level=info msg="Ensure that sandbox a072e0a626a78003d707d9120f0d4982782f1be11bc8ddc22d2f8a1dfb030680 in task-service has been cleanup successfully" Feb 13 19:19:39.181053 containerd[1515]: time="2025-02-13T19:19:39.180355369Z" level=info msg="TearDown network for sandbox \"f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7\" successfully" Feb 13 19:19:39.181053 containerd[1515]: time="2025-02-13T19:19:39.180365598Z" level=info msg="StopPodSandbox for \"f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7\" returns successfully" Feb 13 19:19:39.181053 containerd[1515]: time="2025-02-13T19:19:39.180501444Z" level=info msg="TearDown network for sandbox \"a072e0a626a78003d707d9120f0d4982782f1be11bc8ddc22d2f8a1dfb030680\" successfully" Feb 13 19:19:39.181053 containerd[1515]: time="2025-02-13T19:19:39.180511713Z" level=info msg="StopPodSandbox for \"a072e0a626a78003d707d9120f0d4982782f1be11bc8ddc22d2f8a1dfb030680\" returns successfully" Feb 13 19:19:39.181053 containerd[1515]: time="2025-02-13T19:19:39.180882110Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-frtk4,Uid:f48c8839-9a17-43c4-a0d5-f9e8d9767814,Namespace:kube-system,Attempt:5,}" Feb 13 19:19:39.181481 kubelet[2630]: I0213 19:19:39.179743 2630 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a072e0a626a78003d707d9120f0d4982782f1be11bc8ddc22d2f8a1dfb030680" Feb 13 19:19:39.181481 kubelet[2630]: E0213 19:19:39.180555 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:39.181947 containerd[1515]: time="2025-02-13T19:19:39.181675903Z" level=info msg="StopPodSandbox for \"216f8481f6c55f7a146dde6ed21a0f30b75cfe7059570683b2fb520eadf70207\"" Feb 13 19:19:39.181947 containerd[1515]: time="2025-02-13T19:19:39.181884515Z" level=info msg="TearDown network for sandbox \"216f8481f6c55f7a146dde6ed21a0f30b75cfe7059570683b2fb520eadf70207\" successfully" Feb 13 19:19:39.181947 containerd[1515]: time="2025-02-13T19:19:39.181898932Z" level=info msg="StopPodSandbox for \"216f8481f6c55f7a146dde6ed21a0f30b75cfe7059570683b2fb520eadf70207\" returns successfully" Feb 13 19:19:39.183629 containerd[1515]: time="2025-02-13T19:19:39.183440673Z" level=info msg="StopPodSandbox for \"1d811fd694e600e47efc43338f2c97ee7c9f512c6106b43a1a049dbc81724231\"" Feb 13 19:19:39.183629 containerd[1515]: time="2025-02-13T19:19:39.183518961Z" level=info msg="TearDown network for sandbox \"1d811fd694e600e47efc43338f2c97ee7c9f512c6106b43a1a049dbc81724231\" successfully" Feb 13 19:19:39.183629 containerd[1515]: time="2025-02-13T19:19:39.183528007Z" level=info msg="StopPodSandbox for \"1d811fd694e600e47efc43338f2c97ee7c9f512c6106b43a1a049dbc81724231\" returns successfully" Feb 13 19:19:39.183873 containerd[1515]: time="2025-02-13T19:19:39.183843040Z" level=info msg="StopPodSandbox for \"fa92388357966f3e183c59506d6cfde3e38d36deede0f30821c07e22d4146a49\"" Feb 13 19:19:39.183941 containerd[1515]: time="2025-02-13T19:19:39.183922199Z" level=info msg="TearDown network for sandbox \"fa92388357966f3e183c59506d6cfde3e38d36deede0f30821c07e22d4146a49\" successfully" Feb 13 19:19:39.183941 containerd[1515]: time="2025-02-13T19:19:39.183935904Z" level=info msg="StopPodSandbox for \"fa92388357966f3e183c59506d6cfde3e38d36deede0f30821c07e22d4146a49\" returns successfully" Feb 13 19:19:39.184642 containerd[1515]: time="2025-02-13T19:19:39.184422340Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-drlnv,Uid:84096616-cf79-421c-9e94-8cbcf0a204a5,Namespace:calico-system,Attempt:4,}" Feb 13 19:19:39.184701 kubelet[2630]: I0213 19:19:39.184430 2630 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52abc7c7229296fa61042f80a0bb0c1bc30c4e6c9818f7329496601bde28a193" Feb 13 19:19:39.184748 containerd[1515]: time="2025-02-13T19:19:39.184729818Z" level=info msg="StopPodSandbox for \"52abc7c7229296fa61042f80a0bb0c1bc30c4e6c9818f7329496601bde28a193\"" Feb 13 19:19:39.184903 containerd[1515]: time="2025-02-13T19:19:39.184886804Z" level=info msg="Ensure that sandbox 52abc7c7229296fa61042f80a0bb0c1bc30c4e6c9818f7329496601bde28a193 in task-service has been cleanup successfully" Feb 13 19:19:39.185068 containerd[1515]: time="2025-02-13T19:19:39.185046134Z" level=info msg="TearDown network for sandbox \"52abc7c7229296fa61042f80a0bb0c1bc30c4e6c9818f7329496601bde28a193\" successfully" Feb 13 19:19:39.185068 containerd[1515]: time="2025-02-13T19:19:39.185061262Z" level=info msg="StopPodSandbox for \"52abc7c7229296fa61042f80a0bb0c1bc30c4e6c9818f7329496601bde28a193\" returns successfully" Feb 13 19:19:39.185472 containerd[1515]: time="2025-02-13T19:19:39.185309528Z" level=info msg="StopPodSandbox for \"c03a59deff35e415f0920de8fd1c0462db72392ea452d0eafd388e65b284eb39\"" Feb 13 19:19:39.185472 containerd[1515]: time="2025-02-13T19:19:39.185400550Z" level=info msg="TearDown network for sandbox \"c03a59deff35e415f0920de8fd1c0462db72392ea452d0eafd388e65b284eb39\" successfully" Feb 13 19:19:39.185472 containerd[1515]: time="2025-02-13T19:19:39.185416069Z" level=info msg="StopPodSandbox for \"c03a59deff35e415f0920de8fd1c0462db72392ea452d0eafd388e65b284eb39\" returns successfully" Feb 13 19:19:39.185927 containerd[1515]: time="2025-02-13T19:19:39.185760357Z" level=info msg="StopPodSandbox for \"787aa28bb52b7073a96ced7755cf59d24334348abd3a51ce97858110663e1200\"" Feb 13 19:19:39.185927 containerd[1515]: time="2025-02-13T19:19:39.185851378Z" level=info msg="TearDown network for sandbox \"787aa28bb52b7073a96ced7755cf59d24334348abd3a51ce97858110663e1200\" successfully" Feb 13 19:19:39.185927 containerd[1515]: time="2025-02-13T19:19:39.185865524Z" level=info msg="StopPodSandbox for \"787aa28bb52b7073a96ced7755cf59d24334348abd3a51ce97858110663e1200\" returns successfully" Feb 13 19:19:39.186686 containerd[1515]: time="2025-02-13T19:19:39.186059831Z" level=info msg="StopPodSandbox for \"b4101238369440d1fc253b3e442bebf940a9f5fa609c91f1252b6dcea502c067\"" Feb 13 19:19:39.186686 containerd[1515]: time="2025-02-13T19:19:39.186130102Z" level=info msg="TearDown network for sandbox \"b4101238369440d1fc253b3e442bebf940a9f5fa609c91f1252b6dcea502c067\" successfully" Feb 13 19:19:39.186686 containerd[1515]: time="2025-02-13T19:19:39.186138508Z" level=info msg="StopPodSandbox for \"b4101238369440d1fc253b3e442bebf940a9f5fa609c91f1252b6dcea502c067\" returns successfully" Feb 13 19:19:39.186686 containerd[1515]: time="2025-02-13T19:19:39.186555263Z" level=info msg="StopPodSandbox for \"f7214e7c2d9f81df76f610a79140dcb0085636bf81f33d1eb139eb52cb12a6af\"" Feb 13 19:19:39.186686 containerd[1515]: time="2025-02-13T19:19:39.186627678Z" level=info msg="TearDown network for sandbox \"f7214e7c2d9f81df76f610a79140dcb0085636bf81f33d1eb139eb52cb12a6af\" successfully" Feb 13 19:19:39.186686 containerd[1515]: time="2025-02-13T19:19:39.186635553Z" level=info msg="StopPodSandbox for \"f7214e7c2d9f81df76f610a79140dcb0085636bf81f33d1eb139eb52cb12a6af\" returns successfully" Feb 13 19:19:39.187127 containerd[1515]: time="2025-02-13T19:19:39.187104115Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86475b859c-kwlfl,Uid:f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1,Namespace:calico-apiserver,Attempt:5,}" Feb 13 19:19:39.394358 systemd[1]: run-netns-cni\x2da893aed3\x2d8f53\x2d789a\x2d684e\x2dc7a89e45d324.mount: Deactivated successfully. Feb 13 19:19:39.394473 systemd[1]: run-netns-cni\x2dc1b3e227\x2de648\x2d798f\x2d2c00\x2d8ae04dde137a.mount: Deactivated successfully. Feb 13 19:19:39.394553 systemd[1]: run-netns-cni\x2ddbade362\x2d9537\x2dc6b4\x2d0f20\x2d1880ad9c3122.mount: Deactivated successfully. Feb 13 19:19:39.946073 containerd[1515]: time="2025-02-13T19:19:39.946019232Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:19:39.963570 containerd[1515]: time="2025-02-13T19:19:39.963522459Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.29.1: active requests=0, bytes read=142742010" Feb 13 19:19:39.974969 containerd[1515]: time="2025-02-13T19:19:39.974937674Z" level=info msg="ImageCreate event name:\"sha256:feb26d4585d68e875d9bd9bd6c27ea9f2d5c9ed9ef70f8b8cb0ebb0559a1d664\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:19:39.998405 containerd[1515]: time="2025-02-13T19:19:39.998350879Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:99c3917516efe1f807a0cfdf2d14b628b7c5cc6bd8a9ee5a253154f31756bea1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:19:40.007208 containerd[1515]: time="2025-02-13T19:19:40.005115914Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.29.1\" with image id \"sha256:feb26d4585d68e875d9bd9bd6c27ea9f2d5c9ed9ef70f8b8cb0ebb0559a1d664\", repo tag \"ghcr.io/flatcar/calico/node:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/node@sha256:99c3917516efe1f807a0cfdf2d14b628b7c5cc6bd8a9ee5a253154f31756bea1\", size \"142741872\" in 5.920435691s" Feb 13 19:19:40.007208 containerd[1515]: time="2025-02-13T19:19:40.005150148Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.1\" returns image reference \"sha256:feb26d4585d68e875d9bd9bd6c27ea9f2d5c9ed9ef70f8b8cb0ebb0559a1d664\"" Feb 13 19:19:40.024343 containerd[1515]: time="2025-02-13T19:19:40.024291892Z" level=info msg="CreateContainer within sandbox \"de4c4714d324a399cc66586668f0f0dccf310109ef8a4b507b491a5bbf423afa\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Feb 13 19:19:40.049177 containerd[1515]: time="2025-02-13T19:19:40.049057592Z" level=info msg="CreateContainer within sandbox \"de4c4714d324a399cc66586668f0f0dccf310109ef8a4b507b491a5bbf423afa\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"0afce0c7b685ae300a6ba2b2717dbc064399ca1dd4b6216e93698c5fc55790a6\"" Feb 13 19:19:40.049624 containerd[1515]: time="2025-02-13T19:19:40.049599792Z" level=info msg="StartContainer for \"0afce0c7b685ae300a6ba2b2717dbc064399ca1dd4b6216e93698c5fc55790a6\"" Feb 13 19:19:40.099110 containerd[1515]: time="2025-02-13T19:19:40.099066301Z" level=error msg="Failed to destroy network for sandbox \"4d2e4f99dab68adc0de82a705f6ff963ae519cfd54bd2435b6836ee53a8c4cf9\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:40.099607 containerd[1515]: time="2025-02-13T19:19:40.099585728Z" level=error msg="encountered an error cleaning up failed sandbox \"4d2e4f99dab68adc0de82a705f6ff963ae519cfd54bd2435b6836ee53a8c4cf9\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:40.099791 containerd[1515]: time="2025-02-13T19:19:40.099772539Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86475b859c-swqrn,Uid:5a39efb7-aca1-4893-a36f-840c3c504ece,Namespace:calico-apiserver,Attempt:5,} failed, error" error="failed to setup network for sandbox \"4d2e4f99dab68adc0de82a705f6ff963ae519cfd54bd2435b6836ee53a8c4cf9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:40.100220 kubelet[2630]: E0213 19:19:40.100067 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4d2e4f99dab68adc0de82a705f6ff963ae519cfd54bd2435b6836ee53a8c4cf9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:40.100220 kubelet[2630]: E0213 19:19:40.100137 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4d2e4f99dab68adc0de82a705f6ff963ae519cfd54bd2435b6836ee53a8c4cf9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86475b859c-swqrn" Feb 13 19:19:40.100220 kubelet[2630]: E0213 19:19:40.100159 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4d2e4f99dab68adc0de82a705f6ff963ae519cfd54bd2435b6836ee53a8c4cf9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86475b859c-swqrn" Feb 13 19:19:40.100411 kubelet[2630]: E0213 19:19:40.100366 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-86475b859c-swqrn_calico-apiserver(5a39efb7-aca1-4893-a36f-840c3c504ece)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-86475b859c-swqrn_calico-apiserver(5a39efb7-aca1-4893-a36f-840c3c504ece)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"4d2e4f99dab68adc0de82a705f6ff963ae519cfd54bd2435b6836ee53a8c4cf9\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-86475b859c-swqrn" podUID="5a39efb7-aca1-4893-a36f-840c3c504ece" Feb 13 19:19:40.114496 containerd[1515]: time="2025-02-13T19:19:40.114450556Z" level=error msg="Failed to destroy network for sandbox \"535ac305f91d459b29fcac754f456e877bb89217b5f89ed4a0f736398ab4e183\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:40.115139 containerd[1515]: time="2025-02-13T19:19:40.115106711Z" level=error msg="encountered an error cleaning up failed sandbox \"535ac305f91d459b29fcac754f456e877bb89217b5f89ed4a0f736398ab4e183\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:40.115289 containerd[1515]: time="2025-02-13T19:19:40.115250131Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-frtk4,Uid:f48c8839-9a17-43c4-a0d5-f9e8d9767814,Namespace:kube-system,Attempt:5,} failed, error" error="failed to setup network for sandbox \"535ac305f91d459b29fcac754f456e877bb89217b5f89ed4a0f736398ab4e183\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:40.115577 kubelet[2630]: E0213 19:19:40.115541 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"535ac305f91d459b29fcac754f456e877bb89217b5f89ed4a0f736398ab4e183\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:40.115971 kubelet[2630]: E0213 19:19:40.115678 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"535ac305f91d459b29fcac754f456e877bb89217b5f89ed4a0f736398ab4e183\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-frtk4" Feb 13 19:19:40.115971 kubelet[2630]: E0213 19:19:40.115700 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"535ac305f91d459b29fcac754f456e877bb89217b5f89ed4a0f736398ab4e183\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-frtk4" Feb 13 19:19:40.115971 kubelet[2630]: E0213 19:19:40.115746 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-6f6b679f8f-frtk4_kube-system(f48c8839-9a17-43c4-a0d5-f9e8d9767814)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-6f6b679f8f-frtk4_kube-system(f48c8839-9a17-43c4-a0d5-f9e8d9767814)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"535ac305f91d459b29fcac754f456e877bb89217b5f89ed4a0f736398ab4e183\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-6f6b679f8f-frtk4" podUID="f48c8839-9a17-43c4-a0d5-f9e8d9767814" Feb 13 19:19:40.117125 containerd[1515]: time="2025-02-13T19:19:40.117080584Z" level=error msg="Failed to destroy network for sandbox \"e941b39f4d4a62064805d7ae584462196d1704f1eaaa88f9ee2cc17abc84598e\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:40.117686 containerd[1515]: time="2025-02-13T19:19:40.117586214Z" level=error msg="encountered an error cleaning up failed sandbox \"e941b39f4d4a62064805d7ae584462196d1704f1eaaa88f9ee2cc17abc84598e\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:40.117798 containerd[1515]: time="2025-02-13T19:19:40.117750343Z" level=error msg="Failed to destroy network for sandbox \"45253c1f9735dfb2a3494283160983fb9ed43ea60532211ecb2b069637dde093\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:40.118386 containerd[1515]: time="2025-02-13T19:19:40.118347166Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-s8264,Uid:f2fdfd8a-780b-4320-8f24-ca003c5d0c77,Namespace:kube-system,Attempt:5,} failed, error" error="failed to setup network for sandbox \"e941b39f4d4a62064805d7ae584462196d1704f1eaaa88f9ee2cc17abc84598e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:40.118441 containerd[1515]: time="2025-02-13T19:19:40.118419932Z" level=error msg="encountered an error cleaning up failed sandbox \"45253c1f9735dfb2a3494283160983fb9ed43ea60532211ecb2b069637dde093\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:40.118672 kubelet[2630]: E0213 19:19:40.118627 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e941b39f4d4a62064805d7ae584462196d1704f1eaaa88f9ee2cc17abc84598e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:40.118721 kubelet[2630]: E0213 19:19:40.118689 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e941b39f4d4a62064805d7ae584462196d1704f1eaaa88f9ee2cc17abc84598e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-s8264" Feb 13 19:19:40.118721 kubelet[2630]: E0213 19:19:40.118710 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e941b39f4d4a62064805d7ae584462196d1704f1eaaa88f9ee2cc17abc84598e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-s8264" Feb 13 19:19:40.118792 kubelet[2630]: E0213 19:19:40.118749 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-6f6b679f8f-s8264_kube-system(f2fdfd8a-780b-4320-8f24-ca003c5d0c77)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-6f6b679f8f-s8264_kube-system(f2fdfd8a-780b-4320-8f24-ca003c5d0c77)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"e941b39f4d4a62064805d7ae584462196d1704f1eaaa88f9ee2cc17abc84598e\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-6f6b679f8f-s8264" podUID="f2fdfd8a-780b-4320-8f24-ca003c5d0c77" Feb 13 19:19:40.118853 containerd[1515]: time="2025-02-13T19:19:40.118451832Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-664f6f5454-mnbf7,Uid:4e40d638-ead9-4cd4-859c-b3835db75e0c,Namespace:calico-system,Attempt:5,} failed, error" error="failed to setup network for sandbox \"45253c1f9735dfb2a3494283160983fb9ed43ea60532211ecb2b069637dde093\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:40.120044 kubelet[2630]: E0213 19:19:40.118959 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"45253c1f9735dfb2a3494283160983fb9ed43ea60532211ecb2b069637dde093\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:40.120044 kubelet[2630]: E0213 19:19:40.118985 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"45253c1f9735dfb2a3494283160983fb9ed43ea60532211ecb2b069637dde093\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-664f6f5454-mnbf7" Feb 13 19:19:40.120044 kubelet[2630]: E0213 19:19:40.119002 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"45253c1f9735dfb2a3494283160983fb9ed43ea60532211ecb2b069637dde093\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-664f6f5454-mnbf7" Feb 13 19:19:40.120209 kubelet[2630]: E0213 19:19:40.119028 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-664f6f5454-mnbf7_calico-system(4e40d638-ead9-4cd4-859c-b3835db75e0c)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-664f6f5454-mnbf7_calico-system(4e40d638-ead9-4cd4-859c-b3835db75e0c)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"45253c1f9735dfb2a3494283160983fb9ed43ea60532211ecb2b069637dde093\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-664f6f5454-mnbf7" podUID="4e40d638-ead9-4cd4-859c-b3835db75e0c" Feb 13 19:19:40.127519 containerd[1515]: time="2025-02-13T19:19:40.127378682Z" level=error msg="Failed to destroy network for sandbox \"41f8a27c833a329de6525bb7ea4ae60918c875126c9b0eb7529120508e6289e1\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:40.127931 containerd[1515]: time="2025-02-13T19:19:40.127890625Z" level=error msg="encountered an error cleaning up failed sandbox \"41f8a27c833a329de6525bb7ea4ae60918c875126c9b0eb7529120508e6289e1\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:40.128056 containerd[1515]: time="2025-02-13T19:19:40.127947242Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-drlnv,Uid:84096616-cf79-421c-9e94-8cbcf0a204a5,Namespace:calico-system,Attempt:4,} failed, error" error="failed to setup network for sandbox \"41f8a27c833a329de6525bb7ea4ae60918c875126c9b0eb7529120508e6289e1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:40.128169 kubelet[2630]: E0213 19:19:40.128120 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"41f8a27c833a329de6525bb7ea4ae60918c875126c9b0eb7529120508e6289e1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:40.128259 kubelet[2630]: E0213 19:19:40.128179 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"41f8a27c833a329de6525bb7ea4ae60918c875126c9b0eb7529120508e6289e1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-drlnv" Feb 13 19:19:40.128259 kubelet[2630]: E0213 19:19:40.128213 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"41f8a27c833a329de6525bb7ea4ae60918c875126c9b0eb7529120508e6289e1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-drlnv" Feb 13 19:19:40.128387 kubelet[2630]: E0213 19:19:40.128261 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-drlnv_calico-system(84096616-cf79-421c-9e94-8cbcf0a204a5)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-drlnv_calico-system(84096616-cf79-421c-9e94-8cbcf0a204a5)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"41f8a27c833a329de6525bb7ea4ae60918c875126c9b0eb7529120508e6289e1\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-drlnv" podUID="84096616-cf79-421c-9e94-8cbcf0a204a5" Feb 13 19:19:40.128981 containerd[1515]: time="2025-02-13T19:19:40.128922045Z" level=error msg="Failed to destroy network for sandbox \"216ea206995c57165c7ec894ec603cac89445bad8cd05143541089c3e5d6db12\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:40.129378 containerd[1515]: time="2025-02-13T19:19:40.129342917Z" level=error msg="encountered an error cleaning up failed sandbox \"216ea206995c57165c7ec894ec603cac89445bad8cd05143541089c3e5d6db12\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:40.129414 containerd[1515]: time="2025-02-13T19:19:40.129401287Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86475b859c-kwlfl,Uid:f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1,Namespace:calico-apiserver,Attempt:5,} failed, error" error="failed to setup network for sandbox \"216ea206995c57165c7ec894ec603cac89445bad8cd05143541089c3e5d6db12\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:40.129662 kubelet[2630]: E0213 19:19:40.129589 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"216ea206995c57165c7ec894ec603cac89445bad8cd05143541089c3e5d6db12\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:40.129734 kubelet[2630]: E0213 19:19:40.129681 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"216ea206995c57165c7ec894ec603cac89445bad8cd05143541089c3e5d6db12\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86475b859c-kwlfl" Feb 13 19:19:40.129734 kubelet[2630]: E0213 19:19:40.129705 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"216ea206995c57165c7ec894ec603cac89445bad8cd05143541089c3e5d6db12\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86475b859c-kwlfl" Feb 13 19:19:40.129786 kubelet[2630]: E0213 19:19:40.129746 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-86475b859c-kwlfl_calico-apiserver(f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-86475b859c-kwlfl_calico-apiserver(f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"216ea206995c57165c7ec894ec603cac89445bad8cd05143541089c3e5d6db12\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-86475b859c-kwlfl" podUID="f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1" Feb 13 19:19:40.147321 systemd[1]: Started cri-containerd-0afce0c7b685ae300a6ba2b2717dbc064399ca1dd4b6216e93698c5fc55790a6.scope - libcontainer container 0afce0c7b685ae300a6ba2b2717dbc064399ca1dd4b6216e93698c5fc55790a6. Feb 13 19:19:40.178980 containerd[1515]: time="2025-02-13T19:19:40.178940322Z" level=info msg="StartContainer for \"0afce0c7b685ae300a6ba2b2717dbc064399ca1dd4b6216e93698c5fc55790a6\" returns successfully" Feb 13 19:19:40.193121 kubelet[2630]: I0213 19:19:40.192349 2630 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="535ac305f91d459b29fcac754f456e877bb89217b5f89ed4a0f736398ab4e183" Feb 13 19:19:40.193490 containerd[1515]: time="2025-02-13T19:19:40.192772209Z" level=info msg="StopPodSandbox for \"535ac305f91d459b29fcac754f456e877bb89217b5f89ed4a0f736398ab4e183\"" Feb 13 19:19:40.193490 containerd[1515]: time="2025-02-13T19:19:40.192983917Z" level=info msg="Ensure that sandbox 535ac305f91d459b29fcac754f456e877bb89217b5f89ed4a0f736398ab4e183 in task-service has been cleanup successfully" Feb 13 19:19:40.193640 containerd[1515]: time="2025-02-13T19:19:40.193624070Z" level=info msg="TearDown network for sandbox \"535ac305f91d459b29fcac754f456e877bb89217b5f89ed4a0f736398ab4e183\" successfully" Feb 13 19:19:40.193696 containerd[1515]: time="2025-02-13T19:19:40.193684384Z" level=info msg="StopPodSandbox for \"535ac305f91d459b29fcac754f456e877bb89217b5f89ed4a0f736398ab4e183\" returns successfully" Feb 13 19:19:40.194006 containerd[1515]: time="2025-02-13T19:19:40.193968097Z" level=info msg="StopPodSandbox for \"b9281cb17530bfce23ceee39b94a1f286e88137aa4b1dc73b5faa8a4e7379e99\"" Feb 13 19:19:40.194577 containerd[1515]: time="2025-02-13T19:19:40.194512992Z" level=info msg="TearDown network for sandbox \"b9281cb17530bfce23ceee39b94a1f286e88137aa4b1dc73b5faa8a4e7379e99\" successfully" Feb 13 19:19:40.194626 containerd[1515]: time="2025-02-13T19:19:40.194578576Z" level=info msg="StopPodSandbox for \"b9281cb17530bfce23ceee39b94a1f286e88137aa4b1dc73b5faa8a4e7379e99\" returns successfully" Feb 13 19:19:40.195439 containerd[1515]: time="2025-02-13T19:19:40.195011871Z" level=info msg="StopPodSandbox for \"ee0d44b3886e78bb5043872859e92bfd2620af442c433a402780bc0453826443\"" Feb 13 19:19:40.195499 containerd[1515]: time="2025-02-13T19:19:40.195486784Z" level=info msg="TearDown network for sandbox \"ee0d44b3886e78bb5043872859e92bfd2620af442c433a402780bc0453826443\" successfully" Feb 13 19:19:40.195533 containerd[1515]: time="2025-02-13T19:19:40.195499789Z" level=info msg="StopPodSandbox for \"ee0d44b3886e78bb5043872859e92bfd2620af442c433a402780bc0453826443\" returns successfully" Feb 13 19:19:40.195964 containerd[1515]: time="2025-02-13T19:19:40.195941249Z" level=info msg="StopPodSandbox for \"28a752266e9ffa0b0f35649f5eba403b238d216dc8ce20a8fe3622cc28799217\"" Feb 13 19:19:40.196304 containerd[1515]: time="2025-02-13T19:19:40.196105778Z" level=info msg="TearDown network for sandbox \"28a752266e9ffa0b0f35649f5eba403b238d216dc8ce20a8fe3622cc28799217\" successfully" Feb 13 19:19:40.196304 containerd[1515]: time="2025-02-13T19:19:40.196179977Z" level=info msg="StopPodSandbox for \"28a752266e9ffa0b0f35649f5eba403b238d216dc8ce20a8fe3622cc28799217\" returns successfully" Feb 13 19:19:40.197178 kubelet[2630]: I0213 19:19:40.197140 2630 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41f8a27c833a329de6525bb7ea4ae60918c875126c9b0eb7529120508e6289e1" Feb 13 19:19:40.198623 containerd[1515]: time="2025-02-13T19:19:40.198577968Z" level=info msg="StopPodSandbox for \"a20bc39d31c8f248a56fc8b89d2d3ce62836f574c54b75502431e23bb908b697\"" Feb 13 19:19:40.198720 containerd[1515]: time="2025-02-13T19:19:40.198672866Z" level=info msg="TearDown network for sandbox \"a20bc39d31c8f248a56fc8b89d2d3ce62836f574c54b75502431e23bb908b697\" successfully" Feb 13 19:19:40.198720 containerd[1515]: time="2025-02-13T19:19:40.198693275Z" level=info msg="StopPodSandbox for \"a20bc39d31c8f248a56fc8b89d2d3ce62836f574c54b75502431e23bb908b697\" returns successfully" Feb 13 19:19:40.199441 containerd[1515]: time="2025-02-13T19:19:40.199414000Z" level=info msg="StopPodSandbox for \"41f8a27c833a329de6525bb7ea4ae60918c875126c9b0eb7529120508e6289e1\"" Feb 13 19:19:40.200876 containerd[1515]: time="2025-02-13T19:19:40.200853598Z" level=info msg="Ensure that sandbox 41f8a27c833a329de6525bb7ea4ae60918c875126c9b0eb7529120508e6289e1 in task-service has been cleanup successfully" Feb 13 19:19:40.201124 containerd[1515]: time="2025-02-13T19:19:40.200575255Z" level=info msg="StopPodSandbox for \"f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7\"" Feb 13 19:19:40.201702 containerd[1515]: time="2025-02-13T19:19:40.201234855Z" level=info msg="TearDown network for sandbox \"41f8a27c833a329de6525bb7ea4ae60918c875126c9b0eb7529120508e6289e1\" successfully" Feb 13 19:19:40.201880 containerd[1515]: time="2025-02-13T19:19:40.201776664Z" level=info msg="StopPodSandbox for \"41f8a27c833a329de6525bb7ea4ae60918c875126c9b0eb7529120508e6289e1\" returns successfully" Feb 13 19:19:40.201880 containerd[1515]: time="2025-02-13T19:19:40.201716952Z" level=info msg="TearDown network for sandbox \"f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7\" successfully" Feb 13 19:19:40.201880 containerd[1515]: time="2025-02-13T19:19:40.201824965Z" level=info msg="StopPodSandbox for \"f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7\" returns successfully" Feb 13 19:19:40.202392 kubelet[2630]: E0213 19:19:40.202112 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:40.203181 kubelet[2630]: E0213 19:19:40.203154 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:40.203355 containerd[1515]: time="2025-02-13T19:19:40.203327642Z" level=info msg="StopPodSandbox for \"a072e0a626a78003d707d9120f0d4982782f1be11bc8ddc22d2f8a1dfb030680\"" Feb 13 19:19:40.203741 containerd[1515]: time="2025-02-13T19:19:40.203542736Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-frtk4,Uid:f48c8839-9a17-43c4-a0d5-f9e8d9767814,Namespace:kube-system,Attempt:6,}" Feb 13 19:19:40.204334 containerd[1515]: time="2025-02-13T19:19:40.204313716Z" level=info msg="TearDown network for sandbox \"a072e0a626a78003d707d9120f0d4982782f1be11bc8ddc22d2f8a1dfb030680\" successfully" Feb 13 19:19:40.204441 containerd[1515]: time="2025-02-13T19:19:40.204424204Z" level=info msg="StopPodSandbox for \"a072e0a626a78003d707d9120f0d4982782f1be11bc8ddc22d2f8a1dfb030680\" returns successfully" Feb 13 19:19:40.204786 containerd[1515]: time="2025-02-13T19:19:40.204763202Z" level=info msg="StopPodSandbox for \"216f8481f6c55f7a146dde6ed21a0f30b75cfe7059570683b2fb520eadf70207\"" Feb 13 19:19:40.204873 containerd[1515]: time="2025-02-13T19:19:40.204849774Z" level=info msg="TearDown network for sandbox \"216f8481f6c55f7a146dde6ed21a0f30b75cfe7059570683b2fb520eadf70207\" successfully" Feb 13 19:19:40.204873 containerd[1515]: time="2025-02-13T19:19:40.204866265Z" level=info msg="StopPodSandbox for \"216f8481f6c55f7a146dde6ed21a0f30b75cfe7059570683b2fb520eadf70207\" returns successfully" Feb 13 19:19:40.205158 containerd[1515]: time="2025-02-13T19:19:40.205132687Z" level=info msg="StopPodSandbox for \"1d811fd694e600e47efc43338f2c97ee7c9f512c6106b43a1a049dbc81724231\"" Feb 13 19:19:40.205303 containerd[1515]: time="2025-02-13T19:19:40.205256218Z" level=info msg="TearDown network for sandbox \"1d811fd694e600e47efc43338f2c97ee7c9f512c6106b43a1a049dbc81724231\" successfully" Feb 13 19:19:40.205303 containerd[1515]: time="2025-02-13T19:19:40.205267409Z" level=info msg="StopPodSandbox for \"1d811fd694e600e47efc43338f2c97ee7c9f512c6106b43a1a049dbc81724231\" returns successfully" Feb 13 19:19:40.205680 containerd[1515]: time="2025-02-13T19:19:40.205617698Z" level=info msg="StopPodSandbox for \"fa92388357966f3e183c59506d6cfde3e38d36deede0f30821c07e22d4146a49\"" Feb 13 19:19:40.206892 kubelet[2630]: I0213 19:19:40.206855 2630 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="216ea206995c57165c7ec894ec603cac89445bad8cd05143541089c3e5d6db12" Feb 13 19:19:40.207111 containerd[1515]: time="2025-02-13T19:19:40.207015187Z" level=info msg="TearDown network for sandbox \"fa92388357966f3e183c59506d6cfde3e38d36deede0f30821c07e22d4146a49\" successfully" Feb 13 19:19:40.207111 containerd[1515]: time="2025-02-13T19:19:40.207060392Z" level=info msg="StopPodSandbox for \"fa92388357966f3e183c59506d6cfde3e38d36deede0f30821c07e22d4146a49\" returns successfully" Feb 13 19:19:40.207436 containerd[1515]: time="2025-02-13T19:19:40.207382177Z" level=info msg="StopPodSandbox for \"216ea206995c57165c7ec894ec603cac89445bad8cd05143541089c3e5d6db12\"" Feb 13 19:19:40.207799 containerd[1515]: time="2025-02-13T19:19:40.207778042Z" level=info msg="Ensure that sandbox 216ea206995c57165c7ec894ec603cac89445bad8cd05143541089c3e5d6db12 in task-service has been cleanup successfully" Feb 13 19:19:40.208248 containerd[1515]: time="2025-02-13T19:19:40.208231535Z" level=info msg="TearDown network for sandbox \"216ea206995c57165c7ec894ec603cac89445bad8cd05143541089c3e5d6db12\" successfully" Feb 13 19:19:40.208395 containerd[1515]: time="2025-02-13T19:19:40.208301056Z" level=info msg="StopPodSandbox for \"216ea206995c57165c7ec894ec603cac89445bad8cd05143541089c3e5d6db12\" returns successfully" Feb 13 19:19:40.208877 containerd[1515]: time="2025-02-13T19:19:40.208762674Z" level=info msg="StopPodSandbox for \"52abc7c7229296fa61042f80a0bb0c1bc30c4e6c9818f7329496601bde28a193\"" Feb 13 19:19:40.209037 containerd[1515]: time="2025-02-13T19:19:40.208826584Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-drlnv,Uid:84096616-cf79-421c-9e94-8cbcf0a204a5,Namespace:calico-system,Attempt:5,}" Feb 13 19:19:40.209493 containerd[1515]: time="2025-02-13T19:19:40.209474032Z" level=info msg="TearDown network for sandbox \"52abc7c7229296fa61042f80a0bb0c1bc30c4e6c9818f7329496601bde28a193\" successfully" Feb 13 19:19:40.209493 containerd[1515]: time="2025-02-13T19:19:40.209490012Z" level=info msg="StopPodSandbox for \"52abc7c7229296fa61042f80a0bb0c1bc30c4e6c9818f7329496601bde28a193\" returns successfully" Feb 13 19:19:40.210176 containerd[1515]: time="2025-02-13T19:19:40.210156535Z" level=info msg="StopPodSandbox for \"c03a59deff35e415f0920de8fd1c0462db72392ea452d0eafd388e65b284eb39\"" Feb 13 19:19:40.210284 containerd[1515]: time="2025-02-13T19:19:40.210244171Z" level=info msg="TearDown network for sandbox \"c03a59deff35e415f0920de8fd1c0462db72392ea452d0eafd388e65b284eb39\" successfully" Feb 13 19:19:40.210284 containerd[1515]: time="2025-02-13T19:19:40.210282963Z" level=info msg="StopPodSandbox for \"c03a59deff35e415f0920de8fd1c0462db72392ea452d0eafd388e65b284eb39\" returns successfully" Feb 13 19:19:40.211385 containerd[1515]: time="2025-02-13T19:19:40.211278697Z" level=info msg="StopPodSandbox for \"787aa28bb52b7073a96ced7755cf59d24334348abd3a51ce97858110663e1200\"" Feb 13 19:19:40.211385 containerd[1515]: time="2025-02-13T19:19:40.211371030Z" level=info msg="TearDown network for sandbox \"787aa28bb52b7073a96ced7755cf59d24334348abd3a51ce97858110663e1200\" successfully" Feb 13 19:19:40.211385 containerd[1515]: time="2025-02-13T19:19:40.211381289Z" level=info msg="StopPodSandbox for \"787aa28bb52b7073a96ced7755cf59d24334348abd3a51ce97858110663e1200\" returns successfully" Feb 13 19:19:40.212268 containerd[1515]: time="2025-02-13T19:19:40.212130749Z" level=info msg="StopPodSandbox for \"b4101238369440d1fc253b3e442bebf940a9f5fa609c91f1252b6dcea502c067\"" Feb 13 19:19:40.212419 containerd[1515]: time="2025-02-13T19:19:40.212392932Z" level=info msg="TearDown network for sandbox \"b4101238369440d1fc253b3e442bebf940a9f5fa609c91f1252b6dcea502c067\" successfully" Feb 13 19:19:40.212687 containerd[1515]: time="2025-02-13T19:19:40.212657509Z" level=info msg="StopPodSandbox for \"b4101238369440d1fc253b3e442bebf940a9f5fa609c91f1252b6dcea502c067\" returns successfully" Feb 13 19:19:40.212971 kubelet[2630]: I0213 19:19:40.212940 2630 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d2e4f99dab68adc0de82a705f6ff963ae519cfd54bd2435b6836ee53a8c4cf9" Feb 13 19:19:40.213548 containerd[1515]: time="2025-02-13T19:19:40.213529319Z" level=info msg="StopPodSandbox for \"f7214e7c2d9f81df76f610a79140dcb0085636bf81f33d1eb139eb52cb12a6af\"" Feb 13 19:19:40.213716 containerd[1515]: time="2025-02-13T19:19:40.213662851Z" level=info msg="TearDown network for sandbox \"f7214e7c2d9f81df76f610a79140dcb0085636bf81f33d1eb139eb52cb12a6af\" successfully" Feb 13 19:19:40.213716 containerd[1515]: time="2025-02-13T19:19:40.213675985Z" level=info msg="StopPodSandbox for \"f7214e7c2d9f81df76f610a79140dcb0085636bf81f33d1eb139eb52cb12a6af\" returns successfully" Feb 13 19:19:40.214045 containerd[1515]: time="2025-02-13T19:19:40.213949950Z" level=info msg="StopPodSandbox for \"4d2e4f99dab68adc0de82a705f6ff963ae519cfd54bd2435b6836ee53a8c4cf9\"" Feb 13 19:19:40.214243 containerd[1515]: time="2025-02-13T19:19:40.214135098Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86475b859c-kwlfl,Uid:f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1,Namespace:calico-apiserver,Attempt:6,}" Feb 13 19:19:40.214402 containerd[1515]: time="2025-02-13T19:19:40.214373197Z" level=info msg="Ensure that sandbox 4d2e4f99dab68adc0de82a705f6ff963ae519cfd54bd2435b6836ee53a8c4cf9 in task-service has been cleanup successfully" Feb 13 19:19:40.214968 containerd[1515]: time="2025-02-13T19:19:40.214914185Z" level=info msg="TearDown network for sandbox \"4d2e4f99dab68adc0de82a705f6ff963ae519cfd54bd2435b6836ee53a8c4cf9\" successfully" Feb 13 19:19:40.214968 containerd[1515]: time="2025-02-13T19:19:40.214934132Z" level=info msg="StopPodSandbox for \"4d2e4f99dab68adc0de82a705f6ff963ae519cfd54bd2435b6836ee53a8c4cf9\" returns successfully" Feb 13 19:19:40.215382 containerd[1515]: time="2025-02-13T19:19:40.215334124Z" level=info msg="StopPodSandbox for \"70821f98f9d2702e4d63972ae34f265d6851151a70f0070211ef163eb0840795\"" Feb 13 19:19:40.215536 containerd[1515]: time="2025-02-13T19:19:40.215419584Z" level=info msg="TearDown network for sandbox \"70821f98f9d2702e4d63972ae34f265d6851151a70f0070211ef163eb0840795\" successfully" Feb 13 19:19:40.215536 containerd[1515]: time="2025-02-13T19:19:40.215430415Z" level=info msg="StopPodSandbox for \"70821f98f9d2702e4d63972ae34f265d6851151a70f0070211ef163eb0840795\" returns successfully" Feb 13 19:19:40.216427 containerd[1515]: time="2025-02-13T19:19:40.216230741Z" level=info msg="StopPodSandbox for \"8a568a618a4c0eab0195822f2ccafce3787a78e075bf7c3f5b50c840e692188a\"" Feb 13 19:19:40.216427 containerd[1515]: time="2025-02-13T19:19:40.216324837Z" level=info msg="TearDown network for sandbox \"8a568a618a4c0eab0195822f2ccafce3787a78e075bf7c3f5b50c840e692188a\" successfully" Feb 13 19:19:40.216427 containerd[1515]: time="2025-02-13T19:19:40.216336829Z" level=info msg="StopPodSandbox for \"8a568a618a4c0eab0195822f2ccafce3787a78e075bf7c3f5b50c840e692188a\" returns successfully" Feb 13 19:19:40.217005 containerd[1515]: time="2025-02-13T19:19:40.216792437Z" level=info msg="StopPodSandbox for \"154203fe3c4258e99eeb24eefabf42a73b9226503cd839391e62422056e09db6\"" Feb 13 19:19:40.217005 containerd[1515]: time="2025-02-13T19:19:40.216884880Z" level=info msg="TearDown network for sandbox \"154203fe3c4258e99eeb24eefabf42a73b9226503cd839391e62422056e09db6\" successfully" Feb 13 19:19:40.217005 containerd[1515]: time="2025-02-13T19:19:40.216897063Z" level=info msg="StopPodSandbox for \"154203fe3c4258e99eeb24eefabf42a73b9226503cd839391e62422056e09db6\" returns successfully" Feb 13 19:19:40.217630 containerd[1515]: time="2025-02-13T19:19:40.217459552Z" level=info msg="StopPodSandbox for \"7d827c446a1e808002e0f19943157ab97bcc163d150e79960814ed79c5b4fbf3\"" Feb 13 19:19:40.217630 containerd[1515]: time="2025-02-13T19:19:40.217562164Z" level=info msg="TearDown network for sandbox \"7d827c446a1e808002e0f19943157ab97bcc163d150e79960814ed79c5b4fbf3\" successfully" Feb 13 19:19:40.217630 containerd[1515]: time="2025-02-13T19:19:40.217576181Z" level=info msg="StopPodSandbox for \"7d827c446a1e808002e0f19943157ab97bcc163d150e79960814ed79c5b4fbf3\" returns successfully" Feb 13 19:19:40.217776 kubelet[2630]: I0213 19:19:40.217726 2630 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e941b39f4d4a62064805d7ae584462196d1704f1eaaa88f9ee2cc17abc84598e" Feb 13 19:19:40.219670 containerd[1515]: time="2025-02-13T19:19:40.218544863Z" level=info msg="StopPodSandbox for \"e941b39f4d4a62064805d7ae584462196d1704f1eaaa88f9ee2cc17abc84598e\"" Feb 13 19:19:40.219670 containerd[1515]: time="2025-02-13T19:19:40.218709022Z" level=info msg="Ensure that sandbox e941b39f4d4a62064805d7ae584462196d1704f1eaaa88f9ee2cc17abc84598e in task-service has been cleanup successfully" Feb 13 19:19:40.219670 containerd[1515]: time="2025-02-13T19:19:40.218935297Z" level=info msg="TearDown network for sandbox \"e941b39f4d4a62064805d7ae584462196d1704f1eaaa88f9ee2cc17abc84598e\" successfully" Feb 13 19:19:40.219670 containerd[1515]: time="2025-02-13T19:19:40.218947230Z" level=info msg="StopPodSandbox for \"e941b39f4d4a62064805d7ae584462196d1704f1eaaa88f9ee2cc17abc84598e\" returns successfully" Feb 13 19:19:40.219670 containerd[1515]: time="2025-02-13T19:19:40.219111619Z" level=info msg="StopPodSandbox for \"d195bcf8f9551fdb26f6b42c8b0e4b63d6953f1bbdc1c4b497262fc12be34248\"" Feb 13 19:19:40.219670 containerd[1515]: time="2025-02-13T19:19:40.219181530Z" level=info msg="TearDown network for sandbox \"d195bcf8f9551fdb26f6b42c8b0e4b63d6953f1bbdc1c4b497262fc12be34248\" successfully" Feb 13 19:19:40.219670 containerd[1515]: time="2025-02-13T19:19:40.219204143Z" level=info msg="StopPodSandbox for \"d195bcf8f9551fdb26f6b42c8b0e4b63d6953f1bbdc1c4b497262fc12be34248\" returns successfully" Feb 13 19:19:40.221009 containerd[1515]: time="2025-02-13T19:19:40.220283543Z" level=info msg="StopPodSandbox for \"913a65e5860c4b13ea05638c4cdf916cfe4d5179d77f6f08a60f9aa7332fc559\"" Feb 13 19:19:40.221009 containerd[1515]: time="2025-02-13T19:19:40.220330771Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86475b859c-swqrn,Uid:5a39efb7-aca1-4893-a36f-840c3c504ece,Namespace:calico-apiserver,Attempt:6,}" Feb 13 19:19:40.221009 containerd[1515]: time="2025-02-13T19:19:40.220363093Z" level=info msg="TearDown network for sandbox \"913a65e5860c4b13ea05638c4cdf916cfe4d5179d77f6f08a60f9aa7332fc559\" successfully" Feb 13 19:19:40.221009 containerd[1515]: time="2025-02-13T19:19:40.220373893Z" level=info msg="StopPodSandbox for \"913a65e5860c4b13ea05638c4cdf916cfe4d5179d77f6f08a60f9aa7332fc559\" returns successfully" Feb 13 19:19:40.221125 kubelet[2630]: I0213 19:19:40.220676 2630 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-4n8v4" podStartSLOduration=1.221086824 podStartE2EDuration="18.220663878s" podCreationTimestamp="2025-02-13 19:19:22 +0000 UTC" firstStartedPulling="2025-02-13 19:19:23.010045646 +0000 UTC m=+11.431898737" lastFinishedPulling="2025-02-13 19:19:40.009622701 +0000 UTC m=+28.431475791" observedRunningTime="2025-02-13 19:19:40.218393679 +0000 UTC m=+28.640246779" watchObservedRunningTime="2025-02-13 19:19:40.220663878 +0000 UTC m=+28.642516968" Feb 13 19:19:40.222924 containerd[1515]: time="2025-02-13T19:19:40.222514679Z" level=info msg="StopPodSandbox for \"fd762dfcc2c780c334b7852f3da6760c5d53d6cb6d243b5ae4cf1085971b8f57\"" Feb 13 19:19:40.223182 containerd[1515]: time="2025-02-13T19:19:40.223158781Z" level=info msg="TearDown network for sandbox \"fd762dfcc2c780c334b7852f3da6760c5d53d6cb6d243b5ae4cf1085971b8f57\" successfully" Feb 13 19:19:40.223454 containerd[1515]: time="2025-02-13T19:19:40.223267446Z" level=info msg="StopPodSandbox for \"fd762dfcc2c780c334b7852f3da6760c5d53d6cb6d243b5ae4cf1085971b8f57\" returns successfully" Feb 13 19:19:40.225547 containerd[1515]: time="2025-02-13T19:19:40.225495115Z" level=info msg="StopPodSandbox for \"8509c778c3294467a0ad8b8059bdecd145729001ee04e3c09d32d7f0d60b9468\"" Feb 13 19:19:40.225714 containerd[1515]: time="2025-02-13T19:19:40.225589522Z" level=info msg="TearDown network for sandbox \"8509c778c3294467a0ad8b8059bdecd145729001ee04e3c09d32d7f0d60b9468\" successfully" Feb 13 19:19:40.225714 containerd[1515]: time="2025-02-13T19:19:40.225605292Z" level=info msg="StopPodSandbox for \"8509c778c3294467a0ad8b8059bdecd145729001ee04e3c09d32d7f0d60b9468\" returns successfully" Feb 13 19:19:40.227017 containerd[1515]: time="2025-02-13T19:19:40.226896331Z" level=info msg="StopPodSandbox for \"706e09ab546bf3d89157cd9a2c32b16d198260f6686345817b1acf65f680fdd2\"" Feb 13 19:19:40.227017 containerd[1515]: time="2025-02-13T19:19:40.226976832Z" level=info msg="TearDown network for sandbox \"706e09ab546bf3d89157cd9a2c32b16d198260f6686345817b1acf65f680fdd2\" successfully" Feb 13 19:19:40.227017 containerd[1515]: time="2025-02-13T19:19:40.226986140Z" level=info msg="StopPodSandbox for \"706e09ab546bf3d89157cd9a2c32b16d198260f6686345817b1acf65f680fdd2\" returns successfully" Feb 13 19:19:40.227664 containerd[1515]: time="2025-02-13T19:19:40.227500487Z" level=info msg="StopPodSandbox for \"e2995529ca797998e1cf161d883bb684a3464a8a8d57602758f8705313f0d601\"" Feb 13 19:19:40.227664 containerd[1515]: time="2025-02-13T19:19:40.227589715Z" level=info msg="TearDown network for sandbox \"e2995529ca797998e1cf161d883bb684a3464a8a8d57602758f8705313f0d601\" successfully" Feb 13 19:19:40.227664 containerd[1515]: time="2025-02-13T19:19:40.227603951Z" level=info msg="StopPodSandbox for \"e2995529ca797998e1cf161d883bb684a3464a8a8d57602758f8705313f0d601\" returns successfully" Feb 13 19:19:40.228009 kubelet[2630]: E0213 19:19:40.227986 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:40.228740 containerd[1515]: time="2025-02-13T19:19:40.228659517Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-s8264,Uid:f2fdfd8a-780b-4320-8f24-ca003c5d0c77,Namespace:kube-system,Attempt:6,}" Feb 13 19:19:40.229786 kubelet[2630]: I0213 19:19:40.229479 2630 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45253c1f9735dfb2a3494283160983fb9ed43ea60532211ecb2b069637dde093" Feb 13 19:19:40.230253 containerd[1515]: time="2025-02-13T19:19:40.230074819Z" level=info msg="StopPodSandbox for \"45253c1f9735dfb2a3494283160983fb9ed43ea60532211ecb2b069637dde093\"" Feb 13 19:19:40.230897 containerd[1515]: time="2025-02-13T19:19:40.230877929Z" level=info msg="Ensure that sandbox 45253c1f9735dfb2a3494283160983fb9ed43ea60532211ecb2b069637dde093 in task-service has been cleanup successfully" Feb 13 19:19:40.231335 containerd[1515]: time="2025-02-13T19:19:40.231317456Z" level=info msg="TearDown network for sandbox \"45253c1f9735dfb2a3494283160983fb9ed43ea60532211ecb2b069637dde093\" successfully" Feb 13 19:19:40.231391 containerd[1515]: time="2025-02-13T19:19:40.231379192Z" level=info msg="StopPodSandbox for \"45253c1f9735dfb2a3494283160983fb9ed43ea60532211ecb2b069637dde093\" returns successfully" Feb 13 19:19:40.231715 containerd[1515]: time="2025-02-13T19:19:40.231691309Z" level=info msg="StopPodSandbox for \"9cfc4e8026fe9d3540dd0b1d209b241ede8d2453bfec90606b125f81c566e640\"" Feb 13 19:19:40.231866 containerd[1515]: time="2025-02-13T19:19:40.231844357Z" level=info msg="TearDown network for sandbox \"9cfc4e8026fe9d3540dd0b1d209b241ede8d2453bfec90606b125f81c566e640\" successfully" Feb 13 19:19:40.231866 containerd[1515]: time="2025-02-13T19:19:40.231859326Z" level=info msg="StopPodSandbox for \"9cfc4e8026fe9d3540dd0b1d209b241ede8d2453bfec90606b125f81c566e640\" returns successfully" Feb 13 19:19:40.232168 containerd[1515]: time="2025-02-13T19:19:40.232119434Z" level=info msg="StopPodSandbox for \"334cbee9ef51499256254b9201c8014a8e1e8be612d9ae143735bc19f2a5f6a4\"" Feb 13 19:19:40.232352 containerd[1515]: time="2025-02-13T19:19:40.232292059Z" level=info msg="TearDown network for sandbox \"334cbee9ef51499256254b9201c8014a8e1e8be612d9ae143735bc19f2a5f6a4\" successfully" Feb 13 19:19:40.232352 containerd[1515]: time="2025-02-13T19:19:40.232302529Z" level=info msg="StopPodSandbox for \"334cbee9ef51499256254b9201c8014a8e1e8be612d9ae143735bc19f2a5f6a4\" returns successfully" Feb 13 19:19:40.232756 containerd[1515]: time="2025-02-13T19:19:40.232720646Z" level=info msg="StopPodSandbox for \"e5e6531fa4d44fae3b14be640ca5608599c77b930300d189004cbcc636c08e1b\"" Feb 13 19:19:40.233085 containerd[1515]: time="2025-02-13T19:19:40.232956519Z" level=info msg="TearDown network for sandbox \"e5e6531fa4d44fae3b14be640ca5608599c77b930300d189004cbcc636c08e1b\" successfully" Feb 13 19:19:40.233085 containerd[1515]: time="2025-02-13T19:19:40.232991354Z" level=info msg="StopPodSandbox for \"e5e6531fa4d44fae3b14be640ca5608599c77b930300d189004cbcc636c08e1b\" returns successfully" Feb 13 19:19:40.233676 containerd[1515]: time="2025-02-13T19:19:40.233550726Z" level=info msg="StopPodSandbox for \"4304f46dfb22ea4aced73143fa3f59c4ed426a670211c8559cc8ed545f1b3804\"" Feb 13 19:19:40.233676 containerd[1515]: time="2025-02-13T19:19:40.233642219Z" level=info msg="TearDown network for sandbox \"4304f46dfb22ea4aced73143fa3f59c4ed426a670211c8559cc8ed545f1b3804\" successfully" Feb 13 19:19:40.233676 containerd[1515]: time="2025-02-13T19:19:40.233652428Z" level=info msg="StopPodSandbox for \"4304f46dfb22ea4aced73143fa3f59c4ed426a670211c8559cc8ed545f1b3804\" returns successfully" Feb 13 19:19:40.234300 containerd[1515]: time="2025-02-13T19:19:40.234279567Z" level=info msg="StopPodSandbox for \"aa373a8af75a289d061969e5a22e647e175a6a451ba7c421e5c44fa363ec90f8\"" Feb 13 19:19:40.234575 containerd[1515]: time="2025-02-13T19:19:40.234526532Z" level=info msg="TearDown network for sandbox \"aa373a8af75a289d061969e5a22e647e175a6a451ba7c421e5c44fa363ec90f8\" successfully" Feb 13 19:19:40.234575 containerd[1515]: time="2025-02-13T19:19:40.234539407Z" level=info msg="StopPodSandbox for \"aa373a8af75a289d061969e5a22e647e175a6a451ba7c421e5c44fa363ec90f8\" returns successfully" Feb 13 19:19:40.235338 containerd[1515]: time="2025-02-13T19:19:40.235269370Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-664f6f5454-mnbf7,Uid:4e40d638-ead9-4cd4-859c-b3835db75e0c,Namespace:calico-system,Attempt:6,}" Feb 13 19:19:40.264289 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Feb 13 19:19:40.264377 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Feb 13 19:19:40.407008 systemd[1]: run-netns-cni\x2d41b71d1b\x2db147\x2d4987\x2da73e\x2d31c1558d14c7.mount: Deactivated successfully. Feb 13 19:19:40.409343 systemd[1]: run-netns-cni\x2df6912c4e\x2d6bce\x2d846c\x2d0ce5\x2d668dc696f718.mount: Deactivated successfully. Feb 13 19:19:40.409420 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-e941b39f4d4a62064805d7ae584462196d1704f1eaaa88f9ee2cc17abc84598e-shm.mount: Deactivated successfully. Feb 13 19:19:40.409507 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-216ea206995c57165c7ec894ec603cac89445bad8cd05143541089c3e5d6db12-shm.mount: Deactivated successfully. Feb 13 19:19:40.409581 systemd[1]: run-netns-cni\x2d7401e3d3\x2d3ff4\x2dcd93\x2d3271\x2da49e5907e710.mount: Deactivated successfully. Feb 13 19:19:40.409655 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-41f8a27c833a329de6525bb7ea4ae60918c875126c9b0eb7529120508e6289e1-shm.mount: Deactivated successfully. Feb 13 19:19:40.409730 systemd[1]: run-netns-cni\x2de6cf1900\x2d4d87\x2df2ce\x2df985\x2dcdfa9ba1dd41.mount: Deactivated successfully. Feb 13 19:19:40.409799 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-535ac305f91d459b29fcac754f456e877bb89217b5f89ed4a0f736398ab4e183-shm.mount: Deactivated successfully. Feb 13 19:19:40.409874 systemd[1]: run-netns-cni\x2df6091b33\x2d0db5\x2defa2\x2d9781\x2d3cf59365f354.mount: Deactivated successfully. Feb 13 19:19:40.409944 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-45253c1f9735dfb2a3494283160983fb9ed43ea60532211ecb2b069637dde093-shm.mount: Deactivated successfully. Feb 13 19:19:40.410017 systemd[1]: run-netns-cni\x2d834df11a\x2d65d4\x2d555e\x2dd223\x2dae6487c72441.mount: Deactivated successfully. Feb 13 19:19:40.410087 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-4d2e4f99dab68adc0de82a705f6ff963ae519cfd54bd2435b6836ee53a8c4cf9-shm.mount: Deactivated successfully. Feb 13 19:19:40.509165 containerd[1515]: 2025-02-13 19:19:40.433 [INFO][4809] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="8cd7633242013303826371ffd207e7636573b1425696065f0f279c08885ceb81" Feb 13 19:19:40.509165 containerd[1515]: 2025-02-13 19:19:40.433 [INFO][4809] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="8cd7633242013303826371ffd207e7636573b1425696065f0f279c08885ceb81" iface="eth0" netns="/var/run/netns/cni-af4560f4-fd12-5260-b8e8-5c85cfdb511c" Feb 13 19:19:40.509165 containerd[1515]: 2025-02-13 19:19:40.434 [INFO][4809] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="8cd7633242013303826371ffd207e7636573b1425696065f0f279c08885ceb81" iface="eth0" netns="/var/run/netns/cni-af4560f4-fd12-5260-b8e8-5c85cfdb511c" Feb 13 19:19:40.509165 containerd[1515]: 2025-02-13 19:19:40.434 [INFO][4809] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="8cd7633242013303826371ffd207e7636573b1425696065f0f279c08885ceb81" iface="eth0" netns="/var/run/netns/cni-af4560f4-fd12-5260-b8e8-5c85cfdb511c" Feb 13 19:19:40.509165 containerd[1515]: 2025-02-13 19:19:40.434 [INFO][4809] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="8cd7633242013303826371ffd207e7636573b1425696065f0f279c08885ceb81" Feb 13 19:19:40.509165 containerd[1515]: 2025-02-13 19:19:40.434 [INFO][4809] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="8cd7633242013303826371ffd207e7636573b1425696065f0f279c08885ceb81" Feb 13 19:19:40.509165 containerd[1515]: 2025-02-13 19:19:40.492 [INFO][4880] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="8cd7633242013303826371ffd207e7636573b1425696065f0f279c08885ceb81" HandleID="k8s-pod-network.8cd7633242013303826371ffd207e7636573b1425696065f0f279c08885ceb81" Workload="localhost-k8s-coredns--6f6b679f8f--frtk4-eth0" Feb 13 19:19:40.509165 containerd[1515]: 2025-02-13 19:19:40.492 [INFO][4880] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Feb 13 19:19:40.509165 containerd[1515]: 2025-02-13 19:19:40.492 [INFO][4880] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Feb 13 19:19:40.509165 containerd[1515]: 2025-02-13 19:19:40.500 [WARNING][4880] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="8cd7633242013303826371ffd207e7636573b1425696065f0f279c08885ceb81" HandleID="k8s-pod-network.8cd7633242013303826371ffd207e7636573b1425696065f0f279c08885ceb81" Workload="localhost-k8s-coredns--6f6b679f8f--frtk4-eth0" Feb 13 19:19:40.509165 containerd[1515]: 2025-02-13 19:19:40.500 [INFO][4880] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="8cd7633242013303826371ffd207e7636573b1425696065f0f279c08885ceb81" HandleID="k8s-pod-network.8cd7633242013303826371ffd207e7636573b1425696065f0f279c08885ceb81" Workload="localhost-k8s-coredns--6f6b679f8f--frtk4-eth0" Feb 13 19:19:40.509165 containerd[1515]: 2025-02-13 19:19:40.502 [INFO][4880] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Feb 13 19:19:40.509165 containerd[1515]: 2025-02-13 19:19:40.504 [INFO][4809] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="8cd7633242013303826371ffd207e7636573b1425696065f0f279c08885ceb81" Feb 13 19:19:40.512884 systemd[1]: run-netns-cni\x2daf4560f4\x2dfd12\x2d5260\x2db8e8\x2d5c85cfdb511c.mount: Deactivated successfully. Feb 13 19:19:40.513010 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-8cd7633242013303826371ffd207e7636573b1425696065f0f279c08885ceb81-shm.mount: Deactivated successfully. Feb 13 19:19:40.526117 containerd[1515]: 2025-02-13 19:19:40.387 [INFO][4824] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="8948cb49fa01c71bbaa0121b2db137701a7a9b09a5bc374f63c2f6562d9f53a5" Feb 13 19:19:40.526117 containerd[1515]: 2025-02-13 19:19:40.388 [INFO][4824] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="8948cb49fa01c71bbaa0121b2db137701a7a9b09a5bc374f63c2f6562d9f53a5" iface="eth0" netns="/var/run/netns/cni-9b2397f6-296a-cbcb-89be-e30d6ab9788f" Feb 13 19:19:40.526117 containerd[1515]: 2025-02-13 19:19:40.389 [INFO][4824] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="8948cb49fa01c71bbaa0121b2db137701a7a9b09a5bc374f63c2f6562d9f53a5" iface="eth0" netns="/var/run/netns/cni-9b2397f6-296a-cbcb-89be-e30d6ab9788f" Feb 13 19:19:40.526117 containerd[1515]: 2025-02-13 19:19:40.390 [INFO][4824] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="8948cb49fa01c71bbaa0121b2db137701a7a9b09a5bc374f63c2f6562d9f53a5" iface="eth0" netns="/var/run/netns/cni-9b2397f6-296a-cbcb-89be-e30d6ab9788f" Feb 13 19:19:40.526117 containerd[1515]: 2025-02-13 19:19:40.390 [INFO][4824] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="8948cb49fa01c71bbaa0121b2db137701a7a9b09a5bc374f63c2f6562d9f53a5" Feb 13 19:19:40.526117 containerd[1515]: 2025-02-13 19:19:40.390 [INFO][4824] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="8948cb49fa01c71bbaa0121b2db137701a7a9b09a5bc374f63c2f6562d9f53a5" Feb 13 19:19:40.526117 containerd[1515]: 2025-02-13 19:19:40.507 [INFO][4853] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="8948cb49fa01c71bbaa0121b2db137701a7a9b09a5bc374f63c2f6562d9f53a5" HandleID="k8s-pod-network.8948cb49fa01c71bbaa0121b2db137701a7a9b09a5bc374f63c2f6562d9f53a5" Workload="localhost-k8s-csi--node--driver--drlnv-eth0" Feb 13 19:19:40.526117 containerd[1515]: 2025-02-13 19:19:40.507 [INFO][4853] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Feb 13 19:19:40.526117 containerd[1515]: 2025-02-13 19:19:40.508 [INFO][4853] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Feb 13 19:19:40.526117 containerd[1515]: 2025-02-13 19:19:40.519 [WARNING][4853] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="8948cb49fa01c71bbaa0121b2db137701a7a9b09a5bc374f63c2f6562d9f53a5" HandleID="k8s-pod-network.8948cb49fa01c71bbaa0121b2db137701a7a9b09a5bc374f63c2f6562d9f53a5" Workload="localhost-k8s-csi--node--driver--drlnv-eth0" Feb 13 19:19:40.526117 containerd[1515]: 2025-02-13 19:19:40.520 [INFO][4853] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="8948cb49fa01c71bbaa0121b2db137701a7a9b09a5bc374f63c2f6562d9f53a5" HandleID="k8s-pod-network.8948cb49fa01c71bbaa0121b2db137701a7a9b09a5bc374f63c2f6562d9f53a5" Workload="localhost-k8s-csi--node--driver--drlnv-eth0" Feb 13 19:19:40.526117 containerd[1515]: 2025-02-13 19:19:40.521 [INFO][4853] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Feb 13 19:19:40.526117 containerd[1515]: 2025-02-13 19:19:40.523 [INFO][4824] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="8948cb49fa01c71bbaa0121b2db137701a7a9b09a5bc374f63c2f6562d9f53a5" Feb 13 19:19:40.528296 systemd[1]: run-netns-cni\x2d9b2397f6\x2d296a\x2dcbcb\x2d89be\x2de30d6ab9788f.mount: Deactivated successfully. Feb 13 19:19:40.528402 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-8948cb49fa01c71bbaa0121b2db137701a7a9b09a5bc374f63c2f6562d9f53a5-shm.mount: Deactivated successfully. Feb 13 19:19:40.925895 containerd[1515]: time="2025-02-13T19:19:40.925837557Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-drlnv,Uid:84096616-cf79-421c-9e94-8cbcf0a204a5,Namespace:calico-system,Attempt:5,} failed, error" error="failed to setup network for sandbox \"8948cb49fa01c71bbaa0121b2db137701a7a9b09a5bc374f63c2f6562d9f53a5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:40.926138 kubelet[2630]: E0213 19:19:40.926096 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8948cb49fa01c71bbaa0121b2db137701a7a9b09a5bc374f63c2f6562d9f53a5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:40.926269 kubelet[2630]: E0213 19:19:40.926157 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8948cb49fa01c71bbaa0121b2db137701a7a9b09a5bc374f63c2f6562d9f53a5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-drlnv" Feb 13 19:19:40.926269 kubelet[2630]: E0213 19:19:40.926216 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8948cb49fa01c71bbaa0121b2db137701a7a9b09a5bc374f63c2f6562d9f53a5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-drlnv" Feb 13 19:19:40.926319 kubelet[2630]: E0213 19:19:40.926260 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-drlnv_calico-system(84096616-cf79-421c-9e94-8cbcf0a204a5)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-drlnv_calico-system(84096616-cf79-421c-9e94-8cbcf0a204a5)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"8948cb49fa01c71bbaa0121b2db137701a7a9b09a5bc374f63c2f6562d9f53a5\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-drlnv" podUID="84096616-cf79-421c-9e94-8cbcf0a204a5" Feb 13 19:19:40.927479 containerd[1515]: time="2025-02-13T19:19:40.927333712Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-frtk4,Uid:f48c8839-9a17-43c4-a0d5-f9e8d9767814,Namespace:kube-system,Attempt:6,} failed, error" error="failed to setup network for sandbox \"8cd7633242013303826371ffd207e7636573b1425696065f0f279c08885ceb81\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:40.927791 kubelet[2630]: E0213 19:19:40.927621 2630 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8cd7633242013303826371ffd207e7636573b1425696065f0f279c08885ceb81\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Feb 13 19:19:40.927791 kubelet[2630]: E0213 19:19:40.927676 2630 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8cd7633242013303826371ffd207e7636573b1425696065f0f279c08885ceb81\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-frtk4" Feb 13 19:19:40.927791 kubelet[2630]: E0213 19:19:40.927700 2630 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8cd7633242013303826371ffd207e7636573b1425696065f0f279c08885ceb81\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-frtk4" Feb 13 19:19:40.928100 kubelet[2630]: E0213 19:19:40.927748 2630 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-6f6b679f8f-frtk4_kube-system(f48c8839-9a17-43c4-a0d5-f9e8d9767814)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-6f6b679f8f-frtk4_kube-system(f48c8839-9a17-43c4-a0d5-f9e8d9767814)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"8cd7633242013303826371ffd207e7636573b1425696065f0f279c08885ceb81\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-6f6b679f8f-frtk4" podUID="f48c8839-9a17-43c4-a0d5-f9e8d9767814" Feb 13 19:19:40.952877 systemd-networkd[1422]: cali8b6a1ad33a3: Link UP Feb 13 19:19:40.953667 systemd-networkd[1422]: cali8b6a1ad33a3: Gained carrier Feb 13 19:19:40.968686 containerd[1515]: 2025-02-13 19:19:40.346 [INFO][4785] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Feb 13 19:19:40.968686 containerd[1515]: 2025-02-13 19:19:40.376 [INFO][4785] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--kube--controllers--664f6f5454--mnbf7-eth0 calico-kube-controllers-664f6f5454- calico-system 4e40d638-ead9-4cd4-859c-b3835db75e0c 691 0 2025-02-13 19:19:22 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:664f6f5454 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s localhost calico-kube-controllers-664f6f5454-mnbf7 eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali8b6a1ad33a3 [] []}} ContainerID="46203ac9ba0fbbbb33ec052a083f76a39f1007ec3dc0a1ebe440dc0ca7551d85" Namespace="calico-system" Pod="calico-kube-controllers-664f6f5454-mnbf7" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--664f6f5454--mnbf7-" Feb 13 19:19:40.968686 containerd[1515]: 2025-02-13 19:19:40.379 [INFO][4785] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="46203ac9ba0fbbbb33ec052a083f76a39f1007ec3dc0a1ebe440dc0ca7551d85" Namespace="calico-system" Pod="calico-kube-controllers-664f6f5454-mnbf7" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--664f6f5454--mnbf7-eth0" Feb 13 19:19:40.968686 containerd[1515]: 2025-02-13 19:19:40.514 [INFO][4873] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="46203ac9ba0fbbbb33ec052a083f76a39f1007ec3dc0a1ebe440dc0ca7551d85" HandleID="k8s-pod-network.46203ac9ba0fbbbb33ec052a083f76a39f1007ec3dc0a1ebe440dc0ca7551d85" Workload="localhost-k8s-calico--kube--controllers--664f6f5454--mnbf7-eth0" Feb 13 19:19:40.968686 containerd[1515]: 2025-02-13 19:19:40.806 [INFO][4873] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="46203ac9ba0fbbbb33ec052a083f76a39f1007ec3dc0a1ebe440dc0ca7551d85" HandleID="k8s-pod-network.46203ac9ba0fbbbb33ec052a083f76a39f1007ec3dc0a1ebe440dc0ca7551d85" Workload="localhost-k8s-calico--kube--controllers--664f6f5454--mnbf7-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000399040), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"calico-kube-controllers-664f6f5454-mnbf7", "timestamp":"2025-02-13 19:19:40.514444196 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Feb 13 19:19:40.968686 containerd[1515]: 2025-02-13 19:19:40.806 [INFO][4873] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Feb 13 19:19:40.968686 containerd[1515]: 2025-02-13 19:19:40.806 [INFO][4873] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Feb 13 19:19:40.968686 containerd[1515]: 2025-02-13 19:19:40.806 [INFO][4873] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Feb 13 19:19:40.968686 containerd[1515]: 2025-02-13 19:19:40.809 [INFO][4873] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.46203ac9ba0fbbbb33ec052a083f76a39f1007ec3dc0a1ebe440dc0ca7551d85" host="localhost" Feb 13 19:19:40.968686 containerd[1515]: 2025-02-13 19:19:40.848 [INFO][4873] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" Feb 13 19:19:40.968686 containerd[1515]: 2025-02-13 19:19:40.852 [INFO][4873] ipam/ipam.go 521: Ran out of existing affine blocks for host host="localhost" Feb 13 19:19:40.968686 containerd[1515]: 2025-02-13 19:19:40.861 [INFO][4873] ipam/ipam.go 538: Tried all affine blocks. Looking for an affine block with space, or a new unclaimed block host="localhost" Feb 13 19:19:40.968686 containerd[1515]: 2025-02-13 19:19:40.877 [INFO][4873] ipam/ipam.go 550: Found unclaimed block host="localhost" subnet=192.168.88.128/26 Feb 13 19:19:40.968686 containerd[1515]: 2025-02-13 19:19:40.878 [INFO][4873] ipam/ipam_block_reader_writer.go 171: Trying to create affinity in pending state host="localhost" subnet=192.168.88.128/26 Feb 13 19:19:40.968686 containerd[1515]: 2025-02-13 19:19:40.906 [INFO][4873] ipam/ipam_block_reader_writer.go 182: Block affinity already exists, getting existing affinity host="localhost" subnet=192.168.88.128/26 Feb 13 19:19:40.968686 containerd[1515]: 2025-02-13 19:19:40.909 [INFO][4873] ipam/ipam_block_reader_writer.go 190: Got existing affinity host="localhost" subnet=192.168.88.128/26 Feb 13 19:19:40.968686 containerd[1515]: 2025-02-13 19:19:40.909 [INFO][4873] ipam/ipam_block_reader_writer.go 198: Existing affinity is already confirmed host="localhost" subnet=192.168.88.128/26 Feb 13 19:19:40.968686 containerd[1515]: 2025-02-13 19:19:40.909 [INFO][4873] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Feb 13 19:19:40.968686 containerd[1515]: 2025-02-13 19:19:40.910 [INFO][4873] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Feb 13 19:19:40.968686 containerd[1515]: 2025-02-13 19:19:40.910 [INFO][4873] ipam/ipam.go 585: Block '192.168.88.128/26' has 64 free ips which is more than 1 ips required. host="localhost" subnet=192.168.88.128/26 Feb 13 19:19:40.968686 containerd[1515]: 2025-02-13 19:19:40.910 [INFO][4873] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.46203ac9ba0fbbbb33ec052a083f76a39f1007ec3dc0a1ebe440dc0ca7551d85" host="localhost" Feb 13 19:19:40.968686 containerd[1515]: 2025-02-13 19:19:40.912 [INFO][4873] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.46203ac9ba0fbbbb33ec052a083f76a39f1007ec3dc0a1ebe440dc0ca7551d85 Feb 13 19:19:40.968686 containerd[1515]: 2025-02-13 19:19:40.917 [INFO][4873] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.46203ac9ba0fbbbb33ec052a083f76a39f1007ec3dc0a1ebe440dc0ca7551d85" host="localhost" Feb 13 19:19:40.971232 containerd[1515]: 2025-02-13 19:19:40.920 [ERROR][4873] ipam/customresource.go 184: Error updating resource Key=IPAMBlock(192-168-88-128-26) Name="192-168-88-128-26" Resource="IPAMBlocks" Value=&v3.IPAMBlock{TypeMeta:v1.TypeMeta{Kind:"IPAMBlock", APIVersion:"crd.projectcalico.org/v1"}, ObjectMeta:v1.ObjectMeta{Name:"192-168-88-128-26", GenerateName:"", Namespace:"", SelfLink:"", UID:"", ResourceVersion:"807", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.IPAMBlockSpec{CIDR:"192.168.88.128/26", Affinity:(*string)(0xc000265910), Allocations:[]*int{(*int)(0xc000267340), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil), (*int)(nil)}, Unallocated:[]int{1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63}, Attributes:[]v3.AllocationAttribute{v3.AllocationAttribute{AttrPrimary:(*string)(0xc000399040), AttrSecondary:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"calico-kube-controllers-664f6f5454-mnbf7", "timestamp":"2025-02-13 19:19:40.514444196 +0000 UTC"}}}, SequenceNumber:0x1823dabbb7741505, SequenceNumberForAllocation:map[string]uint64{"0":0x1823dabbb7741504}, Deleted:false, DeprecatedStrictAffinity:false}} error=Operation cannot be fulfilled on ipamblocks.crd.projectcalico.org "192-168-88-128-26": the object has been modified; please apply your changes to the latest version and try again Feb 13 19:19:40.971232 containerd[1515]: 2025-02-13 19:19:40.920 [INFO][4873] ipam/ipam.go 1207: Failed to update block block=192.168.88.128/26 error=update conflict: IPAMBlock(192-168-88-128-26) handle="k8s-pod-network.46203ac9ba0fbbbb33ec052a083f76a39f1007ec3dc0a1ebe440dc0ca7551d85" host="localhost" Feb 13 19:19:40.971232 containerd[1515]: 2025-02-13 19:19:40.935 [INFO][4873] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.46203ac9ba0fbbbb33ec052a083f76a39f1007ec3dc0a1ebe440dc0ca7551d85" host="localhost" Feb 13 19:19:40.971232 containerd[1515]: 2025-02-13 19:19:40.936 [INFO][4873] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.46203ac9ba0fbbbb33ec052a083f76a39f1007ec3dc0a1ebe440dc0ca7551d85 Feb 13 19:19:40.971232 containerd[1515]: 2025-02-13 19:19:40.939 [INFO][4873] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.46203ac9ba0fbbbb33ec052a083f76a39f1007ec3dc0a1ebe440dc0ca7551d85" host="localhost" Feb 13 19:19:40.971232 containerd[1515]: 2025-02-13 19:19:40.943 [INFO][4873] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.129/26] block=192.168.88.128/26 handle="k8s-pod-network.46203ac9ba0fbbbb33ec052a083f76a39f1007ec3dc0a1ebe440dc0ca7551d85" host="localhost" Feb 13 19:19:40.971232 containerd[1515]: 2025-02-13 19:19:40.943 [INFO][4873] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.129/26] handle="k8s-pod-network.46203ac9ba0fbbbb33ec052a083f76a39f1007ec3dc0a1ebe440dc0ca7551d85" host="localhost" Feb 13 19:19:40.971232 containerd[1515]: 2025-02-13 19:19:40.943 [INFO][4873] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Feb 13 19:19:40.971232 containerd[1515]: 2025-02-13 19:19:40.943 [INFO][4873] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.129/26] IPv6=[] ContainerID="46203ac9ba0fbbbb33ec052a083f76a39f1007ec3dc0a1ebe440dc0ca7551d85" HandleID="k8s-pod-network.46203ac9ba0fbbbb33ec052a083f76a39f1007ec3dc0a1ebe440dc0ca7551d85" Workload="localhost-k8s-calico--kube--controllers--664f6f5454--mnbf7-eth0" Feb 13 19:19:40.971232 containerd[1515]: 2025-02-13 19:19:40.945 [INFO][4785] cni-plugin/k8s.go 386: Populated endpoint ContainerID="46203ac9ba0fbbbb33ec052a083f76a39f1007ec3dc0a1ebe440dc0ca7551d85" Namespace="calico-system" Pod="calico-kube-controllers-664f6f5454-mnbf7" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--664f6f5454--mnbf7-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--664f6f5454--mnbf7-eth0", GenerateName:"calico-kube-controllers-664f6f5454-", Namespace:"calico-system", SelfLink:"", UID:"4e40d638-ead9-4cd4-859c-b3835db75e0c", ResourceVersion:"691", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 19, 19, 22, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"664f6f5454", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-kube-controllers-664f6f5454-mnbf7", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali8b6a1ad33a3", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 19:19:40.971535 containerd[1515]: 2025-02-13 19:19:40.946 [INFO][4785] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.129/32] ContainerID="46203ac9ba0fbbbb33ec052a083f76a39f1007ec3dc0a1ebe440dc0ca7551d85" Namespace="calico-system" Pod="calico-kube-controllers-664f6f5454-mnbf7" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--664f6f5454--mnbf7-eth0" Feb 13 19:19:40.971535 containerd[1515]: 2025-02-13 19:19:40.946 [INFO][4785] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali8b6a1ad33a3 ContainerID="46203ac9ba0fbbbb33ec052a083f76a39f1007ec3dc0a1ebe440dc0ca7551d85" Namespace="calico-system" Pod="calico-kube-controllers-664f6f5454-mnbf7" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--664f6f5454--mnbf7-eth0" Feb 13 19:19:40.971535 containerd[1515]: 2025-02-13 19:19:40.953 [INFO][4785] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="46203ac9ba0fbbbb33ec052a083f76a39f1007ec3dc0a1ebe440dc0ca7551d85" Namespace="calico-system" Pod="calico-kube-controllers-664f6f5454-mnbf7" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--664f6f5454--mnbf7-eth0" Feb 13 19:19:40.971535 containerd[1515]: 2025-02-13 19:19:40.953 [INFO][4785] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="46203ac9ba0fbbbb33ec052a083f76a39f1007ec3dc0a1ebe440dc0ca7551d85" Namespace="calico-system" Pod="calico-kube-controllers-664f6f5454-mnbf7" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--664f6f5454--mnbf7-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--664f6f5454--mnbf7-eth0", GenerateName:"calico-kube-controllers-664f6f5454-", Namespace:"calico-system", SelfLink:"", UID:"4e40d638-ead9-4cd4-859c-b3835db75e0c", ResourceVersion:"691", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 19, 19, 22, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"664f6f5454", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"46203ac9ba0fbbbb33ec052a083f76a39f1007ec3dc0a1ebe440dc0ca7551d85", Pod:"calico-kube-controllers-664f6f5454-mnbf7", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali8b6a1ad33a3", MAC:"5a:f1:10:98:02:52", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 19:19:40.971535 containerd[1515]: 2025-02-13 19:19:40.965 [INFO][4785] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="46203ac9ba0fbbbb33ec052a083f76a39f1007ec3dc0a1ebe440dc0ca7551d85" Namespace="calico-system" Pod="calico-kube-controllers-664f6f5454-mnbf7" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--664f6f5454--mnbf7-eth0" Feb 13 19:19:40.980882 systemd-networkd[1422]: calied5a4c5c7f2: Link UP Feb 13 19:19:40.981834 systemd-networkd[1422]: calied5a4c5c7f2: Gained carrier Feb 13 19:19:40.991874 containerd[1515]: 2025-02-13 19:19:40.359 [INFO][4781] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Feb 13 19:19:40.991874 containerd[1515]: 2025-02-13 19:19:40.379 [INFO][4781] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--6f6b679f8f--s8264-eth0 coredns-6f6b679f8f- kube-system f2fdfd8a-780b-4320-8f24-ca003c5d0c77 690 0 2025-02-13 19:19:17 +0000 UTC map[k8s-app:kube-dns pod-template-hash:6f6b679f8f projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-6f6b679f8f-s8264 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] calied5a4c5c7f2 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="143d63705a6cccab6c533569f1bb8f3f4c9da1f189033dff08d90b2c776e7d3d" Namespace="kube-system" Pod="coredns-6f6b679f8f-s8264" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--s8264-" Feb 13 19:19:40.991874 containerd[1515]: 2025-02-13 19:19:40.379 [INFO][4781] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="143d63705a6cccab6c533569f1bb8f3f4c9da1f189033dff08d90b2c776e7d3d" Namespace="kube-system" Pod="coredns-6f6b679f8f-s8264" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--s8264-eth0" Feb 13 19:19:40.991874 containerd[1515]: 2025-02-13 19:19:40.507 [INFO][4855] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="143d63705a6cccab6c533569f1bb8f3f4c9da1f189033dff08d90b2c776e7d3d" HandleID="k8s-pod-network.143d63705a6cccab6c533569f1bb8f3f4c9da1f189033dff08d90b2c776e7d3d" Workload="localhost-k8s-coredns--6f6b679f8f--s8264-eth0" Feb 13 19:19:40.991874 containerd[1515]: 2025-02-13 19:19:40.807 [INFO][4855] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="143d63705a6cccab6c533569f1bb8f3f4c9da1f189033dff08d90b2c776e7d3d" HandleID="k8s-pod-network.143d63705a6cccab6c533569f1bb8f3f4c9da1f189033dff08d90b2c776e7d3d" Workload="localhost-k8s-coredns--6f6b679f8f--s8264-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0001aa910), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-6f6b679f8f-s8264", "timestamp":"2025-02-13 19:19:40.50610433 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Feb 13 19:19:40.991874 containerd[1515]: 2025-02-13 19:19:40.807 [INFO][4855] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Feb 13 19:19:40.991874 containerd[1515]: 2025-02-13 19:19:40.943 [INFO][4855] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Feb 13 19:19:40.991874 containerd[1515]: 2025-02-13 19:19:40.943 [INFO][4855] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Feb 13 19:19:40.991874 containerd[1515]: 2025-02-13 19:19:40.945 [INFO][4855] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.143d63705a6cccab6c533569f1bb8f3f4c9da1f189033dff08d90b2c776e7d3d" host="localhost" Feb 13 19:19:40.991874 containerd[1515]: 2025-02-13 19:19:40.949 [INFO][4855] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" Feb 13 19:19:40.991874 containerd[1515]: 2025-02-13 19:19:40.954 [INFO][4855] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Feb 13 19:19:40.991874 containerd[1515]: 2025-02-13 19:19:40.956 [INFO][4855] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Feb 13 19:19:40.991874 containerd[1515]: 2025-02-13 19:19:40.958 [INFO][4855] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Feb 13 19:19:40.991874 containerd[1515]: 2025-02-13 19:19:40.958 [INFO][4855] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.143d63705a6cccab6c533569f1bb8f3f4c9da1f189033dff08d90b2c776e7d3d" host="localhost" Feb 13 19:19:40.991874 containerd[1515]: 2025-02-13 19:19:40.963 [INFO][4855] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.143d63705a6cccab6c533569f1bb8f3f4c9da1f189033dff08d90b2c776e7d3d Feb 13 19:19:40.991874 containerd[1515]: 2025-02-13 19:19:40.970 [INFO][4855] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.143d63705a6cccab6c533569f1bb8f3f4c9da1f189033dff08d90b2c776e7d3d" host="localhost" Feb 13 19:19:40.991874 containerd[1515]: 2025-02-13 19:19:40.975 [INFO][4855] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.130/26] block=192.168.88.128/26 handle="k8s-pod-network.143d63705a6cccab6c533569f1bb8f3f4c9da1f189033dff08d90b2c776e7d3d" host="localhost" Feb 13 19:19:40.991874 containerd[1515]: 2025-02-13 19:19:40.975 [INFO][4855] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.130/26] handle="k8s-pod-network.143d63705a6cccab6c533569f1bb8f3f4c9da1f189033dff08d90b2c776e7d3d" host="localhost" Feb 13 19:19:40.991874 containerd[1515]: 2025-02-13 19:19:40.975 [INFO][4855] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Feb 13 19:19:40.991874 containerd[1515]: 2025-02-13 19:19:40.975 [INFO][4855] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.130/26] IPv6=[] ContainerID="143d63705a6cccab6c533569f1bb8f3f4c9da1f189033dff08d90b2c776e7d3d" HandleID="k8s-pod-network.143d63705a6cccab6c533569f1bb8f3f4c9da1f189033dff08d90b2c776e7d3d" Workload="localhost-k8s-coredns--6f6b679f8f--s8264-eth0" Feb 13 19:19:40.992559 containerd[1515]: 2025-02-13 19:19:40.978 [INFO][4781] cni-plugin/k8s.go 386: Populated endpoint ContainerID="143d63705a6cccab6c533569f1bb8f3f4c9da1f189033dff08d90b2c776e7d3d" Namespace="kube-system" Pod="coredns-6f6b679f8f-s8264" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--s8264-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--6f6b679f8f--s8264-eth0", GenerateName:"coredns-6f6b679f8f-", Namespace:"kube-system", SelfLink:"", UID:"f2fdfd8a-780b-4320-8f24-ca003c5d0c77", ResourceVersion:"690", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 19, 19, 17, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"6f6b679f8f", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-6f6b679f8f-s8264", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calied5a4c5c7f2", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 19:19:40.992559 containerd[1515]: 2025-02-13 19:19:40.978 [INFO][4781] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.130/32] ContainerID="143d63705a6cccab6c533569f1bb8f3f4c9da1f189033dff08d90b2c776e7d3d" Namespace="kube-system" Pod="coredns-6f6b679f8f-s8264" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--s8264-eth0" Feb 13 19:19:40.992559 containerd[1515]: 2025-02-13 19:19:40.978 [INFO][4781] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calied5a4c5c7f2 ContainerID="143d63705a6cccab6c533569f1bb8f3f4c9da1f189033dff08d90b2c776e7d3d" Namespace="kube-system" Pod="coredns-6f6b679f8f-s8264" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--s8264-eth0" Feb 13 19:19:40.992559 containerd[1515]: 2025-02-13 19:19:40.981 [INFO][4781] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="143d63705a6cccab6c533569f1bb8f3f4c9da1f189033dff08d90b2c776e7d3d" Namespace="kube-system" Pod="coredns-6f6b679f8f-s8264" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--s8264-eth0" Feb 13 19:19:40.992559 containerd[1515]: 2025-02-13 19:19:40.981 [INFO][4781] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="143d63705a6cccab6c533569f1bb8f3f4c9da1f189033dff08d90b2c776e7d3d" Namespace="kube-system" Pod="coredns-6f6b679f8f-s8264" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--s8264-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--6f6b679f8f--s8264-eth0", GenerateName:"coredns-6f6b679f8f-", Namespace:"kube-system", SelfLink:"", UID:"f2fdfd8a-780b-4320-8f24-ca003c5d0c77", ResourceVersion:"690", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 19, 19, 17, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"6f6b679f8f", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"143d63705a6cccab6c533569f1bb8f3f4c9da1f189033dff08d90b2c776e7d3d", Pod:"coredns-6f6b679f8f-s8264", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calied5a4c5c7f2", MAC:"9e:75:d8:1f:44:94", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 19:19:40.992559 containerd[1515]: 2025-02-13 19:19:40.988 [INFO][4781] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="143d63705a6cccab6c533569f1bb8f3f4c9da1f189033dff08d90b2c776e7d3d" Namespace="kube-system" Pod="coredns-6f6b679f8f-s8264" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--s8264-eth0" Feb 13 19:19:41.037084 containerd[1515]: time="2025-02-13T19:19:41.036933611Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:19:41.037084 containerd[1515]: time="2025-02-13T19:19:41.037031024Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:19:41.037084 containerd[1515]: time="2025-02-13T19:19:41.037047484Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:19:41.037326 containerd[1515]: time="2025-02-13T19:19:41.037155378Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:19:41.050868 containerd[1515]: time="2025-02-13T19:19:41.050563892Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:19:41.050868 containerd[1515]: time="2025-02-13T19:19:41.050632010Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:19:41.050868 containerd[1515]: time="2025-02-13T19:19:41.050650224Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:19:41.052756 containerd[1515]: time="2025-02-13T19:19:41.052667809Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:19:41.058823 systemd[1]: Started cri-containerd-143d63705a6cccab6c533569f1bb8f3f4c9da1f189033dff08d90b2c776e7d3d.scope - libcontainer container 143d63705a6cccab6c533569f1bb8f3f4c9da1f189033dff08d90b2c776e7d3d. Feb 13 19:19:41.068146 systemd[1]: Started cri-containerd-46203ac9ba0fbbbb33ec052a083f76a39f1007ec3dc0a1ebe440dc0ca7551d85.scope - libcontainer container 46203ac9ba0fbbbb33ec052a083f76a39f1007ec3dc0a1ebe440dc0ca7551d85. Feb 13 19:19:41.073231 systemd-resolved[1350]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Feb 13 19:19:41.080788 systemd-resolved[1350]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Feb 13 19:19:41.099503 containerd[1515]: time="2025-02-13T19:19:41.099448039Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-s8264,Uid:f2fdfd8a-780b-4320-8f24-ca003c5d0c77,Namespace:kube-system,Attempt:6,} returns sandbox id \"143d63705a6cccab6c533569f1bb8f3f4c9da1f189033dff08d90b2c776e7d3d\"" Feb 13 19:19:41.101011 kubelet[2630]: E0213 19:19:41.100655 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:41.102903 containerd[1515]: time="2025-02-13T19:19:41.102872579Z" level=info msg="CreateContainer within sandbox \"143d63705a6cccab6c533569f1bb8f3f4c9da1f189033dff08d90b2c776e7d3d\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Feb 13 19:19:41.107144 systemd-networkd[1422]: cali562458c7813: Link UP Feb 13 19:19:41.107950 containerd[1515]: time="2025-02-13T19:19:41.107878082Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-664f6f5454-mnbf7,Uid:4e40d638-ead9-4cd4-859c-b3835db75e0c,Namespace:calico-system,Attempt:6,} returns sandbox id \"46203ac9ba0fbbbb33ec052a083f76a39f1007ec3dc0a1ebe440dc0ca7551d85\"" Feb 13 19:19:41.108651 systemd-networkd[1422]: cali562458c7813: Gained carrier Feb 13 19:19:41.110005 containerd[1515]: time="2025-02-13T19:19:41.109985736Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.29.1\"" Feb 13 19:19:41.231295 containerd[1515]: 2025-02-13 19:19:40.340 [INFO][4741] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Feb 13 19:19:41.231295 containerd[1515]: 2025-02-13 19:19:40.383 [INFO][4741] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--86475b859c--kwlfl-eth0 calico-apiserver-86475b859c- calico-apiserver f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1 693 0 2025-02-13 19:19:22 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:86475b859c projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-86475b859c-kwlfl eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali562458c7813 [] []}} ContainerID="b9a6fd62cd77a87bbc457ad2d1331edef05b0ed9b12abaf49f0386fbdf40fa17" Namespace="calico-apiserver" Pod="calico-apiserver-86475b859c-kwlfl" WorkloadEndpoint="localhost-k8s-calico--apiserver--86475b859c--kwlfl-" Feb 13 19:19:41.231295 containerd[1515]: 2025-02-13 19:19:40.383 [INFO][4741] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="b9a6fd62cd77a87bbc457ad2d1331edef05b0ed9b12abaf49f0386fbdf40fa17" Namespace="calico-apiserver" Pod="calico-apiserver-86475b859c-kwlfl" WorkloadEndpoint="localhost-k8s-calico--apiserver--86475b859c--kwlfl-eth0" Feb 13 19:19:41.231295 containerd[1515]: 2025-02-13 19:19:40.507 [INFO][4864] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="b9a6fd62cd77a87bbc457ad2d1331edef05b0ed9b12abaf49f0386fbdf40fa17" HandleID="k8s-pod-network.b9a6fd62cd77a87bbc457ad2d1331edef05b0ed9b12abaf49f0386fbdf40fa17" Workload="localhost-k8s-calico--apiserver--86475b859c--kwlfl-eth0" Feb 13 19:19:41.231295 containerd[1515]: 2025-02-13 19:19:40.806 [INFO][4864] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="b9a6fd62cd77a87bbc457ad2d1331edef05b0ed9b12abaf49f0386fbdf40fa17" HandleID="k8s-pod-network.b9a6fd62cd77a87bbc457ad2d1331edef05b0ed9b12abaf49f0386fbdf40fa17" Workload="localhost-k8s-calico--apiserver--86475b859c--kwlfl-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000442330), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-86475b859c-kwlfl", "timestamp":"2025-02-13 19:19:40.506973816 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Feb 13 19:19:41.231295 containerd[1515]: 2025-02-13 19:19:40.807 [INFO][4864] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Feb 13 19:19:41.231295 containerd[1515]: 2025-02-13 19:19:40.975 [INFO][4864] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Feb 13 19:19:41.231295 containerd[1515]: 2025-02-13 19:19:40.975 [INFO][4864] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Feb 13 19:19:41.231295 containerd[1515]: 2025-02-13 19:19:41.046 [INFO][4864] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.b9a6fd62cd77a87bbc457ad2d1331edef05b0ed9b12abaf49f0386fbdf40fa17" host="localhost" Feb 13 19:19:41.231295 containerd[1515]: 2025-02-13 19:19:41.052 [INFO][4864] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" Feb 13 19:19:41.231295 containerd[1515]: 2025-02-13 19:19:41.055 [INFO][4864] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Feb 13 19:19:41.231295 containerd[1515]: 2025-02-13 19:19:41.057 [INFO][4864] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Feb 13 19:19:41.231295 containerd[1515]: 2025-02-13 19:19:41.059 [INFO][4864] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Feb 13 19:19:41.231295 containerd[1515]: 2025-02-13 19:19:41.059 [INFO][4864] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.b9a6fd62cd77a87bbc457ad2d1331edef05b0ed9b12abaf49f0386fbdf40fa17" host="localhost" Feb 13 19:19:41.231295 containerd[1515]: 2025-02-13 19:19:41.061 [INFO][4864] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.b9a6fd62cd77a87bbc457ad2d1331edef05b0ed9b12abaf49f0386fbdf40fa17 Feb 13 19:19:41.231295 containerd[1515]: 2025-02-13 19:19:41.067 [INFO][4864] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.b9a6fd62cd77a87bbc457ad2d1331edef05b0ed9b12abaf49f0386fbdf40fa17" host="localhost" Feb 13 19:19:41.231295 containerd[1515]: 2025-02-13 19:19:41.100 [INFO][4864] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.131/26] block=192.168.88.128/26 handle="k8s-pod-network.b9a6fd62cd77a87bbc457ad2d1331edef05b0ed9b12abaf49f0386fbdf40fa17" host="localhost" Feb 13 19:19:41.231295 containerd[1515]: 2025-02-13 19:19:41.100 [INFO][4864] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.131/26] handle="k8s-pod-network.b9a6fd62cd77a87bbc457ad2d1331edef05b0ed9b12abaf49f0386fbdf40fa17" host="localhost" Feb 13 19:19:41.231295 containerd[1515]: 2025-02-13 19:19:41.100 [INFO][4864] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Feb 13 19:19:41.231295 containerd[1515]: 2025-02-13 19:19:41.100 [INFO][4864] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.131/26] IPv6=[] ContainerID="b9a6fd62cd77a87bbc457ad2d1331edef05b0ed9b12abaf49f0386fbdf40fa17" HandleID="k8s-pod-network.b9a6fd62cd77a87bbc457ad2d1331edef05b0ed9b12abaf49f0386fbdf40fa17" Workload="localhost-k8s-calico--apiserver--86475b859c--kwlfl-eth0" Feb 13 19:19:41.232375 containerd[1515]: 2025-02-13 19:19:41.104 [INFO][4741] cni-plugin/k8s.go 386: Populated endpoint ContainerID="b9a6fd62cd77a87bbc457ad2d1331edef05b0ed9b12abaf49f0386fbdf40fa17" Namespace="calico-apiserver" Pod="calico-apiserver-86475b859c-kwlfl" WorkloadEndpoint="localhost-k8s-calico--apiserver--86475b859c--kwlfl-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--86475b859c--kwlfl-eth0", GenerateName:"calico-apiserver-86475b859c-", Namespace:"calico-apiserver", SelfLink:"", UID:"f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1", ResourceVersion:"693", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 19, 19, 22, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"86475b859c", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-86475b859c-kwlfl", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali562458c7813", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 19:19:41.232375 containerd[1515]: 2025-02-13 19:19:41.104 [INFO][4741] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.131/32] ContainerID="b9a6fd62cd77a87bbc457ad2d1331edef05b0ed9b12abaf49f0386fbdf40fa17" Namespace="calico-apiserver" Pod="calico-apiserver-86475b859c-kwlfl" WorkloadEndpoint="localhost-k8s-calico--apiserver--86475b859c--kwlfl-eth0" Feb 13 19:19:41.232375 containerd[1515]: 2025-02-13 19:19:41.104 [INFO][4741] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali562458c7813 ContainerID="b9a6fd62cd77a87bbc457ad2d1331edef05b0ed9b12abaf49f0386fbdf40fa17" Namespace="calico-apiserver" Pod="calico-apiserver-86475b859c-kwlfl" WorkloadEndpoint="localhost-k8s-calico--apiserver--86475b859c--kwlfl-eth0" Feb 13 19:19:41.232375 containerd[1515]: 2025-02-13 19:19:41.108 [INFO][4741] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="b9a6fd62cd77a87bbc457ad2d1331edef05b0ed9b12abaf49f0386fbdf40fa17" Namespace="calico-apiserver" Pod="calico-apiserver-86475b859c-kwlfl" WorkloadEndpoint="localhost-k8s-calico--apiserver--86475b859c--kwlfl-eth0" Feb 13 19:19:41.232375 containerd[1515]: 2025-02-13 19:19:41.109 [INFO][4741] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="b9a6fd62cd77a87bbc457ad2d1331edef05b0ed9b12abaf49f0386fbdf40fa17" Namespace="calico-apiserver" Pod="calico-apiserver-86475b859c-kwlfl" WorkloadEndpoint="localhost-k8s-calico--apiserver--86475b859c--kwlfl-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--86475b859c--kwlfl-eth0", GenerateName:"calico-apiserver-86475b859c-", Namespace:"calico-apiserver", SelfLink:"", UID:"f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1", ResourceVersion:"693", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 19, 19, 22, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"86475b859c", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"b9a6fd62cd77a87bbc457ad2d1331edef05b0ed9b12abaf49f0386fbdf40fa17", Pod:"calico-apiserver-86475b859c-kwlfl", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali562458c7813", MAC:"ce:da:f2:6a:ba:0f", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 19:19:41.232375 containerd[1515]: 2025-02-13 19:19:41.228 [INFO][4741] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="b9a6fd62cd77a87bbc457ad2d1331edef05b0ed9b12abaf49f0386fbdf40fa17" Namespace="calico-apiserver" Pod="calico-apiserver-86475b859c-kwlfl" WorkloadEndpoint="localhost-k8s-calico--apiserver--86475b859c--kwlfl-eth0" Feb 13 19:19:41.256993 containerd[1515]: time="2025-02-13T19:19:41.256827143Z" level=info msg="CreateContainer within sandbox \"143d63705a6cccab6c533569f1bb8f3f4c9da1f189033dff08d90b2c776e7d3d\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"94f829fb3f33d0d126bcd9d3209740151b4b3be6c68b9155711539013958be04\"" Feb 13 19:19:41.257658 containerd[1515]: time="2025-02-13T19:19:41.257606017Z" level=info msg="StartContainer for \"94f829fb3f33d0d126bcd9d3209740151b4b3be6c68b9155711539013958be04\"" Feb 13 19:19:41.271388 containerd[1515]: time="2025-02-13T19:19:41.263034325Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:19:41.271388 containerd[1515]: time="2025-02-13T19:19:41.263094408Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:19:41.271388 containerd[1515]: time="2025-02-13T19:19:41.263108524Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:19:41.271388 containerd[1515]: time="2025-02-13T19:19:41.263491214Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:19:41.272698 systemd-networkd[1422]: cali0e9e4da435f: Link UP Feb 13 19:19:41.273542 systemd-networkd[1422]: cali0e9e4da435f: Gained carrier Feb 13 19:19:41.277792 kubelet[2630]: I0213 19:19:41.277477 2630 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 13 19:19:41.277792 kubelet[2630]: E0213 19:19:41.277741 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:41.281524 containerd[1515]: time="2025-02-13T19:19:41.280764067Z" level=info msg="StopPodSandbox for \"535ac305f91d459b29fcac754f456e877bb89217b5f89ed4a0f736398ab4e183\"" Feb 13 19:19:41.281524 containerd[1515]: time="2025-02-13T19:19:41.281089899Z" level=info msg="TearDown network for sandbox \"535ac305f91d459b29fcac754f456e877bb89217b5f89ed4a0f736398ab4e183\" successfully" Feb 13 19:19:41.281524 containerd[1515]: time="2025-02-13T19:19:41.281218721Z" level=info msg="StopPodSandbox for \"535ac305f91d459b29fcac754f456e877bb89217b5f89ed4a0f736398ab4e183\" returns successfully" Feb 13 19:19:41.283605 containerd[1515]: time="2025-02-13T19:19:41.283381007Z" level=info msg="StopPodSandbox for \"41f8a27c833a329de6525bb7ea4ae60918c875126c9b0eb7529120508e6289e1\"" Feb 13 19:19:41.283605 containerd[1515]: time="2025-02-13T19:19:41.283474133Z" level=info msg="TearDown network for sandbox \"41f8a27c833a329de6525bb7ea4ae60918c875126c9b0eb7529120508e6289e1\" successfully" Feb 13 19:19:41.283605 containerd[1515]: time="2025-02-13T19:19:41.283501825Z" level=info msg="StopPodSandbox for \"41f8a27c833a329de6525bb7ea4ae60918c875126c9b0eb7529120508e6289e1\" returns successfully" Feb 13 19:19:41.284048 containerd[1515]: time="2025-02-13T19:19:41.283803522Z" level=info msg="StopPodSandbox for \"b9281cb17530bfce23ceee39b94a1f286e88137aa4b1dc73b5faa8a4e7379e99\"" Feb 13 19:19:41.284048 containerd[1515]: time="2025-02-13T19:19:41.283902959Z" level=info msg="TearDown network for sandbox \"b9281cb17530bfce23ceee39b94a1f286e88137aa4b1dc73b5faa8a4e7379e99\" successfully" Feb 13 19:19:41.284048 containerd[1515]: time="2025-02-13T19:19:41.283913319Z" level=info msg="StopPodSandbox for \"b9281cb17530bfce23ceee39b94a1f286e88137aa4b1dc73b5faa8a4e7379e99\" returns successfully" Feb 13 19:19:41.284140 containerd[1515]: time="2025-02-13T19:19:41.284094850Z" level=info msg="StopPodSandbox for \"a072e0a626a78003d707d9120f0d4982782f1be11bc8ddc22d2f8a1dfb030680\"" Feb 13 19:19:41.284258 containerd[1515]: time="2025-02-13T19:19:41.284173518Z" level=info msg="StopPodSandbox for \"ee0d44b3886e78bb5043872859e92bfd2620af442c433a402780bc0453826443\"" Feb 13 19:19:41.284288 containerd[1515]: time="2025-02-13T19:19:41.284277704Z" level=info msg="TearDown network for sandbox \"ee0d44b3886e78bb5043872859e92bfd2620af442c433a402780bc0453826443\" successfully" Feb 13 19:19:41.284320 containerd[1515]: time="2025-02-13T19:19:41.284310355Z" level=info msg="StopPodSandbox for \"ee0d44b3886e78bb5043872859e92bfd2620af442c433a402780bc0453826443\" returns successfully" Feb 13 19:19:41.284513 containerd[1515]: time="2025-02-13T19:19:41.284403049Z" level=info msg="TearDown network for sandbox \"a072e0a626a78003d707d9120f0d4982782f1be11bc8ddc22d2f8a1dfb030680\" successfully" Feb 13 19:19:41.284513 containerd[1515]: time="2025-02-13T19:19:41.284427115Z" level=info msg="StopPodSandbox for \"a072e0a626a78003d707d9120f0d4982782f1be11bc8ddc22d2f8a1dfb030680\" returns successfully" Feb 13 19:19:41.285084 containerd[1515]: time="2025-02-13T19:19:41.285052180Z" level=info msg="StopPodSandbox for \"28a752266e9ffa0b0f35649f5eba403b238d216dc8ce20a8fe3622cc28799217\"" Feb 13 19:19:41.285790 containerd[1515]: time="2025-02-13T19:19:41.285469776Z" level=info msg="TearDown network for sandbox \"28a752266e9ffa0b0f35649f5eba403b238d216dc8ce20a8fe3622cc28799217\" successfully" Feb 13 19:19:41.285790 containerd[1515]: time="2025-02-13T19:19:41.285486968Z" level=info msg="StopPodSandbox for \"28a752266e9ffa0b0f35649f5eba403b238d216dc8ce20a8fe3622cc28799217\" returns successfully" Feb 13 19:19:41.286707 systemd[1]: Started cri-containerd-b9a6fd62cd77a87bbc457ad2d1331edef05b0ed9b12abaf49f0386fbdf40fa17.scope - libcontainer container b9a6fd62cd77a87bbc457ad2d1331edef05b0ed9b12abaf49f0386fbdf40fa17. Feb 13 19:19:41.287660 containerd[1515]: time="2025-02-13T19:19:41.287636089Z" level=info msg="StopPodSandbox for \"a20bc39d31c8f248a56fc8b89d2d3ce62836f574c54b75502431e23bb908b697\"" Feb 13 19:19:41.288963 containerd[1515]: time="2025-02-13T19:19:41.288943919Z" level=info msg="TearDown network for sandbox \"a20bc39d31c8f248a56fc8b89d2d3ce62836f574c54b75502431e23bb908b697\" successfully" Feb 13 19:19:41.289070 containerd[1515]: time="2025-02-13T19:19:41.289035310Z" level=info msg="StopPodSandbox for \"a20bc39d31c8f248a56fc8b89d2d3ce62836f574c54b75502431e23bb908b697\" returns successfully" Feb 13 19:19:41.289426 containerd[1515]: time="2025-02-13T19:19:41.289309897Z" level=info msg="StopPodSandbox for \"216f8481f6c55f7a146dde6ed21a0f30b75cfe7059570683b2fb520eadf70207\"" Feb 13 19:19:41.289426 containerd[1515]: time="2025-02-13T19:19:41.289383856Z" level=info msg="TearDown network for sandbox \"216f8481f6c55f7a146dde6ed21a0f30b75cfe7059570683b2fb520eadf70207\" successfully" Feb 13 19:19:41.289426 containerd[1515]: time="2025-02-13T19:19:41.289392813Z" level=info msg="StopPodSandbox for \"216f8481f6c55f7a146dde6ed21a0f30b75cfe7059570683b2fb520eadf70207\" returns successfully" Feb 13 19:19:41.290116 containerd[1515]: time="2025-02-13T19:19:41.289851666Z" level=info msg="StopPodSandbox for \"1d811fd694e600e47efc43338f2c97ee7c9f512c6106b43a1a049dbc81724231\"" Feb 13 19:19:41.290116 containerd[1515]: time="2025-02-13T19:19:41.289922238Z" level=info msg="TearDown network for sandbox \"1d811fd694e600e47efc43338f2c97ee7c9f512c6106b43a1a049dbc81724231\" successfully" Feb 13 19:19:41.290116 containerd[1515]: time="2025-02-13T19:19:41.289930093Z" level=info msg="StopPodSandbox for \"1d811fd694e600e47efc43338f2c97ee7c9f512c6106b43a1a049dbc81724231\" returns successfully" Feb 13 19:19:41.290116 containerd[1515]: time="2025-02-13T19:19:41.289966682Z" level=info msg="StopPodSandbox for \"f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7\"" Feb 13 19:19:41.290116 containerd[1515]: time="2025-02-13T19:19:41.290022887Z" level=info msg="TearDown network for sandbox \"f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7\" successfully" Feb 13 19:19:41.290116 containerd[1515]: time="2025-02-13T19:19:41.290030311Z" level=info msg="StopPodSandbox for \"f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7\" returns successfully" Feb 13 19:19:41.290429 kubelet[2630]: E0213 19:19:41.290411 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:41.291885 containerd[1515]: time="2025-02-13T19:19:41.291740878Z" level=info msg="StopPodSandbox for \"fa92388357966f3e183c59506d6cfde3e38d36deede0f30821c07e22d4146a49\"" Feb 13 19:19:41.292305 containerd[1515]: time="2025-02-13T19:19:41.292077291Z" level=info msg="TearDown network for sandbox \"fa92388357966f3e183c59506d6cfde3e38d36deede0f30821c07e22d4146a49\" successfully" Feb 13 19:19:41.292384 containerd[1515]: time="2025-02-13T19:19:41.292369009Z" level=info msg="StopPodSandbox for \"fa92388357966f3e183c59506d6cfde3e38d36deede0f30821c07e22d4146a49\" returns successfully" Feb 13 19:19:41.293087 containerd[1515]: 2025-02-13 19:19:40.352 [INFO][4775] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Feb 13 19:19:41.293087 containerd[1515]: 2025-02-13 19:19:40.378 [INFO][4775] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--86475b859c--swqrn-eth0 calico-apiserver-86475b859c- calico-apiserver 5a39efb7-aca1-4893-a36f-840c3c504ece 692 0 2025-02-13 19:19:22 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:86475b859c projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-86475b859c-swqrn eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali0e9e4da435f [] []}} ContainerID="f51a92f3451c546adf1a5f0d8406c487942da1a756304d943b422e13980bcaba" Namespace="calico-apiserver" Pod="calico-apiserver-86475b859c-swqrn" WorkloadEndpoint="localhost-k8s-calico--apiserver--86475b859c--swqrn-" Feb 13 19:19:41.293087 containerd[1515]: 2025-02-13 19:19:40.378 [INFO][4775] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="f51a92f3451c546adf1a5f0d8406c487942da1a756304d943b422e13980bcaba" Namespace="calico-apiserver" Pod="calico-apiserver-86475b859c-swqrn" WorkloadEndpoint="localhost-k8s-calico--apiserver--86475b859c--swqrn-eth0" Feb 13 19:19:41.293087 containerd[1515]: 2025-02-13 19:19:40.496 [INFO][4854] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="f51a92f3451c546adf1a5f0d8406c487942da1a756304d943b422e13980bcaba" HandleID="k8s-pod-network.f51a92f3451c546adf1a5f0d8406c487942da1a756304d943b422e13980bcaba" Workload="localhost-k8s-calico--apiserver--86475b859c--swqrn-eth0" Feb 13 19:19:41.293087 containerd[1515]: 2025-02-13 19:19:40.808 [INFO][4854] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="f51a92f3451c546adf1a5f0d8406c487942da1a756304d943b422e13980bcaba" HandleID="k8s-pod-network.f51a92f3451c546adf1a5f0d8406c487942da1a756304d943b422e13980bcaba" Workload="localhost-k8s-calico--apiserver--86475b859c--swqrn-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0004dfde0), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-86475b859c-swqrn", "timestamp":"2025-02-13 19:19:40.496419004 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Feb 13 19:19:41.293087 containerd[1515]: 2025-02-13 19:19:40.808 [INFO][4854] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Feb 13 19:19:41.293087 containerd[1515]: 2025-02-13 19:19:41.101 [INFO][4854] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Feb 13 19:19:41.293087 containerd[1515]: 2025-02-13 19:19:41.101 [INFO][4854] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Feb 13 19:19:41.293087 containerd[1515]: 2025-02-13 19:19:41.227 [INFO][4854] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.f51a92f3451c546adf1a5f0d8406c487942da1a756304d943b422e13980bcaba" host="localhost" Feb 13 19:19:41.293087 containerd[1515]: 2025-02-13 19:19:41.233 [INFO][4854] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" Feb 13 19:19:41.293087 containerd[1515]: 2025-02-13 19:19:41.239 [INFO][4854] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Feb 13 19:19:41.293087 containerd[1515]: 2025-02-13 19:19:41.241 [INFO][4854] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Feb 13 19:19:41.293087 containerd[1515]: 2025-02-13 19:19:41.243 [INFO][4854] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Feb 13 19:19:41.293087 containerd[1515]: 2025-02-13 19:19:41.243 [INFO][4854] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.f51a92f3451c546adf1a5f0d8406c487942da1a756304d943b422e13980bcaba" host="localhost" Feb 13 19:19:41.293087 containerd[1515]: 2025-02-13 19:19:41.245 [INFO][4854] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.f51a92f3451c546adf1a5f0d8406c487942da1a756304d943b422e13980bcaba Feb 13 19:19:41.293087 containerd[1515]: 2025-02-13 19:19:41.248 [INFO][4854] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.f51a92f3451c546adf1a5f0d8406c487942da1a756304d943b422e13980bcaba" host="localhost" Feb 13 19:19:41.293087 containerd[1515]: 2025-02-13 19:19:41.256 [INFO][4854] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.132/26] block=192.168.88.128/26 handle="k8s-pod-network.f51a92f3451c546adf1a5f0d8406c487942da1a756304d943b422e13980bcaba" host="localhost" Feb 13 19:19:41.293087 containerd[1515]: 2025-02-13 19:19:41.257 [INFO][4854] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.132/26] handle="k8s-pod-network.f51a92f3451c546adf1a5f0d8406c487942da1a756304d943b422e13980bcaba" host="localhost" Feb 13 19:19:41.293087 containerd[1515]: 2025-02-13 19:19:41.257 [INFO][4854] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Feb 13 19:19:41.293087 containerd[1515]: 2025-02-13 19:19:41.257 [INFO][4854] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.132/26] IPv6=[] ContainerID="f51a92f3451c546adf1a5f0d8406c487942da1a756304d943b422e13980bcaba" HandleID="k8s-pod-network.f51a92f3451c546adf1a5f0d8406c487942da1a756304d943b422e13980bcaba" Workload="localhost-k8s-calico--apiserver--86475b859c--swqrn-eth0" Feb 13 19:19:41.294308 containerd[1515]: 2025-02-13 19:19:41.262 [INFO][4775] cni-plugin/k8s.go 386: Populated endpoint ContainerID="f51a92f3451c546adf1a5f0d8406c487942da1a756304d943b422e13980bcaba" Namespace="calico-apiserver" Pod="calico-apiserver-86475b859c-swqrn" WorkloadEndpoint="localhost-k8s-calico--apiserver--86475b859c--swqrn-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--86475b859c--swqrn-eth0", GenerateName:"calico-apiserver-86475b859c-", Namespace:"calico-apiserver", SelfLink:"", UID:"5a39efb7-aca1-4893-a36f-840c3c504ece", ResourceVersion:"692", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 19, 19, 22, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"86475b859c", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-86475b859c-swqrn", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali0e9e4da435f", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 19:19:41.294308 containerd[1515]: 2025-02-13 19:19:41.262 [INFO][4775] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.132/32] ContainerID="f51a92f3451c546adf1a5f0d8406c487942da1a756304d943b422e13980bcaba" Namespace="calico-apiserver" Pod="calico-apiserver-86475b859c-swqrn" WorkloadEndpoint="localhost-k8s-calico--apiserver--86475b859c--swqrn-eth0" Feb 13 19:19:41.294308 containerd[1515]: 2025-02-13 19:19:41.262 [INFO][4775] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali0e9e4da435f ContainerID="f51a92f3451c546adf1a5f0d8406c487942da1a756304d943b422e13980bcaba" Namespace="calico-apiserver" Pod="calico-apiserver-86475b859c-swqrn" WorkloadEndpoint="localhost-k8s-calico--apiserver--86475b859c--swqrn-eth0" Feb 13 19:19:41.294308 containerd[1515]: 2025-02-13 19:19:41.273 [INFO][4775] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="f51a92f3451c546adf1a5f0d8406c487942da1a756304d943b422e13980bcaba" Namespace="calico-apiserver" Pod="calico-apiserver-86475b859c-swqrn" WorkloadEndpoint="localhost-k8s-calico--apiserver--86475b859c--swqrn-eth0" Feb 13 19:19:41.294308 containerd[1515]: 2025-02-13 19:19:41.274 [INFO][4775] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="f51a92f3451c546adf1a5f0d8406c487942da1a756304d943b422e13980bcaba" Namespace="calico-apiserver" Pod="calico-apiserver-86475b859c-swqrn" WorkloadEndpoint="localhost-k8s-calico--apiserver--86475b859c--swqrn-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--86475b859c--swqrn-eth0", GenerateName:"calico-apiserver-86475b859c-", Namespace:"calico-apiserver", SelfLink:"", UID:"5a39efb7-aca1-4893-a36f-840c3c504ece", ResourceVersion:"692", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 19, 19, 22, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"86475b859c", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"f51a92f3451c546adf1a5f0d8406c487942da1a756304d943b422e13980bcaba", Pod:"calico-apiserver-86475b859c-swqrn", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali0e9e4da435f", MAC:"16:06:5d:96:8f:05", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 19:19:41.294308 containerd[1515]: 2025-02-13 19:19:41.286 [INFO][4775] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="f51a92f3451c546adf1a5f0d8406c487942da1a756304d943b422e13980bcaba" Namespace="calico-apiserver" Pod="calico-apiserver-86475b859c-swqrn" WorkloadEndpoint="localhost-k8s-calico--apiserver--86475b859c--swqrn-eth0" Feb 13 19:19:41.294308 containerd[1515]: time="2025-02-13T19:19:41.293296223Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-frtk4,Uid:f48c8839-9a17-43c4-a0d5-f9e8d9767814,Namespace:kube-system,Attempt:6,}" Feb 13 19:19:41.295763 containerd[1515]: time="2025-02-13T19:19:41.294974569Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-drlnv,Uid:84096616-cf79-421c-9e94-8cbcf0a204a5,Namespace:calico-system,Attempt:5,}" Feb 13 19:19:41.303376 systemd[1]: Started cri-containerd-94f829fb3f33d0d126bcd9d3209740151b4b3be6c68b9155711539013958be04.scope - libcontainer container 94f829fb3f33d0d126bcd9d3209740151b4b3be6c68b9155711539013958be04. Feb 13 19:19:41.307862 systemd-resolved[1350]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Feb 13 19:19:41.324964 containerd[1515]: time="2025-02-13T19:19:41.324127050Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:19:41.325703 containerd[1515]: time="2025-02-13T19:19:41.325664792Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:19:41.327536 containerd[1515]: time="2025-02-13T19:19:41.326279389Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:19:41.327536 containerd[1515]: time="2025-02-13T19:19:41.326361493Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:19:41.347428 systemd[1]: Started cri-containerd-f51a92f3451c546adf1a5f0d8406c487942da1a756304d943b422e13980bcaba.scope - libcontainer container f51a92f3451c546adf1a5f0d8406c487942da1a756304d943b422e13980bcaba. Feb 13 19:19:41.357325 containerd[1515]: time="2025-02-13T19:19:41.357277441Z" level=info msg="StartContainer for \"94f829fb3f33d0d126bcd9d3209740151b4b3be6c68b9155711539013958be04\" returns successfully" Feb 13 19:19:41.363781 containerd[1515]: time="2025-02-13T19:19:41.363744572Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86475b859c-kwlfl,Uid:f6e023ee-9ef0-491c-bd3c-e70ebf96e0c1,Namespace:calico-apiserver,Attempt:6,} returns sandbox id \"b9a6fd62cd77a87bbc457ad2d1331edef05b0ed9b12abaf49f0386fbdf40fa17\"" Feb 13 19:19:41.367103 systemd-resolved[1350]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Feb 13 19:19:41.416792 containerd[1515]: time="2025-02-13T19:19:41.416747214Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86475b859c-swqrn,Uid:5a39efb7-aca1-4893-a36f-840c3c504ece,Namespace:calico-apiserver,Attempt:6,} returns sandbox id \"f51a92f3451c546adf1a5f0d8406c487942da1a756304d943b422e13980bcaba\"" Feb 13 19:19:41.535792 systemd-networkd[1422]: cali35413c5757a: Link UP Feb 13 19:19:41.537316 systemd-networkd[1422]: cali35413c5757a: Gained carrier Feb 13 19:19:41.548035 containerd[1515]: 2025-02-13 19:19:41.348 [INFO][5117] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Feb 13 19:19:41.548035 containerd[1515]: 2025-02-13 19:19:41.362 [INFO][5117] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-csi--node--driver--drlnv-eth0 csi-node-driver- calico-system 84096616-cf79-421c-9e94-8cbcf0a204a5 802 0 2025-02-13 19:19:22 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:56747c9949 k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s localhost csi-node-driver-drlnv eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali35413c5757a [] []}} ContainerID="e22b101a1c7150c18eeb3428d7c5a80aba462dd305769816a18673cbfcf0ec44" Namespace="calico-system" Pod="csi-node-driver-drlnv" WorkloadEndpoint="localhost-k8s-csi--node--driver--drlnv-" Feb 13 19:19:41.548035 containerd[1515]: 2025-02-13 19:19:41.362 [INFO][5117] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="e22b101a1c7150c18eeb3428d7c5a80aba462dd305769816a18673cbfcf0ec44" Namespace="calico-system" Pod="csi-node-driver-drlnv" WorkloadEndpoint="localhost-k8s-csi--node--driver--drlnv-eth0" Feb 13 19:19:41.548035 containerd[1515]: 2025-02-13 19:19:41.400 [INFO][5175] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="e22b101a1c7150c18eeb3428d7c5a80aba462dd305769816a18673cbfcf0ec44" HandleID="k8s-pod-network.e22b101a1c7150c18eeb3428d7c5a80aba462dd305769816a18673cbfcf0ec44" Workload="localhost-k8s-csi--node--driver--drlnv-eth0" Feb 13 19:19:41.548035 containerd[1515]: 2025-02-13 19:19:41.416 [INFO][5175] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="e22b101a1c7150c18eeb3428d7c5a80aba462dd305769816a18673cbfcf0ec44" HandleID="k8s-pod-network.e22b101a1c7150c18eeb3428d7c5a80aba462dd305769816a18673cbfcf0ec44" Workload="localhost-k8s-csi--node--driver--drlnv-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0004379e0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"csi-node-driver-drlnv", "timestamp":"2025-02-13 19:19:41.400150714 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Feb 13 19:19:41.548035 containerd[1515]: 2025-02-13 19:19:41.416 [INFO][5175] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Feb 13 19:19:41.548035 containerd[1515]: 2025-02-13 19:19:41.416 [INFO][5175] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Feb 13 19:19:41.548035 containerd[1515]: 2025-02-13 19:19:41.416 [INFO][5175] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Feb 13 19:19:41.548035 containerd[1515]: 2025-02-13 19:19:41.418 [INFO][5175] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.e22b101a1c7150c18eeb3428d7c5a80aba462dd305769816a18673cbfcf0ec44" host="localhost" Feb 13 19:19:41.548035 containerd[1515]: 2025-02-13 19:19:41.511 [INFO][5175] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" Feb 13 19:19:41.548035 containerd[1515]: 2025-02-13 19:19:41.517 [INFO][5175] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Feb 13 19:19:41.548035 containerd[1515]: 2025-02-13 19:19:41.518 [INFO][5175] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Feb 13 19:19:41.548035 containerd[1515]: 2025-02-13 19:19:41.521 [INFO][5175] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Feb 13 19:19:41.548035 containerd[1515]: 2025-02-13 19:19:41.521 [INFO][5175] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.e22b101a1c7150c18eeb3428d7c5a80aba462dd305769816a18673cbfcf0ec44" host="localhost" Feb 13 19:19:41.548035 containerd[1515]: 2025-02-13 19:19:41.522 [INFO][5175] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.e22b101a1c7150c18eeb3428d7c5a80aba462dd305769816a18673cbfcf0ec44 Feb 13 19:19:41.548035 containerd[1515]: 2025-02-13 19:19:41.526 [INFO][5175] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.e22b101a1c7150c18eeb3428d7c5a80aba462dd305769816a18673cbfcf0ec44" host="localhost" Feb 13 19:19:41.548035 containerd[1515]: 2025-02-13 19:19:41.530 [INFO][5175] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.133/26] block=192.168.88.128/26 handle="k8s-pod-network.e22b101a1c7150c18eeb3428d7c5a80aba462dd305769816a18673cbfcf0ec44" host="localhost" Feb 13 19:19:41.548035 containerd[1515]: 2025-02-13 19:19:41.530 [INFO][5175] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.133/26] handle="k8s-pod-network.e22b101a1c7150c18eeb3428d7c5a80aba462dd305769816a18673cbfcf0ec44" host="localhost" Feb 13 19:19:41.548035 containerd[1515]: 2025-02-13 19:19:41.530 [INFO][5175] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Feb 13 19:19:41.548035 containerd[1515]: 2025-02-13 19:19:41.530 [INFO][5175] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.133/26] IPv6=[] ContainerID="e22b101a1c7150c18eeb3428d7c5a80aba462dd305769816a18673cbfcf0ec44" HandleID="k8s-pod-network.e22b101a1c7150c18eeb3428d7c5a80aba462dd305769816a18673cbfcf0ec44" Workload="localhost-k8s-csi--node--driver--drlnv-eth0" Feb 13 19:19:41.548692 containerd[1515]: 2025-02-13 19:19:41.533 [INFO][5117] cni-plugin/k8s.go 386: Populated endpoint ContainerID="e22b101a1c7150c18eeb3428d7c5a80aba462dd305769816a18673cbfcf0ec44" Namespace="calico-system" Pod="csi-node-driver-drlnv" WorkloadEndpoint="localhost-k8s-csi--node--driver--drlnv-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--drlnv-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"84096616-cf79-421c-9e94-8cbcf0a204a5", ResourceVersion:"802", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 19, 19, 22, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"56747c9949", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"csi-node-driver-drlnv", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali35413c5757a", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 19:19:41.548692 containerd[1515]: 2025-02-13 19:19:41.533 [INFO][5117] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.133/32] ContainerID="e22b101a1c7150c18eeb3428d7c5a80aba462dd305769816a18673cbfcf0ec44" Namespace="calico-system" Pod="csi-node-driver-drlnv" WorkloadEndpoint="localhost-k8s-csi--node--driver--drlnv-eth0" Feb 13 19:19:41.548692 containerd[1515]: 2025-02-13 19:19:41.533 [INFO][5117] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali35413c5757a ContainerID="e22b101a1c7150c18eeb3428d7c5a80aba462dd305769816a18673cbfcf0ec44" Namespace="calico-system" Pod="csi-node-driver-drlnv" WorkloadEndpoint="localhost-k8s-csi--node--driver--drlnv-eth0" Feb 13 19:19:41.548692 containerd[1515]: 2025-02-13 19:19:41.536 [INFO][5117] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="e22b101a1c7150c18eeb3428d7c5a80aba462dd305769816a18673cbfcf0ec44" Namespace="calico-system" Pod="csi-node-driver-drlnv" WorkloadEndpoint="localhost-k8s-csi--node--driver--drlnv-eth0" Feb 13 19:19:41.548692 containerd[1515]: 2025-02-13 19:19:41.537 [INFO][5117] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="e22b101a1c7150c18eeb3428d7c5a80aba462dd305769816a18673cbfcf0ec44" Namespace="calico-system" Pod="csi-node-driver-drlnv" WorkloadEndpoint="localhost-k8s-csi--node--driver--drlnv-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--drlnv-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"84096616-cf79-421c-9e94-8cbcf0a204a5", ResourceVersion:"802", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 19, 19, 22, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"56747c9949", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"e22b101a1c7150c18eeb3428d7c5a80aba462dd305769816a18673cbfcf0ec44", Pod:"csi-node-driver-drlnv", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali35413c5757a", MAC:"32:74:91:3b:39:db", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 19:19:41.548692 containerd[1515]: 2025-02-13 19:19:41.545 [INFO][5117] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="e22b101a1c7150c18eeb3428d7c5a80aba462dd305769816a18673cbfcf0ec44" Namespace="calico-system" Pod="csi-node-driver-drlnv" WorkloadEndpoint="localhost-k8s-csi--node--driver--drlnv-eth0" Feb 13 19:19:41.568860 containerd[1515]: time="2025-02-13T19:19:41.568730601Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:19:41.568860 containerd[1515]: time="2025-02-13T19:19:41.568806143Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:19:41.568860 containerd[1515]: time="2025-02-13T19:19:41.568818586Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:19:41.569071 containerd[1515]: time="2025-02-13T19:19:41.568914517Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:19:41.591346 systemd[1]: Started cri-containerd-e22b101a1c7150c18eeb3428d7c5a80aba462dd305769816a18673cbfcf0ec44.scope - libcontainer container e22b101a1c7150c18eeb3428d7c5a80aba462dd305769816a18673cbfcf0ec44. Feb 13 19:19:41.602165 systemd-resolved[1350]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Feb 13 19:19:41.613478 containerd[1515]: time="2025-02-13T19:19:41.613428266Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-drlnv,Uid:84096616-cf79-421c-9e94-8cbcf0a204a5,Namespace:calico-system,Attempt:5,} returns sandbox id \"e22b101a1c7150c18eeb3428d7c5a80aba462dd305769816a18673cbfcf0ec44\"" Feb 13 19:19:41.636622 systemd-networkd[1422]: caliaf855a05ee0: Link UP Feb 13 19:19:41.636818 systemd-networkd[1422]: caliaf855a05ee0: Gained carrier Feb 13 19:19:41.647176 containerd[1515]: 2025-02-13 19:19:41.340 [INFO][5101] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Feb 13 19:19:41.647176 containerd[1515]: 2025-02-13 19:19:41.353 [INFO][5101] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--6f6b679f8f--frtk4-eth0 coredns-6f6b679f8f- kube-system f48c8839-9a17-43c4-a0d5-f9e8d9767814 803 0 2025-02-13 19:19:17 +0000 UTC map[k8s-app:kube-dns pod-template-hash:6f6b679f8f projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-6f6b679f8f-frtk4 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] caliaf855a05ee0 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="dab92c3285663c9858cae987e8ebef281609cd0ce1ed367cc6269f31674f335c" Namespace="kube-system" Pod="coredns-6f6b679f8f-frtk4" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--frtk4-" Feb 13 19:19:41.647176 containerd[1515]: 2025-02-13 19:19:41.353 [INFO][5101] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="dab92c3285663c9858cae987e8ebef281609cd0ce1ed367cc6269f31674f335c" Namespace="kube-system" Pod="coredns-6f6b679f8f-frtk4" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--frtk4-eth0" Feb 13 19:19:41.647176 containerd[1515]: 2025-02-13 19:19:41.395 [INFO][5169] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="dab92c3285663c9858cae987e8ebef281609cd0ce1ed367cc6269f31674f335c" HandleID="k8s-pod-network.dab92c3285663c9858cae987e8ebef281609cd0ce1ed367cc6269f31674f335c" Workload="localhost-k8s-coredns--6f6b679f8f--frtk4-eth0" Feb 13 19:19:41.647176 containerd[1515]: 2025-02-13 19:19:41.416 [INFO][5169] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="dab92c3285663c9858cae987e8ebef281609cd0ce1ed367cc6269f31674f335c" HandleID="k8s-pod-network.dab92c3285663c9858cae987e8ebef281609cd0ce1ed367cc6269f31674f335c" Workload="localhost-k8s-coredns--6f6b679f8f--frtk4-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0001a05a0), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-6f6b679f8f-frtk4", "timestamp":"2025-02-13 19:19:41.395058218 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Feb 13 19:19:41.647176 containerd[1515]: 2025-02-13 19:19:41.416 [INFO][5169] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Feb 13 19:19:41.647176 containerd[1515]: 2025-02-13 19:19:41.530 [INFO][5169] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Feb 13 19:19:41.647176 containerd[1515]: 2025-02-13 19:19:41.530 [INFO][5169] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Feb 13 19:19:41.647176 containerd[1515]: 2025-02-13 19:19:41.534 [INFO][5169] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.dab92c3285663c9858cae987e8ebef281609cd0ce1ed367cc6269f31674f335c" host="localhost" Feb 13 19:19:41.647176 containerd[1515]: 2025-02-13 19:19:41.612 [INFO][5169] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" Feb 13 19:19:41.647176 containerd[1515]: 2025-02-13 19:19:41.618 [INFO][5169] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Feb 13 19:19:41.647176 containerd[1515]: 2025-02-13 19:19:41.619 [INFO][5169] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Feb 13 19:19:41.647176 containerd[1515]: 2025-02-13 19:19:41.621 [INFO][5169] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Feb 13 19:19:41.647176 containerd[1515]: 2025-02-13 19:19:41.621 [INFO][5169] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.dab92c3285663c9858cae987e8ebef281609cd0ce1ed367cc6269f31674f335c" host="localhost" Feb 13 19:19:41.647176 containerd[1515]: 2025-02-13 19:19:41.622 [INFO][5169] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.dab92c3285663c9858cae987e8ebef281609cd0ce1ed367cc6269f31674f335c Feb 13 19:19:41.647176 containerd[1515]: 2025-02-13 19:19:41.625 [INFO][5169] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.dab92c3285663c9858cae987e8ebef281609cd0ce1ed367cc6269f31674f335c" host="localhost" Feb 13 19:19:41.647176 containerd[1515]: 2025-02-13 19:19:41.631 [INFO][5169] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.134/26] block=192.168.88.128/26 handle="k8s-pod-network.dab92c3285663c9858cae987e8ebef281609cd0ce1ed367cc6269f31674f335c" host="localhost" Feb 13 19:19:41.647176 containerd[1515]: 2025-02-13 19:19:41.631 [INFO][5169] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.134/26] handle="k8s-pod-network.dab92c3285663c9858cae987e8ebef281609cd0ce1ed367cc6269f31674f335c" host="localhost" Feb 13 19:19:41.647176 containerd[1515]: 2025-02-13 19:19:41.631 [INFO][5169] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Feb 13 19:19:41.647176 containerd[1515]: 2025-02-13 19:19:41.631 [INFO][5169] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.134/26] IPv6=[] ContainerID="dab92c3285663c9858cae987e8ebef281609cd0ce1ed367cc6269f31674f335c" HandleID="k8s-pod-network.dab92c3285663c9858cae987e8ebef281609cd0ce1ed367cc6269f31674f335c" Workload="localhost-k8s-coredns--6f6b679f8f--frtk4-eth0" Feb 13 19:19:41.647787 containerd[1515]: 2025-02-13 19:19:41.634 [INFO][5101] cni-plugin/k8s.go 386: Populated endpoint ContainerID="dab92c3285663c9858cae987e8ebef281609cd0ce1ed367cc6269f31674f335c" Namespace="kube-system" Pod="coredns-6f6b679f8f-frtk4" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--frtk4-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--6f6b679f8f--frtk4-eth0", GenerateName:"coredns-6f6b679f8f-", Namespace:"kube-system", SelfLink:"", UID:"f48c8839-9a17-43c4-a0d5-f9e8d9767814", ResourceVersion:"803", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 19, 19, 17, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"6f6b679f8f", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-6f6b679f8f-frtk4", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"caliaf855a05ee0", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 19:19:41.647787 containerd[1515]: 2025-02-13 19:19:41.634 [INFO][5101] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.134/32] ContainerID="dab92c3285663c9858cae987e8ebef281609cd0ce1ed367cc6269f31674f335c" Namespace="kube-system" Pod="coredns-6f6b679f8f-frtk4" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--frtk4-eth0" Feb 13 19:19:41.647787 containerd[1515]: 2025-02-13 19:19:41.634 [INFO][5101] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to caliaf855a05ee0 ContainerID="dab92c3285663c9858cae987e8ebef281609cd0ce1ed367cc6269f31674f335c" Namespace="kube-system" Pod="coredns-6f6b679f8f-frtk4" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--frtk4-eth0" Feb 13 19:19:41.647787 containerd[1515]: 2025-02-13 19:19:41.636 [INFO][5101] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="dab92c3285663c9858cae987e8ebef281609cd0ce1ed367cc6269f31674f335c" Namespace="kube-system" Pod="coredns-6f6b679f8f-frtk4" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--frtk4-eth0" Feb 13 19:19:41.647787 containerd[1515]: 2025-02-13 19:19:41.637 [INFO][5101] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="dab92c3285663c9858cae987e8ebef281609cd0ce1ed367cc6269f31674f335c" Namespace="kube-system" Pod="coredns-6f6b679f8f-frtk4" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--frtk4-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--6f6b679f8f--frtk4-eth0", GenerateName:"coredns-6f6b679f8f-", Namespace:"kube-system", SelfLink:"", UID:"f48c8839-9a17-43c4-a0d5-f9e8d9767814", ResourceVersion:"803", Generation:0, CreationTimestamp:time.Date(2025, time.February, 13, 19, 19, 17, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"6f6b679f8f", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"dab92c3285663c9858cae987e8ebef281609cd0ce1ed367cc6269f31674f335c", Pod:"coredns-6f6b679f8f-frtk4", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"caliaf855a05ee0", MAC:"3a:0b:0c:28:ad:f0", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Feb 13 19:19:41.647787 containerd[1515]: 2025-02-13 19:19:41.644 [INFO][5101] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="dab92c3285663c9858cae987e8ebef281609cd0ce1ed367cc6269f31674f335c" Namespace="kube-system" Pod="coredns-6f6b679f8f-frtk4" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--frtk4-eth0" Feb 13 19:19:41.669453 containerd[1515]: time="2025-02-13T19:19:41.669353723Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 13 19:19:41.669453 containerd[1515]: time="2025-02-13T19:19:41.669405451Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 13 19:19:41.669453 containerd[1515]: time="2025-02-13T19:19:41.669416282Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:19:41.669613 containerd[1515]: time="2025-02-13T19:19:41.669521008Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 13 19:19:41.694350 systemd[1]: Started cri-containerd-dab92c3285663c9858cae987e8ebef281609cd0ce1ed367cc6269f31674f335c.scope - libcontainer container dab92c3285663c9858cae987e8ebef281609cd0ce1ed367cc6269f31674f335c. Feb 13 19:19:41.708923 systemd-resolved[1350]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Feb 13 19:19:41.732680 containerd[1515]: time="2025-02-13T19:19:41.732628593Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-frtk4,Uid:f48c8839-9a17-43c4-a0d5-f9e8d9767814,Namespace:kube-system,Attempt:6,} returns sandbox id \"dab92c3285663c9858cae987e8ebef281609cd0ce1ed367cc6269f31674f335c\"" Feb 13 19:19:41.733484 kubelet[2630]: E0213 19:19:41.733453 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:41.736003 containerd[1515]: time="2025-02-13T19:19:41.735927286Z" level=info msg="CreateContainer within sandbox \"dab92c3285663c9858cae987e8ebef281609cd0ce1ed367cc6269f31674f335c\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Feb 13 19:19:41.754604 containerd[1515]: time="2025-02-13T19:19:41.754545718Z" level=info msg="CreateContainer within sandbox \"dab92c3285663c9858cae987e8ebef281609cd0ce1ed367cc6269f31674f335c\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"7354310dc1f9c1b0014cee6fc9f33c308bacd1d8fc4e3bb99efd52655d753add\"" Feb 13 19:19:41.755089 containerd[1515]: time="2025-02-13T19:19:41.755056018Z" level=info msg="StartContainer for \"7354310dc1f9c1b0014cee6fc9f33c308bacd1d8fc4e3bb99efd52655d753add\"" Feb 13 19:19:41.785325 systemd[1]: Started cri-containerd-7354310dc1f9c1b0014cee6fc9f33c308bacd1d8fc4e3bb99efd52655d753add.scope - libcontainer container 7354310dc1f9c1b0014cee6fc9f33c308bacd1d8fc4e3bb99efd52655d753add. Feb 13 19:19:41.812913 containerd[1515]: time="2025-02-13T19:19:41.812770991Z" level=info msg="StartContainer for \"7354310dc1f9c1b0014cee6fc9f33c308bacd1d8fc4e3bb99efd52655d753add\" returns successfully" Feb 13 19:19:42.281319 kubelet[2630]: E0213 19:19:42.281283 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:42.286474 kubelet[2630]: E0213 19:19:42.286437 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:42.291725 kubelet[2630]: I0213 19:19:42.290822 2630 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-6f6b679f8f-s8264" podStartSLOduration=25.290804915 podStartE2EDuration="25.290804915s" podCreationTimestamp="2025-02-13 19:19:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 19:19:42.290228923 +0000 UTC m=+30.712082013" watchObservedRunningTime="2025-02-13 19:19:42.290804915 +0000 UTC m=+30.712658005" Feb 13 19:19:42.365406 systemd-networkd[1422]: cali8b6a1ad33a3: Gained IPv6LL Feb 13 19:19:42.749321 systemd-networkd[1422]: calied5a4c5c7f2: Gained IPv6LL Feb 13 19:19:42.941348 systemd-networkd[1422]: cali562458c7813: Gained IPv6LL Feb 13 19:19:43.135826 systemd-networkd[1422]: cali0e9e4da435f: Gained IPv6LL Feb 13 19:19:43.261319 systemd-networkd[1422]: caliaf855a05ee0: Gained IPv6LL Feb 13 19:19:43.298215 kubelet[2630]: E0213 19:19:43.298165 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:43.298711 kubelet[2630]: E0213 19:19:43.298225 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:43.517397 systemd-networkd[1422]: cali35413c5757a: Gained IPv6LL Feb 13 19:19:44.304090 kubelet[2630]: E0213 19:19:44.303683 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:44.305033 kubelet[2630]: E0213 19:19:44.304886 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:44.325679 kubelet[2630]: I0213 19:19:44.325631 2630 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 13 19:19:44.326345 kubelet[2630]: E0213 19:19:44.326081 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:44.492113 systemd[1]: run-containerd-runc-k8s.io-0afce0c7b685ae300a6ba2b2717dbc064399ca1dd4b6216e93698c5fc55790a6-runc.i2wh8l.mount: Deactivated successfully. Feb 13 19:19:44.806142 containerd[1515]: time="2025-02-13T19:19:44.806068544Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:19:44.841083 containerd[1515]: time="2025-02-13T19:19:44.841010851Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.29.1: active requests=0, bytes read=34141192" Feb 13 19:19:44.881953 containerd[1515]: time="2025-02-13T19:19:44.881877334Z" level=info msg="ImageCreate event name:\"sha256:6331715a2ae96b18a770a395cac108321d108e445e08b616e5bc9fbd1f9c21da\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:19:44.921903 containerd[1515]: time="2025-02-13T19:19:44.921789854Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:1072d6a98167a14ca361e9ce757733f9bae36d1f1c6a9621ea10934b6b1e10d9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:19:44.922350 containerd[1515]: time="2025-02-13T19:19:44.922319820Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.29.1\" with image id \"sha256:6331715a2ae96b18a770a395cac108321d108e445e08b616e5bc9fbd1f9c21da\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:1072d6a98167a14ca361e9ce757733f9bae36d1f1c6a9621ea10934b6b1e10d9\", size \"35634244\" in 3.81223107s" Feb 13 19:19:44.922416 containerd[1515]: time="2025-02-13T19:19:44.922355336Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.29.1\" returns image reference \"sha256:6331715a2ae96b18a770a395cac108321d108e445e08b616e5bc9fbd1f9c21da\"" Feb 13 19:19:44.923401 containerd[1515]: time="2025-02-13T19:19:44.923352590Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.1\"" Feb 13 19:19:44.936542 containerd[1515]: time="2025-02-13T19:19:44.936500714Z" level=info msg="CreateContainer within sandbox \"46203ac9ba0fbbbb33ec052a083f76a39f1007ec3dc0a1ebe440dc0ca7551d85\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Feb 13 19:19:45.210985 containerd[1515]: time="2025-02-13T19:19:45.210921941Z" level=info msg="CreateContainer within sandbox \"46203ac9ba0fbbbb33ec052a083f76a39f1007ec3dc0a1ebe440dc0ca7551d85\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"e804e0d1aa238fd8e2019036ce03d112b27f4d3d39984d39232a389ae6244fec\"" Feb 13 19:19:45.211536 containerd[1515]: time="2025-02-13T19:19:45.211493265Z" level=info msg="StartContainer for \"e804e0d1aa238fd8e2019036ce03d112b27f4d3d39984d39232a389ae6244fec\"" Feb 13 19:19:45.245504 systemd[1]: Started cri-containerd-e804e0d1aa238fd8e2019036ce03d112b27f4d3d39984d39232a389ae6244fec.scope - libcontainer container e804e0d1aa238fd8e2019036ce03d112b27f4d3d39984d39232a389ae6244fec. Feb 13 19:19:45.248681 systemd[1]: Started sshd@9-10.0.0.46:22-10.0.0.1:48804.service - OpenSSH per-connection server daemon (10.0.0.1:48804). Feb 13 19:19:45.301564 containerd[1515]: time="2025-02-13T19:19:45.301518332Z" level=info msg="StartContainer for \"e804e0d1aa238fd8e2019036ce03d112b27f4d3d39984d39232a389ae6244fec\" returns successfully" Feb 13 19:19:45.319529 sshd[5565]: Accepted publickey for core from 10.0.0.1 port 48804 ssh2: RSA SHA256:xgLbxCKtIvCmXzj7C6d4ih050Hrbkh61XCRduaX62E8 Feb 13 19:19:45.323427 kubelet[2630]: I0213 19:19:45.323356 2630 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-6f6b679f8f-frtk4" podStartSLOduration=28.323313914 podStartE2EDuration="28.323313914s" podCreationTimestamp="2025-02-13 19:19:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-02-13 19:19:42.322119242 +0000 UTC m=+30.743972333" watchObservedRunningTime="2025-02-13 19:19:45.323313914 +0000 UTC m=+33.745167004" Feb 13 19:19:45.324280 sshd-session[5565]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:19:45.331801 systemd-logind[1495]: New session 10 of user core. Feb 13 19:19:45.337398 systemd[1]: Started session-10.scope - Session 10 of User core. Feb 13 19:19:45.387199 kubelet[2630]: I0213 19:19:45.387118 2630 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-664f6f5454-mnbf7" podStartSLOduration=19.573157842 podStartE2EDuration="23.387097054s" podCreationTimestamp="2025-02-13 19:19:22 +0000 UTC" firstStartedPulling="2025-02-13 19:19:41.109263728 +0000 UTC m=+29.531116808" lastFinishedPulling="2025-02-13 19:19:44.92320293 +0000 UTC m=+33.345056020" observedRunningTime="2025-02-13 19:19:45.323792814 +0000 UTC m=+33.745645904" watchObservedRunningTime="2025-02-13 19:19:45.387097054 +0000 UTC m=+33.808950154" Feb 13 19:19:45.506839 sshd[5603]: Connection closed by 10.0.0.1 port 48804 Feb 13 19:19:45.507118 sshd-session[5565]: pam_unix(sshd:session): session closed for user core Feb 13 19:19:45.510863 systemd[1]: sshd@9-10.0.0.46:22-10.0.0.1:48804.service: Deactivated successfully. Feb 13 19:19:45.513214 systemd[1]: session-10.scope: Deactivated successfully. Feb 13 19:19:45.513917 systemd-logind[1495]: Session 10 logged out. Waiting for processes to exit. Feb 13 19:19:45.514730 systemd-logind[1495]: Removed session 10. Feb 13 19:19:47.760592 kubelet[2630]: I0213 19:19:47.760531 2630 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 13 19:19:47.761318 kubelet[2630]: E0213 19:19:47.760908 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:48.058243 kernel: bpftool[5720]: memfd_create() called without MFD_EXEC or MFD_NOEXEC_SEAL set Feb 13 19:19:48.074143 containerd[1515]: time="2025-02-13T19:19:48.074098486Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:19:48.074950 containerd[1515]: time="2025-02-13T19:19:48.074928244Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.29.1: active requests=0, bytes read=42001404" Feb 13 19:19:48.076244 containerd[1515]: time="2025-02-13T19:19:48.076218318Z" level=info msg="ImageCreate event name:\"sha256:421726ace5ed13894f7edf594dd3a462947aedc13d0f69d08525d7369477fb70\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:19:48.078816 containerd[1515]: time="2025-02-13T19:19:48.078792333Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:b8c43e264fe52e0c327b0bf3ac882a0224b33bdd7f4ff58a74242da7d9b00486\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:19:48.079625 containerd[1515]: time="2025-02-13T19:19:48.079592156Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.29.1\" with image id \"sha256:421726ace5ed13894f7edf594dd3a462947aedc13d0f69d08525d7369477fb70\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:b8c43e264fe52e0c327b0bf3ac882a0224b33bdd7f4ff58a74242da7d9b00486\", size \"43494504\" in 3.156201804s" Feb 13 19:19:48.079625 containerd[1515]: time="2025-02-13T19:19:48.079625679Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.1\" returns image reference \"sha256:421726ace5ed13894f7edf594dd3a462947aedc13d0f69d08525d7369477fb70\"" Feb 13 19:19:48.080532 containerd[1515]: time="2025-02-13T19:19:48.080505863Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.1\"" Feb 13 19:19:48.081579 containerd[1515]: time="2025-02-13T19:19:48.081553059Z" level=info msg="CreateContainer within sandbox \"b9a6fd62cd77a87bbc457ad2d1331edef05b0ed9b12abaf49f0386fbdf40fa17\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Feb 13 19:19:48.097646 containerd[1515]: time="2025-02-13T19:19:48.097599766Z" level=info msg="CreateContainer within sandbox \"b9a6fd62cd77a87bbc457ad2d1331edef05b0ed9b12abaf49f0386fbdf40fa17\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"a0fa3e745d3c05f69338c63b283f88509b117c51cf06a9522418fe72bc8b7328\"" Feb 13 19:19:48.098138 containerd[1515]: time="2025-02-13T19:19:48.098048228Z" level=info msg="StartContainer for \"a0fa3e745d3c05f69338c63b283f88509b117c51cf06a9522418fe72bc8b7328\"" Feb 13 19:19:48.132326 systemd[1]: Started cri-containerd-a0fa3e745d3c05f69338c63b283f88509b117c51cf06a9522418fe72bc8b7328.scope - libcontainer container a0fa3e745d3c05f69338c63b283f88509b117c51cf06a9522418fe72bc8b7328. Feb 13 19:19:48.174494 containerd[1515]: time="2025-02-13T19:19:48.174439850Z" level=info msg="StartContainer for \"a0fa3e745d3c05f69338c63b283f88509b117c51cf06a9522418fe72bc8b7328\" returns successfully" Feb 13 19:19:48.320135 kubelet[2630]: E0213 19:19:48.319208 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:19:48.331397 systemd-networkd[1422]: vxlan.calico: Link UP Feb 13 19:19:48.331405 systemd-networkd[1422]: vxlan.calico: Gained carrier Feb 13 19:19:48.743359 containerd[1515]: time="2025-02-13T19:19:48.743301325Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:19:48.744354 containerd[1515]: time="2025-02-13T19:19:48.744253543Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.29.1: active requests=0, bytes read=77" Feb 13 19:19:48.746502 containerd[1515]: time="2025-02-13T19:19:48.746469427Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.29.1\" with image id \"sha256:421726ace5ed13894f7edf594dd3a462947aedc13d0f69d08525d7369477fb70\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:b8c43e264fe52e0c327b0bf3ac882a0224b33bdd7f4ff58a74242da7d9b00486\", size \"43494504\" in 665.932476ms" Feb 13 19:19:48.746556 containerd[1515]: time="2025-02-13T19:19:48.746502649Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.1\" returns image reference \"sha256:421726ace5ed13894f7edf594dd3a462947aedc13d0f69d08525d7369477fb70\"" Feb 13 19:19:48.747681 containerd[1515]: time="2025-02-13T19:19:48.747648692Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.1\"" Feb 13 19:19:48.749533 containerd[1515]: time="2025-02-13T19:19:48.749381928Z" level=info msg="CreateContainer within sandbox \"f51a92f3451c546adf1a5f0d8406c487942da1a756304d943b422e13980bcaba\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Feb 13 19:19:48.766666 containerd[1515]: time="2025-02-13T19:19:48.766621595Z" level=info msg="CreateContainer within sandbox \"f51a92f3451c546adf1a5f0d8406c487942da1a756304d943b422e13980bcaba\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"bdd4562a1fc794aca7ce9f4cf5ff55230f1e20d82fc53095f4246fc95f867ae9\"" Feb 13 19:19:48.767234 containerd[1515]: time="2025-02-13T19:19:48.767184393Z" level=info msg="StartContainer for \"bdd4562a1fc794aca7ce9f4cf5ff55230f1e20d82fc53095f4246fc95f867ae9\"" Feb 13 19:19:48.795334 systemd[1]: Started cri-containerd-bdd4562a1fc794aca7ce9f4cf5ff55230f1e20d82fc53095f4246fc95f867ae9.scope - libcontainer container bdd4562a1fc794aca7ce9f4cf5ff55230f1e20d82fc53095f4246fc95f867ae9. Feb 13 19:19:48.856772 containerd[1515]: time="2025-02-13T19:19:48.856729275Z" level=info msg="StartContainer for \"bdd4562a1fc794aca7ce9f4cf5ff55230f1e20d82fc53095f4246fc95f867ae9\" returns successfully" Feb 13 19:19:49.064708 kubelet[2630]: I0213 19:19:49.064571 2630 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-86475b859c-kwlfl" podStartSLOduration=20.349822135 podStartE2EDuration="27.064556694s" podCreationTimestamp="2025-02-13 19:19:22 +0000 UTC" firstStartedPulling="2025-02-13 19:19:41.365666176 +0000 UTC m=+29.787519266" lastFinishedPulling="2025-02-13 19:19:48.080400734 +0000 UTC m=+36.502253825" observedRunningTime="2025-02-13 19:19:48.333616987 +0000 UTC m=+36.755470087" watchObservedRunningTime="2025-02-13 19:19:49.064556694 +0000 UTC m=+37.486409784" Feb 13 19:19:49.337097 kubelet[2630]: I0213 19:19:49.336950 2630 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-86475b859c-swqrn" podStartSLOduration=20.007672331 podStartE2EDuration="27.336934066s" podCreationTimestamp="2025-02-13 19:19:22 +0000 UTC" firstStartedPulling="2025-02-13 19:19:41.418223039 +0000 UTC m=+29.840076129" lastFinishedPulling="2025-02-13 19:19:48.747484774 +0000 UTC m=+37.169337864" observedRunningTime="2025-02-13 19:19:49.336753838 +0000 UTC m=+37.758606918" watchObservedRunningTime="2025-02-13 19:19:49.336934066 +0000 UTC m=+37.758787156" Feb 13 19:19:49.789348 systemd-networkd[1422]: vxlan.calico: Gained IPv6LL Feb 13 19:19:50.326460 kubelet[2630]: I0213 19:19:50.326434 2630 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 13 19:19:50.523375 systemd[1]: Started sshd@10-10.0.0.46:22-10.0.0.1:43610.service - OpenSSH per-connection server daemon (10.0.0.1:43610). Feb 13 19:19:50.580559 sshd[5931]: Accepted publickey for core from 10.0.0.1 port 43610 ssh2: RSA SHA256:xgLbxCKtIvCmXzj7C6d4ih050Hrbkh61XCRduaX62E8 Feb 13 19:19:50.582228 sshd-session[5931]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:19:50.586889 systemd-logind[1495]: New session 11 of user core. Feb 13 19:19:50.599364 systemd[1]: Started session-11.scope - Session 11 of User core. Feb 13 19:19:50.750726 sshd[5933]: Connection closed by 10.0.0.1 port 43610 Feb 13 19:19:50.751503 sshd-session[5931]: pam_unix(sshd:session): session closed for user core Feb 13 19:19:50.754458 systemd[1]: sshd@10-10.0.0.46:22-10.0.0.1:43610.service: Deactivated successfully. Feb 13 19:19:50.757308 systemd[1]: session-11.scope: Deactivated successfully. Feb 13 19:19:50.759700 systemd-logind[1495]: Session 11 logged out. Waiting for processes to exit. Feb 13 19:19:50.760640 systemd-logind[1495]: Removed session 11. Feb 13 19:19:50.838546 containerd[1515]: time="2025-02-13T19:19:50.838404547Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:19:50.839825 containerd[1515]: time="2025-02-13T19:19:50.839766445Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.29.1: active requests=0, bytes read=7902632" Feb 13 19:19:50.841091 containerd[1515]: time="2025-02-13T19:19:50.841050576Z" level=info msg="ImageCreate event name:\"sha256:bda8c42e04758c4f061339e213f50ccdc7502c4176fbf631aa12357e62b63540\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:19:50.843375 containerd[1515]: time="2025-02-13T19:19:50.843349765Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:eaa7e01fb16b603c155a67b81f16992281db7f831684c7b2081d3434587a7ff3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:19:50.843911 containerd[1515]: time="2025-02-13T19:19:50.843877547Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.29.1\" with image id \"sha256:bda8c42e04758c4f061339e213f50ccdc7502c4176fbf631aa12357e62b63540\", repo tag \"ghcr.io/flatcar/calico/csi:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:eaa7e01fb16b603c155a67b81f16992281db7f831684c7b2081d3434587a7ff3\", size \"9395716\" in 2.096187617s" Feb 13 19:19:50.843911 containerd[1515]: time="2025-02-13T19:19:50.843903886Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.1\" returns image reference \"sha256:bda8c42e04758c4f061339e213f50ccdc7502c4176fbf631aa12357e62b63540\"" Feb 13 19:19:50.845807 containerd[1515]: time="2025-02-13T19:19:50.845774970Z" level=info msg="CreateContainer within sandbox \"e22b101a1c7150c18eeb3428d7c5a80aba462dd305769816a18673cbfcf0ec44\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Feb 13 19:19:50.872853 containerd[1515]: time="2025-02-13T19:19:50.872795121Z" level=info msg="CreateContainer within sandbox \"e22b101a1c7150c18eeb3428d7c5a80aba462dd305769816a18673cbfcf0ec44\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"d54dbdf78a2a3bab62314e1bc992eebc1001064db2a0c747f99494567ec6b49f\"" Feb 13 19:19:50.873373 containerd[1515]: time="2025-02-13T19:19:50.873342208Z" level=info msg="StartContainer for \"d54dbdf78a2a3bab62314e1bc992eebc1001064db2a0c747f99494567ec6b49f\"" Feb 13 19:19:50.915042 systemd[1]: Started cri-containerd-d54dbdf78a2a3bab62314e1bc992eebc1001064db2a0c747f99494567ec6b49f.scope - libcontainer container d54dbdf78a2a3bab62314e1bc992eebc1001064db2a0c747f99494567ec6b49f. Feb 13 19:19:51.151279 containerd[1515]: time="2025-02-13T19:19:51.151143188Z" level=info msg="StartContainer for \"d54dbdf78a2a3bab62314e1bc992eebc1001064db2a0c747f99494567ec6b49f\" returns successfully" Feb 13 19:19:51.152357 containerd[1515]: time="2025-02-13T19:19:51.152325158Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\"" Feb 13 19:19:53.961531 containerd[1515]: time="2025-02-13T19:19:53.961468654Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:19:53.962559 containerd[1515]: time="2025-02-13T19:19:53.962522634Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1: active requests=0, bytes read=10501081" Feb 13 19:19:53.963953 containerd[1515]: time="2025-02-13T19:19:53.963925658Z" level=info msg="ImageCreate event name:\"sha256:8b7d18f262d5cf6a6343578ad0db68a140c4c9989d9e02c58c27cb5d2c70320f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:19:53.966640 containerd[1515]: time="2025-02-13T19:19:53.966581926Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:a338da9488cbaa83c78457c3d7354d84149969c0480e88dd768e036632ff5b76\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Feb 13 19:19:53.967201 containerd[1515]: time="2025-02-13T19:19:53.967139974Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\" with image id \"sha256:8b7d18f262d5cf6a6343578ad0db68a140c4c9989d9e02c58c27cb5d2c70320f\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:a338da9488cbaa83c78457c3d7354d84149969c0480e88dd768e036632ff5b76\", size \"11994117\" in 2.814776044s" Feb 13 19:19:53.967201 containerd[1515]: time="2025-02-13T19:19:53.967179378Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\" returns image reference \"sha256:8b7d18f262d5cf6a6343578ad0db68a140c4c9989d9e02c58c27cb5d2c70320f\"" Feb 13 19:19:53.969110 containerd[1515]: time="2025-02-13T19:19:53.969074236Z" level=info msg="CreateContainer within sandbox \"e22b101a1c7150c18eeb3428d7c5a80aba462dd305769816a18673cbfcf0ec44\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Feb 13 19:19:53.983493 containerd[1515]: time="2025-02-13T19:19:53.983420454Z" level=info msg="CreateContainer within sandbox \"e22b101a1c7150c18eeb3428d7c5a80aba462dd305769816a18673cbfcf0ec44\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"5b939046940f79abbf26e315bad7f3a95c2762ca1e2dc4581a58f97685e6d5a5\"" Feb 13 19:19:53.984258 containerd[1515]: time="2025-02-13T19:19:53.984222581Z" level=info msg="StartContainer for \"5b939046940f79abbf26e315bad7f3a95c2762ca1e2dc4581a58f97685e6d5a5\"" Feb 13 19:19:54.020391 systemd[1]: Started cri-containerd-5b939046940f79abbf26e315bad7f3a95c2762ca1e2dc4581a58f97685e6d5a5.scope - libcontainer container 5b939046940f79abbf26e315bad7f3a95c2762ca1e2dc4581a58f97685e6d5a5. Feb 13 19:19:54.055803 containerd[1515]: time="2025-02-13T19:19:54.055748910Z" level=info msg="StartContainer for \"5b939046940f79abbf26e315bad7f3a95c2762ca1e2dc4581a58f97685e6d5a5\" returns successfully" Feb 13 19:19:55.064406 kubelet[2630]: I0213 19:19:55.064374 2630 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Feb 13 19:19:55.064406 kubelet[2630]: I0213 19:19:55.064408 2630 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Feb 13 19:19:55.767325 systemd[1]: Started sshd@11-10.0.0.46:22-10.0.0.1:43618.service - OpenSSH per-connection server daemon (10.0.0.1:43618). Feb 13 19:19:55.835761 sshd[6040]: Accepted publickey for core from 10.0.0.1 port 43618 ssh2: RSA SHA256:xgLbxCKtIvCmXzj7C6d4ih050Hrbkh61XCRduaX62E8 Feb 13 19:19:55.837648 sshd-session[6040]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:19:55.842468 systemd-logind[1495]: New session 12 of user core. Feb 13 19:19:55.853502 systemd[1]: Started session-12.scope - Session 12 of User core. Feb 13 19:19:55.986172 sshd[6042]: Connection closed by 10.0.0.1 port 43618 Feb 13 19:19:55.986576 sshd-session[6040]: pam_unix(sshd:session): session closed for user core Feb 13 19:19:55.990848 systemd[1]: sshd@11-10.0.0.46:22-10.0.0.1:43618.service: Deactivated successfully. Feb 13 19:19:55.993238 systemd[1]: session-12.scope: Deactivated successfully. Feb 13 19:19:55.994215 systemd-logind[1495]: Session 12 logged out. Waiting for processes to exit. Feb 13 19:19:55.995558 systemd-logind[1495]: Removed session 12. Feb 13 19:20:01.018932 systemd[1]: Started sshd@12-10.0.0.46:22-10.0.0.1:43192.service - OpenSSH per-connection server daemon (10.0.0.1:43192). Feb 13 19:20:01.058152 sshd[6065]: Accepted publickey for core from 10.0.0.1 port 43192 ssh2: RSA SHA256:xgLbxCKtIvCmXzj7C6d4ih050Hrbkh61XCRduaX62E8 Feb 13 19:20:01.059932 sshd-session[6065]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:20:01.065056 systemd-logind[1495]: New session 13 of user core. Feb 13 19:20:01.074348 systemd[1]: Started session-13.scope - Session 13 of User core. Feb 13 19:20:01.211324 sshd[6067]: Connection closed by 10.0.0.1 port 43192 Feb 13 19:20:01.211719 sshd-session[6065]: pam_unix(sshd:session): session closed for user core Feb 13 19:20:01.225273 systemd[1]: sshd@12-10.0.0.46:22-10.0.0.1:43192.service: Deactivated successfully. Feb 13 19:20:01.228055 systemd[1]: session-13.scope: Deactivated successfully. Feb 13 19:20:01.230217 systemd-logind[1495]: Session 13 logged out. Waiting for processes to exit. Feb 13 19:20:01.235552 systemd[1]: Started sshd@13-10.0.0.46:22-10.0.0.1:43202.service - OpenSSH per-connection server daemon (10.0.0.1:43202). Feb 13 19:20:01.236691 systemd-logind[1495]: Removed session 13. Feb 13 19:20:01.278315 sshd[6080]: Accepted publickey for core from 10.0.0.1 port 43202 ssh2: RSA SHA256:xgLbxCKtIvCmXzj7C6d4ih050Hrbkh61XCRduaX62E8 Feb 13 19:20:01.279639 sshd-session[6080]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:20:01.283719 systemd-logind[1495]: New session 14 of user core. Feb 13 19:20:01.293331 systemd[1]: Started session-14.scope - Session 14 of User core. Feb 13 19:20:01.516339 sshd[6083]: Connection closed by 10.0.0.1 port 43202 Feb 13 19:20:01.516890 sshd-session[6080]: pam_unix(sshd:session): session closed for user core Feb 13 19:20:01.528402 systemd[1]: sshd@13-10.0.0.46:22-10.0.0.1:43202.service: Deactivated successfully. Feb 13 19:20:01.530823 systemd[1]: session-14.scope: Deactivated successfully. Feb 13 19:20:01.532626 systemd-logind[1495]: Session 14 logged out. Waiting for processes to exit. Feb 13 19:20:01.540608 systemd[1]: Started sshd@14-10.0.0.46:22-10.0.0.1:43206.service - OpenSSH per-connection server daemon (10.0.0.1:43206). Feb 13 19:20:01.542635 systemd-logind[1495]: Removed session 14. Feb 13 19:20:01.579059 sshd[6093]: Accepted publickey for core from 10.0.0.1 port 43206 ssh2: RSA SHA256:xgLbxCKtIvCmXzj7C6d4ih050Hrbkh61XCRduaX62E8 Feb 13 19:20:01.580844 sshd-session[6093]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:20:01.585382 systemd-logind[1495]: New session 15 of user core. Feb 13 19:20:01.594382 systemd[1]: Started session-15.scope - Session 15 of User core. Feb 13 19:20:01.786003 sshd[6096]: Connection closed by 10.0.0.1 port 43206 Feb 13 19:20:01.786317 sshd-session[6093]: pam_unix(sshd:session): session closed for user core Feb 13 19:20:01.790874 systemd[1]: sshd@14-10.0.0.46:22-10.0.0.1:43206.service: Deactivated successfully. Feb 13 19:20:01.793167 systemd[1]: session-15.scope: Deactivated successfully. Feb 13 19:20:01.793851 systemd-logind[1495]: Session 15 logged out. Waiting for processes to exit. Feb 13 19:20:01.794724 systemd-logind[1495]: Removed session 15. Feb 13 19:20:06.799443 systemd[1]: Started sshd@15-10.0.0.46:22-10.0.0.1:43268.service - OpenSSH per-connection server daemon (10.0.0.1:43268). Feb 13 19:20:06.846625 sshd[6125]: Accepted publickey for core from 10.0.0.1 port 43268 ssh2: RSA SHA256:xgLbxCKtIvCmXzj7C6d4ih050Hrbkh61XCRduaX62E8 Feb 13 19:20:06.848559 sshd-session[6125]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:20:06.853079 systemd-logind[1495]: New session 16 of user core. Feb 13 19:20:06.862325 systemd[1]: Started session-16.scope - Session 16 of User core. Feb 13 19:20:06.985740 sshd[6131]: Connection closed by 10.0.0.1 port 43268 Feb 13 19:20:06.986118 sshd-session[6125]: pam_unix(sshd:session): session closed for user core Feb 13 19:20:06.990748 systemd[1]: sshd@15-10.0.0.46:22-10.0.0.1:43268.service: Deactivated successfully. Feb 13 19:20:06.993628 systemd[1]: session-16.scope: Deactivated successfully. Feb 13 19:20:06.994411 systemd-logind[1495]: Session 16 logged out. Waiting for processes to exit. Feb 13 19:20:06.995334 systemd-logind[1495]: Removed session 16. Feb 13 19:20:07.779949 kubelet[2630]: I0213 19:20:07.779900 2630 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 13 19:20:07.810936 kubelet[2630]: I0213 19:20:07.810746 2630 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-drlnv" podStartSLOduration=33.457397338 podStartE2EDuration="45.810728219s" podCreationTimestamp="2025-02-13 19:19:22 +0000 UTC" firstStartedPulling="2025-02-13 19:19:41.614571384 +0000 UTC m=+30.036424474" lastFinishedPulling="2025-02-13 19:19:53.967902265 +0000 UTC m=+42.389755355" observedRunningTime="2025-02-13 19:19:54.35295573 +0000 UTC m=+42.774808820" watchObservedRunningTime="2025-02-13 19:20:07.810728219 +0000 UTC m=+56.232581309" Feb 13 19:20:11.651733 containerd[1515]: time="2025-02-13T19:20:11.651689649Z" level=info msg="StopPodSandbox for \"d195bcf8f9551fdb26f6b42c8b0e4b63d6953f1bbdc1c4b497262fc12be34248\"" Feb 13 19:20:11.652356 containerd[1515]: time="2025-02-13T19:20:11.651817297Z" level=info msg="TearDown network for sandbox \"d195bcf8f9551fdb26f6b42c8b0e4b63d6953f1bbdc1c4b497262fc12be34248\" successfully" Feb 13 19:20:11.652356 containerd[1515]: time="2025-02-13T19:20:11.651842856Z" level=info msg="StopPodSandbox for \"d195bcf8f9551fdb26f6b42c8b0e4b63d6953f1bbdc1c4b497262fc12be34248\" returns successfully" Feb 13 19:20:11.652356 containerd[1515]: time="2025-02-13T19:20:11.652215016Z" level=info msg="RemovePodSandbox for \"d195bcf8f9551fdb26f6b42c8b0e4b63d6953f1bbdc1c4b497262fc12be34248\"" Feb 13 19:20:11.656030 containerd[1515]: time="2025-02-13T19:20:11.655997747Z" level=info msg="Forcibly stopping sandbox \"d195bcf8f9551fdb26f6b42c8b0e4b63d6953f1bbdc1c4b497262fc12be34248\"" Feb 13 19:20:11.656143 containerd[1515]: time="2025-02-13T19:20:11.656091770Z" level=info msg="TearDown network for sandbox \"d195bcf8f9551fdb26f6b42c8b0e4b63d6953f1bbdc1c4b497262fc12be34248\" successfully" Feb 13 19:20:12.002468 systemd[1]: Started sshd@16-10.0.0.46:22-10.0.0.1:43280.service - OpenSSH per-connection server daemon (10.0.0.1:43280). Feb 13 19:20:12.027452 containerd[1515]: time="2025-02-13T19:20:12.027252995Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"d195bcf8f9551fdb26f6b42c8b0e4b63d6953f1bbdc1c4b497262fc12be34248\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:20:12.027452 containerd[1515]: time="2025-02-13T19:20:12.027324613Z" level=info msg="RemovePodSandbox \"d195bcf8f9551fdb26f6b42c8b0e4b63d6953f1bbdc1c4b497262fc12be34248\" returns successfully" Feb 13 19:20:12.028159 containerd[1515]: time="2025-02-13T19:20:12.027931746Z" level=info msg="StopPodSandbox for \"7d827c446a1e808002e0f19943157ab97bcc163d150e79960814ed79c5b4fbf3\"" Feb 13 19:20:12.028159 containerd[1515]: time="2025-02-13T19:20:12.028020899Z" level=info msg="TearDown network for sandbox \"7d827c446a1e808002e0f19943157ab97bcc163d150e79960814ed79c5b4fbf3\" successfully" Feb 13 19:20:12.028159 containerd[1515]: time="2025-02-13T19:20:12.028031479Z" level=info msg="StopPodSandbox for \"7d827c446a1e808002e0f19943157ab97bcc163d150e79960814ed79c5b4fbf3\" returns successfully" Feb 13 19:20:12.028640 containerd[1515]: time="2025-02-13T19:20:12.028551754Z" level=info msg="RemovePodSandbox for \"7d827c446a1e808002e0f19943157ab97bcc163d150e79960814ed79c5b4fbf3\"" Feb 13 19:20:12.028640 containerd[1515]: time="2025-02-13T19:20:12.028572795Z" level=info msg="Forcibly stopping sandbox \"7d827c446a1e808002e0f19943157ab97bcc163d150e79960814ed79c5b4fbf3\"" Feb 13 19:20:12.030229 containerd[1515]: time="2025-02-13T19:20:12.028764965Z" level=info msg="TearDown network for sandbox \"7d827c446a1e808002e0f19943157ab97bcc163d150e79960814ed79c5b4fbf3\" successfully" Feb 13 19:20:12.049719 sshd[6181]: Accepted publickey for core from 10.0.0.1 port 43280 ssh2: RSA SHA256:xgLbxCKtIvCmXzj7C6d4ih050Hrbkh61XCRduaX62E8 Feb 13 19:20:12.073310 sshd-session[6181]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:20:12.078798 systemd-logind[1495]: New session 17 of user core. Feb 13 19:20:12.087322 systemd[1]: Started session-17.scope - Session 17 of User core. Feb 13 19:20:12.157168 containerd[1515]: time="2025-02-13T19:20:12.157005888Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"7d827c446a1e808002e0f19943157ab97bcc163d150e79960814ed79c5b4fbf3\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:20:12.157168 containerd[1515]: time="2025-02-13T19:20:12.157080623Z" level=info msg="RemovePodSandbox \"7d827c446a1e808002e0f19943157ab97bcc163d150e79960814ed79c5b4fbf3\" returns successfully" Feb 13 19:20:12.159236 containerd[1515]: time="2025-02-13T19:20:12.158817047Z" level=info msg="StopPodSandbox for \"154203fe3c4258e99eeb24eefabf42a73b9226503cd839391e62422056e09db6\"" Feb 13 19:20:12.159236 containerd[1515]: time="2025-02-13T19:20:12.158956005Z" level=info msg="TearDown network for sandbox \"154203fe3c4258e99eeb24eefabf42a73b9226503cd839391e62422056e09db6\" successfully" Feb 13 19:20:12.159236 containerd[1515]: time="2025-02-13T19:20:12.158966635Z" level=info msg="StopPodSandbox for \"154203fe3c4258e99eeb24eefabf42a73b9226503cd839391e62422056e09db6\" returns successfully" Feb 13 19:20:12.159520 containerd[1515]: time="2025-02-13T19:20:12.159443066Z" level=info msg="RemovePodSandbox for \"154203fe3c4258e99eeb24eefabf42a73b9226503cd839391e62422056e09db6\"" Feb 13 19:20:12.159591 containerd[1515]: time="2025-02-13T19:20:12.159561505Z" level=info msg="Forcibly stopping sandbox \"154203fe3c4258e99eeb24eefabf42a73b9226503cd839391e62422056e09db6\"" Feb 13 19:20:12.159888 containerd[1515]: time="2025-02-13T19:20:12.159815756Z" level=info msg="TearDown network for sandbox \"154203fe3c4258e99eeb24eefabf42a73b9226503cd839391e62422056e09db6\" successfully" Feb 13 19:20:12.233690 containerd[1515]: time="2025-02-13T19:20:12.233585159Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"154203fe3c4258e99eeb24eefabf42a73b9226503cd839391e62422056e09db6\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:20:12.233690 containerd[1515]: time="2025-02-13T19:20:12.233664331Z" level=info msg="RemovePodSandbox \"154203fe3c4258e99eeb24eefabf42a73b9226503cd839391e62422056e09db6\" returns successfully" Feb 13 19:20:12.234495 containerd[1515]: time="2025-02-13T19:20:12.234461531Z" level=info msg="StopPodSandbox for \"8a568a618a4c0eab0195822f2ccafce3787a78e075bf7c3f5b50c840e692188a\"" Feb 13 19:20:12.234706 containerd[1515]: time="2025-02-13T19:20:12.234645718Z" level=info msg="TearDown network for sandbox \"8a568a618a4c0eab0195822f2ccafce3787a78e075bf7c3f5b50c840e692188a\" successfully" Feb 13 19:20:12.234786 containerd[1515]: time="2025-02-13T19:20:12.234662149Z" level=info msg="StopPodSandbox for \"8a568a618a4c0eab0195822f2ccafce3787a78e075bf7c3f5b50c840e692188a\" returns successfully" Feb 13 19:20:12.236272 containerd[1515]: time="2025-02-13T19:20:12.235367331Z" level=info msg="RemovePodSandbox for \"8a568a618a4c0eab0195822f2ccafce3787a78e075bf7c3f5b50c840e692188a\"" Feb 13 19:20:12.236272 containerd[1515]: time="2025-02-13T19:20:12.235401707Z" level=info msg="Forcibly stopping sandbox \"8a568a618a4c0eab0195822f2ccafce3787a78e075bf7c3f5b50c840e692188a\"" Feb 13 19:20:12.236272 containerd[1515]: time="2025-02-13T19:20:12.235498304Z" level=info msg="TearDown network for sandbox \"8a568a618a4c0eab0195822f2ccafce3787a78e075bf7c3f5b50c840e692188a\" successfully" Feb 13 19:20:12.258350 containerd[1515]: time="2025-02-13T19:20:12.258160435Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"8a568a618a4c0eab0195822f2ccafce3787a78e075bf7c3f5b50c840e692188a\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:20:12.259234 containerd[1515]: time="2025-02-13T19:20:12.259030926Z" level=info msg="RemovePodSandbox \"8a568a618a4c0eab0195822f2ccafce3787a78e075bf7c3f5b50c840e692188a\" returns successfully" Feb 13 19:20:12.260326 containerd[1515]: time="2025-02-13T19:20:12.260288255Z" level=info msg="StopPodSandbox for \"70821f98f9d2702e4d63972ae34f265d6851151a70f0070211ef163eb0840795\"" Feb 13 19:20:12.260478 containerd[1515]: time="2025-02-13T19:20:12.260394340Z" level=info msg="TearDown network for sandbox \"70821f98f9d2702e4d63972ae34f265d6851151a70f0070211ef163eb0840795\" successfully" Feb 13 19:20:12.260478 containerd[1515]: time="2025-02-13T19:20:12.260404770Z" level=info msg="StopPodSandbox for \"70821f98f9d2702e4d63972ae34f265d6851151a70f0070211ef163eb0840795\" returns successfully" Feb 13 19:20:12.260671 containerd[1515]: time="2025-02-13T19:20:12.260652088Z" level=info msg="RemovePodSandbox for \"70821f98f9d2702e4d63972ae34f265d6851151a70f0070211ef163eb0840795\"" Feb 13 19:20:12.260702 containerd[1515]: time="2025-02-13T19:20:12.260672417Z" level=info msg="Forcibly stopping sandbox \"70821f98f9d2702e4d63972ae34f265d6851151a70f0070211ef163eb0840795\"" Feb 13 19:20:12.260776 containerd[1515]: time="2025-02-13T19:20:12.260738554Z" level=info msg="TearDown network for sandbox \"70821f98f9d2702e4d63972ae34f265d6851151a70f0070211ef163eb0840795\" successfully" Feb 13 19:20:12.263718 sshd[6183]: Connection closed by 10.0.0.1 port 43280 Feb 13 19:20:12.264069 sshd-session[6181]: pam_unix(sshd:session): session closed for user core Feb 13 19:20:12.269407 systemd[1]: sshd@16-10.0.0.46:22-10.0.0.1:43280.service: Deactivated successfully. Feb 13 19:20:12.269576 containerd[1515]: time="2025-02-13T19:20:12.269503724Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"70821f98f9d2702e4d63972ae34f265d6851151a70f0070211ef163eb0840795\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:20:12.269576 containerd[1515]: time="2025-02-13T19:20:12.269575553Z" level=info msg="RemovePodSandbox \"70821f98f9d2702e4d63972ae34f265d6851151a70f0070211ef163eb0840795\" returns successfully" Feb 13 19:20:12.270220 containerd[1515]: time="2025-02-13T19:20:12.270099515Z" level=info msg="StopPodSandbox for \"4d2e4f99dab68adc0de82a705f6ff963ae519cfd54bd2435b6836ee53a8c4cf9\"" Feb 13 19:20:12.270336 containerd[1515]: time="2025-02-13T19:20:12.270312858Z" level=info msg="TearDown network for sandbox \"4d2e4f99dab68adc0de82a705f6ff963ae519cfd54bd2435b6836ee53a8c4cf9\" successfully" Feb 13 19:20:12.270336 containerd[1515]: time="2025-02-13T19:20:12.270329400Z" level=info msg="StopPodSandbox for \"4d2e4f99dab68adc0de82a705f6ff963ae519cfd54bd2435b6836ee53a8c4cf9\" returns successfully" Feb 13 19:20:12.270692 containerd[1515]: time="2025-02-13T19:20:12.270667923Z" level=info msg="RemovePodSandbox for \"4d2e4f99dab68adc0de82a705f6ff963ae519cfd54bd2435b6836ee53a8c4cf9\"" Feb 13 19:20:12.270692 containerd[1515]: time="2025-02-13T19:20:12.270690557Z" level=info msg="Forcibly stopping sandbox \"4d2e4f99dab68adc0de82a705f6ff963ae519cfd54bd2435b6836ee53a8c4cf9\"" Feb 13 19:20:12.271013 containerd[1515]: time="2025-02-13T19:20:12.270900623Z" level=info msg="TearDown network for sandbox \"4d2e4f99dab68adc0de82a705f6ff963ae519cfd54bd2435b6836ee53a8c4cf9\" successfully" Feb 13 19:20:12.272263 systemd[1]: session-17.scope: Deactivated successfully. Feb 13 19:20:12.273711 systemd-logind[1495]: Session 17 logged out. Waiting for processes to exit. Feb 13 19:20:12.275214 systemd-logind[1495]: Removed session 17. Feb 13 19:20:12.279136 containerd[1515]: time="2025-02-13T19:20:12.279081995Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"4d2e4f99dab68adc0de82a705f6ff963ae519cfd54bd2435b6836ee53a8c4cf9\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:20:12.279256 containerd[1515]: time="2025-02-13T19:20:12.279158803Z" level=info msg="RemovePodSandbox \"4d2e4f99dab68adc0de82a705f6ff963ae519cfd54bd2435b6836ee53a8c4cf9\" returns successfully" Feb 13 19:20:12.279943 containerd[1515]: time="2025-02-13T19:20:12.279734484Z" level=info msg="StopPodSandbox for \"e2995529ca797998e1cf161d883bb684a3464a8a8d57602758f8705313f0d601\"" Feb 13 19:20:12.279943 containerd[1515]: time="2025-02-13T19:20:12.279879344Z" level=info msg="TearDown network for sandbox \"e2995529ca797998e1cf161d883bb684a3464a8a8d57602758f8705313f0d601\" successfully" Feb 13 19:20:12.279943 containerd[1515]: time="2025-02-13T19:20:12.279892299Z" level=info msg="StopPodSandbox for \"e2995529ca797998e1cf161d883bb684a3464a8a8d57602758f8705313f0d601\" returns successfully" Feb 13 19:20:12.280509 containerd[1515]: time="2025-02-13T19:20:12.280439276Z" level=info msg="RemovePodSandbox for \"e2995529ca797998e1cf161d883bb684a3464a8a8d57602758f8705313f0d601\"" Feb 13 19:20:12.280645 containerd[1515]: time="2025-02-13T19:20:12.280518689Z" level=info msg="Forcibly stopping sandbox \"e2995529ca797998e1cf161d883bb684a3464a8a8d57602758f8705313f0d601\"" Feb 13 19:20:12.280645 containerd[1515]: time="2025-02-13T19:20:12.280610877Z" level=info msg="TearDown network for sandbox \"e2995529ca797998e1cf161d883bb684a3464a8a8d57602758f8705313f0d601\" successfully" Feb 13 19:20:12.285899 containerd[1515]: time="2025-02-13T19:20:12.285749011Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"e2995529ca797998e1cf161d883bb684a3464a8a8d57602758f8705313f0d601\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:20:12.285899 containerd[1515]: time="2025-02-13T19:20:12.285797565Z" level=info msg="RemovePodSandbox \"e2995529ca797998e1cf161d883bb684a3464a8a8d57602758f8705313f0d601\" returns successfully" Feb 13 19:20:12.287043 containerd[1515]: time="2025-02-13T19:20:12.287007041Z" level=info msg="StopPodSandbox for \"706e09ab546bf3d89157cd9a2c32b16d198260f6686345817b1acf65f680fdd2\"" Feb 13 19:20:12.287162 containerd[1515]: time="2025-02-13T19:20:12.287141001Z" level=info msg="TearDown network for sandbox \"706e09ab546bf3d89157cd9a2c32b16d198260f6686345817b1acf65f680fdd2\" successfully" Feb 13 19:20:12.287162 containerd[1515]: time="2025-02-13T19:20:12.287157191Z" level=info msg="StopPodSandbox for \"706e09ab546bf3d89157cd9a2c32b16d198260f6686345817b1acf65f680fdd2\" returns successfully" Feb 13 19:20:12.287470 containerd[1515]: time="2025-02-13T19:20:12.287440618Z" level=info msg="RemovePodSandbox for \"706e09ab546bf3d89157cd9a2c32b16d198260f6686345817b1acf65f680fdd2\"" Feb 13 19:20:12.287516 containerd[1515]: time="2025-02-13T19:20:12.287470226Z" level=info msg="Forcibly stopping sandbox \"706e09ab546bf3d89157cd9a2c32b16d198260f6686345817b1acf65f680fdd2\"" Feb 13 19:20:12.287571 containerd[1515]: time="2025-02-13T19:20:12.287530934Z" level=info msg="TearDown network for sandbox \"706e09ab546bf3d89157cd9a2c32b16d198260f6686345817b1acf65f680fdd2\" successfully" Feb 13 19:20:12.374172 containerd[1515]: time="2025-02-13T19:20:12.374084779Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"706e09ab546bf3d89157cd9a2c32b16d198260f6686345817b1acf65f680fdd2\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:20:12.374172 containerd[1515]: time="2025-02-13T19:20:12.374171627Z" level=info msg="RemovePodSandbox \"706e09ab546bf3d89157cd9a2c32b16d198260f6686345817b1acf65f680fdd2\" returns successfully" Feb 13 19:20:12.374803 containerd[1515]: time="2025-02-13T19:20:12.374747559Z" level=info msg="StopPodSandbox for \"8509c778c3294467a0ad8b8059bdecd145729001ee04e3c09d32d7f0d60b9468\"" Feb 13 19:20:12.374941 containerd[1515]: time="2025-02-13T19:20:12.374898471Z" level=info msg="TearDown network for sandbox \"8509c778c3294467a0ad8b8059bdecd145729001ee04e3c09d32d7f0d60b9468\" successfully" Feb 13 19:20:12.374941 containerd[1515]: time="2025-02-13T19:20:12.374926775Z" level=info msg="StopPodSandbox for \"8509c778c3294467a0ad8b8059bdecd145729001ee04e3c09d32d7f0d60b9468\" returns successfully" Feb 13 19:20:12.375270 containerd[1515]: time="2025-02-13T19:20:12.375231854Z" level=info msg="RemovePodSandbox for \"8509c778c3294467a0ad8b8059bdecd145729001ee04e3c09d32d7f0d60b9468\"" Feb 13 19:20:12.375270 containerd[1515]: time="2025-02-13T19:20:12.375264147Z" level=info msg="Forcibly stopping sandbox \"8509c778c3294467a0ad8b8059bdecd145729001ee04e3c09d32d7f0d60b9468\"" Feb 13 19:20:12.375410 containerd[1515]: time="2025-02-13T19:20:12.375352267Z" level=info msg="TearDown network for sandbox \"8509c778c3294467a0ad8b8059bdecd145729001ee04e3c09d32d7f0d60b9468\" successfully" Feb 13 19:20:12.390569 containerd[1515]: time="2025-02-13T19:20:12.390520704Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"8509c778c3294467a0ad8b8059bdecd145729001ee04e3c09d32d7f0d60b9468\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:20:12.391248 containerd[1515]: time="2025-02-13T19:20:12.390698818Z" level=info msg="RemovePodSandbox \"8509c778c3294467a0ad8b8059bdecd145729001ee04e3c09d32d7f0d60b9468\" returns successfully" Feb 13 19:20:12.392563 containerd[1515]: time="2025-02-13T19:20:12.391960626Z" level=info msg="StopPodSandbox for \"fd762dfcc2c780c334b7852f3da6760c5d53d6cb6d243b5ae4cf1085971b8f57\"" Feb 13 19:20:12.392563 containerd[1515]: time="2025-02-13T19:20:12.392455462Z" level=info msg="TearDown network for sandbox \"fd762dfcc2c780c334b7852f3da6760c5d53d6cb6d243b5ae4cf1085971b8f57\" successfully" Feb 13 19:20:12.392563 containerd[1515]: time="2025-02-13T19:20:12.392470229Z" level=info msg="StopPodSandbox for \"fd762dfcc2c780c334b7852f3da6760c5d53d6cb6d243b5ae4cf1085971b8f57\" returns successfully" Feb 13 19:20:12.392795 containerd[1515]: time="2025-02-13T19:20:12.392731604Z" level=info msg="RemovePodSandbox for \"fd762dfcc2c780c334b7852f3da6760c5d53d6cb6d243b5ae4cf1085971b8f57\"" Feb 13 19:20:12.392795 containerd[1515]: time="2025-02-13T19:20:12.392754018Z" level=info msg="Forcibly stopping sandbox \"fd762dfcc2c780c334b7852f3da6760c5d53d6cb6d243b5ae4cf1085971b8f57\"" Feb 13 19:20:12.392937 containerd[1515]: time="2025-02-13T19:20:12.392836808Z" level=info msg="TearDown network for sandbox \"fd762dfcc2c780c334b7852f3da6760c5d53d6cb6d243b5ae4cf1085971b8f57\" successfully" Feb 13 19:20:12.399736 containerd[1515]: time="2025-02-13T19:20:12.399625871Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"fd762dfcc2c780c334b7852f3da6760c5d53d6cb6d243b5ae4cf1085971b8f57\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:20:12.399736 containerd[1515]: time="2025-02-13T19:20:12.399728249Z" level=info msg="RemovePodSandbox \"fd762dfcc2c780c334b7852f3da6760c5d53d6cb6d243b5ae4cf1085971b8f57\" returns successfully" Feb 13 19:20:12.400373 containerd[1515]: time="2025-02-13T19:20:12.400334189Z" level=info msg="StopPodSandbox for \"913a65e5860c4b13ea05638c4cdf916cfe4d5179d77f6f08a60f9aa7332fc559\"" Feb 13 19:20:12.400499 containerd[1515]: time="2025-02-13T19:20:12.400448890Z" level=info msg="TearDown network for sandbox \"913a65e5860c4b13ea05638c4cdf916cfe4d5179d77f6f08a60f9aa7332fc559\" successfully" Feb 13 19:20:12.400499 containerd[1515]: time="2025-02-13T19:20:12.400458710Z" level=info msg="StopPodSandbox for \"913a65e5860c4b13ea05638c4cdf916cfe4d5179d77f6f08a60f9aa7332fc559\" returns successfully" Feb 13 19:20:12.401321 containerd[1515]: time="2025-02-13T19:20:12.400924179Z" level=info msg="RemovePodSandbox for \"913a65e5860c4b13ea05638c4cdf916cfe4d5179d77f6f08a60f9aa7332fc559\"" Feb 13 19:20:12.401321 containerd[1515]: time="2025-02-13T19:20:12.400946692Z" level=info msg="Forcibly stopping sandbox \"913a65e5860c4b13ea05638c4cdf916cfe4d5179d77f6f08a60f9aa7332fc559\"" Feb 13 19:20:12.401321 containerd[1515]: time="2025-02-13T19:20:12.401011698Z" level=info msg="TearDown network for sandbox \"913a65e5860c4b13ea05638c4cdf916cfe4d5179d77f6f08a60f9aa7332fc559\" successfully" Feb 13 19:20:12.405406 containerd[1515]: time="2025-02-13T19:20:12.405369715Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"913a65e5860c4b13ea05638c4cdf916cfe4d5179d77f6f08a60f9aa7332fc559\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:20:12.405547 containerd[1515]: time="2025-02-13T19:20:12.405520025Z" level=info msg="RemovePodSandbox \"913a65e5860c4b13ea05638c4cdf916cfe4d5179d77f6f08a60f9aa7332fc559\" returns successfully" Feb 13 19:20:12.405823 containerd[1515]: time="2025-02-13T19:20:12.405802040Z" level=info msg="StopPodSandbox for \"e941b39f4d4a62064805d7ae584462196d1704f1eaaa88f9ee2cc17abc84598e\"" Feb 13 19:20:12.405914 containerd[1515]: time="2025-02-13T19:20:12.405894478Z" level=info msg="TearDown network for sandbox \"e941b39f4d4a62064805d7ae584462196d1704f1eaaa88f9ee2cc17abc84598e\" successfully" Feb 13 19:20:12.405914 containerd[1515]: time="2025-02-13T19:20:12.405909908Z" level=info msg="StopPodSandbox for \"e941b39f4d4a62064805d7ae584462196d1704f1eaaa88f9ee2cc17abc84598e\" returns successfully" Feb 13 19:20:12.406216 containerd[1515]: time="2025-02-13T19:20:12.406161654Z" level=info msg="RemovePodSandbox for \"e941b39f4d4a62064805d7ae584462196d1704f1eaaa88f9ee2cc17abc84598e\"" Feb 13 19:20:12.406320 containerd[1515]: time="2025-02-13T19:20:12.406301144Z" level=info msg="Forcibly stopping sandbox \"e941b39f4d4a62064805d7ae584462196d1704f1eaaa88f9ee2cc17abc84598e\"" Feb 13 19:20:12.406409 containerd[1515]: time="2025-02-13T19:20:12.406372482Z" level=info msg="TearDown network for sandbox \"e941b39f4d4a62064805d7ae584462196d1704f1eaaa88f9ee2cc17abc84598e\" successfully" Feb 13 19:20:12.410307 containerd[1515]: time="2025-02-13T19:20:12.410173093Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"e941b39f4d4a62064805d7ae584462196d1704f1eaaa88f9ee2cc17abc84598e\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:20:12.410307 containerd[1515]: time="2025-02-13T19:20:12.410234821Z" level=info msg="RemovePodSandbox \"e941b39f4d4a62064805d7ae584462196d1704f1eaaa88f9ee2cc17abc84598e\" returns successfully" Feb 13 19:20:12.411140 containerd[1515]: time="2025-02-13T19:20:12.411079874Z" level=info msg="StopPodSandbox for \"f7214e7c2d9f81df76f610a79140dcb0085636bf81f33d1eb139eb52cb12a6af\"" Feb 13 19:20:12.411202 containerd[1515]: time="2025-02-13T19:20:12.411176170Z" level=info msg="TearDown network for sandbox \"f7214e7c2d9f81df76f610a79140dcb0085636bf81f33d1eb139eb52cb12a6af\" successfully" Feb 13 19:20:12.411259 containerd[1515]: time="2025-02-13T19:20:12.411202030Z" level=info msg="StopPodSandbox for \"f7214e7c2d9f81df76f610a79140dcb0085636bf81f33d1eb139eb52cb12a6af\" returns successfully" Feb 13 19:20:12.412487 containerd[1515]: time="2025-02-13T19:20:12.411450560Z" level=info msg="RemovePodSandbox for \"f7214e7c2d9f81df76f610a79140dcb0085636bf81f33d1eb139eb52cb12a6af\"" Feb 13 19:20:12.412487 containerd[1515]: time="2025-02-13T19:20:12.411470639Z" level=info msg="Forcibly stopping sandbox \"f7214e7c2d9f81df76f610a79140dcb0085636bf81f33d1eb139eb52cb12a6af\"" Feb 13 19:20:12.412487 containerd[1515]: time="2025-02-13T19:20:12.411529061Z" level=info msg="TearDown network for sandbox \"f7214e7c2d9f81df76f610a79140dcb0085636bf81f33d1eb139eb52cb12a6af\" successfully" Feb 13 19:20:12.416390 containerd[1515]: time="2025-02-13T19:20:12.416343440Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"f7214e7c2d9f81df76f610a79140dcb0085636bf81f33d1eb139eb52cb12a6af\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:20:12.416522 containerd[1515]: time="2025-02-13T19:20:12.416406362Z" level=info msg="RemovePodSandbox \"f7214e7c2d9f81df76f610a79140dcb0085636bf81f33d1eb139eb52cb12a6af\" returns successfully" Feb 13 19:20:12.416976 containerd[1515]: time="2025-02-13T19:20:12.416822926Z" level=info msg="StopPodSandbox for \"b4101238369440d1fc253b3e442bebf940a9f5fa609c91f1252b6dcea502c067\"" Feb 13 19:20:12.416976 containerd[1515]: time="2025-02-13T19:20:12.416916788Z" level=info msg="TearDown network for sandbox \"b4101238369440d1fc253b3e442bebf940a9f5fa609c91f1252b6dcea502c067\" successfully" Feb 13 19:20:12.416976 containerd[1515]: time="2025-02-13T19:20:12.416926256Z" level=info msg="StopPodSandbox for \"b4101238369440d1fc253b3e442bebf940a9f5fa609c91f1252b6dcea502c067\" returns successfully" Feb 13 19:20:12.417268 containerd[1515]: time="2025-02-13T19:20:12.417224311Z" level=info msg="RemovePodSandbox for \"b4101238369440d1fc253b3e442bebf940a9f5fa609c91f1252b6dcea502c067\"" Feb 13 19:20:12.417268 containerd[1515]: time="2025-02-13T19:20:12.417254310Z" level=info msg="Forcibly stopping sandbox \"b4101238369440d1fc253b3e442bebf940a9f5fa609c91f1252b6dcea502c067\"" Feb 13 19:20:12.417532 containerd[1515]: time="2025-02-13T19:20:12.417344845Z" level=info msg="TearDown network for sandbox \"b4101238369440d1fc253b3e442bebf940a9f5fa609c91f1252b6dcea502c067\" successfully" Feb 13 19:20:12.423310 containerd[1515]: time="2025-02-13T19:20:12.423265851Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"b4101238369440d1fc253b3e442bebf940a9f5fa609c91f1252b6dcea502c067\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:20:12.423310 containerd[1515]: time="2025-02-13T19:20:12.423311729Z" level=info msg="RemovePodSandbox \"b4101238369440d1fc253b3e442bebf940a9f5fa609c91f1252b6dcea502c067\" returns successfully" Feb 13 19:20:12.423783 containerd[1515]: time="2025-02-13T19:20:12.423738603Z" level=info msg="StopPodSandbox for \"787aa28bb52b7073a96ced7755cf59d24334348abd3a51ce97858110663e1200\"" Feb 13 19:20:12.423941 containerd[1515]: time="2025-02-13T19:20:12.423909364Z" level=info msg="TearDown network for sandbox \"787aa28bb52b7073a96ced7755cf59d24334348abd3a51ce97858110663e1200\" successfully" Feb 13 19:20:12.423941 containerd[1515]: time="2025-02-13T19:20:12.423933099Z" level=info msg="StopPodSandbox for \"787aa28bb52b7073a96ced7755cf59d24334348abd3a51ce97858110663e1200\" returns successfully" Feb 13 19:20:12.424403 containerd[1515]: time="2025-02-13T19:20:12.424376536Z" level=info msg="RemovePodSandbox for \"787aa28bb52b7073a96ced7755cf59d24334348abd3a51ce97858110663e1200\"" Feb 13 19:20:12.424503 containerd[1515]: time="2025-02-13T19:20:12.424481979Z" level=info msg="Forcibly stopping sandbox \"787aa28bb52b7073a96ced7755cf59d24334348abd3a51ce97858110663e1200\"" Feb 13 19:20:12.424613 containerd[1515]: time="2025-02-13T19:20:12.424574608Z" level=info msg="TearDown network for sandbox \"787aa28bb52b7073a96ced7755cf59d24334348abd3a51ce97858110663e1200\" successfully" Feb 13 19:20:12.429332 containerd[1515]: time="2025-02-13T19:20:12.429285788Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"787aa28bb52b7073a96ced7755cf59d24334348abd3a51ce97858110663e1200\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:20:12.429332 containerd[1515]: time="2025-02-13T19:20:12.429320705Z" level=info msg="RemovePodSandbox \"787aa28bb52b7073a96ced7755cf59d24334348abd3a51ce97858110663e1200\" returns successfully" Feb 13 19:20:12.429776 containerd[1515]: time="2025-02-13T19:20:12.429750656Z" level=info msg="StopPodSandbox for \"c03a59deff35e415f0920de8fd1c0462db72392ea452d0eafd388e65b284eb39\"" Feb 13 19:20:12.429877 containerd[1515]: time="2025-02-13T19:20:12.429848214Z" level=info msg="TearDown network for sandbox \"c03a59deff35e415f0920de8fd1c0462db72392ea452d0eafd388e65b284eb39\" successfully" Feb 13 19:20:12.429925 containerd[1515]: time="2025-02-13T19:20:12.429874845Z" level=info msg="StopPodSandbox for \"c03a59deff35e415f0920de8fd1c0462db72392ea452d0eafd388e65b284eb39\" returns successfully" Feb 13 19:20:12.430282 containerd[1515]: time="2025-02-13T19:20:12.430237126Z" level=info msg="RemovePodSandbox for \"c03a59deff35e415f0920de8fd1c0462db72392ea452d0eafd388e65b284eb39\"" Feb 13 19:20:12.430282 containerd[1515]: time="2025-02-13T19:20:12.430261563Z" level=info msg="Forcibly stopping sandbox \"c03a59deff35e415f0920de8fd1c0462db72392ea452d0eafd388e65b284eb39\"" Feb 13 19:20:12.430380 containerd[1515]: time="2025-02-13T19:20:12.430337549Z" level=info msg="TearDown network for sandbox \"c03a59deff35e415f0920de8fd1c0462db72392ea452d0eafd388e65b284eb39\" successfully" Feb 13 19:20:12.434025 containerd[1515]: time="2025-02-13T19:20:12.433970586Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"c03a59deff35e415f0920de8fd1c0462db72392ea452d0eafd388e65b284eb39\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:20:12.434025 containerd[1515]: time="2025-02-13T19:20:12.434006726Z" level=info msg="RemovePodSandbox \"c03a59deff35e415f0920de8fd1c0462db72392ea452d0eafd388e65b284eb39\" returns successfully" Feb 13 19:20:12.434417 containerd[1515]: time="2025-02-13T19:20:12.434378624Z" level=info msg="StopPodSandbox for \"52abc7c7229296fa61042f80a0bb0c1bc30c4e6c9818f7329496601bde28a193\"" Feb 13 19:20:12.434520 containerd[1515]: time="2025-02-13T19:20:12.434469971Z" level=info msg="TearDown network for sandbox \"52abc7c7229296fa61042f80a0bb0c1bc30c4e6c9818f7329496601bde28a193\" successfully" Feb 13 19:20:12.434615 containerd[1515]: time="2025-02-13T19:20:12.434583691Z" level=info msg="StopPodSandbox for \"52abc7c7229296fa61042f80a0bb0c1bc30c4e6c9818f7329496601bde28a193\" returns successfully" Feb 13 19:20:12.434957 containerd[1515]: time="2025-02-13T19:20:12.434913627Z" level=info msg="RemovePodSandbox for \"52abc7c7229296fa61042f80a0bb0c1bc30c4e6c9818f7329496601bde28a193\"" Feb 13 19:20:12.434957 containerd[1515]: time="2025-02-13T19:20:12.434959867Z" level=info msg="Forcibly stopping sandbox \"52abc7c7229296fa61042f80a0bb0c1bc30c4e6c9818f7329496601bde28a193\"" Feb 13 19:20:12.435107 containerd[1515]: time="2025-02-13T19:20:12.435058658Z" level=info msg="TearDown network for sandbox \"52abc7c7229296fa61042f80a0bb0c1bc30c4e6c9818f7329496601bde28a193\" successfully" Feb 13 19:20:12.440109 containerd[1515]: time="2025-02-13T19:20:12.440074416Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"52abc7c7229296fa61042f80a0bb0c1bc30c4e6c9818f7329496601bde28a193\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:20:12.440276 containerd[1515]: time="2025-02-13T19:20:12.440130114Z" level=info msg="RemovePodSandbox \"52abc7c7229296fa61042f80a0bb0c1bc30c4e6c9818f7329496601bde28a193\" returns successfully" Feb 13 19:20:12.440720 containerd[1515]: time="2025-02-13T19:20:12.440664386Z" level=info msg="StopPodSandbox for \"216ea206995c57165c7ec894ec603cac89445bad8cd05143541089c3e5d6db12\"" Feb 13 19:20:12.440945 containerd[1515]: time="2025-02-13T19:20:12.440880042Z" level=info msg="TearDown network for sandbox \"216ea206995c57165c7ec894ec603cac89445bad8cd05143541089c3e5d6db12\" successfully" Feb 13 19:20:12.440945 containerd[1515]: time="2025-02-13T19:20:12.440898988Z" level=info msg="StopPodSandbox for \"216ea206995c57165c7ec894ec603cac89445bad8cd05143541089c3e5d6db12\" returns successfully" Feb 13 19:20:12.441246 containerd[1515]: time="2025-02-13T19:20:12.441217404Z" level=info msg="RemovePodSandbox for \"216ea206995c57165c7ec894ec603cac89445bad8cd05143541089c3e5d6db12\"" Feb 13 19:20:12.441246 containerd[1515]: time="2025-02-13T19:20:12.441243774Z" level=info msg="Forcibly stopping sandbox \"216ea206995c57165c7ec894ec603cac89445bad8cd05143541089c3e5d6db12\"" Feb 13 19:20:12.441349 containerd[1515]: time="2025-02-13T19:20:12.441320082Z" level=info msg="TearDown network for sandbox \"216ea206995c57165c7ec894ec603cac89445bad8cd05143541089c3e5d6db12\" successfully" Feb 13 19:20:12.446430 containerd[1515]: time="2025-02-13T19:20:12.446393190Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"216ea206995c57165c7ec894ec603cac89445bad8cd05143541089c3e5d6db12\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:20:12.446551 containerd[1515]: time="2025-02-13T19:20:12.446452374Z" level=info msg="RemovePodSandbox \"216ea206995c57165c7ec894ec603cac89445bad8cd05143541089c3e5d6db12\" returns successfully" Feb 13 19:20:12.447054 containerd[1515]: time="2025-02-13T19:20:12.447023818Z" level=info msg="StopPodSandbox for \"fa92388357966f3e183c59506d6cfde3e38d36deede0f30821c07e22d4146a49\"" Feb 13 19:20:12.447284 containerd[1515]: time="2025-02-13T19:20:12.447262309Z" level=info msg="TearDown network for sandbox \"fa92388357966f3e183c59506d6cfde3e38d36deede0f30821c07e22d4146a49\" successfully" Feb 13 19:20:12.447284 containerd[1515]: time="2025-02-13T19:20:12.447278170Z" level=info msg="StopPodSandbox for \"fa92388357966f3e183c59506d6cfde3e38d36deede0f30821c07e22d4146a49\" returns successfully" Feb 13 19:20:12.447651 containerd[1515]: time="2025-02-13T19:20:12.447622334Z" level=info msg="RemovePodSandbox for \"fa92388357966f3e183c59506d6cfde3e38d36deede0f30821c07e22d4146a49\"" Feb 13 19:20:12.447651 containerd[1515]: time="2025-02-13T19:20:12.447659757Z" level=info msg="Forcibly stopping sandbox \"fa92388357966f3e183c59506d6cfde3e38d36deede0f30821c07e22d4146a49\"" Feb 13 19:20:12.447798 containerd[1515]: time="2025-02-13T19:20:12.447744090Z" level=info msg="TearDown network for sandbox \"fa92388357966f3e183c59506d6cfde3e38d36deede0f30821c07e22d4146a49\" successfully" Feb 13 19:20:12.452304 containerd[1515]: time="2025-02-13T19:20:12.452274720Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"fa92388357966f3e183c59506d6cfde3e38d36deede0f30821c07e22d4146a49\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:20:12.452404 containerd[1515]: time="2025-02-13T19:20:12.452322673Z" level=info msg="RemovePodSandbox \"fa92388357966f3e183c59506d6cfde3e38d36deede0f30821c07e22d4146a49\" returns successfully" Feb 13 19:20:12.452672 containerd[1515]: time="2025-02-13T19:20:12.452648412Z" level=info msg="StopPodSandbox for \"1d811fd694e600e47efc43338f2c97ee7c9f512c6106b43a1a049dbc81724231\"" Feb 13 19:20:12.453075 containerd[1515]: time="2025-02-13T19:20:12.452918524Z" level=info msg="TearDown network for sandbox \"1d811fd694e600e47efc43338f2c97ee7c9f512c6106b43a1a049dbc81724231\" successfully" Feb 13 19:20:12.453075 containerd[1515]: time="2025-02-13T19:20:12.452934885Z" level=info msg="StopPodSandbox for \"1d811fd694e600e47efc43338f2c97ee7c9f512c6106b43a1a049dbc81724231\" returns successfully" Feb 13 19:20:12.453467 containerd[1515]: time="2025-02-13T19:20:12.453415142Z" level=info msg="RemovePodSandbox for \"1d811fd694e600e47efc43338f2c97ee7c9f512c6106b43a1a049dbc81724231\"" Feb 13 19:20:12.453467 containerd[1515]: time="2025-02-13T19:20:12.453450641Z" level=info msg="Forcibly stopping sandbox \"1d811fd694e600e47efc43338f2c97ee7c9f512c6106b43a1a049dbc81724231\"" Feb 13 19:20:12.453966 containerd[1515]: time="2025-02-13T19:20:12.453547739Z" level=info msg="TearDown network for sandbox \"1d811fd694e600e47efc43338f2c97ee7c9f512c6106b43a1a049dbc81724231\" successfully" Feb 13 19:20:12.458368 containerd[1515]: time="2025-02-13T19:20:12.458320487Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"1d811fd694e600e47efc43338f2c97ee7c9f512c6106b43a1a049dbc81724231\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:20:12.458368 containerd[1515]: time="2025-02-13T19:20:12.458369472Z" level=info msg="RemovePodSandbox \"1d811fd694e600e47efc43338f2c97ee7c9f512c6106b43a1a049dbc81724231\" returns successfully" Feb 13 19:20:12.458879 containerd[1515]: time="2025-02-13T19:20:12.458829050Z" level=info msg="StopPodSandbox for \"216f8481f6c55f7a146dde6ed21a0f30b75cfe7059570683b2fb520eadf70207\"" Feb 13 19:20:12.458989 containerd[1515]: time="2025-02-13T19:20:12.458969972Z" level=info msg="TearDown network for sandbox \"216f8481f6c55f7a146dde6ed21a0f30b75cfe7059570683b2fb520eadf70207\" successfully" Feb 13 19:20:12.459030 containerd[1515]: time="2025-02-13T19:20:12.458983529Z" level=info msg="StopPodSandbox for \"216f8481f6c55f7a146dde6ed21a0f30b75cfe7059570683b2fb520eadf70207\" returns successfully" Feb 13 19:20:12.459412 containerd[1515]: time="2025-02-13T19:20:12.459383120Z" level=info msg="RemovePodSandbox for \"216f8481f6c55f7a146dde6ed21a0f30b75cfe7059570683b2fb520eadf70207\"" Feb 13 19:20:12.459472 containerd[1515]: time="2025-02-13T19:20:12.459412687Z" level=info msg="Forcibly stopping sandbox \"216f8481f6c55f7a146dde6ed21a0f30b75cfe7059570683b2fb520eadf70207\"" Feb 13 19:20:12.459543 containerd[1515]: time="2025-02-13T19:20:12.459496458Z" level=info msg="TearDown network for sandbox \"216f8481f6c55f7a146dde6ed21a0f30b75cfe7059570683b2fb520eadf70207\" successfully" Feb 13 19:20:12.464106 containerd[1515]: time="2025-02-13T19:20:12.464055705Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"216f8481f6c55f7a146dde6ed21a0f30b75cfe7059570683b2fb520eadf70207\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:20:12.464106 containerd[1515]: time="2025-02-13T19:20:12.464105070Z" level=info msg="RemovePodSandbox \"216f8481f6c55f7a146dde6ed21a0f30b75cfe7059570683b2fb520eadf70207\" returns successfully" Feb 13 19:20:12.464498 containerd[1515]: time="2025-02-13T19:20:12.464462982Z" level=info msg="StopPodSandbox for \"a072e0a626a78003d707d9120f0d4982782f1be11bc8ddc22d2f8a1dfb030680\"" Feb 13 19:20:12.464597 containerd[1515]: time="2025-02-13T19:20:12.464567242Z" level=info msg="TearDown network for sandbox \"a072e0a626a78003d707d9120f0d4982782f1be11bc8ddc22d2f8a1dfb030680\" successfully" Feb 13 19:20:12.464597 containerd[1515]: time="2025-02-13T19:20:12.464581931Z" level=info msg="StopPodSandbox for \"a072e0a626a78003d707d9120f0d4982782f1be11bc8ddc22d2f8a1dfb030680\" returns successfully" Feb 13 19:20:12.464863 containerd[1515]: time="2025-02-13T19:20:12.464829239Z" level=info msg="RemovePodSandbox for \"a072e0a626a78003d707d9120f0d4982782f1be11bc8ddc22d2f8a1dfb030680\"" Feb 13 19:20:12.464863 containerd[1515]: time="2025-02-13T19:20:12.464849788Z" level=info msg="Forcibly stopping sandbox \"a072e0a626a78003d707d9120f0d4982782f1be11bc8ddc22d2f8a1dfb030680\"" Feb 13 19:20:12.464955 containerd[1515]: time="2025-02-13T19:20:12.464920014Z" level=info msg="TearDown network for sandbox \"a072e0a626a78003d707d9120f0d4982782f1be11bc8ddc22d2f8a1dfb030680\" successfully" Feb 13 19:20:12.468997 containerd[1515]: time="2025-02-13T19:20:12.468922715Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"a072e0a626a78003d707d9120f0d4982782f1be11bc8ddc22d2f8a1dfb030680\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:20:12.468997 containerd[1515]: time="2025-02-13T19:20:12.468971980Z" level=info msg="RemovePodSandbox \"a072e0a626a78003d707d9120f0d4982782f1be11bc8ddc22d2f8a1dfb030680\" returns successfully" Feb 13 19:20:12.469299 containerd[1515]: time="2025-02-13T19:20:12.469241300Z" level=info msg="StopPodSandbox for \"41f8a27c833a329de6525bb7ea4ae60918c875126c9b0eb7529120508e6289e1\"" Feb 13 19:20:12.469424 containerd[1515]: time="2025-02-13T19:20:12.469325664Z" level=info msg="TearDown network for sandbox \"41f8a27c833a329de6525bb7ea4ae60918c875126c9b0eb7529120508e6289e1\" successfully" Feb 13 19:20:12.469424 containerd[1515]: time="2025-02-13T19:20:12.469337095Z" level=info msg="StopPodSandbox for \"41f8a27c833a329de6525bb7ea4ae60918c875126c9b0eb7529120508e6289e1\" returns successfully" Feb 13 19:20:12.469650 containerd[1515]: time="2025-02-13T19:20:12.469623399Z" level=info msg="RemovePodSandbox for \"41f8a27c833a329de6525bb7ea4ae60918c875126c9b0eb7529120508e6289e1\"" Feb 13 19:20:12.469720 containerd[1515]: time="2025-02-13T19:20:12.469651923Z" level=info msg="Forcibly stopping sandbox \"41f8a27c833a329de6525bb7ea4ae60918c875126c9b0eb7529120508e6289e1\"" Feb 13 19:20:12.469807 containerd[1515]: time="2025-02-13T19:20:12.469764612Z" level=info msg="TearDown network for sandbox \"41f8a27c833a329de6525bb7ea4ae60918c875126c9b0eb7529120508e6289e1\" successfully" Feb 13 19:20:12.473451 containerd[1515]: time="2025-02-13T19:20:12.473424911Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"41f8a27c833a329de6525bb7ea4ae60918c875126c9b0eb7529120508e6289e1\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:20:12.473513 containerd[1515]: time="2025-02-13T19:20:12.473464498Z" level=info msg="RemovePodSandbox \"41f8a27c833a329de6525bb7ea4ae60918c875126c9b0eb7529120508e6289e1\" returns successfully" Feb 13 19:20:12.473793 containerd[1515]: time="2025-02-13T19:20:12.473768094Z" level=info msg="StopPodSandbox for \"aa373a8af75a289d061969e5a22e647e175a6a451ba7c421e5c44fa363ec90f8\"" Feb 13 19:20:12.473880 containerd[1515]: time="2025-02-13T19:20:12.473851875Z" level=info msg="TearDown network for sandbox \"aa373a8af75a289d061969e5a22e647e175a6a451ba7c421e5c44fa363ec90f8\" successfully" Feb 13 19:20:12.473880 containerd[1515]: time="2025-02-13T19:20:12.473876854Z" level=info msg="StopPodSandbox for \"aa373a8af75a289d061969e5a22e647e175a6a451ba7c421e5c44fa363ec90f8\" returns successfully" Feb 13 19:20:12.474207 containerd[1515]: time="2025-02-13T19:20:12.474148128Z" level=info msg="RemovePodSandbox for \"aa373a8af75a289d061969e5a22e647e175a6a451ba7c421e5c44fa363ec90f8\"" Feb 13 19:20:12.474207 containerd[1515]: time="2025-02-13T19:20:12.474169890Z" level=info msg="Forcibly stopping sandbox \"aa373a8af75a289d061969e5a22e647e175a6a451ba7c421e5c44fa363ec90f8\"" Feb 13 19:20:12.474353 containerd[1515]: time="2025-02-13T19:20:12.474254283Z" level=info msg="TearDown network for sandbox \"aa373a8af75a289d061969e5a22e647e175a6a451ba7c421e5c44fa363ec90f8\" successfully" Feb 13 19:20:12.478462 containerd[1515]: time="2025-02-13T19:20:12.478405921Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"aa373a8af75a289d061969e5a22e647e175a6a451ba7c421e5c44fa363ec90f8\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:20:12.478520 containerd[1515]: time="2025-02-13T19:20:12.478465116Z" level=info msg="RemovePodSandbox \"aa373a8af75a289d061969e5a22e647e175a6a451ba7c421e5c44fa363ec90f8\" returns successfully" Feb 13 19:20:12.478792 containerd[1515]: time="2025-02-13T19:20:12.478770466Z" level=info msg="StopPodSandbox for \"4304f46dfb22ea4aced73143fa3f59c4ed426a670211c8559cc8ed545f1b3804\"" Feb 13 19:20:12.478894 containerd[1515]: time="2025-02-13T19:20:12.478876260Z" level=info msg="TearDown network for sandbox \"4304f46dfb22ea4aced73143fa3f59c4ed426a670211c8559cc8ed545f1b3804\" successfully" Feb 13 19:20:12.478939 containerd[1515]: time="2025-02-13T19:20:12.478892512Z" level=info msg="StopPodSandbox for \"4304f46dfb22ea4aced73143fa3f59c4ed426a670211c8559cc8ed545f1b3804\" returns successfully" Feb 13 19:20:12.479523 containerd[1515]: time="2025-02-13T19:20:12.479293145Z" level=info msg="RemovePodSandbox for \"4304f46dfb22ea4aced73143fa3f59c4ed426a670211c8559cc8ed545f1b3804\"" Feb 13 19:20:12.479523 containerd[1515]: time="2025-02-13T19:20:12.479332471Z" level=info msg="Forcibly stopping sandbox \"4304f46dfb22ea4aced73143fa3f59c4ed426a670211c8559cc8ed545f1b3804\"" Feb 13 19:20:12.479523 containerd[1515]: time="2025-02-13T19:20:12.479407807Z" level=info msg="TearDown network for sandbox \"4304f46dfb22ea4aced73143fa3f59c4ed426a670211c8559cc8ed545f1b3804\" successfully" Feb 13 19:20:12.483243 containerd[1515]: time="2025-02-13T19:20:12.483205973Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"4304f46dfb22ea4aced73143fa3f59c4ed426a670211c8559cc8ed545f1b3804\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:20:12.483339 containerd[1515]: time="2025-02-13T19:20:12.483248265Z" level=info msg="RemovePodSandbox \"4304f46dfb22ea4aced73143fa3f59c4ed426a670211c8559cc8ed545f1b3804\" returns successfully" Feb 13 19:20:12.483670 containerd[1515]: time="2025-02-13T19:20:12.483640863Z" level=info msg="StopPodSandbox for \"e5e6531fa4d44fae3b14be640ca5608599c77b930300d189004cbcc636c08e1b\"" Feb 13 19:20:12.483758 containerd[1515]: time="2025-02-13T19:20:12.483737630Z" level=info msg="TearDown network for sandbox \"e5e6531fa4d44fae3b14be640ca5608599c77b930300d189004cbcc636c08e1b\" successfully" Feb 13 19:20:12.483802 containerd[1515]: time="2025-02-13T19:20:12.483756015Z" level=info msg="StopPodSandbox for \"e5e6531fa4d44fae3b14be640ca5608599c77b930300d189004cbcc636c08e1b\" returns successfully" Feb 13 19:20:12.484085 containerd[1515]: time="2025-02-13T19:20:12.484061335Z" level=info msg="RemovePodSandbox for \"e5e6531fa4d44fae3b14be640ca5608599c77b930300d189004cbcc636c08e1b\"" Feb 13 19:20:12.484085 containerd[1515]: time="2025-02-13T19:20:12.484081423Z" level=info msg="Forcibly stopping sandbox \"e5e6531fa4d44fae3b14be640ca5608599c77b930300d189004cbcc636c08e1b\"" Feb 13 19:20:12.484171 containerd[1515]: time="2025-02-13T19:20:12.484144856Z" level=info msg="TearDown network for sandbox \"e5e6531fa4d44fae3b14be640ca5608599c77b930300d189004cbcc636c08e1b\" successfully" Feb 13 19:20:12.489984 containerd[1515]: time="2025-02-13T19:20:12.489902377Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"e5e6531fa4d44fae3b14be640ca5608599c77b930300d189004cbcc636c08e1b\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:20:12.489984 containerd[1515]: time="2025-02-13T19:20:12.489964036Z" level=info msg="RemovePodSandbox \"e5e6531fa4d44fae3b14be640ca5608599c77b930300d189004cbcc636c08e1b\" returns successfully" Feb 13 19:20:12.490514 containerd[1515]: time="2025-02-13T19:20:12.490469131Z" level=info msg="StopPodSandbox for \"334cbee9ef51499256254b9201c8014a8e1e8be612d9ae143735bc19f2a5f6a4\"" Feb 13 19:20:12.490653 containerd[1515]: time="2025-02-13T19:20:12.490630142Z" level=info msg="TearDown network for sandbox \"334cbee9ef51499256254b9201c8014a8e1e8be612d9ae143735bc19f2a5f6a4\" successfully" Feb 13 19:20:12.490653 containerd[1515]: time="2025-02-13T19:20:12.490649280Z" level=info msg="StopPodSandbox for \"334cbee9ef51499256254b9201c8014a8e1e8be612d9ae143735bc19f2a5f6a4\" returns successfully" Feb 13 19:20:12.491020 containerd[1515]: time="2025-02-13T19:20:12.490992593Z" level=info msg="RemovePodSandbox for \"334cbee9ef51499256254b9201c8014a8e1e8be612d9ae143735bc19f2a5f6a4\"" Feb 13 19:20:12.491070 containerd[1515]: time="2025-02-13T19:20:12.491019885Z" level=info msg="Forcibly stopping sandbox \"334cbee9ef51499256254b9201c8014a8e1e8be612d9ae143735bc19f2a5f6a4\"" Feb 13 19:20:12.491147 containerd[1515]: time="2025-02-13T19:20:12.491097565Z" level=info msg="TearDown network for sandbox \"334cbee9ef51499256254b9201c8014a8e1e8be612d9ae143735bc19f2a5f6a4\" successfully" Feb 13 19:20:12.495348 containerd[1515]: time="2025-02-13T19:20:12.495320802Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"334cbee9ef51499256254b9201c8014a8e1e8be612d9ae143735bc19f2a5f6a4\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:20:12.495420 containerd[1515]: time="2025-02-13T19:20:12.495372191Z" level=info msg="RemovePodSandbox \"334cbee9ef51499256254b9201c8014a8e1e8be612d9ae143735bc19f2a5f6a4\" returns successfully" Feb 13 19:20:12.495934 containerd[1515]: time="2025-02-13T19:20:12.495765190Z" level=info msg="StopPodSandbox for \"9cfc4e8026fe9d3540dd0b1d209b241ede8d2453bfec90606b125f81c566e640\"" Feb 13 19:20:12.495934 containerd[1515]: time="2025-02-13T19:20:12.495870794Z" level=info msg="TearDown network for sandbox \"9cfc4e8026fe9d3540dd0b1d209b241ede8d2453bfec90606b125f81c566e640\" successfully" Feb 13 19:20:12.495934 containerd[1515]: time="2025-02-13T19:20:12.495880714Z" level=info msg="StopPodSandbox for \"9cfc4e8026fe9d3540dd0b1d209b241ede8d2453bfec90606b125f81c566e640\" returns successfully" Feb 13 19:20:12.496239 containerd[1515]: time="2025-02-13T19:20:12.496208176Z" level=info msg="RemovePodSandbox for \"9cfc4e8026fe9d3540dd0b1d209b241ede8d2453bfec90606b125f81c566e640\"" Feb 13 19:20:12.496272 containerd[1515]: time="2025-02-13T19:20:12.496246370Z" level=info msg="Forcibly stopping sandbox \"9cfc4e8026fe9d3540dd0b1d209b241ede8d2453bfec90606b125f81c566e640\"" Feb 13 19:20:12.496400 containerd[1515]: time="2025-02-13T19:20:12.496349990Z" level=info msg="TearDown network for sandbox \"9cfc4e8026fe9d3540dd0b1d209b241ede8d2453bfec90606b125f81c566e640\" successfully" Feb 13 19:20:12.500754 containerd[1515]: time="2025-02-13T19:20:12.500693519Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"9cfc4e8026fe9d3540dd0b1d209b241ede8d2453bfec90606b125f81c566e640\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:20:12.500754 containerd[1515]: time="2025-02-13T19:20:12.500748315Z" level=info msg="RemovePodSandbox \"9cfc4e8026fe9d3540dd0b1d209b241ede8d2453bfec90606b125f81c566e640\" returns successfully" Feb 13 19:20:12.501176 containerd[1515]: time="2025-02-13T19:20:12.501146124Z" level=info msg="StopPodSandbox for \"45253c1f9735dfb2a3494283160983fb9ed43ea60532211ecb2b069637dde093\"" Feb 13 19:20:12.501287 containerd[1515]: time="2025-02-13T19:20:12.501256186Z" level=info msg="TearDown network for sandbox \"45253c1f9735dfb2a3494283160983fb9ed43ea60532211ecb2b069637dde093\" successfully" Feb 13 19:20:12.501287 containerd[1515]: time="2025-02-13T19:20:12.501274031Z" level=info msg="StopPodSandbox for \"45253c1f9735dfb2a3494283160983fb9ed43ea60532211ecb2b069637dde093\" returns successfully" Feb 13 19:20:12.501540 containerd[1515]: time="2025-02-13T19:20:12.501498243Z" level=info msg="RemovePodSandbox for \"45253c1f9735dfb2a3494283160983fb9ed43ea60532211ecb2b069637dde093\"" Feb 13 19:20:12.501540 containerd[1515]: time="2025-02-13T19:20:12.501517872Z" level=info msg="Forcibly stopping sandbox \"45253c1f9735dfb2a3494283160983fb9ed43ea60532211ecb2b069637dde093\"" Feb 13 19:20:12.501621 containerd[1515]: time="2025-02-13T19:20:12.501576235Z" level=info msg="TearDown network for sandbox \"45253c1f9735dfb2a3494283160983fb9ed43ea60532211ecb2b069637dde093\" successfully" Feb 13 19:20:12.505400 containerd[1515]: time="2025-02-13T19:20:12.505362598Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"45253c1f9735dfb2a3494283160983fb9ed43ea60532211ecb2b069637dde093\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:20:12.505446 containerd[1515]: time="2025-02-13T19:20:12.505402705Z" level=info msg="RemovePodSandbox \"45253c1f9735dfb2a3494283160983fb9ed43ea60532211ecb2b069637dde093\" returns successfully" Feb 13 19:20:12.505654 containerd[1515]: time="2025-02-13T19:20:12.505618021Z" level=info msg="StopPodSandbox for \"f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7\"" Feb 13 19:20:12.505732 containerd[1515]: time="2025-02-13T19:20:12.505696332Z" level=info msg="TearDown network for sandbox \"f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7\" successfully" Feb 13 19:20:12.505732 containerd[1515]: time="2025-02-13T19:20:12.505712834Z" level=info msg="StopPodSandbox for \"f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7\" returns successfully" Feb 13 19:20:12.506023 containerd[1515]: time="2025-02-13T19:20:12.505988336Z" level=info msg="RemovePodSandbox for \"f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7\"" Feb 13 19:20:12.506023 containerd[1515]: time="2025-02-13T19:20:12.506016280Z" level=info msg="Forcibly stopping sandbox \"f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7\"" Feb 13 19:20:12.506175 containerd[1515]: time="2025-02-13T19:20:12.506088499Z" level=info msg="TearDown network for sandbox \"f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7\" successfully" Feb 13 19:20:12.510213 containerd[1515]: time="2025-02-13T19:20:12.510112743Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:20:12.510213 containerd[1515]: time="2025-02-13T19:20:12.510150706Z" level=info msg="RemovePodSandbox \"f9e67eaf9fb0c85af5635a6d52b58b6ca19e3ce4c14ab158565a18ea0bd422e7\" returns successfully" Feb 13 19:20:12.510434 containerd[1515]: time="2025-02-13T19:20:12.510407010Z" level=info msg="StopPodSandbox for \"a20bc39d31c8f248a56fc8b89d2d3ce62836f574c54b75502431e23bb908b697\"" Feb 13 19:20:12.510495 containerd[1515]: time="2025-02-13T19:20:12.510484861Z" level=info msg="TearDown network for sandbox \"a20bc39d31c8f248a56fc8b89d2d3ce62836f574c54b75502431e23bb908b697\" successfully" Feb 13 19:20:12.510914 containerd[1515]: time="2025-02-13T19:20:12.510494339Z" level=info msg="StopPodSandbox for \"a20bc39d31c8f248a56fc8b89d2d3ce62836f574c54b75502431e23bb908b697\" returns successfully" Feb 13 19:20:12.515266 containerd[1515]: time="2025-02-13T19:20:12.513447483Z" level=info msg="RemovePodSandbox for \"a20bc39d31c8f248a56fc8b89d2d3ce62836f574c54b75502431e23bb908b697\"" Feb 13 19:20:12.515266 containerd[1515]: time="2025-02-13T19:20:12.513470477Z" level=info msg="Forcibly stopping sandbox \"a20bc39d31c8f248a56fc8b89d2d3ce62836f574c54b75502431e23bb908b697\"" Feb 13 19:20:12.515266 containerd[1515]: time="2025-02-13T19:20:12.513554991Z" level=info msg="TearDown network for sandbox \"a20bc39d31c8f248a56fc8b89d2d3ce62836f574c54b75502431e23bb908b697\" successfully" Feb 13 19:20:12.518021 containerd[1515]: time="2025-02-13T19:20:12.517991400Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"a20bc39d31c8f248a56fc8b89d2d3ce62836f574c54b75502431e23bb908b697\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:20:12.518090 containerd[1515]: time="2025-02-13T19:20:12.518034042Z" level=info msg="RemovePodSandbox \"a20bc39d31c8f248a56fc8b89d2d3ce62836f574c54b75502431e23bb908b697\" returns successfully" Feb 13 19:20:12.518437 containerd[1515]: time="2025-02-13T19:20:12.518410920Z" level=info msg="StopPodSandbox for \"28a752266e9ffa0b0f35649f5eba403b238d216dc8ce20a8fe3622cc28799217\"" Feb 13 19:20:12.518499 containerd[1515]: time="2025-02-13T19:20:12.518486606Z" level=info msg="TearDown network for sandbox \"28a752266e9ffa0b0f35649f5eba403b238d216dc8ce20a8fe3622cc28799217\" successfully" Feb 13 19:20:12.518499 containerd[1515]: time="2025-02-13T19:20:12.518496235Z" level=info msg="StopPodSandbox for \"28a752266e9ffa0b0f35649f5eba403b238d216dc8ce20a8fe3622cc28799217\" returns successfully" Feb 13 19:20:12.518754 containerd[1515]: time="2025-02-13T19:20:12.518732110Z" level=info msg="RemovePodSandbox for \"28a752266e9ffa0b0f35649f5eba403b238d216dc8ce20a8fe3622cc28799217\"" Feb 13 19:20:12.518864 containerd[1515]: time="2025-02-13T19:20:12.518825711Z" level=info msg="Forcibly stopping sandbox \"28a752266e9ffa0b0f35649f5eba403b238d216dc8ce20a8fe3622cc28799217\"" Feb 13 19:20:12.518941 containerd[1515]: time="2025-02-13T19:20:12.518903431Z" level=info msg="TearDown network for sandbox \"28a752266e9ffa0b0f35649f5eba403b238d216dc8ce20a8fe3622cc28799217\" successfully" Feb 13 19:20:12.522629 containerd[1515]: time="2025-02-13T19:20:12.522583359Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"28a752266e9ffa0b0f35649f5eba403b238d216dc8ce20a8fe3622cc28799217\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:20:12.522629 containerd[1515]: time="2025-02-13T19:20:12.522619790Z" level=info msg="RemovePodSandbox \"28a752266e9ffa0b0f35649f5eba403b238d216dc8ce20a8fe3622cc28799217\" returns successfully" Feb 13 19:20:12.523016 containerd[1515]: time="2025-02-13T19:20:12.522982159Z" level=info msg="StopPodSandbox for \"ee0d44b3886e78bb5043872859e92bfd2620af442c433a402780bc0453826443\"" Feb 13 19:20:12.523132 containerd[1515]: time="2025-02-13T19:20:12.523080630Z" level=info msg="TearDown network for sandbox \"ee0d44b3886e78bb5043872859e92bfd2620af442c433a402780bc0453826443\" successfully" Feb 13 19:20:12.523132 containerd[1515]: time="2025-02-13T19:20:12.523119354Z" level=info msg="StopPodSandbox for \"ee0d44b3886e78bb5043872859e92bfd2620af442c433a402780bc0453826443\" returns successfully" Feb 13 19:20:12.523391 containerd[1515]: time="2025-02-13T19:20:12.523367604Z" level=info msg="RemovePodSandbox for \"ee0d44b3886e78bb5043872859e92bfd2620af442c433a402780bc0453826443\"" Feb 13 19:20:12.523391 containerd[1515]: time="2025-02-13T19:20:12.523387903Z" level=info msg="Forcibly stopping sandbox \"ee0d44b3886e78bb5043872859e92bfd2620af442c433a402780bc0453826443\"" Feb 13 19:20:12.523490 containerd[1515]: time="2025-02-13T19:20:12.523451957Z" level=info msg="TearDown network for sandbox \"ee0d44b3886e78bb5043872859e92bfd2620af442c433a402780bc0453826443\" successfully" Feb 13 19:20:12.527223 containerd[1515]: time="2025-02-13T19:20:12.527170810Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"ee0d44b3886e78bb5043872859e92bfd2620af442c433a402780bc0453826443\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:20:12.527223 containerd[1515]: time="2025-02-13T19:20:12.527222549Z" level=info msg="RemovePodSandbox \"ee0d44b3886e78bb5043872859e92bfd2620af442c433a402780bc0453826443\" returns successfully" Feb 13 19:20:12.527558 containerd[1515]: time="2025-02-13T19:20:12.527515465Z" level=info msg="StopPodSandbox for \"b9281cb17530bfce23ceee39b94a1f286e88137aa4b1dc73b5faa8a4e7379e99\"" Feb 13 19:20:12.527629 containerd[1515]: time="2025-02-13T19:20:12.527604667Z" level=info msg="TearDown network for sandbox \"b9281cb17530bfce23ceee39b94a1f286e88137aa4b1dc73b5faa8a4e7379e99\" successfully" Feb 13 19:20:12.527629 containerd[1515]: time="2025-02-13T19:20:12.527615358Z" level=info msg="StopPodSandbox for \"b9281cb17530bfce23ceee39b94a1f286e88137aa4b1dc73b5faa8a4e7379e99\" returns successfully" Feb 13 19:20:12.527919 containerd[1515]: time="2025-02-13T19:20:12.527888936Z" level=info msg="RemovePodSandbox for \"b9281cb17530bfce23ceee39b94a1f286e88137aa4b1dc73b5faa8a4e7379e99\"" Feb 13 19:20:12.527978 containerd[1515]: time="2025-02-13T19:20:12.527920047Z" level=info msg="Forcibly stopping sandbox \"b9281cb17530bfce23ceee39b94a1f286e88137aa4b1dc73b5faa8a4e7379e99\"" Feb 13 19:20:12.528046 containerd[1515]: time="2025-02-13T19:20:12.528006764Z" level=info msg="TearDown network for sandbox \"b9281cb17530bfce23ceee39b94a1f286e88137aa4b1dc73b5faa8a4e7379e99\" successfully" Feb 13 19:20:12.532619 containerd[1515]: time="2025-02-13T19:20:12.532571490Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"b9281cb17530bfce23ceee39b94a1f286e88137aa4b1dc73b5faa8a4e7379e99\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:20:12.532679 containerd[1515]: time="2025-02-13T19:20:12.532629031Z" level=info msg="RemovePodSandbox \"b9281cb17530bfce23ceee39b94a1f286e88137aa4b1dc73b5faa8a4e7379e99\" returns successfully" Feb 13 19:20:12.532960 containerd[1515]: time="2025-02-13T19:20:12.532927588Z" level=info msg="StopPodSandbox for \"535ac305f91d459b29fcac754f456e877bb89217b5f89ed4a0f736398ab4e183\"" Feb 13 19:20:12.533047 containerd[1515]: time="2025-02-13T19:20:12.533028604Z" level=info msg="TearDown network for sandbox \"535ac305f91d459b29fcac754f456e877bb89217b5f89ed4a0f736398ab4e183\" successfully" Feb 13 19:20:12.533072 containerd[1515]: time="2025-02-13T19:20:12.533045677Z" level=info msg="StopPodSandbox for \"535ac305f91d459b29fcac754f456e877bb89217b5f89ed4a0f736398ab4e183\" returns successfully" Feb 13 19:20:12.533320 containerd[1515]: time="2025-02-13T19:20:12.533292784Z" level=info msg="RemovePodSandbox for \"535ac305f91d459b29fcac754f456e877bb89217b5f89ed4a0f736398ab4e183\"" Feb 13 19:20:12.533320 containerd[1515]: time="2025-02-13T19:20:12.533317541Z" level=info msg="Forcibly stopping sandbox \"535ac305f91d459b29fcac754f456e877bb89217b5f89ed4a0f736398ab4e183\"" Feb 13 19:20:12.533440 containerd[1515]: time="2025-02-13T19:20:12.533397847Z" level=info msg="TearDown network for sandbox \"535ac305f91d459b29fcac754f456e877bb89217b5f89ed4a0f736398ab4e183\" successfully" Feb 13 19:20:12.537148 containerd[1515]: time="2025-02-13T19:20:12.537122551Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"535ac305f91d459b29fcac754f456e877bb89217b5f89ed4a0f736398ab4e183\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Feb 13 19:20:12.537237 containerd[1515]: time="2025-02-13T19:20:12.537163349Z" level=info msg="RemovePodSandbox \"535ac305f91d459b29fcac754f456e877bb89217b5f89ed4a0f736398ab4e183\" returns successfully" Feb 13 19:20:14.393407 kubelet[2630]: E0213 19:20:14.393374 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:20:17.278284 systemd[1]: Started sshd@17-10.0.0.46:22-10.0.0.1:53576.service - OpenSSH per-connection server daemon (10.0.0.1:53576). Feb 13 19:20:17.337063 sshd[6220]: Accepted publickey for core from 10.0.0.1 port 53576 ssh2: RSA SHA256:xgLbxCKtIvCmXzj7C6d4ih050Hrbkh61XCRduaX62E8 Feb 13 19:20:17.338970 sshd-session[6220]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:20:17.343704 systemd-logind[1495]: New session 18 of user core. Feb 13 19:20:17.349404 systemd[1]: Started session-18.scope - Session 18 of User core. Feb 13 19:20:17.463435 sshd[6222]: Connection closed by 10.0.0.1 port 53576 Feb 13 19:20:17.463804 sshd-session[6220]: pam_unix(sshd:session): session closed for user core Feb 13 19:20:17.467624 systemd[1]: sshd@17-10.0.0.46:22-10.0.0.1:53576.service: Deactivated successfully. Feb 13 19:20:17.469937 systemd[1]: session-18.scope: Deactivated successfully. Feb 13 19:20:17.470731 systemd-logind[1495]: Session 18 logged out. Waiting for processes to exit. Feb 13 19:20:17.471738 systemd-logind[1495]: Removed session 18. Feb 13 19:20:21.657938 kubelet[2630]: E0213 19:20:21.657892 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:20:22.480123 systemd[1]: Started sshd@18-10.0.0.46:22-10.0.0.1:53582.service - OpenSSH per-connection server daemon (10.0.0.1:53582). Feb 13 19:20:22.524037 sshd[6237]: Accepted publickey for core from 10.0.0.1 port 53582 ssh2: RSA SHA256:xgLbxCKtIvCmXzj7C6d4ih050Hrbkh61XCRduaX62E8 Feb 13 19:20:22.525451 sshd-session[6237]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:20:22.529428 systemd-logind[1495]: New session 19 of user core. Feb 13 19:20:22.543330 systemd[1]: Started session-19.scope - Session 19 of User core. Feb 13 19:20:22.653859 sshd[6239]: Connection closed by 10.0.0.1 port 53582 Feb 13 19:20:22.656554 sshd-session[6237]: pam_unix(sshd:session): session closed for user core Feb 13 19:20:22.670169 systemd[1]: sshd@18-10.0.0.46:22-10.0.0.1:53582.service: Deactivated successfully. Feb 13 19:20:22.672283 systemd[1]: session-19.scope: Deactivated successfully. Feb 13 19:20:22.674580 systemd-logind[1495]: Session 19 logged out. Waiting for processes to exit. Feb 13 19:20:22.681603 systemd[1]: Started sshd@19-10.0.0.46:22-10.0.0.1:53598.service - OpenSSH per-connection server daemon (10.0.0.1:53598). Feb 13 19:20:22.682767 systemd-logind[1495]: Removed session 19. Feb 13 19:20:22.726646 sshd[6251]: Accepted publickey for core from 10.0.0.1 port 53598 ssh2: RSA SHA256:xgLbxCKtIvCmXzj7C6d4ih050Hrbkh61XCRduaX62E8 Feb 13 19:20:22.728769 sshd-session[6251]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:20:22.735033 systemd-logind[1495]: New session 20 of user core. Feb 13 19:20:22.740456 systemd[1]: Started session-20.scope - Session 20 of User core. Feb 13 19:20:23.037637 sshd[6254]: Connection closed by 10.0.0.1 port 53598 Feb 13 19:20:23.038050 sshd-session[6251]: pam_unix(sshd:session): session closed for user core Feb 13 19:20:23.048452 systemd[1]: sshd@19-10.0.0.46:22-10.0.0.1:53598.service: Deactivated successfully. Feb 13 19:20:23.050727 systemd[1]: session-20.scope: Deactivated successfully. Feb 13 19:20:23.051490 systemd-logind[1495]: Session 20 logged out. Waiting for processes to exit. Feb 13 19:20:23.067619 systemd[1]: Started sshd@20-10.0.0.46:22-10.0.0.1:53612.service - OpenSSH per-connection server daemon (10.0.0.1:53612). Feb 13 19:20:23.068451 systemd-logind[1495]: Removed session 20. Feb 13 19:20:23.119336 sshd[6265]: Accepted publickey for core from 10.0.0.1 port 53612 ssh2: RSA SHA256:xgLbxCKtIvCmXzj7C6d4ih050Hrbkh61XCRduaX62E8 Feb 13 19:20:23.120885 sshd-session[6265]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:20:23.125653 systemd-logind[1495]: New session 21 of user core. Feb 13 19:20:23.136329 systemd[1]: Started session-21.scope - Session 21 of User core. Feb 13 19:20:25.508039 sshd[6268]: Connection closed by 10.0.0.1 port 53612 Feb 13 19:20:25.508645 sshd-session[6265]: pam_unix(sshd:session): session closed for user core Feb 13 19:20:25.519455 systemd[1]: sshd@20-10.0.0.46:22-10.0.0.1:53612.service: Deactivated successfully. Feb 13 19:20:25.521312 systemd[1]: session-21.scope: Deactivated successfully. Feb 13 19:20:25.521540 systemd[1]: session-21.scope: Consumed 596ms CPU time, 73.9M memory peak. Feb 13 19:20:25.522149 systemd-logind[1495]: Session 21 logged out. Waiting for processes to exit. Feb 13 19:20:25.527441 systemd[1]: Started sshd@21-10.0.0.46:22-10.0.0.1:53624.service - OpenSSH per-connection server daemon (10.0.0.1:53624). Feb 13 19:20:25.528378 systemd-logind[1495]: Removed session 21. Feb 13 19:20:25.567921 sshd[6286]: Accepted publickey for core from 10.0.0.1 port 53624 ssh2: RSA SHA256:xgLbxCKtIvCmXzj7C6d4ih050Hrbkh61XCRduaX62E8 Feb 13 19:20:25.569585 sshd-session[6286]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:20:25.574285 systemd-logind[1495]: New session 22 of user core. Feb 13 19:20:25.580360 systemd[1]: Started session-22.scope - Session 22 of User core. Feb 13 19:20:25.799212 sshd[6289]: Connection closed by 10.0.0.1 port 53624 Feb 13 19:20:25.799799 sshd-session[6286]: pam_unix(sshd:session): session closed for user core Feb 13 19:20:25.811517 systemd[1]: sshd@21-10.0.0.46:22-10.0.0.1:53624.service: Deactivated successfully. Feb 13 19:20:25.813666 systemd[1]: session-22.scope: Deactivated successfully. Feb 13 19:20:25.815311 systemd-logind[1495]: Session 22 logged out. Waiting for processes to exit. Feb 13 19:20:25.827524 systemd[1]: Started sshd@22-10.0.0.46:22-10.0.0.1:53638.service - OpenSSH per-connection server daemon (10.0.0.1:53638). Feb 13 19:20:25.828416 systemd-logind[1495]: Removed session 22. Feb 13 19:20:25.866603 sshd[6299]: Accepted publickey for core from 10.0.0.1 port 53638 ssh2: RSA SHA256:xgLbxCKtIvCmXzj7C6d4ih050Hrbkh61XCRduaX62E8 Feb 13 19:20:25.868001 sshd-session[6299]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:20:25.872089 systemd-logind[1495]: New session 23 of user core. Feb 13 19:20:25.878302 systemd[1]: Started session-23.scope - Session 23 of User core. Feb 13 19:20:25.988303 sshd[6302]: Connection closed by 10.0.0.1 port 53638 Feb 13 19:20:25.989070 sshd-session[6299]: pam_unix(sshd:session): session closed for user core Feb 13 19:20:25.992373 systemd[1]: sshd@22-10.0.0.46:22-10.0.0.1:53638.service: Deactivated successfully. Feb 13 19:20:25.994394 systemd[1]: session-23.scope: Deactivated successfully. Feb 13 19:20:25.996394 systemd-logind[1495]: Session 23 logged out. Waiting for processes to exit. Feb 13 19:20:25.997280 systemd-logind[1495]: Removed session 23. Feb 13 19:20:30.658480 kubelet[2630]: E0213 19:20:30.658409 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:20:31.009500 systemd[1]: Started sshd@23-10.0.0.46:22-10.0.0.1:53906.service - OpenSSH per-connection server daemon (10.0.0.1:53906). Feb 13 19:20:31.066473 sshd[6323]: Accepted publickey for core from 10.0.0.1 port 53906 ssh2: RSA SHA256:xgLbxCKtIvCmXzj7C6d4ih050Hrbkh61XCRduaX62E8 Feb 13 19:20:31.067915 sshd-session[6323]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:20:31.072419 systemd-logind[1495]: New session 24 of user core. Feb 13 19:20:31.084344 systemd[1]: Started session-24.scope - Session 24 of User core. Feb 13 19:20:31.199993 sshd[6325]: Connection closed by 10.0.0.1 port 53906 Feb 13 19:20:31.200351 sshd-session[6323]: pam_unix(sshd:session): session closed for user core Feb 13 19:20:31.204313 systemd[1]: sshd@23-10.0.0.46:22-10.0.0.1:53906.service: Deactivated successfully. Feb 13 19:20:31.206441 systemd[1]: session-24.scope: Deactivated successfully. Feb 13 19:20:31.207082 systemd-logind[1495]: Session 24 logged out. Waiting for processes to exit. Feb 13 19:20:31.207940 systemd-logind[1495]: Removed session 24. Feb 13 19:20:36.219662 systemd[1]: Started sshd@24-10.0.0.46:22-10.0.0.1:53910.service - OpenSSH per-connection server daemon (10.0.0.1:53910). Feb 13 19:20:36.263053 sshd[6341]: Accepted publickey for core from 10.0.0.1 port 53910 ssh2: RSA SHA256:xgLbxCKtIvCmXzj7C6d4ih050Hrbkh61XCRduaX62E8 Feb 13 19:20:36.264673 sshd-session[6341]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:20:36.268960 systemd-logind[1495]: New session 25 of user core. Feb 13 19:20:36.275311 systemd[1]: Started session-25.scope - Session 25 of User core. Feb 13 19:20:36.385957 sshd[6343]: Connection closed by 10.0.0.1 port 53910 Feb 13 19:20:36.386453 sshd-session[6341]: pam_unix(sshd:session): session closed for user core Feb 13 19:20:36.391514 systemd[1]: sshd@24-10.0.0.46:22-10.0.0.1:53910.service: Deactivated successfully. Feb 13 19:20:36.393831 systemd[1]: session-25.scope: Deactivated successfully. Feb 13 19:20:36.394723 systemd-logind[1495]: Session 25 logged out. Waiting for processes to exit. Feb 13 19:20:36.396021 systemd-logind[1495]: Removed session 25. Feb 13 19:20:41.398395 systemd[1]: Started sshd@25-10.0.0.46:22-10.0.0.1:36880.service - OpenSSH per-connection server daemon (10.0.0.1:36880). Feb 13 19:20:41.439559 sshd[6376]: Accepted publickey for core from 10.0.0.1 port 36880 ssh2: RSA SHA256:xgLbxCKtIvCmXzj7C6d4ih050Hrbkh61XCRduaX62E8 Feb 13 19:20:41.440941 sshd-session[6376]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:20:41.445163 systemd-logind[1495]: New session 26 of user core. Feb 13 19:20:41.458361 systemd[1]: Started session-26.scope - Session 26 of User core. Feb 13 19:20:41.571169 sshd[6378]: Connection closed by 10.0.0.1 port 36880 Feb 13 19:20:41.571600 sshd-session[6376]: pam_unix(sshd:session): session closed for user core Feb 13 19:20:41.576739 systemd[1]: sshd@25-10.0.0.46:22-10.0.0.1:36880.service: Deactivated successfully. Feb 13 19:20:41.579547 systemd[1]: session-26.scope: Deactivated successfully. Feb 13 19:20:41.580596 systemd-logind[1495]: Session 26 logged out. Waiting for processes to exit. Feb 13 19:20:41.581793 systemd-logind[1495]: Removed session 26. Feb 13 19:20:42.657829 kubelet[2630]: E0213 19:20:42.657794 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:20:43.658082 kubelet[2630]: E0213 19:20:43.658030 2630 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Feb 13 19:20:46.584913 systemd[1]: Started sshd@26-10.0.0.46:22-10.0.0.1:42512.service - OpenSSH per-connection server daemon (10.0.0.1:42512). Feb 13 19:20:46.631941 sshd[6413]: Accepted publickey for core from 10.0.0.1 port 42512 ssh2: RSA SHA256:xgLbxCKtIvCmXzj7C6d4ih050Hrbkh61XCRduaX62E8 Feb 13 19:20:46.633476 sshd-session[6413]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Feb 13 19:20:46.637790 systemd-logind[1495]: New session 27 of user core. Feb 13 19:20:46.647366 systemd[1]: Started session-27.scope - Session 27 of User core. Feb 13 19:20:46.764928 sshd[6415]: Connection closed by 10.0.0.1 port 42512 Feb 13 19:20:46.765333 sshd-session[6413]: pam_unix(sshd:session): session closed for user core Feb 13 19:20:46.769626 systemd[1]: sshd@26-10.0.0.46:22-10.0.0.1:42512.service: Deactivated successfully. Feb 13 19:20:46.771874 systemd[1]: session-27.scope: Deactivated successfully. Feb 13 19:20:46.772594 systemd-logind[1495]: Session 27 logged out. Waiting for processes to exit. Feb 13 19:20:46.773495 systemd-logind[1495]: Removed session 27.