Mar 17 17:41:19.874197 kernel: Linux version 6.6.83-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p1) 13.3.1 20240614, GNU ld (Gentoo 2.42 p6) 2.42.0) #1 SMP PREEMPT_DYNAMIC Mon Mar 17 16:07:40 -00 2025 Mar 17 17:41:19.874218 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=d4b838cd9a6f58e8c4a6b615c32b0b28ee0df1660e34033a8fbd0429c6de5fd0 Mar 17 17:41:19.874229 kernel: BIOS-provided physical RAM map: Mar 17 17:41:19.874235 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable Mar 17 17:41:19.874241 kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000007fffff] usable Mar 17 17:41:19.874247 kernel: BIOS-e820: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS Mar 17 17:41:19.874254 kernel: BIOS-e820: [mem 0x0000000000808000-0x000000000080afff] usable Mar 17 17:41:19.874261 kernel: BIOS-e820: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS Mar 17 17:41:19.874267 kernel: BIOS-e820: [mem 0x000000000080c000-0x0000000000810fff] usable Mar 17 17:41:19.874273 kernel: BIOS-e820: [mem 0x0000000000811000-0x00000000008fffff] ACPI NVS Mar 17 17:41:19.874282 kernel: BIOS-e820: [mem 0x0000000000900000-0x000000009bd3efff] usable Mar 17 17:41:19.874288 kernel: BIOS-e820: [mem 0x000000009bd3f000-0x000000009bdfffff] reserved Mar 17 17:41:19.874294 kernel: BIOS-e820: [mem 0x000000009be00000-0x000000009c8ecfff] usable Mar 17 17:41:19.874300 kernel: BIOS-e820: [mem 0x000000009c8ed000-0x000000009cb6cfff] reserved Mar 17 17:41:19.874312 kernel: BIOS-e820: [mem 0x000000009cb6d000-0x000000009cb7efff] ACPI data Mar 17 17:41:19.874319 kernel: BIOS-e820: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS Mar 17 17:41:19.874328 kernel: BIOS-e820: [mem 0x000000009cbff000-0x000000009ce91fff] usable Mar 17 17:41:19.874335 kernel: BIOS-e820: [mem 0x000000009ce92000-0x000000009ce95fff] reserved Mar 17 17:41:19.874341 kernel: BIOS-e820: [mem 0x000000009ce96000-0x000000009ce97fff] ACPI NVS Mar 17 17:41:19.874348 kernel: BIOS-e820: [mem 0x000000009ce98000-0x000000009cedbfff] usable Mar 17 17:41:19.874354 kernel: BIOS-e820: [mem 0x000000009cedc000-0x000000009cf5ffff] reserved Mar 17 17:41:19.874361 kernel: BIOS-e820: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS Mar 17 17:41:19.874368 kernel: BIOS-e820: [mem 0x00000000e0000000-0x00000000efffffff] reserved Mar 17 17:41:19.874374 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Mar 17 17:41:19.874381 kernel: BIOS-e820: [mem 0x00000000ffc00000-0x00000000ffffffff] reserved Mar 17 17:41:19.874388 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Mar 17 17:41:19.874394 kernel: NX (Execute Disable) protection: active Mar 17 17:41:19.874403 kernel: APIC: Static calls initialized Mar 17 17:41:19.874410 kernel: e820: update [mem 0x9b351018-0x9b35ac57] usable ==> usable Mar 17 17:41:19.874417 kernel: e820: update [mem 0x9b351018-0x9b35ac57] usable ==> usable Mar 17 17:41:19.874423 kernel: e820: update [mem 0x9b314018-0x9b350e57] usable ==> usable Mar 17 17:41:19.874429 kernel: e820: update [mem 0x9b314018-0x9b350e57] usable ==> usable Mar 17 17:41:19.874436 kernel: extended physical RAM map: Mar 17 17:41:19.874442 kernel: reserve setup_data: [mem 0x0000000000000000-0x000000000009ffff] usable Mar 17 17:41:19.874449 kernel: reserve setup_data: [mem 0x0000000000100000-0x00000000007fffff] usable Mar 17 17:41:19.874456 kernel: reserve setup_data: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS Mar 17 17:41:19.874463 kernel: reserve setup_data: [mem 0x0000000000808000-0x000000000080afff] usable Mar 17 17:41:19.874469 kernel: reserve setup_data: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS Mar 17 17:41:19.874478 kernel: reserve setup_data: [mem 0x000000000080c000-0x0000000000810fff] usable Mar 17 17:41:19.874485 kernel: reserve setup_data: [mem 0x0000000000811000-0x00000000008fffff] ACPI NVS Mar 17 17:41:19.874495 kernel: reserve setup_data: [mem 0x0000000000900000-0x000000009b314017] usable Mar 17 17:41:19.874502 kernel: reserve setup_data: [mem 0x000000009b314018-0x000000009b350e57] usable Mar 17 17:41:19.874509 kernel: reserve setup_data: [mem 0x000000009b350e58-0x000000009b351017] usable Mar 17 17:41:19.874516 kernel: reserve setup_data: [mem 0x000000009b351018-0x000000009b35ac57] usable Mar 17 17:41:19.874523 kernel: reserve setup_data: [mem 0x000000009b35ac58-0x000000009bd3efff] usable Mar 17 17:41:19.874532 kernel: reserve setup_data: [mem 0x000000009bd3f000-0x000000009bdfffff] reserved Mar 17 17:41:19.874539 kernel: reserve setup_data: [mem 0x000000009be00000-0x000000009c8ecfff] usable Mar 17 17:41:19.874546 kernel: reserve setup_data: [mem 0x000000009c8ed000-0x000000009cb6cfff] reserved Mar 17 17:41:19.874553 kernel: reserve setup_data: [mem 0x000000009cb6d000-0x000000009cb7efff] ACPI data Mar 17 17:41:19.874561 kernel: reserve setup_data: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS Mar 17 17:41:19.874568 kernel: reserve setup_data: [mem 0x000000009cbff000-0x000000009ce91fff] usable Mar 17 17:41:19.874575 kernel: reserve setup_data: [mem 0x000000009ce92000-0x000000009ce95fff] reserved Mar 17 17:41:19.874582 kernel: reserve setup_data: [mem 0x000000009ce96000-0x000000009ce97fff] ACPI NVS Mar 17 17:41:19.874589 kernel: reserve setup_data: [mem 0x000000009ce98000-0x000000009cedbfff] usable Mar 17 17:41:19.874598 kernel: reserve setup_data: [mem 0x000000009cedc000-0x000000009cf5ffff] reserved Mar 17 17:41:19.874605 kernel: reserve setup_data: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS Mar 17 17:41:19.874612 kernel: reserve setup_data: [mem 0x00000000e0000000-0x00000000efffffff] reserved Mar 17 17:41:19.874619 kernel: reserve setup_data: [mem 0x00000000feffc000-0x00000000feffffff] reserved Mar 17 17:41:19.874626 kernel: reserve setup_data: [mem 0x00000000ffc00000-0x00000000ffffffff] reserved Mar 17 17:41:19.874633 kernel: reserve setup_data: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Mar 17 17:41:19.874640 kernel: efi: EFI v2.7 by EDK II Mar 17 17:41:19.874647 kernel: efi: SMBIOS=0x9c988000 ACPI=0x9cb7e000 ACPI 2.0=0x9cb7e014 MEMATTR=0x9ba0d198 RNG=0x9cb73018 Mar 17 17:41:19.874654 kernel: random: crng init done Mar 17 17:41:19.874661 kernel: efi: Remove mem142: MMIO range=[0xffc00000-0xffffffff] (4MB) from e820 map Mar 17 17:41:19.874668 kernel: e820: remove [mem 0xffc00000-0xffffffff] reserved Mar 17 17:41:19.874679 kernel: secureboot: Secure boot disabled Mar 17 17:41:19.874686 kernel: SMBIOS 2.8 present. Mar 17 17:41:19.874693 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS unknown 02/02/2022 Mar 17 17:41:19.874700 kernel: Hypervisor detected: KVM Mar 17 17:41:19.874707 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Mar 17 17:41:19.874724 kernel: kvm-clock: using sched offset of 2744428258 cycles Mar 17 17:41:19.874731 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Mar 17 17:41:19.874739 kernel: tsc: Detected 2794.746 MHz processor Mar 17 17:41:19.874746 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Mar 17 17:41:19.874754 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Mar 17 17:41:19.874761 kernel: last_pfn = 0x9cedc max_arch_pfn = 0x400000000 Mar 17 17:41:19.874771 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs Mar 17 17:41:19.874778 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Mar 17 17:41:19.874785 kernel: Using GB pages for direct mapping Mar 17 17:41:19.874793 kernel: ACPI: Early table checksum verification disabled Mar 17 17:41:19.874885 kernel: ACPI: RSDP 0x000000009CB7E014 000024 (v02 BOCHS ) Mar 17 17:41:19.874893 kernel: ACPI: XSDT 0x000000009CB7D0E8 000054 (v01 BOCHS BXPC 00000001 01000013) Mar 17 17:41:19.874900 kernel: ACPI: FACP 0x000000009CB79000 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Mar 17 17:41:19.874907 kernel: ACPI: DSDT 0x000000009CB7A000 0021A8 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 17 17:41:19.874914 kernel: ACPI: FACS 0x000000009CBDD000 000040 Mar 17 17:41:19.874925 kernel: ACPI: APIC 0x000000009CB78000 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 17 17:41:19.874932 kernel: ACPI: HPET 0x000000009CB77000 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 17 17:41:19.874939 kernel: ACPI: MCFG 0x000000009CB76000 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 17 17:41:19.874947 kernel: ACPI: WAET 0x000000009CB75000 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 17 17:41:19.874954 kernel: ACPI: BGRT 0x000000009CB74000 000038 (v01 INTEL EDK2 00000002 01000013) Mar 17 17:41:19.874961 kernel: ACPI: Reserving FACP table memory at [mem 0x9cb79000-0x9cb790f3] Mar 17 17:41:19.874968 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cb7a000-0x9cb7c1a7] Mar 17 17:41:19.874975 kernel: ACPI: Reserving FACS table memory at [mem 0x9cbdd000-0x9cbdd03f] Mar 17 17:41:19.874985 kernel: ACPI: Reserving APIC table memory at [mem 0x9cb78000-0x9cb7808f] Mar 17 17:41:19.874992 kernel: ACPI: Reserving HPET table memory at [mem 0x9cb77000-0x9cb77037] Mar 17 17:41:19.874999 kernel: ACPI: Reserving MCFG table memory at [mem 0x9cb76000-0x9cb7603b] Mar 17 17:41:19.875006 kernel: ACPI: Reserving WAET table memory at [mem 0x9cb75000-0x9cb75027] Mar 17 17:41:19.875013 kernel: ACPI: Reserving BGRT table memory at [mem 0x9cb74000-0x9cb74037] Mar 17 17:41:19.875020 kernel: No NUMA configuration found Mar 17 17:41:19.875027 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cedbfff] Mar 17 17:41:19.875034 kernel: NODE_DATA(0) allocated [mem 0x9ce3a000-0x9ce3ffff] Mar 17 17:41:19.875042 kernel: Zone ranges: Mar 17 17:41:19.875049 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Mar 17 17:41:19.875058 kernel: DMA32 [mem 0x0000000001000000-0x000000009cedbfff] Mar 17 17:41:19.875065 kernel: Normal empty Mar 17 17:41:19.875072 kernel: Movable zone start for each node Mar 17 17:41:19.875079 kernel: Early memory node ranges Mar 17 17:41:19.875086 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] Mar 17 17:41:19.875094 kernel: node 0: [mem 0x0000000000100000-0x00000000007fffff] Mar 17 17:41:19.875101 kernel: node 0: [mem 0x0000000000808000-0x000000000080afff] Mar 17 17:41:19.875108 kernel: node 0: [mem 0x000000000080c000-0x0000000000810fff] Mar 17 17:41:19.875115 kernel: node 0: [mem 0x0000000000900000-0x000000009bd3efff] Mar 17 17:41:19.875124 kernel: node 0: [mem 0x000000009be00000-0x000000009c8ecfff] Mar 17 17:41:19.875132 kernel: node 0: [mem 0x000000009cbff000-0x000000009ce91fff] Mar 17 17:41:19.875139 kernel: node 0: [mem 0x000000009ce98000-0x000000009cedbfff] Mar 17 17:41:19.875146 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cedbfff] Mar 17 17:41:19.875153 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Mar 17 17:41:19.875160 kernel: On node 0, zone DMA: 96 pages in unavailable ranges Mar 17 17:41:19.875175 kernel: On node 0, zone DMA: 8 pages in unavailable ranges Mar 17 17:41:19.875185 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Mar 17 17:41:19.875192 kernel: On node 0, zone DMA: 239 pages in unavailable ranges Mar 17 17:41:19.875200 kernel: On node 0, zone DMA32: 193 pages in unavailable ranges Mar 17 17:41:19.875207 kernel: On node 0, zone DMA32: 786 pages in unavailable ranges Mar 17 17:41:19.875217 kernel: On node 0, zone DMA32: 6 pages in unavailable ranges Mar 17 17:41:19.875227 kernel: On node 0, zone DMA32: 12580 pages in unavailable ranges Mar 17 17:41:19.875235 kernel: ACPI: PM-Timer IO Port: 0x608 Mar 17 17:41:19.875242 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Mar 17 17:41:19.875250 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Mar 17 17:41:19.875257 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Mar 17 17:41:19.875267 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Mar 17 17:41:19.875274 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Mar 17 17:41:19.875282 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Mar 17 17:41:19.875289 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Mar 17 17:41:19.875297 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Mar 17 17:41:19.875304 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Mar 17 17:41:19.875312 kernel: TSC deadline timer available Mar 17 17:41:19.875320 kernel: smpboot: Allowing 4 CPUs, 0 hotplug CPUs Mar 17 17:41:19.875327 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Mar 17 17:41:19.875337 kernel: kvm-guest: KVM setup pv remote TLB flush Mar 17 17:41:19.875345 kernel: kvm-guest: setup PV sched yield Mar 17 17:41:19.875352 kernel: [mem 0x9d000000-0xdfffffff] available for PCI devices Mar 17 17:41:19.875360 kernel: Booting paravirtualized kernel on KVM Mar 17 17:41:19.875367 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Mar 17 17:41:19.875375 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:4 nr_cpu_ids:4 nr_node_ids:1 Mar 17 17:41:19.875383 kernel: percpu: Embedded 58 pages/cpu s197032 r8192 d32344 u524288 Mar 17 17:41:19.875390 kernel: pcpu-alloc: s197032 r8192 d32344 u524288 alloc=1*2097152 Mar 17 17:41:19.875398 kernel: pcpu-alloc: [0] 0 1 2 3 Mar 17 17:41:19.875407 kernel: kvm-guest: PV spinlocks enabled Mar 17 17:41:19.875415 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Mar 17 17:41:19.875424 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=d4b838cd9a6f58e8c4a6b615c32b0b28ee0df1660e34033a8fbd0429c6de5fd0 Mar 17 17:41:19.875432 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Mar 17 17:41:19.875439 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Mar 17 17:41:19.875447 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Mar 17 17:41:19.875454 kernel: Fallback order for Node 0: 0 Mar 17 17:41:19.875462 kernel: Built 1 zonelists, mobility grouping on. Total pages: 629460 Mar 17 17:41:19.875472 kernel: Policy zone: DMA32 Mar 17 17:41:19.875479 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Mar 17 17:41:19.875487 kernel: Memory: 2389768K/2565800K available (12288K kernel code, 2303K rwdata, 22744K rodata, 42992K init, 2196K bss, 175776K reserved, 0K cma-reserved) Mar 17 17:41:19.875495 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Mar 17 17:41:19.875502 kernel: ftrace: allocating 37938 entries in 149 pages Mar 17 17:41:19.875510 kernel: ftrace: allocated 149 pages with 4 groups Mar 17 17:41:19.875517 kernel: Dynamic Preempt: voluntary Mar 17 17:41:19.875525 kernel: rcu: Preemptible hierarchical RCU implementation. Mar 17 17:41:19.875533 kernel: rcu: RCU event tracing is enabled. Mar 17 17:41:19.875543 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Mar 17 17:41:19.875551 kernel: Trampoline variant of Tasks RCU enabled. Mar 17 17:41:19.875559 kernel: Rude variant of Tasks RCU enabled. Mar 17 17:41:19.875566 kernel: Tracing variant of Tasks RCU enabled. Mar 17 17:41:19.875574 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Mar 17 17:41:19.875581 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Mar 17 17:41:19.875589 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 Mar 17 17:41:19.875596 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Mar 17 17:41:19.875604 kernel: Console: colour dummy device 80x25 Mar 17 17:41:19.875614 kernel: printk: console [ttyS0] enabled Mar 17 17:41:19.875621 kernel: ACPI: Core revision 20230628 Mar 17 17:41:19.875629 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Mar 17 17:41:19.875636 kernel: APIC: Switch to symmetric I/O mode setup Mar 17 17:41:19.875644 kernel: x2apic enabled Mar 17 17:41:19.875651 kernel: APIC: Switched APIC routing to: physical x2apic Mar 17 17:41:19.875662 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Mar 17 17:41:19.875669 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Mar 17 17:41:19.875677 kernel: kvm-guest: setup PV IPIs Mar 17 17:41:19.875684 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Mar 17 17:41:19.875695 kernel: tsc: Marking TSC unstable due to TSCs unsynchronized Mar 17 17:41:19.875702 kernel: Calibrating delay loop (skipped) preset value.. 5589.49 BogoMIPS (lpj=2794746) Mar 17 17:41:19.875710 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Mar 17 17:41:19.875725 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Mar 17 17:41:19.875733 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Mar 17 17:41:19.875740 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Mar 17 17:41:19.875750 kernel: Spectre V2 : Mitigation: Retpolines Mar 17 17:41:19.875761 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch Mar 17 17:41:19.875774 kernel: Spectre V2 : Spectre v2 / SpectreRSB : Filling RSB on VMEXIT Mar 17 17:41:19.875782 kernel: Spectre V2 : Enabling Speculation Barrier for firmware calls Mar 17 17:41:19.875789 kernel: RETBleed: Mitigation: untrained return thunk Mar 17 17:41:19.875807 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Mar 17 17:41:19.875815 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Mar 17 17:41:19.875823 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! Mar 17 17:41:19.875831 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. Mar 17 17:41:19.875838 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode Mar 17 17:41:19.875846 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Mar 17 17:41:19.875856 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Mar 17 17:41:19.875864 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Mar 17 17:41:19.875871 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Mar 17 17:41:19.875879 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Mar 17 17:41:19.875886 kernel: Freeing SMP alternatives memory: 32K Mar 17 17:41:19.875894 kernel: pid_max: default: 32768 minimum: 301 Mar 17 17:41:19.875901 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Mar 17 17:41:19.875909 kernel: landlock: Up and running. Mar 17 17:41:19.875916 kernel: SELinux: Initializing. Mar 17 17:41:19.875926 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Mar 17 17:41:19.875933 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Mar 17 17:41:19.875941 kernel: smpboot: CPU0: AMD EPYC 7402P 24-Core Processor (family: 0x17, model: 0x31, stepping: 0x0) Mar 17 17:41:19.875948 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Mar 17 17:41:19.875956 kernel: RCU Tasks Rude: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Mar 17 17:41:19.875963 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Mar 17 17:41:19.875971 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. Mar 17 17:41:19.875978 kernel: ... version: 0 Mar 17 17:41:19.875988 kernel: ... bit width: 48 Mar 17 17:41:19.875995 kernel: ... generic registers: 6 Mar 17 17:41:19.876003 kernel: ... value mask: 0000ffffffffffff Mar 17 17:41:19.876010 kernel: ... max period: 00007fffffffffff Mar 17 17:41:19.876018 kernel: ... fixed-purpose events: 0 Mar 17 17:41:19.876025 kernel: ... event mask: 000000000000003f Mar 17 17:41:19.876032 kernel: signal: max sigframe size: 1776 Mar 17 17:41:19.876040 kernel: rcu: Hierarchical SRCU implementation. Mar 17 17:41:19.876047 kernel: rcu: Max phase no-delay instances is 400. Mar 17 17:41:19.876055 kernel: smp: Bringing up secondary CPUs ... Mar 17 17:41:19.876064 kernel: smpboot: x86: Booting SMP configuration: Mar 17 17:41:19.876071 kernel: .... node #0, CPUs: #1 #2 #3 Mar 17 17:41:19.876079 kernel: smp: Brought up 1 node, 4 CPUs Mar 17 17:41:19.876086 kernel: smpboot: Max logical packages: 1 Mar 17 17:41:19.876094 kernel: smpboot: Total of 4 processors activated (22357.96 BogoMIPS) Mar 17 17:41:19.876101 kernel: devtmpfs: initialized Mar 17 17:41:19.876109 kernel: x86/mm: Memory block size: 128MB Mar 17 17:41:19.876116 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00800000-0x00807fff] (32768 bytes) Mar 17 17:41:19.876124 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x0080b000-0x0080bfff] (4096 bytes) Mar 17 17:41:19.876134 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00811000-0x008fffff] (978944 bytes) Mar 17 17:41:19.876142 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cb7f000-0x9cbfefff] (524288 bytes) Mar 17 17:41:19.876149 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9ce96000-0x9ce97fff] (8192 bytes) Mar 17 17:41:19.876157 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cf60000-0x9cffffff] (655360 bytes) Mar 17 17:41:19.876165 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Mar 17 17:41:19.876172 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Mar 17 17:41:19.876180 kernel: pinctrl core: initialized pinctrl subsystem Mar 17 17:41:19.876187 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Mar 17 17:41:19.876194 kernel: audit: initializing netlink subsys (disabled) Mar 17 17:41:19.876204 kernel: audit: type=2000 audit(1742233279.027:1): state=initialized audit_enabled=0 res=1 Mar 17 17:41:19.876211 kernel: thermal_sys: Registered thermal governor 'step_wise' Mar 17 17:41:19.876219 kernel: thermal_sys: Registered thermal governor 'user_space' Mar 17 17:41:19.876226 kernel: cpuidle: using governor menu Mar 17 17:41:19.876234 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Mar 17 17:41:19.876241 kernel: dca service started, version 1.12.1 Mar 17 17:41:19.876249 kernel: PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xe0000000-0xefffffff] (base 0xe0000000) Mar 17 17:41:19.876256 kernel: PCI: Using configuration type 1 for base access Mar 17 17:41:19.876266 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Mar 17 17:41:19.876273 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Mar 17 17:41:19.876281 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Mar 17 17:41:19.876288 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Mar 17 17:41:19.876296 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Mar 17 17:41:19.876303 kernel: ACPI: Added _OSI(Module Device) Mar 17 17:41:19.876311 kernel: ACPI: Added _OSI(Processor Device) Mar 17 17:41:19.876318 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Mar 17 17:41:19.876325 kernel: ACPI: Added _OSI(Processor Aggregator Device) Mar 17 17:41:19.876335 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Mar 17 17:41:19.876342 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Mar 17 17:41:19.876350 kernel: ACPI: Interpreter enabled Mar 17 17:41:19.876357 kernel: ACPI: PM: (supports S0 S3 S5) Mar 17 17:41:19.876364 kernel: ACPI: Using IOAPIC for interrupt routing Mar 17 17:41:19.876372 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Mar 17 17:41:19.876379 kernel: PCI: Using E820 reservations for host bridge windows Mar 17 17:41:19.876387 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Mar 17 17:41:19.876394 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Mar 17 17:41:19.876578 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Mar 17 17:41:19.876708 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Mar 17 17:41:19.876870 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Mar 17 17:41:19.876881 kernel: PCI host bridge to bus 0000:00 Mar 17 17:41:19.877012 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Mar 17 17:41:19.877124 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Mar 17 17:41:19.877234 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Mar 17 17:41:19.877442 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xdfffffff window] Mar 17 17:41:19.877580 kernel: pci_bus 0000:00: root bus resource [mem 0xf0000000-0xfebfffff window] Mar 17 17:41:19.877692 kernel: pci_bus 0000:00: root bus resource [mem 0x380000000000-0x3807ffffffff window] Mar 17 17:41:19.877836 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Mar 17 17:41:19.877975 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 Mar 17 17:41:19.878107 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 Mar 17 17:41:19.878234 kernel: pci 0000:00:01.0: reg 0x10: [mem 0xc0000000-0xc0ffffff pref] Mar 17 17:41:19.878354 kernel: pci 0000:00:01.0: reg 0x18: [mem 0xc1044000-0xc1044fff] Mar 17 17:41:19.878472 kernel: pci 0000:00:01.0: reg 0x30: [mem 0xffff0000-0xffffffff pref] Mar 17 17:41:19.878622 kernel: pci 0000:00:01.0: BAR 0: assigned to efifb Mar 17 17:41:19.878839 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Mar 17 17:41:19.879096 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 Mar 17 17:41:19.879258 kernel: pci 0000:00:02.0: reg 0x10: [io 0x6100-0x611f] Mar 17 17:41:19.879397 kernel: pci 0000:00:02.0: reg 0x14: [mem 0xc1043000-0xc1043fff] Mar 17 17:41:19.879532 kernel: pci 0000:00:02.0: reg 0x20: [mem 0x380000000000-0x380000003fff 64bit pref] Mar 17 17:41:19.879669 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 Mar 17 17:41:19.879830 kernel: pci 0000:00:03.0: reg 0x10: [io 0x6000-0x607f] Mar 17 17:41:19.880597 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xc1042000-0xc1042fff] Mar 17 17:41:19.880743 kernel: pci 0000:00:03.0: reg 0x20: [mem 0x380000004000-0x380000007fff 64bit pref] Mar 17 17:41:19.880895 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 Mar 17 17:41:19.881024 kernel: pci 0000:00:04.0: reg 0x10: [io 0x60e0-0x60ff] Mar 17 17:41:19.881145 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xc1041000-0xc1041fff] Mar 17 17:41:19.881272 kernel: pci 0000:00:04.0: reg 0x20: [mem 0x380000008000-0x38000000bfff 64bit pref] Mar 17 17:41:19.881395 kernel: pci 0000:00:04.0: reg 0x30: [mem 0xfffc0000-0xffffffff pref] Mar 17 17:41:19.881526 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 Mar 17 17:41:19.881647 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Mar 17 17:41:19.881824 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 Mar 17 17:41:19.881957 kernel: pci 0000:00:1f.2: reg 0x20: [io 0x60c0-0x60df] Mar 17 17:41:19.882076 kernel: pci 0000:00:1f.2: reg 0x24: [mem 0xc1040000-0xc1040fff] Mar 17 17:41:19.882203 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 Mar 17 17:41:19.882324 kernel: pci 0000:00:1f.3: reg 0x20: [io 0x6080-0x60bf] Mar 17 17:41:19.882334 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Mar 17 17:41:19.882342 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Mar 17 17:41:19.882350 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Mar 17 17:41:19.882362 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Mar 17 17:41:19.882370 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Mar 17 17:41:19.882378 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Mar 17 17:41:19.882386 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Mar 17 17:41:19.882393 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Mar 17 17:41:19.882401 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Mar 17 17:41:19.882409 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Mar 17 17:41:19.882417 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Mar 17 17:41:19.882424 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Mar 17 17:41:19.882435 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Mar 17 17:41:19.882443 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Mar 17 17:41:19.882450 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Mar 17 17:41:19.882458 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Mar 17 17:41:19.882466 kernel: iommu: Default domain type: Translated Mar 17 17:41:19.882474 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Mar 17 17:41:19.882481 kernel: efivars: Registered efivars operations Mar 17 17:41:19.882489 kernel: PCI: Using ACPI for IRQ routing Mar 17 17:41:19.882497 kernel: PCI: pci_cache_line_size set to 64 bytes Mar 17 17:41:19.882509 kernel: e820: reserve RAM buffer [mem 0x0080b000-0x008fffff] Mar 17 17:41:19.882517 kernel: e820: reserve RAM buffer [mem 0x00811000-0x008fffff] Mar 17 17:41:19.882524 kernel: e820: reserve RAM buffer [mem 0x9b314018-0x9bffffff] Mar 17 17:41:19.882532 kernel: e820: reserve RAM buffer [mem 0x9b351018-0x9bffffff] Mar 17 17:41:19.882540 kernel: e820: reserve RAM buffer [mem 0x9bd3f000-0x9bffffff] Mar 17 17:41:19.882547 kernel: e820: reserve RAM buffer [mem 0x9c8ed000-0x9fffffff] Mar 17 17:41:19.882555 kernel: e820: reserve RAM buffer [mem 0x9ce92000-0x9fffffff] Mar 17 17:41:19.882563 kernel: e820: reserve RAM buffer [mem 0x9cedc000-0x9fffffff] Mar 17 17:41:19.882728 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Mar 17 17:41:19.882898 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Mar 17 17:41:19.883022 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Mar 17 17:41:19.883032 kernel: vgaarb: loaded Mar 17 17:41:19.883040 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Mar 17 17:41:19.883048 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Mar 17 17:41:19.883056 kernel: clocksource: Switched to clocksource kvm-clock Mar 17 17:41:19.883064 kernel: VFS: Disk quotas dquot_6.6.0 Mar 17 17:41:19.883072 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Mar 17 17:41:19.883085 kernel: pnp: PnP ACPI init Mar 17 17:41:19.883226 kernel: system 00:05: [mem 0xe0000000-0xefffffff window] has been reserved Mar 17 17:41:19.883237 kernel: pnp: PnP ACPI: found 6 devices Mar 17 17:41:19.883245 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Mar 17 17:41:19.883254 kernel: NET: Registered PF_INET protocol family Mar 17 17:41:19.883284 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Mar 17 17:41:19.883295 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Mar 17 17:41:19.883303 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Mar 17 17:41:19.883314 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Mar 17 17:41:19.883322 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Mar 17 17:41:19.883330 kernel: TCP: Hash tables configured (established 32768 bind 32768) Mar 17 17:41:19.883338 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Mar 17 17:41:19.883346 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Mar 17 17:41:19.883354 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Mar 17 17:41:19.883363 kernel: NET: Registered PF_XDP protocol family Mar 17 17:41:19.883489 kernel: pci 0000:00:04.0: can't claim BAR 6 [mem 0xfffc0000-0xffffffff pref]: no compatible bridge window Mar 17 17:41:19.883663 kernel: pci 0000:00:04.0: BAR 6: assigned [mem 0x9d000000-0x9d03ffff pref] Mar 17 17:41:19.883826 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Mar 17 17:41:19.883942 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Mar 17 17:41:19.884053 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Mar 17 17:41:19.884170 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xdfffffff window] Mar 17 17:41:19.884281 kernel: pci_bus 0000:00: resource 8 [mem 0xf0000000-0xfebfffff window] Mar 17 17:41:19.884392 kernel: pci_bus 0000:00: resource 9 [mem 0x380000000000-0x3807ffffffff window] Mar 17 17:41:19.884402 kernel: PCI: CLS 0 bytes, default 64 Mar 17 17:41:19.884414 kernel: Initialise system trusted keyrings Mar 17 17:41:19.884422 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Mar 17 17:41:19.884430 kernel: Key type asymmetric registered Mar 17 17:41:19.884438 kernel: Asymmetric key parser 'x509' registered Mar 17 17:41:19.884447 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Mar 17 17:41:19.884455 kernel: io scheduler mq-deadline registered Mar 17 17:41:19.884463 kernel: io scheduler kyber registered Mar 17 17:41:19.884471 kernel: io scheduler bfq registered Mar 17 17:41:19.884479 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Mar 17 17:41:19.884487 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Mar 17 17:41:19.884499 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Mar 17 17:41:19.884510 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Mar 17 17:41:19.884518 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Mar 17 17:41:19.884526 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Mar 17 17:41:19.884534 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Mar 17 17:41:19.884545 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Mar 17 17:41:19.884553 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Mar 17 17:41:19.884678 kernel: rtc_cmos 00:04: RTC can wake from S4 Mar 17 17:41:19.884690 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Mar 17 17:41:19.884831 kernel: rtc_cmos 00:04: registered as rtc0 Mar 17 17:41:19.884948 kernel: rtc_cmos 00:04: setting system clock to 2025-03-17T17:41:19 UTC (1742233279) Mar 17 17:41:19.885063 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram Mar 17 17:41:19.885073 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Mar 17 17:41:19.885086 kernel: efifb: probing for efifb Mar 17 17:41:19.885094 kernel: efifb: framebuffer at 0xc0000000, using 4000k, total 4000k Mar 17 17:41:19.885102 kernel: efifb: mode is 1280x800x32, linelength=5120, pages=1 Mar 17 17:41:19.885110 kernel: efifb: scrolling: redraw Mar 17 17:41:19.885119 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Mar 17 17:41:19.885127 kernel: Console: switching to colour frame buffer device 160x50 Mar 17 17:41:19.885135 kernel: fb0: EFI VGA frame buffer device Mar 17 17:41:19.885143 kernel: pstore: Using crash dump compression: deflate Mar 17 17:41:19.885152 kernel: pstore: Registered efi_pstore as persistent store backend Mar 17 17:41:19.885163 kernel: NET: Registered PF_INET6 protocol family Mar 17 17:41:19.885171 kernel: Segment Routing with IPv6 Mar 17 17:41:19.885180 kernel: In-situ OAM (IOAM) with IPv6 Mar 17 17:41:19.885190 kernel: NET: Registered PF_PACKET protocol family Mar 17 17:41:19.885199 kernel: Key type dns_resolver registered Mar 17 17:41:19.885208 kernel: IPI shorthand broadcast: enabled Mar 17 17:41:19.885216 kernel: sched_clock: Marking stable (900002872, 183086540)->(1168261712, -85172300) Mar 17 17:41:19.885225 kernel: registered taskstats version 1 Mar 17 17:41:19.885233 kernel: Loading compiled-in X.509 certificates Mar 17 17:41:19.885244 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.83-flatcar: 608fb88224bc0ea76afefc598557abb0413f36c0' Mar 17 17:41:19.885252 kernel: Key type .fscrypt registered Mar 17 17:41:19.885260 kernel: Key type fscrypt-provisioning registered Mar 17 17:41:19.885268 kernel: ima: No TPM chip found, activating TPM-bypass! Mar 17 17:41:19.885276 kernel: ima: Allocated hash algorithm: sha1 Mar 17 17:41:19.885284 kernel: ima: No architecture policies found Mar 17 17:41:19.885292 kernel: clk: Disabling unused clocks Mar 17 17:41:19.885301 kernel: Freeing unused kernel image (initmem) memory: 42992K Mar 17 17:41:19.885309 kernel: Write protecting the kernel read-only data: 36864k Mar 17 17:41:19.885320 kernel: Freeing unused kernel image (rodata/data gap) memory: 1832K Mar 17 17:41:19.885328 kernel: Run /init as init process Mar 17 17:41:19.885336 kernel: with arguments: Mar 17 17:41:19.885344 kernel: /init Mar 17 17:41:19.885352 kernel: with environment: Mar 17 17:41:19.885360 kernel: HOME=/ Mar 17 17:41:19.885367 kernel: TERM=linux Mar 17 17:41:19.885375 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Mar 17 17:41:19.885387 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Mar 17 17:41:19.885401 systemd[1]: Detected virtualization kvm. Mar 17 17:41:19.885409 systemd[1]: Detected architecture x86-64. Mar 17 17:41:19.885418 systemd[1]: Running in initrd. Mar 17 17:41:19.885426 systemd[1]: No hostname configured, using default hostname. Mar 17 17:41:19.885434 systemd[1]: Hostname set to . Mar 17 17:41:19.885443 systemd[1]: Initializing machine ID from VM UUID. Mar 17 17:41:19.885451 systemd[1]: Queued start job for default target initrd.target. Mar 17 17:41:19.885463 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 17 17:41:19.885471 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 17 17:41:19.885480 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Mar 17 17:41:19.885489 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Mar 17 17:41:19.885497 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Mar 17 17:41:19.885506 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Mar 17 17:41:19.885517 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Mar 17 17:41:19.885528 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Mar 17 17:41:19.885537 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 17 17:41:19.885545 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Mar 17 17:41:19.885554 systemd[1]: Reached target paths.target - Path Units. Mar 17 17:41:19.885562 systemd[1]: Reached target slices.target - Slice Units. Mar 17 17:41:19.885570 systemd[1]: Reached target swap.target - Swaps. Mar 17 17:41:19.885579 systemd[1]: Reached target timers.target - Timer Units. Mar 17 17:41:19.885587 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Mar 17 17:41:19.885598 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Mar 17 17:41:19.885607 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Mar 17 17:41:19.885615 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Mar 17 17:41:19.885624 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Mar 17 17:41:19.885633 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Mar 17 17:41:19.885641 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Mar 17 17:41:19.885650 systemd[1]: Reached target sockets.target - Socket Units. Mar 17 17:41:19.885659 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Mar 17 17:41:19.885667 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Mar 17 17:41:19.885678 systemd[1]: Finished network-cleanup.service - Network Cleanup. Mar 17 17:41:19.885687 systemd[1]: Starting systemd-fsck-usr.service... Mar 17 17:41:19.885696 systemd[1]: Starting systemd-journald.service - Journal Service... Mar 17 17:41:19.885704 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Mar 17 17:41:19.885722 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 17 17:41:19.885731 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Mar 17 17:41:19.885740 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Mar 17 17:41:19.885749 systemd[1]: Finished systemd-fsck-usr.service. Mar 17 17:41:19.885780 systemd-journald[192]: Collecting audit messages is disabled. Mar 17 17:41:19.885819 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Mar 17 17:41:19.885828 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 17 17:41:19.885836 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 17 17:41:19.885845 systemd-journald[192]: Journal started Mar 17 17:41:19.885864 systemd-journald[192]: Runtime Journal (/run/log/journal/1490813c1f8f4ac78a5a7071c37f2d7a) is 6.0M, max 48.3M, 42.2M free. Mar 17 17:41:19.880466 systemd-modules-load[195]: Inserted module 'overlay' Mar 17 17:41:19.890461 systemd[1]: Started systemd-journald.service - Journal Service. Mar 17 17:41:19.891076 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Mar 17 17:41:19.897558 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Mar 17 17:41:19.900945 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Mar 17 17:41:19.909829 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Mar 17 17:41:19.911647 systemd-modules-load[195]: Inserted module 'br_netfilter' Mar 17 17:41:19.912325 kernel: Bridge firewalling registered Mar 17 17:41:19.912875 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Mar 17 17:41:19.914812 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 17 17:41:19.917667 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 17 17:41:19.932016 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Mar 17 17:41:19.935553 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 17 17:41:19.938101 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 17 17:41:19.944816 dracut-cmdline[223]: dracut-dracut-053 Mar 17 17:41:19.948542 dracut-cmdline[223]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=d4b838cd9a6f58e8c4a6b615c32b0b28ee0df1660e34033a8fbd0429c6de5fd0 Mar 17 17:41:19.948639 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 17 17:41:19.960025 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Mar 17 17:41:20.001524 systemd-resolved[247]: Positive Trust Anchors: Mar 17 17:41:20.001545 systemd-resolved[247]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 17 17:41:20.001576 systemd-resolved[247]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Mar 17 17:41:20.012314 systemd-resolved[247]: Defaulting to hostname 'linux'. Mar 17 17:41:20.014225 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Mar 17 17:41:20.016389 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Mar 17 17:41:20.035823 kernel: SCSI subsystem initialized Mar 17 17:41:20.044820 kernel: Loading iSCSI transport class v2.0-870. Mar 17 17:41:20.055834 kernel: iscsi: registered transport (tcp) Mar 17 17:41:20.078825 kernel: iscsi: registered transport (qla4xxx) Mar 17 17:41:20.078854 kernel: QLogic iSCSI HBA Driver Mar 17 17:41:20.131776 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Mar 17 17:41:20.142023 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Mar 17 17:41:20.168197 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Mar 17 17:41:20.168227 kernel: device-mapper: uevent: version 1.0.3 Mar 17 17:41:20.169240 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Mar 17 17:41:20.210826 kernel: raid6: avx2x4 gen() 29556 MB/s Mar 17 17:41:20.227823 kernel: raid6: avx2x2 gen() 30341 MB/s Mar 17 17:41:20.244937 kernel: raid6: avx2x1 gen() 25161 MB/s Mar 17 17:41:20.244976 kernel: raid6: using algorithm avx2x2 gen() 30341 MB/s Mar 17 17:41:20.262937 kernel: raid6: .... xor() 19258 MB/s, rmw enabled Mar 17 17:41:20.262994 kernel: raid6: using avx2x2 recovery algorithm Mar 17 17:41:20.283852 kernel: xor: automatically using best checksumming function avx Mar 17 17:41:20.436834 kernel: Btrfs loaded, zoned=no, fsverity=no Mar 17 17:41:20.451213 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Mar 17 17:41:20.467955 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 17 17:41:20.481912 systemd-udevd[414]: Using default interface naming scheme 'v255'. Mar 17 17:41:20.486521 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 17 17:41:20.498027 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Mar 17 17:41:20.513033 dracut-pre-trigger[423]: rd.md=0: removing MD RAID activation Mar 17 17:41:20.545390 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Mar 17 17:41:20.561921 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Mar 17 17:41:20.623419 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Mar 17 17:41:20.628119 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Mar 17 17:41:20.642113 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Mar 17 17:41:20.644924 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Mar 17 17:41:20.647502 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 17 17:41:20.650178 systemd[1]: Reached target remote-fs.target - Remote File Systems. Mar 17 17:41:20.659057 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Mar 17 17:41:20.661898 kernel: cryptd: max_cpu_qlen set to 1000 Mar 17 17:41:20.667846 kernel: virtio_blk virtio1: 4/0/0 default/read/poll queues Mar 17 17:41:20.689385 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Mar 17 17:41:20.689592 kernel: AVX2 version of gcm_enc/dec engaged. Mar 17 17:41:20.689607 kernel: AES CTR mode by8 optimization enabled Mar 17 17:41:20.689620 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Mar 17 17:41:20.689632 kernel: GPT:9289727 != 19775487 Mar 17 17:41:20.689652 kernel: GPT:Alternate GPT header not at the end of the disk. Mar 17 17:41:20.689664 kernel: GPT:9289727 != 19775487 Mar 17 17:41:20.689676 kernel: GPT: Use GNU Parted to correct GPT errors. Mar 17 17:41:20.689688 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Mar 17 17:41:20.677386 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Mar 17 17:41:20.694053 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Mar 17 17:41:20.695156 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 17 17:41:20.697127 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 17 17:41:20.698622 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 17 17:41:20.702770 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 17 17:41:20.706878 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Mar 17 17:41:20.712302 kernel: libata version 3.00 loaded. Mar 17 17:41:20.719032 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 17 17:41:20.728822 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/vda6 scanned by (udev-worker) (457) Mar 17 17:41:20.730973 kernel: BTRFS: device fsid 2b8ebefd-e897-48f6-96d5-0893fbb7c64a devid 1 transid 40 /dev/vda3 scanned by (udev-worker) (475) Mar 17 17:41:20.731444 kernel: ahci 0000:00:1f.2: version 3.0 Mar 17 17:41:20.747607 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Mar 17 17:41:20.747625 kernel: ahci 0000:00:1f.2: AHCI 0001.0000 32 slots 6 ports 1.5 Gbps 0x3f impl SATA mode Mar 17 17:41:20.747823 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Mar 17 17:41:20.747970 kernel: scsi host0: ahci Mar 17 17:41:20.748116 kernel: scsi host1: ahci Mar 17 17:41:20.748257 kernel: scsi host2: ahci Mar 17 17:41:20.748394 kernel: scsi host3: ahci Mar 17 17:41:20.749651 kernel: scsi host4: ahci Mar 17 17:41:20.749852 kernel: scsi host5: ahci Mar 17 17:41:20.749998 kernel: ata1: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040100 irq 34 Mar 17 17:41:20.750010 kernel: ata2: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040180 irq 34 Mar 17 17:41:20.750021 kernel: ata3: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040200 irq 34 Mar 17 17:41:20.750031 kernel: ata4: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040280 irq 34 Mar 17 17:41:20.750041 kernel: ata5: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040300 irq 34 Mar 17 17:41:20.750051 kernel: ata6: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040380 irq 34 Mar 17 17:41:20.743251 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Mar 17 17:41:20.752577 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Mar 17 17:41:20.767060 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Mar 17 17:41:20.771178 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Mar 17 17:41:20.771581 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Mar 17 17:41:20.785932 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Mar 17 17:41:20.786359 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 17 17:41:20.786410 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 17 17:41:20.789177 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Mar 17 17:41:20.794250 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 17 17:41:20.814527 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 17 17:41:20.821550 disk-uuid[568]: Primary Header is updated. Mar 17 17:41:20.821550 disk-uuid[568]: Secondary Entries is updated. Mar 17 17:41:20.821550 disk-uuid[568]: Secondary Header is updated. Mar 17 17:41:20.829854 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Mar 17 17:41:20.835841 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Mar 17 17:41:20.837837 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 17 17:41:20.868926 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 17 17:41:21.055856 kernel: ata2: SATA link down (SStatus 0 SControl 300) Mar 17 17:41:21.055951 kernel: ata1: SATA link down (SStatus 0 SControl 300) Mar 17 17:41:21.056837 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Mar 17 17:41:21.058502 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Mar 17 17:41:21.058541 kernel: ata3.00: applying bridge limits Mar 17 17:41:21.059840 kernel: ata3.00: configured for UDMA/100 Mar 17 17:41:21.062269 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Mar 17 17:41:21.062380 kernel: ata6: SATA link down (SStatus 0 SControl 300) Mar 17 17:41:21.063145 kernel: ata5: SATA link down (SStatus 0 SControl 300) Mar 17 17:41:21.065843 kernel: ata4: SATA link down (SStatus 0 SControl 300) Mar 17 17:41:21.142875 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Mar 17 17:41:21.163123 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Mar 17 17:41:21.163148 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Mar 17 17:41:21.841824 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Mar 17 17:41:21.842414 disk-uuid[572]: The operation has completed successfully. Mar 17 17:41:21.869289 systemd[1]: disk-uuid.service: Deactivated successfully. Mar 17 17:41:21.869414 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Mar 17 17:41:21.903013 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Mar 17 17:41:21.907281 sh[597]: Success Mar 17 17:41:21.921839 kernel: device-mapper: verity: sha256 using implementation "sha256-ni" Mar 17 17:41:21.955503 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Mar 17 17:41:21.969265 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Mar 17 17:41:21.972067 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Mar 17 17:41:21.984056 kernel: BTRFS info (device dm-0): first mount of filesystem 2b8ebefd-e897-48f6-96d5-0893fbb7c64a Mar 17 17:41:21.984085 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Mar 17 17:41:21.984103 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Mar 17 17:41:21.985816 kernel: BTRFS info (device dm-0): disabling log replay at mount time Mar 17 17:41:21.985830 kernel: BTRFS info (device dm-0): using free space tree Mar 17 17:41:21.990624 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Mar 17 17:41:21.991737 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Mar 17 17:41:22.005959 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Mar 17 17:41:22.008921 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Mar 17 17:41:22.016172 kernel: BTRFS info (device vda6): first mount of filesystem 7b241d32-136b-4fe3-b105-cecff2b2cf64 Mar 17 17:41:22.016202 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Mar 17 17:41:22.016215 kernel: BTRFS info (device vda6): using free space tree Mar 17 17:41:22.019826 kernel: BTRFS info (device vda6): auto enabling async discard Mar 17 17:41:22.028518 systemd[1]: mnt-oem.mount: Deactivated successfully. Mar 17 17:41:22.030269 kernel: BTRFS info (device vda6): last unmount of filesystem 7b241d32-136b-4fe3-b105-cecff2b2cf64 Mar 17 17:41:22.102892 systemd[1]: Finished ignition-setup.service - Ignition (setup). Mar 17 17:41:22.109067 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Mar 17 17:41:22.125083 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Mar 17 17:41:22.135987 systemd[1]: Starting systemd-networkd.service - Network Configuration... Mar 17 17:41:22.158558 systemd-networkd[779]: lo: Link UP Mar 17 17:41:22.159101 systemd-networkd[779]: lo: Gained carrier Mar 17 17:41:22.162059 systemd-networkd[779]: Enumeration completed Mar 17 17:41:22.162411 systemd[1]: Started systemd-networkd.service - Network Configuration. Mar 17 17:41:22.162494 systemd-networkd[779]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 17 17:41:22.162498 systemd-networkd[779]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 17 17:41:22.163694 systemd-networkd[779]: eth0: Link UP Mar 17 17:41:22.163699 systemd-networkd[779]: eth0: Gained carrier Mar 17 17:41:22.163706 systemd-networkd[779]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 17 17:41:22.172199 ignition[763]: Ignition 2.20.0 Mar 17 17:41:22.168372 systemd[1]: Reached target network.target - Network. Mar 17 17:41:22.172206 ignition[763]: Stage: fetch-offline Mar 17 17:41:22.174842 systemd-networkd[779]: eth0: DHCPv4 address 10.0.0.61/16, gateway 10.0.0.1 acquired from 10.0.0.1 Mar 17 17:41:22.172244 ignition[763]: no configs at "/usr/lib/ignition/base.d" Mar 17 17:41:22.172255 ignition[763]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 17 17:41:22.172371 ignition[763]: parsed url from cmdline: "" Mar 17 17:41:22.172376 ignition[763]: no config URL provided Mar 17 17:41:22.172382 ignition[763]: reading system config file "/usr/lib/ignition/user.ign" Mar 17 17:41:22.172392 ignition[763]: no config at "/usr/lib/ignition/user.ign" Mar 17 17:41:22.172428 ignition[763]: op(1): [started] loading QEMU firmware config module Mar 17 17:41:22.172434 ignition[763]: op(1): executing: "modprobe" "qemu_fw_cfg" Mar 17 17:41:22.181522 ignition[763]: op(1): [finished] loading QEMU firmware config module Mar 17 17:41:22.224594 ignition[763]: parsing config with SHA512: 2d2c2431972a7e27838ac0ca3e96b1d8e8898aaad0bad157d29a9d6571fdd0678b94ce0f284a8787239ae75cebf00555e82ed81af1c8a659b47ed7aca28c03d0 Mar 17 17:41:22.229672 unknown[763]: fetched base config from "system" Mar 17 17:41:22.229685 unknown[763]: fetched user config from "qemu" Mar 17 17:41:22.230068 ignition[763]: fetch-offline: fetch-offline passed Mar 17 17:41:22.230137 ignition[763]: Ignition finished successfully Mar 17 17:41:22.232584 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Mar 17 17:41:22.234759 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Mar 17 17:41:22.248054 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Mar 17 17:41:22.262248 ignition[791]: Ignition 2.20.0 Mar 17 17:41:22.262260 ignition[791]: Stage: kargs Mar 17 17:41:22.262445 ignition[791]: no configs at "/usr/lib/ignition/base.d" Mar 17 17:41:22.262458 ignition[791]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 17 17:41:22.263280 ignition[791]: kargs: kargs passed Mar 17 17:41:22.263324 ignition[791]: Ignition finished successfully Mar 17 17:41:22.267183 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Mar 17 17:41:22.284973 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Mar 17 17:41:22.296078 ignition[801]: Ignition 2.20.0 Mar 17 17:41:22.296091 ignition[801]: Stage: disks Mar 17 17:41:22.296261 ignition[801]: no configs at "/usr/lib/ignition/base.d" Mar 17 17:41:22.296274 ignition[801]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 17 17:41:22.300199 ignition[801]: disks: disks passed Mar 17 17:41:22.300255 ignition[801]: Ignition finished successfully Mar 17 17:41:22.304085 systemd[1]: Finished ignition-disks.service - Ignition (disks). Mar 17 17:41:22.304840 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Mar 17 17:41:22.306417 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Mar 17 17:41:22.306745 systemd[1]: Reached target local-fs.target - Local File Systems. Mar 17 17:41:22.307239 systemd[1]: Reached target sysinit.target - System Initialization. Mar 17 17:41:22.307573 systemd[1]: Reached target basic.target - Basic System. Mar 17 17:41:22.324980 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Mar 17 17:41:22.346522 systemd-fsck[812]: ROOT: clean, 14/553520 files, 52654/553472 blocks Mar 17 17:41:22.468376 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Mar 17 17:41:22.473969 systemd[1]: Mounting sysroot.mount - /sysroot... Mar 17 17:41:22.562847 kernel: EXT4-fs (vda9): mounted filesystem 345fc709-8965-4219-b368-16e508c3d632 r/w with ordered data mode. Quota mode: none. Mar 17 17:41:22.563220 systemd[1]: Mounted sysroot.mount - /sysroot. Mar 17 17:41:22.564677 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Mar 17 17:41:22.578885 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Mar 17 17:41:22.580510 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Mar 17 17:41:22.581738 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Mar 17 17:41:22.590349 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/vda6 scanned by mount (820) Mar 17 17:41:22.590369 kernel: BTRFS info (device vda6): first mount of filesystem 7b241d32-136b-4fe3-b105-cecff2b2cf64 Mar 17 17:41:22.590380 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Mar 17 17:41:22.590391 kernel: BTRFS info (device vda6): using free space tree Mar 17 17:41:22.581775 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Mar 17 17:41:22.581806 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Mar 17 17:41:22.595576 kernel: BTRFS info (device vda6): auto enabling async discard Mar 17 17:41:22.589277 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Mar 17 17:41:22.591166 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Mar 17 17:41:22.596108 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Mar 17 17:41:22.627396 initrd-setup-root[844]: cut: /sysroot/etc/passwd: No such file or directory Mar 17 17:41:22.632545 initrd-setup-root[851]: cut: /sysroot/etc/group: No such file or directory Mar 17 17:41:22.636504 initrd-setup-root[858]: cut: /sysroot/etc/shadow: No such file or directory Mar 17 17:41:22.640191 initrd-setup-root[865]: cut: /sysroot/etc/gshadow: No such file or directory Mar 17 17:41:22.743643 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Mar 17 17:41:22.755954 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Mar 17 17:41:22.757773 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Mar 17 17:41:22.763825 kernel: BTRFS info (device vda6): last unmount of filesystem 7b241d32-136b-4fe3-b105-cecff2b2cf64 Mar 17 17:41:22.782492 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Mar 17 17:41:22.811686 ignition[937]: INFO : Ignition 2.20.0 Mar 17 17:41:22.811686 ignition[937]: INFO : Stage: mount Mar 17 17:41:22.813372 ignition[937]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 17 17:41:22.813372 ignition[937]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 17 17:41:22.813372 ignition[937]: INFO : mount: mount passed Mar 17 17:41:22.813372 ignition[937]: INFO : Ignition finished successfully Mar 17 17:41:22.817402 systemd[1]: Finished ignition-mount.service - Ignition (mount). Mar 17 17:41:22.831922 systemd[1]: Starting ignition-files.service - Ignition (files)... Mar 17 17:41:22.983499 systemd[1]: sysroot-oem.mount: Deactivated successfully. Mar 17 17:41:22.995948 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Mar 17 17:41:23.004886 kernel: BTRFS: device label OEM devid 1 transid 16 /dev/vda6 scanned by mount (947) Mar 17 17:41:23.004913 kernel: BTRFS info (device vda6): first mount of filesystem 7b241d32-136b-4fe3-b105-cecff2b2cf64 Mar 17 17:41:23.004924 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Mar 17 17:41:23.006824 kernel: BTRFS info (device vda6): using free space tree Mar 17 17:41:23.009816 kernel: BTRFS info (device vda6): auto enabling async discard Mar 17 17:41:23.010856 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Mar 17 17:41:23.045330 ignition[964]: INFO : Ignition 2.20.0 Mar 17 17:41:23.045330 ignition[964]: INFO : Stage: files Mar 17 17:41:23.051255 ignition[964]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 17 17:41:23.051255 ignition[964]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 17 17:41:23.054020 ignition[964]: DEBUG : files: compiled without relabeling support, skipping Mar 17 17:41:23.055335 ignition[964]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Mar 17 17:41:23.055335 ignition[964]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Mar 17 17:41:23.059434 ignition[964]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Mar 17 17:41:23.061006 ignition[964]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Mar 17 17:41:23.061006 ignition[964]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Mar 17 17:41:23.060132 unknown[964]: wrote ssh authorized keys file for user: core Mar 17 17:41:23.065024 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Mar 17 17:41:23.065024 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-amd64.tar.gz: attempt #1 Mar 17 17:41:23.099523 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Mar 17 17:41:23.420280 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Mar 17 17:41:23.420280 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Mar 17 17:41:23.424608 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Mar 17 17:41:23.424608 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Mar 17 17:41:23.424608 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Mar 17 17:41:23.424608 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Mar 17 17:41:23.424608 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Mar 17 17:41:23.424608 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Mar 17 17:41:23.424608 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Mar 17 17:41:23.424608 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Mar 17 17:41:23.424608 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Mar 17 17:41:23.424608 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" Mar 17 17:41:23.424608 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" Mar 17 17:41:23.424608 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" Mar 17 17:41:23.424608 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.30.1-x86-64.raw: attempt #1 Mar 17 17:41:23.786704 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Mar 17 17:41:23.968062 systemd-networkd[779]: eth0: Gained IPv6LL Mar 17 17:41:24.153495 ignition[964]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" Mar 17 17:41:24.153495 ignition[964]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Mar 17 17:41:24.157514 ignition[964]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Mar 17 17:41:24.159970 ignition[964]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Mar 17 17:41:24.159970 ignition[964]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Mar 17 17:41:24.159970 ignition[964]: INFO : files: op(d): [started] processing unit "coreos-metadata.service" Mar 17 17:41:24.164850 ignition[964]: INFO : files: op(d): op(e): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Mar 17 17:41:24.167006 ignition[964]: INFO : files: op(d): op(e): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Mar 17 17:41:24.167006 ignition[964]: INFO : files: op(d): [finished] processing unit "coreos-metadata.service" Mar 17 17:41:24.167006 ignition[964]: INFO : files: op(f): [started] setting preset to disabled for "coreos-metadata.service" Mar 17 17:41:24.192288 ignition[964]: INFO : files: op(f): op(10): [started] removing enablement symlink(s) for "coreos-metadata.service" Mar 17 17:41:24.197071 ignition[964]: INFO : files: op(f): op(10): [finished] removing enablement symlink(s) for "coreos-metadata.service" Mar 17 17:41:24.198780 ignition[964]: INFO : files: op(f): [finished] setting preset to disabled for "coreos-metadata.service" Mar 17 17:41:24.198780 ignition[964]: INFO : files: op(11): [started] setting preset to enabled for "prepare-helm.service" Mar 17 17:41:24.198780 ignition[964]: INFO : files: op(11): [finished] setting preset to enabled for "prepare-helm.service" Mar 17 17:41:24.198780 ignition[964]: INFO : files: createResultFile: createFiles: op(12): [started] writing file "/sysroot/etc/.ignition-result.json" Mar 17 17:41:24.198780 ignition[964]: INFO : files: createResultFile: createFiles: op(12): [finished] writing file "/sysroot/etc/.ignition-result.json" Mar 17 17:41:24.198780 ignition[964]: INFO : files: files passed Mar 17 17:41:24.198780 ignition[964]: INFO : Ignition finished successfully Mar 17 17:41:24.199710 systemd[1]: Finished ignition-files.service - Ignition (files). Mar 17 17:41:24.209975 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Mar 17 17:41:24.212836 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Mar 17 17:41:24.215069 systemd[1]: ignition-quench.service: Deactivated successfully. Mar 17 17:41:24.215195 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Mar 17 17:41:24.222699 initrd-setup-root-after-ignition[993]: grep: /sysroot/oem/oem-release: No such file or directory Mar 17 17:41:24.225128 initrd-setup-root-after-ignition[995]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Mar 17 17:41:24.226767 initrd-setup-root-after-ignition[995]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Mar 17 17:41:24.228294 initrd-setup-root-after-ignition[999]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Mar 17 17:41:24.227947 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Mar 17 17:41:24.229746 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Mar 17 17:41:24.241938 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Mar 17 17:41:24.264741 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Mar 17 17:41:24.264884 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Mar 17 17:41:24.265647 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Mar 17 17:41:24.268604 systemd[1]: Reached target initrd.target - Initrd Default Target. Mar 17 17:41:24.271190 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Mar 17 17:41:24.273319 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Mar 17 17:41:24.291435 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Mar 17 17:41:24.306967 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Mar 17 17:41:24.318414 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Mar 17 17:41:24.320736 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 17 17:41:24.321297 systemd[1]: Stopped target timers.target - Timer Units. Mar 17 17:41:24.321574 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Mar 17 17:41:24.321699 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Mar 17 17:41:24.326715 systemd[1]: Stopped target initrd.target - Initrd Default Target. Mar 17 17:41:24.328712 systemd[1]: Stopped target basic.target - Basic System. Mar 17 17:41:24.329254 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Mar 17 17:41:24.329582 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Mar 17 17:41:24.333690 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Mar 17 17:41:24.334137 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Mar 17 17:41:24.337888 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Mar 17 17:41:24.340502 systemd[1]: Stopped target sysinit.target - System Initialization. Mar 17 17:41:24.342676 systemd[1]: Stopped target local-fs.target - Local File Systems. Mar 17 17:41:24.344505 systemd[1]: Stopped target swap.target - Swaps. Mar 17 17:41:24.346212 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Mar 17 17:41:24.346369 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Mar 17 17:41:24.347748 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Mar 17 17:41:24.350832 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 17 17:41:24.351317 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Mar 17 17:41:24.351425 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 17 17:41:24.351850 systemd[1]: dracut-initqueue.service: Deactivated successfully. Mar 17 17:41:24.351951 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Mar 17 17:41:24.358212 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Mar 17 17:41:24.358316 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Mar 17 17:41:24.358820 systemd[1]: Stopped target paths.target - Path Units. Mar 17 17:41:24.361880 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Mar 17 17:41:24.365937 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 17 17:41:24.368980 systemd[1]: Stopped target slices.target - Slice Units. Mar 17 17:41:24.371040 systemd[1]: Stopped target sockets.target - Socket Units. Mar 17 17:41:24.371606 systemd[1]: iscsid.socket: Deactivated successfully. Mar 17 17:41:24.371757 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Mar 17 17:41:24.373476 systemd[1]: iscsiuio.socket: Deactivated successfully. Mar 17 17:41:24.373590 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Mar 17 17:41:24.375296 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Mar 17 17:41:24.375450 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Mar 17 17:41:24.378205 systemd[1]: ignition-files.service: Deactivated successfully. Mar 17 17:41:24.378341 systemd[1]: Stopped ignition-files.service - Ignition (files). Mar 17 17:41:24.388960 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Mar 17 17:41:24.389431 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Mar 17 17:41:24.389581 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Mar 17 17:41:24.390901 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Mar 17 17:41:24.395487 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Mar 17 17:41:24.397671 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Mar 17 17:41:24.400105 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Mar 17 17:41:24.400257 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Mar 17 17:41:24.405173 ignition[1019]: INFO : Ignition 2.20.0 Mar 17 17:41:24.405173 ignition[1019]: INFO : Stage: umount Mar 17 17:41:24.405173 ignition[1019]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 17 17:41:24.405173 ignition[1019]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 17 17:41:24.405173 ignition[1019]: INFO : umount: umount passed Mar 17 17:41:24.405173 ignition[1019]: INFO : Ignition finished successfully Mar 17 17:41:24.405358 systemd[1]: ignition-mount.service: Deactivated successfully. Mar 17 17:41:24.405477 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Mar 17 17:41:24.407827 systemd[1]: initrd-cleanup.service: Deactivated successfully. Mar 17 17:41:24.407939 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Mar 17 17:41:24.409087 systemd[1]: Stopped target network.target - Network. Mar 17 17:41:24.411745 systemd[1]: ignition-disks.service: Deactivated successfully. Mar 17 17:41:24.411824 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Mar 17 17:41:24.414469 systemd[1]: ignition-kargs.service: Deactivated successfully. Mar 17 17:41:24.414517 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Mar 17 17:41:24.415021 systemd[1]: ignition-setup.service: Deactivated successfully. Mar 17 17:41:24.415065 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Mar 17 17:41:24.415341 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Mar 17 17:41:24.415381 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Mar 17 17:41:24.415843 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Mar 17 17:41:24.421717 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Mar 17 17:41:24.428857 systemd-networkd[779]: eth0: DHCPv6 lease lost Mar 17 17:41:24.432469 systemd[1]: systemd-networkd.service: Deactivated successfully. Mar 17 17:41:24.432599 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Mar 17 17:41:24.433852 systemd[1]: systemd-resolved.service: Deactivated successfully. Mar 17 17:41:24.433965 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Mar 17 17:41:24.438041 systemd[1]: systemd-networkd.socket: Deactivated successfully. Mar 17 17:41:24.438096 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Mar 17 17:41:24.446962 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Mar 17 17:41:24.449340 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Mar 17 17:41:24.449410 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Mar 17 17:41:24.451726 systemd[1]: systemd-sysctl.service: Deactivated successfully. Mar 17 17:41:24.453127 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Mar 17 17:41:24.456094 systemd[1]: systemd-modules-load.service: Deactivated successfully. Mar 17 17:41:24.456149 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Mar 17 17:41:24.459341 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Mar 17 17:41:24.460376 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 17 17:41:24.462969 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 17 17:41:24.466172 systemd[1]: sysroot-boot.mount: Deactivated successfully. Mar 17 17:41:24.476792 systemd[1]: systemd-udevd.service: Deactivated successfully. Mar 17 17:41:24.477013 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 17 17:41:24.477926 systemd[1]: network-cleanup.service: Deactivated successfully. Mar 17 17:41:24.478029 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Mar 17 17:41:24.481714 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Mar 17 17:41:24.481792 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Mar 17 17:41:24.483760 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Mar 17 17:41:24.483816 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Mar 17 17:41:24.485778 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Mar 17 17:41:24.485843 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Mar 17 17:41:24.488125 systemd[1]: dracut-cmdline.service: Deactivated successfully. Mar 17 17:41:24.488182 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Mar 17 17:41:24.490311 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Mar 17 17:41:24.490375 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 17 17:41:24.502086 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Mar 17 17:41:24.504852 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Mar 17 17:41:24.504946 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 17 17:41:24.507672 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 17 17:41:24.507723 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 17 17:41:24.510814 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Mar 17 17:41:24.510927 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Mar 17 17:41:24.653230 systemd[1]: sysroot-boot.service: Deactivated successfully. Mar 17 17:41:24.653490 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Mar 17 17:41:24.655527 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Mar 17 17:41:24.656702 systemd[1]: initrd-setup-root.service: Deactivated successfully. Mar 17 17:41:24.656764 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Mar 17 17:41:24.674073 systemd[1]: Starting initrd-switch-root.service - Switch Root... Mar 17 17:41:24.682982 systemd[1]: Switching root. Mar 17 17:41:24.713159 systemd-journald[192]: Journal stopped Mar 17 17:41:25.853056 systemd-journald[192]: Received SIGTERM from PID 1 (systemd). Mar 17 17:41:25.853149 kernel: SELinux: policy capability network_peer_controls=1 Mar 17 17:41:25.853168 kernel: SELinux: policy capability open_perms=1 Mar 17 17:41:25.853183 kernel: SELinux: policy capability extended_socket_class=1 Mar 17 17:41:25.853198 kernel: SELinux: policy capability always_check_network=0 Mar 17 17:41:25.853213 kernel: SELinux: policy capability cgroup_seclabel=1 Mar 17 17:41:25.853227 kernel: SELinux: policy capability nnp_nosuid_transition=1 Mar 17 17:41:25.853242 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Mar 17 17:41:25.853268 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Mar 17 17:41:25.853283 kernel: audit: type=1403 audit(1742233285.123:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Mar 17 17:41:25.853298 systemd[1]: Successfully loaded SELinux policy in 41.960ms. Mar 17 17:41:25.853323 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 12.538ms. Mar 17 17:41:25.853340 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Mar 17 17:41:25.853355 systemd[1]: Detected virtualization kvm. Mar 17 17:41:25.853371 systemd[1]: Detected architecture x86-64. Mar 17 17:41:25.853386 systemd[1]: Detected first boot. Mar 17 17:41:25.853401 systemd[1]: Initializing machine ID from VM UUID. Mar 17 17:41:25.853422 zram_generator::config[1066]: No configuration found. Mar 17 17:41:25.853444 systemd[1]: Populated /etc with preset unit settings. Mar 17 17:41:25.853464 systemd[1]: initrd-switch-root.service: Deactivated successfully. Mar 17 17:41:25.853479 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Mar 17 17:41:25.853495 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Mar 17 17:41:25.853511 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Mar 17 17:41:25.853528 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Mar 17 17:41:25.853544 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Mar 17 17:41:25.853570 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Mar 17 17:41:25.853594 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Mar 17 17:41:25.853610 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Mar 17 17:41:25.853626 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Mar 17 17:41:25.853642 systemd[1]: Created slice user.slice - User and Session Slice. Mar 17 17:41:25.853657 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 17 17:41:25.853673 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 17 17:41:25.853689 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Mar 17 17:41:25.853711 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Mar 17 17:41:25.853727 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Mar 17 17:41:25.853743 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Mar 17 17:41:25.853758 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Mar 17 17:41:25.853774 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 17 17:41:25.853789 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Mar 17 17:41:25.853817 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Mar 17 17:41:25.853833 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Mar 17 17:41:25.853855 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Mar 17 17:41:25.853871 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 17 17:41:25.853886 systemd[1]: Reached target remote-fs.target - Remote File Systems. Mar 17 17:41:25.853904 systemd[1]: Reached target slices.target - Slice Units. Mar 17 17:41:25.853919 systemd[1]: Reached target swap.target - Swaps. Mar 17 17:41:25.853935 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Mar 17 17:41:25.853951 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Mar 17 17:41:25.853966 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Mar 17 17:41:25.853982 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Mar 17 17:41:25.854003 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Mar 17 17:41:25.854018 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Mar 17 17:41:25.854034 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Mar 17 17:41:25.854049 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Mar 17 17:41:25.854065 systemd[1]: Mounting media.mount - External Media Directory... Mar 17 17:41:25.854080 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 17 17:41:25.854096 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Mar 17 17:41:25.854112 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Mar 17 17:41:25.854128 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Mar 17 17:41:25.854149 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Mar 17 17:41:25.854165 systemd[1]: Reached target machines.target - Containers. Mar 17 17:41:25.854181 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Mar 17 17:41:25.854197 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 17 17:41:25.854213 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Mar 17 17:41:25.854229 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Mar 17 17:41:25.854245 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 17 17:41:25.854261 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Mar 17 17:41:25.854282 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 17 17:41:25.854296 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Mar 17 17:41:25.854312 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 17 17:41:25.854329 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Mar 17 17:41:25.854344 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Mar 17 17:41:25.854361 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Mar 17 17:41:25.854377 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Mar 17 17:41:25.854392 systemd[1]: Stopped systemd-fsck-usr.service. Mar 17 17:41:25.854408 systemd[1]: Starting systemd-journald.service - Journal Service... Mar 17 17:41:25.854429 kernel: loop: module loaded Mar 17 17:41:25.854444 kernel: fuse: init (API version 7.39) Mar 17 17:41:25.854459 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Mar 17 17:41:25.854475 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Mar 17 17:41:25.854510 systemd-journald[1129]: Collecting audit messages is disabled. Mar 17 17:41:25.854538 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Mar 17 17:41:25.854554 systemd-journald[1129]: Journal started Mar 17 17:41:25.854598 systemd-journald[1129]: Runtime Journal (/run/log/journal/1490813c1f8f4ac78a5a7071c37f2d7a) is 6.0M, max 48.3M, 42.2M free. Mar 17 17:41:25.639769 systemd[1]: Queued start job for default target multi-user.target. Mar 17 17:41:25.659212 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Mar 17 17:41:25.659648 systemd[1]: systemd-journald.service: Deactivated successfully. Mar 17 17:41:25.864140 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Mar 17 17:41:25.865893 systemd[1]: verity-setup.service: Deactivated successfully. Mar 17 17:41:25.865925 systemd[1]: Stopped verity-setup.service. Mar 17 17:41:25.868821 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 17 17:41:25.870936 kernel: ACPI: bus type drm_connector registered Mar 17 17:41:25.871037 systemd[1]: Started systemd-journald.service - Journal Service. Mar 17 17:41:25.873042 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Mar 17 17:41:25.874375 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Mar 17 17:41:25.879495 systemd[1]: Mounted media.mount - External Media Directory. Mar 17 17:41:25.880656 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Mar 17 17:41:25.881905 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Mar 17 17:41:25.883212 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Mar 17 17:41:25.884477 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Mar 17 17:41:25.886192 systemd[1]: modprobe@configfs.service: Deactivated successfully. Mar 17 17:41:25.886366 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Mar 17 17:41:25.887887 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 17 17:41:25.888051 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 17 17:41:25.889543 systemd[1]: modprobe@drm.service: Deactivated successfully. Mar 17 17:41:25.889726 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Mar 17 17:41:25.893747 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 17 17:41:25.893926 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 17 17:41:25.895569 systemd[1]: modprobe@fuse.service: Deactivated successfully. Mar 17 17:41:25.895753 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Mar 17 17:41:25.897262 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 17 17:41:25.897424 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 17 17:41:25.899031 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Mar 17 17:41:25.900557 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Mar 17 17:41:25.902128 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Mar 17 17:41:25.917547 systemd[1]: Reached target network-pre.target - Preparation for Network. Mar 17 17:41:25.930284 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Mar 17 17:41:25.932860 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Mar 17 17:41:25.934080 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Mar 17 17:41:25.934113 systemd[1]: Reached target local-fs.target - Local File Systems. Mar 17 17:41:25.936107 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Mar 17 17:41:25.938435 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Mar 17 17:41:25.941853 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Mar 17 17:41:25.944176 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 17 17:41:25.947044 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Mar 17 17:41:25.951430 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Mar 17 17:41:25.953488 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 17 17:41:25.956940 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Mar 17 17:41:25.958602 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 17 17:41:25.963191 systemd-journald[1129]: Time spent on flushing to /var/log/journal/1490813c1f8f4ac78a5a7071c37f2d7a is 33.093ms for 1038 entries. Mar 17 17:41:25.963191 systemd-journald[1129]: System Journal (/var/log/journal/1490813c1f8f4ac78a5a7071c37f2d7a) is 8.0M, max 195.6M, 187.6M free. Mar 17 17:41:26.142087 systemd-journald[1129]: Received client request to flush runtime journal. Mar 17 17:41:26.142148 kernel: loop0: detected capacity change from 0 to 138184 Mar 17 17:41:26.142181 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Mar 17 17:41:26.142199 kernel: loop1: detected capacity change from 0 to 140992 Mar 17 17:41:25.963959 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 17 17:41:25.976320 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Mar 17 17:41:25.989664 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Mar 17 17:41:25.991521 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Mar 17 17:41:25.993151 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Mar 17 17:41:26.004099 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Mar 17 17:41:26.005889 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Mar 17 17:41:26.026022 systemd[1]: Starting systemd-sysusers.service - Create System Users... Mar 17 17:41:26.043263 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Mar 17 17:41:26.045223 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 17 17:41:26.057917 udevadm[1190]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. Mar 17 17:41:26.064917 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Mar 17 17:41:26.073826 systemd[1]: Finished systemd-sysusers.service - Create System Users. Mar 17 17:41:26.076845 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Mar 17 17:41:26.087993 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Mar 17 17:41:26.098662 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Mar 17 17:41:26.144214 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Mar 17 17:41:26.175143 systemd-tmpfiles[1196]: ACLs are not supported, ignoring. Mar 17 17:41:26.175167 systemd-tmpfiles[1196]: ACLs are not supported, ignoring. Mar 17 17:41:26.181587 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 17 17:41:26.213853 kernel: loop2: detected capacity change from 0 to 210664 Mar 17 17:41:26.254515 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Mar 17 17:41:26.255403 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Mar 17 17:41:26.263825 kernel: loop3: detected capacity change from 0 to 138184 Mar 17 17:41:26.275009 kernel: loop4: detected capacity change from 0 to 140992 Mar 17 17:41:26.286835 kernel: loop5: detected capacity change from 0 to 210664 Mar 17 17:41:26.293153 (sd-merge)[1205]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. Mar 17 17:41:26.294705 (sd-merge)[1205]: Merged extensions into '/usr'. Mar 17 17:41:26.299322 systemd[1]: Reloading requested from client PID 1179 ('systemd-sysext') (unit systemd-sysext.service)... Mar 17 17:41:26.299340 systemd[1]: Reloading... Mar 17 17:41:26.361336 zram_generator::config[1234]: No configuration found. Mar 17 17:41:26.411147 ldconfig[1174]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Mar 17 17:41:26.487015 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 17 17:41:26.536461 systemd[1]: Reloading finished in 236 ms. Mar 17 17:41:26.572788 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Mar 17 17:41:26.574342 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Mar 17 17:41:26.589969 systemd[1]: Starting ensure-sysext.service... Mar 17 17:41:26.592215 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Mar 17 17:41:26.596584 systemd[1]: Reloading requested from client PID 1268 ('systemctl') (unit ensure-sysext.service)... Mar 17 17:41:26.596597 systemd[1]: Reloading... Mar 17 17:41:26.616104 systemd-tmpfiles[1269]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Mar 17 17:41:26.616839 systemd-tmpfiles[1269]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Mar 17 17:41:26.618096 systemd-tmpfiles[1269]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Mar 17 17:41:26.618475 systemd-tmpfiles[1269]: ACLs are not supported, ignoring. Mar 17 17:41:26.618575 systemd-tmpfiles[1269]: ACLs are not supported, ignoring. Mar 17 17:41:26.625723 systemd-tmpfiles[1269]: Detected autofs mount point /boot during canonicalization of boot. Mar 17 17:41:26.625735 systemd-tmpfiles[1269]: Skipping /boot Mar 17 17:41:26.641283 systemd-tmpfiles[1269]: Detected autofs mount point /boot during canonicalization of boot. Mar 17 17:41:26.641380 systemd-tmpfiles[1269]: Skipping /boot Mar 17 17:41:26.649827 zram_generator::config[1295]: No configuration found. Mar 17 17:41:26.774730 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 17 17:41:26.824629 systemd[1]: Reloading finished in 227 ms. Mar 17 17:41:26.843057 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Mar 17 17:41:26.856494 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 17 17:41:26.866676 systemd[1]: Starting audit-rules.service - Load Audit Rules... Mar 17 17:41:26.869332 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Mar 17 17:41:26.871790 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Mar 17 17:41:26.876699 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Mar 17 17:41:26.879537 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 17 17:41:26.884093 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Mar 17 17:41:26.889574 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 17 17:41:26.889776 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 17 17:41:26.891331 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 17 17:41:26.894968 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 17 17:41:26.897319 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 17 17:41:26.898654 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 17 17:41:26.902060 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Mar 17 17:41:26.903251 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 17 17:41:26.905397 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 17 17:41:26.905619 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 17 17:41:26.910524 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 17 17:41:26.910742 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 17 17:41:26.916206 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 17 17:41:26.916662 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 17 17:41:26.920065 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Mar 17 17:41:26.924564 systemd-udevd[1338]: Using default interface naming scheme 'v255'. Mar 17 17:41:26.928461 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Mar 17 17:41:26.933171 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 17 17:41:26.933461 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 17 17:41:26.939538 augenrules[1369]: No rules Mar 17 17:41:26.940031 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 17 17:41:26.942670 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Mar 17 17:41:26.944790 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 17 17:41:26.949277 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 17 17:41:26.950631 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 17 17:41:26.954068 systemd[1]: Starting systemd-update-done.service - Update is Completed... Mar 17 17:41:26.955258 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Mar 17 17:41:26.956465 systemd[1]: audit-rules.service: Deactivated successfully. Mar 17 17:41:26.956742 systemd[1]: Finished audit-rules.service - Load Audit Rules. Mar 17 17:41:26.958356 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 17 17:41:26.958583 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 17 17:41:26.960465 systemd[1]: modprobe@drm.service: Deactivated successfully. Mar 17 17:41:26.960724 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Mar 17 17:41:26.962371 systemd[1]: Started systemd-userdbd.service - User Database Manager. Mar 17 17:41:26.964205 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 17 17:41:26.966273 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 17 17:41:26.966485 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 17 17:41:26.968394 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 17 17:41:26.968585 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 17 17:41:26.976572 systemd[1]: Finished ensure-sysext.service. Mar 17 17:41:26.983500 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Mar 17 17:41:26.986757 systemd[1]: Finished systemd-update-done.service - Update is Completed. Mar 17 17:41:27.000975 systemd[1]: Starting systemd-networkd.service - Network Configuration... Mar 17 17:41:27.004124 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 17 17:41:27.004199 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 17 17:41:27.006243 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Mar 17 17:41:27.007442 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Mar 17 17:41:27.026044 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Mar 17 17:41:27.056859 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 40 scanned by (udev-worker) (1384) Mar 17 17:41:27.068722 systemd-resolved[1337]: Positive Trust Anchors: Mar 17 17:41:27.068745 systemd-resolved[1337]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 17 17:41:27.068778 systemd-resolved[1337]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Mar 17 17:41:27.080741 systemd-resolved[1337]: Defaulting to hostname 'linux'. Mar 17 17:41:27.083094 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Mar 17 17:41:27.084491 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Mar 17 17:41:27.088212 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Mar 17 17:41:27.092814 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input2 Mar 17 17:41:27.096735 systemd-networkd[1410]: lo: Link UP Mar 17 17:41:27.101364 kernel: ACPI: button: Power Button [PWRF] Mar 17 17:41:27.096751 systemd-networkd[1410]: lo: Gained carrier Mar 17 17:41:27.099867 systemd-networkd[1410]: Enumeration completed Mar 17 17:41:27.101002 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Mar 17 17:41:27.102043 systemd-networkd[1410]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 17 17:41:27.102056 systemd-networkd[1410]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 17 17:41:27.102290 systemd[1]: Started systemd-networkd.service - Network Configuration. Mar 17 17:41:27.103588 systemd-networkd[1410]: eth0: Link UP Mar 17 17:41:27.103601 systemd-networkd[1410]: eth0: Gained carrier Mar 17 17:41:27.103613 systemd-networkd[1410]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 17 17:41:27.107285 systemd[1]: Reached target network.target - Network. Mar 17 17:41:27.123872 systemd-networkd[1410]: eth0: DHCPv4 address 10.0.0.61/16, gateway 10.0.0.1 acquired from 10.0.0.1 Mar 17 17:41:27.123968 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Mar 17 17:41:27.125567 systemd-timesyncd[1411]: Network configuration changed, trying to establish connection. Mar 17 17:41:27.126736 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Mar 17 17:41:27.586839 systemd-resolved[1337]: Clock change detected. Flushing caches. Mar 17 17:41:27.586979 systemd-timesyncd[1411]: Contacted time server 10.0.0.1:123 (10.0.0.1). Mar 17 17:41:27.587073 systemd-timesyncd[1411]: Initial clock synchronization to Mon 2025-03-17 17:41:27.586791 UTC. Mar 17 17:41:27.591062 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input3 Mar 17 17:41:27.591093 kernel: i801_smbus 0000:00:1f.3: Enabling SMBus device Mar 17 17:41:27.591837 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Mar 17 17:41:27.592000 kernel: i2c i2c-0: 1/1 memory slots populated (from DMI) Mar 17 17:41:27.592186 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Mar 17 17:41:27.588689 systemd[1]: Reached target time-set.target - System Time Set. Mar 17 17:41:27.596772 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Mar 17 17:41:27.625665 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 17 17:41:27.631540 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 17 17:41:27.631916 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 17 17:41:27.642581 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 17 17:41:27.694384 kernel: mousedev: PS/2 mouse device common for all mice Mar 17 17:41:27.705648 kernel: kvm_amd: TSC scaling supported Mar 17 17:41:27.705692 kernel: kvm_amd: Nested Virtualization enabled Mar 17 17:41:27.705705 kernel: kvm_amd: Nested Paging enabled Mar 17 17:41:27.705717 kernel: kvm_amd: LBR virtualization supported Mar 17 17:41:27.706907 kernel: kvm_amd: Virtual VMLOAD VMSAVE supported Mar 17 17:41:27.706925 kernel: kvm_amd: Virtual GIF supported Mar 17 17:41:27.731394 kernel: EDAC MC: Ver: 3.0.0 Mar 17 17:41:27.738400 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 17 17:41:27.760785 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Mar 17 17:41:27.778719 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Mar 17 17:41:27.787061 lvm[1439]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Mar 17 17:41:27.817616 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Mar 17 17:41:27.823628 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Mar 17 17:41:27.824778 systemd[1]: Reached target sysinit.target - System Initialization. Mar 17 17:41:27.825969 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Mar 17 17:41:27.827256 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Mar 17 17:41:27.828861 systemd[1]: Started logrotate.timer - Daily rotation of log files. Mar 17 17:41:27.830122 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Mar 17 17:41:27.831402 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Mar 17 17:41:27.832699 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Mar 17 17:41:27.832726 systemd[1]: Reached target paths.target - Path Units. Mar 17 17:41:27.833660 systemd[1]: Reached target timers.target - Timer Units. Mar 17 17:41:27.835395 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Mar 17 17:41:27.838218 systemd[1]: Starting docker.socket - Docker Socket for the API... Mar 17 17:41:27.852051 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Mar 17 17:41:27.861959 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Mar 17 17:41:27.863614 systemd[1]: Listening on docker.socket - Docker Socket for the API. Mar 17 17:41:27.864784 systemd[1]: Reached target sockets.target - Socket Units. Mar 17 17:41:27.865782 systemd[1]: Reached target basic.target - Basic System. Mar 17 17:41:27.866767 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Mar 17 17:41:27.866795 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Mar 17 17:41:27.867793 systemd[1]: Starting containerd.service - containerd container runtime... Mar 17 17:41:27.869874 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Mar 17 17:41:27.871490 lvm[1443]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Mar 17 17:41:27.873794 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Mar 17 17:41:27.877304 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Mar 17 17:41:27.878519 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Mar 17 17:41:27.879922 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Mar 17 17:41:27.883473 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Mar 17 17:41:27.886210 jq[1446]: false Mar 17 17:41:27.887469 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Mar 17 17:41:27.889928 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Mar 17 17:41:27.895536 systemd[1]: Starting systemd-logind.service - User Login Management... Mar 17 17:41:27.897068 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Mar 17 17:41:27.897601 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Mar 17 17:41:27.898571 systemd[1]: Starting update-engine.service - Update Engine... Mar 17 17:41:27.902712 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Mar 17 17:41:27.902769 dbus-daemon[1445]: [system] SELinux support is enabled Mar 17 17:41:27.905143 systemd[1]: Started dbus.service - D-Bus System Message Bus. Mar 17 17:41:27.912450 extend-filesystems[1447]: Found loop3 Mar 17 17:41:27.912450 extend-filesystems[1447]: Found loop4 Mar 17 17:41:27.912450 extend-filesystems[1447]: Found loop5 Mar 17 17:41:27.912450 extend-filesystems[1447]: Found sr0 Mar 17 17:41:27.912450 extend-filesystems[1447]: Found vda Mar 17 17:41:27.912450 extend-filesystems[1447]: Found vda1 Mar 17 17:41:27.912450 extend-filesystems[1447]: Found vda2 Mar 17 17:41:27.912450 extend-filesystems[1447]: Found vda3 Mar 17 17:41:27.912450 extend-filesystems[1447]: Found usr Mar 17 17:41:27.912450 extend-filesystems[1447]: Found vda4 Mar 17 17:41:27.912450 extend-filesystems[1447]: Found vda6 Mar 17 17:41:27.912450 extend-filesystems[1447]: Found vda7 Mar 17 17:41:27.910461 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Mar 17 17:41:27.939823 update_engine[1454]: I20250317 17:41:27.919126 1454 main.cc:92] Flatcar Update Engine starting Mar 17 17:41:27.939823 update_engine[1454]: I20250317 17:41:27.920303 1454 update_check_scheduler.cc:74] Next update check in 3m43s Mar 17 17:41:27.940140 extend-filesystems[1447]: Found vda9 Mar 17 17:41:27.940140 extend-filesystems[1447]: Checking size of /dev/vda9 Mar 17 17:41:27.920892 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Mar 17 17:41:27.921158 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Mar 17 17:41:27.922172 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Mar 17 17:41:27.945113 jq[1456]: true Mar 17 17:41:27.922427 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Mar 17 17:41:27.928964 systemd[1]: motdgen.service: Deactivated successfully. Mar 17 17:41:27.929193 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Mar 17 17:41:27.932065 (ntainerd)[1467]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Mar 17 17:41:27.942146 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Mar 17 17:41:27.942181 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Mar 17 17:41:27.943705 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Mar 17 17:41:27.943735 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Mar 17 17:41:27.949958 extend-filesystems[1447]: Resized partition /dev/vda9 Mar 17 17:41:27.950260 systemd[1]: Started update-engine.service - Update Engine. Mar 17 17:41:27.957611 extend-filesystems[1480]: resize2fs 1.47.1 (20-May-2024) Mar 17 17:41:27.959728 systemd[1]: Started locksmithd.service - Cluster reboot manager. Mar 17 17:41:27.964887 tar[1466]: linux-amd64/helm Mar 17 17:41:27.967521 jq[1471]: true Mar 17 17:41:27.970479 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 40 scanned by (udev-worker) (1407) Mar 17 17:41:27.980417 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Mar 17 17:41:27.979836 systemd-logind[1453]: Watching system buttons on /dev/input/event1 (Power Button) Mar 17 17:41:27.979856 systemd-logind[1453]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Mar 17 17:41:27.982284 systemd-logind[1453]: New seat seat0. Mar 17 17:41:27.984054 systemd[1]: Started systemd-logind.service - User Login Management. Mar 17 17:41:28.045682 locksmithd[1482]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Mar 17 17:41:28.105374 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Mar 17 17:41:28.110050 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Mar 17 17:41:28.277497 extend-filesystems[1480]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Mar 17 17:41:28.277497 extend-filesystems[1480]: old_desc_blocks = 1, new_desc_blocks = 1 Mar 17 17:41:28.277497 extend-filesystems[1480]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Mar 17 17:41:28.285318 extend-filesystems[1447]: Resized filesystem in /dev/vda9 Mar 17 17:41:28.279208 systemd[1]: extend-filesystems.service: Deactivated successfully. Mar 17 17:41:28.286307 sshd_keygen[1470]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Mar 17 17:41:28.279443 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Mar 17 17:41:28.305399 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Mar 17 17:41:28.322665 systemd[1]: Starting issuegen.service - Generate /run/issue... Mar 17 17:41:28.331527 systemd[1]: Started sshd@0-10.0.0.61:22-10.0.0.1:49346.service - OpenSSH per-connection server daemon (10.0.0.1:49346). Mar 17 17:41:28.333456 systemd[1]: issuegen.service: Deactivated successfully. Mar 17 17:41:28.334051 systemd[1]: Finished issuegen.service - Generate /run/issue. Mar 17 17:41:28.341464 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Mar 17 17:41:28.350248 containerd[1467]: time="2025-03-17T17:41:28.350161533Z" level=info msg="starting containerd" revision=9b2ad7760328148397346d10c7b2004271249db4 version=v1.7.23 Mar 17 17:41:28.359075 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Mar 17 17:41:28.373886 systemd[1]: Started getty@tty1.service - Getty on tty1. Mar 17 17:41:28.374160 containerd[1467]: time="2025-03-17T17:41:28.373858149Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Mar 17 17:41:28.376760 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Mar 17 17:41:28.378530 systemd[1]: Reached target getty.target - Login Prompts. Mar 17 17:41:28.380680 bash[1499]: Updated "/home/core/.ssh/authorized_keys" Mar 17 17:41:28.380805 containerd[1467]: time="2025-03-17T17:41:28.380620225Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.83-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Mar 17 17:41:28.380805 containerd[1467]: time="2025-03-17T17:41:28.380658517Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Mar 17 17:41:28.380805 containerd[1467]: time="2025-03-17T17:41:28.380677573Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Mar 17 17:41:28.381234 containerd[1467]: time="2025-03-17T17:41:28.380875534Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Mar 17 17:41:28.381234 containerd[1467]: time="2025-03-17T17:41:28.380899660Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Mar 17 17:41:28.381234 containerd[1467]: time="2025-03-17T17:41:28.380966886Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Mar 17 17:41:28.381234 containerd[1467]: time="2025-03-17T17:41:28.380980311Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Mar 17 17:41:28.381234 containerd[1467]: time="2025-03-17T17:41:28.381180947Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Mar 17 17:41:28.381234 containerd[1467]: time="2025-03-17T17:41:28.381194282Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Mar 17 17:41:28.381234 containerd[1467]: time="2025-03-17T17:41:28.381207687Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Mar 17 17:41:28.381234 containerd[1467]: time="2025-03-17T17:41:28.381217115Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Mar 17 17:41:28.381437 containerd[1467]: time="2025-03-17T17:41:28.381309408Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Mar 17 17:41:28.382407 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Mar 17 17:41:28.383198 containerd[1467]: time="2025-03-17T17:41:28.383170319Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Mar 17 17:41:28.383325 containerd[1467]: time="2025-03-17T17:41:28.383301055Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Mar 17 17:41:28.383325 containerd[1467]: time="2025-03-17T17:41:28.383320682Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Mar 17 17:41:28.383529 containerd[1467]: time="2025-03-17T17:41:28.383438613Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Mar 17 17:41:28.383529 containerd[1467]: time="2025-03-17T17:41:28.383496672Z" level=info msg="metadata content store policy set" policy=shared Mar 17 17:41:28.384556 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Mar 17 17:41:28.441320 sshd[1524]: Accepted publickey for core from 10.0.0.1 port 49346 ssh2: RSA SHA256:j201F9FRK1q3ChnxQf0adNdYppDp+g37vmaXPvsVhek Mar 17 17:41:28.444021 sshd-session[1524]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 17 17:41:28.453532 systemd-logind[1453]: New session 1 of user core. Mar 17 17:41:28.454811 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Mar 17 17:41:28.464616 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Mar 17 17:41:28.472977 containerd[1467]: time="2025-03-17T17:41:28.472943100Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Mar 17 17:41:28.473030 containerd[1467]: time="2025-03-17T17:41:28.473008933Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Mar 17 17:41:28.473030 containerd[1467]: time="2025-03-17T17:41:28.473024102Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Mar 17 17:41:28.473067 containerd[1467]: time="2025-03-17T17:41:28.473040202Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Mar 17 17:41:28.473067 containerd[1467]: time="2025-03-17T17:41:28.473054048Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Mar 17 17:41:28.473238 containerd[1467]: time="2025-03-17T17:41:28.473217765Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Mar 17 17:41:28.473465 containerd[1467]: time="2025-03-17T17:41:28.473445963Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Mar 17 17:41:28.473568 containerd[1467]: time="2025-03-17T17:41:28.473549207Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Mar 17 17:41:28.473568 containerd[1467]: time="2025-03-17T17:41:28.473565808Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Mar 17 17:41:28.473619 containerd[1467]: time="2025-03-17T17:41:28.473578612Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Mar 17 17:41:28.473619 containerd[1467]: time="2025-03-17T17:41:28.473591356Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Mar 17 17:41:28.473619 containerd[1467]: time="2025-03-17T17:41:28.473604320Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Mar 17 17:41:28.473619 containerd[1467]: time="2025-03-17T17:41:28.473615591Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Mar 17 17:41:28.473684 containerd[1467]: time="2025-03-17T17:41:28.473628015Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Mar 17 17:41:28.473684 containerd[1467]: time="2025-03-17T17:41:28.473642181Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Mar 17 17:41:28.473684 containerd[1467]: time="2025-03-17T17:41:28.473655476Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Mar 17 17:41:28.473684 containerd[1467]: time="2025-03-17T17:41:28.473667378Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Mar 17 17:41:28.473684 containerd[1467]: time="2025-03-17T17:41:28.473678159Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Mar 17 17:41:28.473772 containerd[1467]: time="2025-03-17T17:41:28.473696483Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Mar 17 17:41:28.473772 containerd[1467]: time="2025-03-17T17:41:28.473708776Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Mar 17 17:41:28.473772 containerd[1467]: time="2025-03-17T17:41:28.473721049Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Mar 17 17:41:28.473772 containerd[1467]: time="2025-03-17T17:41:28.473732861Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Mar 17 17:41:28.473772 containerd[1467]: time="2025-03-17T17:41:28.473744774Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Mar 17 17:41:28.473772 containerd[1467]: time="2025-03-17T17:41:28.473761816Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Mar 17 17:41:28.473772 containerd[1467]: time="2025-03-17T17:41:28.473772896Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Mar 17 17:41:28.473899 containerd[1467]: time="2025-03-17T17:41:28.473785480Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Mar 17 17:41:28.473899 containerd[1467]: time="2025-03-17T17:41:28.473797142Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Mar 17 17:41:28.473899 containerd[1467]: time="2025-03-17T17:41:28.473810918Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Mar 17 17:41:28.473899 containerd[1467]: time="2025-03-17T17:41:28.473821818Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Mar 17 17:41:28.473899 containerd[1467]: time="2025-03-17T17:41:28.473833250Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Mar 17 17:41:28.473899 containerd[1467]: time="2025-03-17T17:41:28.473845422Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Mar 17 17:41:28.473899 containerd[1467]: time="2025-03-17T17:41:28.473857705Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Mar 17 17:41:28.473899 containerd[1467]: time="2025-03-17T17:41:28.473875278Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Mar 17 17:41:28.473899 containerd[1467]: time="2025-03-17T17:41:28.473887121Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Mar 17 17:41:28.473899 containerd[1467]: time="2025-03-17T17:41:28.473896929Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Mar 17 17:41:28.474067 containerd[1467]: time="2025-03-17T17:41:28.473940280Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Mar 17 17:41:28.474067 containerd[1467]: time="2025-03-17T17:41:28.473956381Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Mar 17 17:41:28.474067 containerd[1467]: time="2025-03-17T17:41:28.473965828Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Mar 17 17:41:28.474067 containerd[1467]: time="2025-03-17T17:41:28.473977370Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Mar 17 17:41:28.474067 containerd[1467]: time="2025-03-17T17:41:28.473985976Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Mar 17 17:41:28.474067 containerd[1467]: time="2025-03-17T17:41:28.473997748Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Mar 17 17:41:28.474067 containerd[1467]: time="2025-03-17T17:41:28.474007647Z" level=info msg="NRI interface is disabled by configuration." Mar 17 17:41:28.474067 containerd[1467]: time="2025-03-17T17:41:28.474017405Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Mar 17 17:41:28.474315 containerd[1467]: time="2025-03-17T17:41:28.474272223Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Mar 17 17:41:28.474315 containerd[1467]: time="2025-03-17T17:41:28.474314292Z" level=info msg="Connect containerd service" Mar 17 17:41:28.474485 containerd[1467]: time="2025-03-17T17:41:28.474344829Z" level=info msg="using legacy CRI server" Mar 17 17:41:28.474485 containerd[1467]: time="2025-03-17T17:41:28.474351171Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Mar 17 17:41:28.474485 containerd[1467]: time="2025-03-17T17:41:28.474457942Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Mar 17 17:41:28.475152 containerd[1467]: time="2025-03-17T17:41:28.474971114Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Mar 17 17:41:28.475641 containerd[1467]: time="2025-03-17T17:41:28.475605845Z" level=info msg="Start subscribing containerd event" Mar 17 17:41:28.475922 containerd[1467]: time="2025-03-17T17:41:28.475810659Z" level=info msg="Start recovering state" Mar 17 17:41:28.475922 containerd[1467]: time="2025-03-17T17:41:28.475885329Z" level=info msg="Start event monitor" Mar 17 17:41:28.475922 containerd[1467]: time="2025-03-17T17:41:28.475895559Z" level=info msg="Start snapshots syncer" Mar 17 17:41:28.476092 containerd[1467]: time="2025-03-17T17:41:28.475906148Z" level=info msg="Start cni network conf syncer for default" Mar 17 17:41:28.476092 containerd[1467]: time="2025-03-17T17:41:28.476052804Z" level=info msg="Start streaming server" Mar 17 17:41:28.477535 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Mar 17 17:41:28.479946 containerd[1467]: time="2025-03-17T17:41:28.479927723Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Mar 17 17:41:28.481371 containerd[1467]: time="2025-03-17T17:41:28.479985462Z" level=info msg=serving... address=/run/containerd/containerd.sock Mar 17 17:41:28.481371 containerd[1467]: time="2025-03-17T17:41:28.480043200Z" level=info msg="containerd successfully booted in 0.136623s" Mar 17 17:41:28.481340 systemd[1]: Started containerd.service - containerd container runtime. Mar 17 17:41:28.490594 systemd[1]: Starting user@500.service - User Manager for UID 500... Mar 17 17:41:28.496200 (systemd)[1538]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Mar 17 17:41:28.575916 tar[1466]: linux-amd64/LICENSE Mar 17 17:41:28.576003 tar[1466]: linux-amd64/README.md Mar 17 17:41:28.588827 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Mar 17 17:41:28.599125 systemd[1538]: Queued start job for default target default.target. Mar 17 17:41:28.610600 systemd[1538]: Created slice app.slice - User Application Slice. Mar 17 17:41:28.610626 systemd[1538]: Reached target paths.target - Paths. Mar 17 17:41:28.610640 systemd[1538]: Reached target timers.target - Timers. Mar 17 17:41:28.612211 systemd[1538]: Starting dbus.socket - D-Bus User Message Bus Socket... Mar 17 17:41:28.624432 systemd[1538]: Listening on dbus.socket - D-Bus User Message Bus Socket. Mar 17 17:41:28.624557 systemd[1538]: Reached target sockets.target - Sockets. Mar 17 17:41:28.624577 systemd[1538]: Reached target basic.target - Basic System. Mar 17 17:41:28.624614 systemd[1538]: Reached target default.target - Main User Target. Mar 17 17:41:28.624648 systemd[1538]: Startup finished in 120ms. Mar 17 17:41:28.625058 systemd[1]: Started user@500.service - User Manager for UID 500. Mar 17 17:41:28.635526 systemd[1]: Started session-1.scope - Session 1 of User core. Mar 17 17:41:28.696732 systemd[1]: Started sshd@1-10.0.0.61:22-10.0.0.1:49352.service - OpenSSH per-connection server daemon (10.0.0.1:49352). Mar 17 17:41:28.745990 sshd[1552]: Accepted publickey for core from 10.0.0.1 port 49352 ssh2: RSA SHA256:j201F9FRK1q3ChnxQf0adNdYppDp+g37vmaXPvsVhek Mar 17 17:41:28.747288 sshd-session[1552]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 17 17:41:28.751202 systemd-logind[1453]: New session 2 of user core. Mar 17 17:41:28.767489 systemd[1]: Started session-2.scope - Session 2 of User core. Mar 17 17:41:28.821206 sshd[1554]: Connection closed by 10.0.0.1 port 49352 Mar 17 17:41:28.821588 sshd-session[1552]: pam_unix(sshd:session): session closed for user core Mar 17 17:41:28.839900 systemd[1]: sshd@1-10.0.0.61:22-10.0.0.1:49352.service: Deactivated successfully. Mar 17 17:41:28.841468 systemd[1]: session-2.scope: Deactivated successfully. Mar 17 17:41:28.842996 systemd-logind[1453]: Session 2 logged out. Waiting for processes to exit. Mar 17 17:41:28.855580 systemd[1]: Started sshd@2-10.0.0.61:22-10.0.0.1:49362.service - OpenSSH per-connection server daemon (10.0.0.1:49362). Mar 17 17:41:28.857686 systemd-logind[1453]: Removed session 2. Mar 17 17:41:28.886815 sshd[1559]: Accepted publickey for core from 10.0.0.1 port 49362 ssh2: RSA SHA256:j201F9FRK1q3ChnxQf0adNdYppDp+g37vmaXPvsVhek Mar 17 17:41:28.888107 sshd-session[1559]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 17 17:41:28.891523 systemd-logind[1453]: New session 3 of user core. Mar 17 17:41:28.900469 systemd[1]: Started session-3.scope - Session 3 of User core. Mar 17 17:41:28.953587 sshd[1561]: Connection closed by 10.0.0.1 port 49362 Mar 17 17:41:28.953942 sshd-session[1559]: pam_unix(sshd:session): session closed for user core Mar 17 17:41:28.957715 systemd[1]: sshd@2-10.0.0.61:22-10.0.0.1:49362.service: Deactivated successfully. Mar 17 17:41:28.959228 systemd[1]: session-3.scope: Deactivated successfully. Mar 17 17:41:28.959871 systemd-logind[1453]: Session 3 logged out. Waiting for processes to exit. Mar 17 17:41:28.960628 systemd-logind[1453]: Removed session 3. Mar 17 17:41:28.971467 systemd-networkd[1410]: eth0: Gained IPv6LL Mar 17 17:41:28.974107 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Mar 17 17:41:28.975891 systemd[1]: Reached target network-online.target - Network is Online. Mar 17 17:41:28.987544 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Mar 17 17:41:28.989766 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 17 17:41:28.991885 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Mar 17 17:41:29.008676 systemd[1]: coreos-metadata.service: Deactivated successfully. Mar 17 17:41:29.008923 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Mar 17 17:41:29.010540 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Mar 17 17:41:29.014008 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Mar 17 17:41:29.608427 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 17 17:41:29.610255 systemd[1]: Reached target multi-user.target - Multi-User System. Mar 17 17:41:29.611632 systemd[1]: Startup finished in 1.030s (kernel) + 5.418s (initrd) + 4.068s (userspace) = 10.517s. Mar 17 17:41:29.615258 (kubelet)[1587]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 17 17:41:30.054374 kubelet[1587]: E0317 17:41:30.054198 1587 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 17 17:41:30.058888 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 17 17:41:30.059122 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 17 17:41:38.970279 systemd[1]: Started sshd@3-10.0.0.61:22-10.0.0.1:52930.service - OpenSSH per-connection server daemon (10.0.0.1:52930). Mar 17 17:41:39.006638 sshd[1601]: Accepted publickey for core from 10.0.0.1 port 52930 ssh2: RSA SHA256:j201F9FRK1q3ChnxQf0adNdYppDp+g37vmaXPvsVhek Mar 17 17:41:39.008094 sshd-session[1601]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 17 17:41:39.011942 systemd-logind[1453]: New session 4 of user core. Mar 17 17:41:39.019628 systemd[1]: Started session-4.scope - Session 4 of User core. Mar 17 17:41:39.074119 sshd[1603]: Connection closed by 10.0.0.1 port 52930 Mar 17 17:41:39.074497 sshd-session[1601]: pam_unix(sshd:session): session closed for user core Mar 17 17:41:39.089270 systemd[1]: sshd@3-10.0.0.61:22-10.0.0.1:52930.service: Deactivated successfully. Mar 17 17:41:39.091314 systemd[1]: session-4.scope: Deactivated successfully. Mar 17 17:41:39.093137 systemd-logind[1453]: Session 4 logged out. Waiting for processes to exit. Mar 17 17:41:39.100608 systemd[1]: Started sshd@4-10.0.0.61:22-10.0.0.1:52946.service - OpenSSH per-connection server daemon (10.0.0.1:52946). Mar 17 17:41:39.101510 systemd-logind[1453]: Removed session 4. Mar 17 17:41:39.131798 sshd[1608]: Accepted publickey for core from 10.0.0.1 port 52946 ssh2: RSA SHA256:j201F9FRK1q3ChnxQf0adNdYppDp+g37vmaXPvsVhek Mar 17 17:41:39.133407 sshd-session[1608]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 17 17:41:39.137877 systemd-logind[1453]: New session 5 of user core. Mar 17 17:41:39.155493 systemd[1]: Started session-5.scope - Session 5 of User core. Mar 17 17:41:39.206113 sshd[1610]: Connection closed by 10.0.0.1 port 52946 Mar 17 17:41:39.206461 sshd-session[1608]: pam_unix(sshd:session): session closed for user core Mar 17 17:41:39.217152 systemd[1]: sshd@4-10.0.0.61:22-10.0.0.1:52946.service: Deactivated successfully. Mar 17 17:41:39.219038 systemd[1]: session-5.scope: Deactivated successfully. Mar 17 17:41:39.220307 systemd-logind[1453]: Session 5 logged out. Waiting for processes to exit. Mar 17 17:41:39.238573 systemd[1]: Started sshd@5-10.0.0.61:22-10.0.0.1:52958.service - OpenSSH per-connection server daemon (10.0.0.1:52958). Mar 17 17:41:39.239483 systemd-logind[1453]: Removed session 5. Mar 17 17:41:39.270172 sshd[1615]: Accepted publickey for core from 10.0.0.1 port 52958 ssh2: RSA SHA256:j201F9FRK1q3ChnxQf0adNdYppDp+g37vmaXPvsVhek Mar 17 17:41:39.271517 sshd-session[1615]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 17 17:41:39.275288 systemd-logind[1453]: New session 6 of user core. Mar 17 17:41:39.285462 systemd[1]: Started session-6.scope - Session 6 of User core. Mar 17 17:41:39.339572 sshd[1617]: Connection closed by 10.0.0.1 port 52958 Mar 17 17:41:39.339856 sshd-session[1615]: pam_unix(sshd:session): session closed for user core Mar 17 17:41:39.349112 systemd[1]: sshd@5-10.0.0.61:22-10.0.0.1:52958.service: Deactivated successfully. Mar 17 17:41:39.350930 systemd[1]: session-6.scope: Deactivated successfully. Mar 17 17:41:39.352684 systemd-logind[1453]: Session 6 logged out. Waiting for processes to exit. Mar 17 17:41:39.362568 systemd[1]: Started sshd@6-10.0.0.61:22-10.0.0.1:52962.service - OpenSSH per-connection server daemon (10.0.0.1:52962). Mar 17 17:41:39.363340 systemd-logind[1453]: Removed session 6. Mar 17 17:41:39.393295 sshd[1622]: Accepted publickey for core from 10.0.0.1 port 52962 ssh2: RSA SHA256:j201F9FRK1q3ChnxQf0adNdYppDp+g37vmaXPvsVhek Mar 17 17:41:39.394599 sshd-session[1622]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 17 17:41:39.397988 systemd-logind[1453]: New session 7 of user core. Mar 17 17:41:39.408466 systemd[1]: Started session-7.scope - Session 7 of User core. Mar 17 17:41:39.465843 sudo[1625]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Mar 17 17:41:39.466170 sudo[1625]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 17 17:41:39.492069 sudo[1625]: pam_unix(sudo:session): session closed for user root Mar 17 17:41:39.493670 sshd[1624]: Connection closed by 10.0.0.1 port 52962 Mar 17 17:41:39.494054 sshd-session[1622]: pam_unix(sshd:session): session closed for user core Mar 17 17:41:39.505243 systemd[1]: sshd@6-10.0.0.61:22-10.0.0.1:52962.service: Deactivated successfully. Mar 17 17:41:39.507022 systemd[1]: session-7.scope: Deactivated successfully. Mar 17 17:41:39.508457 systemd-logind[1453]: Session 7 logged out. Waiting for processes to exit. Mar 17 17:41:39.521592 systemd[1]: Started sshd@7-10.0.0.61:22-10.0.0.1:52976.service - OpenSSH per-connection server daemon (10.0.0.1:52976). Mar 17 17:41:39.522438 systemd-logind[1453]: Removed session 7. Mar 17 17:41:39.552915 sshd[1630]: Accepted publickey for core from 10.0.0.1 port 52976 ssh2: RSA SHA256:j201F9FRK1q3ChnxQf0adNdYppDp+g37vmaXPvsVhek Mar 17 17:41:39.554302 sshd-session[1630]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 17 17:41:39.557952 systemd-logind[1453]: New session 8 of user core. Mar 17 17:41:39.574469 systemd[1]: Started session-8.scope - Session 8 of User core. Mar 17 17:41:39.627451 sudo[1634]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Mar 17 17:41:39.627790 sudo[1634]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 17 17:41:39.631137 sudo[1634]: pam_unix(sudo:session): session closed for user root Mar 17 17:41:39.637621 sudo[1633]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Mar 17 17:41:39.637962 sudo[1633]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 17 17:41:39.657604 systemd[1]: Starting audit-rules.service - Load Audit Rules... Mar 17 17:41:39.687093 augenrules[1656]: No rules Mar 17 17:41:39.688839 systemd[1]: audit-rules.service: Deactivated successfully. Mar 17 17:41:39.689068 systemd[1]: Finished audit-rules.service - Load Audit Rules. Mar 17 17:41:39.690550 sudo[1633]: pam_unix(sudo:session): session closed for user root Mar 17 17:41:39.692049 sshd[1632]: Connection closed by 10.0.0.1 port 52976 Mar 17 17:41:39.692426 sshd-session[1630]: pam_unix(sshd:session): session closed for user core Mar 17 17:41:39.703034 systemd[1]: sshd@7-10.0.0.61:22-10.0.0.1:52976.service: Deactivated successfully. Mar 17 17:41:39.704898 systemd[1]: session-8.scope: Deactivated successfully. Mar 17 17:41:39.706577 systemd-logind[1453]: Session 8 logged out. Waiting for processes to exit. Mar 17 17:41:39.715601 systemd[1]: Started sshd@8-10.0.0.61:22-10.0.0.1:52988.service - OpenSSH per-connection server daemon (10.0.0.1:52988). Mar 17 17:41:39.716430 systemd-logind[1453]: Removed session 8. Mar 17 17:41:39.747268 sshd[1664]: Accepted publickey for core from 10.0.0.1 port 52988 ssh2: RSA SHA256:j201F9FRK1q3ChnxQf0adNdYppDp+g37vmaXPvsVhek Mar 17 17:41:39.748771 sshd-session[1664]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 17 17:41:39.752490 systemd-logind[1453]: New session 9 of user core. Mar 17 17:41:39.766477 systemd[1]: Started session-9.scope - Session 9 of User core. Mar 17 17:41:39.818725 sudo[1668]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Mar 17 17:41:39.819064 sudo[1668]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 17 17:41:40.073040 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Mar 17 17:41:40.081556 systemd[1]: Starting docker.service - Docker Application Container Engine... Mar 17 17:41:40.081719 (dockerd)[1688]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Mar 17 17:41:40.082619 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 17 17:41:40.248252 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 17 17:41:40.254074 (kubelet)[1703]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 17 17:41:40.295266 kubelet[1703]: E0317 17:41:40.295088 1703 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 17 17:41:40.302106 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 17 17:41:40.302312 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 17 17:41:40.336219 dockerd[1688]: time="2025-03-17T17:41:40.336100920Z" level=info msg="Starting up" Mar 17 17:41:41.182642 dockerd[1688]: time="2025-03-17T17:41:41.182568865Z" level=info msg="Loading containers: start." Mar 17 17:41:41.515385 kernel: Initializing XFRM netlink socket Mar 17 17:41:41.597024 systemd-networkd[1410]: docker0: Link UP Mar 17 17:41:41.697823 dockerd[1688]: time="2025-03-17T17:41:41.697761052Z" level=info msg="Loading containers: done." Mar 17 17:41:41.712382 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck2885099140-merged.mount: Deactivated successfully. Mar 17 17:41:41.784331 dockerd[1688]: time="2025-03-17T17:41:41.784293313Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Mar 17 17:41:41.784416 dockerd[1688]: time="2025-03-17T17:41:41.784386828Z" level=info msg="Docker daemon" commit=8b539b8df24032dabeaaa099cf1d0535ef0286a3 containerd-snapshotter=false storage-driver=overlay2 version=27.2.1 Mar 17 17:41:41.784534 dockerd[1688]: time="2025-03-17T17:41:41.784511292Z" level=info msg="Daemon has completed initialization" Mar 17 17:41:42.069317 dockerd[1688]: time="2025-03-17T17:41:42.068989012Z" level=info msg="API listen on /run/docker.sock" Mar 17 17:41:42.069281 systemd[1]: Started docker.service - Docker Application Container Engine. Mar 17 17:41:42.899605 containerd[1467]: time="2025-03-17T17:41:42.899550926Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.30.11\"" Mar 17 17:41:43.504708 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1090382049.mount: Deactivated successfully. Mar 17 17:41:44.461402 containerd[1467]: time="2025-03-17T17:41:44.461324384Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.30.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:41:44.462027 containerd[1467]: time="2025-03-17T17:41:44.461965287Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.30.11: active requests=0, bytes read=32674573" Mar 17 17:41:44.463144 containerd[1467]: time="2025-03-17T17:41:44.463112739Z" level=info msg="ImageCreate event name:\"sha256:4db5a05c271eac8f5da2f95895ea1ccb9a38f48db3135ba3bdfe35941a396ea8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:41:44.465684 containerd[1467]: time="2025-03-17T17:41:44.465658125Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:77c54346965036acc7ac95c3200597ede36db9246179248dde21c1a3ecc1caf0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:41:44.466751 containerd[1467]: time="2025-03-17T17:41:44.466705439Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.30.11\" with image id \"sha256:4db5a05c271eac8f5da2f95895ea1ccb9a38f48db3135ba3bdfe35941a396ea8\", repo tag \"registry.k8s.io/kube-apiserver:v1.30.11\", repo digest \"registry.k8s.io/kube-apiserver@sha256:77c54346965036acc7ac95c3200597ede36db9246179248dde21c1a3ecc1caf0\", size \"32671373\" in 1.567113116s" Mar 17 17:41:44.466796 containerd[1467]: time="2025-03-17T17:41:44.466753610Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.30.11\" returns image reference \"sha256:4db5a05c271eac8f5da2f95895ea1ccb9a38f48db3135ba3bdfe35941a396ea8\"" Mar 17 17:41:44.487726 containerd[1467]: time="2025-03-17T17:41:44.487677955Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.30.11\"" Mar 17 17:41:46.302876 containerd[1467]: time="2025-03-17T17:41:46.302805294Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.30.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:41:46.303507 containerd[1467]: time="2025-03-17T17:41:46.303425457Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.30.11: active requests=0, bytes read=29619772" Mar 17 17:41:46.304625 containerd[1467]: time="2025-03-17T17:41:46.304590663Z" level=info msg="ImageCreate event name:\"sha256:de1025c2d496829d3250130380737609ffcdd10a4dce6f2dcd03f23a85a15e6a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:41:46.307887 containerd[1467]: time="2025-03-17T17:41:46.307859696Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:d8874f3fb45591ecdac67a3035c730808f18b3ab13147495c7d77eb1960d4f6f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:41:46.311196 containerd[1467]: time="2025-03-17T17:41:46.310116180Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.30.11\" with image id \"sha256:de1025c2d496829d3250130380737609ffcdd10a4dce6f2dcd03f23a85a15e6a\", repo tag \"registry.k8s.io/kube-controller-manager:v1.30.11\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:d8874f3fb45591ecdac67a3035c730808f18b3ab13147495c7d77eb1960d4f6f\", size \"31107380\" in 1.822410232s" Mar 17 17:41:46.311196 containerd[1467]: time="2025-03-17T17:41:46.310146567Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.30.11\" returns image reference \"sha256:de1025c2d496829d3250130380737609ffcdd10a4dce6f2dcd03f23a85a15e6a\"" Mar 17 17:41:46.332429 containerd[1467]: time="2025-03-17T17:41:46.332387693Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.30.11\"" Mar 17 17:41:47.320267 containerd[1467]: time="2025-03-17T17:41:47.320180172Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.30.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:41:47.321139 containerd[1467]: time="2025-03-17T17:41:47.321096541Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.30.11: active requests=0, bytes read=17903309" Mar 17 17:41:47.322131 containerd[1467]: time="2025-03-17T17:41:47.322099773Z" level=info msg="ImageCreate event name:\"sha256:11492f0faf138e933cadd6f533f03e401da9a35e53711e833f18afa6b185b2b7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:41:47.324855 containerd[1467]: time="2025-03-17T17:41:47.324826579Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:c699f8c97ae7ec819c8bd878d3db104ba72fc440d810d9030e09286b696017b5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:41:47.325974 containerd[1467]: time="2025-03-17T17:41:47.325925570Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.30.11\" with image id \"sha256:11492f0faf138e933cadd6f533f03e401da9a35e53711e833f18afa6b185b2b7\", repo tag \"registry.k8s.io/kube-scheduler:v1.30.11\", repo digest \"registry.k8s.io/kube-scheduler@sha256:c699f8c97ae7ec819c8bd878d3db104ba72fc440d810d9030e09286b696017b5\", size \"19390935\" in 993.499386ms" Mar 17 17:41:47.325974 containerd[1467]: time="2025-03-17T17:41:47.325971777Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.30.11\" returns image reference \"sha256:11492f0faf138e933cadd6f533f03e401da9a35e53711e833f18afa6b185b2b7\"" Mar 17 17:41:47.348217 containerd[1467]: time="2025-03-17T17:41:47.348145226Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.11\"" Mar 17 17:41:48.285819 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2683431612.mount: Deactivated successfully. Mar 17 17:41:48.534745 containerd[1467]: time="2025-03-17T17:41:48.534678066Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.30.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:41:48.535540 containerd[1467]: time="2025-03-17T17:41:48.535474481Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.30.11: active requests=0, bytes read=29185372" Mar 17 17:41:48.536821 containerd[1467]: time="2025-03-17T17:41:48.536739313Z" level=info msg="ImageCreate event name:\"sha256:01045f200a8856c3f5ccfa7be03d72274f1f16fc7a047659e709d603d5c019dc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:41:48.538604 containerd[1467]: time="2025-03-17T17:41:48.538575248Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:ea4da798040a18ed3f302e8d5f67307c7275a2a53bcf3d51bcec223acda84a55\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:41:48.539100 containerd[1467]: time="2025-03-17T17:41:48.539057332Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.30.11\" with image id \"sha256:01045f200a8856c3f5ccfa7be03d72274f1f16fc7a047659e709d603d5c019dc\", repo tag \"registry.k8s.io/kube-proxy:v1.30.11\", repo digest \"registry.k8s.io/kube-proxy@sha256:ea4da798040a18ed3f302e8d5f67307c7275a2a53bcf3d51bcec223acda84a55\", size \"29184391\" in 1.190850892s" Mar 17 17:41:48.539127 containerd[1467]: time="2025-03-17T17:41:48.539100283Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.11\" returns image reference \"sha256:01045f200a8856c3f5ccfa7be03d72274f1f16fc7a047659e709d603d5c019dc\"" Mar 17 17:41:48.561364 containerd[1467]: time="2025-03-17T17:41:48.561322654Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\"" Mar 17 17:41:49.100852 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2176240215.mount: Deactivated successfully. Mar 17 17:41:49.996546 containerd[1467]: time="2025-03-17T17:41:49.996485101Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:41:49.997222 containerd[1467]: time="2025-03-17T17:41:49.997165638Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.1: active requests=0, bytes read=18185761" Mar 17 17:41:49.998372 containerd[1467]: time="2025-03-17T17:41:49.998318831Z" level=info msg="ImageCreate event name:\"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:41:50.000954 containerd[1467]: time="2025-03-17T17:41:50.000921404Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:41:50.002013 containerd[1467]: time="2025-03-17T17:41:50.001984398Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.1\" with image id \"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\", size \"18182961\" in 1.44062705s" Mar 17 17:41:50.002055 containerd[1467]: time="2025-03-17T17:41:50.002012611Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\" returns image reference \"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\"" Mar 17 17:41:50.024527 containerd[1467]: time="2025-03-17T17:41:50.024493236Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\"" Mar 17 17:41:50.447141 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Mar 17 17:41:50.453510 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 17 17:41:50.454799 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2523711000.mount: Deactivated successfully. Mar 17 17:41:50.455495 containerd[1467]: time="2025-03-17T17:41:50.455254285Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:41:50.456945 containerd[1467]: time="2025-03-17T17:41:50.456895574Z" level=info msg="stop pulling image registry.k8s.io/pause:3.9: active requests=0, bytes read=322290" Mar 17 17:41:50.458119 containerd[1467]: time="2025-03-17T17:41:50.458077271Z" level=info msg="ImageCreate event name:\"sha256:e6f1816883972d4be47bd48879a08919b96afcd344132622e4d444987919323c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:41:50.460500 containerd[1467]: time="2025-03-17T17:41:50.460446926Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:41:50.461137 containerd[1467]: time="2025-03-17T17:41:50.461107876Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.9\" with image id \"sha256:e6f1816883972d4be47bd48879a08919b96afcd344132622e4d444987919323c\", repo tag \"registry.k8s.io/pause:3.9\", repo digest \"registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097\", size \"321520\" in 436.583382ms" Mar 17 17:41:50.461137 containerd[1467]: time="2025-03-17T17:41:50.461136189Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\" returns image reference \"sha256:e6f1816883972d4be47bd48879a08919b96afcd344132622e4d444987919323c\"" Mar 17 17:41:50.483509 containerd[1467]: time="2025-03-17T17:41:50.483404787Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.12-0\"" Mar 17 17:41:50.598805 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 17 17:41:50.603162 (kubelet)[2076]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 17 17:41:50.642961 kubelet[2076]: E0317 17:41:50.642906 2076 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 17 17:41:50.647193 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 17 17:41:50.647468 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 17 17:41:51.157785 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2453560143.mount: Deactivated successfully. Mar 17 17:41:52.737861 containerd[1467]: time="2025-03-17T17:41:52.737780206Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.12-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:41:52.738500 containerd[1467]: time="2025-03-17T17:41:52.738446576Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.12-0: active requests=0, bytes read=57238571" Mar 17 17:41:52.739610 containerd[1467]: time="2025-03-17T17:41:52.739584010Z" level=info msg="ImageCreate event name:\"sha256:3861cfcd7c04ccac1f062788eca39487248527ef0c0cfd477a83d7691a75a899\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:41:52.742761 containerd[1467]: time="2025-03-17T17:41:52.742730133Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:44a8e24dcbba3470ee1fee21d5e88d128c936e9b55d4bc51fbef8086f8ed123b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:41:52.743836 containerd[1467]: time="2025-03-17T17:41:52.743799519Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.12-0\" with image id \"sha256:3861cfcd7c04ccac1f062788eca39487248527ef0c0cfd477a83d7691a75a899\", repo tag \"registry.k8s.io/etcd:3.5.12-0\", repo digest \"registry.k8s.io/etcd@sha256:44a8e24dcbba3470ee1fee21d5e88d128c936e9b55d4bc51fbef8086f8ed123b\", size \"57236178\" in 2.260367551s" Mar 17 17:41:52.743836 containerd[1467]: time="2025-03-17T17:41:52.743828102Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.12-0\" returns image reference \"sha256:3861cfcd7c04ccac1f062788eca39487248527ef0c0cfd477a83d7691a75a899\"" Mar 17 17:41:55.193198 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 17 17:41:55.204560 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 17 17:41:55.221950 systemd[1]: Reloading requested from client PID 2212 ('systemctl') (unit session-9.scope)... Mar 17 17:41:55.221970 systemd[1]: Reloading... Mar 17 17:41:55.310453 zram_generator::config[2252]: No configuration found. Mar 17 17:41:55.511864 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 17 17:41:55.597768 systemd[1]: Reloading finished in 375 ms. Mar 17 17:41:55.647515 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Mar 17 17:41:55.647642 systemd[1]: kubelet.service: Failed with result 'signal'. Mar 17 17:41:55.647959 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 17 17:41:55.649674 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 17 17:41:55.792988 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 17 17:41:55.798052 (kubelet)[2300]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Mar 17 17:41:55.834810 kubelet[2300]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 17 17:41:55.834810 kubelet[2300]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Mar 17 17:41:55.834810 kubelet[2300]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 17 17:41:55.835744 kubelet[2300]: I0317 17:41:55.835698 2300 server.go:205] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 17 17:41:56.115289 kubelet[2300]: I0317 17:41:56.115172 2300 server.go:484] "Kubelet version" kubeletVersion="v1.30.1" Mar 17 17:41:56.115289 kubelet[2300]: I0317 17:41:56.115209 2300 server.go:486] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 17 17:41:56.115469 kubelet[2300]: I0317 17:41:56.115448 2300 server.go:927] "Client rotation is on, will bootstrap in background" Mar 17 17:41:56.132363 kubelet[2300]: I0317 17:41:56.132309 2300 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 17 17:41:56.132940 kubelet[2300]: E0317 17:41:56.132909 2300 certificate_manager.go:562] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://10.0.0.61:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 10.0.0.61:6443: connect: connection refused Mar 17 17:41:56.145559 kubelet[2300]: I0317 17:41:56.145532 2300 server.go:742] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Mar 17 17:41:56.146755 kubelet[2300]: I0317 17:41:56.146712 2300 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 17 17:41:56.146922 kubelet[2300]: I0317 17:41:56.146747 2300 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Mar 17 17:41:56.147022 kubelet[2300]: I0317 17:41:56.146932 2300 topology_manager.go:138] "Creating topology manager with none policy" Mar 17 17:41:56.147022 kubelet[2300]: I0317 17:41:56.146943 2300 container_manager_linux.go:301] "Creating device plugin manager" Mar 17 17:41:56.147094 kubelet[2300]: I0317 17:41:56.147077 2300 state_mem.go:36] "Initialized new in-memory state store" Mar 17 17:41:56.147674 kubelet[2300]: I0317 17:41:56.147651 2300 kubelet.go:400] "Attempting to sync node with API server" Mar 17 17:41:56.147674 kubelet[2300]: I0317 17:41:56.147668 2300 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 17 17:41:56.147729 kubelet[2300]: I0317 17:41:56.147688 2300 kubelet.go:312] "Adding apiserver pod source" Mar 17 17:41:56.147729 kubelet[2300]: I0317 17:41:56.147699 2300 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 17 17:41:56.148292 kubelet[2300]: W0317 17:41:56.148194 2300 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.61:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.61:6443: connect: connection refused Mar 17 17:41:56.148292 kubelet[2300]: E0317 17:41:56.148254 2300 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.0.0.61:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.61:6443: connect: connection refused Mar 17 17:41:56.148292 kubelet[2300]: W0317 17:41:56.148256 2300 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.61:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.61:6443: connect: connection refused Mar 17 17:41:56.148292 kubelet[2300]: E0317 17:41:56.148296 2300 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.0.0.61:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.61:6443: connect: connection refused Mar 17 17:41:56.150731 kubelet[2300]: I0317 17:41:56.150701 2300 kuberuntime_manager.go:261] "Container runtime initialized" containerRuntime="containerd" version="v1.7.23" apiVersion="v1" Mar 17 17:41:56.152340 kubelet[2300]: I0317 17:41:56.152314 2300 kubelet.go:815] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Mar 17 17:41:56.152409 kubelet[2300]: W0317 17:41:56.152383 2300 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Mar 17 17:41:56.153006 kubelet[2300]: I0317 17:41:56.152985 2300 server.go:1264] "Started kubelet" Mar 17 17:41:56.153670 kubelet[2300]: I0317 17:41:56.153080 2300 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Mar 17 17:41:56.153670 kubelet[2300]: I0317 17:41:56.153256 2300 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 17 17:41:56.153670 kubelet[2300]: I0317 17:41:56.153561 2300 server.go:227] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 17 17:41:56.159263 kubelet[2300]: I0317 17:41:56.158650 2300 server.go:455] "Adding debug handlers to kubelet server" Mar 17 17:41:56.159263 kubelet[2300]: I0317 17:41:56.159229 2300 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 17 17:41:56.162369 kubelet[2300]: E0317 17:41:56.160789 2300 kubelet.go:1467] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Mar 17 17:41:56.162369 kubelet[2300]: E0317 17:41:56.160762 2300 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.61:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.61:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.182da7f86525dec9 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-03-17 17:41:56.152958665 +0000 UTC m=+0.350947684,LastTimestamp:2025-03-17 17:41:56.152958665 +0000 UTC m=+0.350947684,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Mar 17 17:41:56.162369 kubelet[2300]: E0317 17:41:56.160941 2300 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 17 17:41:56.162369 kubelet[2300]: I0317 17:41:56.160981 2300 volume_manager.go:291] "Starting Kubelet Volume Manager" Mar 17 17:41:56.162369 kubelet[2300]: I0317 17:41:56.161085 2300 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Mar 17 17:41:56.162369 kubelet[2300]: I0317 17:41:56.161132 2300 reconciler.go:26] "Reconciler: start to sync state" Mar 17 17:41:56.162369 kubelet[2300]: W0317 17:41:56.161425 2300 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.61:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.61:6443: connect: connection refused Mar 17 17:41:56.162369 kubelet[2300]: E0317 17:41:56.161468 2300 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.0.0.61:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.61:6443: connect: connection refused Mar 17 17:41:56.162622 kubelet[2300]: E0317 17:41:56.161644 2300 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.61:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.61:6443: connect: connection refused" interval="200ms" Mar 17 17:41:56.162850 kubelet[2300]: I0317 17:41:56.162822 2300 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Mar 17 17:41:56.164566 kubelet[2300]: I0317 17:41:56.164551 2300 factory.go:221] Registration of the containerd container factory successfully Mar 17 17:41:56.164622 kubelet[2300]: I0317 17:41:56.164613 2300 factory.go:221] Registration of the systemd container factory successfully Mar 17 17:41:56.175973 kubelet[2300]: I0317 17:41:56.175798 2300 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Mar 17 17:41:56.177081 kubelet[2300]: I0317 17:41:56.177058 2300 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Mar 17 17:41:56.177081 kubelet[2300]: I0317 17:41:56.177083 2300 status_manager.go:217] "Starting to sync pod status with apiserver" Mar 17 17:41:56.177180 kubelet[2300]: I0317 17:41:56.177102 2300 kubelet.go:2337] "Starting kubelet main sync loop" Mar 17 17:41:56.177180 kubelet[2300]: E0317 17:41:56.177147 2300 kubelet.go:2361] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 17 17:41:56.181116 kubelet[2300]: W0317 17:41:56.180954 2300 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.61:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.61:6443: connect: connection refused Mar 17 17:41:56.181116 kubelet[2300]: E0317 17:41:56.181003 2300 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.0.0.61:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.61:6443: connect: connection refused Mar 17 17:41:56.181545 kubelet[2300]: I0317 17:41:56.181524 2300 cpu_manager.go:214] "Starting CPU manager" policy="none" Mar 17 17:41:56.181545 kubelet[2300]: I0317 17:41:56.181541 2300 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Mar 17 17:41:56.181711 kubelet[2300]: I0317 17:41:56.181675 2300 state_mem.go:36] "Initialized new in-memory state store" Mar 17 17:41:56.262334 kubelet[2300]: I0317 17:41:56.262300 2300 kubelet_node_status.go:73] "Attempting to register node" node="localhost" Mar 17 17:41:56.262658 kubelet[2300]: E0317 17:41:56.262617 2300 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.61:6443/api/v1/nodes\": dial tcp 10.0.0.61:6443: connect: connection refused" node="localhost" Mar 17 17:41:56.277784 kubelet[2300]: E0317 17:41:56.277753 2300 kubelet.go:2361] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Mar 17 17:41:56.362720 kubelet[2300]: E0317 17:41:56.362676 2300 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.61:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.61:6443: connect: connection refused" interval="400ms" Mar 17 17:41:56.460625 kubelet[2300]: I0317 17:41:56.460546 2300 policy_none.go:49] "None policy: Start" Mar 17 17:41:56.461286 kubelet[2300]: I0317 17:41:56.461208 2300 memory_manager.go:170] "Starting memorymanager" policy="None" Mar 17 17:41:56.461286 kubelet[2300]: I0317 17:41:56.461234 2300 state_mem.go:35] "Initializing new in-memory state store" Mar 17 17:41:56.464520 kubelet[2300]: I0317 17:41:56.464494 2300 kubelet_node_status.go:73] "Attempting to register node" node="localhost" Mar 17 17:41:56.464838 kubelet[2300]: E0317 17:41:56.464808 2300 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.61:6443/api/v1/nodes\": dial tcp 10.0.0.61:6443: connect: connection refused" node="localhost" Mar 17 17:41:56.468924 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Mar 17 17:41:56.478509 kubelet[2300]: E0317 17:41:56.478478 2300 kubelet.go:2361] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Mar 17 17:41:56.483623 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Mar 17 17:41:56.486617 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Mar 17 17:41:56.502202 kubelet[2300]: I0317 17:41:56.502179 2300 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Mar 17 17:41:56.502481 kubelet[2300]: I0317 17:41:56.502402 2300 container_log_manager.go:186] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 17 17:41:56.502942 kubelet[2300]: I0317 17:41:56.502547 2300 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 17 17:41:56.503386 kubelet[2300]: E0317 17:41:56.503342 2300 eviction_manager.go:282] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Mar 17 17:41:56.763721 kubelet[2300]: E0317 17:41:56.763631 2300 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.61:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.61:6443: connect: connection refused" interval="800ms" Mar 17 17:41:56.866293 kubelet[2300]: I0317 17:41:56.866252 2300 kubelet_node_status.go:73] "Attempting to register node" node="localhost" Mar 17 17:41:56.866718 kubelet[2300]: E0317 17:41:56.866680 2300 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.61:6443/api/v1/nodes\": dial tcp 10.0.0.61:6443: connect: connection refused" node="localhost" Mar 17 17:41:56.878790 kubelet[2300]: I0317 17:41:56.878751 2300 topology_manager.go:215] "Topology Admit Handler" podUID="c29c7b4ac07fd8117c1d69a65632e6c1" podNamespace="kube-system" podName="kube-apiserver-localhost" Mar 17 17:41:56.879793 kubelet[2300]: I0317 17:41:56.879764 2300 topology_manager.go:215] "Topology Admit Handler" podUID="23a18e2dc14f395c5f1bea711a5a9344" podNamespace="kube-system" podName="kube-controller-manager-localhost" Mar 17 17:41:56.880608 kubelet[2300]: I0317 17:41:56.880567 2300 topology_manager.go:215] "Topology Admit Handler" podUID="d79ab404294384d4bcc36fb5b5509bbb" podNamespace="kube-system" podName="kube-scheduler-localhost" Mar 17 17:41:56.885708 systemd[1]: Created slice kubepods-burstable-podc29c7b4ac07fd8117c1d69a65632e6c1.slice - libcontainer container kubepods-burstable-podc29c7b4ac07fd8117c1d69a65632e6c1.slice. Mar 17 17:41:56.899169 systemd[1]: Created slice kubepods-burstable-pod23a18e2dc14f395c5f1bea711a5a9344.slice - libcontainer container kubepods-burstable-pod23a18e2dc14f395c5f1bea711a5a9344.slice. Mar 17 17:41:56.902975 systemd[1]: Created slice kubepods-burstable-podd79ab404294384d4bcc36fb5b5509bbb.slice - libcontainer container kubepods-burstable-podd79ab404294384d4bcc36fb5b5509bbb.slice. Mar 17 17:41:56.966675 kubelet[2300]: I0317 17:41:56.966632 2300 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/23a18e2dc14f395c5f1bea711a5a9344-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"23a18e2dc14f395c5f1bea711a5a9344\") " pod="kube-system/kube-controller-manager-localhost" Mar 17 17:41:56.966675 kubelet[2300]: I0317 17:41:56.966666 2300 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/23a18e2dc14f395c5f1bea711a5a9344-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"23a18e2dc14f395c5f1bea711a5a9344\") " pod="kube-system/kube-controller-manager-localhost" Mar 17 17:41:56.966842 kubelet[2300]: I0317 17:41:56.966685 2300 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/23a18e2dc14f395c5f1bea711a5a9344-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"23a18e2dc14f395c5f1bea711a5a9344\") " pod="kube-system/kube-controller-manager-localhost" Mar 17 17:41:56.966842 kubelet[2300]: I0317 17:41:56.966702 2300 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/23a18e2dc14f395c5f1bea711a5a9344-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"23a18e2dc14f395c5f1bea711a5a9344\") " pod="kube-system/kube-controller-manager-localhost" Mar 17 17:41:56.966842 kubelet[2300]: I0317 17:41:56.966717 2300 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/23a18e2dc14f395c5f1bea711a5a9344-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"23a18e2dc14f395c5f1bea711a5a9344\") " pod="kube-system/kube-controller-manager-localhost" Mar 17 17:41:56.966842 kubelet[2300]: I0317 17:41:56.966731 2300 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/d79ab404294384d4bcc36fb5b5509bbb-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"d79ab404294384d4bcc36fb5b5509bbb\") " pod="kube-system/kube-scheduler-localhost" Mar 17 17:41:56.966842 kubelet[2300]: I0317 17:41:56.966744 2300 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/c29c7b4ac07fd8117c1d69a65632e6c1-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"c29c7b4ac07fd8117c1d69a65632e6c1\") " pod="kube-system/kube-apiserver-localhost" Mar 17 17:41:56.966956 kubelet[2300]: I0317 17:41:56.966757 2300 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/c29c7b4ac07fd8117c1d69a65632e6c1-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"c29c7b4ac07fd8117c1d69a65632e6c1\") " pod="kube-system/kube-apiserver-localhost" Mar 17 17:41:56.966956 kubelet[2300]: I0317 17:41:56.966776 2300 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/c29c7b4ac07fd8117c1d69a65632e6c1-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"c29c7b4ac07fd8117c1d69a65632e6c1\") " pod="kube-system/kube-apiserver-localhost" Mar 17 17:41:57.196812 kubelet[2300]: E0317 17:41:57.196690 2300 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:41:57.197424 containerd[1467]: time="2025-03-17T17:41:57.197346611Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:c29c7b4ac07fd8117c1d69a65632e6c1,Namespace:kube-system,Attempt:0,}" Mar 17 17:41:57.201492 kubelet[2300]: E0317 17:41:57.201470 2300 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:41:57.201917 containerd[1467]: time="2025-03-17T17:41:57.201878463Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:23a18e2dc14f395c5f1bea711a5a9344,Namespace:kube-system,Attempt:0,}" Mar 17 17:41:57.205114 kubelet[2300]: E0317 17:41:57.205087 2300 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:41:57.205431 containerd[1467]: time="2025-03-17T17:41:57.205394760Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:d79ab404294384d4bcc36fb5b5509bbb,Namespace:kube-system,Attempt:0,}" Mar 17 17:41:57.239917 kubelet[2300]: W0317 17:41:57.239879 2300 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.61:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.61:6443: connect: connection refused Mar 17 17:41:57.239917 kubelet[2300]: E0317 17:41:57.239914 2300 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.0.0.61:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.61:6443: connect: connection refused Mar 17 17:41:57.564842 kubelet[2300]: E0317 17:41:57.564755 2300 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.61:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.61:6443: connect: connection refused" interval="1.6s" Mar 17 17:41:57.591344 kubelet[2300]: W0317 17:41:57.591280 2300 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.61:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.61:6443: connect: connection refused Mar 17 17:41:57.591426 kubelet[2300]: E0317 17:41:57.591350 2300 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.0.0.61:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.61:6443: connect: connection refused Mar 17 17:41:57.667593 kubelet[2300]: I0317 17:41:57.667574 2300 kubelet_node_status.go:73] "Attempting to register node" node="localhost" Mar 17 17:41:57.667892 kubelet[2300]: E0317 17:41:57.667845 2300 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.61:6443/api/v1/nodes\": dial tcp 10.0.0.61:6443: connect: connection refused" node="localhost" Mar 17 17:41:57.687477 kubelet[2300]: W0317 17:41:57.687423 2300 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.61:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.61:6443: connect: connection refused Mar 17 17:41:57.687477 kubelet[2300]: E0317 17:41:57.687475 2300 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.0.0.61:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.61:6443: connect: connection refused Mar 17 17:41:57.710703 kubelet[2300]: W0317 17:41:57.710667 2300 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.61:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.61:6443: connect: connection refused Mar 17 17:41:57.710703 kubelet[2300]: E0317 17:41:57.710696 2300 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.0.0.61:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.61:6443: connect: connection refused Mar 17 17:41:57.729605 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount53526676.mount: Deactivated successfully. Mar 17 17:41:57.737273 containerd[1467]: time="2025-03-17T17:41:57.737223238Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 17 17:41:57.739875 containerd[1467]: time="2025-03-17T17:41:57.739819148Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312056" Mar 17 17:41:57.740738 containerd[1467]: time="2025-03-17T17:41:57.740690483Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 17 17:41:57.742489 containerd[1467]: time="2025-03-17T17:41:57.742443712Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 17 17:41:57.743189 containerd[1467]: time="2025-03-17T17:41:57.743145929Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Mar 17 17:41:57.744158 containerd[1467]: time="2025-03-17T17:41:57.744119115Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 17 17:41:57.745106 containerd[1467]: time="2025-03-17T17:41:57.745068236Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Mar 17 17:41:57.745913 containerd[1467]: time="2025-03-17T17:41:57.745869920Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 17 17:41:57.747555 containerd[1467]: time="2025-03-17T17:41:57.747513884Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 542.043122ms" Mar 17 17:41:57.748303 containerd[1467]: time="2025-03-17T17:41:57.748271986Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 550.785763ms" Mar 17 17:41:57.752865 containerd[1467]: time="2025-03-17T17:41:57.752823375Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 550.857558ms" Mar 17 17:41:57.885820 containerd[1467]: time="2025-03-17T17:41:57.883194581Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 17:41:57.885820 containerd[1467]: time="2025-03-17T17:41:57.885461113Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 17:41:57.885820 containerd[1467]: time="2025-03-17T17:41:57.885480910Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:41:57.885820 containerd[1467]: time="2025-03-17T17:41:57.885300542Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 17:41:57.885820 containerd[1467]: time="2025-03-17T17:41:57.885383056Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 17:41:57.885820 containerd[1467]: time="2025-03-17T17:41:57.885396532Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:41:57.885820 containerd[1467]: time="2025-03-17T17:41:57.885492893Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:41:57.885820 containerd[1467]: time="2025-03-17T17:41:57.885251750Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 17:41:57.885820 containerd[1467]: time="2025-03-17T17:41:57.885322403Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 17:41:57.885820 containerd[1467]: time="2025-03-17T17:41:57.885333724Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:41:57.885820 containerd[1467]: time="2025-03-17T17:41:57.885435825Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:41:57.885820 containerd[1467]: time="2025-03-17T17:41:57.885575908Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:41:57.909512 systemd[1]: Started cri-containerd-326fb1035c4c8c401f84754a7b05ae51a98a624584b81b1ec5435387519acc53.scope - libcontainer container 326fb1035c4c8c401f84754a7b05ae51a98a624584b81b1ec5435387519acc53. Mar 17 17:41:57.911899 systemd[1]: Started cri-containerd-e2cc825235505f1101b6c2deb626545d7e80565ed853aa213e81c640396c8513.scope - libcontainer container e2cc825235505f1101b6c2deb626545d7e80565ed853aa213e81c640396c8513. Mar 17 17:41:57.915597 systemd[1]: Started cri-containerd-069e068773a6a358853407e1774aa837c14ae44c161888c8bd64013a9184a653.scope - libcontainer container 069e068773a6a358853407e1774aa837c14ae44c161888c8bd64013a9184a653. Mar 17 17:41:57.953689 containerd[1467]: time="2025-03-17T17:41:57.953635613Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:23a18e2dc14f395c5f1bea711a5a9344,Namespace:kube-system,Attempt:0,} returns sandbox id \"326fb1035c4c8c401f84754a7b05ae51a98a624584b81b1ec5435387519acc53\"" Mar 17 17:41:57.955670 containerd[1467]: time="2025-03-17T17:41:57.955602032Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:d79ab404294384d4bcc36fb5b5509bbb,Namespace:kube-system,Attempt:0,} returns sandbox id \"e2cc825235505f1101b6c2deb626545d7e80565ed853aa213e81c640396c8513\"" Mar 17 17:41:57.956250 kubelet[2300]: E0317 17:41:57.955938 2300 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:41:57.957068 containerd[1467]: time="2025-03-17T17:41:57.956972253Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:c29c7b4ac07fd8117c1d69a65632e6c1,Namespace:kube-system,Attempt:0,} returns sandbox id \"069e068773a6a358853407e1774aa837c14ae44c161888c8bd64013a9184a653\"" Mar 17 17:41:57.957567 kubelet[2300]: E0317 17:41:57.957313 2300 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:41:57.958058 kubelet[2300]: E0317 17:41:57.957976 2300 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:41:57.959954 containerd[1467]: time="2025-03-17T17:41:57.959928800Z" level=info msg="CreateContainer within sandbox \"e2cc825235505f1101b6c2deb626545d7e80565ed853aa213e81c640396c8513\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Mar 17 17:41:57.960142 containerd[1467]: time="2025-03-17T17:41:57.960108757Z" level=info msg="CreateContainer within sandbox \"069e068773a6a358853407e1774aa837c14ae44c161888c8bd64013a9184a653\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Mar 17 17:41:57.960200 containerd[1467]: time="2025-03-17T17:41:57.960113526Z" level=info msg="CreateContainer within sandbox \"326fb1035c4c8c401f84754a7b05ae51a98a624584b81b1ec5435387519acc53\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Mar 17 17:41:58.302875 containerd[1467]: time="2025-03-17T17:41:58.302752231Z" level=info msg="CreateContainer within sandbox \"326fb1035c4c8c401f84754a7b05ae51a98a624584b81b1ec5435387519acc53\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"2dcaf586964baeec65fdafe1693855e5c5ad10ee71dce356cee101b381503080\"" Mar 17 17:41:58.303645 containerd[1467]: time="2025-03-17T17:41:58.303612765Z" level=info msg="StartContainer for \"2dcaf586964baeec65fdafe1693855e5c5ad10ee71dce356cee101b381503080\"" Mar 17 17:41:58.321982 containerd[1467]: time="2025-03-17T17:41:58.321936632Z" level=info msg="CreateContainer within sandbox \"e2cc825235505f1101b6c2deb626545d7e80565ed853aa213e81c640396c8513\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"e6f5343892b24ede0359e041b14d850b61bf7722147ccfdf2ef9ba2115c04743\"" Mar 17 17:41:58.322819 containerd[1467]: time="2025-03-17T17:41:58.322615115Z" level=info msg="StartContainer for \"e6f5343892b24ede0359e041b14d850b61bf7722147ccfdf2ef9ba2115c04743\"" Mar 17 17:41:58.327500 systemd[1]: Started cri-containerd-2dcaf586964baeec65fdafe1693855e5c5ad10ee71dce356cee101b381503080.scope - libcontainer container 2dcaf586964baeec65fdafe1693855e5c5ad10ee71dce356cee101b381503080. Mar 17 17:41:58.329946 kubelet[2300]: E0317 17:41:58.329872 2300 certificate_manager.go:562] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://10.0.0.61:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 10.0.0.61:6443: connect: connection refused Mar 17 17:41:58.348495 systemd[1]: Started cri-containerd-e6f5343892b24ede0359e041b14d850b61bf7722147ccfdf2ef9ba2115c04743.scope - libcontainer container e6f5343892b24ede0359e041b14d850b61bf7722147ccfdf2ef9ba2115c04743. Mar 17 17:41:58.350833 containerd[1467]: time="2025-03-17T17:41:58.350794982Z" level=info msg="CreateContainer within sandbox \"069e068773a6a358853407e1774aa837c14ae44c161888c8bd64013a9184a653\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"c3d2467a0010ea3e2e12eabcf4bb78db444bcd75d1a9303171517d1732c4ca74\"" Mar 17 17:41:58.351811 containerd[1467]: time="2025-03-17T17:41:58.351329495Z" level=info msg="StartContainer for \"c3d2467a0010ea3e2e12eabcf4bb78db444bcd75d1a9303171517d1732c4ca74\"" Mar 17 17:41:58.381613 systemd[1]: Started cri-containerd-c3d2467a0010ea3e2e12eabcf4bb78db444bcd75d1a9303171517d1732c4ca74.scope - libcontainer container c3d2467a0010ea3e2e12eabcf4bb78db444bcd75d1a9303171517d1732c4ca74. Mar 17 17:41:58.383890 containerd[1467]: time="2025-03-17T17:41:58.383848483Z" level=info msg="StartContainer for \"2dcaf586964baeec65fdafe1693855e5c5ad10ee71dce356cee101b381503080\" returns successfully" Mar 17 17:41:58.402140 containerd[1467]: time="2025-03-17T17:41:58.402048137Z" level=info msg="StartContainer for \"e6f5343892b24ede0359e041b14d850b61bf7722147ccfdf2ef9ba2115c04743\" returns successfully" Mar 17 17:41:58.423719 containerd[1467]: time="2025-03-17T17:41:58.423578610Z" level=info msg="StartContainer for \"c3d2467a0010ea3e2e12eabcf4bb78db444bcd75d1a9303171517d1732c4ca74\" returns successfully" Mar 17 17:41:59.190813 kubelet[2300]: E0317 17:41:59.190768 2300 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:41:59.193124 kubelet[2300]: E0317 17:41:59.193098 2300 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:41:59.194247 kubelet[2300]: E0317 17:41:59.194223 2300 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:41:59.232101 kubelet[2300]: E0317 17:41:59.232058 2300 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Mar 17 17:41:59.269389 kubelet[2300]: I0317 17:41:59.269333 2300 kubelet_node_status.go:73] "Attempting to register node" node="localhost" Mar 17 17:41:59.379310 kubelet[2300]: I0317 17:41:59.379262 2300 kubelet_node_status.go:76] "Successfully registered node" node="localhost" Mar 17 17:41:59.385295 kubelet[2300]: E0317 17:41:59.385268 2300 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 17 17:41:59.485744 kubelet[2300]: E0317 17:41:59.485644 2300 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 17 17:41:59.586554 kubelet[2300]: E0317 17:41:59.586500 2300 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 17 17:41:59.687448 kubelet[2300]: E0317 17:41:59.687405 2300 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 17 17:41:59.788262 kubelet[2300]: E0317 17:41:59.788143 2300 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 17 17:41:59.888691 kubelet[2300]: E0317 17:41:59.888646 2300 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 17 17:41:59.989181 kubelet[2300]: E0317 17:41:59.989143 2300 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 17 17:42:00.089760 kubelet[2300]: E0317 17:42:00.089650 2300 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 17 17:42:00.190561 kubelet[2300]: E0317 17:42:00.190526 2300 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 17 17:42:00.196022 kubelet[2300]: E0317 17:42:00.195996 2300 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:00.196412 kubelet[2300]: E0317 17:42:00.196040 2300 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:00.196412 kubelet[2300]: E0317 17:42:00.196040 2300 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:00.291139 kubelet[2300]: E0317 17:42:00.291083 2300 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 17 17:42:00.391692 kubelet[2300]: E0317 17:42:00.391580 2300 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 17 17:42:00.492274 kubelet[2300]: E0317 17:42:00.492232 2300 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 17 17:42:00.593019 kubelet[2300]: E0317 17:42:00.592975 2300 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 17 17:42:00.693550 kubelet[2300]: E0317 17:42:00.693406 2300 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 17 17:42:00.793916 kubelet[2300]: E0317 17:42:00.793881 2300 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 17 17:42:01.151048 kubelet[2300]: I0317 17:42:01.150938 2300 apiserver.go:52] "Watching apiserver" Mar 17 17:42:01.161975 kubelet[2300]: I0317 17:42:01.161936 2300 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" Mar 17 17:42:01.204178 kubelet[2300]: E0317 17:42:01.203810 2300 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:01.204972 kubelet[2300]: E0317 17:42:01.204943 2300 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:01.226180 systemd[1]: Reloading requested from client PID 2578 ('systemctl') (unit session-9.scope)... Mar 17 17:42:01.226195 systemd[1]: Reloading... Mar 17 17:42:01.308396 zram_generator::config[2617]: No configuration found. Mar 17 17:42:01.417331 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 17 17:42:01.509876 systemd[1]: Reloading finished in 283 ms. Mar 17 17:42:01.551147 kubelet[2300]: I0317 17:42:01.551063 2300 dynamic_cafile_content.go:171] "Shutting down controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 17 17:42:01.551163 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Mar 17 17:42:01.568797 systemd[1]: kubelet.service: Deactivated successfully. Mar 17 17:42:01.569102 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 17 17:42:01.583589 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 17 17:42:01.730961 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 17 17:42:01.736342 (kubelet)[2662]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Mar 17 17:42:01.776506 kubelet[2662]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 17 17:42:01.776506 kubelet[2662]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Mar 17 17:42:01.776506 kubelet[2662]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 17 17:42:01.776893 kubelet[2662]: I0317 17:42:01.776551 2662 server.go:205] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 17 17:42:01.783288 kubelet[2662]: I0317 17:42:01.783252 2662 server.go:484] "Kubelet version" kubeletVersion="v1.30.1" Mar 17 17:42:01.783288 kubelet[2662]: I0317 17:42:01.783281 2662 server.go:486] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 17 17:42:01.783509 kubelet[2662]: I0317 17:42:01.783487 2662 server.go:927] "Client rotation is on, will bootstrap in background" Mar 17 17:42:01.784782 kubelet[2662]: I0317 17:42:01.784760 2662 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Mar 17 17:42:01.785969 kubelet[2662]: I0317 17:42:01.785846 2662 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 17 17:42:01.792805 kubelet[2662]: I0317 17:42:01.792777 2662 server.go:742] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Mar 17 17:42:01.793021 kubelet[2662]: I0317 17:42:01.792983 2662 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 17 17:42:01.793168 kubelet[2662]: I0317 17:42:01.793012 2662 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Mar 17 17:42:01.793246 kubelet[2662]: I0317 17:42:01.793180 2662 topology_manager.go:138] "Creating topology manager with none policy" Mar 17 17:42:01.793246 kubelet[2662]: I0317 17:42:01.793191 2662 container_manager_linux.go:301] "Creating device plugin manager" Mar 17 17:42:01.793246 kubelet[2662]: I0317 17:42:01.793237 2662 state_mem.go:36] "Initialized new in-memory state store" Mar 17 17:42:01.793348 kubelet[2662]: I0317 17:42:01.793333 2662 kubelet.go:400] "Attempting to sync node with API server" Mar 17 17:42:01.793348 kubelet[2662]: I0317 17:42:01.793346 2662 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 17 17:42:01.793429 kubelet[2662]: I0317 17:42:01.793383 2662 kubelet.go:312] "Adding apiserver pod source" Mar 17 17:42:01.793429 kubelet[2662]: I0317 17:42:01.793409 2662 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 17 17:42:01.794403 kubelet[2662]: I0317 17:42:01.794215 2662 kuberuntime_manager.go:261] "Container runtime initialized" containerRuntime="containerd" version="v1.7.23" apiVersion="v1" Mar 17 17:42:01.794536 kubelet[2662]: I0317 17:42:01.794444 2662 kubelet.go:815] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Mar 17 17:42:01.794874 kubelet[2662]: I0317 17:42:01.794850 2662 server.go:1264] "Started kubelet" Mar 17 17:42:01.796875 kubelet[2662]: I0317 17:42:01.796826 2662 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 17 17:42:01.797278 kubelet[2662]: I0317 17:42:01.797258 2662 server.go:227] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 17 17:42:01.798841 kubelet[2662]: I0317 17:42:01.798795 2662 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Mar 17 17:42:01.800301 kubelet[2662]: I0317 17:42:01.800281 2662 server.go:455] "Adding debug handlers to kubelet server" Mar 17 17:42:01.805502 kubelet[2662]: I0317 17:42:01.805464 2662 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 17 17:42:01.805860 kubelet[2662]: E0317 17:42:01.805750 2662 kubelet.go:1467] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Mar 17 17:42:01.805994 kubelet[2662]: I0317 17:42:01.805980 2662 volume_manager.go:291] "Starting Kubelet Volume Manager" Mar 17 17:42:01.806715 kubelet[2662]: I0317 17:42:01.806493 2662 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Mar 17 17:42:01.806715 kubelet[2662]: I0317 17:42:01.806665 2662 reconciler.go:26] "Reconciler: start to sync state" Mar 17 17:42:01.807557 kubelet[2662]: I0317 17:42:01.807536 2662 factory.go:221] Registration of the systemd container factory successfully Mar 17 17:42:01.807659 kubelet[2662]: I0317 17:42:01.807632 2662 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Mar 17 17:42:01.810596 kubelet[2662]: I0317 17:42:01.810559 2662 factory.go:221] Registration of the containerd container factory successfully Mar 17 17:42:01.818787 kubelet[2662]: I0317 17:42:01.818717 2662 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Mar 17 17:42:01.820371 kubelet[2662]: I0317 17:42:01.820326 2662 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Mar 17 17:42:01.820427 kubelet[2662]: I0317 17:42:01.820384 2662 status_manager.go:217] "Starting to sync pod status with apiserver" Mar 17 17:42:01.820427 kubelet[2662]: I0317 17:42:01.820408 2662 kubelet.go:2337] "Starting kubelet main sync loop" Mar 17 17:42:01.820484 kubelet[2662]: E0317 17:42:01.820455 2662 kubelet.go:2361] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 17 17:42:01.846787 kubelet[2662]: I0317 17:42:01.846756 2662 cpu_manager.go:214] "Starting CPU manager" policy="none" Mar 17 17:42:01.846787 kubelet[2662]: I0317 17:42:01.846771 2662 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Mar 17 17:42:01.846787 kubelet[2662]: I0317 17:42:01.846787 2662 state_mem.go:36] "Initialized new in-memory state store" Mar 17 17:42:01.846966 kubelet[2662]: I0317 17:42:01.846944 2662 state_mem.go:88] "Updated default CPUSet" cpuSet="" Mar 17 17:42:01.846989 kubelet[2662]: I0317 17:42:01.846959 2662 state_mem.go:96] "Updated CPUSet assignments" assignments={} Mar 17 17:42:01.846989 kubelet[2662]: I0317 17:42:01.846975 2662 policy_none.go:49] "None policy: Start" Mar 17 17:42:01.847554 kubelet[2662]: I0317 17:42:01.847530 2662 memory_manager.go:170] "Starting memorymanager" policy="None" Mar 17 17:42:01.847554 kubelet[2662]: I0317 17:42:01.847551 2662 state_mem.go:35] "Initializing new in-memory state store" Mar 17 17:42:01.847663 kubelet[2662]: I0317 17:42:01.847650 2662 state_mem.go:75] "Updated machine memory state" Mar 17 17:42:01.851491 kubelet[2662]: I0317 17:42:01.851465 2662 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Mar 17 17:42:01.852577 kubelet[2662]: I0317 17:42:01.851829 2662 container_log_manager.go:186] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 17 17:42:01.852577 kubelet[2662]: I0317 17:42:01.852211 2662 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 17 17:42:01.910428 kubelet[2662]: I0317 17:42:01.910379 2662 kubelet_node_status.go:73] "Attempting to register node" node="localhost" Mar 17 17:42:01.916285 kubelet[2662]: I0317 17:42:01.916250 2662 kubelet_node_status.go:112] "Node was previously registered" node="localhost" Mar 17 17:42:01.916285 kubelet[2662]: I0317 17:42:01.916304 2662 kubelet_node_status.go:76] "Successfully registered node" node="localhost" Mar 17 17:42:01.921163 kubelet[2662]: I0317 17:42:01.921124 2662 topology_manager.go:215] "Topology Admit Handler" podUID="23a18e2dc14f395c5f1bea711a5a9344" podNamespace="kube-system" podName="kube-controller-manager-localhost" Mar 17 17:42:01.921227 kubelet[2662]: I0317 17:42:01.921211 2662 topology_manager.go:215] "Topology Admit Handler" podUID="d79ab404294384d4bcc36fb5b5509bbb" podNamespace="kube-system" podName="kube-scheduler-localhost" Mar 17 17:42:01.921274 kubelet[2662]: I0317 17:42:01.921260 2662 topology_manager.go:215] "Topology Admit Handler" podUID="c29c7b4ac07fd8117c1d69a65632e6c1" podNamespace="kube-system" podName="kube-apiserver-localhost" Mar 17 17:42:01.926235 kubelet[2662]: E0317 17:42:01.926180 2662 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Mar 17 17:42:01.926235 kubelet[2662]: E0317 17:42:01.926220 2662 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-localhost\" already exists" pod="kube-system/kube-controller-manager-localhost" Mar 17 17:42:02.007631 kubelet[2662]: I0317 17:42:02.007540 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/c29c7b4ac07fd8117c1d69a65632e6c1-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"c29c7b4ac07fd8117c1d69a65632e6c1\") " pod="kube-system/kube-apiserver-localhost" Mar 17 17:42:02.007631 kubelet[2662]: I0317 17:42:02.007567 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/c29c7b4ac07fd8117c1d69a65632e6c1-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"c29c7b4ac07fd8117c1d69a65632e6c1\") " pod="kube-system/kube-apiserver-localhost" Mar 17 17:42:02.007631 kubelet[2662]: I0317 17:42:02.007591 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/23a18e2dc14f395c5f1bea711a5a9344-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"23a18e2dc14f395c5f1bea711a5a9344\") " pod="kube-system/kube-controller-manager-localhost" Mar 17 17:42:02.007631 kubelet[2662]: I0317 17:42:02.007606 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/23a18e2dc14f395c5f1bea711a5a9344-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"23a18e2dc14f395c5f1bea711a5a9344\") " pod="kube-system/kube-controller-manager-localhost" Mar 17 17:42:02.007631 kubelet[2662]: I0317 17:42:02.007627 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/23a18e2dc14f395c5f1bea711a5a9344-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"23a18e2dc14f395c5f1bea711a5a9344\") " pod="kube-system/kube-controller-manager-localhost" Mar 17 17:42:02.007786 kubelet[2662]: I0317 17:42:02.007691 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/d79ab404294384d4bcc36fb5b5509bbb-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"d79ab404294384d4bcc36fb5b5509bbb\") " pod="kube-system/kube-scheduler-localhost" Mar 17 17:42:02.007786 kubelet[2662]: I0317 17:42:02.007722 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/23a18e2dc14f395c5f1bea711a5a9344-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"23a18e2dc14f395c5f1bea711a5a9344\") " pod="kube-system/kube-controller-manager-localhost" Mar 17 17:42:02.007786 kubelet[2662]: I0317 17:42:02.007741 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/23a18e2dc14f395c5f1bea711a5a9344-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"23a18e2dc14f395c5f1bea711a5a9344\") " pod="kube-system/kube-controller-manager-localhost" Mar 17 17:42:02.007786 kubelet[2662]: I0317 17:42:02.007755 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/c29c7b4ac07fd8117c1d69a65632e6c1-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"c29c7b4ac07fd8117c1d69a65632e6c1\") " pod="kube-system/kube-apiserver-localhost" Mar 17 17:42:02.226047 kubelet[2662]: E0317 17:42:02.226012 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:02.226752 kubelet[2662]: E0317 17:42:02.226638 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:02.226867 kubelet[2662]: E0317 17:42:02.226763 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:02.794015 kubelet[2662]: I0317 17:42:02.793952 2662 apiserver.go:52] "Watching apiserver" Mar 17 17:42:02.807736 kubelet[2662]: I0317 17:42:02.806759 2662 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" Mar 17 17:42:02.835422 kubelet[2662]: E0317 17:42:02.835340 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:02.839776 kubelet[2662]: E0317 17:42:02.839734 2662 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost" Mar 17 17:42:02.840009 kubelet[2662]: E0317 17:42:02.839950 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:02.845969 kubelet[2662]: E0317 17:42:02.845496 2662 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Mar 17 17:42:02.845969 kubelet[2662]: E0317 17:42:02.845900 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:02.854769 kubelet[2662]: I0317 17:42:02.854702 2662 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.8546865769999998 podStartE2EDuration="1.854686577s" podCreationTimestamp="2025-03-17 17:42:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-17 17:42:02.854034779 +0000 UTC m=+1.114001505" watchObservedRunningTime="2025-03-17 17:42:02.854686577 +0000 UTC m=+1.114653303" Mar 17 17:42:02.862289 kubelet[2662]: I0317 17:42:02.862239 2662 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=1.862221541 podStartE2EDuration="1.862221541s" podCreationTimestamp="2025-03-17 17:42:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-17 17:42:02.862066805 +0000 UTC m=+1.122033531" watchObservedRunningTime="2025-03-17 17:42:02.862221541 +0000 UTC m=+1.122188267" Mar 17 17:42:03.836089 kubelet[2662]: E0317 17:42:03.836047 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:03.836608 kubelet[2662]: E0317 17:42:03.836412 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:04.837675 kubelet[2662]: E0317 17:42:04.837626 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:06.183965 sudo[1668]: pam_unix(sudo:session): session closed for user root Mar 17 17:42:06.185438 sshd[1667]: Connection closed by 10.0.0.1 port 52988 Mar 17 17:42:06.185913 sshd-session[1664]: pam_unix(sshd:session): session closed for user core Mar 17 17:42:06.190033 systemd[1]: sshd@8-10.0.0.61:22-10.0.0.1:52988.service: Deactivated successfully. Mar 17 17:42:06.192108 systemd[1]: session-9.scope: Deactivated successfully. Mar 17 17:42:06.192309 systemd[1]: session-9.scope: Consumed 4.166s CPU time, 189.0M memory peak, 0B memory swap peak. Mar 17 17:42:06.192823 systemd-logind[1453]: Session 9 logged out. Waiting for processes to exit. Mar 17 17:42:06.193748 systemd-logind[1453]: Removed session 9. Mar 17 17:42:10.422125 kubelet[2662]: E0317 17:42:10.422077 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:10.485907 kubelet[2662]: I0317 17:42:10.485842 2662 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=9.485826482 podStartE2EDuration="9.485826482s" podCreationTimestamp="2025-03-17 17:42:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-17 17:42:02.869656463 +0000 UTC m=+1.129623189" watchObservedRunningTime="2025-03-17 17:42:10.485826482 +0000 UTC m=+8.745793208" Mar 17 17:42:10.848942 kubelet[2662]: E0317 17:42:10.848778 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:11.843183 kubelet[2662]: E0317 17:42:11.843150 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:11.849579 kubelet[2662]: E0317 17:42:11.849564 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:13.101939 update_engine[1454]: I20250317 17:42:13.101832 1454 update_attempter.cc:509] Updating boot flags... Mar 17 17:42:13.133396 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 40 scanned by (udev-worker) (2758) Mar 17 17:42:13.172848 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 40 scanned by (udev-worker) (2758) Mar 17 17:42:13.211421 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 40 scanned by (udev-worker) (2758) Mar 17 17:42:14.788174 kubelet[2662]: E0317 17:42:14.788138 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:15.397011 kubelet[2662]: I0317 17:42:15.396979 2662 kuberuntime_manager.go:1523] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Mar 17 17:42:15.397298 containerd[1467]: time="2025-03-17T17:42:15.397261047Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Mar 17 17:42:15.397714 kubelet[2662]: I0317 17:42:15.397398 2662 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Mar 17 17:42:16.234307 kubelet[2662]: I0317 17:42:16.234244 2662 topology_manager.go:215] "Topology Admit Handler" podUID="1720f620-1a0d-43dc-b178-f98905c67388" podNamespace="kube-system" podName="kube-proxy-kzfsj" Mar 17 17:42:16.242313 systemd[1]: Created slice kubepods-besteffort-pod1720f620_1a0d_43dc_b178_f98905c67388.slice - libcontainer container kubepods-besteffort-pod1720f620_1a0d_43dc_b178_f98905c67388.slice. Mar 17 17:42:16.394783 kubelet[2662]: I0317 17:42:16.394665 2662 topology_manager.go:215] "Topology Admit Handler" podUID="e2c3de02-a4aa-4f16-8bec-0f7b0cdfcf86" podNamespace="tigera-operator" podName="tigera-operator-6479d6dc54-8khns" Mar 17 17:42:16.396238 kubelet[2662]: I0317 17:42:16.395994 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/1720f620-1a0d-43dc-b178-f98905c67388-kube-proxy\") pod \"kube-proxy-kzfsj\" (UID: \"1720f620-1a0d-43dc-b178-f98905c67388\") " pod="kube-system/kube-proxy-kzfsj" Mar 17 17:42:16.396238 kubelet[2662]: I0317 17:42:16.396025 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1720f620-1a0d-43dc-b178-f98905c67388-lib-modules\") pod \"kube-proxy-kzfsj\" (UID: \"1720f620-1a0d-43dc-b178-f98905c67388\") " pod="kube-system/kube-proxy-kzfsj" Mar 17 17:42:16.396238 kubelet[2662]: I0317 17:42:16.396043 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/1720f620-1a0d-43dc-b178-f98905c67388-xtables-lock\") pod \"kube-proxy-kzfsj\" (UID: \"1720f620-1a0d-43dc-b178-f98905c67388\") " pod="kube-system/kube-proxy-kzfsj" Mar 17 17:42:16.396238 kubelet[2662]: I0317 17:42:16.396057 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ft2nt\" (UniqueName: \"kubernetes.io/projected/1720f620-1a0d-43dc-b178-f98905c67388-kube-api-access-ft2nt\") pod \"kube-proxy-kzfsj\" (UID: \"1720f620-1a0d-43dc-b178-f98905c67388\") " pod="kube-system/kube-proxy-kzfsj" Mar 17 17:42:16.402440 systemd[1]: Created slice kubepods-besteffort-pode2c3de02_a4aa_4f16_8bec_0f7b0cdfcf86.slice - libcontainer container kubepods-besteffort-pode2c3de02_a4aa_4f16_8bec_0f7b0cdfcf86.slice. Mar 17 17:42:16.496932 kubelet[2662]: I0317 17:42:16.496782 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/e2c3de02-a4aa-4f16-8bec-0f7b0cdfcf86-var-lib-calico\") pod \"tigera-operator-6479d6dc54-8khns\" (UID: \"e2c3de02-a4aa-4f16-8bec-0f7b0cdfcf86\") " pod="tigera-operator/tigera-operator-6479d6dc54-8khns" Mar 17 17:42:16.496932 kubelet[2662]: I0317 17:42:16.496856 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmwwv\" (UniqueName: \"kubernetes.io/projected/e2c3de02-a4aa-4f16-8bec-0f7b0cdfcf86-kube-api-access-fmwwv\") pod \"tigera-operator-6479d6dc54-8khns\" (UID: \"e2c3de02-a4aa-4f16-8bec-0f7b0cdfcf86\") " pod="tigera-operator/tigera-operator-6479d6dc54-8khns" Mar 17 17:42:16.549321 kubelet[2662]: E0317 17:42:16.549265 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:16.550002 containerd[1467]: time="2025-03-17T17:42:16.549955926Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-kzfsj,Uid:1720f620-1a0d-43dc-b178-f98905c67388,Namespace:kube-system,Attempt:0,}" Mar 17 17:42:16.577518 containerd[1467]: time="2025-03-17T17:42:16.577376061Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 17:42:16.577518 containerd[1467]: time="2025-03-17T17:42:16.577445282Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 17:42:16.577518 containerd[1467]: time="2025-03-17T17:42:16.577466302Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:42:16.577744 containerd[1467]: time="2025-03-17T17:42:16.577580718Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:42:16.604651 systemd[1]: Started cri-containerd-bf24ebce1199a8030567837adecc765abfd9af044aefee591ac2c97c43012305.scope - libcontainer container bf24ebce1199a8030567837adecc765abfd9af044aefee591ac2c97c43012305. Mar 17 17:42:16.632751 containerd[1467]: time="2025-03-17T17:42:16.632703062Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-kzfsj,Uid:1720f620-1a0d-43dc-b178-f98905c67388,Namespace:kube-system,Attempt:0,} returns sandbox id \"bf24ebce1199a8030567837adecc765abfd9af044aefee591ac2c97c43012305\"" Mar 17 17:42:16.633579 kubelet[2662]: E0317 17:42:16.633531 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:16.637135 containerd[1467]: time="2025-03-17T17:42:16.637081479Z" level=info msg="CreateContainer within sandbox \"bf24ebce1199a8030567837adecc765abfd9af044aefee591ac2c97c43012305\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Mar 17 17:42:16.661474 containerd[1467]: time="2025-03-17T17:42:16.661389524Z" level=info msg="CreateContainer within sandbox \"bf24ebce1199a8030567837adecc765abfd9af044aefee591ac2c97c43012305\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"26963989e14b252700c09716602c5f9b18a2f4f189d33f8710948fcaa15fc2b1\"" Mar 17 17:42:16.662202 containerd[1467]: time="2025-03-17T17:42:16.662154482Z" level=info msg="StartContainer for \"26963989e14b252700c09716602c5f9b18a2f4f189d33f8710948fcaa15fc2b1\"" Mar 17 17:42:16.698716 systemd[1]: Started cri-containerd-26963989e14b252700c09716602c5f9b18a2f4f189d33f8710948fcaa15fc2b1.scope - libcontainer container 26963989e14b252700c09716602c5f9b18a2f4f189d33f8710948fcaa15fc2b1. Mar 17 17:42:16.705464 containerd[1467]: time="2025-03-17T17:42:16.705278488Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-6479d6dc54-8khns,Uid:e2c3de02-a4aa-4f16-8bec-0f7b0cdfcf86,Namespace:tigera-operator,Attempt:0,}" Mar 17 17:42:16.732433 containerd[1467]: time="2025-03-17T17:42:16.732227532Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 17:42:16.732433 containerd[1467]: time="2025-03-17T17:42:16.732309497Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 17:42:16.732433 containerd[1467]: time="2025-03-17T17:42:16.732325989Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:42:16.732756 containerd[1467]: time="2025-03-17T17:42:16.732462978Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:42:16.736180 containerd[1467]: time="2025-03-17T17:42:16.736055186Z" level=info msg="StartContainer for \"26963989e14b252700c09716602c5f9b18a2f4f189d33f8710948fcaa15fc2b1\" returns successfully" Mar 17 17:42:16.756573 systemd[1]: Started cri-containerd-60d65e0f9f758d3c96c8ccfe41a18eeb0656d79babfcfc505361c0ef2c01cb0f.scope - libcontainer container 60d65e0f9f758d3c96c8ccfe41a18eeb0656d79babfcfc505361c0ef2c01cb0f. Mar 17 17:42:16.794667 containerd[1467]: time="2025-03-17T17:42:16.794619224Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-6479d6dc54-8khns,Uid:e2c3de02-a4aa-4f16-8bec-0f7b0cdfcf86,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"60d65e0f9f758d3c96c8ccfe41a18eeb0656d79babfcfc505361c0ef2c01cb0f\"" Mar 17 17:42:16.796784 containerd[1467]: time="2025-03-17T17:42:16.796738114Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.36.5\"" Mar 17 17:42:16.862002 kubelet[2662]: E0317 17:42:16.861953 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:16.871106 kubelet[2662]: I0317 17:42:16.871034 2662 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-kzfsj" podStartSLOduration=0.871010942 podStartE2EDuration="871.010942ms" podCreationTimestamp="2025-03-17 17:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-17 17:42:16.870795745 +0000 UTC m=+15.130762472" watchObservedRunningTime="2025-03-17 17:42:16.871010942 +0000 UTC m=+15.130977668" Mar 17 17:42:18.321383 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4288895504.mount: Deactivated successfully. Mar 17 17:42:19.788057 containerd[1467]: time="2025-03-17T17:42:19.787997611Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.36.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:42:19.788798 containerd[1467]: time="2025-03-17T17:42:19.788739734Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.36.5: active requests=0, bytes read=21945008" Mar 17 17:42:19.789991 containerd[1467]: time="2025-03-17T17:42:19.789961081Z" level=info msg="ImageCreate event name:\"sha256:dc4a8a56c133edb1bc4c3d6bc94bcd96f2bde82413370cb1783ac2d7f3a46d53\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:42:19.792023 containerd[1467]: time="2025-03-17T17:42:19.791994664Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:3341fa9475c0325b86228c8726389f9bae9fd6c430c66fe5cd5dc39d7bb6ad4b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:42:19.792761 containerd[1467]: time="2025-03-17T17:42:19.792718893Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.36.5\" with image id \"sha256:dc4a8a56c133edb1bc4c3d6bc94bcd96f2bde82413370cb1783ac2d7f3a46d53\", repo tag \"quay.io/tigera/operator:v1.36.5\", repo digest \"quay.io/tigera/operator@sha256:3341fa9475c0325b86228c8726389f9bae9fd6c430c66fe5cd5dc39d7bb6ad4b\", size \"21941003\" in 2.995939321s" Mar 17 17:42:19.792795 containerd[1467]: time="2025-03-17T17:42:19.792759279Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.36.5\" returns image reference \"sha256:dc4a8a56c133edb1bc4c3d6bc94bcd96f2bde82413370cb1783ac2d7f3a46d53\"" Mar 17 17:42:19.794917 containerd[1467]: time="2025-03-17T17:42:19.794864047Z" level=info msg="CreateContainer within sandbox \"60d65e0f9f758d3c96c8ccfe41a18eeb0656d79babfcfc505361c0ef2c01cb0f\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Mar 17 17:42:19.807732 containerd[1467]: time="2025-03-17T17:42:19.807699542Z" level=info msg="CreateContainer within sandbox \"60d65e0f9f758d3c96c8ccfe41a18eeb0656d79babfcfc505361c0ef2c01cb0f\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"70dc8952d933cafdbf8aba2e43e7fb2471a24a2464a32754676d95108da50ff1\"" Mar 17 17:42:19.808071 containerd[1467]: time="2025-03-17T17:42:19.808039935Z" level=info msg="StartContainer for \"70dc8952d933cafdbf8aba2e43e7fb2471a24a2464a32754676d95108da50ff1\"" Mar 17 17:42:19.836494 systemd[1]: Started cri-containerd-70dc8952d933cafdbf8aba2e43e7fb2471a24a2464a32754676d95108da50ff1.scope - libcontainer container 70dc8952d933cafdbf8aba2e43e7fb2471a24a2464a32754676d95108da50ff1. Mar 17 17:42:19.861099 containerd[1467]: time="2025-03-17T17:42:19.861053905Z" level=info msg="StartContainer for \"70dc8952d933cafdbf8aba2e43e7fb2471a24a2464a32754676d95108da50ff1\" returns successfully" Mar 17 17:42:22.688931 kubelet[2662]: I0317 17:42:22.688801 2662 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-6479d6dc54-8khns" podStartSLOduration=3.691291758 podStartE2EDuration="6.688778213s" podCreationTimestamp="2025-03-17 17:42:16 +0000 UTC" firstStartedPulling="2025-03-17 17:42:16.796004346 +0000 UTC m=+15.055971072" lastFinishedPulling="2025-03-17 17:42:19.793490801 +0000 UTC m=+18.053457527" observedRunningTime="2025-03-17 17:42:19.87656235 +0000 UTC m=+18.136529076" watchObservedRunningTime="2025-03-17 17:42:22.688778213 +0000 UTC m=+20.948744939" Mar 17 17:42:22.691507 kubelet[2662]: I0317 17:42:22.691470 2662 topology_manager.go:215] "Topology Admit Handler" podUID="c1834c71-5ae6-4d93-b7af-2c16faf191c5" podNamespace="calico-system" podName="calico-typha-7b7cf6c67c-kdn9m" Mar 17 17:42:22.701551 systemd[1]: Created slice kubepods-besteffort-podc1834c71_5ae6_4d93_b7af_2c16faf191c5.slice - libcontainer container kubepods-besteffort-podc1834c71_5ae6_4d93_b7af_2c16faf191c5.slice. Mar 17 17:42:22.736121 kubelet[2662]: I0317 17:42:22.736057 2662 topology_manager.go:215] "Topology Admit Handler" podUID="77e9b5a5-a1f1-40db-82d5-f5ed3dcf3fa2" podNamespace="calico-system" podName="calico-node-fbl5x" Mar 17 17:42:22.744329 systemd[1]: Created slice kubepods-besteffort-pod77e9b5a5_a1f1_40db_82d5_f5ed3dcf3fa2.slice - libcontainer container kubepods-besteffort-pod77e9b5a5_a1f1_40db_82d5_f5ed3dcf3fa2.slice. Mar 17 17:42:22.838467 kubelet[2662]: I0317 17:42:22.838422 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/77e9b5a5-a1f1-40db-82d5-f5ed3dcf3fa2-cni-net-dir\") pod \"calico-node-fbl5x\" (UID: \"77e9b5a5-a1f1-40db-82d5-f5ed3dcf3fa2\") " pod="calico-system/calico-node-fbl5x" Mar 17 17:42:22.838467 kubelet[2662]: I0317 17:42:22.838469 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fd2mj\" (UniqueName: \"kubernetes.io/projected/c1834c71-5ae6-4d93-b7af-2c16faf191c5-kube-api-access-fd2mj\") pod \"calico-typha-7b7cf6c67c-kdn9m\" (UID: \"c1834c71-5ae6-4d93-b7af-2c16faf191c5\") " pod="calico-system/calico-typha-7b7cf6c67c-kdn9m" Mar 17 17:42:22.838677 kubelet[2662]: I0317 17:42:22.838488 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/77e9b5a5-a1f1-40db-82d5-f5ed3dcf3fa2-xtables-lock\") pod \"calico-node-fbl5x\" (UID: \"77e9b5a5-a1f1-40db-82d5-f5ed3dcf3fa2\") " pod="calico-system/calico-node-fbl5x" Mar 17 17:42:22.838677 kubelet[2662]: I0317 17:42:22.838503 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/77e9b5a5-a1f1-40db-82d5-f5ed3dcf3fa2-tigera-ca-bundle\") pod \"calico-node-fbl5x\" (UID: \"77e9b5a5-a1f1-40db-82d5-f5ed3dcf3fa2\") " pod="calico-system/calico-node-fbl5x" Mar 17 17:42:22.838677 kubelet[2662]: I0317 17:42:22.838518 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/77e9b5a5-a1f1-40db-82d5-f5ed3dcf3fa2-node-certs\") pod \"calico-node-fbl5x\" (UID: \"77e9b5a5-a1f1-40db-82d5-f5ed3dcf3fa2\") " pod="calico-system/calico-node-fbl5x" Mar 17 17:42:22.838677 kubelet[2662]: I0317 17:42:22.838533 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/77e9b5a5-a1f1-40db-82d5-f5ed3dcf3fa2-cni-bin-dir\") pod \"calico-node-fbl5x\" (UID: \"77e9b5a5-a1f1-40db-82d5-f5ed3dcf3fa2\") " pod="calico-system/calico-node-fbl5x" Mar 17 17:42:22.838677 kubelet[2662]: I0317 17:42:22.838549 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/77e9b5a5-a1f1-40db-82d5-f5ed3dcf3fa2-cni-log-dir\") pod \"calico-node-fbl5x\" (UID: \"77e9b5a5-a1f1-40db-82d5-f5ed3dcf3fa2\") " pod="calico-system/calico-node-fbl5x" Mar 17 17:42:22.838844 kubelet[2662]: I0317 17:42:22.838652 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/77e9b5a5-a1f1-40db-82d5-f5ed3dcf3fa2-var-lib-calico\") pod \"calico-node-fbl5x\" (UID: \"77e9b5a5-a1f1-40db-82d5-f5ed3dcf3fa2\") " pod="calico-system/calico-node-fbl5x" Mar 17 17:42:22.838844 kubelet[2662]: I0317 17:42:22.838748 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/c1834c71-5ae6-4d93-b7af-2c16faf191c5-typha-certs\") pod \"calico-typha-7b7cf6c67c-kdn9m\" (UID: \"c1834c71-5ae6-4d93-b7af-2c16faf191c5\") " pod="calico-system/calico-typha-7b7cf6c67c-kdn9m" Mar 17 17:42:22.838844 kubelet[2662]: I0317 17:42:22.838798 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/77e9b5a5-a1f1-40db-82d5-f5ed3dcf3fa2-lib-modules\") pod \"calico-node-fbl5x\" (UID: \"77e9b5a5-a1f1-40db-82d5-f5ed3dcf3fa2\") " pod="calico-system/calico-node-fbl5x" Mar 17 17:42:22.838844 kubelet[2662]: I0317 17:42:22.838817 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/77e9b5a5-a1f1-40db-82d5-f5ed3dcf3fa2-var-run-calico\") pod \"calico-node-fbl5x\" (UID: \"77e9b5a5-a1f1-40db-82d5-f5ed3dcf3fa2\") " pod="calico-system/calico-node-fbl5x" Mar 17 17:42:22.838844 kubelet[2662]: I0317 17:42:22.838836 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvf9s\" (UniqueName: \"kubernetes.io/projected/77e9b5a5-a1f1-40db-82d5-f5ed3dcf3fa2-kube-api-access-lvf9s\") pod \"calico-node-fbl5x\" (UID: \"77e9b5a5-a1f1-40db-82d5-f5ed3dcf3fa2\") " pod="calico-system/calico-node-fbl5x" Mar 17 17:42:22.839015 kubelet[2662]: I0317 17:42:22.838885 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c1834c71-5ae6-4d93-b7af-2c16faf191c5-tigera-ca-bundle\") pod \"calico-typha-7b7cf6c67c-kdn9m\" (UID: \"c1834c71-5ae6-4d93-b7af-2c16faf191c5\") " pod="calico-system/calico-typha-7b7cf6c67c-kdn9m" Mar 17 17:42:22.839015 kubelet[2662]: I0317 17:42:22.838925 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/77e9b5a5-a1f1-40db-82d5-f5ed3dcf3fa2-policysync\") pod \"calico-node-fbl5x\" (UID: \"77e9b5a5-a1f1-40db-82d5-f5ed3dcf3fa2\") " pod="calico-system/calico-node-fbl5x" Mar 17 17:42:22.839015 kubelet[2662]: I0317 17:42:22.838964 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/77e9b5a5-a1f1-40db-82d5-f5ed3dcf3fa2-flexvol-driver-host\") pod \"calico-node-fbl5x\" (UID: \"77e9b5a5-a1f1-40db-82d5-f5ed3dcf3fa2\") " pod="calico-system/calico-node-fbl5x" Mar 17 17:42:22.845006 kubelet[2662]: I0317 17:42:22.844070 2662 topology_manager.go:215] "Topology Admit Handler" podUID="27eb0c6c-40f4-4b19-b513-0003d63df5d0" podNamespace="calico-system" podName="csi-node-driver-j9gdg" Mar 17 17:42:22.845006 kubelet[2662]: E0317 17:42:22.844810 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-j9gdg" podUID="27eb0c6c-40f4-4b19-b513-0003d63df5d0" Mar 17 17:42:22.942810 kubelet[2662]: E0317 17:42:22.941722 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.942810 kubelet[2662]: W0317 17:42:22.941769 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.942810 kubelet[2662]: E0317 17:42:22.941794 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.942810 kubelet[2662]: E0317 17:42:22.942063 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.942810 kubelet[2662]: W0317 17:42:22.942093 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.942810 kubelet[2662]: E0317 17:42:22.942112 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.942810 kubelet[2662]: E0317 17:42:22.942380 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.942810 kubelet[2662]: W0317 17:42:22.942389 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.942810 kubelet[2662]: E0317 17:42:22.942414 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.942810 kubelet[2662]: E0317 17:42:22.942665 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.943124 kubelet[2662]: W0317 17:42:22.942674 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.943124 kubelet[2662]: E0317 17:42:22.942722 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.943124 kubelet[2662]: E0317 17:42:22.943045 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.943124 kubelet[2662]: W0317 17:42:22.943053 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.943124 kubelet[2662]: E0317 17:42:22.943120 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.943584 kubelet[2662]: E0317 17:42:22.943332 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.943584 kubelet[2662]: W0317 17:42:22.943341 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.943584 kubelet[2662]: E0317 17:42:22.943470 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.943675 kubelet[2662]: E0317 17:42:22.943630 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.943675 kubelet[2662]: W0317 17:42:22.943637 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.945385 kubelet[2662]: E0317 17:42:22.943742 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.945385 kubelet[2662]: E0317 17:42:22.943911 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.945385 kubelet[2662]: W0317 17:42:22.943918 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.945385 kubelet[2662]: E0317 17:42:22.943971 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.945385 kubelet[2662]: E0317 17:42:22.944903 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.945385 kubelet[2662]: W0317 17:42:22.944911 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.945385 kubelet[2662]: E0317 17:42:22.944923 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.945570 kubelet[2662]: E0317 17:42:22.945463 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.945570 kubelet[2662]: W0317 17:42:22.945471 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.945638 kubelet[2662]: E0317 17:42:22.945578 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.946823 kubelet[2662]: E0317 17:42:22.946411 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.946823 kubelet[2662]: W0317 17:42:22.946438 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.946823 kubelet[2662]: E0317 17:42:22.946458 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.946823 kubelet[2662]: E0317 17:42:22.946650 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.946823 kubelet[2662]: W0317 17:42:22.946658 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.946823 kubelet[2662]: E0317 17:42:22.946666 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.946990 kubelet[2662]: E0317 17:42:22.946885 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.946990 kubelet[2662]: W0317 17:42:22.946893 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.946990 kubelet[2662]: E0317 17:42:22.946901 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.947141 kubelet[2662]: E0317 17:42:22.947115 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.947141 kubelet[2662]: W0317 17:42:22.947127 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.947141 kubelet[2662]: E0317 17:42:22.947136 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.947326 kubelet[2662]: E0317 17:42:22.947283 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.947326 kubelet[2662]: W0317 17:42:22.947290 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.947326 kubelet[2662]: E0317 17:42:22.947298 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.947534 kubelet[2662]: E0317 17:42:22.947494 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.947534 kubelet[2662]: W0317 17:42:22.947502 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.947534 kubelet[2662]: E0317 17:42:22.947512 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.947920 kubelet[2662]: E0317 17:42:22.947896 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.947920 kubelet[2662]: W0317 17:42:22.947912 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.947920 kubelet[2662]: E0317 17:42:22.947921 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.948093 kubelet[2662]: E0317 17:42:22.948072 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.948093 kubelet[2662]: W0317 17:42:22.948084 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.948093 kubelet[2662]: E0317 17:42:22.948091 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.948428 kubelet[2662]: E0317 17:42:22.948236 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.948428 kubelet[2662]: W0317 17:42:22.948249 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.948428 kubelet[2662]: E0317 17:42:22.948256 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.948830 kubelet[2662]: E0317 17:42:22.948478 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.948830 kubelet[2662]: W0317 17:42:22.948486 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.948830 kubelet[2662]: E0317 17:42:22.948494 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.948830 kubelet[2662]: E0317 17:42:22.948653 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.948830 kubelet[2662]: W0317 17:42:22.948660 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.948830 kubelet[2662]: E0317 17:42:22.948667 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.948830 kubelet[2662]: E0317 17:42:22.948820 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.948830 kubelet[2662]: W0317 17:42:22.948828 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.948830 kubelet[2662]: E0317 17:42:22.948836 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.949069 kubelet[2662]: E0317 17:42:22.948980 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.949069 kubelet[2662]: W0317 17:42:22.948988 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.949069 kubelet[2662]: E0317 17:42:22.948995 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.949403 kubelet[2662]: E0317 17:42:22.949137 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.949403 kubelet[2662]: W0317 17:42:22.949151 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.949403 kubelet[2662]: E0317 17:42:22.949158 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.949403 kubelet[2662]: E0317 17:42:22.949317 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.949403 kubelet[2662]: W0317 17:42:22.949324 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.949403 kubelet[2662]: E0317 17:42:22.949331 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.951851 kubelet[2662]: E0317 17:42:22.951826 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.951851 kubelet[2662]: W0317 17:42:22.951844 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.951923 kubelet[2662]: E0317 17:42:22.951855 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.953460 kubelet[2662]: E0317 17:42:22.952020 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.953460 kubelet[2662]: W0317 17:42:22.952036 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.953460 kubelet[2662]: E0317 17:42:22.952044 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.953460 kubelet[2662]: E0317 17:42:22.952192 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.953460 kubelet[2662]: W0317 17:42:22.952199 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.953460 kubelet[2662]: E0317 17:42:22.952207 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.953460 kubelet[2662]: E0317 17:42:22.952377 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.953460 kubelet[2662]: W0317 17:42:22.952384 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.953460 kubelet[2662]: E0317 17:42:22.952401 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.953460 kubelet[2662]: E0317 17:42:22.952553 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.953914 kubelet[2662]: W0317 17:42:22.952560 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.953914 kubelet[2662]: E0317 17:42:22.952568 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.953914 kubelet[2662]: E0317 17:42:22.952719 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.953914 kubelet[2662]: W0317 17:42:22.952727 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.953914 kubelet[2662]: E0317 17:42:22.952735 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.953914 kubelet[2662]: E0317 17:42:22.952880 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.953914 kubelet[2662]: W0317 17:42:22.952888 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.953914 kubelet[2662]: E0317 17:42:22.952895 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.953914 kubelet[2662]: E0317 17:42:22.953047 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.953914 kubelet[2662]: W0317 17:42:22.953054 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.954259 kubelet[2662]: E0317 17:42:22.953061 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.954259 kubelet[2662]: E0317 17:42:22.953229 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.954259 kubelet[2662]: W0317 17:42:22.953236 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.954259 kubelet[2662]: E0317 17:42:22.953244 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.954259 kubelet[2662]: E0317 17:42:22.953496 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.954259 kubelet[2662]: W0317 17:42:22.953503 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.954259 kubelet[2662]: E0317 17:42:22.953668 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.954259 kubelet[2662]: W0317 17:42:22.953674 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.954259 kubelet[2662]: E0317 17:42:22.953835 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.954259 kubelet[2662]: W0317 17:42:22.953842 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.954259 kubelet[2662]: E0317 17:42:22.954045 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.954605 kubelet[2662]: W0317 17:42:22.954052 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.954605 kubelet[2662]: E0317 17:42:22.954061 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.954605 kubelet[2662]: E0317 17:42:22.954249 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.954605 kubelet[2662]: W0317 17:42:22.954256 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.954605 kubelet[2662]: E0317 17:42:22.954264 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.954605 kubelet[2662]: E0317 17:42:22.954280 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.954605 kubelet[2662]: E0317 17:42:22.954468 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.954605 kubelet[2662]: W0317 17:42:22.954475 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.954605 kubelet[2662]: E0317 17:42:22.954483 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.954605 kubelet[2662]: E0317 17:42:22.954547 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.954855 kubelet[2662]: E0317 17:42:22.954573 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.954855 kubelet[2662]: E0317 17:42:22.954811 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.954855 kubelet[2662]: W0317 17:42:22.954818 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.954855 kubelet[2662]: E0317 17:42:22.954826 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:22.962462 kubelet[2662]: E0317 17:42:22.962425 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:22.962462 kubelet[2662]: W0317 17:42:22.962446 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:22.962462 kubelet[2662]: E0317 17:42:22.962466 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.007162 kubelet[2662]: E0317 17:42:23.007122 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:23.007668 containerd[1467]: time="2025-03-17T17:42:23.007623963Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-7b7cf6c67c-kdn9m,Uid:c1834c71-5ae6-4d93-b7af-2c16faf191c5,Namespace:calico-system,Attempt:0,}" Mar 17 17:42:23.036935 containerd[1467]: time="2025-03-17T17:42:23.036839080Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 17:42:23.037104 containerd[1467]: time="2025-03-17T17:42:23.036953125Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 17:42:23.037104 containerd[1467]: time="2025-03-17T17:42:23.036968203Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:42:23.037104 containerd[1467]: time="2025-03-17T17:42:23.037057632Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:42:23.040789 kubelet[2662]: E0317 17:42:23.040751 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.040789 kubelet[2662]: W0317 17:42:23.040775 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.040789 kubelet[2662]: E0317 17:42:23.040806 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.041078 kubelet[2662]: I0317 17:42:23.040837 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/27eb0c6c-40f4-4b19-b513-0003d63df5d0-socket-dir\") pod \"csi-node-driver-j9gdg\" (UID: \"27eb0c6c-40f4-4b19-b513-0003d63df5d0\") " pod="calico-system/csi-node-driver-j9gdg" Mar 17 17:42:23.041078 kubelet[2662]: E0317 17:42:23.041064 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.041078 kubelet[2662]: W0317 17:42:23.041073 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.041154 kubelet[2662]: E0317 17:42:23.041091 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.041154 kubelet[2662]: I0317 17:42:23.041105 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/27eb0c6c-40f4-4b19-b513-0003d63df5d0-registration-dir\") pod \"csi-node-driver-j9gdg\" (UID: \"27eb0c6c-40f4-4b19-b513-0003d63df5d0\") " pod="calico-system/csi-node-driver-j9gdg" Mar 17 17:42:23.041369 kubelet[2662]: E0317 17:42:23.041335 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.041422 kubelet[2662]: W0317 17:42:23.041349 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.041449 kubelet[2662]: E0317 17:42:23.041438 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.041473 kubelet[2662]: I0317 17:42:23.041455 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/27eb0c6c-40f4-4b19-b513-0003d63df5d0-varrun\") pod \"csi-node-driver-j9gdg\" (UID: \"27eb0c6c-40f4-4b19-b513-0003d63df5d0\") " pod="calico-system/csi-node-driver-j9gdg" Mar 17 17:42:23.041759 kubelet[2662]: E0317 17:42:23.041742 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.041759 kubelet[2662]: W0317 17:42:23.041755 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.041819 kubelet[2662]: E0317 17:42:23.041781 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.041819 kubelet[2662]: I0317 17:42:23.041795 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/27eb0c6c-40f4-4b19-b513-0003d63df5d0-kubelet-dir\") pod \"csi-node-driver-j9gdg\" (UID: \"27eb0c6c-40f4-4b19-b513-0003d63df5d0\") " pod="calico-system/csi-node-driver-j9gdg" Mar 17 17:42:23.042052 kubelet[2662]: E0317 17:42:23.042038 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.042052 kubelet[2662]: W0317 17:42:23.042049 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.042106 kubelet[2662]: E0317 17:42:23.042075 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.042131 kubelet[2662]: I0317 17:42:23.042104 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-975n9\" (UniqueName: \"kubernetes.io/projected/27eb0c6c-40f4-4b19-b513-0003d63df5d0-kube-api-access-975n9\") pod \"csi-node-driver-j9gdg\" (UID: \"27eb0c6c-40f4-4b19-b513-0003d63df5d0\") " pod="calico-system/csi-node-driver-j9gdg" Mar 17 17:42:23.042439 kubelet[2662]: E0317 17:42:23.042414 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.042486 kubelet[2662]: W0317 17:42:23.042436 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.042515 kubelet[2662]: E0317 17:42:23.042492 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.042737 kubelet[2662]: E0317 17:42:23.042721 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.042737 kubelet[2662]: W0317 17:42:23.042732 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.042811 kubelet[2662]: E0317 17:42:23.042754 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.042996 kubelet[2662]: E0317 17:42:23.042971 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.042996 kubelet[2662]: W0317 17:42:23.042993 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.043051 kubelet[2662]: E0317 17:42:23.043017 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.043206 kubelet[2662]: E0317 17:42:23.043191 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.043206 kubelet[2662]: W0317 17:42:23.043202 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.043260 kubelet[2662]: E0317 17:42:23.043230 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.043437 kubelet[2662]: E0317 17:42:23.043422 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.043437 kubelet[2662]: W0317 17:42:23.043433 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.043499 kubelet[2662]: E0317 17:42:23.043455 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.043642 kubelet[2662]: E0317 17:42:23.043628 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.043642 kubelet[2662]: W0317 17:42:23.043638 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.043695 kubelet[2662]: E0317 17:42:23.043660 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.043842 kubelet[2662]: E0317 17:42:23.043828 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.043842 kubelet[2662]: W0317 17:42:23.043838 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.043899 kubelet[2662]: E0317 17:42:23.043846 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.044068 kubelet[2662]: E0317 17:42:23.044053 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.044106 kubelet[2662]: W0317 17:42:23.044073 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.044106 kubelet[2662]: E0317 17:42:23.044083 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.044285 kubelet[2662]: E0317 17:42:23.044271 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.044285 kubelet[2662]: W0317 17:42:23.044280 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.044436 kubelet[2662]: E0317 17:42:23.044288 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.044592 kubelet[2662]: E0317 17:42:23.044576 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.044592 kubelet[2662]: W0317 17:42:23.044586 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.044644 kubelet[2662]: E0317 17:42:23.044595 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.046625 kubelet[2662]: E0317 17:42:23.046600 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:23.048020 containerd[1467]: time="2025-03-17T17:42:23.047955732Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-fbl5x,Uid:77e9b5a5-a1f1-40db-82d5-f5ed3dcf3fa2,Namespace:calico-system,Attempt:0,}" Mar 17 17:42:23.063557 systemd[1]: Started cri-containerd-bb99fb828c93ee7bc6a9461658e09177fb1be27bd9bd10af7677933352452186.scope - libcontainer container bb99fb828c93ee7bc6a9461658e09177fb1be27bd9bd10af7677933352452186. Mar 17 17:42:23.086589 containerd[1467]: time="2025-03-17T17:42:23.086335589Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 17:42:23.086589 containerd[1467]: time="2025-03-17T17:42:23.086417794Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 17:42:23.086589 containerd[1467]: time="2025-03-17T17:42:23.086431259Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:42:23.087398 containerd[1467]: time="2025-03-17T17:42:23.086575993Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:42:23.107547 systemd[1]: Started cri-containerd-0fa693deaf0f5a67ec99e9c752c255e61ecc989170bfbc0ea77ac17bb35fd6b3.scope - libcontainer container 0fa693deaf0f5a67ec99e9c752c255e61ecc989170bfbc0ea77ac17bb35fd6b3. Mar 17 17:42:23.108489 containerd[1467]: time="2025-03-17T17:42:23.108454448Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-7b7cf6c67c-kdn9m,Uid:c1834c71-5ae6-4d93-b7af-2c16faf191c5,Namespace:calico-system,Attempt:0,} returns sandbox id \"bb99fb828c93ee7bc6a9461658e09177fb1be27bd9bd10af7677933352452186\"" Mar 17 17:42:23.111479 kubelet[2662]: E0317 17:42:23.111447 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:23.117256 containerd[1467]: time="2025-03-17T17:42:23.117184628Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.29.2\"" Mar 17 17:42:23.134753 containerd[1467]: time="2025-03-17T17:42:23.134706164Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-fbl5x,Uid:77e9b5a5-a1f1-40db-82d5-f5ed3dcf3fa2,Namespace:calico-system,Attempt:0,} returns sandbox id \"0fa693deaf0f5a67ec99e9c752c255e61ecc989170bfbc0ea77ac17bb35fd6b3\"" Mar 17 17:42:23.135312 kubelet[2662]: E0317 17:42:23.135289 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:23.143215 kubelet[2662]: E0317 17:42:23.143163 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.143215 kubelet[2662]: W0317 17:42:23.143183 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.143215 kubelet[2662]: E0317 17:42:23.143204 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.143515 kubelet[2662]: E0317 17:42:23.143507 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.143515 kubelet[2662]: W0317 17:42:23.143516 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.143654 kubelet[2662]: E0317 17:42:23.143531 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.143840 kubelet[2662]: E0317 17:42:23.143823 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.143840 kubelet[2662]: W0317 17:42:23.143836 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.143936 kubelet[2662]: E0317 17:42:23.143849 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.144133 kubelet[2662]: E0317 17:42:23.144119 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.144133 kubelet[2662]: W0317 17:42:23.144143 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.144133 kubelet[2662]: E0317 17:42:23.144155 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.144423 kubelet[2662]: E0317 17:42:23.144387 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.144423 kubelet[2662]: W0317 17:42:23.144417 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.144503 kubelet[2662]: E0317 17:42:23.144429 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.144774 kubelet[2662]: E0317 17:42:23.144752 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.144833 kubelet[2662]: W0317 17:42:23.144773 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.144833 kubelet[2662]: E0317 17:42:23.144803 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.145049 kubelet[2662]: E0317 17:42:23.145034 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.145049 kubelet[2662]: W0317 17:42:23.145046 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.145152 kubelet[2662]: E0317 17:42:23.145076 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.145315 kubelet[2662]: E0317 17:42:23.145266 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.145315 kubelet[2662]: W0317 17:42:23.145281 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.145315 kubelet[2662]: E0317 17:42:23.145308 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.145559 kubelet[2662]: E0317 17:42:23.145524 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.145559 kubelet[2662]: W0317 17:42:23.145537 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.145609 kubelet[2662]: E0317 17:42:23.145565 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.145761 kubelet[2662]: E0317 17:42:23.145747 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.145761 kubelet[2662]: W0317 17:42:23.145759 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.145817 kubelet[2662]: E0317 17:42:23.145786 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.145973 kubelet[2662]: E0317 17:42:23.145958 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.145973 kubelet[2662]: W0317 17:42:23.145970 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.146025 kubelet[2662]: E0317 17:42:23.145995 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.146188 kubelet[2662]: E0317 17:42:23.146172 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.146188 kubelet[2662]: W0317 17:42:23.146185 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.146245 kubelet[2662]: E0317 17:42:23.146200 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.146524 kubelet[2662]: E0317 17:42:23.146501 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.146524 kubelet[2662]: W0317 17:42:23.146514 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.146583 kubelet[2662]: E0317 17:42:23.146529 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.146726 kubelet[2662]: E0317 17:42:23.146711 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.146726 kubelet[2662]: W0317 17:42:23.146721 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.146782 kubelet[2662]: E0317 17:42:23.146769 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.146927 kubelet[2662]: E0317 17:42:23.146913 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.146927 kubelet[2662]: W0317 17:42:23.146923 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.147002 kubelet[2662]: E0317 17:42:23.146977 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.147143 kubelet[2662]: E0317 17:42:23.147130 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.147143 kubelet[2662]: W0317 17:42:23.147140 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.147293 kubelet[2662]: E0317 17:42:23.147178 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.147349 kubelet[2662]: E0317 17:42:23.147335 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.147349 kubelet[2662]: W0317 17:42:23.147345 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.147446 kubelet[2662]: E0317 17:42:23.147424 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.147618 kubelet[2662]: E0317 17:42:23.147604 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.147618 kubelet[2662]: W0317 17:42:23.147615 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.147670 kubelet[2662]: E0317 17:42:23.147631 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.147861 kubelet[2662]: E0317 17:42:23.147848 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.147861 kubelet[2662]: W0317 17:42:23.147857 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.147911 kubelet[2662]: E0317 17:42:23.147869 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.148097 kubelet[2662]: E0317 17:42:23.148082 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.148097 kubelet[2662]: W0317 17:42:23.148095 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.148147 kubelet[2662]: E0317 17:42:23.148110 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.148382 kubelet[2662]: E0317 17:42:23.148366 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.148382 kubelet[2662]: W0317 17:42:23.148380 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.148449 kubelet[2662]: E0317 17:42:23.148406 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.148623 kubelet[2662]: E0317 17:42:23.148609 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.148623 kubelet[2662]: W0317 17:42:23.148619 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.148676 kubelet[2662]: E0317 17:42:23.148632 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.148905 kubelet[2662]: E0317 17:42:23.148870 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.148905 kubelet[2662]: W0317 17:42:23.148892 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.149007 kubelet[2662]: E0317 17:42:23.148930 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.149277 kubelet[2662]: E0317 17:42:23.149257 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.149277 kubelet[2662]: W0317 17:42:23.149271 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.149417 kubelet[2662]: E0317 17:42:23.149281 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.158672 kubelet[2662]: E0317 17:42:23.158623 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.158672 kubelet[2662]: W0317 17:42:23.158653 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.158672 kubelet[2662]: E0317 17:42:23.158666 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:23.162979 kubelet[2662]: E0317 17:42:23.162957 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:23.162979 kubelet[2662]: W0317 17:42:23.162976 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:23.163077 kubelet[2662]: E0317 17:42:23.162995 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:24.825304 kubelet[2662]: E0317 17:42:24.825229 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-j9gdg" podUID="27eb0c6c-40f4-4b19-b513-0003d63df5d0" Mar 17 17:42:25.943336 kubelet[2662]: E0317 17:42:25.943224 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-j9gdg" podUID="27eb0c6c-40f4-4b19-b513-0003d63df5d0" Mar 17 17:42:25.945161 containerd[1467]: time="2025-03-17T17:42:25.944629243Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.29.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:42:25.947932 containerd[1467]: time="2025-03-17T17:42:25.946716148Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.29.2: active requests=0, bytes read=30414075" Mar 17 17:42:25.948073 containerd[1467]: time="2025-03-17T17:42:25.948034042Z" level=info msg="ImageCreate event name:\"sha256:1d6f9d005866d74e6f0a8b0b8b743d0eaf4efcb7c7032fd2215da9c6ca131cb5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:42:25.951079 containerd[1467]: time="2025-03-17T17:42:25.951050399Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:9839fd34b4c1bad50beed72aec59c64893487a46eea57dc2d7d66c3041d7bcce\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:42:25.951649 containerd[1467]: time="2025-03-17T17:42:25.951626825Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.29.2\" with image id \"sha256:1d6f9d005866d74e6f0a8b0b8b743d0eaf4efcb7c7032fd2215da9c6ca131cb5\", repo tag \"ghcr.io/flatcar/calico/typha:v3.29.2\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:9839fd34b4c1bad50beed72aec59c64893487a46eea57dc2d7d66c3041d7bcce\", size \"31907171\" in 2.834376934s" Mar 17 17:42:25.951702 containerd[1467]: time="2025-03-17T17:42:25.951650820Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.29.2\" returns image reference \"sha256:1d6f9d005866d74e6f0a8b0b8b743d0eaf4efcb7c7032fd2215da9c6ca131cb5\"" Mar 17 17:42:25.952421 containerd[1467]: time="2025-03-17T17:42:25.952396226Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.2\"" Mar 17 17:42:25.960191 containerd[1467]: time="2025-03-17T17:42:25.960111232Z" level=info msg="CreateContainer within sandbox \"bb99fb828c93ee7bc6a9461658e09177fb1be27bd9bd10af7677933352452186\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Mar 17 17:42:25.975825 containerd[1467]: time="2025-03-17T17:42:25.975775844Z" level=info msg="CreateContainer within sandbox \"bb99fb828c93ee7bc6a9461658e09177fb1be27bd9bd10af7677933352452186\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"2aba5f44f580e414cd39320fa4ac98d995827edd374dbd2ff70d00f6046ff1d4\"" Mar 17 17:42:25.976247 containerd[1467]: time="2025-03-17T17:42:25.976225301Z" level=info msg="StartContainer for \"2aba5f44f580e414cd39320fa4ac98d995827edd374dbd2ff70d00f6046ff1d4\"" Mar 17 17:42:26.005535 systemd[1]: Started cri-containerd-2aba5f44f580e414cd39320fa4ac98d995827edd374dbd2ff70d00f6046ff1d4.scope - libcontainer container 2aba5f44f580e414cd39320fa4ac98d995827edd374dbd2ff70d00f6046ff1d4. Mar 17 17:42:26.046718 containerd[1467]: time="2025-03-17T17:42:26.046679373Z" level=info msg="StartContainer for \"2aba5f44f580e414cd39320fa4ac98d995827edd374dbd2ff70d00f6046ff1d4\" returns successfully" Mar 17 17:42:26.950046 kubelet[2662]: E0317 17:42:26.950018 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:26.960614 kubelet[2662]: I0317 17:42:26.960566 2662 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-7b7cf6c67c-kdn9m" podStartSLOduration=2.125077398 podStartE2EDuration="4.960551662s" podCreationTimestamp="2025-03-17 17:42:22 +0000 UTC" firstStartedPulling="2025-03-17 17:42:23.11681473 +0000 UTC m=+21.376781456" lastFinishedPulling="2025-03-17 17:42:25.952288994 +0000 UTC m=+24.212255720" observedRunningTime="2025-03-17 17:42:26.960041902 +0000 UTC m=+25.220008628" watchObservedRunningTime="2025-03-17 17:42:26.960551662 +0000 UTC m=+25.220518388" Mar 17 17:42:27.046584 kubelet[2662]: E0317 17:42:27.046550 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:27.046584 kubelet[2662]: W0317 17:42:27.046574 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:27.046584 kubelet[2662]: E0317 17:42:27.046592 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:27.046816 kubelet[2662]: E0317 17:42:27.046801 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:27.046816 kubelet[2662]: W0317 17:42:27.046812 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:27.046874 kubelet[2662]: E0317 17:42:27.046820 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:27.047035 kubelet[2662]: E0317 17:42:27.047017 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:27.047035 kubelet[2662]: W0317 17:42:27.047027 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:27.047035 kubelet[2662]: E0317 17:42:27.047034 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:27.047277 kubelet[2662]: E0317 17:42:27.047258 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:27.047277 kubelet[2662]: W0317 17:42:27.047269 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:27.047277 kubelet[2662]: E0317 17:42:27.047276 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:27.047549 kubelet[2662]: E0317 17:42:27.047532 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:27.047549 kubelet[2662]: W0317 17:42:27.047542 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:27.047549 kubelet[2662]: E0317 17:42:27.047550 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:27.047756 kubelet[2662]: E0317 17:42:27.047739 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:27.047756 kubelet[2662]: W0317 17:42:27.047749 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:27.047816 kubelet[2662]: E0317 17:42:27.047757 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:27.047957 kubelet[2662]: E0317 17:42:27.047940 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:27.047957 kubelet[2662]: W0317 17:42:27.047950 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:27.048005 kubelet[2662]: E0317 17:42:27.047958 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:27.048155 kubelet[2662]: E0317 17:42:27.048137 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:27.048155 kubelet[2662]: W0317 17:42:27.048148 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:27.048220 kubelet[2662]: E0317 17:42:27.048156 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:27.048407 kubelet[2662]: E0317 17:42:27.048389 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:27.048407 kubelet[2662]: W0317 17:42:27.048401 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:27.048459 kubelet[2662]: E0317 17:42:27.048409 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:27.048602 kubelet[2662]: E0317 17:42:27.048585 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:27.048602 kubelet[2662]: W0317 17:42:27.048595 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:27.048602 kubelet[2662]: E0317 17:42:27.048602 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:27.048789 kubelet[2662]: E0317 17:42:27.048772 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:27.048789 kubelet[2662]: W0317 17:42:27.048782 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:27.048789 kubelet[2662]: E0317 17:42:27.048789 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:27.048973 kubelet[2662]: E0317 17:42:27.048956 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:27.048973 kubelet[2662]: W0317 17:42:27.048966 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:27.048973 kubelet[2662]: E0317 17:42:27.048974 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:27.049203 kubelet[2662]: E0317 17:42:27.049185 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:27.049203 kubelet[2662]: W0317 17:42:27.049195 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:27.049254 kubelet[2662]: E0317 17:42:27.049203 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:27.049422 kubelet[2662]: E0317 17:42:27.049404 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:27.049422 kubelet[2662]: W0317 17:42:27.049415 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:27.049482 kubelet[2662]: E0317 17:42:27.049423 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:27.049610 kubelet[2662]: E0317 17:42:27.049593 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:27.049610 kubelet[2662]: W0317 17:42:27.049603 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:27.049610 kubelet[2662]: E0317 17:42:27.049610 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:27.049879 kubelet[2662]: E0317 17:42:27.049860 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:27.049879 kubelet[2662]: W0317 17:42:27.049872 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:27.049879 kubelet[2662]: E0317 17:42:27.049879 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:27.050116 kubelet[2662]: E0317 17:42:27.050097 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:27.050116 kubelet[2662]: W0317 17:42:27.050108 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:27.050203 kubelet[2662]: E0317 17:42:27.050120 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:27.050432 kubelet[2662]: E0317 17:42:27.050408 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:27.050529 kubelet[2662]: W0317 17:42:27.050432 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:27.050529 kubelet[2662]: E0317 17:42:27.050461 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:27.050755 kubelet[2662]: E0317 17:42:27.050742 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:27.050780 kubelet[2662]: W0317 17:42:27.050754 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:27.050780 kubelet[2662]: E0317 17:42:27.050773 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:27.050987 kubelet[2662]: E0317 17:42:27.050970 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:27.050987 kubelet[2662]: W0317 17:42:27.050983 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:27.051048 kubelet[2662]: E0317 17:42:27.051000 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:27.051242 kubelet[2662]: E0317 17:42:27.051228 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:27.051242 kubelet[2662]: W0317 17:42:27.051238 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:27.051286 kubelet[2662]: E0317 17:42:27.051250 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:27.051509 kubelet[2662]: E0317 17:42:27.051494 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:27.051509 kubelet[2662]: W0317 17:42:27.051504 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:27.051574 kubelet[2662]: E0317 17:42:27.051534 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:27.051718 kubelet[2662]: E0317 17:42:27.051704 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:27.051747 kubelet[2662]: W0317 17:42:27.051717 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:27.051773 kubelet[2662]: E0317 17:42:27.051748 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:27.051928 kubelet[2662]: E0317 17:42:27.051915 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:27.051928 kubelet[2662]: W0317 17:42:27.051925 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:27.051984 kubelet[2662]: E0317 17:42:27.051941 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:27.052157 kubelet[2662]: E0317 17:42:27.052146 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:27.052181 kubelet[2662]: W0317 17:42:27.052156 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:27.052181 kubelet[2662]: E0317 17:42:27.052169 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:27.052439 kubelet[2662]: E0317 17:42:27.052425 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:27.052477 kubelet[2662]: W0317 17:42:27.052438 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:27.052477 kubelet[2662]: E0317 17:42:27.052452 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:27.052733 kubelet[2662]: E0317 17:42:27.052721 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:27.052759 kubelet[2662]: W0317 17:42:27.052733 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:27.052759 kubelet[2662]: E0317 17:42:27.052747 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:27.052945 kubelet[2662]: E0317 17:42:27.052933 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:27.052977 kubelet[2662]: W0317 17:42:27.052944 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:27.052977 kubelet[2662]: E0317 17:42:27.052958 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:27.053192 kubelet[2662]: E0317 17:42:27.053174 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:27.053192 kubelet[2662]: W0317 17:42:27.053187 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:27.053245 kubelet[2662]: E0317 17:42:27.053203 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:27.053553 kubelet[2662]: E0317 17:42:27.053539 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:27.053581 kubelet[2662]: W0317 17:42:27.053553 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:27.053581 kubelet[2662]: E0317 17:42:27.053569 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:27.053770 kubelet[2662]: E0317 17:42:27.053754 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:27.053770 kubelet[2662]: W0317 17:42:27.053765 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:27.053818 kubelet[2662]: E0317 17:42:27.053779 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:27.054082 kubelet[2662]: E0317 17:42:27.054062 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:27.054082 kubelet[2662]: W0317 17:42:27.054075 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:27.054126 kubelet[2662]: E0317 17:42:27.054090 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:27.054322 kubelet[2662]: E0317 17:42:27.054307 2662 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 17 17:42:27.054322 kubelet[2662]: W0317 17:42:27.054317 2662 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 17 17:42:27.054413 kubelet[2662]: E0317 17:42:27.054327 2662 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 17 17:42:27.433523 containerd[1467]: time="2025-03-17T17:42:27.433464254Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:42:27.434159 containerd[1467]: time="2025-03-17T17:42:27.434111343Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.2: active requests=0, bytes read=5364011" Mar 17 17:42:27.435130 containerd[1467]: time="2025-03-17T17:42:27.435106308Z" level=info msg="ImageCreate event name:\"sha256:441bf8ace5b7fa3742b7fafaf6cd60fea340dd307169a18c75a1d78cba3a8365\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:42:27.437078 containerd[1467]: time="2025-03-17T17:42:27.437051413Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:51d9341a4a37e278a906f40ecc73f5076e768612c21621f1b1d4f2b2f0735a1d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:42:27.437673 containerd[1467]: time="2025-03-17T17:42:27.437648849Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.2\" with image id \"sha256:441bf8ace5b7fa3742b7fafaf6cd60fea340dd307169a18c75a1d78cba3a8365\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.2\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:51d9341a4a37e278a906f40ecc73f5076e768612c21621f1b1d4f2b2f0735a1d\", size \"6857075\" in 1.485228436s" Mar 17 17:42:27.437710 containerd[1467]: time="2025-03-17T17:42:27.437673014Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.2\" returns image reference \"sha256:441bf8ace5b7fa3742b7fafaf6cd60fea340dd307169a18c75a1d78cba3a8365\"" Mar 17 17:42:27.439286 containerd[1467]: time="2025-03-17T17:42:27.439264904Z" level=info msg="CreateContainer within sandbox \"0fa693deaf0f5a67ec99e9c752c255e61ecc989170bfbc0ea77ac17bb35fd6b3\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Mar 17 17:42:27.454892 containerd[1467]: time="2025-03-17T17:42:27.454851662Z" level=info msg="CreateContainer within sandbox \"0fa693deaf0f5a67ec99e9c752c255e61ecc989170bfbc0ea77ac17bb35fd6b3\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"a134c4a06e8d686b78017b505f74e4580b26a2578a856090c3bbe8fc66e3c97c\"" Mar 17 17:42:27.455512 containerd[1467]: time="2025-03-17T17:42:27.455293405Z" level=info msg="StartContainer for \"a134c4a06e8d686b78017b505f74e4580b26a2578a856090c3bbe8fc66e3c97c\"" Mar 17 17:42:27.485518 systemd[1]: Started cri-containerd-a134c4a06e8d686b78017b505f74e4580b26a2578a856090c3bbe8fc66e3c97c.scope - libcontainer container a134c4a06e8d686b78017b505f74e4580b26a2578a856090c3bbe8fc66e3c97c. Mar 17 17:42:27.514066 containerd[1467]: time="2025-03-17T17:42:27.514023564Z" level=info msg="StartContainer for \"a134c4a06e8d686b78017b505f74e4580b26a2578a856090c3bbe8fc66e3c97c\" returns successfully" Mar 17 17:42:27.529306 systemd[1]: cri-containerd-a134c4a06e8d686b78017b505f74e4580b26a2578a856090c3bbe8fc66e3c97c.scope: Deactivated successfully. Mar 17 17:42:27.680603 containerd[1467]: time="2025-03-17T17:42:27.680528983Z" level=info msg="shim disconnected" id=a134c4a06e8d686b78017b505f74e4580b26a2578a856090c3bbe8fc66e3c97c namespace=k8s.io Mar 17 17:42:27.680603 containerd[1467]: time="2025-03-17T17:42:27.680595407Z" level=warning msg="cleaning up after shim disconnected" id=a134c4a06e8d686b78017b505f74e4580b26a2578a856090c3bbe8fc66e3c97c namespace=k8s.io Mar 17 17:42:27.680603 containerd[1467]: time="2025-03-17T17:42:27.680607049Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 17 17:42:27.821595 kubelet[2662]: E0317 17:42:27.821397 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-j9gdg" podUID="27eb0c6c-40f4-4b19-b513-0003d63df5d0" Mar 17 17:42:27.952709 kubelet[2662]: I0317 17:42:27.952676 2662 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 17 17:42:27.953106 kubelet[2662]: E0317 17:42:27.952931 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:27.953488 kubelet[2662]: E0317 17:42:27.953454 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:27.953963 containerd[1467]: time="2025-03-17T17:42:27.953732532Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.2\"" Mar 17 17:42:27.957330 systemd[1]: run-containerd-runc-k8s.io-a134c4a06e8d686b78017b505f74e4580b26a2578a856090c3bbe8fc66e3c97c-runc.zW8vPj.mount: Deactivated successfully. Mar 17 17:42:27.957518 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a134c4a06e8d686b78017b505f74e4580b26a2578a856090c3bbe8fc66e3c97c-rootfs.mount: Deactivated successfully. Mar 17 17:42:29.820784 kubelet[2662]: E0317 17:42:29.820721 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-j9gdg" podUID="27eb0c6c-40f4-4b19-b513-0003d63df5d0" Mar 17 17:42:31.715175 systemd[1]: Started sshd@9-10.0.0.61:22-10.0.0.1:51294.service - OpenSSH per-connection server daemon (10.0.0.1:51294). Mar 17 17:42:31.773508 sshd[3392]: Accepted publickey for core from 10.0.0.1 port 51294 ssh2: RSA SHA256:j201F9FRK1q3ChnxQf0adNdYppDp+g37vmaXPvsVhek Mar 17 17:42:31.775651 sshd-session[3392]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 17 17:42:31.782792 systemd-logind[1453]: New session 10 of user core. Mar 17 17:42:31.793622 systemd[1]: Started session-10.scope - Session 10 of User core. Mar 17 17:42:31.822659 kubelet[2662]: E0317 17:42:31.821500 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-j9gdg" podUID="27eb0c6c-40f4-4b19-b513-0003d63df5d0" Mar 17 17:42:31.936211 sshd[3394]: Connection closed by 10.0.0.1 port 51294 Mar 17 17:42:31.936557 sshd-session[3392]: pam_unix(sshd:session): session closed for user core Mar 17 17:42:31.940812 systemd[1]: sshd@9-10.0.0.61:22-10.0.0.1:51294.service: Deactivated successfully. Mar 17 17:42:31.943136 systemd[1]: session-10.scope: Deactivated successfully. Mar 17 17:42:31.944402 systemd-logind[1453]: Session 10 logged out. Waiting for processes to exit. Mar 17 17:42:31.945564 systemd-logind[1453]: Removed session 10. Mar 17 17:42:32.413856 containerd[1467]: time="2025-03-17T17:42:32.413795650Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.29.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:42:32.414515 containerd[1467]: time="2025-03-17T17:42:32.414438059Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.29.2: active requests=0, bytes read=97781477" Mar 17 17:42:32.415584 containerd[1467]: time="2025-03-17T17:42:32.415544271Z" level=info msg="ImageCreate event name:\"sha256:cda13293c895a8a3b06c1e190b70fb6fe61036db2e59764036fc6e65ec374693\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:42:32.417458 containerd[1467]: time="2025-03-17T17:42:32.417426623Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:890e1db6ae363695cfc23ffae4d612cc85cdd99d759bd539af6683969d0c3c25\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:42:32.418077 containerd[1467]: time="2025-03-17T17:42:32.418037994Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.29.2\" with image id \"sha256:cda13293c895a8a3b06c1e190b70fb6fe61036db2e59764036fc6e65ec374693\", repo tag \"ghcr.io/flatcar/calico/cni:v3.29.2\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:890e1db6ae363695cfc23ffae4d612cc85cdd99d759bd539af6683969d0c3c25\", size \"99274581\" in 4.46427216s" Mar 17 17:42:32.418077 containerd[1467]: time="2025-03-17T17:42:32.418071697Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.2\" returns image reference \"sha256:cda13293c895a8a3b06c1e190b70fb6fe61036db2e59764036fc6e65ec374693\"" Mar 17 17:42:32.420498 containerd[1467]: time="2025-03-17T17:42:32.420466284Z" level=info msg="CreateContainer within sandbox \"0fa693deaf0f5a67ec99e9c752c255e61ecc989170bfbc0ea77ac17bb35fd6b3\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Mar 17 17:42:32.436008 containerd[1467]: time="2025-03-17T17:42:32.435970261Z" level=info msg="CreateContainer within sandbox \"0fa693deaf0f5a67ec99e9c752c255e61ecc989170bfbc0ea77ac17bb35fd6b3\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"810000a42d68cd14f876016389b6765245d3cfcba1d84b8b615619ff910b6311\"" Mar 17 17:42:32.436406 containerd[1467]: time="2025-03-17T17:42:32.436309049Z" level=info msg="StartContainer for \"810000a42d68cd14f876016389b6765245d3cfcba1d84b8b615619ff910b6311\"" Mar 17 17:42:32.471657 systemd[1]: Started cri-containerd-810000a42d68cd14f876016389b6765245d3cfcba1d84b8b615619ff910b6311.scope - libcontainer container 810000a42d68cd14f876016389b6765245d3cfcba1d84b8b615619ff910b6311. Mar 17 17:42:32.507923 containerd[1467]: time="2025-03-17T17:42:32.507866968Z" level=info msg="StartContainer for \"810000a42d68cd14f876016389b6765245d3cfcba1d84b8b615619ff910b6311\" returns successfully" Mar 17 17:42:32.969972 kubelet[2662]: E0317 17:42:32.969706 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:33.347962 containerd[1467]: time="2025-03-17T17:42:33.347690090Z" level=error msg="failed to reload cni configuration after receiving fs change event(WRITE \"/etc/cni/net.d/calico-kubeconfig\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Mar 17 17:42:33.350588 systemd[1]: cri-containerd-810000a42d68cd14f876016389b6765245d3cfcba1d84b8b615619ff910b6311.scope: Deactivated successfully. Mar 17 17:42:33.359311 kubelet[2662]: I0317 17:42:33.359232 2662 kubelet_node_status.go:497] "Fast updating node status as it just became ready" Mar 17 17:42:33.373779 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-810000a42d68cd14f876016389b6765245d3cfcba1d84b8b615619ff910b6311-rootfs.mount: Deactivated successfully. Mar 17 17:42:33.382722 kubelet[2662]: I0317 17:42:33.382617 2662 topology_manager.go:215] "Topology Admit Handler" podUID="ae45280b-f87f-4410-991d-e0beb4bdc7c6" podNamespace="kube-system" podName="coredns-7db6d8ff4d-fw5hr" Mar 17 17:42:33.388577 systemd[1]: Created slice kubepods-burstable-podae45280b_f87f_4410_991d_e0beb4bdc7c6.slice - libcontainer container kubepods-burstable-podae45280b_f87f_4410_991d_e0beb4bdc7c6.slice. Mar 17 17:42:33.391691 kubelet[2662]: I0317 17:42:33.390797 2662 topology_manager.go:215] "Topology Admit Handler" podUID="c9eab445-c078-490b-861c-b96d204afef3" podNamespace="calico-apiserver" podName="calico-apiserver-86c8b6f988-ppvkt" Mar 17 17:42:33.391691 kubelet[2662]: I0317 17:42:33.390982 2662 topology_manager.go:215] "Topology Admit Handler" podUID="228c6093-6dc2-4faf-8ec8-87849ef402ab" podNamespace="kube-system" podName="coredns-7db6d8ff4d-sjrwq" Mar 17 17:42:33.393827 kubelet[2662]: I0317 17:42:33.393783 2662 topology_manager.go:215] "Topology Admit Handler" podUID="3265838f-27ca-489d-8c65-7cadce5cca40" podNamespace="calico-apiserver" podName="calico-apiserver-86c8b6f988-79bzl" Mar 17 17:42:33.394301 containerd[1467]: time="2025-03-17T17:42:33.394244165Z" level=info msg="shim disconnected" id=810000a42d68cd14f876016389b6765245d3cfcba1d84b8b615619ff910b6311 namespace=k8s.io Mar 17 17:42:33.394435 containerd[1467]: time="2025-03-17T17:42:33.394303898Z" level=warning msg="cleaning up after shim disconnected" id=810000a42d68cd14f876016389b6765245d3cfcba1d84b8b615619ff910b6311 namespace=k8s.io Mar 17 17:42:33.394435 containerd[1467]: time="2025-03-17T17:42:33.394314779Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 17 17:42:33.394822 kubelet[2662]: I0317 17:42:33.394783 2662 topology_manager.go:215] "Topology Admit Handler" podUID="e531d9e1-a7bb-4958-adfd-8db7b3f84f95" podNamespace="calico-system" podName="calico-kube-controllers-7fddcd7f99-92q58" Mar 17 17:42:33.400954 systemd[1]: Created slice kubepods-burstable-pod228c6093_6dc2_4faf_8ec8_87849ef402ab.slice - libcontainer container kubepods-burstable-pod228c6093_6dc2_4faf_8ec8_87849ef402ab.slice. Mar 17 17:42:33.406681 systemd[1]: Created slice kubepods-besteffort-podc9eab445_c078_490b_861c_b96d204afef3.slice - libcontainer container kubepods-besteffort-podc9eab445_c078_490b_861c_b96d204afef3.slice. Mar 17 17:42:33.411811 containerd[1467]: time="2025-03-17T17:42:33.411764231Z" level=warning msg="cleanup warnings time=\"2025-03-17T17:42:33Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Mar 17 17:42:33.414345 systemd[1]: Created slice kubepods-besteffort-pod3265838f_27ca_489d_8c65_7cadce5cca40.slice - libcontainer container kubepods-besteffort-pod3265838f_27ca_489d_8c65_7cadce5cca40.slice. Mar 17 17:42:33.420112 systemd[1]: Created slice kubepods-besteffort-pode531d9e1_a7bb_4958_adfd_8db7b3f84f95.slice - libcontainer container kubepods-besteffort-pode531d9e1_a7bb_4958_adfd_8db7b3f84f95.slice. Mar 17 17:42:33.497878 kubelet[2662]: I0317 17:42:33.497829 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/c9eab445-c078-490b-861c-b96d204afef3-calico-apiserver-certs\") pod \"calico-apiserver-86c8b6f988-ppvkt\" (UID: \"c9eab445-c078-490b-861c-b96d204afef3\") " pod="calico-apiserver/calico-apiserver-86c8b6f988-ppvkt" Mar 17 17:42:33.498023 kubelet[2662]: I0317 17:42:33.497894 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2626c\" (UniqueName: \"kubernetes.io/projected/e531d9e1-a7bb-4958-adfd-8db7b3f84f95-kube-api-access-2626c\") pod \"calico-kube-controllers-7fddcd7f99-92q58\" (UID: \"e531d9e1-a7bb-4958-adfd-8db7b3f84f95\") " pod="calico-system/calico-kube-controllers-7fddcd7f99-92q58" Mar 17 17:42:33.498023 kubelet[2662]: I0317 17:42:33.497925 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkjs9\" (UniqueName: \"kubernetes.io/projected/3265838f-27ca-489d-8c65-7cadce5cca40-kube-api-access-gkjs9\") pod \"calico-apiserver-86c8b6f988-79bzl\" (UID: \"3265838f-27ca-489d-8c65-7cadce5cca40\") " pod="calico-apiserver/calico-apiserver-86c8b6f988-79bzl" Mar 17 17:42:33.498023 kubelet[2662]: I0317 17:42:33.497947 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xs57m\" (UniqueName: \"kubernetes.io/projected/ae45280b-f87f-4410-991d-e0beb4bdc7c6-kube-api-access-xs57m\") pod \"coredns-7db6d8ff4d-fw5hr\" (UID: \"ae45280b-f87f-4410-991d-e0beb4bdc7c6\") " pod="kube-system/coredns-7db6d8ff4d-fw5hr" Mar 17 17:42:33.498023 kubelet[2662]: I0317 17:42:33.497962 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ccrh\" (UniqueName: \"kubernetes.io/projected/c9eab445-c078-490b-861c-b96d204afef3-kube-api-access-6ccrh\") pod \"calico-apiserver-86c8b6f988-ppvkt\" (UID: \"c9eab445-c078-490b-861c-b96d204afef3\") " pod="calico-apiserver/calico-apiserver-86c8b6f988-ppvkt" Mar 17 17:42:33.498023 kubelet[2662]: I0317 17:42:33.497992 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7jws\" (UniqueName: \"kubernetes.io/projected/228c6093-6dc2-4faf-8ec8-87849ef402ab-kube-api-access-r7jws\") pod \"coredns-7db6d8ff4d-sjrwq\" (UID: \"228c6093-6dc2-4faf-8ec8-87849ef402ab\") " pod="kube-system/coredns-7db6d8ff4d-sjrwq" Mar 17 17:42:33.498158 kubelet[2662]: I0317 17:42:33.498008 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/228c6093-6dc2-4faf-8ec8-87849ef402ab-config-volume\") pod \"coredns-7db6d8ff4d-sjrwq\" (UID: \"228c6093-6dc2-4faf-8ec8-87849ef402ab\") " pod="kube-system/coredns-7db6d8ff4d-sjrwq" Mar 17 17:42:33.498158 kubelet[2662]: I0317 17:42:33.498025 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/3265838f-27ca-489d-8c65-7cadce5cca40-calico-apiserver-certs\") pod \"calico-apiserver-86c8b6f988-79bzl\" (UID: \"3265838f-27ca-489d-8c65-7cadce5cca40\") " pod="calico-apiserver/calico-apiserver-86c8b6f988-79bzl" Mar 17 17:42:33.498158 kubelet[2662]: I0317 17:42:33.498041 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ae45280b-f87f-4410-991d-e0beb4bdc7c6-config-volume\") pod \"coredns-7db6d8ff4d-fw5hr\" (UID: \"ae45280b-f87f-4410-991d-e0beb4bdc7c6\") " pod="kube-system/coredns-7db6d8ff4d-fw5hr" Mar 17 17:42:33.498158 kubelet[2662]: I0317 17:42:33.498093 2662 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e531d9e1-a7bb-4958-adfd-8db7b3f84f95-tigera-ca-bundle\") pod \"calico-kube-controllers-7fddcd7f99-92q58\" (UID: \"e531d9e1-a7bb-4958-adfd-8db7b3f84f95\") " pod="calico-system/calico-kube-controllers-7fddcd7f99-92q58" Mar 17 17:42:33.693581 kubelet[2662]: E0317 17:42:33.693545 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:33.694165 containerd[1467]: time="2025-03-17T17:42:33.694122834Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-fw5hr,Uid:ae45280b-f87f-4410-991d-e0beb4bdc7c6,Namespace:kube-system,Attempt:0,}" Mar 17 17:42:33.704615 kubelet[2662]: E0317 17:42:33.704562 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:33.705106 containerd[1467]: time="2025-03-17T17:42:33.705025698Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-sjrwq,Uid:228c6093-6dc2-4faf-8ec8-87849ef402ab,Namespace:kube-system,Attempt:0,}" Mar 17 17:42:33.713107 containerd[1467]: time="2025-03-17T17:42:33.713068201Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86c8b6f988-ppvkt,Uid:c9eab445-c078-490b-861c-b96d204afef3,Namespace:calico-apiserver,Attempt:0,}" Mar 17 17:42:33.719379 containerd[1467]: time="2025-03-17T17:42:33.719320756Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86c8b6f988-79bzl,Uid:3265838f-27ca-489d-8c65-7cadce5cca40,Namespace:calico-apiserver,Attempt:0,}" Mar 17 17:42:33.723443 containerd[1467]: time="2025-03-17T17:42:33.723414258Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7fddcd7f99-92q58,Uid:e531d9e1-a7bb-4958-adfd-8db7b3f84f95,Namespace:calico-system,Attempt:0,}" Mar 17 17:42:33.775835 containerd[1467]: time="2025-03-17T17:42:33.775747357Z" level=error msg="Failed to destroy network for sandbox \"e983d9d5aee8c472a9806708fa057e3b2de9217794b9f28b9fc2133b98942f76\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:33.777031 containerd[1467]: time="2025-03-17T17:42:33.776717172Z" level=error msg="encountered an error cleaning up failed sandbox \"e983d9d5aee8c472a9806708fa057e3b2de9217794b9f28b9fc2133b98942f76\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:33.777210 containerd[1467]: time="2025-03-17T17:42:33.777078432Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-sjrwq,Uid:228c6093-6dc2-4faf-8ec8-87849ef402ab,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"e983d9d5aee8c472a9806708fa057e3b2de9217794b9f28b9fc2133b98942f76\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:33.779163 kubelet[2662]: E0317 17:42:33.778713 2662 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e983d9d5aee8c472a9806708fa057e3b2de9217794b9f28b9fc2133b98942f76\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:33.779163 kubelet[2662]: E0317 17:42:33.778786 2662 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e983d9d5aee8c472a9806708fa057e3b2de9217794b9f28b9fc2133b98942f76\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-sjrwq" Mar 17 17:42:33.779163 kubelet[2662]: E0317 17:42:33.778805 2662 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e983d9d5aee8c472a9806708fa057e3b2de9217794b9f28b9fc2133b98942f76\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-sjrwq" Mar 17 17:42:33.779329 kubelet[2662]: E0317 17:42:33.778852 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-sjrwq_kube-system(228c6093-6dc2-4faf-8ec8-87849ef402ab)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-sjrwq_kube-system(228c6093-6dc2-4faf-8ec8-87849ef402ab)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"e983d9d5aee8c472a9806708fa057e3b2de9217794b9f28b9fc2133b98942f76\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-sjrwq" podUID="228c6093-6dc2-4faf-8ec8-87849ef402ab" Mar 17 17:42:33.782095 containerd[1467]: time="2025-03-17T17:42:33.781843809Z" level=error msg="Failed to destroy network for sandbox \"6c94b5b9e7a3d7b63ccc9268b52bbe37ef401118cf01d372890eeea51b3ec01a\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:33.782276 containerd[1467]: time="2025-03-17T17:42:33.782246586Z" level=error msg="encountered an error cleaning up failed sandbox \"6c94b5b9e7a3d7b63ccc9268b52bbe37ef401118cf01d372890eeea51b3ec01a\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:33.782328 containerd[1467]: time="2025-03-17T17:42:33.782311919Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-fw5hr,Uid:ae45280b-f87f-4410-991d-e0beb4bdc7c6,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"6c94b5b9e7a3d7b63ccc9268b52bbe37ef401118cf01d372890eeea51b3ec01a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:33.782602 kubelet[2662]: E0317 17:42:33.782545 2662 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6c94b5b9e7a3d7b63ccc9268b52bbe37ef401118cf01d372890eeea51b3ec01a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:33.782647 kubelet[2662]: E0317 17:42:33.782626 2662 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6c94b5b9e7a3d7b63ccc9268b52bbe37ef401118cf01d372890eeea51b3ec01a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-fw5hr" Mar 17 17:42:33.782675 kubelet[2662]: E0317 17:42:33.782649 2662 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6c94b5b9e7a3d7b63ccc9268b52bbe37ef401118cf01d372890eeea51b3ec01a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-fw5hr" Mar 17 17:42:33.782786 kubelet[2662]: E0317 17:42:33.782690 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-fw5hr_kube-system(ae45280b-f87f-4410-991d-e0beb4bdc7c6)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-fw5hr_kube-system(ae45280b-f87f-4410-991d-e0beb4bdc7c6)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"6c94b5b9e7a3d7b63ccc9268b52bbe37ef401118cf01d372890eeea51b3ec01a\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-fw5hr" podUID="ae45280b-f87f-4410-991d-e0beb4bdc7c6" Mar 17 17:42:33.825645 containerd[1467]: time="2025-03-17T17:42:33.825288804Z" level=error msg="Failed to destroy network for sandbox \"e93aa7d450e66ea71243d70e3ef70d32837865b756a4bd00a6f296f08f7881d8\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:33.827969 containerd[1467]: time="2025-03-17T17:42:33.826973243Z" level=error msg="encountered an error cleaning up failed sandbox \"e93aa7d450e66ea71243d70e3ef70d32837865b756a4bd00a6f296f08f7881d8\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:33.829092 systemd[1]: Created slice kubepods-besteffort-pod27eb0c6c_40f4_4b19_b513_0003d63df5d0.slice - libcontainer container kubepods-besteffort-pod27eb0c6c_40f4_4b19_b513_0003d63df5d0.slice. Mar 17 17:42:33.830270 containerd[1467]: time="2025-03-17T17:42:33.830115096Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86c8b6f988-ppvkt,Uid:c9eab445-c078-490b-861c-b96d204afef3,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"e93aa7d450e66ea71243d70e3ef70d32837865b756a4bd00a6f296f08f7881d8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:33.831471 kubelet[2662]: E0317 17:42:33.831138 2662 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e93aa7d450e66ea71243d70e3ef70d32837865b756a4bd00a6f296f08f7881d8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:33.831471 kubelet[2662]: E0317 17:42:33.831183 2662 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e93aa7d450e66ea71243d70e3ef70d32837865b756a4bd00a6f296f08f7881d8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86c8b6f988-ppvkt" Mar 17 17:42:33.831471 kubelet[2662]: E0317 17:42:33.831200 2662 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e93aa7d450e66ea71243d70e3ef70d32837865b756a4bd00a6f296f08f7881d8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86c8b6f988-ppvkt" Mar 17 17:42:33.831589 kubelet[2662]: E0317 17:42:33.831242 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-86c8b6f988-ppvkt_calico-apiserver(c9eab445-c078-490b-861c-b96d204afef3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-86c8b6f988-ppvkt_calico-apiserver(c9eab445-c078-490b-861c-b96d204afef3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"e93aa7d450e66ea71243d70e3ef70d32837865b756a4bd00a6f296f08f7881d8\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-86c8b6f988-ppvkt" podUID="c9eab445-c078-490b-861c-b96d204afef3" Mar 17 17:42:33.831925 containerd[1467]: time="2025-03-17T17:42:33.831907617Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-j9gdg,Uid:27eb0c6c-40f4-4b19-b513-0003d63df5d0,Namespace:calico-system,Attempt:0,}" Mar 17 17:42:33.834337 containerd[1467]: time="2025-03-17T17:42:33.834293918Z" level=error msg="Failed to destroy network for sandbox \"a45255c5e842412af83dd06ba8b258ef311550cba3da5240506cd00cf0678895\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:33.834748 containerd[1467]: time="2025-03-17T17:42:33.834641643Z" level=error msg="encountered an error cleaning up failed sandbox \"a45255c5e842412af83dd06ba8b258ef311550cba3da5240506cd00cf0678895\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:33.834748 containerd[1467]: time="2025-03-17T17:42:33.834703499Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86c8b6f988-79bzl,Uid:3265838f-27ca-489d-8c65-7cadce5cca40,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"a45255c5e842412af83dd06ba8b258ef311550cba3da5240506cd00cf0678895\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:33.835098 kubelet[2662]: E0317 17:42:33.835065 2662 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a45255c5e842412af83dd06ba8b258ef311550cba3da5240506cd00cf0678895\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:33.835148 kubelet[2662]: E0317 17:42:33.835098 2662 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a45255c5e842412af83dd06ba8b258ef311550cba3da5240506cd00cf0678895\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86c8b6f988-79bzl" Mar 17 17:42:33.835148 kubelet[2662]: E0317 17:42:33.835115 2662 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a45255c5e842412af83dd06ba8b258ef311550cba3da5240506cd00cf0678895\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86c8b6f988-79bzl" Mar 17 17:42:33.835148 kubelet[2662]: E0317 17:42:33.835139 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-86c8b6f988-79bzl_calico-apiserver(3265838f-27ca-489d-8c65-7cadce5cca40)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-86c8b6f988-79bzl_calico-apiserver(3265838f-27ca-489d-8c65-7cadce5cca40)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"a45255c5e842412af83dd06ba8b258ef311550cba3da5240506cd00cf0678895\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-86c8b6f988-79bzl" podUID="3265838f-27ca-489d-8c65-7cadce5cca40" Mar 17 17:42:33.839986 containerd[1467]: time="2025-03-17T17:42:33.839933119Z" level=error msg="Failed to destroy network for sandbox \"cc57dbc98644c56867a38eb1c7016e6e03a565b54b9bea52d164d9463bd478d9\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:33.840534 containerd[1467]: time="2025-03-17T17:42:33.840500546Z" level=error msg="encountered an error cleaning up failed sandbox \"cc57dbc98644c56867a38eb1c7016e6e03a565b54b9bea52d164d9463bd478d9\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:33.840603 containerd[1467]: time="2025-03-17T17:42:33.840560178Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7fddcd7f99-92q58,Uid:e531d9e1-a7bb-4958-adfd-8db7b3f84f95,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"cc57dbc98644c56867a38eb1c7016e6e03a565b54b9bea52d164d9463bd478d9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:33.840821 kubelet[2662]: E0317 17:42:33.840780 2662 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"cc57dbc98644c56867a38eb1c7016e6e03a565b54b9bea52d164d9463bd478d9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:33.840888 kubelet[2662]: E0317 17:42:33.840831 2662 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"cc57dbc98644c56867a38eb1c7016e6e03a565b54b9bea52d164d9463bd478d9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-7fddcd7f99-92q58" Mar 17 17:42:33.840888 kubelet[2662]: E0317 17:42:33.840850 2662 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"cc57dbc98644c56867a38eb1c7016e6e03a565b54b9bea52d164d9463bd478d9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-7fddcd7f99-92q58" Mar 17 17:42:33.840937 kubelet[2662]: E0317 17:42:33.840890 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-7fddcd7f99-92q58_calico-system(e531d9e1-a7bb-4958-adfd-8db7b3f84f95)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-7fddcd7f99-92q58_calico-system(e531d9e1-a7bb-4958-adfd-8db7b3f84f95)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"cc57dbc98644c56867a38eb1c7016e6e03a565b54b9bea52d164d9463bd478d9\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-7fddcd7f99-92q58" podUID="e531d9e1-a7bb-4958-adfd-8db7b3f84f95" Mar 17 17:42:33.891130 containerd[1467]: time="2025-03-17T17:42:33.891066049Z" level=error msg="Failed to destroy network for sandbox \"b33f35d6e3bd6b1a5f1d8cb21ebceacfc304ffe6c1b7ed3aa40b78efa59e6a03\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:33.891539 containerd[1467]: time="2025-03-17T17:42:33.891499185Z" level=error msg="encountered an error cleaning up failed sandbox \"b33f35d6e3bd6b1a5f1d8cb21ebceacfc304ffe6c1b7ed3aa40b78efa59e6a03\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:33.891584 containerd[1467]: time="2025-03-17T17:42:33.891561422Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-j9gdg,Uid:27eb0c6c-40f4-4b19-b513-0003d63df5d0,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"b33f35d6e3bd6b1a5f1d8cb21ebceacfc304ffe6c1b7ed3aa40b78efa59e6a03\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:33.891855 kubelet[2662]: E0317 17:42:33.891798 2662 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b33f35d6e3bd6b1a5f1d8cb21ebceacfc304ffe6c1b7ed3aa40b78efa59e6a03\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:33.891855 kubelet[2662]: E0317 17:42:33.891874 2662 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b33f35d6e3bd6b1a5f1d8cb21ebceacfc304ffe6c1b7ed3aa40b78efa59e6a03\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-j9gdg" Mar 17 17:42:33.892038 kubelet[2662]: E0317 17:42:33.891899 2662 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b33f35d6e3bd6b1a5f1d8cb21ebceacfc304ffe6c1b7ed3aa40b78efa59e6a03\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-j9gdg" Mar 17 17:42:33.892038 kubelet[2662]: E0317 17:42:33.891962 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-j9gdg_calico-system(27eb0c6c-40f4-4b19-b513-0003d63df5d0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-j9gdg_calico-system(27eb0c6c-40f4-4b19-b513-0003d63df5d0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"b33f35d6e3bd6b1a5f1d8cb21ebceacfc304ffe6c1b7ed3aa40b78efa59e6a03\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-j9gdg" podUID="27eb0c6c-40f4-4b19-b513-0003d63df5d0" Mar 17 17:42:33.973379 kubelet[2662]: I0317 17:42:33.972473 2662 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e983d9d5aee8c472a9806708fa057e3b2de9217794b9f28b9fc2133b98942f76" Mar 17 17:42:33.973912 containerd[1467]: time="2025-03-17T17:42:33.973314267Z" level=info msg="StopPodSandbox for \"e983d9d5aee8c472a9806708fa057e3b2de9217794b9f28b9fc2133b98942f76\"" Mar 17 17:42:33.973912 containerd[1467]: time="2025-03-17T17:42:33.973515706Z" level=info msg="Ensure that sandbox e983d9d5aee8c472a9806708fa057e3b2de9217794b9f28b9fc2133b98942f76 in task-service has been cleanup successfully" Mar 17 17:42:33.974068 kubelet[2662]: I0317 17:42:33.973765 2662 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c94b5b9e7a3d7b63ccc9268b52bbe37ef401118cf01d372890eeea51b3ec01a" Mar 17 17:42:33.974841 kubelet[2662]: E0317 17:42:33.974349 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:33.974899 containerd[1467]: time="2025-03-17T17:42:33.974122357Z" level=info msg="TearDown network for sandbox \"e983d9d5aee8c472a9806708fa057e3b2de9217794b9f28b9fc2133b98942f76\" successfully" Mar 17 17:42:33.974899 containerd[1467]: time="2025-03-17T17:42:33.974137396Z" level=info msg="StopPodSandbox for \"e983d9d5aee8c472a9806708fa057e3b2de9217794b9f28b9fc2133b98942f76\" returns successfully" Mar 17 17:42:33.974899 containerd[1467]: time="2025-03-17T17:42:33.974323086Z" level=info msg="StopPodSandbox for \"6c94b5b9e7a3d7b63ccc9268b52bbe37ef401118cf01d372890eeea51b3ec01a\"" Mar 17 17:42:33.974899 containerd[1467]: time="2025-03-17T17:42:33.974564600Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-sjrwq,Uid:228c6093-6dc2-4faf-8ec8-87849ef402ab,Namespace:kube-system,Attempt:1,}" Mar 17 17:42:33.974899 containerd[1467]: time="2025-03-17T17:42:33.974632697Z" level=info msg="Ensure that sandbox 6c94b5b9e7a3d7b63ccc9268b52bbe37ef401118cf01d372890eeea51b3ec01a in task-service has been cleanup successfully" Mar 17 17:42:33.975082 kubelet[2662]: I0317 17:42:33.975065 2662 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b33f35d6e3bd6b1a5f1d8cb21ebceacfc304ffe6c1b7ed3aa40b78efa59e6a03" Mar 17 17:42:33.975232 containerd[1467]: time="2025-03-17T17:42:33.975187531Z" level=info msg="TearDown network for sandbox \"6c94b5b9e7a3d7b63ccc9268b52bbe37ef401118cf01d372890eeea51b3ec01a\" successfully" Mar 17 17:42:33.975232 containerd[1467]: time="2025-03-17T17:42:33.975211827Z" level=info msg="StopPodSandbox for \"6c94b5b9e7a3d7b63ccc9268b52bbe37ef401118cf01d372890eeea51b3ec01a\" returns successfully" Mar 17 17:42:33.975470 containerd[1467]: time="2025-03-17T17:42:33.975444055Z" level=info msg="StopPodSandbox for \"b33f35d6e3bd6b1a5f1d8cb21ebceacfc304ffe6c1b7ed3aa40b78efa59e6a03\"" Mar 17 17:42:33.975523 kubelet[2662]: E0317 17:42:33.975508 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:33.975625 containerd[1467]: time="2025-03-17T17:42:33.975605158Z" level=info msg="Ensure that sandbox b33f35d6e3bd6b1a5f1d8cb21ebceacfc304ffe6c1b7ed3aa40b78efa59e6a03 in task-service has been cleanup successfully" Mar 17 17:42:33.975868 containerd[1467]: time="2025-03-17T17:42:33.975842765Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-fw5hr,Uid:ae45280b-f87f-4410-991d-e0beb4bdc7c6,Namespace:kube-system,Attempt:1,}" Mar 17 17:42:33.976390 containerd[1467]: time="2025-03-17T17:42:33.976096231Z" level=info msg="TearDown network for sandbox \"b33f35d6e3bd6b1a5f1d8cb21ebceacfc304ffe6c1b7ed3aa40b78efa59e6a03\" successfully" Mar 17 17:42:33.976390 containerd[1467]: time="2025-03-17T17:42:33.976111630Z" level=info msg="StopPodSandbox for \"b33f35d6e3bd6b1a5f1d8cb21ebceacfc304ffe6c1b7ed3aa40b78efa59e6a03\" returns successfully" Mar 17 17:42:33.976543 containerd[1467]: time="2025-03-17T17:42:33.976513888Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-j9gdg,Uid:27eb0c6c-40f4-4b19-b513-0003d63df5d0,Namespace:calico-system,Attempt:1,}" Mar 17 17:42:33.977453 kubelet[2662]: I0317 17:42:33.977101 2662 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc57dbc98644c56867a38eb1c7016e6e03a565b54b9bea52d164d9463bd478d9" Mar 17 17:42:33.977836 containerd[1467]: time="2025-03-17T17:42:33.977726198Z" level=info msg="StopPodSandbox for \"cc57dbc98644c56867a38eb1c7016e6e03a565b54b9bea52d164d9463bd478d9\"" Mar 17 17:42:33.978021 containerd[1467]: time="2025-03-17T17:42:33.978000525Z" level=info msg="Ensure that sandbox cc57dbc98644c56867a38eb1c7016e6e03a565b54b9bea52d164d9463bd478d9 in task-service has been cleanup successfully" Mar 17 17:42:33.978376 containerd[1467]: time="2025-03-17T17:42:33.978303094Z" level=info msg="TearDown network for sandbox \"cc57dbc98644c56867a38eb1c7016e6e03a565b54b9bea52d164d9463bd478d9\" successfully" Mar 17 17:42:33.978376 containerd[1467]: time="2025-03-17T17:42:33.978325586Z" level=info msg="StopPodSandbox for \"cc57dbc98644c56867a38eb1c7016e6e03a565b54b9bea52d164d9463bd478d9\" returns successfully" Mar 17 17:42:33.978717 containerd[1467]: time="2025-03-17T17:42:33.978694531Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7fddcd7f99-92q58,Uid:e531d9e1-a7bb-4958-adfd-8db7b3f84f95,Namespace:calico-system,Attempt:1,}" Mar 17 17:42:33.979149 kubelet[2662]: E0317 17:42:33.978831 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:33.979676 containerd[1467]: time="2025-03-17T17:42:33.979644097Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.2\"" Mar 17 17:42:33.981638 kubelet[2662]: I0317 17:42:33.980994 2662 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a45255c5e842412af83dd06ba8b258ef311550cba3da5240506cd00cf0678895" Mar 17 17:42:33.982001 containerd[1467]: time="2025-03-17T17:42:33.981385294Z" level=info msg="StopPodSandbox for \"a45255c5e842412af83dd06ba8b258ef311550cba3da5240506cd00cf0678895\"" Mar 17 17:42:33.982001 containerd[1467]: time="2025-03-17T17:42:33.981571915Z" level=info msg="Ensure that sandbox a45255c5e842412af83dd06ba8b258ef311550cba3da5240506cd00cf0678895 in task-service has been cleanup successfully" Mar 17 17:42:33.982001 containerd[1467]: time="2025-03-17T17:42:33.981844477Z" level=info msg="TearDown network for sandbox \"a45255c5e842412af83dd06ba8b258ef311550cba3da5240506cd00cf0678895\" successfully" Mar 17 17:42:33.982001 containerd[1467]: time="2025-03-17T17:42:33.981886366Z" level=info msg="StopPodSandbox for \"a45255c5e842412af83dd06ba8b258ef311550cba3da5240506cd00cf0678895\" returns successfully" Mar 17 17:42:33.982153 kubelet[2662]: I0317 17:42:33.981991 2662 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e93aa7d450e66ea71243d70e3ef70d32837865b756a4bd00a6f296f08f7881d8" Mar 17 17:42:33.982383 containerd[1467]: time="2025-03-17T17:42:33.982333798Z" level=info msg="StopPodSandbox for \"e93aa7d450e66ea71243d70e3ef70d32837865b756a4bd00a6f296f08f7881d8\"" Mar 17 17:42:33.982516 containerd[1467]: time="2025-03-17T17:42:33.982493519Z" level=info msg="Ensure that sandbox e93aa7d450e66ea71243d70e3ef70d32837865b756a4bd00a6f296f08f7881d8 in task-service has been cleanup successfully" Mar 17 17:42:33.982772 containerd[1467]: time="2025-03-17T17:42:33.982674820Z" level=info msg="TearDown network for sandbox \"e93aa7d450e66ea71243d70e3ef70d32837865b756a4bd00a6f296f08f7881d8\" successfully" Mar 17 17:42:33.982772 containerd[1467]: time="2025-03-17T17:42:33.982699616Z" level=info msg="StopPodSandbox for \"e93aa7d450e66ea71243d70e3ef70d32837865b756a4bd00a6f296f08f7881d8\" returns successfully" Mar 17 17:42:33.982973 containerd[1467]: time="2025-03-17T17:42:33.982930671Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86c8b6f988-79bzl,Uid:3265838f-27ca-489d-8c65-7cadce5cca40,Namespace:calico-apiserver,Attempt:1,}" Mar 17 17:42:33.983803 containerd[1467]: time="2025-03-17T17:42:33.983583139Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86c8b6f988-ppvkt,Uid:c9eab445-c078-490b-861c-b96d204afef3,Namespace:calico-apiserver,Attempt:1,}" Mar 17 17:42:34.130168 containerd[1467]: time="2025-03-17T17:42:34.130107262Z" level=error msg="Failed to destroy network for sandbox \"3f78c7317a0c3224f4039412ac38faee93bb589fa11754f46d7d13d15315c3d5\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:34.130921 containerd[1467]: time="2025-03-17T17:42:34.130793533Z" level=error msg="encountered an error cleaning up failed sandbox \"3f78c7317a0c3224f4039412ac38faee93bb589fa11754f46d7d13d15315c3d5\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:34.130921 containerd[1467]: time="2025-03-17T17:42:34.130871941Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-fw5hr,Uid:ae45280b-f87f-4410-991d-e0beb4bdc7c6,Namespace:kube-system,Attempt:1,} failed, error" error="failed to setup network for sandbox \"3f78c7317a0c3224f4039412ac38faee93bb589fa11754f46d7d13d15315c3d5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:34.131549 kubelet[2662]: E0317 17:42:34.131302 2662 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3f78c7317a0c3224f4039412ac38faee93bb589fa11754f46d7d13d15315c3d5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:34.131549 kubelet[2662]: E0317 17:42:34.131396 2662 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3f78c7317a0c3224f4039412ac38faee93bb589fa11754f46d7d13d15315c3d5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-fw5hr" Mar 17 17:42:34.131549 kubelet[2662]: E0317 17:42:34.131438 2662 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3f78c7317a0c3224f4039412ac38faee93bb589fa11754f46d7d13d15315c3d5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-fw5hr" Mar 17 17:42:34.131709 kubelet[2662]: E0317 17:42:34.131502 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-fw5hr_kube-system(ae45280b-f87f-4410-991d-e0beb4bdc7c6)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-fw5hr_kube-system(ae45280b-f87f-4410-991d-e0beb4bdc7c6)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"3f78c7317a0c3224f4039412ac38faee93bb589fa11754f46d7d13d15315c3d5\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-fw5hr" podUID="ae45280b-f87f-4410-991d-e0beb4bdc7c6" Mar 17 17:42:34.134131 containerd[1467]: time="2025-03-17T17:42:34.134096177Z" level=error msg="Failed to destroy network for sandbox \"47b88b63e50185559458c168aee26da05df22f8d62d10c49042400bc79ba64b6\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:34.136511 containerd[1467]: time="2025-03-17T17:42:34.136464081Z" level=error msg="encountered an error cleaning up failed sandbox \"47b88b63e50185559458c168aee26da05df22f8d62d10c49042400bc79ba64b6\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:34.136586 containerd[1467]: time="2025-03-17T17:42:34.136537639Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-j9gdg,Uid:27eb0c6c-40f4-4b19-b513-0003d63df5d0,Namespace:calico-system,Attempt:1,} failed, error" error="failed to setup network for sandbox \"47b88b63e50185559458c168aee26da05df22f8d62d10c49042400bc79ba64b6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:34.136867 kubelet[2662]: E0317 17:42:34.136831 2662 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"47b88b63e50185559458c168aee26da05df22f8d62d10c49042400bc79ba64b6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:34.136976 kubelet[2662]: E0317 17:42:34.136888 2662 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"47b88b63e50185559458c168aee26da05df22f8d62d10c49042400bc79ba64b6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-j9gdg" Mar 17 17:42:34.136976 kubelet[2662]: E0317 17:42:34.136912 2662 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"47b88b63e50185559458c168aee26da05df22f8d62d10c49042400bc79ba64b6\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-j9gdg" Mar 17 17:42:34.136976 kubelet[2662]: E0317 17:42:34.136957 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-j9gdg_calico-system(27eb0c6c-40f4-4b19-b513-0003d63df5d0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-j9gdg_calico-system(27eb0c6c-40f4-4b19-b513-0003d63df5d0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"47b88b63e50185559458c168aee26da05df22f8d62d10c49042400bc79ba64b6\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-j9gdg" podUID="27eb0c6c-40f4-4b19-b513-0003d63df5d0" Mar 17 17:42:34.144867 containerd[1467]: time="2025-03-17T17:42:34.144821543Z" level=error msg="Failed to destroy network for sandbox \"3f5345f62ce3def9878f952bc3b9ba8515b066f140a947d652235852cb08fae5\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:34.145246 containerd[1467]: time="2025-03-17T17:42:34.145215174Z" level=error msg="encountered an error cleaning up failed sandbox \"3f5345f62ce3def9878f952bc3b9ba8515b066f140a947d652235852cb08fae5\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:34.145306 containerd[1467]: time="2025-03-17T17:42:34.145277531Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86c8b6f988-ppvkt,Uid:c9eab445-c078-490b-861c-b96d204afef3,Namespace:calico-apiserver,Attempt:1,} failed, error" error="failed to setup network for sandbox \"3f5345f62ce3def9878f952bc3b9ba8515b066f140a947d652235852cb08fae5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:34.145537 kubelet[2662]: E0317 17:42:34.145496 2662 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3f5345f62ce3def9878f952bc3b9ba8515b066f140a947d652235852cb08fae5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:34.145592 kubelet[2662]: E0317 17:42:34.145562 2662 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3f5345f62ce3def9878f952bc3b9ba8515b066f140a947d652235852cb08fae5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86c8b6f988-ppvkt" Mar 17 17:42:34.145630 kubelet[2662]: E0317 17:42:34.145589 2662 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3f5345f62ce3def9878f952bc3b9ba8515b066f140a947d652235852cb08fae5\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86c8b6f988-ppvkt" Mar 17 17:42:34.145918 kubelet[2662]: E0317 17:42:34.145654 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-86c8b6f988-ppvkt_calico-apiserver(c9eab445-c078-490b-861c-b96d204afef3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-86c8b6f988-ppvkt_calico-apiserver(c9eab445-c078-490b-861c-b96d204afef3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"3f5345f62ce3def9878f952bc3b9ba8515b066f140a947d652235852cb08fae5\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-86c8b6f988-ppvkt" podUID="c9eab445-c078-490b-861c-b96d204afef3" Mar 17 17:42:34.147859 containerd[1467]: time="2025-03-17T17:42:34.147683728Z" level=error msg="Failed to destroy network for sandbox \"f1d105acdc9b53357e4e621942df86a7fba22a3900a30edbb033cb1322dbced2\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:34.148159 containerd[1467]: time="2025-03-17T17:42:34.148137311Z" level=error msg="encountered an error cleaning up failed sandbox \"f1d105acdc9b53357e4e621942df86a7fba22a3900a30edbb033cb1322dbced2\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:34.148283 containerd[1467]: time="2025-03-17T17:42:34.148256405Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-sjrwq,Uid:228c6093-6dc2-4faf-8ec8-87849ef402ab,Namespace:kube-system,Attempt:1,} failed, error" error="failed to setup network for sandbox \"f1d105acdc9b53357e4e621942df86a7fba22a3900a30edbb033cb1322dbced2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:34.148656 kubelet[2662]: E0317 17:42:34.148608 2662 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f1d105acdc9b53357e4e621942df86a7fba22a3900a30edbb033cb1322dbced2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:34.148754 kubelet[2662]: E0317 17:42:34.148738 2662 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f1d105acdc9b53357e4e621942df86a7fba22a3900a30edbb033cb1322dbced2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-sjrwq" Mar 17 17:42:34.148831 kubelet[2662]: E0317 17:42:34.148809 2662 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f1d105acdc9b53357e4e621942df86a7fba22a3900a30edbb033cb1322dbced2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-sjrwq" Mar 17 17:42:34.148892 kubelet[2662]: E0317 17:42:34.148851 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-sjrwq_kube-system(228c6093-6dc2-4faf-8ec8-87849ef402ab)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-sjrwq_kube-system(228c6093-6dc2-4faf-8ec8-87849ef402ab)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"f1d105acdc9b53357e4e621942df86a7fba22a3900a30edbb033cb1322dbced2\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-sjrwq" podUID="228c6093-6dc2-4faf-8ec8-87849ef402ab" Mar 17 17:42:34.156430 containerd[1467]: time="2025-03-17T17:42:34.156379567Z" level=error msg="Failed to destroy network for sandbox \"317763565ecd927b3fe20fd547beaf2029d8ad181454207c47710163dbd09a88\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:34.156937 containerd[1467]: time="2025-03-17T17:42:34.156789498Z" level=error msg="encountered an error cleaning up failed sandbox \"317763565ecd927b3fe20fd547beaf2029d8ad181454207c47710163dbd09a88\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:34.156937 containerd[1467]: time="2025-03-17T17:42:34.156844260Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7fddcd7f99-92q58,Uid:e531d9e1-a7bb-4958-adfd-8db7b3f84f95,Namespace:calico-system,Attempt:1,} failed, error" error="failed to setup network for sandbox \"317763565ecd927b3fe20fd547beaf2029d8ad181454207c47710163dbd09a88\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:34.157242 kubelet[2662]: E0317 17:42:34.157061 2662 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"317763565ecd927b3fe20fd547beaf2029d8ad181454207c47710163dbd09a88\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:34.157242 kubelet[2662]: E0317 17:42:34.157119 2662 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"317763565ecd927b3fe20fd547beaf2029d8ad181454207c47710163dbd09a88\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-7fddcd7f99-92q58" Mar 17 17:42:34.157242 kubelet[2662]: E0317 17:42:34.157138 2662 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"317763565ecd927b3fe20fd547beaf2029d8ad181454207c47710163dbd09a88\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-7fddcd7f99-92q58" Mar 17 17:42:34.157340 kubelet[2662]: E0317 17:42:34.157176 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-7fddcd7f99-92q58_calico-system(e531d9e1-a7bb-4958-adfd-8db7b3f84f95)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-7fddcd7f99-92q58_calico-system(e531d9e1-a7bb-4958-adfd-8db7b3f84f95)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"317763565ecd927b3fe20fd547beaf2029d8ad181454207c47710163dbd09a88\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-7fddcd7f99-92q58" podUID="e531d9e1-a7bb-4958-adfd-8db7b3f84f95" Mar 17 17:42:34.162729 containerd[1467]: time="2025-03-17T17:42:34.162685129Z" level=error msg="Failed to destroy network for sandbox \"f905ddf13c6f6a970be32393ab7c951bb1766cec8d75bdc0aca8519fcc2783c7\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:34.163065 containerd[1467]: time="2025-03-17T17:42:34.163036981Z" level=error msg="encountered an error cleaning up failed sandbox \"f905ddf13c6f6a970be32393ab7c951bb1766cec8d75bdc0aca8519fcc2783c7\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:34.163123 containerd[1467]: time="2025-03-17T17:42:34.163098817Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86c8b6f988-79bzl,Uid:3265838f-27ca-489d-8c65-7cadce5cca40,Namespace:calico-apiserver,Attempt:1,} failed, error" error="failed to setup network for sandbox \"f905ddf13c6f6a970be32393ab7c951bb1766cec8d75bdc0aca8519fcc2783c7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:34.163317 kubelet[2662]: E0317 17:42:34.163272 2662 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f905ddf13c6f6a970be32393ab7c951bb1766cec8d75bdc0aca8519fcc2783c7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:34.163408 kubelet[2662]: E0317 17:42:34.163324 2662 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f905ddf13c6f6a970be32393ab7c951bb1766cec8d75bdc0aca8519fcc2783c7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86c8b6f988-79bzl" Mar 17 17:42:34.163408 kubelet[2662]: E0317 17:42:34.163344 2662 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f905ddf13c6f6a970be32393ab7c951bb1766cec8d75bdc0aca8519fcc2783c7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86c8b6f988-79bzl" Mar 17 17:42:34.163408 kubelet[2662]: E0317 17:42:34.163388 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-86c8b6f988-79bzl_calico-apiserver(3265838f-27ca-489d-8c65-7cadce5cca40)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-86c8b6f988-79bzl_calico-apiserver(3265838f-27ca-489d-8c65-7cadce5cca40)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"f905ddf13c6f6a970be32393ab7c951bb1766cec8d75bdc0aca8519fcc2783c7\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-86c8b6f988-79bzl" podUID="3265838f-27ca-489d-8c65-7cadce5cca40" Mar 17 17:42:34.985112 kubelet[2662]: I0317 17:42:34.985055 2662 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f5345f62ce3def9878f952bc3b9ba8515b066f140a947d652235852cb08fae5" Mar 17 17:42:34.985775 containerd[1467]: time="2025-03-17T17:42:34.985744394Z" level=info msg="StopPodSandbox for \"3f5345f62ce3def9878f952bc3b9ba8515b066f140a947d652235852cb08fae5\"" Mar 17 17:42:34.986005 containerd[1467]: time="2025-03-17T17:42:34.985941495Z" level=info msg="Ensure that sandbox 3f5345f62ce3def9878f952bc3b9ba8515b066f140a947d652235852cb08fae5 in task-service has been cleanup successfully" Mar 17 17:42:34.986952 containerd[1467]: time="2025-03-17T17:42:34.986847739Z" level=info msg="TearDown network for sandbox \"3f5345f62ce3def9878f952bc3b9ba8515b066f140a947d652235852cb08fae5\" successfully" Mar 17 17:42:34.986952 containerd[1467]: time="2025-03-17T17:42:34.986876984Z" level=info msg="StopPodSandbox for \"3f5345f62ce3def9878f952bc3b9ba8515b066f140a947d652235852cb08fae5\" returns successfully" Mar 17 17:42:34.987886 kubelet[2662]: I0317 17:42:34.987853 2662 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1d105acdc9b53357e4e621942df86a7fba22a3900a30edbb033cb1322dbced2" Mar 17 17:42:34.988156 containerd[1467]: time="2025-03-17T17:42:34.988117739Z" level=info msg="StopPodSandbox for \"e93aa7d450e66ea71243d70e3ef70d32837865b756a4bd00a6f296f08f7881d8\"" Mar 17 17:42:34.988216 systemd[1]: run-netns-cni\x2d566f19a0\x2db057\x2d22c0\x2d1d80\x2d2dc69d51d534.mount: Deactivated successfully. Mar 17 17:42:34.988552 containerd[1467]: time="2025-03-17T17:42:34.988253945Z" level=info msg="TearDown network for sandbox \"e93aa7d450e66ea71243d70e3ef70d32837865b756a4bd00a6f296f08f7881d8\" successfully" Mar 17 17:42:34.988552 containerd[1467]: time="2025-03-17T17:42:34.988265216Z" level=info msg="StopPodSandbox for \"e93aa7d450e66ea71243d70e3ef70d32837865b756a4bd00a6f296f08f7881d8\" returns successfully" Mar 17 17:42:34.988552 containerd[1467]: time="2025-03-17T17:42:34.988348673Z" level=info msg="StopPodSandbox for \"f1d105acdc9b53357e4e621942df86a7fba22a3900a30edbb033cb1322dbced2\"" Mar 17 17:42:34.988880 containerd[1467]: time="2025-03-17T17:42:34.988540524Z" level=info msg="Ensure that sandbox f1d105acdc9b53357e4e621942df86a7fba22a3900a30edbb033cb1322dbced2 in task-service has been cleanup successfully" Mar 17 17:42:34.989429 containerd[1467]: time="2025-03-17T17:42:34.989118982Z" level=info msg="TearDown network for sandbox \"f1d105acdc9b53357e4e621942df86a7fba22a3900a30edbb033cb1322dbced2\" successfully" Mar 17 17:42:34.989429 containerd[1467]: time="2025-03-17T17:42:34.989140452Z" level=info msg="StopPodSandbox for \"f1d105acdc9b53357e4e621942df86a7fba22a3900a30edbb033cb1322dbced2\" returns successfully" Mar 17 17:42:34.990983 containerd[1467]: time="2025-03-17T17:42:34.990066445Z" level=info msg="StopPodSandbox for \"e983d9d5aee8c472a9806708fa057e3b2de9217794b9f28b9fc2133b98942f76\"" Mar 17 17:42:34.990983 containerd[1467]: time="2025-03-17T17:42:34.990163667Z" level=info msg="TearDown network for sandbox \"e983d9d5aee8c472a9806708fa057e3b2de9217794b9f28b9fc2133b98942f76\" successfully" Mar 17 17:42:34.990983 containerd[1467]: time="2025-03-17T17:42:34.990185489Z" level=info msg="StopPodSandbox for \"e983d9d5aee8c472a9806708fa057e3b2de9217794b9f28b9fc2133b98942f76\" returns successfully" Mar 17 17:42:34.990983 containerd[1467]: time="2025-03-17T17:42:34.990565895Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86c8b6f988-ppvkt,Uid:c9eab445-c078-490b-861c-b96d204afef3,Namespace:calico-apiserver,Attempt:2,}" Mar 17 17:42:34.991092 systemd[1]: run-netns-cni\x2dde3a2d8e\x2df16b\x2d5169\x2de8e1\x2d158f82dd9f92.mount: Deactivated successfully. Mar 17 17:42:34.991940 kubelet[2662]: E0317 17:42:34.991897 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:34.992478 containerd[1467]: time="2025-03-17T17:42:34.992165634Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-sjrwq,Uid:228c6093-6dc2-4faf-8ec8-87849ef402ab,Namespace:kube-system,Attempt:2,}" Mar 17 17:42:34.993862 kubelet[2662]: I0317 17:42:34.993832 2662 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f78c7317a0c3224f4039412ac38faee93bb589fa11754f46d7d13d15315c3d5" Mar 17 17:42:34.994504 containerd[1467]: time="2025-03-17T17:42:34.994475379Z" level=info msg="StopPodSandbox for \"3f78c7317a0c3224f4039412ac38faee93bb589fa11754f46d7d13d15315c3d5\"" Mar 17 17:42:34.994699 containerd[1467]: time="2025-03-17T17:42:34.994677760Z" level=info msg="Ensure that sandbox 3f78c7317a0c3224f4039412ac38faee93bb589fa11754f46d7d13d15315c3d5 in task-service has been cleanup successfully" Mar 17 17:42:34.996735 systemd[1]: run-netns-cni\x2d1cc7617c\x2d0954\x2d5b58\x2d9988\x2d9c842266d9a3.mount: Deactivated successfully. Mar 17 17:42:34.996909 containerd[1467]: time="2025-03-17T17:42:34.996871116Z" level=info msg="TearDown network for sandbox \"3f78c7317a0c3224f4039412ac38faee93bb589fa11754f46d7d13d15315c3d5\" successfully" Mar 17 17:42:34.996974 containerd[1467]: time="2025-03-17T17:42:34.996953972Z" level=info msg="StopPodSandbox for \"3f78c7317a0c3224f4039412ac38faee93bb589fa11754f46d7d13d15315c3d5\" returns successfully" Mar 17 17:42:34.997825 containerd[1467]: time="2025-03-17T17:42:34.997593515Z" level=info msg="StopPodSandbox for \"6c94b5b9e7a3d7b63ccc9268b52bbe37ef401118cf01d372890eeea51b3ec01a\"" Mar 17 17:42:34.997825 containerd[1467]: time="2025-03-17T17:42:34.997745551Z" level=info msg="TearDown network for sandbox \"6c94b5b9e7a3d7b63ccc9268b52bbe37ef401118cf01d372890eeea51b3ec01a\" successfully" Mar 17 17:42:34.997825 containerd[1467]: time="2025-03-17T17:42:34.997782360Z" level=info msg="StopPodSandbox for \"6c94b5b9e7a3d7b63ccc9268b52bbe37ef401118cf01d372890eeea51b3ec01a\" returns successfully" Mar 17 17:42:34.998573 kubelet[2662]: E0317 17:42:34.998548 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:34.999028 containerd[1467]: time="2025-03-17T17:42:34.998959875Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-fw5hr,Uid:ae45280b-f87f-4410-991d-e0beb4bdc7c6,Namespace:kube-system,Attempt:2,}" Mar 17 17:42:34.999426 kubelet[2662]: I0317 17:42:34.999400 2662 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47b88b63e50185559458c168aee26da05df22f8d62d10c49042400bc79ba64b6" Mar 17 17:42:34.999824 containerd[1467]: time="2025-03-17T17:42:34.999805977Z" level=info msg="StopPodSandbox for \"47b88b63e50185559458c168aee26da05df22f8d62d10c49042400bc79ba64b6\"" Mar 17 17:42:34.999967 containerd[1467]: time="2025-03-17T17:42:34.999950749Z" level=info msg="Ensure that sandbox 47b88b63e50185559458c168aee26da05df22f8d62d10c49042400bc79ba64b6 in task-service has been cleanup successfully" Mar 17 17:42:35.000158 containerd[1467]: time="2025-03-17T17:42:35.000137731Z" level=info msg="TearDown network for sandbox \"47b88b63e50185559458c168aee26da05df22f8d62d10c49042400bc79ba64b6\" successfully" Mar 17 17:42:35.000158 containerd[1467]: time="2025-03-17T17:42:35.000155374Z" level=info msg="StopPodSandbox for \"47b88b63e50185559458c168aee26da05df22f8d62d10c49042400bc79ba64b6\" returns successfully" Mar 17 17:42:35.000920 containerd[1467]: time="2025-03-17T17:42:35.000607755Z" level=info msg="StopPodSandbox for \"b33f35d6e3bd6b1a5f1d8cb21ebceacfc304ffe6c1b7ed3aa40b78efa59e6a03\"" Mar 17 17:42:35.000920 containerd[1467]: time="2025-03-17T17:42:35.000689549Z" level=info msg="TearDown network for sandbox \"b33f35d6e3bd6b1a5f1d8cb21ebceacfc304ffe6c1b7ed3aa40b78efa59e6a03\" successfully" Mar 17 17:42:35.000920 containerd[1467]: time="2025-03-17T17:42:35.000703626Z" level=info msg="StopPodSandbox for \"b33f35d6e3bd6b1a5f1d8cb21ebceacfc304ffe6c1b7ed3aa40b78efa59e6a03\" returns successfully" Mar 17 17:42:35.001432 containerd[1467]: time="2025-03-17T17:42:35.001399325Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-j9gdg,Uid:27eb0c6c-40f4-4b19-b513-0003d63df5d0,Namespace:calico-system,Attempt:2,}" Mar 17 17:42:35.002057 kubelet[2662]: I0317 17:42:35.002034 2662 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="317763565ecd927b3fe20fd547beaf2029d8ad181454207c47710163dbd09a88" Mar 17 17:42:35.002564 containerd[1467]: time="2025-03-17T17:42:35.002517989Z" level=info msg="StopPodSandbox for \"317763565ecd927b3fe20fd547beaf2029d8ad181454207c47710163dbd09a88\"" Mar 17 17:42:35.002599 systemd[1]: run-netns-cni\x2db5598cac\x2dc298\x2d318e\x2d580f\x2dc4d385e7f016.mount: Deactivated successfully. Mar 17 17:42:35.003253 containerd[1467]: time="2025-03-17T17:42:35.003005285Z" level=info msg="Ensure that sandbox 317763565ecd927b3fe20fd547beaf2029d8ad181454207c47710163dbd09a88 in task-service has been cleanup successfully" Mar 17 17:42:35.003716 containerd[1467]: time="2025-03-17T17:42:35.003548487Z" level=info msg="TearDown network for sandbox \"317763565ecd927b3fe20fd547beaf2029d8ad181454207c47710163dbd09a88\" successfully" Mar 17 17:42:35.003716 containerd[1467]: time="2025-03-17T17:42:35.003565639Z" level=info msg="StopPodSandbox for \"317763565ecd927b3fe20fd547beaf2029d8ad181454207c47710163dbd09a88\" returns successfully" Mar 17 17:42:35.003870 containerd[1467]: time="2025-03-17T17:42:35.003833513Z" level=info msg="StopPodSandbox for \"cc57dbc98644c56867a38eb1c7016e6e03a565b54b9bea52d164d9463bd478d9\"" Mar 17 17:42:35.003996 containerd[1467]: time="2025-03-17T17:42:35.003946796Z" level=info msg="TearDown network for sandbox \"cc57dbc98644c56867a38eb1c7016e6e03a565b54b9bea52d164d9463bd478d9\" successfully" Mar 17 17:42:35.003996 containerd[1467]: time="2025-03-17T17:42:35.003992592Z" level=info msg="StopPodSandbox for \"cc57dbc98644c56867a38eb1c7016e6e03a565b54b9bea52d164d9463bd478d9\" returns successfully" Mar 17 17:42:35.004806 containerd[1467]: time="2025-03-17T17:42:35.004526046Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7fddcd7f99-92q58,Uid:e531d9e1-a7bb-4958-adfd-8db7b3f84f95,Namespace:calico-system,Attempt:2,}" Mar 17 17:42:35.005233 kubelet[2662]: I0317 17:42:35.005216 2662 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f905ddf13c6f6a970be32393ab7c951bb1766cec8d75bdc0aca8519fcc2783c7" Mar 17 17:42:35.005659 containerd[1467]: time="2025-03-17T17:42:35.005626235Z" level=info msg="StopPodSandbox for \"f905ddf13c6f6a970be32393ab7c951bb1766cec8d75bdc0aca8519fcc2783c7\"" Mar 17 17:42:35.005813 containerd[1467]: time="2025-03-17T17:42:35.005786156Z" level=info msg="Ensure that sandbox f905ddf13c6f6a970be32393ab7c951bb1766cec8d75bdc0aca8519fcc2783c7 in task-service has been cleanup successfully" Mar 17 17:42:35.005986 containerd[1467]: time="2025-03-17T17:42:35.005963259Z" level=info msg="TearDown network for sandbox \"f905ddf13c6f6a970be32393ab7c951bb1766cec8d75bdc0aca8519fcc2783c7\" successfully" Mar 17 17:42:35.005986 containerd[1467]: time="2025-03-17T17:42:35.005979770Z" level=info msg="StopPodSandbox for \"f905ddf13c6f6a970be32393ab7c951bb1766cec8d75bdc0aca8519fcc2783c7\" returns successfully" Mar 17 17:42:35.006184 containerd[1467]: time="2025-03-17T17:42:35.006156012Z" level=info msg="StopPodSandbox for \"a45255c5e842412af83dd06ba8b258ef311550cba3da5240506cd00cf0678895\"" Mar 17 17:42:35.006270 containerd[1467]: time="2025-03-17T17:42:35.006252823Z" level=info msg="TearDown network for sandbox \"a45255c5e842412af83dd06ba8b258ef311550cba3da5240506cd00cf0678895\" successfully" Mar 17 17:42:35.006270 containerd[1467]: time="2025-03-17T17:42:35.006267391Z" level=info msg="StopPodSandbox for \"a45255c5e842412af83dd06ba8b258ef311550cba3da5240506cd00cf0678895\" returns successfully" Mar 17 17:42:35.006724 containerd[1467]: time="2025-03-17T17:42:35.006641515Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86c8b6f988-79bzl,Uid:3265838f-27ca-489d-8c65-7cadce5cca40,Namespace:calico-apiserver,Attempt:2,}" Mar 17 17:42:35.315883 containerd[1467]: time="2025-03-17T17:42:35.315743069Z" level=error msg="Failed to destroy network for sandbox \"4044ac7223d418a825ef4eedf876ba4769d2804921eba6925b3d0d6d4afbc33a\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:35.316239 containerd[1467]: time="2025-03-17T17:42:35.316209016Z" level=error msg="encountered an error cleaning up failed sandbox \"4044ac7223d418a825ef4eedf876ba4769d2804921eba6925b3d0d6d4afbc33a\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:35.316305 containerd[1467]: time="2025-03-17T17:42:35.316271674Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86c8b6f988-ppvkt,Uid:c9eab445-c078-490b-861c-b96d204afef3,Namespace:calico-apiserver,Attempt:2,} failed, error" error="failed to setup network for sandbox \"4044ac7223d418a825ef4eedf876ba4769d2804921eba6925b3d0d6d4afbc33a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:35.316696 kubelet[2662]: E0317 17:42:35.316629 2662 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4044ac7223d418a825ef4eedf876ba4769d2804921eba6925b3d0d6d4afbc33a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:35.316764 kubelet[2662]: E0317 17:42:35.316717 2662 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4044ac7223d418a825ef4eedf876ba4769d2804921eba6925b3d0d6d4afbc33a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86c8b6f988-ppvkt" Mar 17 17:42:35.316813 kubelet[2662]: E0317 17:42:35.316762 2662 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4044ac7223d418a825ef4eedf876ba4769d2804921eba6925b3d0d6d4afbc33a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86c8b6f988-ppvkt" Mar 17 17:42:35.316854 kubelet[2662]: E0317 17:42:35.316813 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-86c8b6f988-ppvkt_calico-apiserver(c9eab445-c078-490b-861c-b96d204afef3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-86c8b6f988-ppvkt_calico-apiserver(c9eab445-c078-490b-861c-b96d204afef3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"4044ac7223d418a825ef4eedf876ba4769d2804921eba6925b3d0d6d4afbc33a\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-86c8b6f988-ppvkt" podUID="c9eab445-c078-490b-861c-b96d204afef3" Mar 17 17:42:35.552095 containerd[1467]: time="2025-03-17T17:42:35.552025909Z" level=error msg="Failed to destroy network for sandbox \"3edf99fd7e4c4bcd608727ff216f57eed4c9785e1d1f4d8c623f0c2aa6ca839a\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:35.552753 containerd[1467]: time="2025-03-17T17:42:35.552726205Z" level=error msg="encountered an error cleaning up failed sandbox \"3edf99fd7e4c4bcd608727ff216f57eed4c9785e1d1f4d8c623f0c2aa6ca839a\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:35.552931 containerd[1467]: time="2025-03-17T17:42:35.552891216Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-sjrwq,Uid:228c6093-6dc2-4faf-8ec8-87849ef402ab,Namespace:kube-system,Attempt:2,} failed, error" error="failed to setup network for sandbox \"3edf99fd7e4c4bcd608727ff216f57eed4c9785e1d1f4d8c623f0c2aa6ca839a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:35.553821 kubelet[2662]: E0317 17:42:35.553450 2662 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3edf99fd7e4c4bcd608727ff216f57eed4c9785e1d1f4d8c623f0c2aa6ca839a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:35.553821 kubelet[2662]: E0317 17:42:35.553516 2662 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3edf99fd7e4c4bcd608727ff216f57eed4c9785e1d1f4d8c623f0c2aa6ca839a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-sjrwq" Mar 17 17:42:35.553821 kubelet[2662]: E0317 17:42:35.553535 2662 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3edf99fd7e4c4bcd608727ff216f57eed4c9785e1d1f4d8c623f0c2aa6ca839a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-sjrwq" Mar 17 17:42:35.553945 kubelet[2662]: E0317 17:42:35.553573 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-sjrwq_kube-system(228c6093-6dc2-4faf-8ec8-87849ef402ab)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-sjrwq_kube-system(228c6093-6dc2-4faf-8ec8-87849ef402ab)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"3edf99fd7e4c4bcd608727ff216f57eed4c9785e1d1f4d8c623f0c2aa6ca839a\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-sjrwq" podUID="228c6093-6dc2-4faf-8ec8-87849ef402ab" Mar 17 17:42:35.566537 containerd[1467]: time="2025-03-17T17:42:35.566287804Z" level=error msg="Failed to destroy network for sandbox \"284f791e9331db53c5a68aa2679f439389dd260cc6a34b8aff9bc622e96ca04d\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:35.567020 containerd[1467]: time="2025-03-17T17:42:35.566873015Z" level=error msg="encountered an error cleaning up failed sandbox \"284f791e9331db53c5a68aa2679f439389dd260cc6a34b8aff9bc622e96ca04d\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:35.567020 containerd[1467]: time="2025-03-17T17:42:35.566931185Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-fw5hr,Uid:ae45280b-f87f-4410-991d-e0beb4bdc7c6,Namespace:kube-system,Attempt:2,} failed, error" error="failed to setup network for sandbox \"284f791e9331db53c5a68aa2679f439389dd260cc6a34b8aff9bc622e96ca04d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:35.567221 kubelet[2662]: E0317 17:42:35.567158 2662 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"284f791e9331db53c5a68aa2679f439389dd260cc6a34b8aff9bc622e96ca04d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:35.567296 kubelet[2662]: E0317 17:42:35.567244 2662 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"284f791e9331db53c5a68aa2679f439389dd260cc6a34b8aff9bc622e96ca04d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-fw5hr" Mar 17 17:42:35.567296 kubelet[2662]: E0317 17:42:35.567263 2662 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"284f791e9331db53c5a68aa2679f439389dd260cc6a34b8aff9bc622e96ca04d\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-fw5hr" Mar 17 17:42:35.567341 kubelet[2662]: E0317 17:42:35.567305 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-fw5hr_kube-system(ae45280b-f87f-4410-991d-e0beb4bdc7c6)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-fw5hr_kube-system(ae45280b-f87f-4410-991d-e0beb4bdc7c6)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"284f791e9331db53c5a68aa2679f439389dd260cc6a34b8aff9bc622e96ca04d\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-fw5hr" podUID="ae45280b-f87f-4410-991d-e0beb4bdc7c6" Mar 17 17:42:35.575108 containerd[1467]: time="2025-03-17T17:42:35.575052249Z" level=error msg="Failed to destroy network for sandbox \"980554b391378bf7ff18652b7dc7922f5d0a4743b3074f835149be5d75b9c811\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:35.575746 containerd[1467]: time="2025-03-17T17:42:35.575712982Z" level=error msg="encountered an error cleaning up failed sandbox \"980554b391378bf7ff18652b7dc7922f5d0a4743b3074f835149be5d75b9c811\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:35.575798 containerd[1467]: time="2025-03-17T17:42:35.575770039Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7fddcd7f99-92q58,Uid:e531d9e1-a7bb-4958-adfd-8db7b3f84f95,Namespace:calico-system,Attempt:2,} failed, error" error="failed to setup network for sandbox \"980554b391378bf7ff18652b7dc7922f5d0a4743b3074f835149be5d75b9c811\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:35.576074 kubelet[2662]: E0317 17:42:35.576012 2662 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"980554b391378bf7ff18652b7dc7922f5d0a4743b3074f835149be5d75b9c811\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:35.576126 kubelet[2662]: E0317 17:42:35.576107 2662 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"980554b391378bf7ff18652b7dc7922f5d0a4743b3074f835149be5d75b9c811\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-7fddcd7f99-92q58" Mar 17 17:42:35.576166 kubelet[2662]: E0317 17:42:35.576129 2662 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"980554b391378bf7ff18652b7dc7922f5d0a4743b3074f835149be5d75b9c811\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-7fddcd7f99-92q58" Mar 17 17:42:35.576237 kubelet[2662]: E0317 17:42:35.576192 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-7fddcd7f99-92q58_calico-system(e531d9e1-a7bb-4958-adfd-8db7b3f84f95)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-7fddcd7f99-92q58_calico-system(e531d9e1-a7bb-4958-adfd-8db7b3f84f95)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"980554b391378bf7ff18652b7dc7922f5d0a4743b3074f835149be5d75b9c811\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-7fddcd7f99-92q58" podUID="e531d9e1-a7bb-4958-adfd-8db7b3f84f95" Mar 17 17:42:35.600923 containerd[1467]: time="2025-03-17T17:42:35.600866664Z" level=error msg="Failed to destroy network for sandbox \"1033b80b196d9abbe23dc46629630e49f397ddc00f3445a019c576bdd8d54f14\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:35.601282 containerd[1467]: time="2025-03-17T17:42:35.601255756Z" level=error msg="encountered an error cleaning up failed sandbox \"1033b80b196d9abbe23dc46629630e49f397ddc00f3445a019c576bdd8d54f14\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:35.601333 containerd[1467]: time="2025-03-17T17:42:35.601312754Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-j9gdg,Uid:27eb0c6c-40f4-4b19-b513-0003d63df5d0,Namespace:calico-system,Attempt:2,} failed, error" error="failed to setup network for sandbox \"1033b80b196d9abbe23dc46629630e49f397ddc00f3445a019c576bdd8d54f14\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:35.601583 kubelet[2662]: E0317 17:42:35.601539 2662 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1033b80b196d9abbe23dc46629630e49f397ddc00f3445a019c576bdd8d54f14\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:35.601646 kubelet[2662]: E0317 17:42:35.601600 2662 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1033b80b196d9abbe23dc46629630e49f397ddc00f3445a019c576bdd8d54f14\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-j9gdg" Mar 17 17:42:35.601646 kubelet[2662]: E0317 17:42:35.601623 2662 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1033b80b196d9abbe23dc46629630e49f397ddc00f3445a019c576bdd8d54f14\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-j9gdg" Mar 17 17:42:35.601705 kubelet[2662]: E0317 17:42:35.601679 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-j9gdg_calico-system(27eb0c6c-40f4-4b19-b513-0003d63df5d0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-j9gdg_calico-system(27eb0c6c-40f4-4b19-b513-0003d63df5d0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"1033b80b196d9abbe23dc46629630e49f397ddc00f3445a019c576bdd8d54f14\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-j9gdg" podUID="27eb0c6c-40f4-4b19-b513-0003d63df5d0" Mar 17 17:42:35.602233 containerd[1467]: time="2025-03-17T17:42:35.602210402Z" level=error msg="Failed to destroy network for sandbox \"4a6e2b6a3eecd39a342da3235b8bc51d6bca9e9c8c2f588a315458c197bf0390\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:35.602695 containerd[1467]: time="2025-03-17T17:42:35.602618069Z" level=error msg="encountered an error cleaning up failed sandbox \"4a6e2b6a3eecd39a342da3235b8bc51d6bca9e9c8c2f588a315458c197bf0390\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:35.602695 containerd[1467]: time="2025-03-17T17:42:35.602655369Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86c8b6f988-79bzl,Uid:3265838f-27ca-489d-8c65-7cadce5cca40,Namespace:calico-apiserver,Attempt:2,} failed, error" error="failed to setup network for sandbox \"4a6e2b6a3eecd39a342da3235b8bc51d6bca9e9c8c2f588a315458c197bf0390\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:35.602995 kubelet[2662]: E0317 17:42:35.602897 2662 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4a6e2b6a3eecd39a342da3235b8bc51d6bca9e9c8c2f588a315458c197bf0390\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:35.602995 kubelet[2662]: E0317 17:42:35.602926 2662 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4a6e2b6a3eecd39a342da3235b8bc51d6bca9e9c8c2f588a315458c197bf0390\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86c8b6f988-79bzl" Mar 17 17:42:35.602995 kubelet[2662]: E0317 17:42:35.602941 2662 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4a6e2b6a3eecd39a342da3235b8bc51d6bca9e9c8c2f588a315458c197bf0390\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86c8b6f988-79bzl" Mar 17 17:42:35.603138 kubelet[2662]: E0317 17:42:35.602970 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-86c8b6f988-79bzl_calico-apiserver(3265838f-27ca-489d-8c65-7cadce5cca40)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-86c8b6f988-79bzl_calico-apiserver(3265838f-27ca-489d-8c65-7cadce5cca40)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"4a6e2b6a3eecd39a342da3235b8bc51d6bca9e9c8c2f588a315458c197bf0390\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-86c8b6f988-79bzl" podUID="3265838f-27ca-489d-8c65-7cadce5cca40" Mar 17 17:42:35.606146 systemd[1]: run-netns-cni\x2dc976374f\x2d6356\x2d2977\x2ddfb7\x2de6262d6a1557.mount: Deactivated successfully. Mar 17 17:42:35.606262 systemd[1]: run-netns-cni\x2dab5f3fe9\x2da00a\x2df6c4\x2dc538\x2db16bcdbb0775.mount: Deactivated successfully. Mar 17 17:42:36.008534 kubelet[2662]: I0317 17:42:36.008497 2662 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4044ac7223d418a825ef4eedf876ba4769d2804921eba6925b3d0d6d4afbc33a" Mar 17 17:42:36.009485 containerd[1467]: time="2025-03-17T17:42:36.009427922Z" level=info msg="StopPodSandbox for \"4044ac7223d418a825ef4eedf876ba4769d2804921eba6925b3d0d6d4afbc33a\"" Mar 17 17:42:36.009901 containerd[1467]: time="2025-03-17T17:42:36.009766569Z" level=info msg="Ensure that sandbox 4044ac7223d418a825ef4eedf876ba4769d2804921eba6925b3d0d6d4afbc33a in task-service has been cleanup successfully" Mar 17 17:42:36.012491 containerd[1467]: time="2025-03-17T17:42:36.012459544Z" level=info msg="TearDown network for sandbox \"4044ac7223d418a825ef4eedf876ba4769d2804921eba6925b3d0d6d4afbc33a\" successfully" Mar 17 17:42:36.012491 containerd[1467]: time="2025-03-17T17:42:36.012483178Z" level=info msg="StopPodSandbox for \"4044ac7223d418a825ef4eedf876ba4769d2804921eba6925b3d0d6d4afbc33a\" returns successfully" Mar 17 17:42:36.013199 systemd[1]: run-netns-cni\x2ddbf481d6\x2d45e8\x2d0554\x2dce48\x2d553224f26f3a.mount: Deactivated successfully. Mar 17 17:42:36.013834 containerd[1467]: time="2025-03-17T17:42:36.013650844Z" level=info msg="StopPodSandbox for \"3f5345f62ce3def9878f952bc3b9ba8515b066f140a947d652235852cb08fae5\"" Mar 17 17:42:36.013834 containerd[1467]: time="2025-03-17T17:42:36.013770058Z" level=info msg="TearDown network for sandbox \"3f5345f62ce3def9878f952bc3b9ba8515b066f140a947d652235852cb08fae5\" successfully" Mar 17 17:42:36.013834 containerd[1467]: time="2025-03-17T17:42:36.013781139Z" level=info msg="StopPodSandbox for \"3f5345f62ce3def9878f952bc3b9ba8515b066f140a947d652235852cb08fae5\" returns successfully" Mar 17 17:42:36.014750 containerd[1467]: time="2025-03-17T17:42:36.014715165Z" level=info msg="StopPodSandbox for \"e93aa7d450e66ea71243d70e3ef70d32837865b756a4bd00a6f296f08f7881d8\"" Mar 17 17:42:36.014862 containerd[1467]: time="2025-03-17T17:42:36.014826726Z" level=info msg="TearDown network for sandbox \"e93aa7d450e66ea71243d70e3ef70d32837865b756a4bd00a6f296f08f7881d8\" successfully" Mar 17 17:42:36.014862 containerd[1467]: time="2025-03-17T17:42:36.014855370Z" level=info msg="StopPodSandbox for \"e93aa7d450e66ea71243d70e3ef70d32837865b756a4bd00a6f296f08f7881d8\" returns successfully" Mar 17 17:42:36.015612 containerd[1467]: time="2025-03-17T17:42:36.015588358Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86c8b6f988-ppvkt,Uid:c9eab445-c078-490b-861c-b96d204afef3,Namespace:calico-apiserver,Attempt:3,}" Mar 17 17:42:36.016114 kubelet[2662]: I0317 17:42:36.016055 2662 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3edf99fd7e4c4bcd608727ff216f57eed4c9785e1d1f4d8c623f0c2aa6ca839a" Mar 17 17:42:36.016963 containerd[1467]: time="2025-03-17T17:42:36.016593779Z" level=info msg="StopPodSandbox for \"3edf99fd7e4c4bcd608727ff216f57eed4c9785e1d1f4d8c623f0c2aa6ca839a\"" Mar 17 17:42:36.016963 containerd[1467]: time="2025-03-17T17:42:36.016811999Z" level=info msg="Ensure that sandbox 3edf99fd7e4c4bcd608727ff216f57eed4c9785e1d1f4d8c623f0c2aa6ca839a in task-service has been cleanup successfully" Mar 17 17:42:36.017532 containerd[1467]: time="2025-03-17T17:42:36.017509070Z" level=info msg="TearDown network for sandbox \"3edf99fd7e4c4bcd608727ff216f57eed4c9785e1d1f4d8c623f0c2aa6ca839a\" successfully" Mar 17 17:42:36.017751 containerd[1467]: time="2025-03-17T17:42:36.017613566Z" level=info msg="StopPodSandbox for \"3edf99fd7e4c4bcd608727ff216f57eed4c9785e1d1f4d8c623f0c2aa6ca839a\" returns successfully" Mar 17 17:42:36.018221 containerd[1467]: time="2025-03-17T17:42:36.017934600Z" level=info msg="StopPodSandbox for \"f1d105acdc9b53357e4e621942df86a7fba22a3900a30edbb033cb1322dbced2\"" Mar 17 17:42:36.018221 containerd[1467]: time="2025-03-17T17:42:36.018019950Z" level=info msg="TearDown network for sandbox \"f1d105acdc9b53357e4e621942df86a7fba22a3900a30edbb033cb1322dbced2\" successfully" Mar 17 17:42:36.018221 containerd[1467]: time="2025-03-17T17:42:36.018032384Z" level=info msg="StopPodSandbox for \"f1d105acdc9b53357e4e621942df86a7fba22a3900a30edbb033cb1322dbced2\" returns successfully" Mar 17 17:42:36.018446 containerd[1467]: time="2025-03-17T17:42:36.018415845Z" level=info msg="StopPodSandbox for \"e983d9d5aee8c472a9806708fa057e3b2de9217794b9f28b9fc2133b98942f76\"" Mar 17 17:42:36.018532 containerd[1467]: time="2025-03-17T17:42:36.018510803Z" level=info msg="TearDown network for sandbox \"e983d9d5aee8c472a9806708fa057e3b2de9217794b9f28b9fc2133b98942f76\" successfully" Mar 17 17:42:36.018592 containerd[1467]: time="2025-03-17T17:42:36.018530160Z" level=info msg="StopPodSandbox for \"e983d9d5aee8c472a9806708fa057e3b2de9217794b9f28b9fc2133b98942f76\" returns successfully" Mar 17 17:42:36.018760 kubelet[2662]: E0317 17:42:36.018741 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:36.019291 kubelet[2662]: I0317 17:42:36.018879 2662 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="284f791e9331db53c5a68aa2679f439389dd260cc6a34b8aff9bc622e96ca04d" Mar 17 17:42:36.020423 containerd[1467]: time="2025-03-17T17:42:36.019521775Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-sjrwq,Uid:228c6093-6dc2-4faf-8ec8-87849ef402ab,Namespace:kube-system,Attempt:3,}" Mar 17 17:42:36.020423 containerd[1467]: time="2025-03-17T17:42:36.019604691Z" level=info msg="StopPodSandbox for \"284f791e9331db53c5a68aa2679f439389dd260cc6a34b8aff9bc622e96ca04d\"" Mar 17 17:42:36.020154 systemd[1]: run-netns-cni\x2d80fb8bdb\x2d14d6\x2de87b\x2df73a\x2d26fcb3877837.mount: Deactivated successfully. Mar 17 17:42:36.020655 containerd[1467]: time="2025-03-17T17:42:36.020580636Z" level=info msg="Ensure that sandbox 284f791e9331db53c5a68aa2679f439389dd260cc6a34b8aff9bc622e96ca04d in task-service has been cleanup successfully" Mar 17 17:42:36.020844 containerd[1467]: time="2025-03-17T17:42:36.020772968Z" level=info msg="TearDown network for sandbox \"284f791e9331db53c5a68aa2679f439389dd260cc6a34b8aff9bc622e96ca04d\" successfully" Mar 17 17:42:36.020844 containerd[1467]: time="2025-03-17T17:42:36.020792405Z" level=info msg="StopPodSandbox for \"284f791e9331db53c5a68aa2679f439389dd260cc6a34b8aff9bc622e96ca04d\" returns successfully" Mar 17 17:42:36.021120 containerd[1467]: time="2025-03-17T17:42:36.021099552Z" level=info msg="StopPodSandbox for \"3f78c7317a0c3224f4039412ac38faee93bb589fa11754f46d7d13d15315c3d5\"" Mar 17 17:42:36.021199 containerd[1467]: time="2025-03-17T17:42:36.021178320Z" level=info msg="TearDown network for sandbox \"3f78c7317a0c3224f4039412ac38faee93bb589fa11754f46d7d13d15315c3d5\" successfully" Mar 17 17:42:36.021243 containerd[1467]: time="2025-03-17T17:42:36.021198158Z" level=info msg="StopPodSandbox for \"3f78c7317a0c3224f4039412ac38faee93bb589fa11754f46d7d13d15315c3d5\" returns successfully" Mar 17 17:42:36.021449 kubelet[2662]: I0317 17:42:36.021427 2662 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1033b80b196d9abbe23dc46629630e49f397ddc00f3445a019c576bdd8d54f14" Mar 17 17:42:36.026291 systemd[1]: run-netns-cni\x2d288bc3d2\x2d849b\x2d2162\x2d0a20\x2d7eff805ece7c.mount: Deactivated successfully. Mar 17 17:42:36.027517 containerd[1467]: time="2025-03-17T17:42:36.027480191Z" level=info msg="StopPodSandbox for \"1033b80b196d9abbe23dc46629630e49f397ddc00f3445a019c576bdd8d54f14\"" Mar 17 17:42:36.027885 containerd[1467]: time="2025-03-17T17:42:36.027722417Z" level=info msg="StopPodSandbox for \"6c94b5b9e7a3d7b63ccc9268b52bbe37ef401118cf01d372890eeea51b3ec01a\"" Mar 17 17:42:36.028056 containerd[1467]: time="2025-03-17T17:42:36.028008936Z" level=info msg="TearDown network for sandbox \"6c94b5b9e7a3d7b63ccc9268b52bbe37ef401118cf01d372890eeea51b3ec01a\" successfully" Mar 17 17:42:36.028056 containerd[1467]: time="2025-03-17T17:42:36.028025017Z" level=info msg="StopPodSandbox for \"6c94b5b9e7a3d7b63ccc9268b52bbe37ef401118cf01d372890eeea51b3ec01a\" returns successfully" Mar 17 17:42:36.028432 kubelet[2662]: E0317 17:42:36.028404 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:36.028537 containerd[1467]: time="2025-03-17T17:42:36.028501022Z" level=info msg="Ensure that sandbox 1033b80b196d9abbe23dc46629630e49f397ddc00f3445a019c576bdd8d54f14 in task-service has been cleanup successfully" Mar 17 17:42:36.028899 containerd[1467]: time="2025-03-17T17:42:36.028685268Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-fw5hr,Uid:ae45280b-f87f-4410-991d-e0beb4bdc7c6,Namespace:kube-system,Attempt:3,}" Mar 17 17:42:36.029485 containerd[1467]: time="2025-03-17T17:42:36.029418307Z" level=info msg="TearDown network for sandbox \"1033b80b196d9abbe23dc46629630e49f397ddc00f3445a019c576bdd8d54f14\" successfully" Mar 17 17:42:36.029485 containerd[1467]: time="2025-03-17T17:42:36.029436360Z" level=info msg="StopPodSandbox for \"1033b80b196d9abbe23dc46629630e49f397ddc00f3445a019c576bdd8d54f14\" returns successfully" Mar 17 17:42:36.031344 kubelet[2662]: I0317 17:42:36.031314 2662 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="980554b391378bf7ff18652b7dc7922f5d0a4743b3074f835149be5d75b9c811" Mar 17 17:42:36.032316 containerd[1467]: time="2025-03-17T17:42:36.031896216Z" level=info msg="StopPodSandbox for \"980554b391378bf7ff18652b7dc7922f5d0a4743b3074f835149be5d75b9c811\"" Mar 17 17:42:36.032316 containerd[1467]: time="2025-03-17T17:42:36.031984934Z" level=info msg="StopPodSandbox for \"47b88b63e50185559458c168aee26da05df22f8d62d10c49042400bc79ba64b6\"" Mar 17 17:42:36.032316 containerd[1467]: time="2025-03-17T17:42:36.032055466Z" level=info msg="TearDown network for sandbox \"47b88b63e50185559458c168aee26da05df22f8d62d10c49042400bc79ba64b6\" successfully" Mar 17 17:42:36.032316 containerd[1467]: time="2025-03-17T17:42:36.032064713Z" level=info msg="StopPodSandbox for \"47b88b63e50185559458c168aee26da05df22f8d62d10c49042400bc79ba64b6\" returns successfully" Mar 17 17:42:36.032316 containerd[1467]: time="2025-03-17T17:42:36.032140996Z" level=info msg="Ensure that sandbox 980554b391378bf7ff18652b7dc7922f5d0a4743b3074f835149be5d75b9c811 in task-service has been cleanup successfully" Mar 17 17:42:36.032840 containerd[1467]: time="2025-03-17T17:42:36.032761153Z" level=info msg="TearDown network for sandbox \"980554b391378bf7ff18652b7dc7922f5d0a4743b3074f835149be5d75b9c811\" successfully" Mar 17 17:42:36.032840 containerd[1467]: time="2025-03-17T17:42:36.032783916Z" level=info msg="StopPodSandbox for \"980554b391378bf7ff18652b7dc7922f5d0a4743b3074f835149be5d75b9c811\" returns successfully" Mar 17 17:42:36.033091 containerd[1467]: time="2025-03-17T17:42:36.033066157Z" level=info msg="StopPodSandbox for \"317763565ecd927b3fe20fd547beaf2029d8ad181454207c47710163dbd09a88\"" Mar 17 17:42:36.033244 containerd[1467]: time="2025-03-17T17:42:36.033112694Z" level=info msg="StopPodSandbox for \"b33f35d6e3bd6b1a5f1d8cb21ebceacfc304ffe6c1b7ed3aa40b78efa59e6a03\"" Mar 17 17:42:36.033327 containerd[1467]: time="2025-03-17T17:42:36.033309685Z" level=info msg="TearDown network for sandbox \"317763565ecd927b3fe20fd547beaf2029d8ad181454207c47710163dbd09a88\" successfully" Mar 17 17:42:36.033556 containerd[1467]: time="2025-03-17T17:42:36.033431974Z" level=info msg="StopPodSandbox for \"317763565ecd927b3fe20fd547beaf2029d8ad181454207c47710163dbd09a88\" returns successfully" Mar 17 17:42:36.033556 containerd[1467]: time="2025-03-17T17:42:36.033309755Z" level=info msg="TearDown network for sandbox \"b33f35d6e3bd6b1a5f1d8cb21ebceacfc304ffe6c1b7ed3aa40b78efa59e6a03\" successfully" Mar 17 17:42:36.033556 containerd[1467]: time="2025-03-17T17:42:36.033522425Z" level=info msg="StopPodSandbox for \"b33f35d6e3bd6b1a5f1d8cb21ebceacfc304ffe6c1b7ed3aa40b78efa59e6a03\" returns successfully" Mar 17 17:42:36.034543 containerd[1467]: time="2025-03-17T17:42:36.034518949Z" level=info msg="StopPodSandbox for \"cc57dbc98644c56867a38eb1c7016e6e03a565b54b9bea52d164d9463bd478d9\"" Mar 17 17:42:36.034811 containerd[1467]: time="2025-03-17T17:42:36.034616453Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-j9gdg,Uid:27eb0c6c-40f4-4b19-b513-0003d63df5d0,Namespace:calico-system,Attempt:3,}" Mar 17 17:42:36.034811 containerd[1467]: time="2025-03-17T17:42:36.034746867Z" level=info msg="TearDown network for sandbox \"cc57dbc98644c56867a38eb1c7016e6e03a565b54b9bea52d164d9463bd478d9\" successfully" Mar 17 17:42:36.034811 containerd[1467]: time="2025-03-17T17:42:36.034759210Z" level=info msg="StopPodSandbox for \"cc57dbc98644c56867a38eb1c7016e6e03a565b54b9bea52d164d9463bd478d9\" returns successfully" Mar 17 17:42:36.036578 containerd[1467]: time="2025-03-17T17:42:36.036553595Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7fddcd7f99-92q58,Uid:e531d9e1-a7bb-4958-adfd-8db7b3f84f95,Namespace:calico-system,Attempt:3,}" Mar 17 17:42:36.036722 kubelet[2662]: I0317 17:42:36.036695 2662 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a6e2b6a3eecd39a342da3235b8bc51d6bca9e9c8c2f588a315458c197bf0390" Mar 17 17:42:36.037069 containerd[1467]: time="2025-03-17T17:42:36.037046913Z" level=info msg="StopPodSandbox for \"4a6e2b6a3eecd39a342da3235b8bc51d6bca9e9c8c2f588a315458c197bf0390\"" Mar 17 17:42:36.037249 containerd[1467]: time="2025-03-17T17:42:36.037227924Z" level=info msg="Ensure that sandbox 4a6e2b6a3eecd39a342da3235b8bc51d6bca9e9c8c2f588a315458c197bf0390 in task-service has been cleanup successfully" Mar 17 17:42:36.037928 containerd[1467]: time="2025-03-17T17:42:36.037838311Z" level=info msg="TearDown network for sandbox \"4a6e2b6a3eecd39a342da3235b8bc51d6bca9e9c8c2f588a315458c197bf0390\" successfully" Mar 17 17:42:36.037928 containerd[1467]: time="2025-03-17T17:42:36.037866844Z" level=info msg="StopPodSandbox for \"4a6e2b6a3eecd39a342da3235b8bc51d6bca9e9c8c2f588a315458c197bf0390\" returns successfully" Mar 17 17:42:36.038148 containerd[1467]: time="2025-03-17T17:42:36.038126413Z" level=info msg="StopPodSandbox for \"f905ddf13c6f6a970be32393ab7c951bb1766cec8d75bdc0aca8519fcc2783c7\"" Mar 17 17:42:36.038228 containerd[1467]: time="2025-03-17T17:42:36.038210010Z" level=info msg="TearDown network for sandbox \"f905ddf13c6f6a970be32393ab7c951bb1766cec8d75bdc0aca8519fcc2783c7\" successfully" Mar 17 17:42:36.038228 containerd[1467]: time="2025-03-17T17:42:36.038223596Z" level=info msg="StopPodSandbox for \"f905ddf13c6f6a970be32393ab7c951bb1766cec8d75bdc0aca8519fcc2783c7\" returns successfully" Mar 17 17:42:36.038812 containerd[1467]: time="2025-03-17T17:42:36.038793258Z" level=info msg="StopPodSandbox for \"a45255c5e842412af83dd06ba8b258ef311550cba3da5240506cd00cf0678895\"" Mar 17 17:42:36.039181 containerd[1467]: time="2025-03-17T17:42:36.039158694Z" level=info msg="TearDown network for sandbox \"a45255c5e842412af83dd06ba8b258ef311550cba3da5240506cd00cf0678895\" successfully" Mar 17 17:42:36.039181 containerd[1467]: time="2025-03-17T17:42:36.039171518Z" level=info msg="StopPodSandbox for \"a45255c5e842412af83dd06ba8b258ef311550cba3da5240506cd00cf0678895\" returns successfully" Mar 17 17:42:36.039944 containerd[1467]: time="2025-03-17T17:42:36.039919395Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86c8b6f988-79bzl,Uid:3265838f-27ca-489d-8c65-7cadce5cca40,Namespace:calico-apiserver,Attempt:3,}" Mar 17 17:42:36.263475 kubelet[2662]: I0317 17:42:36.262159 2662 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 17 17:42:36.263475 kubelet[2662]: E0317 17:42:36.263066 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:36.313790 containerd[1467]: time="2025-03-17T17:42:36.313739927Z" level=error msg="Failed to destroy network for sandbox \"008cee49813c10fd33682eb193a75ecd740653b4fd1a4a322bfcc088c6078101\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:36.314150 containerd[1467]: time="2025-03-17T17:42:36.314123978Z" level=error msg="encountered an error cleaning up failed sandbox \"008cee49813c10fd33682eb193a75ecd740653b4fd1a4a322bfcc088c6078101\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:36.314209 containerd[1467]: time="2025-03-17T17:42:36.314191977Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-sjrwq,Uid:228c6093-6dc2-4faf-8ec8-87849ef402ab,Namespace:kube-system,Attempt:3,} failed, error" error="failed to setup network for sandbox \"008cee49813c10fd33682eb193a75ecd740653b4fd1a4a322bfcc088c6078101\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:36.314440 kubelet[2662]: E0317 17:42:36.314400 2662 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"008cee49813c10fd33682eb193a75ecd740653b4fd1a4a322bfcc088c6078101\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:36.314504 kubelet[2662]: E0317 17:42:36.314457 2662 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"008cee49813c10fd33682eb193a75ecd740653b4fd1a4a322bfcc088c6078101\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-sjrwq" Mar 17 17:42:36.314504 kubelet[2662]: E0317 17:42:36.314477 2662 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"008cee49813c10fd33682eb193a75ecd740653b4fd1a4a322bfcc088c6078101\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-sjrwq" Mar 17 17:42:36.314561 kubelet[2662]: E0317 17:42:36.314521 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-sjrwq_kube-system(228c6093-6dc2-4faf-8ec8-87849ef402ab)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-sjrwq_kube-system(228c6093-6dc2-4faf-8ec8-87849ef402ab)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"008cee49813c10fd33682eb193a75ecd740653b4fd1a4a322bfcc088c6078101\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-sjrwq" podUID="228c6093-6dc2-4faf-8ec8-87849ef402ab" Mar 17 17:42:36.388219 containerd[1467]: time="2025-03-17T17:42:36.388145524Z" level=error msg="Failed to destroy network for sandbox \"eec0898e435269a73cab252710c67519bb65ec8c70880a65c0d84177e2d5547f\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:36.388595 containerd[1467]: time="2025-03-17T17:42:36.388563390Z" level=error msg="encountered an error cleaning up failed sandbox \"eec0898e435269a73cab252710c67519bb65ec8c70880a65c0d84177e2d5547f\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:36.388655 containerd[1467]: time="2025-03-17T17:42:36.388628273Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86c8b6f988-ppvkt,Uid:c9eab445-c078-490b-861c-b96d204afef3,Namespace:calico-apiserver,Attempt:3,} failed, error" error="failed to setup network for sandbox \"eec0898e435269a73cab252710c67519bb65ec8c70880a65c0d84177e2d5547f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:36.388891 kubelet[2662]: E0317 17:42:36.388855 2662 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"eec0898e435269a73cab252710c67519bb65ec8c70880a65c0d84177e2d5547f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:36.388936 kubelet[2662]: E0317 17:42:36.388915 2662 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"eec0898e435269a73cab252710c67519bb65ec8c70880a65c0d84177e2d5547f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86c8b6f988-ppvkt" Mar 17 17:42:36.388967 kubelet[2662]: E0317 17:42:36.388934 2662 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"eec0898e435269a73cab252710c67519bb65ec8c70880a65c0d84177e2d5547f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86c8b6f988-ppvkt" Mar 17 17:42:36.389006 kubelet[2662]: E0317 17:42:36.388984 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-86c8b6f988-ppvkt_calico-apiserver(c9eab445-c078-490b-861c-b96d204afef3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-86c8b6f988-ppvkt_calico-apiserver(c9eab445-c078-490b-861c-b96d204afef3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"eec0898e435269a73cab252710c67519bb65ec8c70880a65c0d84177e2d5547f\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-86c8b6f988-ppvkt" podUID="c9eab445-c078-490b-861c-b96d204afef3" Mar 17 17:42:36.609327 systemd[1]: run-netns-cni\x2d26dd0cca\x2d18f5\x2d584a\x2d771e\x2d1484a5e1a4fc.mount: Deactivated successfully. Mar 17 17:42:36.609477 systemd[1]: run-netns-cni\x2d68872ed2\x2da51e\x2d8367\x2d55f7\x2d4030b035552a.mount: Deactivated successfully. Mar 17 17:42:36.609562 systemd[1]: run-netns-cni\x2d5928724b\x2dc2bc\x2d57d2\x2d27b1\x2dcb4f98cc3241.mount: Deactivated successfully. Mar 17 17:42:36.946952 systemd[1]: Started sshd@10-10.0.0.61:22-10.0.0.1:51298.service - OpenSSH per-connection server daemon (10.0.0.1:51298). Mar 17 17:42:37.014411 sshd[4222]: Accepted publickey for core from 10.0.0.1 port 51298 ssh2: RSA SHA256:j201F9FRK1q3ChnxQf0adNdYppDp+g37vmaXPvsVhek Mar 17 17:42:37.015819 sshd-session[4222]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 17 17:42:37.022317 systemd-logind[1453]: New session 11 of user core. Mar 17 17:42:37.034478 systemd[1]: Started session-11.scope - Session 11 of User core. Mar 17 17:42:37.039878 kubelet[2662]: I0317 17:42:37.039844 2662 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eec0898e435269a73cab252710c67519bb65ec8c70880a65c0d84177e2d5547f" Mar 17 17:42:37.040397 containerd[1467]: time="2025-03-17T17:42:37.040316086Z" level=info msg="StopPodSandbox for \"eec0898e435269a73cab252710c67519bb65ec8c70880a65c0d84177e2d5547f\"" Mar 17 17:42:37.040708 containerd[1467]: time="2025-03-17T17:42:37.040684009Z" level=info msg="Ensure that sandbox eec0898e435269a73cab252710c67519bb65ec8c70880a65c0d84177e2d5547f in task-service has been cleanup successfully" Mar 17 17:42:37.041482 containerd[1467]: time="2025-03-17T17:42:37.041456701Z" level=info msg="TearDown network for sandbox \"eec0898e435269a73cab252710c67519bb65ec8c70880a65c0d84177e2d5547f\" successfully" Mar 17 17:42:37.041482 containerd[1467]: time="2025-03-17T17:42:37.041476919Z" level=info msg="StopPodSandbox for \"eec0898e435269a73cab252710c67519bb65ec8c70880a65c0d84177e2d5547f\" returns successfully" Mar 17 17:42:37.043181 kubelet[2662]: I0317 17:42:37.042771 2662 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="008cee49813c10fd33682eb193a75ecd740653b4fd1a4a322bfcc088c6078101" Mar 17 17:42:37.043181 kubelet[2662]: E0317 17:42:37.043090 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:37.043342 containerd[1467]: time="2025-03-17T17:42:37.041850251Z" level=info msg="StopPodSandbox for \"4044ac7223d418a825ef4eedf876ba4769d2804921eba6925b3d0d6d4afbc33a\"" Mar 17 17:42:37.043342 containerd[1467]: time="2025-03-17T17:42:37.043294016Z" level=info msg="TearDown network for sandbox \"4044ac7223d418a825ef4eedf876ba4769d2804921eba6925b3d0d6d4afbc33a\" successfully" Mar 17 17:42:37.043342 containerd[1467]: time="2025-03-17T17:42:37.043305878Z" level=info msg="StopPodSandbox for \"4044ac7223d418a825ef4eedf876ba4769d2804921eba6925b3d0d6d4afbc33a\" returns successfully" Mar 17 17:42:37.043455 containerd[1467]: time="2025-03-17T17:42:37.043413099Z" level=info msg="StopPodSandbox for \"008cee49813c10fd33682eb193a75ecd740653b4fd1a4a322bfcc088c6078101\"" Mar 17 17:42:37.043500 systemd[1]: run-netns-cni\x2d47b327d5\x2d35a5\x2de04d\x2dfe2a\x2dd7b6b743302f.mount: Deactivated successfully. Mar 17 17:42:37.043602 containerd[1467]: time="2025-03-17T17:42:37.043569564Z" level=info msg="Ensure that sandbox 008cee49813c10fd33682eb193a75ecd740653b4fd1a4a322bfcc088c6078101 in task-service has been cleanup successfully" Mar 17 17:42:37.043666 containerd[1467]: time="2025-03-17T17:42:37.043640176Z" level=info msg="StopPodSandbox for \"3f5345f62ce3def9878f952bc3b9ba8515b066f140a947d652235852cb08fae5\"" Mar 17 17:42:37.043756 containerd[1467]: time="2025-03-17T17:42:37.043737771Z" level=info msg="TearDown network for sandbox \"008cee49813c10fd33682eb193a75ecd740653b4fd1a4a322bfcc088c6078101\" successfully" Mar 17 17:42:37.043756 containerd[1467]: time="2025-03-17T17:42:37.043751857Z" level=info msg="StopPodSandbox for \"008cee49813c10fd33682eb193a75ecd740653b4fd1a4a322bfcc088c6078101\" returns successfully" Mar 17 17:42:37.043839 containerd[1467]: time="2025-03-17T17:42:37.043800859Z" level=info msg="TearDown network for sandbox \"3f5345f62ce3def9878f952bc3b9ba8515b066f140a947d652235852cb08fae5\" successfully" Mar 17 17:42:37.043867 containerd[1467]: time="2025-03-17T17:42:37.043839371Z" level=info msg="StopPodSandbox for \"3f5345f62ce3def9878f952bc3b9ba8515b066f140a947d652235852cb08fae5\" returns successfully" Mar 17 17:42:37.044027 containerd[1467]: time="2025-03-17T17:42:37.044006426Z" level=info msg="StopPodSandbox for \"e93aa7d450e66ea71243d70e3ef70d32837865b756a4bd00a6f296f08f7881d8\"" Mar 17 17:42:37.044121 containerd[1467]: time="2025-03-17T17:42:37.044080284Z" level=info msg="TearDown network for sandbox \"e93aa7d450e66ea71243d70e3ef70d32837865b756a4bd00a6f296f08f7881d8\" successfully" Mar 17 17:42:37.044121 containerd[1467]: time="2025-03-17T17:42:37.044116662Z" level=info msg="StopPodSandbox for \"e93aa7d450e66ea71243d70e3ef70d32837865b756a4bd00a6f296f08f7881d8\" returns successfully" Mar 17 17:42:37.044184 containerd[1467]: time="2025-03-17T17:42:37.044154825Z" level=info msg="StopPodSandbox for \"3edf99fd7e4c4bcd608727ff216f57eed4c9785e1d1f4d8c623f0c2aa6ca839a\"" Mar 17 17:42:37.044669 containerd[1467]: time="2025-03-17T17:42:37.044332017Z" level=info msg="TearDown network for sandbox \"3edf99fd7e4c4bcd608727ff216f57eed4c9785e1d1f4d8c623f0c2aa6ca839a\" successfully" Mar 17 17:42:37.044669 containerd[1467]: time="2025-03-17T17:42:37.044346695Z" level=info msg="StopPodSandbox for \"3edf99fd7e4c4bcd608727ff216f57eed4c9785e1d1f4d8c623f0c2aa6ca839a\" returns successfully" Mar 17 17:42:37.044669 containerd[1467]: time="2025-03-17T17:42:37.044525381Z" level=info msg="StopPodSandbox for \"f1d105acdc9b53357e4e621942df86a7fba22a3900a30edbb033cb1322dbced2\"" Mar 17 17:42:37.044669 containerd[1467]: time="2025-03-17T17:42:37.044599550Z" level=info msg="TearDown network for sandbox \"f1d105acdc9b53357e4e621942df86a7fba22a3900a30edbb033cb1322dbced2\" successfully" Mar 17 17:42:37.044669 containerd[1467]: time="2025-03-17T17:42:37.044608468Z" level=info msg="StopPodSandbox for \"f1d105acdc9b53357e4e621942df86a7fba22a3900a30edbb033cb1322dbced2\" returns successfully" Mar 17 17:42:37.044918 containerd[1467]: time="2025-03-17T17:42:37.044893814Z" level=info msg="StopPodSandbox for \"e983d9d5aee8c472a9806708fa057e3b2de9217794b9f28b9fc2133b98942f76\"" Mar 17 17:42:37.044967 containerd[1467]: time="2025-03-17T17:42:37.044946964Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86c8b6f988-ppvkt,Uid:c9eab445-c078-490b-861c-b96d204afef3,Namespace:calico-apiserver,Attempt:4,}" Mar 17 17:42:37.045009 containerd[1467]: time="2025-03-17T17:42:37.044970057Z" level=info msg="TearDown network for sandbox \"e983d9d5aee8c472a9806708fa057e3b2de9217794b9f28b9fc2133b98942f76\" successfully" Mar 17 17:42:37.045009 containerd[1467]: time="2025-03-17T17:42:37.044979355Z" level=info msg="StopPodSandbox for \"e983d9d5aee8c472a9806708fa057e3b2de9217794b9f28b9fc2133b98942f76\" returns successfully" Mar 17 17:42:37.045117 kubelet[2662]: E0317 17:42:37.045099 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:37.045317 containerd[1467]: time="2025-03-17T17:42:37.045295279Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-sjrwq,Uid:228c6093-6dc2-4faf-8ec8-87849ef402ab,Namespace:kube-system,Attempt:4,}" Mar 17 17:42:37.047506 systemd[1]: run-netns-cni\x2dc46c1e7a\x2d0a3a\x2ddd6f\x2d0d08\x2d6ea0e2f5a05a.mount: Deactivated successfully. Mar 17 17:42:37.140118 sshd[4224]: Connection closed by 10.0.0.1 port 51298 Mar 17 17:42:37.140550 sshd-session[4222]: pam_unix(sshd:session): session closed for user core Mar 17 17:42:37.144163 systemd[1]: sshd@10-10.0.0.61:22-10.0.0.1:51298.service: Deactivated successfully. Mar 17 17:42:37.145987 systemd[1]: session-11.scope: Deactivated successfully. Mar 17 17:42:37.146706 systemd-logind[1453]: Session 11 logged out. Waiting for processes to exit. Mar 17 17:42:37.147601 systemd-logind[1453]: Removed session 11. Mar 17 17:42:37.350016 containerd[1467]: time="2025-03-17T17:42:37.349885640Z" level=error msg="Failed to destroy network for sandbox \"9eba50d4b991c21e718945f4ae399cbc0ccc0b0ef2c96bdf7d5d5e8361566abd\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:37.350599 containerd[1467]: time="2025-03-17T17:42:37.350571380Z" level=error msg="Failed to destroy network for sandbox \"10ef4549d0b409f4064de73e7cb3238ecdb64a4888527c02814ca83e190159e7\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:37.352029 containerd[1467]: time="2025-03-17T17:42:37.351730880Z" level=error msg="encountered an error cleaning up failed sandbox \"10ef4549d0b409f4064de73e7cb3238ecdb64a4888527c02814ca83e190159e7\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:37.352029 containerd[1467]: time="2025-03-17T17:42:37.351803797Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-j9gdg,Uid:27eb0c6c-40f4-4b19-b513-0003d63df5d0,Namespace:calico-system,Attempt:3,} failed, error" error="failed to setup network for sandbox \"10ef4549d0b409f4064de73e7cb3238ecdb64a4888527c02814ca83e190159e7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:37.352108 kubelet[2662]: E0317 17:42:37.352041 2662 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"10ef4549d0b409f4064de73e7cb3238ecdb64a4888527c02814ca83e190159e7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:37.352156 kubelet[2662]: E0317 17:42:37.352102 2662 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"10ef4549d0b409f4064de73e7cb3238ecdb64a4888527c02814ca83e190159e7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-j9gdg" Mar 17 17:42:37.352156 kubelet[2662]: E0317 17:42:37.352124 2662 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"10ef4549d0b409f4064de73e7cb3238ecdb64a4888527c02814ca83e190159e7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-j9gdg" Mar 17 17:42:37.352216 kubelet[2662]: E0317 17:42:37.352173 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-j9gdg_calico-system(27eb0c6c-40f4-4b19-b513-0003d63df5d0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-j9gdg_calico-system(27eb0c6c-40f4-4b19-b513-0003d63df5d0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"10ef4549d0b409f4064de73e7cb3238ecdb64a4888527c02814ca83e190159e7\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-j9gdg" podUID="27eb0c6c-40f4-4b19-b513-0003d63df5d0" Mar 17 17:42:37.360510 containerd[1467]: time="2025-03-17T17:42:37.360312496Z" level=error msg="Failed to destroy network for sandbox \"b71e2358cd15f8a4cec878737cb78ea057ba63b0bc401daa2c0c7f9f8d7fef30\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:37.360922 containerd[1467]: time="2025-03-17T17:42:37.360900011Z" level=error msg="encountered an error cleaning up failed sandbox \"b71e2358cd15f8a4cec878737cb78ea057ba63b0bc401daa2c0c7f9f8d7fef30\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:37.361224 containerd[1467]: time="2025-03-17T17:42:37.361205625Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7fddcd7f99-92q58,Uid:e531d9e1-a7bb-4958-adfd-8db7b3f84f95,Namespace:calico-system,Attempt:3,} failed, error" error="failed to setup network for sandbox \"b71e2358cd15f8a4cec878737cb78ea057ba63b0bc401daa2c0c7f9f8d7fef30\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:37.361599 kubelet[2662]: E0317 17:42:37.361557 2662 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b71e2358cd15f8a4cec878737cb78ea057ba63b0bc401daa2c0c7f9f8d7fef30\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:37.361650 kubelet[2662]: E0317 17:42:37.361621 2662 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b71e2358cd15f8a4cec878737cb78ea057ba63b0bc401daa2c0c7f9f8d7fef30\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-7fddcd7f99-92q58" Mar 17 17:42:37.361650 kubelet[2662]: E0317 17:42:37.361639 2662 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b71e2358cd15f8a4cec878737cb78ea057ba63b0bc401daa2c0c7f9f8d7fef30\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-7fddcd7f99-92q58" Mar 17 17:42:37.361702 kubelet[2662]: E0317 17:42:37.361679 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-7fddcd7f99-92q58_calico-system(e531d9e1-a7bb-4958-adfd-8db7b3f84f95)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-7fddcd7f99-92q58_calico-system(e531d9e1-a7bb-4958-adfd-8db7b3f84f95)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"b71e2358cd15f8a4cec878737cb78ea057ba63b0bc401daa2c0c7f9f8d7fef30\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-7fddcd7f99-92q58" podUID="e531d9e1-a7bb-4958-adfd-8db7b3f84f95" Mar 17 17:42:37.377458 containerd[1467]: time="2025-03-17T17:42:37.377376742Z" level=error msg="encountered an error cleaning up failed sandbox \"9eba50d4b991c21e718945f4ae399cbc0ccc0b0ef2c96bdf7d5d5e8361566abd\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:37.377458 containerd[1467]: time="2025-03-17T17:42:37.377448969Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-fw5hr,Uid:ae45280b-f87f-4410-991d-e0beb4bdc7c6,Namespace:kube-system,Attempt:3,} failed, error" error="failed to setup network for sandbox \"9eba50d4b991c21e718945f4ae399cbc0ccc0b0ef2c96bdf7d5d5e8361566abd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:37.377710 kubelet[2662]: E0317 17:42:37.377650 2662 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9eba50d4b991c21e718945f4ae399cbc0ccc0b0ef2c96bdf7d5d5e8361566abd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:37.377757 kubelet[2662]: E0317 17:42:37.377707 2662 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9eba50d4b991c21e718945f4ae399cbc0ccc0b0ef2c96bdf7d5d5e8361566abd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-fw5hr" Mar 17 17:42:37.377757 kubelet[2662]: E0317 17:42:37.377725 2662 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9eba50d4b991c21e718945f4ae399cbc0ccc0b0ef2c96bdf7d5d5e8361566abd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-fw5hr" Mar 17 17:42:37.377813 kubelet[2662]: E0317 17:42:37.377764 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-fw5hr_kube-system(ae45280b-f87f-4410-991d-e0beb4bdc7c6)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-fw5hr_kube-system(ae45280b-f87f-4410-991d-e0beb4bdc7c6)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"9eba50d4b991c21e718945f4ae399cbc0ccc0b0ef2c96bdf7d5d5e8361566abd\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-fw5hr" podUID="ae45280b-f87f-4410-991d-e0beb4bdc7c6" Mar 17 17:42:37.392969 containerd[1467]: time="2025-03-17T17:42:37.392897256Z" level=error msg="Failed to destroy network for sandbox \"27220f5bf5e3f9e3a1064ebd76371f3363a6b06ed379f42a7bd72b4e7c55d55a\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:37.394265 containerd[1467]: time="2025-03-17T17:42:37.394240973Z" level=error msg="encountered an error cleaning up failed sandbox \"27220f5bf5e3f9e3a1064ebd76371f3363a6b06ed379f42a7bd72b4e7c55d55a\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:37.394460 containerd[1467]: time="2025-03-17T17:42:37.394429487Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86c8b6f988-79bzl,Uid:3265838f-27ca-489d-8c65-7cadce5cca40,Namespace:calico-apiserver,Attempt:3,} failed, error" error="failed to setup network for sandbox \"27220f5bf5e3f9e3a1064ebd76371f3363a6b06ed379f42a7bd72b4e7c55d55a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:37.395496 kubelet[2662]: E0317 17:42:37.395233 2662 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"27220f5bf5e3f9e3a1064ebd76371f3363a6b06ed379f42a7bd72b4e7c55d55a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:37.395496 kubelet[2662]: E0317 17:42:37.395303 2662 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"27220f5bf5e3f9e3a1064ebd76371f3363a6b06ed379f42a7bd72b4e7c55d55a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86c8b6f988-79bzl" Mar 17 17:42:37.395496 kubelet[2662]: E0317 17:42:37.395321 2662 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"27220f5bf5e3f9e3a1064ebd76371f3363a6b06ed379f42a7bd72b4e7c55d55a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86c8b6f988-79bzl" Mar 17 17:42:37.395645 kubelet[2662]: E0317 17:42:37.395384 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-86c8b6f988-79bzl_calico-apiserver(3265838f-27ca-489d-8c65-7cadce5cca40)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-86c8b6f988-79bzl_calico-apiserver(3265838f-27ca-489d-8c65-7cadce5cca40)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"27220f5bf5e3f9e3a1064ebd76371f3363a6b06ed379f42a7bd72b4e7c55d55a\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-86c8b6f988-79bzl" podUID="3265838f-27ca-489d-8c65-7cadce5cca40" Mar 17 17:42:37.408627 containerd[1467]: time="2025-03-17T17:42:37.408569926Z" level=error msg="Failed to destroy network for sandbox \"1f73f191946516c6303f3e08ba9764f9ef1ed363435b102ed8343391428b48bc\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:37.409994 containerd[1467]: time="2025-03-17T17:42:37.409938730Z" level=error msg="encountered an error cleaning up failed sandbox \"1f73f191946516c6303f3e08ba9764f9ef1ed363435b102ed8343391428b48bc\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:37.409994 containerd[1467]: time="2025-03-17T17:42:37.409999895Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86c8b6f988-ppvkt,Uid:c9eab445-c078-490b-861c-b96d204afef3,Namespace:calico-apiserver,Attempt:4,} failed, error" error="failed to setup network for sandbox \"1f73f191946516c6303f3e08ba9764f9ef1ed363435b102ed8343391428b48bc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:37.410347 kubelet[2662]: E0317 17:42:37.410299 2662 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1f73f191946516c6303f3e08ba9764f9ef1ed363435b102ed8343391428b48bc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:37.410421 kubelet[2662]: E0317 17:42:37.410395 2662 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1f73f191946516c6303f3e08ba9764f9ef1ed363435b102ed8343391428b48bc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86c8b6f988-ppvkt" Mar 17 17:42:37.410421 kubelet[2662]: E0317 17:42:37.410416 2662 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1f73f191946516c6303f3e08ba9764f9ef1ed363435b102ed8343391428b48bc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86c8b6f988-ppvkt" Mar 17 17:42:37.410506 kubelet[2662]: E0317 17:42:37.410462 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-86c8b6f988-ppvkt_calico-apiserver(c9eab445-c078-490b-861c-b96d204afef3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-86c8b6f988-ppvkt_calico-apiserver(c9eab445-c078-490b-861c-b96d204afef3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"1f73f191946516c6303f3e08ba9764f9ef1ed363435b102ed8343391428b48bc\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-86c8b6f988-ppvkt" podUID="c9eab445-c078-490b-861c-b96d204afef3" Mar 17 17:42:37.424672 containerd[1467]: time="2025-03-17T17:42:37.424625686Z" level=error msg="Failed to destroy network for sandbox \"6dd83b0760f3d749a5709564d2613214f6b299b039d6f1b56dcb51a9cb223ea7\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:37.425173 containerd[1467]: time="2025-03-17T17:42:37.425138681Z" level=error msg="encountered an error cleaning up failed sandbox \"6dd83b0760f3d749a5709564d2613214f6b299b039d6f1b56dcb51a9cb223ea7\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:37.425222 containerd[1467]: time="2025-03-17T17:42:37.425209634Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-sjrwq,Uid:228c6093-6dc2-4faf-8ec8-87849ef402ab,Namespace:kube-system,Attempt:4,} failed, error" error="failed to setup network for sandbox \"6dd83b0760f3d749a5709564d2613214f6b299b039d6f1b56dcb51a9cb223ea7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:37.425428 kubelet[2662]: E0317 17:42:37.425391 2662 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6dd83b0760f3d749a5709564d2613214f6b299b039d6f1b56dcb51a9cb223ea7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:37.425475 kubelet[2662]: E0317 17:42:37.425446 2662 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6dd83b0760f3d749a5709564d2613214f6b299b039d6f1b56dcb51a9cb223ea7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-sjrwq" Mar 17 17:42:37.425475 kubelet[2662]: E0317 17:42:37.425463 2662 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6dd83b0760f3d749a5709564d2613214f6b299b039d6f1b56dcb51a9cb223ea7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-sjrwq" Mar 17 17:42:37.425526 kubelet[2662]: E0317 17:42:37.425505 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-sjrwq_kube-system(228c6093-6dc2-4faf-8ec8-87849ef402ab)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-sjrwq_kube-system(228c6093-6dc2-4faf-8ec8-87849ef402ab)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"6dd83b0760f3d749a5709564d2613214f6b299b039d6f1b56dcb51a9cb223ea7\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-sjrwq" podUID="228c6093-6dc2-4faf-8ec8-87849ef402ab" Mar 17 17:42:37.608562 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-27220f5bf5e3f9e3a1064ebd76371f3363a6b06ed379f42a7bd72b4e7c55d55a-shm.mount: Deactivated successfully. Mar 17 17:42:37.608669 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-b71e2358cd15f8a4cec878737cb78ea057ba63b0bc401daa2c0c7f9f8d7fef30-shm.mount: Deactivated successfully. Mar 17 17:42:37.608757 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-10ef4549d0b409f4064de73e7cb3238ecdb64a4888527c02814ca83e190159e7-shm.mount: Deactivated successfully. Mar 17 17:42:37.608844 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-9eba50d4b991c21e718945f4ae399cbc0ccc0b0ef2c96bdf7d5d5e8361566abd-shm.mount: Deactivated successfully. Mar 17 17:42:38.046519 kubelet[2662]: I0317 17:42:38.046484 2662 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b71e2358cd15f8a4cec878737cb78ea057ba63b0bc401daa2c0c7f9f8d7fef30" Mar 17 17:42:38.047269 containerd[1467]: time="2025-03-17T17:42:38.047167309Z" level=info msg="StopPodSandbox for \"b71e2358cd15f8a4cec878737cb78ea057ba63b0bc401daa2c0c7f9f8d7fef30\"" Mar 17 17:42:38.047601 containerd[1467]: time="2025-03-17T17:42:38.047386601Z" level=info msg="Ensure that sandbox b71e2358cd15f8a4cec878737cb78ea057ba63b0bc401daa2c0c7f9f8d7fef30 in task-service has been cleanup successfully" Mar 17 17:42:38.047732 containerd[1467]: time="2025-03-17T17:42:38.047716351Z" level=info msg="TearDown network for sandbox \"b71e2358cd15f8a4cec878737cb78ea057ba63b0bc401daa2c0c7f9f8d7fef30\" successfully" Mar 17 17:42:38.047766 containerd[1467]: time="2025-03-17T17:42:38.047730989Z" level=info msg="StopPodSandbox for \"b71e2358cd15f8a4cec878737cb78ea057ba63b0bc401daa2c0c7f9f8d7fef30\" returns successfully" Mar 17 17:42:38.049840 containerd[1467]: time="2025-03-17T17:42:38.049240798Z" level=info msg="StopPodSandbox for \"980554b391378bf7ff18652b7dc7922f5d0a4743b3074f835149be5d75b9c811\"" Mar 17 17:42:38.049840 containerd[1467]: time="2025-03-17T17:42:38.049323283Z" level=info msg="TearDown network for sandbox \"980554b391378bf7ff18652b7dc7922f5d0a4743b3074f835149be5d75b9c811\" successfully" Mar 17 17:42:38.049840 containerd[1467]: time="2025-03-17T17:42:38.049332831Z" level=info msg="StopPodSandbox for \"980554b391378bf7ff18652b7dc7922f5d0a4743b3074f835149be5d75b9c811\" returns successfully" Mar 17 17:42:38.049966 systemd[1]: run-netns-cni\x2d2884a2fe\x2d8811\x2d1f61\x2d1641\x2def63c0bdda98.mount: Deactivated successfully. Mar 17 17:42:38.050286 containerd[1467]: time="2025-03-17T17:42:38.050108829Z" level=info msg="StopPodSandbox for \"317763565ecd927b3fe20fd547beaf2029d8ad181454207c47710163dbd09a88\"" Mar 17 17:42:38.050286 containerd[1467]: time="2025-03-17T17:42:38.050255556Z" level=info msg="TearDown network for sandbox \"317763565ecd927b3fe20fd547beaf2029d8ad181454207c47710163dbd09a88\" successfully" Mar 17 17:42:38.050286 containerd[1467]: time="2025-03-17T17:42:38.050269592Z" level=info msg="StopPodSandbox for \"317763565ecd927b3fe20fd547beaf2029d8ad181454207c47710163dbd09a88\" returns successfully" Mar 17 17:42:38.051075 containerd[1467]: time="2025-03-17T17:42:38.051047855Z" level=info msg="StopPodSandbox for \"cc57dbc98644c56867a38eb1c7016e6e03a565b54b9bea52d164d9463bd478d9\"" Mar 17 17:42:38.051423 containerd[1467]: time="2025-03-17T17:42:38.051391842Z" level=info msg="TearDown network for sandbox \"cc57dbc98644c56867a38eb1c7016e6e03a565b54b9bea52d164d9463bd478d9\" successfully" Mar 17 17:42:38.051423 containerd[1467]: time="2025-03-17T17:42:38.051409815Z" level=info msg="StopPodSandbox for \"cc57dbc98644c56867a38eb1c7016e6e03a565b54b9bea52d164d9463bd478d9\" returns successfully" Mar 17 17:42:38.052875 kubelet[2662]: I0317 17:42:38.052395 2662 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27220f5bf5e3f9e3a1064ebd76371f3363a6b06ed379f42a7bd72b4e7c55d55a" Mar 17 17:42:38.052940 containerd[1467]: time="2025-03-17T17:42:38.052764513Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7fddcd7f99-92q58,Uid:e531d9e1-a7bb-4958-adfd-8db7b3f84f95,Namespace:calico-system,Attempt:4,}" Mar 17 17:42:38.053009 containerd[1467]: time="2025-03-17T17:42:38.052988793Z" level=info msg="StopPodSandbox for \"27220f5bf5e3f9e3a1064ebd76371f3363a6b06ed379f42a7bd72b4e7c55d55a\"" Mar 17 17:42:38.053240 containerd[1467]: time="2025-03-17T17:42:38.053219317Z" level=info msg="Ensure that sandbox 27220f5bf5e3f9e3a1064ebd76371f3363a6b06ed379f42a7bd72b4e7c55d55a in task-service has been cleanup successfully" Mar 17 17:42:38.053659 containerd[1467]: time="2025-03-17T17:42:38.053632945Z" level=info msg="TearDown network for sandbox \"27220f5bf5e3f9e3a1064ebd76371f3363a6b06ed379f42a7bd72b4e7c55d55a\" successfully" Mar 17 17:42:38.053707 containerd[1467]: time="2025-03-17T17:42:38.053650919Z" level=info msg="StopPodSandbox for \"27220f5bf5e3f9e3a1064ebd76371f3363a6b06ed379f42a7bd72b4e7c55d55a\" returns successfully" Mar 17 17:42:38.054510 containerd[1467]: time="2025-03-17T17:42:38.054489745Z" level=info msg="StopPodSandbox for \"4a6e2b6a3eecd39a342da3235b8bc51d6bca9e9c8c2f588a315458c197bf0390\"" Mar 17 17:42:38.054576 containerd[1467]: time="2025-03-17T17:42:38.054560619Z" level=info msg="TearDown network for sandbox \"4a6e2b6a3eecd39a342da3235b8bc51d6bca9e9c8c2f588a315458c197bf0390\" successfully" Mar 17 17:42:38.054576 containerd[1467]: time="2025-03-17T17:42:38.054573463Z" level=info msg="StopPodSandbox for \"4a6e2b6a3eecd39a342da3235b8bc51d6bca9e9c8c2f588a315458c197bf0390\" returns successfully" Mar 17 17:42:38.055045 containerd[1467]: time="2025-03-17T17:42:38.054988703Z" level=info msg="StopPodSandbox for \"f905ddf13c6f6a970be32393ab7c951bb1766cec8d75bdc0aca8519fcc2783c7\"" Mar 17 17:42:38.055128 containerd[1467]: time="2025-03-17T17:42:38.055106435Z" level=info msg="TearDown network for sandbox \"f905ddf13c6f6a970be32393ab7c951bb1766cec8d75bdc0aca8519fcc2783c7\" successfully" Mar 17 17:42:38.055128 containerd[1467]: time="2025-03-17T17:42:38.055125221Z" level=info msg="StopPodSandbox for \"f905ddf13c6f6a970be32393ab7c951bb1766cec8d75bdc0aca8519fcc2783c7\" returns successfully" Mar 17 17:42:38.055920 containerd[1467]: time="2025-03-17T17:42:38.055872495Z" level=info msg="StopPodSandbox for \"a45255c5e842412af83dd06ba8b258ef311550cba3da5240506cd00cf0678895\"" Mar 17 17:42:38.056105 kubelet[2662]: I0317 17:42:38.056068 2662 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f73f191946516c6303f3e08ba9764f9ef1ed363435b102ed8343391428b48bc" Mar 17 17:42:38.056145 containerd[1467]: time="2025-03-17T17:42:38.056107127Z" level=info msg="TearDown network for sandbox \"a45255c5e842412af83dd06ba8b258ef311550cba3da5240506cd00cf0678895\" successfully" Mar 17 17:42:38.056145 containerd[1467]: time="2025-03-17T17:42:38.056118107Z" level=info msg="StopPodSandbox for \"a45255c5e842412af83dd06ba8b258ef311550cba3da5240506cd00cf0678895\" returns successfully" Mar 17 17:42:38.056311 systemd[1]: run-netns-cni\x2dbcf90909\x2db415\x2df75a\x2d60cc\x2d3275dddadc0b.mount: Deactivated successfully. Mar 17 17:42:38.056553 containerd[1467]: time="2025-03-17T17:42:38.056523390Z" level=info msg="StopPodSandbox for \"1f73f191946516c6303f3e08ba9764f9ef1ed363435b102ed8343391428b48bc\"" Mar 17 17:42:38.056785 containerd[1467]: time="2025-03-17T17:42:38.056719888Z" level=info msg="Ensure that sandbox 1f73f191946516c6303f3e08ba9764f9ef1ed363435b102ed8343391428b48bc in task-service has been cleanup successfully" Mar 17 17:42:38.057792 containerd[1467]: time="2025-03-17T17:42:38.057487622Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86c8b6f988-79bzl,Uid:3265838f-27ca-489d-8c65-7cadce5cca40,Namespace:calico-apiserver,Attempt:4,}" Mar 17 17:42:38.057982 containerd[1467]: time="2025-03-17T17:42:38.057965721Z" level=info msg="TearDown network for sandbox \"1f73f191946516c6303f3e08ba9764f9ef1ed363435b102ed8343391428b48bc\" successfully" Mar 17 17:42:38.058049 containerd[1467]: time="2025-03-17T17:42:38.058035823Z" level=info msg="StopPodSandbox for \"1f73f191946516c6303f3e08ba9764f9ef1ed363435b102ed8343391428b48bc\" returns successfully" Mar 17 17:42:38.058702 containerd[1467]: time="2025-03-17T17:42:38.058686576Z" level=info msg="StopPodSandbox for \"eec0898e435269a73cab252710c67519bb65ec8c70880a65c0d84177e2d5547f\"" Mar 17 17:42:38.058872 containerd[1467]: time="2025-03-17T17:42:38.058856235Z" level=info msg="TearDown network for sandbox \"eec0898e435269a73cab252710c67519bb65ec8c70880a65c0d84177e2d5547f\" successfully" Mar 17 17:42:38.058940 containerd[1467]: time="2025-03-17T17:42:38.058928391Z" level=info msg="StopPodSandbox for \"eec0898e435269a73cab252710c67519bb65ec8c70880a65c0d84177e2d5547f\" returns successfully" Mar 17 17:42:38.060408 containerd[1467]: time="2025-03-17T17:42:38.060329445Z" level=info msg="StopPodSandbox for \"4044ac7223d418a825ef4eedf876ba4769d2804921eba6925b3d0d6d4afbc33a\"" Mar 17 17:42:38.060420 systemd[1]: run-netns-cni\x2d84f15b81\x2d13bb\x2dfaeb\x2dc8e4\x2dc8649b6a04c1.mount: Deactivated successfully. Mar 17 17:42:38.060520 containerd[1467]: time="2025-03-17T17:42:38.060498143Z" level=info msg="TearDown network for sandbox \"4044ac7223d418a825ef4eedf876ba4769d2804921eba6925b3d0d6d4afbc33a\" successfully" Mar 17 17:42:38.060520 containerd[1467]: time="2025-03-17T17:42:38.060514133Z" level=info msg="StopPodSandbox for \"4044ac7223d418a825ef4eedf876ba4769d2804921eba6925b3d0d6d4afbc33a\" returns successfully" Mar 17 17:42:38.061224 containerd[1467]: time="2025-03-17T17:42:38.061066711Z" level=info msg="StopPodSandbox for \"3f5345f62ce3def9878f952bc3b9ba8515b066f140a947d652235852cb08fae5\"" Mar 17 17:42:38.061224 containerd[1467]: time="2025-03-17T17:42:38.061145910Z" level=info msg="TearDown network for sandbox \"3f5345f62ce3def9878f952bc3b9ba8515b066f140a947d652235852cb08fae5\" successfully" Mar 17 17:42:38.061224 containerd[1467]: time="2025-03-17T17:42:38.061164675Z" level=info msg="StopPodSandbox for \"3f5345f62ce3def9878f952bc3b9ba8515b066f140a947d652235852cb08fae5\" returns successfully" Mar 17 17:42:38.062145 containerd[1467]: time="2025-03-17T17:42:38.062112407Z" level=info msg="StopPodSandbox for \"e93aa7d450e66ea71243d70e3ef70d32837865b756a4bd00a6f296f08f7881d8\"" Mar 17 17:42:38.062296 containerd[1467]: time="2025-03-17T17:42:38.062199991Z" level=info msg="TearDown network for sandbox \"e93aa7d450e66ea71243d70e3ef70d32837865b756a4bd00a6f296f08f7881d8\" successfully" Mar 17 17:42:38.062296 containerd[1467]: time="2025-03-17T17:42:38.062211072Z" level=info msg="StopPodSandbox for \"e93aa7d450e66ea71243d70e3ef70d32837865b756a4bd00a6f296f08f7881d8\" returns successfully" Mar 17 17:42:38.062545 kubelet[2662]: I0317 17:42:38.062523 2662 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6dd83b0760f3d749a5709564d2613214f6b299b039d6f1b56dcb51a9cb223ea7" Mar 17 17:42:38.063705 containerd[1467]: time="2025-03-17T17:42:38.063673201Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86c8b6f988-ppvkt,Uid:c9eab445-c078-490b-861c-b96d204afef3,Namespace:calico-apiserver,Attempt:5,}" Mar 17 17:42:38.063783 containerd[1467]: time="2025-03-17T17:42:38.063731572Z" level=info msg="StopPodSandbox for \"6dd83b0760f3d749a5709564d2613214f6b299b039d6f1b56dcb51a9cb223ea7\"" Mar 17 17:42:38.064249 containerd[1467]: time="2025-03-17T17:42:38.064057945Z" level=info msg="Ensure that sandbox 6dd83b0760f3d749a5709564d2613214f6b299b039d6f1b56dcb51a9cb223ea7 in task-service has been cleanup successfully" Mar 17 17:42:38.064318 containerd[1467]: time="2025-03-17T17:42:38.064286044Z" level=info msg="TearDown network for sandbox \"6dd83b0760f3d749a5709564d2613214f6b299b039d6f1b56dcb51a9cb223ea7\" successfully" Mar 17 17:42:38.064346 containerd[1467]: time="2025-03-17T17:42:38.064321611Z" level=info msg="StopPodSandbox for \"6dd83b0760f3d749a5709564d2613214f6b299b039d6f1b56dcb51a9cb223ea7\" returns successfully" Mar 17 17:42:38.065409 containerd[1467]: time="2025-03-17T17:42:38.065295822Z" level=info msg="StopPodSandbox for \"008cee49813c10fd33682eb193a75ecd740653b4fd1a4a322bfcc088c6078101\"" Mar 17 17:42:38.065743 containerd[1467]: time="2025-03-17T17:42:38.065720871Z" level=info msg="TearDown network for sandbox \"008cee49813c10fd33682eb193a75ecd740653b4fd1a4a322bfcc088c6078101\" successfully" Mar 17 17:42:38.065743 containerd[1467]: time="2025-03-17T17:42:38.065738946Z" level=info msg="StopPodSandbox for \"008cee49813c10fd33682eb193a75ecd740653b4fd1a4a322bfcc088c6078101\" returns successfully" Mar 17 17:42:38.066412 containerd[1467]: time="2025-03-17T17:42:38.066382375Z" level=info msg="StopPodSandbox for \"3edf99fd7e4c4bcd608727ff216f57eed4c9785e1d1f4d8c623f0c2aa6ca839a\"" Mar 17 17:42:38.068437 containerd[1467]: time="2025-03-17T17:42:38.066458357Z" level=info msg="TearDown network for sandbox \"3edf99fd7e4c4bcd608727ff216f57eed4c9785e1d1f4d8c623f0c2aa6ca839a\" successfully" Mar 17 17:42:38.068437 containerd[1467]: time="2025-03-17T17:42:38.066469759Z" level=info msg="StopPodSandbox for \"3edf99fd7e4c4bcd608727ff216f57eed4c9785e1d1f4d8c623f0c2aa6ca839a\" returns successfully" Mar 17 17:42:38.068437 containerd[1467]: time="2025-03-17T17:42:38.067183220Z" level=info msg="StopPodSandbox for \"f1d105acdc9b53357e4e621942df86a7fba22a3900a30edbb033cb1322dbced2\"" Mar 17 17:42:38.068437 containerd[1467]: time="2025-03-17T17:42:38.067253102Z" level=info msg="TearDown network for sandbox \"f1d105acdc9b53357e4e621942df86a7fba22a3900a30edbb033cb1322dbced2\" successfully" Mar 17 17:42:38.068437 containerd[1467]: time="2025-03-17T17:42:38.067261438Z" level=info msg="StopPodSandbox for \"f1d105acdc9b53357e4e621942df86a7fba22a3900a30edbb033cb1322dbced2\" returns successfully" Mar 17 17:42:38.068437 containerd[1467]: time="2025-03-17T17:42:38.068010646Z" level=info msg="StopPodSandbox for \"e983d9d5aee8c472a9806708fa057e3b2de9217794b9f28b9fc2133b98942f76\"" Mar 17 17:42:38.068437 containerd[1467]: time="2025-03-17T17:42:38.068109352Z" level=info msg="TearDown network for sandbox \"e983d9d5aee8c472a9806708fa057e3b2de9217794b9f28b9fc2133b98942f76\" successfully" Mar 17 17:42:38.068437 containerd[1467]: time="2025-03-17T17:42:38.068120633Z" level=info msg="StopPodSandbox for \"e983d9d5aee8c472a9806708fa057e3b2de9217794b9f28b9fc2133b98942f76\" returns successfully" Mar 17 17:42:38.066774 systemd[1]: run-netns-cni\x2d87ead328\x2d8223\x2d4b30\x2df392\x2d7337f25b6e20.mount: Deactivated successfully. Mar 17 17:42:38.068791 kubelet[2662]: E0317 17:42:38.068465 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:38.068914 kubelet[2662]: I0317 17:42:38.068885 2662 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9eba50d4b991c21e718945f4ae399cbc0ccc0b0ef2c96bdf7d5d5e8361566abd" Mar 17 17:42:38.069345 containerd[1467]: time="2025-03-17T17:42:38.069324677Z" level=info msg="StopPodSandbox for \"9eba50d4b991c21e718945f4ae399cbc0ccc0b0ef2c96bdf7d5d5e8361566abd\"" Mar 17 17:42:38.069483 containerd[1467]: time="2025-03-17T17:42:38.069445233Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-sjrwq,Uid:228c6093-6dc2-4faf-8ec8-87849ef402ab,Namespace:kube-system,Attempt:5,}" Mar 17 17:42:38.069512 containerd[1467]: time="2025-03-17T17:42:38.069482864Z" level=info msg="Ensure that sandbox 9eba50d4b991c21e718945f4ae399cbc0ccc0b0ef2c96bdf7d5d5e8361566abd in task-service has been cleanup successfully" Mar 17 17:42:38.070144 containerd[1467]: time="2025-03-17T17:42:38.070092490Z" level=info msg="TearDown network for sandbox \"9eba50d4b991c21e718945f4ae399cbc0ccc0b0ef2c96bdf7d5d5e8361566abd\" successfully" Mar 17 17:42:38.070144 containerd[1467]: time="2025-03-17T17:42:38.070111426Z" level=info msg="StopPodSandbox for \"9eba50d4b991c21e718945f4ae399cbc0ccc0b0ef2c96bdf7d5d5e8361566abd\" returns successfully" Mar 17 17:42:38.070578 containerd[1467]: time="2025-03-17T17:42:38.070506950Z" level=info msg="StopPodSandbox for \"284f791e9331db53c5a68aa2679f439389dd260cc6a34b8aff9bc622e96ca04d\"" Mar 17 17:42:38.070648 containerd[1467]: time="2025-03-17T17:42:38.070625091Z" level=info msg="TearDown network for sandbox \"284f791e9331db53c5a68aa2679f439389dd260cc6a34b8aff9bc622e96ca04d\" successfully" Mar 17 17:42:38.070648 containerd[1467]: time="2025-03-17T17:42:38.070636052Z" level=info msg="StopPodSandbox for \"284f791e9331db53c5a68aa2679f439389dd260cc6a34b8aff9bc622e96ca04d\" returns successfully" Mar 17 17:42:38.070897 containerd[1467]: time="2025-03-17T17:42:38.070874130Z" level=info msg="StopPodSandbox for \"3f78c7317a0c3224f4039412ac38faee93bb589fa11754f46d7d13d15315c3d5\"" Mar 17 17:42:38.071218 containerd[1467]: time="2025-03-17T17:42:38.070943269Z" level=info msg="TearDown network for sandbox \"3f78c7317a0c3224f4039412ac38faee93bb589fa11754f46d7d13d15315c3d5\" successfully" Mar 17 17:42:38.071218 containerd[1467]: time="2025-03-17T17:42:38.070955773Z" level=info msg="StopPodSandbox for \"3f78c7317a0c3224f4039412ac38faee93bb589fa11754f46d7d13d15315c3d5\" returns successfully" Mar 17 17:42:38.071733 containerd[1467]: time="2025-03-17T17:42:38.071607629Z" level=info msg="StopPodSandbox for \"6c94b5b9e7a3d7b63ccc9268b52bbe37ef401118cf01d372890eeea51b3ec01a\"" Mar 17 17:42:38.071733 containerd[1467]: time="2025-03-17T17:42:38.071706113Z" level=info msg="TearDown network for sandbox \"6c94b5b9e7a3d7b63ccc9268b52bbe37ef401118cf01d372890eeea51b3ec01a\" successfully" Mar 17 17:42:38.071733 containerd[1467]: time="2025-03-17T17:42:38.071716713Z" level=info msg="StopPodSandbox for \"6c94b5b9e7a3d7b63ccc9268b52bbe37ef401118cf01d372890eeea51b3ec01a\" returns successfully" Mar 17 17:42:38.071957 kubelet[2662]: E0317 17:42:38.071919 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:38.072321 containerd[1467]: time="2025-03-17T17:42:38.072298818Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-fw5hr,Uid:ae45280b-f87f-4410-991d-e0beb4bdc7c6,Namespace:kube-system,Attempt:4,}" Mar 17 17:42:38.072406 kubelet[2662]: I0317 17:42:38.072341 2662 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10ef4549d0b409f4064de73e7cb3238ecdb64a4888527c02814ca83e190159e7" Mar 17 17:42:38.072766 containerd[1467]: time="2025-03-17T17:42:38.072731602Z" level=info msg="StopPodSandbox for \"10ef4549d0b409f4064de73e7cb3238ecdb64a4888527c02814ca83e190159e7\"" Mar 17 17:42:38.072940 containerd[1467]: time="2025-03-17T17:42:38.072920887Z" level=info msg="Ensure that sandbox 10ef4549d0b409f4064de73e7cb3238ecdb64a4888527c02814ca83e190159e7 in task-service has been cleanup successfully" Mar 17 17:42:38.073232 containerd[1467]: time="2025-03-17T17:42:38.073144728Z" level=info msg="TearDown network for sandbox \"10ef4549d0b409f4064de73e7cb3238ecdb64a4888527c02814ca83e190159e7\" successfully" Mar 17 17:42:38.073232 containerd[1467]: time="2025-03-17T17:42:38.073182339Z" level=info msg="StopPodSandbox for \"10ef4549d0b409f4064de73e7cb3238ecdb64a4888527c02814ca83e190159e7\" returns successfully" Mar 17 17:42:38.075377 containerd[1467]: time="2025-03-17T17:42:38.075308517Z" level=info msg="StopPodSandbox for \"1033b80b196d9abbe23dc46629630e49f397ddc00f3445a019c576bdd8d54f14\"" Mar 17 17:42:38.077925 containerd[1467]: time="2025-03-17T17:42:38.077886543Z" level=info msg="TearDown network for sandbox \"1033b80b196d9abbe23dc46629630e49f397ddc00f3445a019c576bdd8d54f14\" successfully" Mar 17 17:42:38.078543 containerd[1467]: time="2025-03-17T17:42:38.078412312Z" level=info msg="StopPodSandbox for \"1033b80b196d9abbe23dc46629630e49f397ddc00f3445a019c576bdd8d54f14\" returns successfully" Mar 17 17:42:38.078929 containerd[1467]: time="2025-03-17T17:42:38.078901551Z" level=info msg="StopPodSandbox for \"47b88b63e50185559458c168aee26da05df22f8d62d10c49042400bc79ba64b6\"" Mar 17 17:42:38.079202 containerd[1467]: time="2025-03-17T17:42:38.079181007Z" level=info msg="TearDown network for sandbox \"47b88b63e50185559458c168aee26da05df22f8d62d10c49042400bc79ba64b6\" successfully" Mar 17 17:42:38.079202 containerd[1467]: time="2025-03-17T17:42:38.079197848Z" level=info msg="StopPodSandbox for \"47b88b63e50185559458c168aee26da05df22f8d62d10c49042400bc79ba64b6\" returns successfully" Mar 17 17:42:38.079711 containerd[1467]: time="2025-03-17T17:42:38.079683461Z" level=info msg="StopPodSandbox for \"b33f35d6e3bd6b1a5f1d8cb21ebceacfc304ffe6c1b7ed3aa40b78efa59e6a03\"" Mar 17 17:42:38.079995 containerd[1467]: time="2025-03-17T17:42:38.079968958Z" level=info msg="TearDown network for sandbox \"b33f35d6e3bd6b1a5f1d8cb21ebceacfc304ffe6c1b7ed3aa40b78efa59e6a03\" successfully" Mar 17 17:42:38.079995 containerd[1467]: time="2025-03-17T17:42:38.079987354Z" level=info msg="StopPodSandbox for \"b33f35d6e3bd6b1a5f1d8cb21ebceacfc304ffe6c1b7ed3aa40b78efa59e6a03\" returns successfully" Mar 17 17:42:38.081958 containerd[1467]: time="2025-03-17T17:42:38.081851628Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-j9gdg,Uid:27eb0c6c-40f4-4b19-b513-0003d63df5d0,Namespace:calico-system,Attempt:4,}" Mar 17 17:42:38.604240 systemd[1]: run-netns-cni\x2d47b76ed6\x2d8169\x2da0eb\x2de275\x2d899d26a21251.mount: Deactivated successfully. Mar 17 17:42:38.606985 systemd[1]: run-netns-cni\x2d9b8f6b2a\x2dc2dc\x2d0300\x2ddb70\x2d0c531ba7999a.mount: Deactivated successfully. Mar 17 17:42:39.013956 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount110644591.mount: Deactivated successfully. Mar 17 17:42:39.282465 containerd[1467]: time="2025-03-17T17:42:39.282332220Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.29.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:42:39.287176 containerd[1467]: time="2025-03-17T17:42:39.287057924Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.29.2: active requests=0, bytes read=142241445" Mar 17 17:42:39.288872 containerd[1467]: time="2025-03-17T17:42:39.288828362Z" level=info msg="ImageCreate event name:\"sha256:048bf7af1f8c697d151dbecc478a18e89d89ed8627da98e17a56c11b3d45d351\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:42:39.292927 containerd[1467]: time="2025-03-17T17:42:39.292885709Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:d9a21be37fe591ee5ab5a2e3dc26408ea165a44a55705102ffaa002de9908b32\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:42:39.295658 containerd[1467]: time="2025-03-17T17:42:39.295623125Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.29.2\" with image id \"sha256:048bf7af1f8c697d151dbecc478a18e89d89ed8627da98e17a56c11b3d45d351\", repo tag \"ghcr.io/flatcar/calico/node:v3.29.2\", repo digest \"ghcr.io/flatcar/calico/node@sha256:d9a21be37fe591ee5ab5a2e3dc26408ea165a44a55705102ffaa002de9908b32\", size \"142241307\" in 5.315798629s" Mar 17 17:42:39.295707 containerd[1467]: time="2025-03-17T17:42:39.295659934Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.2\" returns image reference \"sha256:048bf7af1f8c697d151dbecc478a18e89d89ed8627da98e17a56c11b3d45d351\"" Mar 17 17:42:39.305970 containerd[1467]: time="2025-03-17T17:42:39.305910843Z" level=info msg="CreateContainer within sandbox \"0fa693deaf0f5a67ec99e9c752c255e61ecc989170bfbc0ea77ac17bb35fd6b3\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Mar 17 17:42:39.331534 containerd[1467]: time="2025-03-17T17:42:39.331463588Z" level=error msg="Failed to destroy network for sandbox \"e570c82c935c8544758ef95881b0e0abf15540c0e05ff8eaf1df19398bdcd032\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:39.332597 containerd[1467]: time="2025-03-17T17:42:39.332424254Z" level=error msg="encountered an error cleaning up failed sandbox \"e570c82c935c8544758ef95881b0e0abf15540c0e05ff8eaf1df19398bdcd032\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:39.333667 containerd[1467]: time="2025-03-17T17:42:39.333063996Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86c8b6f988-79bzl,Uid:3265838f-27ca-489d-8c65-7cadce5cca40,Namespace:calico-apiserver,Attempt:4,} failed, error" error="failed to setup network for sandbox \"e570c82c935c8544758ef95881b0e0abf15540c0e05ff8eaf1df19398bdcd032\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:39.333667 containerd[1467]: time="2025-03-17T17:42:39.333183532Z" level=error msg="Failed to destroy network for sandbox \"86089e35a842c8c8e7ea82ce5cbce4b7646d35903cb49ceaccd5d1db0dfbc90c\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:39.334103 containerd[1467]: time="2025-03-17T17:42:39.334056793Z" level=error msg="encountered an error cleaning up failed sandbox \"86089e35a842c8c8e7ea82ce5cbce4b7646d35903cb49ceaccd5d1db0dfbc90c\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:39.334167 containerd[1467]: time="2025-03-17T17:42:39.334142664Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-fw5hr,Uid:ae45280b-f87f-4410-991d-e0beb4bdc7c6,Namespace:kube-system,Attempt:4,} failed, error" error="failed to setup network for sandbox \"86089e35a842c8c8e7ea82ce5cbce4b7646d35903cb49ceaccd5d1db0dfbc90c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:39.337945 kubelet[2662]: E0317 17:42:39.337902 2662 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e570c82c935c8544758ef95881b0e0abf15540c0e05ff8eaf1df19398bdcd032\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:39.338777 kubelet[2662]: E0317 17:42:39.338418 2662 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e570c82c935c8544758ef95881b0e0abf15540c0e05ff8eaf1df19398bdcd032\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86c8b6f988-79bzl" Mar 17 17:42:39.338777 kubelet[2662]: E0317 17:42:39.338459 2662 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e570c82c935c8544758ef95881b0e0abf15540c0e05ff8eaf1df19398bdcd032\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86c8b6f988-79bzl" Mar 17 17:42:39.338777 kubelet[2662]: E0317 17:42:39.338498 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-86c8b6f988-79bzl_calico-apiserver(3265838f-27ca-489d-8c65-7cadce5cca40)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-86c8b6f988-79bzl_calico-apiserver(3265838f-27ca-489d-8c65-7cadce5cca40)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"e570c82c935c8544758ef95881b0e0abf15540c0e05ff8eaf1df19398bdcd032\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-86c8b6f988-79bzl" podUID="3265838f-27ca-489d-8c65-7cadce5cca40" Mar 17 17:42:39.338951 kubelet[2662]: E0317 17:42:39.338367 2662 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"86089e35a842c8c8e7ea82ce5cbce4b7646d35903cb49ceaccd5d1db0dfbc90c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:39.338951 kubelet[2662]: E0317 17:42:39.338712 2662 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"86089e35a842c8c8e7ea82ce5cbce4b7646d35903cb49ceaccd5d1db0dfbc90c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-fw5hr" Mar 17 17:42:39.338951 kubelet[2662]: E0317 17:42:39.338728 2662 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"86089e35a842c8c8e7ea82ce5cbce4b7646d35903cb49ceaccd5d1db0dfbc90c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-fw5hr" Mar 17 17:42:39.339022 kubelet[2662]: E0317 17:42:39.338750 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-fw5hr_kube-system(ae45280b-f87f-4410-991d-e0beb4bdc7c6)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-fw5hr_kube-system(ae45280b-f87f-4410-991d-e0beb4bdc7c6)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"86089e35a842c8c8e7ea82ce5cbce4b7646d35903cb49ceaccd5d1db0dfbc90c\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-fw5hr" podUID="ae45280b-f87f-4410-991d-e0beb4bdc7c6" Mar 17 17:42:39.355911 containerd[1467]: time="2025-03-17T17:42:39.355847576Z" level=error msg="Failed to destroy network for sandbox \"58f98dec59f83fef7090b948443c29832b1906274dd91e7d0c3b3c902d3df763\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:39.356927 containerd[1467]: time="2025-03-17T17:42:39.356869888Z" level=error msg="encountered an error cleaning up failed sandbox \"58f98dec59f83fef7090b948443c29832b1906274dd91e7d0c3b3c902d3df763\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:39.357217 containerd[1467]: time="2025-03-17T17:42:39.357109527Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86c8b6f988-ppvkt,Uid:c9eab445-c078-490b-861c-b96d204afef3,Namespace:calico-apiserver,Attempt:5,} failed, error" error="failed to setup network for sandbox \"58f98dec59f83fef7090b948443c29832b1906274dd91e7d0c3b3c902d3df763\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:39.357287 containerd[1467]: time="2025-03-17T17:42:39.357246495Z" level=info msg="CreateContainer within sandbox \"0fa693deaf0f5a67ec99e9c752c255e61ecc989170bfbc0ea77ac17bb35fd6b3\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"e3755fc1ee54cacb19b1da6347e008cbf9a6ae1adedcb01031253ef03e4cb0b7\"" Mar 17 17:42:39.357685 kubelet[2662]: E0317 17:42:39.357488 2662 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"58f98dec59f83fef7090b948443c29832b1906274dd91e7d0c3b3c902d3df763\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:39.357685 kubelet[2662]: E0317 17:42:39.357549 2662 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"58f98dec59f83fef7090b948443c29832b1906274dd91e7d0c3b3c902d3df763\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86c8b6f988-ppvkt" Mar 17 17:42:39.357685 kubelet[2662]: E0317 17:42:39.357571 2662 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"58f98dec59f83fef7090b948443c29832b1906274dd91e7d0c3b3c902d3df763\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-86c8b6f988-ppvkt" Mar 17 17:42:39.357815 kubelet[2662]: E0317 17:42:39.357609 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-86c8b6f988-ppvkt_calico-apiserver(c9eab445-c078-490b-861c-b96d204afef3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-86c8b6f988-ppvkt_calico-apiserver(c9eab445-c078-490b-861c-b96d204afef3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"58f98dec59f83fef7090b948443c29832b1906274dd91e7d0c3b3c902d3df763\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-86c8b6f988-ppvkt" podUID="c9eab445-c078-490b-861c-b96d204afef3" Mar 17 17:42:39.359782 containerd[1467]: time="2025-03-17T17:42:39.358343187Z" level=info msg="StartContainer for \"e3755fc1ee54cacb19b1da6347e008cbf9a6ae1adedcb01031253ef03e4cb0b7\"" Mar 17 17:42:39.368548 containerd[1467]: time="2025-03-17T17:42:39.368500120Z" level=error msg="Failed to destroy network for sandbox \"84f2f05c82276b89c454b200d6c53ebf68e120d900db1ee2b4b76ce464926d93\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:39.368917 containerd[1467]: time="2025-03-17T17:42:39.368887980Z" level=error msg="encountered an error cleaning up failed sandbox \"84f2f05c82276b89c454b200d6c53ebf68e120d900db1ee2b4b76ce464926d93\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:39.368977 containerd[1467]: time="2025-03-17T17:42:39.368949585Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7fddcd7f99-92q58,Uid:e531d9e1-a7bb-4958-adfd-8db7b3f84f95,Namespace:calico-system,Attempt:4,} failed, error" error="failed to setup network for sandbox \"84f2f05c82276b89c454b200d6c53ebf68e120d900db1ee2b4b76ce464926d93\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:39.369283 kubelet[2662]: E0317 17:42:39.369238 2662 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"84f2f05c82276b89c454b200d6c53ebf68e120d900db1ee2b4b76ce464926d93\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:39.369501 kubelet[2662]: E0317 17:42:39.369472 2662 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"84f2f05c82276b89c454b200d6c53ebf68e120d900db1ee2b4b76ce464926d93\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-7fddcd7f99-92q58" Mar 17 17:42:39.369627 kubelet[2662]: E0317 17:42:39.369610 2662 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"84f2f05c82276b89c454b200d6c53ebf68e120d900db1ee2b4b76ce464926d93\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-7fddcd7f99-92q58" Mar 17 17:42:39.369759 kubelet[2662]: E0317 17:42:39.369721 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-7fddcd7f99-92q58_calico-system(e531d9e1-a7bb-4958-adfd-8db7b3f84f95)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-7fddcd7f99-92q58_calico-system(e531d9e1-a7bb-4958-adfd-8db7b3f84f95)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"84f2f05c82276b89c454b200d6c53ebf68e120d900db1ee2b4b76ce464926d93\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-7fddcd7f99-92q58" podUID="e531d9e1-a7bb-4958-adfd-8db7b3f84f95" Mar 17 17:42:39.378245 containerd[1467]: time="2025-03-17T17:42:39.378180637Z" level=error msg="Failed to destroy network for sandbox \"32d2a6ea6079479e27c144f7baad1a289529d3a29ff0a7f2e40468d0ecdc1347\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:39.378795 containerd[1467]: time="2025-03-17T17:42:39.378629972Z" level=error msg="encountered an error cleaning up failed sandbox \"32d2a6ea6079479e27c144f7baad1a289529d3a29ff0a7f2e40468d0ecdc1347\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:39.378795 containerd[1467]: time="2025-03-17T17:42:39.378689003Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-j9gdg,Uid:27eb0c6c-40f4-4b19-b513-0003d63df5d0,Namespace:calico-system,Attempt:4,} failed, error" error="failed to setup network for sandbox \"32d2a6ea6079479e27c144f7baad1a289529d3a29ff0a7f2e40468d0ecdc1347\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:39.379070 kubelet[2662]: E0317 17:42:39.378933 2662 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"32d2a6ea6079479e27c144f7baad1a289529d3a29ff0a7f2e40468d0ecdc1347\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:39.379070 kubelet[2662]: E0317 17:42:39.379009 2662 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"32d2a6ea6079479e27c144f7baad1a289529d3a29ff0a7f2e40468d0ecdc1347\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-j9gdg" Mar 17 17:42:39.379070 kubelet[2662]: E0317 17:42:39.379026 2662 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"32d2a6ea6079479e27c144f7baad1a289529d3a29ff0a7f2e40468d0ecdc1347\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-j9gdg" Mar 17 17:42:39.379197 kubelet[2662]: E0317 17:42:39.379077 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-j9gdg_calico-system(27eb0c6c-40f4-4b19-b513-0003d63df5d0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-j9gdg_calico-system(27eb0c6c-40f4-4b19-b513-0003d63df5d0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"32d2a6ea6079479e27c144f7baad1a289529d3a29ff0a7f2e40468d0ecdc1347\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-j9gdg" podUID="27eb0c6c-40f4-4b19-b513-0003d63df5d0" Mar 17 17:42:39.381330 containerd[1467]: time="2025-03-17T17:42:39.381302075Z" level=error msg="Failed to destroy network for sandbox \"9906341f06aa2f7e95fd6b7d0f4548be4a6da1921703748eb9f0e054258d0bf4\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:39.381663 containerd[1467]: time="2025-03-17T17:42:39.381618189Z" level=error msg="encountered an error cleaning up failed sandbox \"9906341f06aa2f7e95fd6b7d0f4548be4a6da1921703748eb9f0e054258d0bf4\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:39.381663 containerd[1467]: time="2025-03-17T17:42:39.381659878Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-sjrwq,Uid:228c6093-6dc2-4faf-8ec8-87849ef402ab,Namespace:kube-system,Attempt:5,} failed, error" error="failed to setup network for sandbox \"9906341f06aa2f7e95fd6b7d0f4548be4a6da1921703748eb9f0e054258d0bf4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:39.381920 kubelet[2662]: E0317 17:42:39.381857 2662 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9906341f06aa2f7e95fd6b7d0f4548be4a6da1921703748eb9f0e054258d0bf4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 17 17:42:39.381966 kubelet[2662]: E0317 17:42:39.381922 2662 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9906341f06aa2f7e95fd6b7d0f4548be4a6da1921703748eb9f0e054258d0bf4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-sjrwq" Mar 17 17:42:39.381966 kubelet[2662]: E0317 17:42:39.381945 2662 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"9906341f06aa2f7e95fd6b7d0f4548be4a6da1921703748eb9f0e054258d0bf4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-sjrwq" Mar 17 17:42:39.382030 kubelet[2662]: E0317 17:42:39.381991 2662 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-sjrwq_kube-system(228c6093-6dc2-4faf-8ec8-87849ef402ab)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-sjrwq_kube-system(228c6093-6dc2-4faf-8ec8-87849ef402ab)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"9906341f06aa2f7e95fd6b7d0f4548be4a6da1921703748eb9f0e054258d0bf4\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-sjrwq" podUID="228c6093-6dc2-4faf-8ec8-87849ef402ab" Mar 17 17:42:39.428495 systemd[1]: Started cri-containerd-e3755fc1ee54cacb19b1da6347e008cbf9a6ae1adedcb01031253ef03e4cb0b7.scope - libcontainer container e3755fc1ee54cacb19b1da6347e008cbf9a6ae1adedcb01031253ef03e4cb0b7. Mar 17 17:42:39.461224 containerd[1467]: time="2025-03-17T17:42:39.461188147Z" level=info msg="StartContainer for \"e3755fc1ee54cacb19b1da6347e008cbf9a6ae1adedcb01031253ef03e4cb0b7\" returns successfully" Mar 17 17:42:39.522584 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Mar 17 17:42:39.522687 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Mar 17 17:42:39.610605 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-86089e35a842c8c8e7ea82ce5cbce4b7646d35903cb49ceaccd5d1db0dfbc90c-shm.mount: Deactivated successfully. Mar 17 17:42:39.610739 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-e570c82c935c8544758ef95881b0e0abf15540c0e05ff8eaf1df19398bdcd032-shm.mount: Deactivated successfully. Mar 17 17:42:40.076794 kubelet[2662]: I0317 17:42:40.076764 2662 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="86089e35a842c8c8e7ea82ce5cbce4b7646d35903cb49ceaccd5d1db0dfbc90c" Mar 17 17:42:40.077188 containerd[1467]: time="2025-03-17T17:42:40.077159001Z" level=info msg="StopPodSandbox for \"86089e35a842c8c8e7ea82ce5cbce4b7646d35903cb49ceaccd5d1db0dfbc90c\"" Mar 17 17:42:40.077413 containerd[1467]: time="2025-03-17T17:42:40.077392560Z" level=info msg="Ensure that sandbox 86089e35a842c8c8e7ea82ce5cbce4b7646d35903cb49ceaccd5d1db0dfbc90c in task-service has been cleanup successfully" Mar 17 17:42:40.078480 containerd[1467]: time="2025-03-17T17:42:40.078454095Z" level=info msg="TearDown network for sandbox \"86089e35a842c8c8e7ea82ce5cbce4b7646d35903cb49ceaccd5d1db0dfbc90c\" successfully" Mar 17 17:42:40.078480 containerd[1467]: time="2025-03-17T17:42:40.078476668Z" level=info msg="StopPodSandbox for \"86089e35a842c8c8e7ea82ce5cbce4b7646d35903cb49ceaccd5d1db0dfbc90c\" returns successfully" Mar 17 17:42:40.079120 containerd[1467]: time="2025-03-17T17:42:40.079048282Z" level=info msg="StopPodSandbox for \"9eba50d4b991c21e718945f4ae399cbc0ccc0b0ef2c96bdf7d5d5e8361566abd\"" Mar 17 17:42:40.079250 containerd[1467]: time="2025-03-17T17:42:40.079184738Z" level=info msg="TearDown network for sandbox \"9eba50d4b991c21e718945f4ae399cbc0ccc0b0ef2c96bdf7d5d5e8361566abd\" successfully" Mar 17 17:42:40.079250 containerd[1467]: time="2025-03-17T17:42:40.079203924Z" level=info msg="StopPodSandbox for \"9eba50d4b991c21e718945f4ae399cbc0ccc0b0ef2c96bdf7d5d5e8361566abd\" returns successfully" Mar 17 17:42:40.079921 containerd[1467]: time="2025-03-17T17:42:40.079631699Z" level=info msg="StopPodSandbox for \"284f791e9331db53c5a68aa2679f439389dd260cc6a34b8aff9bc622e96ca04d\"" Mar 17 17:42:40.079921 containerd[1467]: time="2025-03-17T17:42:40.079716548Z" level=info msg="TearDown network for sandbox \"284f791e9331db53c5a68aa2679f439389dd260cc6a34b8aff9bc622e96ca04d\" successfully" Mar 17 17:42:40.079921 containerd[1467]: time="2025-03-17T17:42:40.079726156Z" level=info msg="StopPodSandbox for \"284f791e9331db53c5a68aa2679f439389dd260cc6a34b8aff9bc622e96ca04d\" returns successfully" Mar 17 17:42:40.080607 systemd[1]: run-netns-cni\x2df162d769\x2dadca\x2d03ac\x2d51a2\x2d8abff603d642.mount: Deactivated successfully. Mar 17 17:42:40.080741 kubelet[2662]: E0317 17:42:40.080657 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:40.080944 containerd[1467]: time="2025-03-17T17:42:40.080787210Z" level=info msg="StopPodSandbox for \"3f78c7317a0c3224f4039412ac38faee93bb589fa11754f46d7d13d15315c3d5\"" Mar 17 17:42:40.080944 containerd[1467]: time="2025-03-17T17:42:40.080896405Z" level=info msg="TearDown network for sandbox \"3f78c7317a0c3224f4039412ac38faee93bb589fa11754f46d7d13d15315c3d5\" successfully" Mar 17 17:42:40.080944 containerd[1467]: time="2025-03-17T17:42:40.080907035Z" level=info msg="StopPodSandbox for \"3f78c7317a0c3224f4039412ac38faee93bb589fa11754f46d7d13d15315c3d5\" returns successfully" Mar 17 17:42:40.081538 containerd[1467]: time="2025-03-17T17:42:40.081387499Z" level=info msg="StopPodSandbox for \"6c94b5b9e7a3d7b63ccc9268b52bbe37ef401118cf01d372890eeea51b3ec01a\"" Mar 17 17:42:40.081538 containerd[1467]: time="2025-03-17T17:42:40.081479591Z" level=info msg="TearDown network for sandbox \"6c94b5b9e7a3d7b63ccc9268b52bbe37ef401118cf01d372890eeea51b3ec01a\" successfully" Mar 17 17:42:40.081538 containerd[1467]: time="2025-03-17T17:42:40.081491966Z" level=info msg="StopPodSandbox for \"6c94b5b9e7a3d7b63ccc9268b52bbe37ef401118cf01d372890eeea51b3ec01a\" returns successfully" Mar 17 17:42:40.082274 kubelet[2662]: E0317 17:42:40.081650 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:40.082337 containerd[1467]: time="2025-03-17T17:42:40.081926923Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-fw5hr,Uid:ae45280b-f87f-4410-991d-e0beb4bdc7c6,Namespace:kube-system,Attempt:5,}" Mar 17 17:42:40.083571 kubelet[2662]: I0317 17:42:40.083549 2662 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="32d2a6ea6079479e27c144f7baad1a289529d3a29ff0a7f2e40468d0ecdc1347" Mar 17 17:42:40.083981 containerd[1467]: time="2025-03-17T17:42:40.083958912Z" level=info msg="StopPodSandbox for \"32d2a6ea6079479e27c144f7baad1a289529d3a29ff0a7f2e40468d0ecdc1347\"" Mar 17 17:42:40.084164 containerd[1467]: time="2025-03-17T17:42:40.084128701Z" level=info msg="Ensure that sandbox 32d2a6ea6079479e27c144f7baad1a289529d3a29ff0a7f2e40468d0ecdc1347 in task-service has been cleanup successfully" Mar 17 17:42:40.089414 containerd[1467]: time="2025-03-17T17:42:40.084688043Z" level=info msg="TearDown network for sandbox \"32d2a6ea6079479e27c144f7baad1a289529d3a29ff0a7f2e40468d0ecdc1347\" successfully" Mar 17 17:42:40.089414 containerd[1467]: time="2025-03-17T17:42:40.084704824Z" level=info msg="StopPodSandbox for \"32d2a6ea6079479e27c144f7baad1a289529d3a29ff0a7f2e40468d0ecdc1347\" returns successfully" Mar 17 17:42:40.089414 containerd[1467]: time="2025-03-17T17:42:40.085170509Z" level=info msg="StopPodSandbox for \"10ef4549d0b409f4064de73e7cb3238ecdb64a4888527c02814ca83e190159e7\"" Mar 17 17:42:40.089414 containerd[1467]: time="2025-03-17T17:42:40.085253415Z" level=info msg="TearDown network for sandbox \"10ef4549d0b409f4064de73e7cb3238ecdb64a4888527c02814ca83e190159e7\" successfully" Mar 17 17:42:40.089414 containerd[1467]: time="2025-03-17T17:42:40.085263093Z" level=info msg="StopPodSandbox for \"10ef4549d0b409f4064de73e7cb3238ecdb64a4888527c02814ca83e190159e7\" returns successfully" Mar 17 17:42:40.090402 containerd[1467]: time="2025-03-17T17:42:40.090021036Z" level=info msg="StopPodSandbox for \"1033b80b196d9abbe23dc46629630e49f397ddc00f3445a019c576bdd8d54f14\"" Mar 17 17:42:40.090402 containerd[1467]: time="2025-03-17T17:42:40.090173022Z" level=info msg="TearDown network for sandbox \"1033b80b196d9abbe23dc46629630e49f397ddc00f3445a019c576bdd8d54f14\" successfully" Mar 17 17:42:40.090402 containerd[1467]: time="2025-03-17T17:42:40.090186457Z" level=info msg="StopPodSandbox for \"1033b80b196d9abbe23dc46629630e49f397ddc00f3445a019c576bdd8d54f14\" returns successfully" Mar 17 17:42:40.090724 systemd[1]: run-netns-cni\x2d43a5e868\x2dd36e\x2d5a92\x2df07e\x2d6d9d44edfb83.mount: Deactivated successfully. Mar 17 17:42:40.091686 containerd[1467]: time="2025-03-17T17:42:40.091666378Z" level=info msg="StopPodSandbox for \"47b88b63e50185559458c168aee26da05df22f8d62d10c49042400bc79ba64b6\"" Mar 17 17:42:40.091905 containerd[1467]: time="2025-03-17T17:42:40.091834154Z" level=info msg="TearDown network for sandbox \"47b88b63e50185559458c168aee26da05df22f8d62d10c49042400bc79ba64b6\" successfully" Mar 17 17:42:40.091905 containerd[1467]: time="2025-03-17T17:42:40.091891271Z" level=info msg="StopPodSandbox for \"47b88b63e50185559458c168aee26da05df22f8d62d10c49042400bc79ba64b6\" returns successfully" Mar 17 17:42:40.092587 containerd[1467]: time="2025-03-17T17:42:40.092537566Z" level=info msg="StopPodSandbox for \"b33f35d6e3bd6b1a5f1d8cb21ebceacfc304ffe6c1b7ed3aa40b78efa59e6a03\"" Mar 17 17:42:40.092652 containerd[1467]: time="2025-03-17T17:42:40.092617316Z" level=info msg="TearDown network for sandbox \"b33f35d6e3bd6b1a5f1d8cb21ebceacfc304ffe6c1b7ed3aa40b78efa59e6a03\" successfully" Mar 17 17:42:40.092652 containerd[1467]: time="2025-03-17T17:42:40.092645810Z" level=info msg="StopPodSandbox for \"b33f35d6e3bd6b1a5f1d8cb21ebceacfc304ffe6c1b7ed3aa40b78efa59e6a03\" returns successfully" Mar 17 17:42:40.092751 kubelet[2662]: I0317 17:42:40.092728 2662 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="84f2f05c82276b89c454b200d6c53ebf68e120d900db1ee2b4b76ce464926d93" Mar 17 17:42:40.093147 containerd[1467]: time="2025-03-17T17:42:40.093125260Z" level=info msg="StopPodSandbox for \"84f2f05c82276b89c454b200d6c53ebf68e120d900db1ee2b4b76ce464926d93\"" Mar 17 17:42:40.093309 containerd[1467]: time="2025-03-17T17:42:40.093288427Z" level=info msg="Ensure that sandbox 84f2f05c82276b89c454b200d6c53ebf68e120d900db1ee2b4b76ce464926d93 in task-service has been cleanup successfully" Mar 17 17:42:40.093502 containerd[1467]: time="2025-03-17T17:42:40.093483384Z" level=info msg="TearDown network for sandbox \"84f2f05c82276b89c454b200d6c53ebf68e120d900db1ee2b4b76ce464926d93\" successfully" Mar 17 17:42:40.093502 containerd[1467]: time="2025-03-17T17:42:40.093497841Z" level=info msg="StopPodSandbox for \"84f2f05c82276b89c454b200d6c53ebf68e120d900db1ee2b4b76ce464926d93\" returns successfully" Mar 17 17:42:40.093565 containerd[1467]: time="2025-03-17T17:42:40.093508982Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-j9gdg,Uid:27eb0c6c-40f4-4b19-b513-0003d63df5d0,Namespace:calico-system,Attempt:5,}" Mar 17 17:42:40.094412 containerd[1467]: time="2025-03-17T17:42:40.094379569Z" level=info msg="StopPodSandbox for \"b71e2358cd15f8a4cec878737cb78ea057ba63b0bc401daa2c0c7f9f8d7fef30\"" Mar 17 17:42:40.094524 containerd[1467]: time="2025-03-17T17:42:40.094479376Z" level=info msg="TearDown network for sandbox \"b71e2358cd15f8a4cec878737cb78ea057ba63b0bc401daa2c0c7f9f8d7fef30\" successfully" Mar 17 17:42:40.094524 containerd[1467]: time="2025-03-17T17:42:40.094494795Z" level=info msg="StopPodSandbox for \"b71e2358cd15f8a4cec878737cb78ea057ba63b0bc401daa2c0c7f9f8d7fef30\" returns successfully" Mar 17 17:42:40.095493 containerd[1467]: time="2025-03-17T17:42:40.095465991Z" level=info msg="StopPodSandbox for \"980554b391378bf7ff18652b7dc7922f5d0a4743b3074f835149be5d75b9c811\"" Mar 17 17:42:40.095541 systemd[1]: run-netns-cni\x2d4145619e\x2d72f8\x2d8c69\x2da11c\x2d231330f03e6b.mount: Deactivated successfully. Mar 17 17:42:40.096095 containerd[1467]: time="2025-03-17T17:42:40.095549267Z" level=info msg="TearDown network for sandbox \"980554b391378bf7ff18652b7dc7922f5d0a4743b3074f835149be5d75b9c811\" successfully" Mar 17 17:42:40.096095 containerd[1467]: time="2025-03-17T17:42:40.095560518Z" level=info msg="StopPodSandbox for \"980554b391378bf7ff18652b7dc7922f5d0a4743b3074f835149be5d75b9c811\" returns successfully" Mar 17 17:42:40.096095 containerd[1467]: time="2025-03-17T17:42:40.095778668Z" level=info msg="StopPodSandbox for \"317763565ecd927b3fe20fd547beaf2029d8ad181454207c47710163dbd09a88\"" Mar 17 17:42:40.096095 containerd[1467]: time="2025-03-17T17:42:40.095864379Z" level=info msg="TearDown network for sandbox \"317763565ecd927b3fe20fd547beaf2029d8ad181454207c47710163dbd09a88\" successfully" Mar 17 17:42:40.096095 containerd[1467]: time="2025-03-17T17:42:40.095875069Z" level=info msg="StopPodSandbox for \"317763565ecd927b3fe20fd547beaf2029d8ad181454207c47710163dbd09a88\" returns successfully" Mar 17 17:42:40.096276 containerd[1467]: time="2025-03-17T17:42:40.096192846Z" level=info msg="StopPodSandbox for \"cc57dbc98644c56867a38eb1c7016e6e03a565b54b9bea52d164d9463bd478d9\"" Mar 17 17:42:40.096276 containerd[1467]: time="2025-03-17T17:42:40.096264310Z" level=info msg="TearDown network for sandbox \"cc57dbc98644c56867a38eb1c7016e6e03a565b54b9bea52d164d9463bd478d9\" successfully" Mar 17 17:42:40.096276 containerd[1467]: time="2025-03-17T17:42:40.096273999Z" level=info msg="StopPodSandbox for \"cc57dbc98644c56867a38eb1c7016e6e03a565b54b9bea52d164d9463bd478d9\" returns successfully" Mar 17 17:42:40.096649 containerd[1467]: time="2025-03-17T17:42:40.096627043Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7fddcd7f99-92q58,Uid:e531d9e1-a7bb-4958-adfd-8db7b3f84f95,Namespace:calico-system,Attempt:5,}" Mar 17 17:42:40.096932 kubelet[2662]: I0317 17:42:40.096908 2662 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e570c82c935c8544758ef95881b0e0abf15540c0e05ff8eaf1df19398bdcd032" Mar 17 17:42:40.097343 containerd[1467]: time="2025-03-17T17:42:40.097315997Z" level=info msg="StopPodSandbox for \"e570c82c935c8544758ef95881b0e0abf15540c0e05ff8eaf1df19398bdcd032\"" Mar 17 17:42:40.097540 containerd[1467]: time="2025-03-17T17:42:40.097518659Z" level=info msg="Ensure that sandbox e570c82c935c8544758ef95881b0e0abf15540c0e05ff8eaf1df19398bdcd032 in task-service has been cleanup successfully" Mar 17 17:42:40.097731 containerd[1467]: time="2025-03-17T17:42:40.097711091Z" level=info msg="TearDown network for sandbox \"e570c82c935c8544758ef95881b0e0abf15540c0e05ff8eaf1df19398bdcd032\" successfully" Mar 17 17:42:40.097731 containerd[1467]: time="2025-03-17T17:42:40.097726319Z" level=info msg="StopPodSandbox for \"e570c82c935c8544758ef95881b0e0abf15540c0e05ff8eaf1df19398bdcd032\" returns successfully" Mar 17 17:42:40.098893 containerd[1467]: time="2025-03-17T17:42:40.098724685Z" level=info msg="StopPodSandbox for \"27220f5bf5e3f9e3a1064ebd76371f3363a6b06ed379f42a7bd72b4e7c55d55a\"" Mar 17 17:42:40.098893 containerd[1467]: time="2025-03-17T17:42:40.098797091Z" level=info msg="TearDown network for sandbox \"27220f5bf5e3f9e3a1064ebd76371f3363a6b06ed379f42a7bd72b4e7c55d55a\" successfully" Mar 17 17:42:40.098893 containerd[1467]: time="2025-03-17T17:42:40.098805817Z" level=info msg="StopPodSandbox for \"27220f5bf5e3f9e3a1064ebd76371f3363a6b06ed379f42a7bd72b4e7c55d55a\" returns successfully" Mar 17 17:42:40.099062 containerd[1467]: time="2025-03-17T17:42:40.099043494Z" level=info msg="StopPodSandbox for \"4a6e2b6a3eecd39a342da3235b8bc51d6bca9e9c8c2f588a315458c197bf0390\"" Mar 17 17:42:40.099146 containerd[1467]: time="2025-03-17T17:42:40.099131921Z" level=info msg="TearDown network for sandbox \"4a6e2b6a3eecd39a342da3235b8bc51d6bca9e9c8c2f588a315458c197bf0390\" successfully" Mar 17 17:42:40.099170 containerd[1467]: time="2025-03-17T17:42:40.099144565Z" level=info msg="StopPodSandbox for \"4a6e2b6a3eecd39a342da3235b8bc51d6bca9e9c8c2f588a315458c197bf0390\" returns successfully" Mar 17 17:42:40.100511 containerd[1467]: time="2025-03-17T17:42:40.100395746Z" level=info msg="StopPodSandbox for \"f905ddf13c6f6a970be32393ab7c951bb1766cec8d75bdc0aca8519fcc2783c7\"" Mar 17 17:42:40.100511 containerd[1467]: time="2025-03-17T17:42:40.100469254Z" level=info msg="TearDown network for sandbox \"f905ddf13c6f6a970be32393ab7c951bb1766cec8d75bdc0aca8519fcc2783c7\" successfully" Mar 17 17:42:40.100511 containerd[1467]: time="2025-03-17T17:42:40.100478602Z" level=info msg="StopPodSandbox for \"f905ddf13c6f6a970be32393ab7c951bb1766cec8d75bdc0aca8519fcc2783c7\" returns successfully" Mar 17 17:42:40.100519 systemd[1]: run-netns-cni\x2d8b1d5245\x2d3765\x2d7bd6\x2d6002\x2dae890b8cbe65.mount: Deactivated successfully. Mar 17 17:42:40.101002 containerd[1467]: time="2025-03-17T17:42:40.100976438Z" level=info msg="StopPodSandbox for \"a45255c5e842412af83dd06ba8b258ef311550cba3da5240506cd00cf0678895\"" Mar 17 17:42:40.101077 containerd[1467]: time="2025-03-17T17:42:40.101060836Z" level=info msg="TearDown network for sandbox \"a45255c5e842412af83dd06ba8b258ef311550cba3da5240506cd00cf0678895\" successfully" Mar 17 17:42:40.101077 containerd[1467]: time="2025-03-17T17:42:40.101074542Z" level=info msg="StopPodSandbox for \"a45255c5e842412af83dd06ba8b258ef311550cba3da5240506cd00cf0678895\" returns successfully" Mar 17 17:42:40.101462 containerd[1467]: time="2025-03-17T17:42:40.101442023Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86c8b6f988-79bzl,Uid:3265838f-27ca-489d-8c65-7cadce5cca40,Namespace:calico-apiserver,Attempt:5,}" Mar 17 17:42:40.101903 kubelet[2662]: I0317 17:42:40.101779 2662 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58f98dec59f83fef7090b948443c29832b1906274dd91e7d0c3b3c902d3df763" Mar 17 17:42:40.102166 containerd[1467]: time="2025-03-17T17:42:40.102144924Z" level=info msg="StopPodSandbox for \"58f98dec59f83fef7090b948443c29832b1906274dd91e7d0c3b3c902d3df763\"" Mar 17 17:42:40.102320 containerd[1467]: time="2025-03-17T17:42:40.102305025Z" level=info msg="Ensure that sandbox 58f98dec59f83fef7090b948443c29832b1906274dd91e7d0c3b3c902d3df763 in task-service has been cleanup successfully" Mar 17 17:42:40.102778 containerd[1467]: time="2025-03-17T17:42:40.102621690Z" level=info msg="TearDown network for sandbox \"58f98dec59f83fef7090b948443c29832b1906274dd91e7d0c3b3c902d3df763\" successfully" Mar 17 17:42:40.102778 containerd[1467]: time="2025-03-17T17:42:40.102635646Z" level=info msg="StopPodSandbox for \"58f98dec59f83fef7090b948443c29832b1906274dd91e7d0c3b3c902d3df763\" returns successfully" Mar 17 17:42:40.102847 containerd[1467]: time="2025-03-17T17:42:40.102810134Z" level=info msg="StopPodSandbox for \"1f73f191946516c6303f3e08ba9764f9ef1ed363435b102ed8343391428b48bc\"" Mar 17 17:42:40.102896 containerd[1467]: time="2025-03-17T17:42:40.102879725Z" level=info msg="TearDown network for sandbox \"1f73f191946516c6303f3e08ba9764f9ef1ed363435b102ed8343391428b48bc\" successfully" Mar 17 17:42:40.102896 containerd[1467]: time="2025-03-17T17:42:40.102894162Z" level=info msg="StopPodSandbox for \"1f73f191946516c6303f3e08ba9764f9ef1ed363435b102ed8343391428b48bc\" returns successfully" Mar 17 17:42:40.103116 containerd[1467]: time="2025-03-17T17:42:40.103088968Z" level=info msg="StopPodSandbox for \"eec0898e435269a73cab252710c67519bb65ec8c70880a65c0d84177e2d5547f\"" Mar 17 17:42:40.103236 containerd[1467]: time="2025-03-17T17:42:40.103219153Z" level=info msg="TearDown network for sandbox \"eec0898e435269a73cab252710c67519bb65ec8c70880a65c0d84177e2d5547f\" successfully" Mar 17 17:42:40.103236 containerd[1467]: time="2025-03-17T17:42:40.103233991Z" level=info msg="StopPodSandbox for \"eec0898e435269a73cab252710c67519bb65ec8c70880a65c0d84177e2d5547f\" returns successfully" Mar 17 17:42:40.103416 containerd[1467]: time="2025-03-17T17:42:40.103381228Z" level=info msg="StopPodSandbox for \"4044ac7223d418a825ef4eedf876ba4769d2804921eba6925b3d0d6d4afbc33a\"" Mar 17 17:42:40.103452 containerd[1467]: time="2025-03-17T17:42:40.103446781Z" level=info msg="TearDown network for sandbox \"4044ac7223d418a825ef4eedf876ba4769d2804921eba6925b3d0d6d4afbc33a\" successfully" Mar 17 17:42:40.103476 containerd[1467]: time="2025-03-17T17:42:40.103455137Z" level=info msg="StopPodSandbox for \"4044ac7223d418a825ef4eedf876ba4769d2804921eba6925b3d0d6d4afbc33a\" returns successfully" Mar 17 17:42:40.104717 containerd[1467]: time="2025-03-17T17:42:40.103775820Z" level=info msg="StopPodSandbox for \"3f5345f62ce3def9878f952bc3b9ba8515b066f140a947d652235852cb08fae5\"" Mar 17 17:42:40.104717 containerd[1467]: time="2025-03-17T17:42:40.103871008Z" level=info msg="TearDown network for sandbox \"3f5345f62ce3def9878f952bc3b9ba8515b066f140a947d652235852cb08fae5\" successfully" Mar 17 17:42:40.104717 containerd[1467]: time="2025-03-17T17:42:40.103881247Z" level=info msg="StopPodSandbox for \"3f5345f62ce3def9878f952bc3b9ba8515b066f140a947d652235852cb08fae5\" returns successfully" Mar 17 17:42:40.105137 containerd[1467]: time="2025-03-17T17:42:40.105103966Z" level=info msg="StopPodSandbox for \"e93aa7d450e66ea71243d70e3ef70d32837865b756a4bd00a6f296f08f7881d8\"" Mar 17 17:42:40.105211 containerd[1467]: time="2025-03-17T17:42:40.105193935Z" level=info msg="TearDown network for sandbox \"e93aa7d450e66ea71243d70e3ef70d32837865b756a4bd00a6f296f08f7881d8\" successfully" Mar 17 17:42:40.105211 containerd[1467]: time="2025-03-17T17:42:40.105207821Z" level=info msg="StopPodSandbox for \"e93aa7d450e66ea71243d70e3ef70d32837865b756a4bd00a6f296f08f7881d8\" returns successfully" Mar 17 17:42:40.106583 kubelet[2662]: I0317 17:42:40.106556 2662 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9906341f06aa2f7e95fd6b7d0f4548be4a6da1921703748eb9f0e054258d0bf4" Mar 17 17:42:40.106819 containerd[1467]: time="2025-03-17T17:42:40.106788011Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86c8b6f988-ppvkt,Uid:c9eab445-c078-490b-861c-b96d204afef3,Namespace:calico-apiserver,Attempt:6,}" Mar 17 17:42:40.106983 containerd[1467]: time="2025-03-17T17:42:40.106927513Z" level=info msg="StopPodSandbox for \"9906341f06aa2f7e95fd6b7d0f4548be4a6da1921703748eb9f0e054258d0bf4\"" Mar 17 17:42:40.107128 containerd[1467]: time="2025-03-17T17:42:40.107093035Z" level=info msg="Ensure that sandbox 9906341f06aa2f7e95fd6b7d0f4548be4a6da1921703748eb9f0e054258d0bf4 in task-service has been cleanup successfully" Mar 17 17:42:40.107514 containerd[1467]: time="2025-03-17T17:42:40.107495761Z" level=info msg="TearDown network for sandbox \"9906341f06aa2f7e95fd6b7d0f4548be4a6da1921703748eb9f0e054258d0bf4\" successfully" Mar 17 17:42:40.107514 containerd[1467]: time="2025-03-17T17:42:40.107511952Z" level=info msg="StopPodSandbox for \"9906341f06aa2f7e95fd6b7d0f4548be4a6da1921703748eb9f0e054258d0bf4\" returns successfully" Mar 17 17:42:40.107849 containerd[1467]: time="2025-03-17T17:42:40.107827254Z" level=info msg="StopPodSandbox for \"6dd83b0760f3d749a5709564d2613214f6b299b039d6f1b56dcb51a9cb223ea7\"" Mar 17 17:42:40.107965 containerd[1467]: time="2025-03-17T17:42:40.107949344Z" level=info msg="TearDown network for sandbox \"6dd83b0760f3d749a5709564d2613214f6b299b039d6f1b56dcb51a9cb223ea7\" successfully" Mar 17 17:42:40.108018 containerd[1467]: time="2025-03-17T17:42:40.107964352Z" level=info msg="StopPodSandbox for \"6dd83b0760f3d749a5709564d2613214f6b299b039d6f1b56dcb51a9cb223ea7\" returns successfully" Mar 17 17:42:40.108197 containerd[1467]: time="2025-03-17T17:42:40.108163626Z" level=info msg="StopPodSandbox for \"008cee49813c10fd33682eb193a75ecd740653b4fd1a4a322bfcc088c6078101\"" Mar 17 17:42:40.108254 containerd[1467]: time="2025-03-17T17:42:40.108241493Z" level=info msg="TearDown network for sandbox \"008cee49813c10fd33682eb193a75ecd740653b4fd1a4a322bfcc088c6078101\" successfully" Mar 17 17:42:40.108276 containerd[1467]: time="2025-03-17T17:42:40.108253556Z" level=info msg="StopPodSandbox for \"008cee49813c10fd33682eb193a75ecd740653b4fd1a4a322bfcc088c6078101\" returns successfully" Mar 17 17:42:40.108470 containerd[1467]: time="2025-03-17T17:42:40.108452870Z" level=info msg="StopPodSandbox for \"3edf99fd7e4c4bcd608727ff216f57eed4c9785e1d1f4d8c623f0c2aa6ca839a\"" Mar 17 17:42:40.108537 containerd[1467]: time="2025-03-17T17:42:40.108522462Z" level=info msg="TearDown network for sandbox \"3edf99fd7e4c4bcd608727ff216f57eed4c9785e1d1f4d8c623f0c2aa6ca839a\" successfully" Mar 17 17:42:40.108537 containerd[1467]: time="2025-03-17T17:42:40.108534364Z" level=info msg="StopPodSandbox for \"3edf99fd7e4c4bcd608727ff216f57eed4c9785e1d1f4d8c623f0c2aa6ca839a\" returns successfully" Mar 17 17:42:40.108875 containerd[1467]: time="2025-03-17T17:42:40.108715153Z" level=info msg="StopPodSandbox for \"f1d105acdc9b53357e4e621942df86a7fba22a3900a30edbb033cb1322dbced2\"" Mar 17 17:42:40.108875 containerd[1467]: time="2025-03-17T17:42:40.108795284Z" level=info msg="TearDown network for sandbox \"f1d105acdc9b53357e4e621942df86a7fba22a3900a30edbb033cb1322dbced2\" successfully" Mar 17 17:42:40.108875 containerd[1467]: time="2025-03-17T17:42:40.108815743Z" level=info msg="StopPodSandbox for \"f1d105acdc9b53357e4e621942df86a7fba22a3900a30edbb033cb1322dbced2\" returns successfully" Mar 17 17:42:40.109296 containerd[1467]: time="2025-03-17T17:42:40.109095789Z" level=info msg="StopPodSandbox for \"e983d9d5aee8c472a9806708fa057e3b2de9217794b9f28b9fc2133b98942f76\"" Mar 17 17:42:40.109296 containerd[1467]: time="2025-03-17T17:42:40.109221295Z" level=info msg="TearDown network for sandbox \"e983d9d5aee8c472a9806708fa057e3b2de9217794b9f28b9fc2133b98942f76\" successfully" Mar 17 17:42:40.109296 containerd[1467]: time="2025-03-17T17:42:40.109231263Z" level=info msg="StopPodSandbox for \"e983d9d5aee8c472a9806708fa057e3b2de9217794b9f28b9fc2133b98942f76\" returns successfully" Mar 17 17:42:40.109600 kubelet[2662]: E0317 17:42:40.109575 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:40.128635 containerd[1467]: time="2025-03-17T17:42:40.128584830Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-sjrwq,Uid:228c6093-6dc2-4faf-8ec8-87849ef402ab,Namespace:kube-system,Attempt:6,}" Mar 17 17:42:40.348196 systemd-networkd[1410]: cali0e24b01453e: Link UP Mar 17 17:42:40.349513 systemd-networkd[1410]: cali0e24b01453e: Gained carrier Mar 17 17:42:40.360260 kubelet[2662]: I0317 17:42:40.359894 2662 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-fbl5x" podStartSLOduration=2.199042134 podStartE2EDuration="18.359875981s" podCreationTimestamp="2025-03-17 17:42:22 +0000 UTC" firstStartedPulling="2025-03-17 17:42:23.135731158 +0000 UTC m=+21.395697884" lastFinishedPulling="2025-03-17 17:42:39.296565005 +0000 UTC m=+37.556531731" observedRunningTime="2025-03-17 17:42:40.106426101 +0000 UTC m=+38.366392827" watchObservedRunningTime="2025-03-17 17:42:40.359875981 +0000 UTC m=+38.619842707" Mar 17 17:42:40.367089 containerd[1467]: 2025-03-17 17:42:40.210 [INFO][4816] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Mar 17 17:42:40.367089 containerd[1467]: 2025-03-17 17:42:40.225 [INFO][4816] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--7db6d8ff4d--sjrwq-eth0 coredns-7db6d8ff4d- kube-system 228c6093-6dc2-4faf-8ec8-87849ef402ab 730 0 2025-03-17 17:42:16 +0000 UTC map[k8s-app:kube-dns pod-template-hash:7db6d8ff4d projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-7db6d8ff4d-sjrwq eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali0e24b01453e [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="854e88e2deedcc07c9c8cc51cb5c96cb4732e3e8f7e0a0a40aebc2bc877f541f" Namespace="kube-system" Pod="coredns-7db6d8ff4d-sjrwq" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--sjrwq-" Mar 17 17:42:40.367089 containerd[1467]: 2025-03-17 17:42:40.225 [INFO][4816] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="854e88e2deedcc07c9c8cc51cb5c96cb4732e3e8f7e0a0a40aebc2bc877f541f" Namespace="kube-system" Pod="coredns-7db6d8ff4d-sjrwq" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--sjrwq-eth0" Mar 17 17:42:40.367089 containerd[1467]: 2025-03-17 17:42:40.291 [INFO][4850] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="854e88e2deedcc07c9c8cc51cb5c96cb4732e3e8f7e0a0a40aebc2bc877f541f" HandleID="k8s-pod-network.854e88e2deedcc07c9c8cc51cb5c96cb4732e3e8f7e0a0a40aebc2bc877f541f" Workload="localhost-k8s-coredns--7db6d8ff4d--sjrwq-eth0" Mar 17 17:42:40.367089 containerd[1467]: 2025-03-17 17:42:40.305 [INFO][4850] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="854e88e2deedcc07c9c8cc51cb5c96cb4732e3e8f7e0a0a40aebc2bc877f541f" HandleID="k8s-pod-network.854e88e2deedcc07c9c8cc51cb5c96cb4732e3e8f7e0a0a40aebc2bc877f541f" Workload="localhost-k8s-coredns--7db6d8ff4d--sjrwq-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0003957f0), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-7db6d8ff4d-sjrwq", "timestamp":"2025-03-17 17:42:40.291150145 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Mar 17 17:42:40.367089 containerd[1467]: 2025-03-17 17:42:40.306 [INFO][4850] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Mar 17 17:42:40.367089 containerd[1467]: 2025-03-17 17:42:40.306 [INFO][4850] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Mar 17 17:42:40.367089 containerd[1467]: 2025-03-17 17:42:40.306 [INFO][4850] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Mar 17 17:42:40.367089 containerd[1467]: 2025-03-17 17:42:40.309 [INFO][4850] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.854e88e2deedcc07c9c8cc51cb5c96cb4732e3e8f7e0a0a40aebc2bc877f541f" host="localhost" Mar 17 17:42:40.367089 containerd[1467]: 2025-03-17 17:42:40.318 [INFO][4850] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" Mar 17 17:42:40.367089 containerd[1467]: 2025-03-17 17:42:40.322 [INFO][4850] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Mar 17 17:42:40.367089 containerd[1467]: 2025-03-17 17:42:40.324 [INFO][4850] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Mar 17 17:42:40.367089 containerd[1467]: 2025-03-17 17:42:40.326 [INFO][4850] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Mar 17 17:42:40.367089 containerd[1467]: 2025-03-17 17:42:40.326 [INFO][4850] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.854e88e2deedcc07c9c8cc51cb5c96cb4732e3e8f7e0a0a40aebc2bc877f541f" host="localhost" Mar 17 17:42:40.367089 containerd[1467]: 2025-03-17 17:42:40.327 [INFO][4850] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.854e88e2deedcc07c9c8cc51cb5c96cb4732e3e8f7e0a0a40aebc2bc877f541f Mar 17 17:42:40.367089 containerd[1467]: 2025-03-17 17:42:40.330 [INFO][4850] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.854e88e2deedcc07c9c8cc51cb5c96cb4732e3e8f7e0a0a40aebc2bc877f541f" host="localhost" Mar 17 17:42:40.367089 containerd[1467]: 2025-03-17 17:42:40.336 [INFO][4850] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.129/26] block=192.168.88.128/26 handle="k8s-pod-network.854e88e2deedcc07c9c8cc51cb5c96cb4732e3e8f7e0a0a40aebc2bc877f541f" host="localhost" Mar 17 17:42:40.367089 containerd[1467]: 2025-03-17 17:42:40.336 [INFO][4850] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.129/26] handle="k8s-pod-network.854e88e2deedcc07c9c8cc51cb5c96cb4732e3e8f7e0a0a40aebc2bc877f541f" host="localhost" Mar 17 17:42:40.367089 containerd[1467]: 2025-03-17 17:42:40.336 [INFO][4850] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Mar 17 17:42:40.367089 containerd[1467]: 2025-03-17 17:42:40.336 [INFO][4850] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.129/26] IPv6=[] ContainerID="854e88e2deedcc07c9c8cc51cb5c96cb4732e3e8f7e0a0a40aebc2bc877f541f" HandleID="k8s-pod-network.854e88e2deedcc07c9c8cc51cb5c96cb4732e3e8f7e0a0a40aebc2bc877f541f" Workload="localhost-k8s-coredns--7db6d8ff4d--sjrwq-eth0" Mar 17 17:42:40.368265 containerd[1467]: 2025-03-17 17:42:40.340 [INFO][4816] cni-plugin/k8s.go 386: Populated endpoint ContainerID="854e88e2deedcc07c9c8cc51cb5c96cb4732e3e8f7e0a0a40aebc2bc877f541f" Namespace="kube-system" Pod="coredns-7db6d8ff4d-sjrwq" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--sjrwq-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7db6d8ff4d--sjrwq-eth0", GenerateName:"coredns-7db6d8ff4d-", Namespace:"kube-system", SelfLink:"", UID:"228c6093-6dc2-4faf-8ec8-87849ef402ab", ResourceVersion:"730", Generation:0, CreationTimestamp:time.Date(2025, time.March, 17, 17, 42, 16, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7db6d8ff4d", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-7db6d8ff4d-sjrwq", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali0e24b01453e", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Mar 17 17:42:40.368265 containerd[1467]: 2025-03-17 17:42:40.340 [INFO][4816] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.129/32] ContainerID="854e88e2deedcc07c9c8cc51cb5c96cb4732e3e8f7e0a0a40aebc2bc877f541f" Namespace="kube-system" Pod="coredns-7db6d8ff4d-sjrwq" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--sjrwq-eth0" Mar 17 17:42:40.368265 containerd[1467]: 2025-03-17 17:42:40.340 [INFO][4816] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali0e24b01453e ContainerID="854e88e2deedcc07c9c8cc51cb5c96cb4732e3e8f7e0a0a40aebc2bc877f541f" Namespace="kube-system" Pod="coredns-7db6d8ff4d-sjrwq" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--sjrwq-eth0" Mar 17 17:42:40.368265 containerd[1467]: 2025-03-17 17:42:40.350 [INFO][4816] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="854e88e2deedcc07c9c8cc51cb5c96cb4732e3e8f7e0a0a40aebc2bc877f541f" Namespace="kube-system" Pod="coredns-7db6d8ff4d-sjrwq" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--sjrwq-eth0" Mar 17 17:42:40.368265 containerd[1467]: 2025-03-17 17:42:40.350 [INFO][4816] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="854e88e2deedcc07c9c8cc51cb5c96cb4732e3e8f7e0a0a40aebc2bc877f541f" Namespace="kube-system" Pod="coredns-7db6d8ff4d-sjrwq" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--sjrwq-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7db6d8ff4d--sjrwq-eth0", GenerateName:"coredns-7db6d8ff4d-", Namespace:"kube-system", SelfLink:"", UID:"228c6093-6dc2-4faf-8ec8-87849ef402ab", ResourceVersion:"730", Generation:0, CreationTimestamp:time.Date(2025, time.March, 17, 17, 42, 16, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7db6d8ff4d", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"854e88e2deedcc07c9c8cc51cb5c96cb4732e3e8f7e0a0a40aebc2bc877f541f", Pod:"coredns-7db6d8ff4d-sjrwq", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali0e24b01453e", MAC:"02:2e:93:00:8e:78", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Mar 17 17:42:40.368265 containerd[1467]: 2025-03-17 17:42:40.363 [INFO][4816] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="854e88e2deedcc07c9c8cc51cb5c96cb4732e3e8f7e0a0a40aebc2bc877f541f" Namespace="kube-system" Pod="coredns-7db6d8ff4d-sjrwq" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--sjrwq-eth0" Mar 17 17:42:40.371626 systemd-networkd[1410]: cali47e18c80f12: Link UP Mar 17 17:42:40.371964 systemd-networkd[1410]: cali47e18c80f12: Gained carrier Mar 17 17:42:40.386636 containerd[1467]: 2025-03-17 17:42:40.189 [INFO][4761] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Mar 17 17:42:40.386636 containerd[1467]: 2025-03-17 17:42:40.204 [INFO][4761] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-csi--node--driver--j9gdg-eth0 csi-node-driver- calico-system 27eb0c6c-40f4-4b19-b513-0003d63df5d0 603 0 2025-03-17 17:42:22 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:69ddf5d45d k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s localhost csi-node-driver-j9gdg eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali47e18c80f12 [] []}} ContainerID="b5375e4dccdc52e071442404c4b8a1d48eb22ca0db08a28f6918d5cd419f96d4" Namespace="calico-system" Pod="csi-node-driver-j9gdg" WorkloadEndpoint="localhost-k8s-csi--node--driver--j9gdg-" Mar 17 17:42:40.386636 containerd[1467]: 2025-03-17 17:42:40.204 [INFO][4761] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="b5375e4dccdc52e071442404c4b8a1d48eb22ca0db08a28f6918d5cd419f96d4" Namespace="calico-system" Pod="csi-node-driver-j9gdg" WorkloadEndpoint="localhost-k8s-csi--node--driver--j9gdg-eth0" Mar 17 17:42:40.386636 containerd[1467]: 2025-03-17 17:42:40.291 [INFO][4832] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="b5375e4dccdc52e071442404c4b8a1d48eb22ca0db08a28f6918d5cd419f96d4" HandleID="k8s-pod-network.b5375e4dccdc52e071442404c4b8a1d48eb22ca0db08a28f6918d5cd419f96d4" Workload="localhost-k8s-csi--node--driver--j9gdg-eth0" Mar 17 17:42:40.386636 containerd[1467]: 2025-03-17 17:42:40.306 [INFO][4832] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="b5375e4dccdc52e071442404c4b8a1d48eb22ca0db08a28f6918d5cd419f96d4" HandleID="k8s-pod-network.b5375e4dccdc52e071442404c4b8a1d48eb22ca0db08a28f6918d5cd419f96d4" Workload="localhost-k8s-csi--node--driver--j9gdg-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000311440), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"csi-node-driver-j9gdg", "timestamp":"2025-03-17 17:42:40.291275901 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Mar 17 17:42:40.386636 containerd[1467]: 2025-03-17 17:42:40.306 [INFO][4832] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Mar 17 17:42:40.386636 containerd[1467]: 2025-03-17 17:42:40.336 [INFO][4832] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Mar 17 17:42:40.386636 containerd[1467]: 2025-03-17 17:42:40.336 [INFO][4832] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Mar 17 17:42:40.386636 containerd[1467]: 2025-03-17 17:42:40.338 [INFO][4832] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.b5375e4dccdc52e071442404c4b8a1d48eb22ca0db08a28f6918d5cd419f96d4" host="localhost" Mar 17 17:42:40.386636 containerd[1467]: 2025-03-17 17:42:40.342 [INFO][4832] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" Mar 17 17:42:40.386636 containerd[1467]: 2025-03-17 17:42:40.345 [INFO][4832] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Mar 17 17:42:40.386636 containerd[1467]: 2025-03-17 17:42:40.347 [INFO][4832] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Mar 17 17:42:40.386636 containerd[1467]: 2025-03-17 17:42:40.349 [INFO][4832] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Mar 17 17:42:40.386636 containerd[1467]: 2025-03-17 17:42:40.349 [INFO][4832] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.b5375e4dccdc52e071442404c4b8a1d48eb22ca0db08a28f6918d5cd419f96d4" host="localhost" Mar 17 17:42:40.386636 containerd[1467]: 2025-03-17 17:42:40.351 [INFO][4832] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.b5375e4dccdc52e071442404c4b8a1d48eb22ca0db08a28f6918d5cd419f96d4 Mar 17 17:42:40.386636 containerd[1467]: 2025-03-17 17:42:40.356 [INFO][4832] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.b5375e4dccdc52e071442404c4b8a1d48eb22ca0db08a28f6918d5cd419f96d4" host="localhost" Mar 17 17:42:40.386636 containerd[1467]: 2025-03-17 17:42:40.363 [INFO][4832] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.130/26] block=192.168.88.128/26 handle="k8s-pod-network.b5375e4dccdc52e071442404c4b8a1d48eb22ca0db08a28f6918d5cd419f96d4" host="localhost" Mar 17 17:42:40.386636 containerd[1467]: 2025-03-17 17:42:40.363 [INFO][4832] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.130/26] handle="k8s-pod-network.b5375e4dccdc52e071442404c4b8a1d48eb22ca0db08a28f6918d5cd419f96d4" host="localhost" Mar 17 17:42:40.386636 containerd[1467]: 2025-03-17 17:42:40.363 [INFO][4832] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Mar 17 17:42:40.386636 containerd[1467]: 2025-03-17 17:42:40.363 [INFO][4832] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.130/26] IPv6=[] ContainerID="b5375e4dccdc52e071442404c4b8a1d48eb22ca0db08a28f6918d5cd419f96d4" HandleID="k8s-pod-network.b5375e4dccdc52e071442404c4b8a1d48eb22ca0db08a28f6918d5cd419f96d4" Workload="localhost-k8s-csi--node--driver--j9gdg-eth0" Mar 17 17:42:40.388073 containerd[1467]: 2025-03-17 17:42:40.369 [INFO][4761] cni-plugin/k8s.go 386: Populated endpoint ContainerID="b5375e4dccdc52e071442404c4b8a1d48eb22ca0db08a28f6918d5cd419f96d4" Namespace="calico-system" Pod="csi-node-driver-j9gdg" WorkloadEndpoint="localhost-k8s-csi--node--driver--j9gdg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--j9gdg-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"27eb0c6c-40f4-4b19-b513-0003d63df5d0", ResourceVersion:"603", Generation:0, CreationTimestamp:time.Date(2025, time.March, 17, 17, 42, 22, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"69ddf5d45d", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"csi-node-driver-j9gdg", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali47e18c80f12", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Mar 17 17:42:40.388073 containerd[1467]: 2025-03-17 17:42:40.369 [INFO][4761] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.130/32] ContainerID="b5375e4dccdc52e071442404c4b8a1d48eb22ca0db08a28f6918d5cd419f96d4" Namespace="calico-system" Pod="csi-node-driver-j9gdg" WorkloadEndpoint="localhost-k8s-csi--node--driver--j9gdg-eth0" Mar 17 17:42:40.388073 containerd[1467]: 2025-03-17 17:42:40.369 [INFO][4761] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali47e18c80f12 ContainerID="b5375e4dccdc52e071442404c4b8a1d48eb22ca0db08a28f6918d5cd419f96d4" Namespace="calico-system" Pod="csi-node-driver-j9gdg" WorkloadEndpoint="localhost-k8s-csi--node--driver--j9gdg-eth0" Mar 17 17:42:40.388073 containerd[1467]: 2025-03-17 17:42:40.371 [INFO][4761] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="b5375e4dccdc52e071442404c4b8a1d48eb22ca0db08a28f6918d5cd419f96d4" Namespace="calico-system" Pod="csi-node-driver-j9gdg" WorkloadEndpoint="localhost-k8s-csi--node--driver--j9gdg-eth0" Mar 17 17:42:40.388073 containerd[1467]: 2025-03-17 17:42:40.372 [INFO][4761] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="b5375e4dccdc52e071442404c4b8a1d48eb22ca0db08a28f6918d5cd419f96d4" Namespace="calico-system" Pod="csi-node-driver-j9gdg" WorkloadEndpoint="localhost-k8s-csi--node--driver--j9gdg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--j9gdg-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"27eb0c6c-40f4-4b19-b513-0003d63df5d0", ResourceVersion:"603", Generation:0, CreationTimestamp:time.Date(2025, time.March, 17, 17, 42, 22, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"69ddf5d45d", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"b5375e4dccdc52e071442404c4b8a1d48eb22ca0db08a28f6918d5cd419f96d4", Pod:"csi-node-driver-j9gdg", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali47e18c80f12", MAC:"8e:9e:8f:af:36:e4", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Mar 17 17:42:40.388073 containerd[1467]: 2025-03-17 17:42:40.382 [INFO][4761] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="b5375e4dccdc52e071442404c4b8a1d48eb22ca0db08a28f6918d5cd419f96d4" Namespace="calico-system" Pod="csi-node-driver-j9gdg" WorkloadEndpoint="localhost-k8s-csi--node--driver--j9gdg-eth0" Mar 17 17:42:40.401264 systemd-networkd[1410]: calice8227e9e30: Link UP Mar 17 17:42:40.401524 systemd-networkd[1410]: calice8227e9e30: Gained carrier Mar 17 17:42:40.429580 containerd[1467]: 2025-03-17 17:42:40.216 [INFO][4783] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Mar 17 17:42:40.429580 containerd[1467]: 2025-03-17 17:42:40.231 [INFO][4783] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--86c8b6f988--79bzl-eth0 calico-apiserver-86c8b6f988- calico-apiserver 3265838f-27ca-489d-8c65-7cadce5cca40 733 0 2025-03-17 17:42:22 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:86c8b6f988 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-86c8b6f988-79bzl eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] calice8227e9e30 [] []}} ContainerID="0d14e669762aafcce27fbf8489157e7e276641f1503453d7643a41b088ea7d19" Namespace="calico-apiserver" Pod="calico-apiserver-86c8b6f988-79bzl" WorkloadEndpoint="localhost-k8s-calico--apiserver--86c8b6f988--79bzl-" Mar 17 17:42:40.429580 containerd[1467]: 2025-03-17 17:42:40.231 [INFO][4783] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="0d14e669762aafcce27fbf8489157e7e276641f1503453d7643a41b088ea7d19" Namespace="calico-apiserver" Pod="calico-apiserver-86c8b6f988-79bzl" WorkloadEndpoint="localhost-k8s-calico--apiserver--86c8b6f988--79bzl-eth0" Mar 17 17:42:40.429580 containerd[1467]: 2025-03-17 17:42:40.291 [INFO][4866] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="0d14e669762aafcce27fbf8489157e7e276641f1503453d7643a41b088ea7d19" HandleID="k8s-pod-network.0d14e669762aafcce27fbf8489157e7e276641f1503453d7643a41b088ea7d19" Workload="localhost-k8s-calico--apiserver--86c8b6f988--79bzl-eth0" Mar 17 17:42:40.429580 containerd[1467]: 2025-03-17 17:42:40.309 [INFO][4866] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="0d14e669762aafcce27fbf8489157e7e276641f1503453d7643a41b088ea7d19" HandleID="k8s-pod-network.0d14e669762aafcce27fbf8489157e7e276641f1503453d7643a41b088ea7d19" Workload="localhost-k8s-calico--apiserver--86c8b6f988--79bzl-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000362b60), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-86c8b6f988-79bzl", "timestamp":"2025-03-17 17:42:40.291836955 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Mar 17 17:42:40.429580 containerd[1467]: 2025-03-17 17:42:40.309 [INFO][4866] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Mar 17 17:42:40.429580 containerd[1467]: 2025-03-17 17:42:40.363 [INFO][4866] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Mar 17 17:42:40.429580 containerd[1467]: 2025-03-17 17:42:40.363 [INFO][4866] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Mar 17 17:42:40.429580 containerd[1467]: 2025-03-17 17:42:40.365 [INFO][4866] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.0d14e669762aafcce27fbf8489157e7e276641f1503453d7643a41b088ea7d19" host="localhost" Mar 17 17:42:40.429580 containerd[1467]: 2025-03-17 17:42:40.370 [INFO][4866] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" Mar 17 17:42:40.429580 containerd[1467]: 2025-03-17 17:42:40.375 [INFO][4866] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Mar 17 17:42:40.429580 containerd[1467]: 2025-03-17 17:42:40.380 [INFO][4866] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Mar 17 17:42:40.429580 containerd[1467]: 2025-03-17 17:42:40.383 [INFO][4866] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Mar 17 17:42:40.429580 containerd[1467]: 2025-03-17 17:42:40.383 [INFO][4866] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.0d14e669762aafcce27fbf8489157e7e276641f1503453d7643a41b088ea7d19" host="localhost" Mar 17 17:42:40.429580 containerd[1467]: 2025-03-17 17:42:40.386 [INFO][4866] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.0d14e669762aafcce27fbf8489157e7e276641f1503453d7643a41b088ea7d19 Mar 17 17:42:40.429580 containerd[1467]: 2025-03-17 17:42:40.390 [INFO][4866] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.0d14e669762aafcce27fbf8489157e7e276641f1503453d7643a41b088ea7d19" host="localhost" Mar 17 17:42:40.429580 containerd[1467]: 2025-03-17 17:42:40.394 [INFO][4866] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.131/26] block=192.168.88.128/26 handle="k8s-pod-network.0d14e669762aafcce27fbf8489157e7e276641f1503453d7643a41b088ea7d19" host="localhost" Mar 17 17:42:40.429580 containerd[1467]: 2025-03-17 17:42:40.394 [INFO][4866] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.131/26] handle="k8s-pod-network.0d14e669762aafcce27fbf8489157e7e276641f1503453d7643a41b088ea7d19" host="localhost" Mar 17 17:42:40.429580 containerd[1467]: 2025-03-17 17:42:40.395 [INFO][4866] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Mar 17 17:42:40.429580 containerd[1467]: 2025-03-17 17:42:40.395 [INFO][4866] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.131/26] IPv6=[] ContainerID="0d14e669762aafcce27fbf8489157e7e276641f1503453d7643a41b088ea7d19" HandleID="k8s-pod-network.0d14e669762aafcce27fbf8489157e7e276641f1503453d7643a41b088ea7d19" Workload="localhost-k8s-calico--apiserver--86c8b6f988--79bzl-eth0" Mar 17 17:42:40.430236 containerd[1467]: 2025-03-17 17:42:40.399 [INFO][4783] cni-plugin/k8s.go 386: Populated endpoint ContainerID="0d14e669762aafcce27fbf8489157e7e276641f1503453d7643a41b088ea7d19" Namespace="calico-apiserver" Pod="calico-apiserver-86c8b6f988-79bzl" WorkloadEndpoint="localhost-k8s-calico--apiserver--86c8b6f988--79bzl-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--86c8b6f988--79bzl-eth0", GenerateName:"calico-apiserver-86c8b6f988-", Namespace:"calico-apiserver", SelfLink:"", UID:"3265838f-27ca-489d-8c65-7cadce5cca40", ResourceVersion:"733", Generation:0, CreationTimestamp:time.Date(2025, time.March, 17, 17, 42, 22, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"86c8b6f988", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-86c8b6f988-79bzl", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calice8227e9e30", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Mar 17 17:42:40.430236 containerd[1467]: 2025-03-17 17:42:40.399 [INFO][4783] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.131/32] ContainerID="0d14e669762aafcce27fbf8489157e7e276641f1503453d7643a41b088ea7d19" Namespace="calico-apiserver" Pod="calico-apiserver-86c8b6f988-79bzl" WorkloadEndpoint="localhost-k8s-calico--apiserver--86c8b6f988--79bzl-eth0" Mar 17 17:42:40.430236 containerd[1467]: 2025-03-17 17:42:40.399 [INFO][4783] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calice8227e9e30 ContainerID="0d14e669762aafcce27fbf8489157e7e276641f1503453d7643a41b088ea7d19" Namespace="calico-apiserver" Pod="calico-apiserver-86c8b6f988-79bzl" WorkloadEndpoint="localhost-k8s-calico--apiserver--86c8b6f988--79bzl-eth0" Mar 17 17:42:40.430236 containerd[1467]: 2025-03-17 17:42:40.401 [INFO][4783] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="0d14e669762aafcce27fbf8489157e7e276641f1503453d7643a41b088ea7d19" Namespace="calico-apiserver" Pod="calico-apiserver-86c8b6f988-79bzl" WorkloadEndpoint="localhost-k8s-calico--apiserver--86c8b6f988--79bzl-eth0" Mar 17 17:42:40.430236 containerd[1467]: 2025-03-17 17:42:40.402 [INFO][4783] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="0d14e669762aafcce27fbf8489157e7e276641f1503453d7643a41b088ea7d19" Namespace="calico-apiserver" Pod="calico-apiserver-86c8b6f988-79bzl" WorkloadEndpoint="localhost-k8s-calico--apiserver--86c8b6f988--79bzl-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--86c8b6f988--79bzl-eth0", GenerateName:"calico-apiserver-86c8b6f988-", Namespace:"calico-apiserver", SelfLink:"", UID:"3265838f-27ca-489d-8c65-7cadce5cca40", ResourceVersion:"733", Generation:0, CreationTimestamp:time.Date(2025, time.March, 17, 17, 42, 22, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"86c8b6f988", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"0d14e669762aafcce27fbf8489157e7e276641f1503453d7643a41b088ea7d19", Pod:"calico-apiserver-86c8b6f988-79bzl", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calice8227e9e30", MAC:"fa:67:23:e2:8e:0d", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Mar 17 17:42:40.430236 containerd[1467]: 2025-03-17 17:42:40.426 [INFO][4783] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="0d14e669762aafcce27fbf8489157e7e276641f1503453d7643a41b088ea7d19" Namespace="calico-apiserver" Pod="calico-apiserver-86c8b6f988-79bzl" WorkloadEndpoint="localhost-k8s-calico--apiserver--86c8b6f988--79bzl-eth0" Mar 17 17:42:40.497301 systemd-networkd[1410]: caliba0c337e231: Link UP Mar 17 17:42:40.499256 systemd-networkd[1410]: caliba0c337e231: Gained carrier Mar 17 17:42:40.503012 containerd[1467]: time="2025-03-17T17:42:40.502661578Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 17:42:40.503012 containerd[1467]: time="2025-03-17T17:42:40.502783106Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 17:42:40.503012 containerd[1467]: time="2025-03-17T17:42:40.502802783Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:42:40.507783 containerd[1467]: time="2025-03-17T17:42:40.507699317Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:42:40.511629 containerd[1467]: time="2025-03-17T17:42:40.511348556Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 17:42:40.511804 containerd[1467]: time="2025-03-17T17:42:40.511635565Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 17:42:40.511804 containerd[1467]: time="2025-03-17T17:42:40.511675681Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:42:40.512048 containerd[1467]: time="2025-03-17T17:42:40.511885224Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:42:40.523241 containerd[1467]: 2025-03-17 17:42:40.171 [INFO][4750] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Mar 17 17:42:40.523241 containerd[1467]: 2025-03-17 17:42:40.204 [INFO][4750] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--7db6d8ff4d--fw5hr-eth0 coredns-7db6d8ff4d- kube-system ae45280b-f87f-4410-991d-e0beb4bdc7c6 727 0 2025-03-17 17:42:16 +0000 UTC map[k8s-app:kube-dns pod-template-hash:7db6d8ff4d projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-7db6d8ff4d-fw5hr eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] caliba0c337e231 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="85b06febbb19bd0a58aa28066f2570b6de7ab545b4046bee07d08d94ee9ccc38" Namespace="kube-system" Pod="coredns-7db6d8ff4d-fw5hr" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--fw5hr-" Mar 17 17:42:40.523241 containerd[1467]: 2025-03-17 17:42:40.204 [INFO][4750] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="85b06febbb19bd0a58aa28066f2570b6de7ab545b4046bee07d08d94ee9ccc38" Namespace="kube-system" Pod="coredns-7db6d8ff4d-fw5hr" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--fw5hr-eth0" Mar 17 17:42:40.523241 containerd[1467]: 2025-03-17 17:42:40.297 [INFO][4834] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="85b06febbb19bd0a58aa28066f2570b6de7ab545b4046bee07d08d94ee9ccc38" HandleID="k8s-pod-network.85b06febbb19bd0a58aa28066f2570b6de7ab545b4046bee07d08d94ee9ccc38" Workload="localhost-k8s-coredns--7db6d8ff4d--fw5hr-eth0" Mar 17 17:42:40.523241 containerd[1467]: 2025-03-17 17:42:40.317 [INFO][4834] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="85b06febbb19bd0a58aa28066f2570b6de7ab545b4046bee07d08d94ee9ccc38" HandleID="k8s-pod-network.85b06febbb19bd0a58aa28066f2570b6de7ab545b4046bee07d08d94ee9ccc38" Workload="localhost-k8s-coredns--7db6d8ff4d--fw5hr-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000366980), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-7db6d8ff4d-fw5hr", "timestamp":"2025-03-17 17:42:40.297475894 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Mar 17 17:42:40.523241 containerd[1467]: 2025-03-17 17:42:40.317 [INFO][4834] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Mar 17 17:42:40.523241 containerd[1467]: 2025-03-17 17:42:40.395 [INFO][4834] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Mar 17 17:42:40.523241 containerd[1467]: 2025-03-17 17:42:40.395 [INFO][4834] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Mar 17 17:42:40.523241 containerd[1467]: 2025-03-17 17:42:40.397 [INFO][4834] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.85b06febbb19bd0a58aa28066f2570b6de7ab545b4046bee07d08d94ee9ccc38" host="localhost" Mar 17 17:42:40.523241 containerd[1467]: 2025-03-17 17:42:40.401 [INFO][4834] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" Mar 17 17:42:40.523241 containerd[1467]: 2025-03-17 17:42:40.406 [INFO][4834] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Mar 17 17:42:40.523241 containerd[1467]: 2025-03-17 17:42:40.408 [INFO][4834] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Mar 17 17:42:40.523241 containerd[1467]: 2025-03-17 17:42:40.426 [INFO][4834] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Mar 17 17:42:40.523241 containerd[1467]: 2025-03-17 17:42:40.426 [INFO][4834] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.85b06febbb19bd0a58aa28066f2570b6de7ab545b4046bee07d08d94ee9ccc38" host="localhost" Mar 17 17:42:40.523241 containerd[1467]: 2025-03-17 17:42:40.428 [INFO][4834] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.85b06febbb19bd0a58aa28066f2570b6de7ab545b4046bee07d08d94ee9ccc38 Mar 17 17:42:40.523241 containerd[1467]: 2025-03-17 17:42:40.435 [INFO][4834] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.85b06febbb19bd0a58aa28066f2570b6de7ab545b4046bee07d08d94ee9ccc38" host="localhost" Mar 17 17:42:40.523241 containerd[1467]: 2025-03-17 17:42:40.484 [INFO][4834] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.132/26] block=192.168.88.128/26 handle="k8s-pod-network.85b06febbb19bd0a58aa28066f2570b6de7ab545b4046bee07d08d94ee9ccc38" host="localhost" Mar 17 17:42:40.523241 containerd[1467]: 2025-03-17 17:42:40.484 [INFO][4834] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.132/26] handle="k8s-pod-network.85b06febbb19bd0a58aa28066f2570b6de7ab545b4046bee07d08d94ee9ccc38" host="localhost" Mar 17 17:42:40.523241 containerd[1467]: 2025-03-17 17:42:40.484 [INFO][4834] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Mar 17 17:42:40.523241 containerd[1467]: 2025-03-17 17:42:40.484 [INFO][4834] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.132/26] IPv6=[] ContainerID="85b06febbb19bd0a58aa28066f2570b6de7ab545b4046bee07d08d94ee9ccc38" HandleID="k8s-pod-network.85b06febbb19bd0a58aa28066f2570b6de7ab545b4046bee07d08d94ee9ccc38" Workload="localhost-k8s-coredns--7db6d8ff4d--fw5hr-eth0" Mar 17 17:42:40.524200 containerd[1467]: 2025-03-17 17:42:40.488 [INFO][4750] cni-plugin/k8s.go 386: Populated endpoint ContainerID="85b06febbb19bd0a58aa28066f2570b6de7ab545b4046bee07d08d94ee9ccc38" Namespace="kube-system" Pod="coredns-7db6d8ff4d-fw5hr" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--fw5hr-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7db6d8ff4d--fw5hr-eth0", GenerateName:"coredns-7db6d8ff4d-", Namespace:"kube-system", SelfLink:"", UID:"ae45280b-f87f-4410-991d-e0beb4bdc7c6", ResourceVersion:"727", Generation:0, CreationTimestamp:time.Date(2025, time.March, 17, 17, 42, 16, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7db6d8ff4d", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-7db6d8ff4d-fw5hr", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"caliba0c337e231", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Mar 17 17:42:40.524200 containerd[1467]: 2025-03-17 17:42:40.488 [INFO][4750] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.132/32] ContainerID="85b06febbb19bd0a58aa28066f2570b6de7ab545b4046bee07d08d94ee9ccc38" Namespace="kube-system" Pod="coredns-7db6d8ff4d-fw5hr" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--fw5hr-eth0" Mar 17 17:42:40.524200 containerd[1467]: 2025-03-17 17:42:40.488 [INFO][4750] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to caliba0c337e231 ContainerID="85b06febbb19bd0a58aa28066f2570b6de7ab545b4046bee07d08d94ee9ccc38" Namespace="kube-system" Pod="coredns-7db6d8ff4d-fw5hr" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--fw5hr-eth0" Mar 17 17:42:40.524200 containerd[1467]: 2025-03-17 17:42:40.501 [INFO][4750] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="85b06febbb19bd0a58aa28066f2570b6de7ab545b4046bee07d08d94ee9ccc38" Namespace="kube-system" Pod="coredns-7db6d8ff4d-fw5hr" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--fw5hr-eth0" Mar 17 17:42:40.524200 containerd[1467]: 2025-03-17 17:42:40.502 [INFO][4750] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="85b06febbb19bd0a58aa28066f2570b6de7ab545b4046bee07d08d94ee9ccc38" Namespace="kube-system" Pod="coredns-7db6d8ff4d-fw5hr" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--fw5hr-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7db6d8ff4d--fw5hr-eth0", GenerateName:"coredns-7db6d8ff4d-", Namespace:"kube-system", SelfLink:"", UID:"ae45280b-f87f-4410-991d-e0beb4bdc7c6", ResourceVersion:"727", Generation:0, CreationTimestamp:time.Date(2025, time.March, 17, 17, 42, 16, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7db6d8ff4d", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"85b06febbb19bd0a58aa28066f2570b6de7ab545b4046bee07d08d94ee9ccc38", Pod:"coredns-7db6d8ff4d-fw5hr", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"caliba0c337e231", MAC:"1e:be:3b:ae:17:33", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Mar 17 17:42:40.524200 containerd[1467]: 2025-03-17 17:42:40.516 [INFO][4750] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="85b06febbb19bd0a58aa28066f2570b6de7ab545b4046bee07d08d94ee9ccc38" Namespace="kube-system" Pod="coredns-7db6d8ff4d-fw5hr" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--fw5hr-eth0" Mar 17 17:42:40.526824 containerd[1467]: time="2025-03-17T17:42:40.526566078Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 17:42:40.526824 containerd[1467]: time="2025-03-17T17:42:40.526640628Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 17:42:40.526824 containerd[1467]: time="2025-03-17T17:42:40.526664683Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:42:40.526824 containerd[1467]: time="2025-03-17T17:42:40.526767968Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:42:40.538655 systemd[1]: Started cri-containerd-b5375e4dccdc52e071442404c4b8a1d48eb22ca0db08a28f6918d5cd419f96d4.scope - libcontainer container b5375e4dccdc52e071442404c4b8a1d48eb22ca0db08a28f6918d5cd419f96d4. Mar 17 17:42:40.543782 systemd[1]: Started cri-containerd-854e88e2deedcc07c9c8cc51cb5c96cb4732e3e8f7e0a0a40aebc2bc877f541f.scope - libcontainer container 854e88e2deedcc07c9c8cc51cb5c96cb4732e3e8f7e0a0a40aebc2bc877f541f. Mar 17 17:42:40.550939 systemd-networkd[1410]: cali88aa4b200db: Link UP Mar 17 17:42:40.551432 systemd-networkd[1410]: cali88aa4b200db: Gained carrier Mar 17 17:42:40.558555 systemd[1]: Started cri-containerd-0d14e669762aafcce27fbf8489157e7e276641f1503453d7643a41b088ea7d19.scope - libcontainer container 0d14e669762aafcce27fbf8489157e7e276641f1503453d7643a41b088ea7d19. Mar 17 17:42:40.569424 systemd-resolved[1337]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Mar 17 17:42:40.572481 systemd-resolved[1337]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Mar 17 17:42:40.582076 containerd[1467]: 2025-03-17 17:42:40.203 [INFO][4771] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Mar 17 17:42:40.582076 containerd[1467]: 2025-03-17 17:42:40.219 [INFO][4771] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--kube--controllers--7fddcd7f99--92q58-eth0 calico-kube-controllers-7fddcd7f99- calico-system e531d9e1-a7bb-4958-adfd-8db7b3f84f95 734 0 2025-03-17 17:42:22 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:7fddcd7f99 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s localhost calico-kube-controllers-7fddcd7f99-92q58 eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali88aa4b200db [] []}} ContainerID="05d19de7c0c59db343e5a96552f1b3972e5702566d2b7466ff922ce5b5800748" Namespace="calico-system" Pod="calico-kube-controllers-7fddcd7f99-92q58" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7fddcd7f99--92q58-" Mar 17 17:42:40.582076 containerd[1467]: 2025-03-17 17:42:40.220 [INFO][4771] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="05d19de7c0c59db343e5a96552f1b3972e5702566d2b7466ff922ce5b5800748" Namespace="calico-system" Pod="calico-kube-controllers-7fddcd7f99-92q58" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7fddcd7f99--92q58-eth0" Mar 17 17:42:40.582076 containerd[1467]: 2025-03-17 17:42:40.308 [INFO][4860] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="05d19de7c0c59db343e5a96552f1b3972e5702566d2b7466ff922ce5b5800748" HandleID="k8s-pod-network.05d19de7c0c59db343e5a96552f1b3972e5702566d2b7466ff922ce5b5800748" Workload="localhost-k8s-calico--kube--controllers--7fddcd7f99--92q58-eth0" Mar 17 17:42:40.582076 containerd[1467]: 2025-03-17 17:42:40.317 [INFO][4860] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="05d19de7c0c59db343e5a96552f1b3972e5702566d2b7466ff922ce5b5800748" HandleID="k8s-pod-network.05d19de7c0c59db343e5a96552f1b3972e5702566d2b7466ff922ce5b5800748" Workload="localhost-k8s-calico--kube--controllers--7fddcd7f99--92q58-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0005020d0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"calico-kube-controllers-7fddcd7f99-92q58", "timestamp":"2025-03-17 17:42:40.307061742 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Mar 17 17:42:40.582076 containerd[1467]: 2025-03-17 17:42:40.318 [INFO][4860] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Mar 17 17:42:40.582076 containerd[1467]: 2025-03-17 17:42:40.484 [INFO][4860] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Mar 17 17:42:40.582076 containerd[1467]: 2025-03-17 17:42:40.484 [INFO][4860] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Mar 17 17:42:40.582076 containerd[1467]: 2025-03-17 17:42:40.486 [INFO][4860] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.05d19de7c0c59db343e5a96552f1b3972e5702566d2b7466ff922ce5b5800748" host="localhost" Mar 17 17:42:40.582076 containerd[1467]: 2025-03-17 17:42:40.494 [INFO][4860] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" Mar 17 17:42:40.582076 containerd[1467]: 2025-03-17 17:42:40.503 [INFO][4860] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Mar 17 17:42:40.582076 containerd[1467]: 2025-03-17 17:42:40.511 [INFO][4860] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Mar 17 17:42:40.582076 containerd[1467]: 2025-03-17 17:42:40.516 [INFO][4860] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Mar 17 17:42:40.582076 containerd[1467]: 2025-03-17 17:42:40.516 [INFO][4860] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.05d19de7c0c59db343e5a96552f1b3972e5702566d2b7466ff922ce5b5800748" host="localhost" Mar 17 17:42:40.582076 containerd[1467]: 2025-03-17 17:42:40.518 [INFO][4860] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.05d19de7c0c59db343e5a96552f1b3972e5702566d2b7466ff922ce5b5800748 Mar 17 17:42:40.582076 containerd[1467]: 2025-03-17 17:42:40.523 [INFO][4860] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.05d19de7c0c59db343e5a96552f1b3972e5702566d2b7466ff922ce5b5800748" host="localhost" Mar 17 17:42:40.582076 containerd[1467]: 2025-03-17 17:42:40.532 [INFO][4860] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.133/26] block=192.168.88.128/26 handle="k8s-pod-network.05d19de7c0c59db343e5a96552f1b3972e5702566d2b7466ff922ce5b5800748" host="localhost" Mar 17 17:42:40.582076 containerd[1467]: 2025-03-17 17:42:40.532 [INFO][4860] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.133/26] handle="k8s-pod-network.05d19de7c0c59db343e5a96552f1b3972e5702566d2b7466ff922ce5b5800748" host="localhost" Mar 17 17:42:40.582076 containerd[1467]: 2025-03-17 17:42:40.532 [INFO][4860] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Mar 17 17:42:40.582076 containerd[1467]: 2025-03-17 17:42:40.532 [INFO][4860] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.133/26] IPv6=[] ContainerID="05d19de7c0c59db343e5a96552f1b3972e5702566d2b7466ff922ce5b5800748" HandleID="k8s-pod-network.05d19de7c0c59db343e5a96552f1b3972e5702566d2b7466ff922ce5b5800748" Workload="localhost-k8s-calico--kube--controllers--7fddcd7f99--92q58-eth0" Mar 17 17:42:40.582679 containerd[1467]: 2025-03-17 17:42:40.539 [INFO][4771] cni-plugin/k8s.go 386: Populated endpoint ContainerID="05d19de7c0c59db343e5a96552f1b3972e5702566d2b7466ff922ce5b5800748" Namespace="calico-system" Pod="calico-kube-controllers-7fddcd7f99-92q58" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7fddcd7f99--92q58-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--7fddcd7f99--92q58-eth0", GenerateName:"calico-kube-controllers-7fddcd7f99-", Namespace:"calico-system", SelfLink:"", UID:"e531d9e1-a7bb-4958-adfd-8db7b3f84f95", ResourceVersion:"734", Generation:0, CreationTimestamp:time.Date(2025, time.March, 17, 17, 42, 22, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"7fddcd7f99", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-kube-controllers-7fddcd7f99-92q58", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali88aa4b200db", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Mar 17 17:42:40.582679 containerd[1467]: 2025-03-17 17:42:40.540 [INFO][4771] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.133/32] ContainerID="05d19de7c0c59db343e5a96552f1b3972e5702566d2b7466ff922ce5b5800748" Namespace="calico-system" Pod="calico-kube-controllers-7fddcd7f99-92q58" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7fddcd7f99--92q58-eth0" Mar 17 17:42:40.582679 containerd[1467]: 2025-03-17 17:42:40.540 [INFO][4771] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali88aa4b200db ContainerID="05d19de7c0c59db343e5a96552f1b3972e5702566d2b7466ff922ce5b5800748" Namespace="calico-system" Pod="calico-kube-controllers-7fddcd7f99-92q58" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7fddcd7f99--92q58-eth0" Mar 17 17:42:40.582679 containerd[1467]: 2025-03-17 17:42:40.555 [INFO][4771] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="05d19de7c0c59db343e5a96552f1b3972e5702566d2b7466ff922ce5b5800748" Namespace="calico-system" Pod="calico-kube-controllers-7fddcd7f99-92q58" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7fddcd7f99--92q58-eth0" Mar 17 17:42:40.582679 containerd[1467]: 2025-03-17 17:42:40.555 [INFO][4771] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="05d19de7c0c59db343e5a96552f1b3972e5702566d2b7466ff922ce5b5800748" Namespace="calico-system" Pod="calico-kube-controllers-7fddcd7f99-92q58" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7fddcd7f99--92q58-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--7fddcd7f99--92q58-eth0", GenerateName:"calico-kube-controllers-7fddcd7f99-", Namespace:"calico-system", SelfLink:"", UID:"e531d9e1-a7bb-4958-adfd-8db7b3f84f95", ResourceVersion:"734", Generation:0, CreationTimestamp:time.Date(2025, time.March, 17, 17, 42, 22, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"7fddcd7f99", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"05d19de7c0c59db343e5a96552f1b3972e5702566d2b7466ff922ce5b5800748", Pod:"calico-kube-controllers-7fddcd7f99-92q58", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali88aa4b200db", MAC:"7a:81:9d:61:5c:b5", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Mar 17 17:42:40.582679 containerd[1467]: 2025-03-17 17:42:40.568 [INFO][4771] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="05d19de7c0c59db343e5a96552f1b3972e5702566d2b7466ff922ce5b5800748" Namespace="calico-system" Pod="calico-kube-controllers-7fddcd7f99-92q58" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7fddcd7f99--92q58-eth0" Mar 17 17:42:40.589503 containerd[1467]: time="2025-03-17T17:42:40.589377428Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 17:42:40.589842 containerd[1467]: time="2025-03-17T17:42:40.589792076Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 17:42:40.589842 containerd[1467]: time="2025-03-17T17:42:40.589819328Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:42:40.589983 containerd[1467]: time="2025-03-17T17:42:40.589915439Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:42:40.603966 containerd[1467]: time="2025-03-17T17:42:40.603870579Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-j9gdg,Uid:27eb0c6c-40f4-4b19-b513-0003d63df5d0,Namespace:calico-system,Attempt:5,} returns sandbox id \"b5375e4dccdc52e071442404c4b8a1d48eb22ca0db08a28f6918d5cd419f96d4\"" Mar 17 17:42:40.608095 containerd[1467]: time="2025-03-17T17:42:40.608053782Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.2\"" Mar 17 17:42:40.622634 systemd[1]: run-netns-cni\x2df78a1080\x2d6347\x2d68d7\x2da689\x2de0521ab3d0ed.mount: Deactivated successfully. Mar 17 17:42:40.622878 systemd[1]: run-netns-cni\x2d1eb8ecc6\x2d0ee1\x2dce15\x2d0144\x2d974084a67219.mount: Deactivated successfully. Mar 17 17:42:40.628156 containerd[1467]: time="2025-03-17T17:42:40.627734814Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-sjrwq,Uid:228c6093-6dc2-4faf-8ec8-87849ef402ab,Namespace:kube-system,Attempt:6,} returns sandbox id \"854e88e2deedcc07c9c8cc51cb5c96cb4732e3e8f7e0a0a40aebc2bc877f541f\"" Mar 17 17:42:40.629546 kubelet[2662]: E0317 17:42:40.629006 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:40.632335 systemd-networkd[1410]: calib7bfeedfd56: Link UP Mar 17 17:42:40.633343 systemd-networkd[1410]: calib7bfeedfd56: Gained carrier Mar 17 17:42:40.635137 systemd[1]: Started cri-containerd-85b06febbb19bd0a58aa28066f2570b6de7ab545b4046bee07d08d94ee9ccc38.scope - libcontainer container 85b06febbb19bd0a58aa28066f2570b6de7ab545b4046bee07d08d94ee9ccc38. Mar 17 17:42:40.635253 containerd[1467]: time="2025-03-17T17:42:40.635140303Z" level=info msg="CreateContainer within sandbox \"854e88e2deedcc07c9c8cc51cb5c96cb4732e3e8f7e0a0a40aebc2bc877f541f\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Mar 17 17:42:40.644539 systemd-resolved[1337]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Mar 17 17:42:40.656764 systemd-resolved[1337]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Mar 17 17:42:40.657244 containerd[1467]: 2025-03-17 17:42:40.195 [INFO][4793] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Mar 17 17:42:40.657244 containerd[1467]: 2025-03-17 17:42:40.205 [INFO][4793] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--86c8b6f988--ppvkt-eth0 calico-apiserver-86c8b6f988- calico-apiserver c9eab445-c078-490b-861c-b96d204afef3 732 0 2025-03-17 17:42:22 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:86c8b6f988 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-86c8b6f988-ppvkt eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] calib7bfeedfd56 [] []}} ContainerID="b52450c5eb1a269012006cf7355e70d1ac2959e27862704ab508fe9696d3aec6" Namespace="calico-apiserver" Pod="calico-apiserver-86c8b6f988-ppvkt" WorkloadEndpoint="localhost-k8s-calico--apiserver--86c8b6f988--ppvkt-" Mar 17 17:42:40.657244 containerd[1467]: 2025-03-17 17:42:40.205 [INFO][4793] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="b52450c5eb1a269012006cf7355e70d1ac2959e27862704ab508fe9696d3aec6" Namespace="calico-apiserver" Pod="calico-apiserver-86c8b6f988-ppvkt" WorkloadEndpoint="localhost-k8s-calico--apiserver--86c8b6f988--ppvkt-eth0" Mar 17 17:42:40.657244 containerd[1467]: 2025-03-17 17:42:40.309 [INFO][4839] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="b52450c5eb1a269012006cf7355e70d1ac2959e27862704ab508fe9696d3aec6" HandleID="k8s-pod-network.b52450c5eb1a269012006cf7355e70d1ac2959e27862704ab508fe9696d3aec6" Workload="localhost-k8s-calico--apiserver--86c8b6f988--ppvkt-eth0" Mar 17 17:42:40.657244 containerd[1467]: 2025-03-17 17:42:40.319 [INFO][4839] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="b52450c5eb1a269012006cf7355e70d1ac2959e27862704ab508fe9696d3aec6" HandleID="k8s-pod-network.b52450c5eb1a269012006cf7355e70d1ac2959e27862704ab508fe9696d3aec6" Workload="localhost-k8s-calico--apiserver--86c8b6f988--ppvkt-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000373920), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-86c8b6f988-ppvkt", "timestamp":"2025-03-17 17:42:40.308182479 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Mar 17 17:42:40.657244 containerd[1467]: 2025-03-17 17:42:40.319 [INFO][4839] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Mar 17 17:42:40.657244 containerd[1467]: 2025-03-17 17:42:40.534 [INFO][4839] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Mar 17 17:42:40.657244 containerd[1467]: 2025-03-17 17:42:40.535 [INFO][4839] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Mar 17 17:42:40.657244 containerd[1467]: 2025-03-17 17:42:40.542 [INFO][4839] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.b52450c5eb1a269012006cf7355e70d1ac2959e27862704ab508fe9696d3aec6" host="localhost" Mar 17 17:42:40.657244 containerd[1467]: 2025-03-17 17:42:40.550 [INFO][4839] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" Mar 17 17:42:40.657244 containerd[1467]: 2025-03-17 17:42:40.576 [INFO][4839] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Mar 17 17:42:40.657244 containerd[1467]: 2025-03-17 17:42:40.583 [INFO][4839] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Mar 17 17:42:40.657244 containerd[1467]: 2025-03-17 17:42:40.593 [INFO][4839] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Mar 17 17:42:40.657244 containerd[1467]: 2025-03-17 17:42:40.593 [INFO][4839] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.b52450c5eb1a269012006cf7355e70d1ac2959e27862704ab508fe9696d3aec6" host="localhost" Mar 17 17:42:40.657244 containerd[1467]: 2025-03-17 17:42:40.597 [INFO][4839] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.b52450c5eb1a269012006cf7355e70d1ac2959e27862704ab508fe9696d3aec6 Mar 17 17:42:40.657244 containerd[1467]: 2025-03-17 17:42:40.602 [INFO][4839] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.b52450c5eb1a269012006cf7355e70d1ac2959e27862704ab508fe9696d3aec6" host="localhost" Mar 17 17:42:40.657244 containerd[1467]: 2025-03-17 17:42:40.613 [INFO][4839] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.134/26] block=192.168.88.128/26 handle="k8s-pod-network.b52450c5eb1a269012006cf7355e70d1ac2959e27862704ab508fe9696d3aec6" host="localhost" Mar 17 17:42:40.657244 containerd[1467]: 2025-03-17 17:42:40.613 [INFO][4839] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.134/26] handle="k8s-pod-network.b52450c5eb1a269012006cf7355e70d1ac2959e27862704ab508fe9696d3aec6" host="localhost" Mar 17 17:42:40.657244 containerd[1467]: 2025-03-17 17:42:40.613 [INFO][4839] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Mar 17 17:42:40.657244 containerd[1467]: 2025-03-17 17:42:40.613 [INFO][4839] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.134/26] IPv6=[] ContainerID="b52450c5eb1a269012006cf7355e70d1ac2959e27862704ab508fe9696d3aec6" HandleID="k8s-pod-network.b52450c5eb1a269012006cf7355e70d1ac2959e27862704ab508fe9696d3aec6" Workload="localhost-k8s-calico--apiserver--86c8b6f988--ppvkt-eth0" Mar 17 17:42:40.657850 containerd[1467]: 2025-03-17 17:42:40.624 [INFO][4793] cni-plugin/k8s.go 386: Populated endpoint ContainerID="b52450c5eb1a269012006cf7355e70d1ac2959e27862704ab508fe9696d3aec6" Namespace="calico-apiserver" Pod="calico-apiserver-86c8b6f988-ppvkt" WorkloadEndpoint="localhost-k8s-calico--apiserver--86c8b6f988--ppvkt-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--86c8b6f988--ppvkt-eth0", GenerateName:"calico-apiserver-86c8b6f988-", Namespace:"calico-apiserver", SelfLink:"", UID:"c9eab445-c078-490b-861c-b96d204afef3", ResourceVersion:"732", Generation:0, CreationTimestamp:time.Date(2025, time.March, 17, 17, 42, 22, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"86c8b6f988", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-86c8b6f988-ppvkt", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calib7bfeedfd56", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Mar 17 17:42:40.657850 containerd[1467]: 2025-03-17 17:42:40.625 [INFO][4793] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.134/32] ContainerID="b52450c5eb1a269012006cf7355e70d1ac2959e27862704ab508fe9696d3aec6" Namespace="calico-apiserver" Pod="calico-apiserver-86c8b6f988-ppvkt" WorkloadEndpoint="localhost-k8s-calico--apiserver--86c8b6f988--ppvkt-eth0" Mar 17 17:42:40.657850 containerd[1467]: 2025-03-17 17:42:40.626 [INFO][4793] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calib7bfeedfd56 ContainerID="b52450c5eb1a269012006cf7355e70d1ac2959e27862704ab508fe9696d3aec6" Namespace="calico-apiserver" Pod="calico-apiserver-86c8b6f988-ppvkt" WorkloadEndpoint="localhost-k8s-calico--apiserver--86c8b6f988--ppvkt-eth0" Mar 17 17:42:40.657850 containerd[1467]: 2025-03-17 17:42:40.633 [INFO][4793] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="b52450c5eb1a269012006cf7355e70d1ac2959e27862704ab508fe9696d3aec6" Namespace="calico-apiserver" Pod="calico-apiserver-86c8b6f988-ppvkt" WorkloadEndpoint="localhost-k8s-calico--apiserver--86c8b6f988--ppvkt-eth0" Mar 17 17:42:40.657850 containerd[1467]: 2025-03-17 17:42:40.637 [INFO][4793] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="b52450c5eb1a269012006cf7355e70d1ac2959e27862704ab508fe9696d3aec6" Namespace="calico-apiserver" Pod="calico-apiserver-86c8b6f988-ppvkt" WorkloadEndpoint="localhost-k8s-calico--apiserver--86c8b6f988--ppvkt-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--86c8b6f988--ppvkt-eth0", GenerateName:"calico-apiserver-86c8b6f988-", Namespace:"calico-apiserver", SelfLink:"", UID:"c9eab445-c078-490b-861c-b96d204afef3", ResourceVersion:"732", Generation:0, CreationTimestamp:time.Date(2025, time.March, 17, 17, 42, 22, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"86c8b6f988", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"b52450c5eb1a269012006cf7355e70d1ac2959e27862704ab508fe9696d3aec6", Pod:"calico-apiserver-86c8b6f988-ppvkt", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"calib7bfeedfd56", MAC:"c2:ff:88:33:93:b1", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Mar 17 17:42:40.657850 containerd[1467]: 2025-03-17 17:42:40.652 [INFO][4793] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="b52450c5eb1a269012006cf7355e70d1ac2959e27862704ab508fe9696d3aec6" Namespace="calico-apiserver" Pod="calico-apiserver-86c8b6f988-ppvkt" WorkloadEndpoint="localhost-k8s-calico--apiserver--86c8b6f988--ppvkt-eth0" Mar 17 17:42:40.659424 containerd[1467]: time="2025-03-17T17:42:40.658726174Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 17:42:40.659424 containerd[1467]: time="2025-03-17T17:42:40.658784995Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 17:42:40.659424 containerd[1467]: time="2025-03-17T17:42:40.658795214Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:42:40.659424 containerd[1467]: time="2025-03-17T17:42:40.658890332Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:42:40.676576 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount614869657.mount: Deactivated successfully. Mar 17 17:42:40.696338 containerd[1467]: time="2025-03-17T17:42:40.692844733Z" level=info msg="CreateContainer within sandbox \"854e88e2deedcc07c9c8cc51cb5c96cb4732e3e8f7e0a0a40aebc2bc877f541f\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"264f69901afc26ab78bc0b315a0c8571ea35d75eceb2638c3c2d4c76a6c4898e\"" Mar 17 17:42:40.696851 containerd[1467]: time="2025-03-17T17:42:40.696820766Z" level=info msg="StartContainer for \"264f69901afc26ab78bc0b315a0c8571ea35d75eceb2638c3c2d4c76a6c4898e\"" Mar 17 17:42:40.700630 systemd[1]: Started cri-containerd-05d19de7c0c59db343e5a96552f1b3972e5702566d2b7466ff922ce5b5800748.scope - libcontainer container 05d19de7c0c59db343e5a96552f1b3972e5702566d2b7466ff922ce5b5800748. Mar 17 17:42:40.703527 containerd[1467]: time="2025-03-17T17:42:40.703490562Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-fw5hr,Uid:ae45280b-f87f-4410-991d-e0beb4bdc7c6,Namespace:kube-system,Attempt:5,} returns sandbox id \"85b06febbb19bd0a58aa28066f2570b6de7ab545b4046bee07d08d94ee9ccc38\"" Mar 17 17:42:40.704856 kubelet[2662]: E0317 17:42:40.704837 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:40.708118 containerd[1467]: time="2025-03-17T17:42:40.708036226Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86c8b6f988-79bzl,Uid:3265838f-27ca-489d-8c65-7cadce5cca40,Namespace:calico-apiserver,Attempt:5,} returns sandbox id \"0d14e669762aafcce27fbf8489157e7e276641f1503453d7643a41b088ea7d19\"" Mar 17 17:42:40.712107 containerd[1467]: time="2025-03-17T17:42:40.711403736Z" level=info msg="CreateContainer within sandbox \"85b06febbb19bd0a58aa28066f2570b6de7ab545b4046bee07d08d94ee9ccc38\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Mar 17 17:42:40.718106 containerd[1467]: time="2025-03-17T17:42:40.717450322Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 17:42:40.718106 containerd[1467]: time="2025-03-17T17:42:40.717498662Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 17:42:40.718106 containerd[1467]: time="2025-03-17T17:42:40.717509523Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:42:40.718106 containerd[1467]: time="2025-03-17T17:42:40.717581979Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 17:42:40.720077 systemd-resolved[1337]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Mar 17 17:42:40.732561 systemd[1]: Started cri-containerd-264f69901afc26ab78bc0b315a0c8571ea35d75eceb2638c3c2d4c76a6c4898e.scope - libcontainer container 264f69901afc26ab78bc0b315a0c8571ea35d75eceb2638c3c2d4c76a6c4898e. Mar 17 17:42:40.736746 systemd[1]: Started cri-containerd-b52450c5eb1a269012006cf7355e70d1ac2959e27862704ab508fe9696d3aec6.scope - libcontainer container b52450c5eb1a269012006cf7355e70d1ac2959e27862704ab508fe9696d3aec6. Mar 17 17:42:40.736889 containerd[1467]: time="2025-03-17T17:42:40.736849915Z" level=info msg="CreateContainer within sandbox \"85b06febbb19bd0a58aa28066f2570b6de7ab545b4046bee07d08d94ee9ccc38\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"b827b1c200f0b006fc743c47d92c921d140894daabf676a32dc53e197aa3886a\"" Mar 17 17:42:40.737607 containerd[1467]: time="2025-03-17T17:42:40.737581459Z" level=info msg="StartContainer for \"b827b1c200f0b006fc743c47d92c921d140894daabf676a32dc53e197aa3886a\"" Mar 17 17:42:40.757596 systemd-resolved[1337]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Mar 17 17:42:40.768611 containerd[1467]: time="2025-03-17T17:42:40.768576006Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7fddcd7f99-92q58,Uid:e531d9e1-a7bb-4958-adfd-8db7b3f84f95,Namespace:calico-system,Attempt:5,} returns sandbox id \"05d19de7c0c59db343e5a96552f1b3972e5702566d2b7466ff922ce5b5800748\"" Mar 17 17:42:40.775942 containerd[1467]: time="2025-03-17T17:42:40.775834710Z" level=info msg="StartContainer for \"264f69901afc26ab78bc0b315a0c8571ea35d75eceb2638c3c2d4c76a6c4898e\" returns successfully" Mar 17 17:42:40.777637 systemd[1]: Started cri-containerd-b827b1c200f0b006fc743c47d92c921d140894daabf676a32dc53e197aa3886a.scope - libcontainer container b827b1c200f0b006fc743c47d92c921d140894daabf676a32dc53e197aa3886a. Mar 17 17:42:40.791795 containerd[1467]: time="2025-03-17T17:42:40.791752368Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-86c8b6f988-ppvkt,Uid:c9eab445-c078-490b-861c-b96d204afef3,Namespace:calico-apiserver,Attempt:6,} returns sandbox id \"b52450c5eb1a269012006cf7355e70d1ac2959e27862704ab508fe9696d3aec6\"" Mar 17 17:42:40.809886 containerd[1467]: time="2025-03-17T17:42:40.809799990Z" level=info msg="StartContainer for \"b827b1c200f0b006fc743c47d92c921d140894daabf676a32dc53e197aa3886a\" returns successfully" Mar 17 17:42:41.035548 kernel: bpftool[5416]: memfd_create() called without MFD_EXEC or MFD_NOEXEC_SEAL set Mar 17 17:42:41.111792 kubelet[2662]: E0317 17:42:41.111760 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:41.115925 kubelet[2662]: E0317 17:42:41.115880 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:41.124914 kubelet[2662]: I0317 17:42:41.124888 2662 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 17 17:42:41.125714 kubelet[2662]: E0317 17:42:41.125687 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:41.161167 kubelet[2662]: I0317 17:42:41.161094 2662 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7db6d8ff4d-sjrwq" podStartSLOduration=25.161068092 podStartE2EDuration="25.161068092s" podCreationTimestamp="2025-03-17 17:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-17 17:42:41.160306861 +0000 UTC m=+39.420273587" watchObservedRunningTime="2025-03-17 17:42:41.161068092 +0000 UTC m=+39.421034818" Mar 17 17:42:41.174856 kubelet[2662]: I0317 17:42:41.174745 2662 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7db6d8ff4d-fw5hr" podStartSLOduration=25.174726572 podStartE2EDuration="25.174726572s" podCreationTimestamp="2025-03-17 17:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-17 17:42:41.172669485 +0000 UTC m=+39.432636211" watchObservedRunningTime="2025-03-17 17:42:41.174726572 +0000 UTC m=+39.434693298" Mar 17 17:42:41.271076 systemd-networkd[1410]: vxlan.calico: Link UP Mar 17 17:42:41.271097 systemd-networkd[1410]: vxlan.calico: Gained carrier Mar 17 17:42:41.419477 systemd-networkd[1410]: cali0e24b01453e: Gained IPv6LL Mar 17 17:42:41.548648 systemd-networkd[1410]: caliba0c337e231: Gained IPv6LL Mar 17 17:42:42.059513 systemd-networkd[1410]: calib7bfeedfd56: Gained IPv6LL Mar 17 17:42:42.128106 kubelet[2662]: E0317 17:42:42.127984 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:42.128106 kubelet[2662]: E0317 17:42:42.128057 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:42.156368 systemd[1]: Started sshd@11-10.0.0.61:22-10.0.0.1:33422.service - OpenSSH per-connection server daemon (10.0.0.1:33422). Mar 17 17:42:42.188591 systemd-networkd[1410]: cali47e18c80f12: Gained IPv6LL Mar 17 17:42:42.221807 sshd[5504]: Accepted publickey for core from 10.0.0.1 port 33422 ssh2: RSA SHA256:j201F9FRK1q3ChnxQf0adNdYppDp+g37vmaXPvsVhek Mar 17 17:42:42.224138 sshd-session[5504]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 17 17:42:42.228303 systemd-logind[1453]: New session 12 of user core. Mar 17 17:42:42.233476 systemd[1]: Started session-12.scope - Session 12 of User core. Mar 17 17:42:42.234905 containerd[1467]: time="2025-03-17T17:42:42.234640446Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.29.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:42:42.236459 containerd[1467]: time="2025-03-17T17:42:42.236394883Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.29.2: active requests=0, bytes read=7909887" Mar 17 17:42:42.237909 containerd[1467]: time="2025-03-17T17:42:42.237874493Z" level=info msg="ImageCreate event name:\"sha256:0fae09f861e350c042fe0db9ce9f8cc5ac4df975a5c4e4a9ddc3c6fac1552a9a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:42:42.240316 containerd[1467]: time="2025-03-17T17:42:42.240284792Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:214b4eef7008808bda55ad3cc1d4a3cd8df9e0e8094dff213fa3241104eb892c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:42:42.240939 containerd[1467]: time="2025-03-17T17:42:42.240916449Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.29.2\" with image id \"sha256:0fae09f861e350c042fe0db9ce9f8cc5ac4df975a5c4e4a9ddc3c6fac1552a9a\", repo tag \"ghcr.io/flatcar/calico/csi:v3.29.2\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:214b4eef7008808bda55ad3cc1d4a3cd8df9e0e8094dff213fa3241104eb892c\", size \"9402991\" in 1.632819878s" Mar 17 17:42:42.240982 containerd[1467]: time="2025-03-17T17:42:42.240943220Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.2\" returns image reference \"sha256:0fae09f861e350c042fe0db9ce9f8cc5ac4df975a5c4e4a9ddc3c6fac1552a9a\"" Mar 17 17:42:42.242743 containerd[1467]: time="2025-03-17T17:42:42.242500035Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.2\"" Mar 17 17:42:42.244081 containerd[1467]: time="2025-03-17T17:42:42.244035321Z" level=info msg="CreateContainer within sandbox \"b5375e4dccdc52e071442404c4b8a1d48eb22ca0db08a28f6918d5cd419f96d4\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Mar 17 17:42:42.251471 systemd-networkd[1410]: calice8227e9e30: Gained IPv6LL Mar 17 17:42:42.270250 containerd[1467]: time="2025-03-17T17:42:42.270205298Z" level=info msg="CreateContainer within sandbox \"b5375e4dccdc52e071442404c4b8a1d48eb22ca0db08a28f6918d5cd419f96d4\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"efee84d7630f9c833d494251d694a9517b5a822bbf2e0d5dca3f82cdbe7827f9\"" Mar 17 17:42:42.270718 containerd[1467]: time="2025-03-17T17:42:42.270691010Z" level=info msg="StartContainer for \"efee84d7630f9c833d494251d694a9517b5a822bbf2e0d5dca3f82cdbe7827f9\"" Mar 17 17:42:42.305488 systemd[1]: Started cri-containerd-efee84d7630f9c833d494251d694a9517b5a822bbf2e0d5dca3f82cdbe7827f9.scope - libcontainer container efee84d7630f9c833d494251d694a9517b5a822bbf2e0d5dca3f82cdbe7827f9. Mar 17 17:42:42.340182 containerd[1467]: time="2025-03-17T17:42:42.340052562Z" level=info msg="StartContainer for \"efee84d7630f9c833d494251d694a9517b5a822bbf2e0d5dca3f82cdbe7827f9\" returns successfully" Mar 17 17:42:42.375957 sshd[5507]: Connection closed by 10.0.0.1 port 33422 Mar 17 17:42:42.376312 sshd-session[5504]: pam_unix(sshd:session): session closed for user core Mar 17 17:42:42.380416 systemd[1]: sshd@11-10.0.0.61:22-10.0.0.1:33422.service: Deactivated successfully. Mar 17 17:42:42.382529 systemd[1]: session-12.scope: Deactivated successfully. Mar 17 17:42:42.383140 systemd-logind[1453]: Session 12 logged out. Waiting for processes to exit. Mar 17 17:42:42.383949 systemd-logind[1453]: Removed session 12. Mar 17 17:42:42.507508 systemd-networkd[1410]: cali88aa4b200db: Gained IPv6LL Mar 17 17:42:42.571540 systemd-networkd[1410]: vxlan.calico: Gained IPv6LL Mar 17 17:42:43.131786 kubelet[2662]: E0317 17:42:43.131752 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:43.132257 kubelet[2662]: E0317 17:42:43.132012 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:45.118557 containerd[1467]: time="2025-03-17T17:42:45.118486075Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.29.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:42:45.147240 containerd[1467]: time="2025-03-17T17:42:45.147177566Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.29.2: active requests=0, bytes read=42993204" Mar 17 17:42:45.196060 containerd[1467]: time="2025-03-17T17:42:45.196015704Z" level=info msg="ImageCreate event name:\"sha256:d27fc480d1ad33921c40abef2ab6828fadf6524674fdcc622f571a5abc34ad55\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:42:45.219383 containerd[1467]: time="2025-03-17T17:42:45.219310365Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:3623f5b60fad0da3387a8649371b53171a4b1226f4d989d2acad9145dc0ef56f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:42:45.220111 containerd[1467]: time="2025-03-17T17:42:45.220077677Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.29.2\" with image id \"sha256:d27fc480d1ad33921c40abef2ab6828fadf6524674fdcc622f571a5abc34ad55\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.29.2\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:3623f5b60fad0da3387a8649371b53171a4b1226f4d989d2acad9145dc0ef56f\", size \"44486324\" in 2.977538849s" Mar 17 17:42:45.220111 containerd[1467]: time="2025-03-17T17:42:45.220105589Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.2\" returns image reference \"sha256:d27fc480d1ad33921c40abef2ab6828fadf6524674fdcc622f571a5abc34ad55\"" Mar 17 17:42:45.222163 containerd[1467]: time="2025-03-17T17:42:45.221017291Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.29.2\"" Mar 17 17:42:45.222163 containerd[1467]: time="2025-03-17T17:42:45.221919687Z" level=info msg="CreateContainer within sandbox \"0d14e669762aafcce27fbf8489157e7e276641f1503453d7643a41b088ea7d19\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Mar 17 17:42:45.641332 containerd[1467]: time="2025-03-17T17:42:45.641250798Z" level=info msg="CreateContainer within sandbox \"0d14e669762aafcce27fbf8489157e7e276641f1503453d7643a41b088ea7d19\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"dea0666c03d1afddbb584360cb0b57ffaef375058b6826c293af202d7e3e7a45\"" Mar 17 17:42:45.642399 containerd[1467]: time="2025-03-17T17:42:45.641975099Z" level=info msg="StartContainer for \"dea0666c03d1afddbb584360cb0b57ffaef375058b6826c293af202d7e3e7a45\"" Mar 17 17:42:45.679861 systemd[1]: Started cri-containerd-dea0666c03d1afddbb584360cb0b57ffaef375058b6826c293af202d7e3e7a45.scope - libcontainer container dea0666c03d1afddbb584360cb0b57ffaef375058b6826c293af202d7e3e7a45. Mar 17 17:42:45.807526 containerd[1467]: time="2025-03-17T17:42:45.807462686Z" level=info msg="StartContainer for \"dea0666c03d1afddbb584360cb0b57ffaef375058b6826c293af202d7e3e7a45\" returns successfully" Mar 17 17:42:47.141372 kubelet[2662]: I0317 17:42:47.141333 2662 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 17 17:42:47.389530 systemd[1]: Started sshd@12-10.0.0.61:22-10.0.0.1:33430.service - OpenSSH per-connection server daemon (10.0.0.1:33430). Mar 17 17:42:47.532032 sshd[5619]: Accepted publickey for core from 10.0.0.1 port 33430 ssh2: RSA SHA256:j201F9FRK1q3ChnxQf0adNdYppDp+g37vmaXPvsVhek Mar 17 17:42:47.533967 sshd-session[5619]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 17 17:42:47.537904 systemd-logind[1453]: New session 13 of user core. Mar 17 17:42:47.546490 systemd[1]: Started session-13.scope - Session 13 of User core. Mar 17 17:42:47.782699 sshd[5621]: Connection closed by 10.0.0.1 port 33430 Mar 17 17:42:47.782973 sshd-session[5619]: pam_unix(sshd:session): session closed for user core Mar 17 17:42:47.788315 systemd[1]: sshd@12-10.0.0.61:22-10.0.0.1:33430.service: Deactivated successfully. Mar 17 17:42:47.790917 systemd[1]: session-13.scope: Deactivated successfully. Mar 17 17:42:47.791666 systemd-logind[1453]: Session 13 logged out. Waiting for processes to exit. Mar 17 17:42:47.792601 systemd-logind[1453]: Removed session 13. Mar 17 17:42:49.106082 containerd[1467]: time="2025-03-17T17:42:49.106021390Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.29.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:42:49.129189 containerd[1467]: time="2025-03-17T17:42:49.129147120Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.29.2: active requests=0, bytes read=34792912" Mar 17 17:42:49.141686 containerd[1467]: time="2025-03-17T17:42:49.141644200Z" level=info msg="ImageCreate event name:\"sha256:f6a228558381bc7de7c5296ac6c4e903cfda929899c85806367a726ef6d7ff5f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:42:49.151572 containerd[1467]: time="2025-03-17T17:42:49.151532621Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:6d1f392b747f912366ec5c60ee1130952c2c07e8ce24c53480187daa0e3364aa\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:42:49.152398 containerd[1467]: time="2025-03-17T17:42:49.152328075Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.29.2\" with image id \"sha256:f6a228558381bc7de7c5296ac6c4e903cfda929899c85806367a726ef6d7ff5f\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.29.2\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:6d1f392b747f912366ec5c60ee1130952c2c07e8ce24c53480187daa0e3364aa\", size \"36285984\" in 3.931281698s" Mar 17 17:42:49.152439 containerd[1467]: time="2025-03-17T17:42:49.152398457Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.29.2\" returns image reference \"sha256:f6a228558381bc7de7c5296ac6c4e903cfda929899c85806367a726ef6d7ff5f\"" Mar 17 17:42:49.153224 containerd[1467]: time="2025-03-17T17:42:49.153155579Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.2\"" Mar 17 17:42:49.160148 containerd[1467]: time="2025-03-17T17:42:49.160101984Z" level=info msg="CreateContainer within sandbox \"05d19de7c0c59db343e5a96552f1b3972e5702566d2b7466ff922ce5b5800748\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Mar 17 17:42:49.242783 containerd[1467]: time="2025-03-17T17:42:49.242745392Z" level=info msg="CreateContainer within sandbox \"05d19de7c0c59db343e5a96552f1b3972e5702566d2b7466ff922ce5b5800748\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"6286cc1551c27a6acbf7b9e267ced054e1b1c2966d792f77e99cfec018b6ed8a\"" Mar 17 17:42:49.243334 containerd[1467]: time="2025-03-17T17:42:49.243269045Z" level=info msg="StartContainer for \"6286cc1551c27a6acbf7b9e267ced054e1b1c2966d792f77e99cfec018b6ed8a\"" Mar 17 17:42:49.272506 systemd[1]: Started cri-containerd-6286cc1551c27a6acbf7b9e267ced054e1b1c2966d792f77e99cfec018b6ed8a.scope - libcontainer container 6286cc1551c27a6acbf7b9e267ced054e1b1c2966d792f77e99cfec018b6ed8a. Mar 17 17:42:49.346628 containerd[1467]: time="2025-03-17T17:42:49.346581480Z" level=info msg="StartContainer for \"6286cc1551c27a6acbf7b9e267ced054e1b1c2966d792f77e99cfec018b6ed8a\" returns successfully" Mar 17 17:42:49.646925 kubelet[2662]: I0317 17:42:49.646843 2662 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 17 17:42:49.647559 kubelet[2662]: E0317 17:42:49.647526 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:49.784535 kubelet[2662]: I0317 17:42:49.784131 2662 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-86c8b6f988-79bzl" podStartSLOduration=23.273019582 podStartE2EDuration="27.784113751s" podCreationTimestamp="2025-03-17 17:42:22 +0000 UTC" firstStartedPulling="2025-03-17 17:42:40.7096928 +0000 UTC m=+38.969659526" lastFinishedPulling="2025-03-17 17:42:45.220786969 +0000 UTC m=+43.480753695" observedRunningTime="2025-03-17 17:42:46.262229887 +0000 UTC m=+44.522196613" watchObservedRunningTime="2025-03-17 17:42:49.784113751 +0000 UTC m=+48.044080467" Mar 17 17:42:49.805923 containerd[1467]: time="2025-03-17T17:42:49.805853439Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.29.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:42:49.806795 containerd[1467]: time="2025-03-17T17:42:49.806756906Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.29.2: active requests=0, bytes read=77" Mar 17 17:42:49.809026 containerd[1467]: time="2025-03-17T17:42:49.808982686Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.29.2\" with image id \"sha256:d27fc480d1ad33921c40abef2ab6828fadf6524674fdcc622f571a5abc34ad55\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.29.2\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:3623f5b60fad0da3387a8649371b53171a4b1226f4d989d2acad9145dc0ef56f\", size \"44486324\" in 655.795417ms" Mar 17 17:42:49.809026 containerd[1467]: time="2025-03-17T17:42:49.809023512Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.2\" returns image reference \"sha256:d27fc480d1ad33921c40abef2ab6828fadf6524674fdcc622f571a5abc34ad55\"" Mar 17 17:42:49.810860 containerd[1467]: time="2025-03-17T17:42:49.810626192Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.2\"" Mar 17 17:42:49.811828 containerd[1467]: time="2025-03-17T17:42:49.811745754Z" level=info msg="CreateContainer within sandbox \"b52450c5eb1a269012006cf7355e70d1ac2959e27862704ab508fe9696d3aec6\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Mar 17 17:42:49.829369 containerd[1467]: time="2025-03-17T17:42:49.829172182Z" level=info msg="CreateContainer within sandbox \"b52450c5eb1a269012006cf7355e70d1ac2959e27862704ab508fe9696d3aec6\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"78671a609f970c741a24f3f35845b1e5e3e94dec343fee7df302402bd4030b0c\"" Mar 17 17:42:49.834424 containerd[1467]: time="2025-03-17T17:42:49.833443524Z" level=info msg="StartContainer for \"78671a609f970c741a24f3f35845b1e5e3e94dec343fee7df302402bd4030b0c\"" Mar 17 17:42:49.876608 systemd[1]: Started cri-containerd-78671a609f970c741a24f3f35845b1e5e3e94dec343fee7df302402bd4030b0c.scope - libcontainer container 78671a609f970c741a24f3f35845b1e5e3e94dec343fee7df302402bd4030b0c. Mar 17 17:42:49.916671 containerd[1467]: time="2025-03-17T17:42:49.916571041Z" level=info msg="StartContainer for \"78671a609f970c741a24f3f35845b1e5e3e94dec343fee7df302402bd4030b0c\" returns successfully" Mar 17 17:42:50.161774 kubelet[2662]: E0317 17:42:50.161730 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:42:50.183396 kubelet[2662]: I0317 17:42:50.183183 2662 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-86c8b6f988-ppvkt" podStartSLOduration=19.166276998 podStartE2EDuration="28.183158669s" podCreationTimestamp="2025-03-17 17:42:22 +0000 UTC" firstStartedPulling="2025-03-17 17:42:40.792995134 +0000 UTC m=+39.052961860" lastFinishedPulling="2025-03-17 17:42:49.809876805 +0000 UTC m=+48.069843531" observedRunningTime="2025-03-17 17:42:50.181510704 +0000 UTC m=+48.441477440" watchObservedRunningTime="2025-03-17 17:42:50.183158669 +0000 UTC m=+48.443125395" Mar 17 17:42:50.183396 kubelet[2662]: I0317 17:42:50.183289 2662 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-7fddcd7f99-92q58" podStartSLOduration=19.800644428 podStartE2EDuration="28.183284766s" podCreationTimestamp="2025-03-17 17:42:22 +0000 UTC" firstStartedPulling="2025-03-17 17:42:40.770402008 +0000 UTC m=+39.030368724" lastFinishedPulling="2025-03-17 17:42:49.153042336 +0000 UTC m=+47.413009062" observedRunningTime="2025-03-17 17:42:50.166955821 +0000 UTC m=+48.426922547" watchObservedRunningTime="2025-03-17 17:42:50.183284766 +0000 UTC m=+48.443251492" Mar 17 17:42:51.160437 kubelet[2662]: I0317 17:42:51.160407 2662 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 17 17:42:52.232931 containerd[1467]: time="2025-03-17T17:42:52.232835645Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:42:52.233758 containerd[1467]: time="2025-03-17T17:42:52.233724033Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.29.2: active requests=0, bytes read=13986843" Mar 17 17:42:52.235092 containerd[1467]: time="2025-03-17T17:42:52.235038842Z" level=info msg="ImageCreate event name:\"sha256:09a5a6ea58a48ac826468e05538c78d1378e103737124f1744efea8699fc29a8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:42:52.237143 containerd[1467]: time="2025-03-17T17:42:52.237099852Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:54ef0afa50feb3f691782e8d6df9a7f27d127a3af9bbcbd0bcdadac98e8be8e3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 17 17:42:52.237729 containerd[1467]: time="2025-03-17T17:42:52.237690220Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.2\" with image id \"sha256:09a5a6ea58a48ac826468e05538c78d1378e103737124f1744efea8699fc29a8\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.2\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:54ef0afa50feb3f691782e8d6df9a7f27d127a3af9bbcbd0bcdadac98e8be8e3\", size \"15479899\" in 2.427026778s" Mar 17 17:42:52.237729 containerd[1467]: time="2025-03-17T17:42:52.237721639Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.2\" returns image reference \"sha256:09a5a6ea58a48ac826468e05538c78d1378e103737124f1744efea8699fc29a8\"" Mar 17 17:42:52.240250 containerd[1467]: time="2025-03-17T17:42:52.239905059Z" level=info msg="CreateContainer within sandbox \"b5375e4dccdc52e071442404c4b8a1d48eb22ca0db08a28f6918d5cd419f96d4\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Mar 17 17:42:52.260038 containerd[1467]: time="2025-03-17T17:42:52.259974172Z" level=info msg="CreateContainer within sandbox \"b5375e4dccdc52e071442404c4b8a1d48eb22ca0db08a28f6918d5cd419f96d4\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"b37119fc7d811a132aa4fabeb370e180a012f2db2eeefa5260f13d48dd082941\"" Mar 17 17:42:52.260732 containerd[1467]: time="2025-03-17T17:42:52.260680810Z" level=info msg="StartContainer for \"b37119fc7d811a132aa4fabeb370e180a012f2db2eeefa5260f13d48dd082941\"" Mar 17 17:42:52.305010 systemd[1]: Started cri-containerd-b37119fc7d811a132aa4fabeb370e180a012f2db2eeefa5260f13d48dd082941.scope - libcontainer container b37119fc7d811a132aa4fabeb370e180a012f2db2eeefa5260f13d48dd082941. Mar 17 17:42:52.377663 containerd[1467]: time="2025-03-17T17:42:52.377606099Z" level=info msg="StartContainer for \"b37119fc7d811a132aa4fabeb370e180a012f2db2eeefa5260f13d48dd082941\" returns successfully" Mar 17 17:42:52.798623 systemd[1]: Started sshd@13-10.0.0.61:22-10.0.0.1:48892.service - OpenSSH per-connection server daemon (10.0.0.1:48892). Mar 17 17:42:52.849002 sshd[5839]: Accepted publickey for core from 10.0.0.1 port 48892 ssh2: RSA SHA256:j201F9FRK1q3ChnxQf0adNdYppDp+g37vmaXPvsVhek Mar 17 17:42:52.850938 sshd-session[5839]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 17 17:42:52.855122 systemd-logind[1453]: New session 14 of user core. Mar 17 17:42:52.870544 systemd[1]: Started session-14.scope - Session 14 of User core. Mar 17 17:42:52.957908 kubelet[2662]: I0317 17:42:52.957743 2662 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Mar 17 17:42:52.957908 kubelet[2662]: I0317 17:42:52.957778 2662 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Mar 17 17:42:53.003091 sshd[5841]: Connection closed by 10.0.0.1 port 48892 Mar 17 17:42:53.003614 sshd-session[5839]: pam_unix(sshd:session): session closed for user core Mar 17 17:42:53.012530 systemd[1]: sshd@13-10.0.0.61:22-10.0.0.1:48892.service: Deactivated successfully. Mar 17 17:42:53.015043 systemd[1]: session-14.scope: Deactivated successfully. Mar 17 17:42:53.017390 systemd-logind[1453]: Session 14 logged out. Waiting for processes to exit. Mar 17 17:42:53.022886 systemd[1]: Started sshd@14-10.0.0.61:22-10.0.0.1:48902.service - OpenSSH per-connection server daemon (10.0.0.1:48902). Mar 17 17:42:53.024188 systemd-logind[1453]: Removed session 14. Mar 17 17:42:53.062147 sshd[5854]: Accepted publickey for core from 10.0.0.1 port 48902 ssh2: RSA SHA256:j201F9FRK1q3ChnxQf0adNdYppDp+g37vmaXPvsVhek Mar 17 17:42:53.063715 sshd-session[5854]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 17 17:42:53.069416 systemd-logind[1453]: New session 15 of user core. Mar 17 17:42:53.076491 systemd[1]: Started session-15.scope - Session 15 of User core. Mar 17 17:42:53.287853 kubelet[2662]: I0317 17:42:53.286955 2662 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-j9gdg" podStartSLOduration=19.655340149 podStartE2EDuration="31.286932994s" podCreationTimestamp="2025-03-17 17:42:22 +0000 UTC" firstStartedPulling="2025-03-17 17:42:40.606996795 +0000 UTC m=+38.866963521" lastFinishedPulling="2025-03-17 17:42:52.238589639 +0000 UTC m=+50.498556366" observedRunningTime="2025-03-17 17:42:53.286831223 +0000 UTC m=+51.546797939" watchObservedRunningTime="2025-03-17 17:42:53.286932994 +0000 UTC m=+51.546899720" Mar 17 17:42:53.349923 sshd[5856]: Connection closed by 10.0.0.1 port 48902 Mar 17 17:42:53.350735 sshd-session[5854]: pam_unix(sshd:session): session closed for user core Mar 17 17:42:53.358645 systemd[1]: sshd@14-10.0.0.61:22-10.0.0.1:48902.service: Deactivated successfully. Mar 17 17:42:53.362536 systemd[1]: session-15.scope: Deactivated successfully. Mar 17 17:42:53.364616 systemd-logind[1453]: Session 15 logged out. Waiting for processes to exit. Mar 17 17:42:53.372881 systemd[1]: Started sshd@15-10.0.0.61:22-10.0.0.1:48904.service - OpenSSH per-connection server daemon (10.0.0.1:48904). Mar 17 17:42:53.373961 systemd-logind[1453]: Removed session 15. Mar 17 17:42:53.405825 sshd[5867]: Accepted publickey for core from 10.0.0.1 port 48904 ssh2: RSA SHA256:j201F9FRK1q3ChnxQf0adNdYppDp+g37vmaXPvsVhek Mar 17 17:42:53.407256 sshd-session[5867]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 17 17:42:53.411348 systemd-logind[1453]: New session 16 of user core. Mar 17 17:42:53.420482 systemd[1]: Started session-16.scope - Session 16 of User core. Mar 17 17:42:53.614837 sshd[5869]: Connection closed by 10.0.0.1 port 48904 Mar 17 17:42:53.615704 sshd-session[5867]: pam_unix(sshd:session): session closed for user core Mar 17 17:42:53.620446 systemd[1]: sshd@15-10.0.0.61:22-10.0.0.1:48904.service: Deactivated successfully. Mar 17 17:42:53.622904 systemd[1]: session-16.scope: Deactivated successfully. Mar 17 17:42:53.623755 systemd-logind[1453]: Session 16 logged out. Waiting for processes to exit. Mar 17 17:42:53.624734 systemd-logind[1453]: Removed session 16. Mar 17 17:42:54.368045 kubelet[2662]: I0317 17:42:54.368004 2662 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 17 17:42:58.627909 systemd[1]: Started sshd@16-10.0.0.61:22-10.0.0.1:48916.service - OpenSSH per-connection server daemon (10.0.0.1:48916). Mar 17 17:42:58.674303 sshd[5887]: Accepted publickey for core from 10.0.0.1 port 48916 ssh2: RSA SHA256:j201F9FRK1q3ChnxQf0adNdYppDp+g37vmaXPvsVhek Mar 17 17:42:58.676097 sshd-session[5887]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 17 17:42:58.679860 systemd-logind[1453]: New session 17 of user core. Mar 17 17:42:58.688508 systemd[1]: Started session-17.scope - Session 17 of User core. Mar 17 17:42:58.801414 sshd[5889]: Connection closed by 10.0.0.1 port 48916 Mar 17 17:42:58.801836 sshd-session[5887]: pam_unix(sshd:session): session closed for user core Mar 17 17:42:58.805726 systemd[1]: sshd@16-10.0.0.61:22-10.0.0.1:48916.service: Deactivated successfully. Mar 17 17:42:58.808657 systemd[1]: session-17.scope: Deactivated successfully. Mar 17 17:42:58.809429 systemd-logind[1453]: Session 17 logged out. Waiting for processes to exit. Mar 17 17:42:58.810313 systemd-logind[1453]: Removed session 17. Mar 17 17:43:01.808278 containerd[1467]: time="2025-03-17T17:43:01.808232665Z" level=info msg="StopPodSandbox for \"e93aa7d450e66ea71243d70e3ef70d32837865b756a4bd00a6f296f08f7881d8\"" Mar 17 17:43:01.808843 containerd[1467]: time="2025-03-17T17:43:01.808373930Z" level=info msg="TearDown network for sandbox \"e93aa7d450e66ea71243d70e3ef70d32837865b756a4bd00a6f296f08f7881d8\" successfully" Mar 17 17:43:01.808843 containerd[1467]: time="2025-03-17T17:43:01.808420988Z" level=info msg="StopPodSandbox for \"e93aa7d450e66ea71243d70e3ef70d32837865b756a4bd00a6f296f08f7881d8\" returns successfully" Mar 17 17:43:01.808843 containerd[1467]: time="2025-03-17T17:43:01.808744576Z" level=info msg="RemovePodSandbox for \"e93aa7d450e66ea71243d70e3ef70d32837865b756a4bd00a6f296f08f7881d8\"" Mar 17 17:43:01.818196 containerd[1467]: time="2025-03-17T17:43:01.818167081Z" level=info msg="Forcibly stopping sandbox \"e93aa7d450e66ea71243d70e3ef70d32837865b756a4bd00a6f296f08f7881d8\"" Mar 17 17:43:01.818286 containerd[1467]: time="2025-03-17T17:43:01.818268311Z" level=info msg="TearDown network for sandbox \"e93aa7d450e66ea71243d70e3ef70d32837865b756a4bd00a6f296f08f7881d8\" successfully" Mar 17 17:43:01.936838 containerd[1467]: time="2025-03-17T17:43:01.936771249Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"e93aa7d450e66ea71243d70e3ef70d32837865b756a4bd00a6f296f08f7881d8\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:43:01.936973 containerd[1467]: time="2025-03-17T17:43:01.936855467Z" level=info msg="RemovePodSandbox \"e93aa7d450e66ea71243d70e3ef70d32837865b756a4bd00a6f296f08f7881d8\" returns successfully" Mar 17 17:43:01.937323 containerd[1467]: time="2025-03-17T17:43:01.937285865Z" level=info msg="StopPodSandbox for \"3f5345f62ce3def9878f952bc3b9ba8515b066f140a947d652235852cb08fae5\"" Mar 17 17:43:01.937445 containerd[1467]: time="2025-03-17T17:43:01.937420177Z" level=info msg="TearDown network for sandbox \"3f5345f62ce3def9878f952bc3b9ba8515b066f140a947d652235852cb08fae5\" successfully" Mar 17 17:43:01.937445 containerd[1467]: time="2025-03-17T17:43:01.937436477Z" level=info msg="StopPodSandbox for \"3f5345f62ce3def9878f952bc3b9ba8515b066f140a947d652235852cb08fae5\" returns successfully" Mar 17 17:43:01.937765 containerd[1467]: time="2025-03-17T17:43:01.937730749Z" level=info msg="RemovePodSandbox for \"3f5345f62ce3def9878f952bc3b9ba8515b066f140a947d652235852cb08fae5\"" Mar 17 17:43:01.937765 containerd[1467]: time="2025-03-17T17:43:01.937756037Z" level=info msg="Forcibly stopping sandbox \"3f5345f62ce3def9878f952bc3b9ba8515b066f140a947d652235852cb08fae5\"" Mar 17 17:43:01.937879 containerd[1467]: time="2025-03-17T17:43:01.937822422Z" level=info msg="TearDown network for sandbox \"3f5345f62ce3def9878f952bc3b9ba8515b066f140a947d652235852cb08fae5\" successfully" Mar 17 17:43:02.002772 containerd[1467]: time="2025-03-17T17:43:02.002734820Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"3f5345f62ce3def9878f952bc3b9ba8515b066f140a947d652235852cb08fae5\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:43:02.002772 containerd[1467]: time="2025-03-17T17:43:02.002783031Z" level=info msg="RemovePodSandbox \"3f5345f62ce3def9878f952bc3b9ba8515b066f140a947d652235852cb08fae5\" returns successfully" Mar 17 17:43:02.003208 containerd[1467]: time="2025-03-17T17:43:02.003144250Z" level=info msg="StopPodSandbox for \"4044ac7223d418a825ef4eedf876ba4769d2804921eba6925b3d0d6d4afbc33a\"" Mar 17 17:43:02.003456 containerd[1467]: time="2025-03-17T17:43:02.003265637Z" level=info msg="TearDown network for sandbox \"4044ac7223d418a825ef4eedf876ba4769d2804921eba6925b3d0d6d4afbc33a\" successfully" Mar 17 17:43:02.003456 containerd[1467]: time="2025-03-17T17:43:02.003283280Z" level=info msg="StopPodSandbox for \"4044ac7223d418a825ef4eedf876ba4769d2804921eba6925b3d0d6d4afbc33a\" returns successfully" Mar 17 17:43:02.003756 containerd[1467]: time="2025-03-17T17:43:02.003725881Z" level=info msg="RemovePodSandbox for \"4044ac7223d418a825ef4eedf876ba4769d2804921eba6925b3d0d6d4afbc33a\"" Mar 17 17:43:02.003807 containerd[1467]: time="2025-03-17T17:43:02.003765986Z" level=info msg="Forcibly stopping sandbox \"4044ac7223d418a825ef4eedf876ba4769d2804921eba6925b3d0d6d4afbc33a\"" Mar 17 17:43:02.003948 containerd[1467]: time="2025-03-17T17:43:02.003885180Z" level=info msg="TearDown network for sandbox \"4044ac7223d418a825ef4eedf876ba4769d2804921eba6925b3d0d6d4afbc33a\" successfully" Mar 17 17:43:02.028088 containerd[1467]: time="2025-03-17T17:43:02.028036210Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"4044ac7223d418a825ef4eedf876ba4769d2804921eba6925b3d0d6d4afbc33a\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:43:02.028235 containerd[1467]: time="2025-03-17T17:43:02.028112803Z" level=info msg="RemovePodSandbox \"4044ac7223d418a825ef4eedf876ba4769d2804921eba6925b3d0d6d4afbc33a\" returns successfully" Mar 17 17:43:02.028834 containerd[1467]: time="2025-03-17T17:43:02.028640213Z" level=info msg="StopPodSandbox for \"eec0898e435269a73cab252710c67519bb65ec8c70880a65c0d84177e2d5547f\"" Mar 17 17:43:02.028834 containerd[1467]: time="2025-03-17T17:43:02.028754328Z" level=info msg="TearDown network for sandbox \"eec0898e435269a73cab252710c67519bb65ec8c70880a65c0d84177e2d5547f\" successfully" Mar 17 17:43:02.028834 containerd[1467]: time="2025-03-17T17:43:02.028764527Z" level=info msg="StopPodSandbox for \"eec0898e435269a73cab252710c67519bb65ec8c70880a65c0d84177e2d5547f\" returns successfully" Mar 17 17:43:02.029020 containerd[1467]: time="2025-03-17T17:43:02.029000069Z" level=info msg="RemovePodSandbox for \"eec0898e435269a73cab252710c67519bb65ec8c70880a65c0d84177e2d5547f\"" Mar 17 17:43:02.029020 containerd[1467]: time="2025-03-17T17:43:02.029016480Z" level=info msg="Forcibly stopping sandbox \"eec0898e435269a73cab252710c67519bb65ec8c70880a65c0d84177e2d5547f\"" Mar 17 17:43:02.029123 containerd[1467]: time="2025-03-17T17:43:02.029076943Z" level=info msg="TearDown network for sandbox \"eec0898e435269a73cab252710c67519bb65ec8c70880a65c0d84177e2d5547f\" successfully" Mar 17 17:43:02.077416 containerd[1467]: time="2025-03-17T17:43:02.077243618Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"eec0898e435269a73cab252710c67519bb65ec8c70880a65c0d84177e2d5547f\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:43:02.077416 containerd[1467]: time="2025-03-17T17:43:02.077328477Z" level=info msg="RemovePodSandbox \"eec0898e435269a73cab252710c67519bb65ec8c70880a65c0d84177e2d5547f\" returns successfully" Mar 17 17:43:02.077796 containerd[1467]: time="2025-03-17T17:43:02.077771118Z" level=info msg="StopPodSandbox for \"1f73f191946516c6303f3e08ba9764f9ef1ed363435b102ed8343391428b48bc\"" Mar 17 17:43:02.078143 containerd[1467]: time="2025-03-17T17:43:02.078102110Z" level=info msg="TearDown network for sandbox \"1f73f191946516c6303f3e08ba9764f9ef1ed363435b102ed8343391428b48bc\" successfully" Mar 17 17:43:02.078143 containerd[1467]: time="2025-03-17T17:43:02.078125103Z" level=info msg="StopPodSandbox for \"1f73f191946516c6303f3e08ba9764f9ef1ed363435b102ed8343391428b48bc\" returns successfully" Mar 17 17:43:02.078494 containerd[1467]: time="2025-03-17T17:43:02.078455673Z" level=info msg="RemovePodSandbox for \"1f73f191946516c6303f3e08ba9764f9ef1ed363435b102ed8343391428b48bc\"" Mar 17 17:43:02.078548 containerd[1467]: time="2025-03-17T17:43:02.078495598Z" level=info msg="Forcibly stopping sandbox \"1f73f191946516c6303f3e08ba9764f9ef1ed363435b102ed8343391428b48bc\"" Mar 17 17:43:02.078660 containerd[1467]: time="2025-03-17T17:43:02.078603170Z" level=info msg="TearDown network for sandbox \"1f73f191946516c6303f3e08ba9764f9ef1ed363435b102ed8343391428b48bc\" successfully" Mar 17 17:43:02.143387 containerd[1467]: time="2025-03-17T17:43:02.143302555Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"1f73f191946516c6303f3e08ba9764f9ef1ed363435b102ed8343391428b48bc\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:43:02.143387 containerd[1467]: time="2025-03-17T17:43:02.143395700Z" level=info msg="RemovePodSandbox \"1f73f191946516c6303f3e08ba9764f9ef1ed363435b102ed8343391428b48bc\" returns successfully" Mar 17 17:43:02.143944 containerd[1467]: time="2025-03-17T17:43:02.143887955Z" level=info msg="StopPodSandbox for \"58f98dec59f83fef7090b948443c29832b1906274dd91e7d0c3b3c902d3df763\"" Mar 17 17:43:02.144022 containerd[1467]: time="2025-03-17T17:43:02.144001457Z" level=info msg="TearDown network for sandbox \"58f98dec59f83fef7090b948443c29832b1906274dd91e7d0c3b3c902d3df763\" successfully" Mar 17 17:43:02.144022 containerd[1467]: time="2025-03-17T17:43:02.144016937Z" level=info msg="StopPodSandbox for \"58f98dec59f83fef7090b948443c29832b1906274dd91e7d0c3b3c902d3df763\" returns successfully" Mar 17 17:43:02.144271 containerd[1467]: time="2025-03-17T17:43:02.144229095Z" level=info msg="RemovePodSandbox for \"58f98dec59f83fef7090b948443c29832b1906274dd91e7d0c3b3c902d3df763\"" Mar 17 17:43:02.144271 containerd[1467]: time="2025-03-17T17:43:02.144263649Z" level=info msg="Forcibly stopping sandbox \"58f98dec59f83fef7090b948443c29832b1906274dd91e7d0c3b3c902d3df763\"" Mar 17 17:43:02.144401 containerd[1467]: time="2025-03-17T17:43:02.144327900Z" level=info msg="TearDown network for sandbox \"58f98dec59f83fef7090b948443c29832b1906274dd91e7d0c3b3c902d3df763\" successfully" Mar 17 17:43:02.235902 containerd[1467]: time="2025-03-17T17:43:02.235851332Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"58f98dec59f83fef7090b948443c29832b1906274dd91e7d0c3b3c902d3df763\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:43:02.235981 containerd[1467]: time="2025-03-17T17:43:02.235936171Z" level=info msg="RemovePodSandbox \"58f98dec59f83fef7090b948443c29832b1906274dd91e7d0c3b3c902d3df763\" returns successfully" Mar 17 17:43:02.236284 containerd[1467]: time="2025-03-17T17:43:02.236244550Z" level=info msg="StopPodSandbox for \"cc57dbc98644c56867a38eb1c7016e6e03a565b54b9bea52d164d9463bd478d9\"" Mar 17 17:43:02.236402 containerd[1467]: time="2025-03-17T17:43:02.236377650Z" level=info msg="TearDown network for sandbox \"cc57dbc98644c56867a38eb1c7016e6e03a565b54b9bea52d164d9463bd478d9\" successfully" Mar 17 17:43:02.236433 containerd[1467]: time="2025-03-17T17:43:02.236402536Z" level=info msg="StopPodSandbox for \"cc57dbc98644c56867a38eb1c7016e6e03a565b54b9bea52d164d9463bd478d9\" returns successfully" Mar 17 17:43:02.236648 containerd[1467]: time="2025-03-17T17:43:02.236614474Z" level=info msg="RemovePodSandbox for \"cc57dbc98644c56867a38eb1c7016e6e03a565b54b9bea52d164d9463bd478d9\"" Mar 17 17:43:02.236648 containerd[1467]: time="2025-03-17T17:43:02.236636565Z" level=info msg="Forcibly stopping sandbox \"cc57dbc98644c56867a38eb1c7016e6e03a565b54b9bea52d164d9463bd478d9\"" Mar 17 17:43:02.236769 containerd[1467]: time="2025-03-17T17:43:02.236729801Z" level=info msg="TearDown network for sandbox \"cc57dbc98644c56867a38eb1c7016e6e03a565b54b9bea52d164d9463bd478d9\" successfully" Mar 17 17:43:02.330385 containerd[1467]: time="2025-03-17T17:43:02.330280819Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"cc57dbc98644c56867a38eb1c7016e6e03a565b54b9bea52d164d9463bd478d9\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:43:02.330385 containerd[1467]: time="2025-03-17T17:43:02.330333808Z" level=info msg="RemovePodSandbox \"cc57dbc98644c56867a38eb1c7016e6e03a565b54b9bea52d164d9463bd478d9\" returns successfully" Mar 17 17:43:02.330792 containerd[1467]: time="2025-03-17T17:43:02.330768394Z" level=info msg="StopPodSandbox for \"317763565ecd927b3fe20fd547beaf2029d8ad181454207c47710163dbd09a88\"" Mar 17 17:43:02.330872 containerd[1467]: time="2025-03-17T17:43:02.330853674Z" level=info msg="TearDown network for sandbox \"317763565ecd927b3fe20fd547beaf2029d8ad181454207c47710163dbd09a88\" successfully" Mar 17 17:43:02.330872 containerd[1467]: time="2025-03-17T17:43:02.330867330Z" level=info msg="StopPodSandbox for \"317763565ecd927b3fe20fd547beaf2029d8ad181454207c47710163dbd09a88\" returns successfully" Mar 17 17:43:02.331126 containerd[1467]: time="2025-03-17T17:43:02.331096600Z" level=info msg="RemovePodSandbox for \"317763565ecd927b3fe20fd547beaf2029d8ad181454207c47710163dbd09a88\"" Mar 17 17:43:02.331126 containerd[1467]: time="2025-03-17T17:43:02.331124943Z" level=info msg="Forcibly stopping sandbox \"317763565ecd927b3fe20fd547beaf2029d8ad181454207c47710163dbd09a88\"" Mar 17 17:43:02.331220 containerd[1467]: time="2025-03-17T17:43:02.331186749Z" level=info msg="TearDown network for sandbox \"317763565ecd927b3fe20fd547beaf2029d8ad181454207c47710163dbd09a88\" successfully" Mar 17 17:43:02.410361 containerd[1467]: time="2025-03-17T17:43:02.410328482Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"317763565ecd927b3fe20fd547beaf2029d8ad181454207c47710163dbd09a88\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:43:02.410428 containerd[1467]: time="2025-03-17T17:43:02.410399896Z" level=info msg="RemovePodSandbox \"317763565ecd927b3fe20fd547beaf2029d8ad181454207c47710163dbd09a88\" returns successfully" Mar 17 17:43:02.410798 containerd[1467]: time="2025-03-17T17:43:02.410745724Z" level=info msg="StopPodSandbox for \"980554b391378bf7ff18652b7dc7922f5d0a4743b3074f835149be5d75b9c811\"" Mar 17 17:43:02.410922 containerd[1467]: time="2025-03-17T17:43:02.410854129Z" level=info msg="TearDown network for sandbox \"980554b391378bf7ff18652b7dc7922f5d0a4743b3074f835149be5d75b9c811\" successfully" Mar 17 17:43:02.410922 containerd[1467]: time="2025-03-17T17:43:02.410868606Z" level=info msg="StopPodSandbox for \"980554b391378bf7ff18652b7dc7922f5d0a4743b3074f835149be5d75b9c811\" returns successfully" Mar 17 17:43:02.411113 containerd[1467]: time="2025-03-17T17:43:02.411091284Z" level=info msg="RemovePodSandbox for \"980554b391378bf7ff18652b7dc7922f5d0a4743b3074f835149be5d75b9c811\"" Mar 17 17:43:02.411143 containerd[1467]: time="2025-03-17T17:43:02.411113295Z" level=info msg="Forcibly stopping sandbox \"980554b391378bf7ff18652b7dc7922f5d0a4743b3074f835149be5d75b9c811\"" Mar 17 17:43:02.411208 containerd[1467]: time="2025-03-17T17:43:02.411176834Z" level=info msg="TearDown network for sandbox \"980554b391378bf7ff18652b7dc7922f5d0a4743b3074f835149be5d75b9c811\" successfully" Mar 17 17:43:02.489227 containerd[1467]: time="2025-03-17T17:43:02.489190770Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"980554b391378bf7ff18652b7dc7922f5d0a4743b3074f835149be5d75b9c811\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:43:02.489294 containerd[1467]: time="2025-03-17T17:43:02.489239812Z" level=info msg="RemovePodSandbox \"980554b391378bf7ff18652b7dc7922f5d0a4743b3074f835149be5d75b9c811\" returns successfully" Mar 17 17:43:02.489569 containerd[1467]: time="2025-03-17T17:43:02.489544133Z" level=info msg="StopPodSandbox for \"b71e2358cd15f8a4cec878737cb78ea057ba63b0bc401daa2c0c7f9f8d7fef30\"" Mar 17 17:43:02.489662 containerd[1467]: time="2025-03-17T17:43:02.489624754Z" level=info msg="TearDown network for sandbox \"b71e2358cd15f8a4cec878737cb78ea057ba63b0bc401daa2c0c7f9f8d7fef30\" successfully" Mar 17 17:43:02.489662 containerd[1467]: time="2025-03-17T17:43:02.489645153Z" level=info msg="StopPodSandbox for \"b71e2358cd15f8a4cec878737cb78ea057ba63b0bc401daa2c0c7f9f8d7fef30\" returns successfully" Mar 17 17:43:02.489884 containerd[1467]: time="2025-03-17T17:43:02.489861729Z" level=info msg="RemovePodSandbox for \"b71e2358cd15f8a4cec878737cb78ea057ba63b0bc401daa2c0c7f9f8d7fef30\"" Mar 17 17:43:02.489947 containerd[1467]: time="2025-03-17T17:43:02.489886997Z" level=info msg="Forcibly stopping sandbox \"b71e2358cd15f8a4cec878737cb78ea057ba63b0bc401daa2c0c7f9f8d7fef30\"" Mar 17 17:43:02.490069 containerd[1467]: time="2025-03-17T17:43:02.490030576Z" level=info msg="TearDown network for sandbox \"b71e2358cd15f8a4cec878737cb78ea057ba63b0bc401daa2c0c7f9f8d7fef30\" successfully" Mar 17 17:43:02.544975 containerd[1467]: time="2025-03-17T17:43:02.544935990Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"b71e2358cd15f8a4cec878737cb78ea057ba63b0bc401daa2c0c7f9f8d7fef30\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:43:02.544975 containerd[1467]: time="2025-03-17T17:43:02.544973851Z" level=info msg="RemovePodSandbox \"b71e2358cd15f8a4cec878737cb78ea057ba63b0bc401daa2c0c7f9f8d7fef30\" returns successfully" Mar 17 17:43:02.545244 containerd[1467]: time="2025-03-17T17:43:02.545208982Z" level=info msg="StopPodSandbox for \"84f2f05c82276b89c454b200d6c53ebf68e120d900db1ee2b4b76ce464926d93\"" Mar 17 17:43:02.545295 containerd[1467]: time="2025-03-17T17:43:02.545283211Z" level=info msg="TearDown network for sandbox \"84f2f05c82276b89c454b200d6c53ebf68e120d900db1ee2b4b76ce464926d93\" successfully" Mar 17 17:43:02.545295 containerd[1467]: time="2025-03-17T17:43:02.545291767Z" level=info msg="StopPodSandbox for \"84f2f05c82276b89c454b200d6c53ebf68e120d900db1ee2b4b76ce464926d93\" returns successfully" Mar 17 17:43:02.545500 containerd[1467]: time="2025-03-17T17:43:02.545477846Z" level=info msg="RemovePodSandbox for \"84f2f05c82276b89c454b200d6c53ebf68e120d900db1ee2b4b76ce464926d93\"" Mar 17 17:43:02.545500 containerd[1467]: time="2025-03-17T17:43:02.545495429Z" level=info msg="Forcibly stopping sandbox \"84f2f05c82276b89c454b200d6c53ebf68e120d900db1ee2b4b76ce464926d93\"" Mar 17 17:43:02.545642 containerd[1467]: time="2025-03-17T17:43:02.545575479Z" level=info msg="TearDown network for sandbox \"84f2f05c82276b89c454b200d6c53ebf68e120d900db1ee2b4b76ce464926d93\" successfully" Mar 17 17:43:02.680733 containerd[1467]: time="2025-03-17T17:43:02.680698715Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"84f2f05c82276b89c454b200d6c53ebf68e120d900db1ee2b4b76ce464926d93\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:43:02.680838 containerd[1467]: time="2025-03-17T17:43:02.680751584Z" level=info msg="RemovePodSandbox \"84f2f05c82276b89c454b200d6c53ebf68e120d900db1ee2b4b76ce464926d93\" returns successfully" Mar 17 17:43:02.681086 containerd[1467]: time="2025-03-17T17:43:02.681065643Z" level=info msg="StopPodSandbox for \"a45255c5e842412af83dd06ba8b258ef311550cba3da5240506cd00cf0678895\"" Mar 17 17:43:02.681209 containerd[1467]: time="2025-03-17T17:43:02.681184196Z" level=info msg="TearDown network for sandbox \"a45255c5e842412af83dd06ba8b258ef311550cba3da5240506cd00cf0678895\" successfully" Mar 17 17:43:02.681209 containerd[1467]: time="2025-03-17T17:43:02.681203191Z" level=info msg="StopPodSandbox for \"a45255c5e842412af83dd06ba8b258ef311550cba3da5240506cd00cf0678895\" returns successfully" Mar 17 17:43:02.681456 containerd[1467]: time="2025-03-17T17:43:02.681425909Z" level=info msg="RemovePodSandbox for \"a45255c5e842412af83dd06ba8b258ef311550cba3da5240506cd00cf0678895\"" Mar 17 17:43:02.681456 containerd[1467]: time="2025-03-17T17:43:02.681448231Z" level=info msg="Forcibly stopping sandbox \"a45255c5e842412af83dd06ba8b258ef311550cba3da5240506cd00cf0678895\"" Mar 17 17:43:02.681524 containerd[1467]: time="2025-03-17T17:43:02.681510358Z" level=info msg="TearDown network for sandbox \"a45255c5e842412af83dd06ba8b258ef311550cba3da5240506cd00cf0678895\" successfully" Mar 17 17:43:02.778982 containerd[1467]: time="2025-03-17T17:43:02.778942019Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"a45255c5e842412af83dd06ba8b258ef311550cba3da5240506cd00cf0678895\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:43:02.779035 containerd[1467]: time="2025-03-17T17:43:02.778983367Z" level=info msg="RemovePodSandbox \"a45255c5e842412af83dd06ba8b258ef311550cba3da5240506cd00cf0678895\" returns successfully" Mar 17 17:43:02.779319 containerd[1467]: time="2025-03-17T17:43:02.779296014Z" level=info msg="StopPodSandbox for \"f905ddf13c6f6a970be32393ab7c951bb1766cec8d75bdc0aca8519fcc2783c7\"" Mar 17 17:43:02.779404 containerd[1467]: time="2025-03-17T17:43:02.779388267Z" level=info msg="TearDown network for sandbox \"f905ddf13c6f6a970be32393ab7c951bb1766cec8d75bdc0aca8519fcc2783c7\" successfully" Mar 17 17:43:02.779404 containerd[1467]: time="2025-03-17T17:43:02.779401472Z" level=info msg="StopPodSandbox for \"f905ddf13c6f6a970be32393ab7c951bb1766cec8d75bdc0aca8519fcc2783c7\" returns successfully" Mar 17 17:43:02.779751 containerd[1467]: time="2025-03-17T17:43:02.779713297Z" level=info msg="RemovePodSandbox for \"f905ddf13c6f6a970be32393ab7c951bb1766cec8d75bdc0aca8519fcc2783c7\"" Mar 17 17:43:02.779751 containerd[1467]: time="2025-03-17T17:43:02.779751539Z" level=info msg="Forcibly stopping sandbox \"f905ddf13c6f6a970be32393ab7c951bb1766cec8d75bdc0aca8519fcc2783c7\"" Mar 17 17:43:02.779901 containerd[1467]: time="2025-03-17T17:43:02.779850815Z" level=info msg="TearDown network for sandbox \"f905ddf13c6f6a970be32393ab7c951bb1766cec8d75bdc0aca8519fcc2783c7\" successfully" Mar 17 17:43:02.834961 containerd[1467]: time="2025-03-17T17:43:02.834928502Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"f905ddf13c6f6a970be32393ab7c951bb1766cec8d75bdc0aca8519fcc2783c7\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:43:02.835322 containerd[1467]: time="2025-03-17T17:43:02.834969899Z" level=info msg="RemovePodSandbox \"f905ddf13c6f6a970be32393ab7c951bb1766cec8d75bdc0aca8519fcc2783c7\" returns successfully" Mar 17 17:43:02.835322 containerd[1467]: time="2025-03-17T17:43:02.835185233Z" level=info msg="StopPodSandbox for \"4a6e2b6a3eecd39a342da3235b8bc51d6bca9e9c8c2f588a315458c197bf0390\"" Mar 17 17:43:02.835322 containerd[1467]: time="2025-03-17T17:43:02.835259573Z" level=info msg="TearDown network for sandbox \"4a6e2b6a3eecd39a342da3235b8bc51d6bca9e9c8c2f588a315458c197bf0390\" successfully" Mar 17 17:43:02.835322 containerd[1467]: time="2025-03-17T17:43:02.835269221Z" level=info msg="StopPodSandbox for \"4a6e2b6a3eecd39a342da3235b8bc51d6bca9e9c8c2f588a315458c197bf0390\" returns successfully" Mar 17 17:43:02.835501 containerd[1467]: time="2025-03-17T17:43:02.835479616Z" level=info msg="RemovePodSandbox for \"4a6e2b6a3eecd39a342da3235b8bc51d6bca9e9c8c2f588a315458c197bf0390\"" Mar 17 17:43:02.835538 containerd[1467]: time="2025-03-17T17:43:02.835501888Z" level=info msg="Forcibly stopping sandbox \"4a6e2b6a3eecd39a342da3235b8bc51d6bca9e9c8c2f588a315458c197bf0390\"" Mar 17 17:43:02.835593 containerd[1467]: time="2025-03-17T17:43:02.835558384Z" level=info msg="TearDown network for sandbox \"4a6e2b6a3eecd39a342da3235b8bc51d6bca9e9c8c2f588a315458c197bf0390\" successfully" Mar 17 17:43:02.880236 containerd[1467]: time="2025-03-17T17:43:02.880180866Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"4a6e2b6a3eecd39a342da3235b8bc51d6bca9e9c8c2f588a315458c197bf0390\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:43:02.880316 containerd[1467]: time="2025-03-17T17:43:02.880249907Z" level=info msg="RemovePodSandbox \"4a6e2b6a3eecd39a342da3235b8bc51d6bca9e9c8c2f588a315458c197bf0390\" returns successfully" Mar 17 17:43:02.880797 containerd[1467]: time="2025-03-17T17:43:02.880622135Z" level=info msg="StopPodSandbox for \"27220f5bf5e3f9e3a1064ebd76371f3363a6b06ed379f42a7bd72b4e7c55d55a\"" Mar 17 17:43:02.880797 containerd[1467]: time="2025-03-17T17:43:02.880737442Z" level=info msg="TearDown network for sandbox \"27220f5bf5e3f9e3a1064ebd76371f3363a6b06ed379f42a7bd72b4e7c55d55a\" successfully" Mar 17 17:43:02.880797 containerd[1467]: time="2025-03-17T17:43:02.880750596Z" level=info msg="StopPodSandbox for \"27220f5bf5e3f9e3a1064ebd76371f3363a6b06ed379f42a7bd72b4e7c55d55a\" returns successfully" Mar 17 17:43:02.881039 containerd[1467]: time="2025-03-17T17:43:02.881021274Z" level=info msg="RemovePodSandbox for \"27220f5bf5e3f9e3a1064ebd76371f3363a6b06ed379f42a7bd72b4e7c55d55a\"" Mar 17 17:43:02.881083 containerd[1467]: time="2025-03-17T17:43:02.881042143Z" level=info msg="Forcibly stopping sandbox \"27220f5bf5e3f9e3a1064ebd76371f3363a6b06ed379f42a7bd72b4e7c55d55a\"" Mar 17 17:43:02.881140 containerd[1467]: time="2025-03-17T17:43:02.881104470Z" level=info msg="TearDown network for sandbox \"27220f5bf5e3f9e3a1064ebd76371f3363a6b06ed379f42a7bd72b4e7c55d55a\" successfully" Mar 17 17:43:03.046175 containerd[1467]: time="2025-03-17T17:43:03.046062832Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"27220f5bf5e3f9e3a1064ebd76371f3363a6b06ed379f42a7bd72b4e7c55d55a\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:43:03.046175 containerd[1467]: time="2025-03-17T17:43:03.046110291Z" level=info msg="RemovePodSandbox \"27220f5bf5e3f9e3a1064ebd76371f3363a6b06ed379f42a7bd72b4e7c55d55a\" returns successfully" Mar 17 17:43:03.046550 containerd[1467]: time="2025-03-17T17:43:03.046497307Z" level=info msg="StopPodSandbox for \"e570c82c935c8544758ef95881b0e0abf15540c0e05ff8eaf1df19398bdcd032\"" Mar 17 17:43:03.046670 containerd[1467]: time="2025-03-17T17:43:03.046645956Z" level=info msg="TearDown network for sandbox \"e570c82c935c8544758ef95881b0e0abf15540c0e05ff8eaf1df19398bdcd032\" successfully" Mar 17 17:43:03.046670 containerd[1467]: time="2025-03-17T17:43:03.046663388Z" level=info msg="StopPodSandbox for \"e570c82c935c8544758ef95881b0e0abf15540c0e05ff8eaf1df19398bdcd032\" returns successfully" Mar 17 17:43:03.046964 containerd[1467]: time="2025-03-17T17:43:03.046914941Z" level=info msg="RemovePodSandbox for \"e570c82c935c8544758ef95881b0e0abf15540c0e05ff8eaf1df19398bdcd032\"" Mar 17 17:43:03.046964 containerd[1467]: time="2025-03-17T17:43:03.046946060Z" level=info msg="Forcibly stopping sandbox \"e570c82c935c8544758ef95881b0e0abf15540c0e05ff8eaf1df19398bdcd032\"" Mar 17 17:43:03.047123 containerd[1467]: time="2025-03-17T17:43:03.047042661Z" level=info msg="TearDown network for sandbox \"e570c82c935c8544758ef95881b0e0abf15540c0e05ff8eaf1df19398bdcd032\" successfully" Mar 17 17:43:03.128155 containerd[1467]: time="2025-03-17T17:43:03.128127946Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"e570c82c935c8544758ef95881b0e0abf15540c0e05ff8eaf1df19398bdcd032\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:43:03.128439 containerd[1467]: time="2025-03-17T17:43:03.128164625Z" level=info msg="RemovePodSandbox \"e570c82c935c8544758ef95881b0e0abf15540c0e05ff8eaf1df19398bdcd032\" returns successfully" Mar 17 17:43:03.128586 containerd[1467]: time="2025-03-17T17:43:03.128538307Z" level=info msg="StopPodSandbox for \"e983d9d5aee8c472a9806708fa057e3b2de9217794b9f28b9fc2133b98942f76\"" Mar 17 17:43:03.128720 containerd[1467]: time="2025-03-17T17:43:03.128691254Z" level=info msg="TearDown network for sandbox \"e983d9d5aee8c472a9806708fa057e3b2de9217794b9f28b9fc2133b98942f76\" successfully" Mar 17 17:43:03.128720 containerd[1467]: time="2025-03-17T17:43:03.128707645Z" level=info msg="StopPodSandbox for \"e983d9d5aee8c472a9806708fa057e3b2de9217794b9f28b9fc2133b98942f76\" returns successfully" Mar 17 17:43:03.129029 containerd[1467]: time="2025-03-17T17:43:03.128995806Z" level=info msg="RemovePodSandbox for \"e983d9d5aee8c472a9806708fa057e3b2de9217794b9f28b9fc2133b98942f76\"" Mar 17 17:43:03.129029 containerd[1467]: time="2025-03-17T17:43:03.129021634Z" level=info msg="Forcibly stopping sandbox \"e983d9d5aee8c472a9806708fa057e3b2de9217794b9f28b9fc2133b98942f76\"" Mar 17 17:43:03.129274 containerd[1467]: time="2025-03-17T17:43:03.129107726Z" level=info msg="TearDown network for sandbox \"e983d9d5aee8c472a9806708fa057e3b2de9217794b9f28b9fc2133b98942f76\" successfully" Mar 17 17:43:03.156292 containerd[1467]: time="2025-03-17T17:43:03.156257382Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"e983d9d5aee8c472a9806708fa057e3b2de9217794b9f28b9fc2133b98942f76\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:43:03.156369 containerd[1467]: time="2025-03-17T17:43:03.156298509Z" level=info msg="RemovePodSandbox \"e983d9d5aee8c472a9806708fa057e3b2de9217794b9f28b9fc2133b98942f76\" returns successfully" Mar 17 17:43:03.156759 containerd[1467]: time="2025-03-17T17:43:03.156723016Z" level=info msg="StopPodSandbox for \"f1d105acdc9b53357e4e621942df86a7fba22a3900a30edbb033cb1322dbced2\"" Mar 17 17:43:03.156881 containerd[1467]: time="2025-03-17T17:43:03.156860804Z" level=info msg="TearDown network for sandbox \"f1d105acdc9b53357e4e621942df86a7fba22a3900a30edbb033cb1322dbced2\" successfully" Mar 17 17:43:03.156881 containerd[1467]: time="2025-03-17T17:43:03.156875251Z" level=info msg="StopPodSandbox for \"f1d105acdc9b53357e4e621942df86a7fba22a3900a30edbb033cb1322dbced2\" returns successfully" Mar 17 17:43:03.157158 containerd[1467]: time="2025-03-17T17:43:03.157134127Z" level=info msg="RemovePodSandbox for \"f1d105acdc9b53357e4e621942df86a7fba22a3900a30edbb033cb1322dbced2\"" Mar 17 17:43:03.157198 containerd[1467]: time="2025-03-17T17:43:03.157168341Z" level=info msg="Forcibly stopping sandbox \"f1d105acdc9b53357e4e621942df86a7fba22a3900a30edbb033cb1322dbced2\"" Mar 17 17:43:03.157282 containerd[1467]: time="2025-03-17T17:43:03.157248121Z" level=info msg="TearDown network for sandbox \"f1d105acdc9b53357e4e621942df86a7fba22a3900a30edbb033cb1322dbced2\" successfully" Mar 17 17:43:03.185414 containerd[1467]: time="2025-03-17T17:43:03.185386213Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"f1d105acdc9b53357e4e621942df86a7fba22a3900a30edbb033cb1322dbced2\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:43:03.185468 containerd[1467]: time="2025-03-17T17:43:03.185421599Z" level=info msg="RemovePodSandbox \"f1d105acdc9b53357e4e621942df86a7fba22a3900a30edbb033cb1322dbced2\" returns successfully" Mar 17 17:43:03.185705 containerd[1467]: time="2025-03-17T17:43:03.185638186Z" level=info msg="StopPodSandbox for \"3edf99fd7e4c4bcd608727ff216f57eed4c9785e1d1f4d8c623f0c2aa6ca839a\"" Mar 17 17:43:03.185751 containerd[1467]: time="2025-03-17T17:43:03.185737111Z" level=info msg="TearDown network for sandbox \"3edf99fd7e4c4bcd608727ff216f57eed4c9785e1d1f4d8c623f0c2aa6ca839a\" successfully" Mar 17 17:43:03.185751 containerd[1467]: time="2025-03-17T17:43:03.185747942Z" level=info msg="StopPodSandbox for \"3edf99fd7e4c4bcd608727ff216f57eed4c9785e1d1f4d8c623f0c2aa6ca839a\" returns successfully" Mar 17 17:43:03.186086 containerd[1467]: time="2025-03-17T17:43:03.186060167Z" level=info msg="RemovePodSandbox for \"3edf99fd7e4c4bcd608727ff216f57eed4c9785e1d1f4d8c623f0c2aa6ca839a\"" Mar 17 17:43:03.186207 containerd[1467]: time="2025-03-17T17:43:03.186168591Z" level=info msg="Forcibly stopping sandbox \"3edf99fd7e4c4bcd608727ff216f57eed4c9785e1d1f4d8c623f0c2aa6ca839a\"" Mar 17 17:43:03.186295 containerd[1467]: time="2025-03-17T17:43:03.186255784Z" level=info msg="TearDown network for sandbox \"3edf99fd7e4c4bcd608727ff216f57eed4c9785e1d1f4d8c623f0c2aa6ca839a\" successfully" Mar 17 17:43:03.222899 containerd[1467]: time="2025-03-17T17:43:03.222858504Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"3edf99fd7e4c4bcd608727ff216f57eed4c9785e1d1f4d8c623f0c2aa6ca839a\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:43:03.222964 containerd[1467]: time="2025-03-17T17:43:03.222908077Z" level=info msg="RemovePodSandbox \"3edf99fd7e4c4bcd608727ff216f57eed4c9785e1d1f4d8c623f0c2aa6ca839a\" returns successfully" Mar 17 17:43:03.223179 containerd[1467]: time="2025-03-17T17:43:03.223149540Z" level=info msg="StopPodSandbox for \"008cee49813c10fd33682eb193a75ecd740653b4fd1a4a322bfcc088c6078101\"" Mar 17 17:43:03.223257 containerd[1467]: time="2025-03-17T17:43:03.223239889Z" level=info msg="TearDown network for sandbox \"008cee49813c10fd33682eb193a75ecd740653b4fd1a4a322bfcc088c6078101\" successfully" Mar 17 17:43:03.223257 containerd[1467]: time="2025-03-17T17:43:03.223251932Z" level=info msg="StopPodSandbox for \"008cee49813c10fd33682eb193a75ecd740653b4fd1a4a322bfcc088c6078101\" returns successfully" Mar 17 17:43:03.223524 containerd[1467]: time="2025-03-17T17:43:03.223500939Z" level=info msg="RemovePodSandbox for \"008cee49813c10fd33682eb193a75ecd740653b4fd1a4a322bfcc088c6078101\"" Mar 17 17:43:03.223582 containerd[1467]: time="2025-03-17T17:43:03.223532969Z" level=info msg="Forcibly stopping sandbox \"008cee49813c10fd33682eb193a75ecd740653b4fd1a4a322bfcc088c6078101\"" Mar 17 17:43:03.223654 containerd[1467]: time="2025-03-17T17:43:03.223629109Z" level=info msg="TearDown network for sandbox \"008cee49813c10fd33682eb193a75ecd740653b4fd1a4a322bfcc088c6078101\" successfully" Mar 17 17:43:03.244178 containerd[1467]: time="2025-03-17T17:43:03.244121511Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"008cee49813c10fd33682eb193a75ecd740653b4fd1a4a322bfcc088c6078101\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:43:03.244178 containerd[1467]: time="2025-03-17T17:43:03.244187105Z" level=info msg="RemovePodSandbox \"008cee49813c10fd33682eb193a75ecd740653b4fd1a4a322bfcc088c6078101\" returns successfully" Mar 17 17:43:03.244471 containerd[1467]: time="2025-03-17T17:43:03.244443275Z" level=info msg="StopPodSandbox for \"6dd83b0760f3d749a5709564d2613214f6b299b039d6f1b56dcb51a9cb223ea7\"" Mar 17 17:43:03.244585 containerd[1467]: time="2025-03-17T17:43:03.244524628Z" level=info msg="TearDown network for sandbox \"6dd83b0760f3d749a5709564d2613214f6b299b039d6f1b56dcb51a9cb223ea7\" successfully" Mar 17 17:43:03.244585 containerd[1467]: time="2025-03-17T17:43:03.244543914Z" level=info msg="StopPodSandbox for \"6dd83b0760f3d749a5709564d2613214f6b299b039d6f1b56dcb51a9cb223ea7\" returns successfully" Mar 17 17:43:03.244779 containerd[1467]: time="2025-03-17T17:43:03.244746614Z" level=info msg="RemovePodSandbox for \"6dd83b0760f3d749a5709564d2613214f6b299b039d6f1b56dcb51a9cb223ea7\"" Mar 17 17:43:03.244779 containerd[1467]: time="2025-03-17T17:43:03.244772914Z" level=info msg="Forcibly stopping sandbox \"6dd83b0760f3d749a5709564d2613214f6b299b039d6f1b56dcb51a9cb223ea7\"" Mar 17 17:43:03.244893 containerd[1467]: time="2025-03-17T17:43:03.244854146Z" level=info msg="TearDown network for sandbox \"6dd83b0760f3d749a5709564d2613214f6b299b039d6f1b56dcb51a9cb223ea7\" successfully" Mar 17 17:43:03.284494 containerd[1467]: time="2025-03-17T17:43:03.284436718Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"6dd83b0760f3d749a5709564d2613214f6b299b039d6f1b56dcb51a9cb223ea7\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:43:03.284563 containerd[1467]: time="2025-03-17T17:43:03.284519834Z" level=info msg="RemovePodSandbox \"6dd83b0760f3d749a5709564d2613214f6b299b039d6f1b56dcb51a9cb223ea7\" returns successfully" Mar 17 17:43:03.284879 containerd[1467]: time="2025-03-17T17:43:03.284845104Z" level=info msg="StopPodSandbox for \"9906341f06aa2f7e95fd6b7d0f4548be4a6da1921703748eb9f0e054258d0bf4\"" Mar 17 17:43:03.284974 containerd[1467]: time="2025-03-17T17:43:03.284951964Z" level=info msg="TearDown network for sandbox \"9906341f06aa2f7e95fd6b7d0f4548be4a6da1921703748eb9f0e054258d0bf4\" successfully" Mar 17 17:43:03.284974 containerd[1467]: time="2025-03-17T17:43:03.284962895Z" level=info msg="StopPodSandbox for \"9906341f06aa2f7e95fd6b7d0f4548be4a6da1921703748eb9f0e054258d0bf4\" returns successfully" Mar 17 17:43:03.285249 containerd[1467]: time="2025-03-17T17:43:03.285211321Z" level=info msg="RemovePodSandbox for \"9906341f06aa2f7e95fd6b7d0f4548be4a6da1921703748eb9f0e054258d0bf4\"" Mar 17 17:43:03.285249 containerd[1467]: time="2025-03-17T17:43:03.285241578Z" level=info msg="Forcibly stopping sandbox \"9906341f06aa2f7e95fd6b7d0f4548be4a6da1921703748eb9f0e054258d0bf4\"" Mar 17 17:43:03.285422 containerd[1467]: time="2025-03-17T17:43:03.285319564Z" level=info msg="TearDown network for sandbox \"9906341f06aa2f7e95fd6b7d0f4548be4a6da1921703748eb9f0e054258d0bf4\" successfully" Mar 17 17:43:03.326818 containerd[1467]: time="2025-03-17T17:43:03.326700752Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"9906341f06aa2f7e95fd6b7d0f4548be4a6da1921703748eb9f0e054258d0bf4\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:43:03.332314 containerd[1467]: time="2025-03-17T17:43:03.332278590Z" level=info msg="RemovePodSandbox \"9906341f06aa2f7e95fd6b7d0f4548be4a6da1921703748eb9f0e054258d0bf4\" returns successfully" Mar 17 17:43:03.332726 containerd[1467]: time="2025-03-17T17:43:03.332694881Z" level=info msg="StopPodSandbox for \"6c94b5b9e7a3d7b63ccc9268b52bbe37ef401118cf01d372890eeea51b3ec01a\"" Mar 17 17:43:03.332856 containerd[1467]: time="2025-03-17T17:43:03.332791392Z" level=info msg="TearDown network for sandbox \"6c94b5b9e7a3d7b63ccc9268b52bbe37ef401118cf01d372890eeea51b3ec01a\" successfully" Mar 17 17:43:03.332856 containerd[1467]: time="2025-03-17T17:43:03.332837148Z" level=info msg="StopPodSandbox for \"6c94b5b9e7a3d7b63ccc9268b52bbe37ef401118cf01d372890eeea51b3ec01a\" returns successfully" Mar 17 17:43:03.333253 containerd[1467]: time="2025-03-17T17:43:03.333214167Z" level=info msg="RemovePodSandbox for \"6c94b5b9e7a3d7b63ccc9268b52bbe37ef401118cf01d372890eeea51b3ec01a\"" Mar 17 17:43:03.333253 containerd[1467]: time="2025-03-17T17:43:03.333251967Z" level=info msg="Forcibly stopping sandbox \"6c94b5b9e7a3d7b63ccc9268b52bbe37ef401118cf01d372890eeea51b3ec01a\"" Mar 17 17:43:03.333415 containerd[1467]: time="2025-03-17T17:43:03.333347196Z" level=info msg="TearDown network for sandbox \"6c94b5b9e7a3d7b63ccc9268b52bbe37ef401118cf01d372890eeea51b3ec01a\" successfully" Mar 17 17:43:03.373717 containerd[1467]: time="2025-03-17T17:43:03.373673874Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"6c94b5b9e7a3d7b63ccc9268b52bbe37ef401118cf01d372890eeea51b3ec01a\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:43:03.373867 containerd[1467]: time="2025-03-17T17:43:03.373732203Z" level=info msg="RemovePodSandbox \"6c94b5b9e7a3d7b63ccc9268b52bbe37ef401118cf01d372890eeea51b3ec01a\" returns successfully" Mar 17 17:43:03.381119 containerd[1467]: time="2025-03-17T17:43:03.381082649Z" level=info msg="StopPodSandbox for \"3f78c7317a0c3224f4039412ac38faee93bb589fa11754f46d7d13d15315c3d5\"" Mar 17 17:43:03.381207 containerd[1467]: time="2025-03-17T17:43:03.381181765Z" level=info msg="TearDown network for sandbox \"3f78c7317a0c3224f4039412ac38faee93bb589fa11754f46d7d13d15315c3d5\" successfully" Mar 17 17:43:03.381207 containerd[1467]: time="2025-03-17T17:43:03.381196112Z" level=info msg="StopPodSandbox for \"3f78c7317a0c3224f4039412ac38faee93bb589fa11754f46d7d13d15315c3d5\" returns successfully" Mar 17 17:43:03.381529 containerd[1467]: time="2025-03-17T17:43:03.381494011Z" level=info msg="RemovePodSandbox for \"3f78c7317a0c3224f4039412ac38faee93bb589fa11754f46d7d13d15315c3d5\"" Mar 17 17:43:03.381529 containerd[1467]: time="2025-03-17T17:43:03.381516193Z" level=info msg="Forcibly stopping sandbox \"3f78c7317a0c3224f4039412ac38faee93bb589fa11754f46d7d13d15315c3d5\"" Mar 17 17:43:03.381727 containerd[1467]: time="2025-03-17T17:43:03.381580613Z" level=info msg="TearDown network for sandbox \"3f78c7317a0c3224f4039412ac38faee93bb589fa11754f46d7d13d15315c3d5\" successfully" Mar 17 17:43:03.460983 containerd[1467]: time="2025-03-17T17:43:03.460925504Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"3f78c7317a0c3224f4039412ac38faee93bb589fa11754f46d7d13d15315c3d5\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:43:03.460983 containerd[1467]: time="2025-03-17T17:43:03.460975798Z" level=info msg="RemovePodSandbox \"3f78c7317a0c3224f4039412ac38faee93bb589fa11754f46d7d13d15315c3d5\" returns successfully" Mar 17 17:43:03.461329 containerd[1467]: time="2025-03-17T17:43:03.461301990Z" level=info msg="StopPodSandbox for \"284f791e9331db53c5a68aa2679f439389dd260cc6a34b8aff9bc622e96ca04d\"" Mar 17 17:43:03.461444 containerd[1467]: time="2025-03-17T17:43:03.461415704Z" level=info msg="TearDown network for sandbox \"284f791e9331db53c5a68aa2679f439389dd260cc6a34b8aff9bc622e96ca04d\" successfully" Mar 17 17:43:03.461444 containerd[1467]: time="2025-03-17T17:43:03.461425592Z" level=info msg="StopPodSandbox for \"284f791e9331db53c5a68aa2679f439389dd260cc6a34b8aff9bc622e96ca04d\" returns successfully" Mar 17 17:43:03.461855 containerd[1467]: time="2025-03-17T17:43:03.461820092Z" level=info msg="RemovePodSandbox for \"284f791e9331db53c5a68aa2679f439389dd260cc6a34b8aff9bc622e96ca04d\"" Mar 17 17:43:03.461898 containerd[1467]: time="2025-03-17T17:43:03.461859616Z" level=info msg="Forcibly stopping sandbox \"284f791e9331db53c5a68aa2679f439389dd260cc6a34b8aff9bc622e96ca04d\"" Mar 17 17:43:03.462028 containerd[1467]: time="2025-03-17T17:43:03.461974412Z" level=info msg="TearDown network for sandbox \"284f791e9331db53c5a68aa2679f439389dd260cc6a34b8aff9bc622e96ca04d\" successfully" Mar 17 17:43:03.521825 containerd[1467]: time="2025-03-17T17:43:03.521780530Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"284f791e9331db53c5a68aa2679f439389dd260cc6a34b8aff9bc622e96ca04d\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:43:03.521880 containerd[1467]: time="2025-03-17T17:43:03.521841424Z" level=info msg="RemovePodSandbox \"284f791e9331db53c5a68aa2679f439389dd260cc6a34b8aff9bc622e96ca04d\" returns successfully" Mar 17 17:43:03.522230 containerd[1467]: time="2025-03-17T17:43:03.522205397Z" level=info msg="StopPodSandbox for \"9eba50d4b991c21e718945f4ae399cbc0ccc0b0ef2c96bdf7d5d5e8361566abd\"" Mar 17 17:43:03.522311 containerd[1467]: time="2025-03-17T17:43:03.522294745Z" level=info msg="TearDown network for sandbox \"9eba50d4b991c21e718945f4ae399cbc0ccc0b0ef2c96bdf7d5d5e8361566abd\" successfully" Mar 17 17:43:03.522311 containerd[1467]: time="2025-03-17T17:43:03.522306567Z" level=info msg="StopPodSandbox for \"9eba50d4b991c21e718945f4ae399cbc0ccc0b0ef2c96bdf7d5d5e8361566abd\" returns successfully" Mar 17 17:43:03.522589 containerd[1467]: time="2025-03-17T17:43:03.522569892Z" level=info msg="RemovePodSandbox for \"9eba50d4b991c21e718945f4ae399cbc0ccc0b0ef2c96bdf7d5d5e8361566abd\"" Mar 17 17:43:03.522640 containerd[1467]: time="2025-03-17T17:43:03.522590801Z" level=info msg="Forcibly stopping sandbox \"9eba50d4b991c21e718945f4ae399cbc0ccc0b0ef2c96bdf7d5d5e8361566abd\"" Mar 17 17:43:03.522698 containerd[1467]: time="2025-03-17T17:43:03.522662316Z" level=info msg="TearDown network for sandbox \"9eba50d4b991c21e718945f4ae399cbc0ccc0b0ef2c96bdf7d5d5e8361566abd\" successfully" Mar 17 17:43:03.572673 containerd[1467]: time="2025-03-17T17:43:03.572573722Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"9eba50d4b991c21e718945f4ae399cbc0ccc0b0ef2c96bdf7d5d5e8361566abd\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:43:03.572673 containerd[1467]: time="2025-03-17T17:43:03.572663530Z" level=info msg="RemovePodSandbox \"9eba50d4b991c21e718945f4ae399cbc0ccc0b0ef2c96bdf7d5d5e8361566abd\" returns successfully" Mar 17 17:43:03.573378 containerd[1467]: time="2025-03-17T17:43:03.573193966Z" level=info msg="StopPodSandbox for \"86089e35a842c8c8e7ea82ce5cbce4b7646d35903cb49ceaccd5d1db0dfbc90c\"" Mar 17 17:43:03.573378 containerd[1467]: time="2025-03-17T17:43:03.573299364Z" level=info msg="TearDown network for sandbox \"86089e35a842c8c8e7ea82ce5cbce4b7646d35903cb49ceaccd5d1db0dfbc90c\" successfully" Mar 17 17:43:03.573378 containerd[1467]: time="2025-03-17T17:43:03.573309283Z" level=info msg="StopPodSandbox for \"86089e35a842c8c8e7ea82ce5cbce4b7646d35903cb49ceaccd5d1db0dfbc90c\" returns successfully" Mar 17 17:43:03.573595 containerd[1467]: time="2025-03-17T17:43:03.573560083Z" level=info msg="RemovePodSandbox for \"86089e35a842c8c8e7ea82ce5cbce4b7646d35903cb49ceaccd5d1db0dfbc90c\"" Mar 17 17:43:03.573595 containerd[1467]: time="2025-03-17T17:43:03.573580431Z" level=info msg="Forcibly stopping sandbox \"86089e35a842c8c8e7ea82ce5cbce4b7646d35903cb49ceaccd5d1db0dfbc90c\"" Mar 17 17:43:03.573771 containerd[1467]: time="2025-03-17T17:43:03.573653268Z" level=info msg="TearDown network for sandbox \"86089e35a842c8c8e7ea82ce5cbce4b7646d35903cb49ceaccd5d1db0dfbc90c\" successfully" Mar 17 17:43:03.635833 containerd[1467]: time="2025-03-17T17:43:03.635782777Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"86089e35a842c8c8e7ea82ce5cbce4b7646d35903cb49ceaccd5d1db0dfbc90c\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:43:03.635975 containerd[1467]: time="2025-03-17T17:43:03.635846958Z" level=info msg="RemovePodSandbox \"86089e35a842c8c8e7ea82ce5cbce4b7646d35903cb49ceaccd5d1db0dfbc90c\" returns successfully" Mar 17 17:43:03.636267 containerd[1467]: time="2025-03-17T17:43:03.636232471Z" level=info msg="StopPodSandbox for \"b33f35d6e3bd6b1a5f1d8cb21ebceacfc304ffe6c1b7ed3aa40b78efa59e6a03\"" Mar 17 17:43:03.636411 containerd[1467]: time="2025-03-17T17:43:03.636337188Z" level=info msg="TearDown network for sandbox \"b33f35d6e3bd6b1a5f1d8cb21ebceacfc304ffe6c1b7ed3aa40b78efa59e6a03\" successfully" Mar 17 17:43:03.636411 containerd[1467]: time="2025-03-17T17:43:03.636348379Z" level=info msg="StopPodSandbox for \"b33f35d6e3bd6b1a5f1d8cb21ebceacfc304ffe6c1b7ed3aa40b78efa59e6a03\" returns successfully" Mar 17 17:43:03.636578 containerd[1467]: time="2025-03-17T17:43:03.636551369Z" level=info msg="RemovePodSandbox for \"b33f35d6e3bd6b1a5f1d8cb21ebceacfc304ffe6c1b7ed3aa40b78efa59e6a03\"" Mar 17 17:43:03.636578 containerd[1467]: time="2025-03-17T17:43:03.636568491Z" level=info msg="Forcibly stopping sandbox \"b33f35d6e3bd6b1a5f1d8cb21ebceacfc304ffe6c1b7ed3aa40b78efa59e6a03\"" Mar 17 17:43:03.636719 containerd[1467]: time="2025-03-17T17:43:03.636639885Z" level=info msg="TearDown network for sandbox \"b33f35d6e3bd6b1a5f1d8cb21ebceacfc304ffe6c1b7ed3aa40b78efa59e6a03\" successfully" Mar 17 17:43:03.685220 containerd[1467]: time="2025-03-17T17:43:03.685165572Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"b33f35d6e3bd6b1a5f1d8cb21ebceacfc304ffe6c1b7ed3aa40b78efa59e6a03\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:43:03.685286 containerd[1467]: time="2025-03-17T17:43:03.685247746Z" level=info msg="RemovePodSandbox \"b33f35d6e3bd6b1a5f1d8cb21ebceacfc304ffe6c1b7ed3aa40b78efa59e6a03\" returns successfully" Mar 17 17:43:03.685598 containerd[1467]: time="2025-03-17T17:43:03.685572125Z" level=info msg="StopPodSandbox for \"47b88b63e50185559458c168aee26da05df22f8d62d10c49042400bc79ba64b6\"" Mar 17 17:43:03.685702 containerd[1467]: time="2025-03-17T17:43:03.685682973Z" level=info msg="TearDown network for sandbox \"47b88b63e50185559458c168aee26da05df22f8d62d10c49042400bc79ba64b6\" successfully" Mar 17 17:43:03.685702 containerd[1467]: time="2025-03-17T17:43:03.685698001Z" level=info msg="StopPodSandbox for \"47b88b63e50185559458c168aee26da05df22f8d62d10c49042400bc79ba64b6\" returns successfully" Mar 17 17:43:03.686011 containerd[1467]: time="2025-03-17T17:43:03.685978988Z" level=info msg="RemovePodSandbox for \"47b88b63e50185559458c168aee26da05df22f8d62d10c49042400bc79ba64b6\"" Mar 17 17:43:03.686067 containerd[1467]: time="2025-03-17T17:43:03.686010537Z" level=info msg="Forcibly stopping sandbox \"47b88b63e50185559458c168aee26da05df22f8d62d10c49042400bc79ba64b6\"" Mar 17 17:43:03.686114 containerd[1467]: time="2025-03-17T17:43:03.686085418Z" level=info msg="TearDown network for sandbox \"47b88b63e50185559458c168aee26da05df22f8d62d10c49042400bc79ba64b6\" successfully" Mar 17 17:43:03.813066 systemd[1]: Started sshd@17-10.0.0.61:22-10.0.0.1:59522.service - OpenSSH per-connection server daemon (10.0.0.1:59522). Mar 17 17:43:03.852383 containerd[1467]: time="2025-03-17T17:43:03.852303397Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"47b88b63e50185559458c168aee26da05df22f8d62d10c49042400bc79ba64b6\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:43:03.852383 containerd[1467]: time="2025-03-17T17:43:03.852369932Z" level=info msg="RemovePodSandbox \"47b88b63e50185559458c168aee26da05df22f8d62d10c49042400bc79ba64b6\" returns successfully" Mar 17 17:43:03.853452 containerd[1467]: time="2025-03-17T17:43:03.852914383Z" level=info msg="StopPodSandbox for \"1033b80b196d9abbe23dc46629630e49f397ddc00f3445a019c576bdd8d54f14\"" Mar 17 17:43:03.853452 containerd[1467]: time="2025-03-17T17:43:03.853005755Z" level=info msg="TearDown network for sandbox \"1033b80b196d9abbe23dc46629630e49f397ddc00f3445a019c576bdd8d54f14\" successfully" Mar 17 17:43:03.853452 containerd[1467]: time="2025-03-17T17:43:03.853015523Z" level=info msg="StopPodSandbox for \"1033b80b196d9abbe23dc46629630e49f397ddc00f3445a019c576bdd8d54f14\" returns successfully" Mar 17 17:43:03.853761 containerd[1467]: time="2025-03-17T17:43:03.853716980Z" level=info msg="RemovePodSandbox for \"1033b80b196d9abbe23dc46629630e49f397ddc00f3445a019c576bdd8d54f14\"" Mar 17 17:43:03.853761 containerd[1467]: time="2025-03-17T17:43:03.853758298Z" level=info msg="Forcibly stopping sandbox \"1033b80b196d9abbe23dc46629630e49f397ddc00f3445a019c576bdd8d54f14\"" Mar 17 17:43:03.854701 containerd[1467]: time="2025-03-17T17:43:03.854617330Z" level=info msg="TearDown network for sandbox \"1033b80b196d9abbe23dc46629630e49f397ddc00f3445a019c576bdd8d54f14\" successfully" Mar 17 17:43:03.894195 sshd[5932]: Accepted publickey for core from 10.0.0.1 port 59522 ssh2: RSA SHA256:j201F9FRK1q3ChnxQf0adNdYppDp+g37vmaXPvsVhek Mar 17 17:43:03.896026 sshd-session[5932]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 17 17:43:03.899993 systemd-logind[1453]: New session 18 of user core. Mar 17 17:43:03.904694 containerd[1467]: time="2025-03-17T17:43:03.904647742Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"1033b80b196d9abbe23dc46629630e49f397ddc00f3445a019c576bdd8d54f14\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:43:03.904758 containerd[1467]: time="2025-03-17T17:43:03.904704309Z" level=info msg="RemovePodSandbox \"1033b80b196d9abbe23dc46629630e49f397ddc00f3445a019c576bdd8d54f14\" returns successfully" Mar 17 17:43:03.905173 containerd[1467]: time="2025-03-17T17:43:03.905154677Z" level=info msg="StopPodSandbox for \"10ef4549d0b409f4064de73e7cb3238ecdb64a4888527c02814ca83e190159e7\"" Mar 17 17:43:03.905258 containerd[1467]: time="2025-03-17T17:43:03.905242562Z" level=info msg="TearDown network for sandbox \"10ef4549d0b409f4064de73e7cb3238ecdb64a4888527c02814ca83e190159e7\" successfully" Mar 17 17:43:03.905258 containerd[1467]: time="2025-03-17T17:43:03.905254855Z" level=info msg="StopPodSandbox for \"10ef4549d0b409f4064de73e7cb3238ecdb64a4888527c02814ca83e190159e7\" returns successfully" Mar 17 17:43:03.905637 containerd[1467]: time="2025-03-17T17:43:03.905585969Z" level=info msg="RemovePodSandbox for \"10ef4549d0b409f4064de73e7cb3238ecdb64a4888527c02814ca83e190159e7\"" Mar 17 17:43:03.905637 containerd[1467]: time="2025-03-17T17:43:03.905615374Z" level=info msg="Forcibly stopping sandbox \"10ef4549d0b409f4064de73e7cb3238ecdb64a4888527c02814ca83e190159e7\"" Mar 17 17:43:03.905786 containerd[1467]: time="2025-03-17T17:43:03.905680967Z" level=info msg="TearDown network for sandbox \"10ef4549d0b409f4064de73e7cb3238ecdb64a4888527c02814ca83e190159e7\" successfully" Mar 17 17:43:03.915485 systemd[1]: Started session-18.scope - Session 18 of User core. Mar 17 17:43:04.033531 sshd[5934]: Connection closed by 10.0.0.1 port 59522 Mar 17 17:43:04.033916 sshd-session[5932]: pam_unix(sshd:session): session closed for user core Mar 17 17:43:04.046569 systemd[1]: sshd@17-10.0.0.61:22-10.0.0.1:59522.service: Deactivated successfully. Mar 17 17:43:04.048836 systemd[1]: session-18.scope: Deactivated successfully. Mar 17 17:43:04.050967 systemd-logind[1453]: Session 18 logged out. Waiting for processes to exit. Mar 17 17:43:04.055625 systemd[1]: Started sshd@18-10.0.0.61:22-10.0.0.1:59534.service - OpenSSH per-connection server daemon (10.0.0.1:59534). Mar 17 17:43:04.056770 systemd-logind[1453]: Removed session 18. Mar 17 17:43:04.090962 sshd[5946]: Accepted publickey for core from 10.0.0.1 port 59534 ssh2: RSA SHA256:j201F9FRK1q3ChnxQf0adNdYppDp+g37vmaXPvsVhek Mar 17 17:43:04.092518 sshd-session[5946]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 17 17:43:04.096550 systemd-logind[1453]: New session 19 of user core. Mar 17 17:43:04.106470 systemd[1]: Started session-19.scope - Session 19 of User core. Mar 17 17:43:04.497003 containerd[1467]: time="2025-03-17T17:43:04.496929798Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"10ef4549d0b409f4064de73e7cb3238ecdb64a4888527c02814ca83e190159e7\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:43:04.497194 containerd[1467]: time="2025-03-17T17:43:04.497015103Z" level=info msg="RemovePodSandbox \"10ef4549d0b409f4064de73e7cb3238ecdb64a4888527c02814ca83e190159e7\" returns successfully" Mar 17 17:43:04.497663 containerd[1467]: time="2025-03-17T17:43:04.497563912Z" level=info msg="StopPodSandbox for \"32d2a6ea6079479e27c144f7baad1a289529d3a29ff0a7f2e40468d0ecdc1347\"" Mar 17 17:43:04.497754 containerd[1467]: time="2025-03-17T17:43:04.497689904Z" level=info msg="TearDown network for sandbox \"32d2a6ea6079479e27c144f7baad1a289529d3a29ff0a7f2e40468d0ecdc1347\" successfully" Mar 17 17:43:04.497754 containerd[1467]: time="2025-03-17T17:43:04.497702769Z" level=info msg="StopPodSandbox for \"32d2a6ea6079479e27c144f7baad1a289529d3a29ff0a7f2e40468d0ecdc1347\" returns successfully" Mar 17 17:43:04.498098 containerd[1467]: time="2025-03-17T17:43:04.498047674Z" level=info msg="RemovePodSandbox for \"32d2a6ea6079479e27c144f7baad1a289529d3a29ff0a7f2e40468d0ecdc1347\"" Mar 17 17:43:04.498098 containerd[1467]: time="2025-03-17T17:43:04.498100286Z" level=info msg="Forcibly stopping sandbox \"32d2a6ea6079479e27c144f7baad1a289529d3a29ff0a7f2e40468d0ecdc1347\"" Mar 17 17:43:04.498275 containerd[1467]: time="2025-03-17T17:43:04.498222261Z" level=info msg="TearDown network for sandbox \"32d2a6ea6079479e27c144f7baad1a289529d3a29ff0a7f2e40468d0ecdc1347\" successfully" Mar 17 17:43:04.542606 containerd[1467]: time="2025-03-17T17:43:04.542530199Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"32d2a6ea6079479e27c144f7baad1a289529d3a29ff0a7f2e40468d0ecdc1347\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Mar 17 17:43:04.542606 containerd[1467]: time="2025-03-17T17:43:04.542612477Z" level=info msg="RemovePodSandbox \"32d2a6ea6079479e27c144f7baad1a289529d3a29ff0a7f2e40468d0ecdc1347\" returns successfully" Mar 17 17:43:04.640969 sshd[5948]: Connection closed by 10.0.0.1 port 59534 Mar 17 17:43:04.641454 sshd-session[5946]: pam_unix(sshd:session): session closed for user core Mar 17 17:43:04.653504 systemd[1]: sshd@18-10.0.0.61:22-10.0.0.1:59534.service: Deactivated successfully. Mar 17 17:43:04.655661 systemd[1]: session-19.scope: Deactivated successfully. Mar 17 17:43:04.657555 systemd-logind[1453]: Session 19 logged out. Waiting for processes to exit. Mar 17 17:43:04.664632 systemd[1]: Started sshd@19-10.0.0.61:22-10.0.0.1:59542.service - OpenSSH per-connection server daemon (10.0.0.1:59542). Mar 17 17:43:04.666055 systemd-logind[1453]: Removed session 19. Mar 17 17:43:04.700598 sshd[5958]: Accepted publickey for core from 10.0.0.1 port 59542 ssh2: RSA SHA256:j201F9FRK1q3ChnxQf0adNdYppDp+g37vmaXPvsVhek Mar 17 17:43:04.702425 sshd-session[5958]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 17 17:43:04.706826 systemd-logind[1453]: New session 20 of user core. Mar 17 17:43:04.713527 systemd[1]: Started session-20.scope - Session 20 of User core. Mar 17 17:43:07.136338 sshd[5960]: Connection closed by 10.0.0.1 port 59542 Mar 17 17:43:07.137055 sshd-session[5958]: pam_unix(sshd:session): session closed for user core Mar 17 17:43:07.147563 systemd[1]: sshd@19-10.0.0.61:22-10.0.0.1:59542.service: Deactivated successfully. Mar 17 17:43:07.149467 systemd[1]: session-20.scope: Deactivated successfully. Mar 17 17:43:07.151157 systemd-logind[1453]: Session 20 logged out. Waiting for processes to exit. Mar 17 17:43:07.158640 systemd[1]: Started sshd@20-10.0.0.61:22-10.0.0.1:59546.service - OpenSSH per-connection server daemon (10.0.0.1:59546). Mar 17 17:43:07.159834 systemd-logind[1453]: Removed session 20. Mar 17 17:43:07.196642 sshd[5977]: Accepted publickey for core from 10.0.0.1 port 59546 ssh2: RSA SHA256:j201F9FRK1q3ChnxQf0adNdYppDp+g37vmaXPvsVhek Mar 17 17:43:07.198282 sshd-session[5977]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 17 17:43:07.204676 systemd-logind[1453]: New session 21 of user core. Mar 17 17:43:07.214506 systemd[1]: Started session-21.scope - Session 21 of User core. Mar 17 17:43:07.469130 sshd[5979]: Connection closed by 10.0.0.1 port 59546 Mar 17 17:43:07.469666 sshd-session[5977]: pam_unix(sshd:session): session closed for user core Mar 17 17:43:07.477931 systemd[1]: sshd@20-10.0.0.61:22-10.0.0.1:59546.service: Deactivated successfully. Mar 17 17:43:07.480412 systemd[1]: session-21.scope: Deactivated successfully. Mar 17 17:43:07.482632 systemd-logind[1453]: Session 21 logged out. Waiting for processes to exit. Mar 17 17:43:07.487794 systemd[1]: Started sshd@21-10.0.0.61:22-10.0.0.1:59556.service - OpenSSH per-connection server daemon (10.0.0.1:59556). Mar 17 17:43:07.489439 systemd-logind[1453]: Removed session 21. Mar 17 17:43:07.521243 sshd[5990]: Accepted publickey for core from 10.0.0.1 port 59556 ssh2: RSA SHA256:j201F9FRK1q3ChnxQf0adNdYppDp+g37vmaXPvsVhek Mar 17 17:43:07.522853 sshd-session[5990]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 17 17:43:07.527204 systemd-logind[1453]: New session 22 of user core. Mar 17 17:43:07.534514 systemd[1]: Started session-22.scope - Session 22 of User core. Mar 17 17:43:07.644234 sshd[5992]: Connection closed by 10.0.0.1 port 59556 Mar 17 17:43:07.644584 sshd-session[5990]: pam_unix(sshd:session): session closed for user core Mar 17 17:43:07.648221 systemd[1]: sshd@21-10.0.0.61:22-10.0.0.1:59556.service: Deactivated successfully. Mar 17 17:43:07.650593 systemd[1]: session-22.scope: Deactivated successfully. Mar 17 17:43:07.651296 systemd-logind[1453]: Session 22 logged out. Waiting for processes to exit. Mar 17 17:43:07.652222 systemd-logind[1453]: Removed session 22. Mar 17 17:43:10.474940 kubelet[2662]: I0317 17:43:10.474886 2662 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 17 17:43:12.657652 systemd[1]: Started sshd@22-10.0.0.61:22-10.0.0.1:38570.service - OpenSSH per-connection server daemon (10.0.0.1:38570). Mar 17 17:43:12.693303 sshd[6007]: Accepted publickey for core from 10.0.0.1 port 38570 ssh2: RSA SHA256:j201F9FRK1q3ChnxQf0adNdYppDp+g37vmaXPvsVhek Mar 17 17:43:12.694785 sshd-session[6007]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 17 17:43:12.698976 systemd-logind[1453]: New session 23 of user core. Mar 17 17:43:12.707537 systemd[1]: Started session-23.scope - Session 23 of User core. Mar 17 17:43:12.819542 sshd[6009]: Connection closed by 10.0.0.1 port 38570 Mar 17 17:43:12.819983 sshd-session[6007]: pam_unix(sshd:session): session closed for user core Mar 17 17:43:12.824947 systemd[1]: sshd@22-10.0.0.61:22-10.0.0.1:38570.service: Deactivated successfully. Mar 17 17:43:12.827771 systemd[1]: session-23.scope: Deactivated successfully. Mar 17 17:43:12.828516 systemd-logind[1453]: Session 23 logged out. Waiting for processes to exit. Mar 17 17:43:12.829444 systemd-logind[1453]: Removed session 23. Mar 17 17:43:17.835079 systemd[1]: Started sshd@23-10.0.0.61:22-10.0.0.1:38580.service - OpenSSH per-connection server daemon (10.0.0.1:38580). Mar 17 17:43:17.873914 sshd[6026]: Accepted publickey for core from 10.0.0.1 port 38580 ssh2: RSA SHA256:j201F9FRK1q3ChnxQf0adNdYppDp+g37vmaXPvsVhek Mar 17 17:43:17.875800 sshd-session[6026]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 17 17:43:17.879680 systemd-logind[1453]: New session 24 of user core. Mar 17 17:43:17.886499 systemd[1]: Started session-24.scope - Session 24 of User core. Mar 17 17:43:18.003340 sshd[6028]: Connection closed by 10.0.0.1 port 38580 Mar 17 17:43:18.003767 sshd-session[6026]: pam_unix(sshd:session): session closed for user core Mar 17 17:43:18.008104 systemd[1]: sshd@23-10.0.0.61:22-10.0.0.1:38580.service: Deactivated successfully. Mar 17 17:43:18.010523 systemd[1]: session-24.scope: Deactivated successfully. Mar 17 17:43:18.011265 systemd-logind[1453]: Session 24 logged out. Waiting for processes to exit. Mar 17 17:43:18.012219 systemd-logind[1453]: Removed session 24. Mar 17 17:43:19.822112 kubelet[2662]: E0317 17:43:19.822014 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:43:23.019901 systemd[1]: Started sshd@24-10.0.0.61:22-10.0.0.1:58066.service - OpenSSH per-connection server daemon (10.0.0.1:58066). Mar 17 17:43:23.071772 sshd[6069]: Accepted publickey for core from 10.0.0.1 port 58066 ssh2: RSA SHA256:j201F9FRK1q3ChnxQf0adNdYppDp+g37vmaXPvsVhek Mar 17 17:43:23.073542 sshd-session[6069]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 17 17:43:23.077911 systemd-logind[1453]: New session 25 of user core. Mar 17 17:43:23.091585 systemd[1]: Started session-25.scope - Session 25 of User core. Mar 17 17:43:23.222125 sshd[6071]: Connection closed by 10.0.0.1 port 58066 Mar 17 17:43:23.222710 sshd-session[6069]: pam_unix(sshd:session): session closed for user core Mar 17 17:43:23.227832 systemd[1]: sshd@24-10.0.0.61:22-10.0.0.1:58066.service: Deactivated successfully. Mar 17 17:43:23.230550 systemd[1]: session-25.scope: Deactivated successfully. Mar 17 17:43:23.231413 systemd-logind[1453]: Session 25 logged out. Waiting for processes to exit. Mar 17 17:43:23.232558 systemd-logind[1453]: Removed session 25. Mar 17 17:43:24.821094 kubelet[2662]: E0317 17:43:24.821038 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 17:43:28.233508 systemd[1]: Started sshd@25-10.0.0.61:22-10.0.0.1:58076.service - OpenSSH per-connection server daemon (10.0.0.1:58076). Mar 17 17:43:28.279372 sshd[6083]: Accepted publickey for core from 10.0.0.1 port 58076 ssh2: RSA SHA256:j201F9FRK1q3ChnxQf0adNdYppDp+g37vmaXPvsVhek Mar 17 17:43:28.281078 sshd-session[6083]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 17 17:43:28.285028 systemd-logind[1453]: New session 26 of user core. Mar 17 17:43:28.294478 systemd[1]: Started session-26.scope - Session 26 of User core. Mar 17 17:43:28.410916 sshd[6085]: Connection closed by 10.0.0.1 port 58076 Mar 17 17:43:28.411270 sshd-session[6083]: pam_unix(sshd:session): session closed for user core Mar 17 17:43:28.414995 systemd[1]: sshd@25-10.0.0.61:22-10.0.0.1:58076.service: Deactivated successfully. Mar 17 17:43:28.417220 systemd[1]: session-26.scope: Deactivated successfully. Mar 17 17:43:28.417855 systemd-logind[1453]: Session 26 logged out. Waiting for processes to exit. Mar 17 17:43:28.418663 systemd-logind[1453]: Removed session 26. Mar 17 17:43:29.821727 kubelet[2662]: E0317 17:43:29.821682 2662 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"