Sep 4 17:18:31.957923 kernel: Booting Linux on physical CPU 0x0000000000 [0x413fd0c1] Sep 4 17:18:31.957947 kernel: Linux version 6.6.48-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 13.2.1_p20240210 p14) 13.2.1 20240210, GNU ld (Gentoo 2.41 p5) 2.41.0) #1 SMP PREEMPT Wed Sep 4 15:52:28 -00 2024 Sep 4 17:18:31.957957 kernel: KASLR enabled Sep 4 17:18:31.957963 kernel: efi: EFI v2.7 by EDK II Sep 4 17:18:31.957969 kernel: efi: SMBIOS 3.0=0xdced0000 MEMATTR=0xdb8fd018 ACPI 2.0=0xd9710018 RNG=0xd971e498 MEMRESERVE=0xd9b43d18 Sep 4 17:18:31.957975 kernel: random: crng init done Sep 4 17:18:31.957982 kernel: ACPI: Early table checksum verification disabled Sep 4 17:18:31.957988 kernel: ACPI: RSDP 0x00000000D9710018 000024 (v02 BOCHS ) Sep 4 17:18:31.957995 kernel: ACPI: XSDT 0x00000000D971FE98 000064 (v01 BOCHS BXPC 00000001 01000013) Sep 4 17:18:31.958002 kernel: ACPI: FACP 0x00000000D971FA98 000114 (v06 BOCHS BXPC 00000001 BXPC 00000001) Sep 4 17:18:31.958008 kernel: ACPI: DSDT 0x00000000D9717518 0014A2 (v02 BOCHS BXPC 00000001 BXPC 00000001) Sep 4 17:18:31.958014 kernel: ACPI: APIC 0x00000000D971FC18 0001A8 (v04 BOCHS BXPC 00000001 BXPC 00000001) Sep 4 17:18:31.958020 kernel: ACPI: PPTT 0x00000000D971D898 00009C (v02 BOCHS BXPC 00000001 BXPC 00000001) Sep 4 17:18:31.958026 kernel: ACPI: GTDT 0x00000000D971E818 000060 (v02 BOCHS BXPC 00000001 BXPC 00000001) Sep 4 17:18:31.958033 kernel: ACPI: MCFG 0x00000000D971E918 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Sep 4 17:18:31.958041 kernel: ACPI: SPCR 0x00000000D971FF98 000050 (v02 BOCHS BXPC 00000001 BXPC 00000001) Sep 4 17:18:31.958048 kernel: ACPI: DBG2 0x00000000D971E418 000057 (v00 BOCHS BXPC 00000001 BXPC 00000001) Sep 4 17:18:31.958054 kernel: ACPI: IORT 0x00000000D971E718 000080 (v03 BOCHS BXPC 00000001 BXPC 00000001) Sep 4 17:18:31.958060 kernel: ACPI: SPCR: console: pl011,mmio,0x9000000,9600 Sep 4 17:18:31.958066 kernel: NUMA: Failed to initialise from firmware Sep 4 17:18:31.958073 kernel: NUMA: Faking a node at [mem 0x0000000040000000-0x00000000dcffffff] Sep 4 17:18:31.958079 kernel: NUMA: NODE_DATA [mem 0xdc958800-0xdc95dfff] Sep 4 17:18:31.958086 kernel: Zone ranges: Sep 4 17:18:31.958092 kernel: DMA [mem 0x0000000040000000-0x00000000dcffffff] Sep 4 17:18:31.958098 kernel: DMA32 empty Sep 4 17:18:31.958106 kernel: Normal empty Sep 4 17:18:31.958113 kernel: Movable zone start for each node Sep 4 17:18:31.958119 kernel: Early memory node ranges Sep 4 17:18:31.958125 kernel: node 0: [mem 0x0000000040000000-0x00000000d976ffff] Sep 4 17:18:31.958132 kernel: node 0: [mem 0x00000000d9770000-0x00000000d9b3ffff] Sep 4 17:18:31.958138 kernel: node 0: [mem 0x00000000d9b40000-0x00000000dce1ffff] Sep 4 17:18:31.958144 kernel: node 0: [mem 0x00000000dce20000-0x00000000dceaffff] Sep 4 17:18:31.958151 kernel: node 0: [mem 0x00000000dceb0000-0x00000000dcebffff] Sep 4 17:18:31.958157 kernel: node 0: [mem 0x00000000dcec0000-0x00000000dcfdffff] Sep 4 17:18:31.958163 kernel: node 0: [mem 0x00000000dcfe0000-0x00000000dcffffff] Sep 4 17:18:31.958170 kernel: Initmem setup node 0 [mem 0x0000000040000000-0x00000000dcffffff] Sep 4 17:18:31.958176 kernel: On node 0, zone DMA: 12288 pages in unavailable ranges Sep 4 17:18:31.958184 kernel: psci: probing for conduit method from ACPI. Sep 4 17:18:31.958191 kernel: psci: PSCIv1.1 detected in firmware. Sep 4 17:18:31.958197 kernel: psci: Using standard PSCI v0.2 function IDs Sep 4 17:18:31.958206 kernel: psci: Trusted OS migration not required Sep 4 17:18:31.958213 kernel: psci: SMC Calling Convention v1.1 Sep 4 17:18:31.958220 kernel: smccc: KVM: hypervisor services detected (0x00000000 0x00000000 0x00000000 0x00000003) Sep 4 17:18:31.958228 kernel: percpu: Embedded 31 pages/cpu s86632 r8192 d32152 u126976 Sep 4 17:18:31.958235 kernel: pcpu-alloc: s86632 r8192 d32152 u126976 alloc=31*4096 Sep 4 17:18:31.958242 kernel: pcpu-alloc: [0] 0 [0] 1 [0] 2 [0] 3 Sep 4 17:18:31.958249 kernel: Detected PIPT I-cache on CPU0 Sep 4 17:18:31.958255 kernel: CPU features: detected: GIC system register CPU interface Sep 4 17:18:31.958262 kernel: CPU features: detected: Hardware dirty bit management Sep 4 17:18:31.958269 kernel: CPU features: detected: Spectre-v4 Sep 4 17:18:31.958275 kernel: CPU features: detected: Spectre-BHB Sep 4 17:18:31.958282 kernel: CPU features: kernel page table isolation forced ON by KASLR Sep 4 17:18:31.958289 kernel: CPU features: detected: Kernel page table isolation (KPTI) Sep 4 17:18:31.958306 kernel: CPU features: detected: ARM erratum 1418040 Sep 4 17:18:31.958313 kernel: CPU features: detected: SSBS not fully self-synchronizing Sep 4 17:18:31.958320 kernel: alternatives: applying boot alternatives Sep 4 17:18:31.958327 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected acpi=force verity.usrhash=7913866621ae0af53522ae1b4ff4e1e453dd69d966d437a439147039341ecbbc Sep 4 17:18:31.958335 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Sep 4 17:18:31.958341 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Sep 4 17:18:31.958348 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Sep 4 17:18:31.958355 kernel: Fallback order for Node 0: 0 Sep 4 17:18:31.958362 kernel: Built 1 zonelists, mobility grouping on. Total pages: 633024 Sep 4 17:18:31.958368 kernel: Policy zone: DMA Sep 4 17:18:31.958375 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Sep 4 17:18:31.958384 kernel: software IO TLB: area num 4. Sep 4 17:18:31.958392 kernel: software IO TLB: mapped [mem 0x00000000d2e00000-0x00000000d6e00000] (64MB) Sep 4 17:18:31.958399 kernel: Memory: 2386852K/2572288K available (10240K kernel code, 2182K rwdata, 8076K rodata, 39040K init, 897K bss, 185436K reserved, 0K cma-reserved) Sep 4 17:18:31.958406 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Sep 4 17:18:31.958412 kernel: trace event string verifier disabled Sep 4 17:18:31.958419 kernel: rcu: Preemptible hierarchical RCU implementation. Sep 4 17:18:31.958426 kernel: rcu: RCU event tracing is enabled. Sep 4 17:18:31.958433 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Sep 4 17:18:31.958440 kernel: Trampoline variant of Tasks RCU enabled. Sep 4 17:18:31.958447 kernel: Tracing variant of Tasks RCU enabled. Sep 4 17:18:31.958454 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Sep 4 17:18:31.958461 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Sep 4 17:18:31.958469 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Sep 4 17:18:31.958476 kernel: GICv3: 256 SPIs implemented Sep 4 17:18:31.958482 kernel: GICv3: 0 Extended SPIs implemented Sep 4 17:18:31.958489 kernel: Root IRQ handler: gic_handle_irq Sep 4 17:18:31.958495 kernel: GICv3: GICv3 features: 16 PPIs, DirectLPI Sep 4 17:18:31.958502 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000080a0000 Sep 4 17:18:31.958509 kernel: ITS [mem 0x08080000-0x0809ffff] Sep 4 17:18:31.958516 kernel: ITS@0x0000000008080000: allocated 8192 Devices @400d0000 (indirect, esz 8, psz 64K, shr 1) Sep 4 17:18:31.958523 kernel: ITS@0x0000000008080000: allocated 8192 Interrupt Collections @400e0000 (flat, esz 8, psz 64K, shr 1) Sep 4 17:18:31.958530 kernel: GICv3: using LPI property table @0x00000000400f0000 Sep 4 17:18:31.958537 kernel: GICv3: CPU0: using allocated LPI pending table @0x0000000040100000 Sep 4 17:18:31.958545 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Sep 4 17:18:31.958552 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Sep 4 17:18:31.958559 kernel: arch_timer: cp15 timer(s) running at 25.00MHz (virt). Sep 4 17:18:31.958566 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns Sep 4 17:18:31.958573 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns Sep 4 17:18:31.958579 kernel: arm-pv: using stolen time PV Sep 4 17:18:31.958586 kernel: Console: colour dummy device 80x25 Sep 4 17:18:31.958593 kernel: ACPI: Core revision 20230628 Sep 4 17:18:31.958600 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000) Sep 4 17:18:31.958607 kernel: pid_max: default: 32768 minimum: 301 Sep 4 17:18:31.958615 kernel: LSM: initializing lsm=lockdown,capability,selinux,integrity Sep 4 17:18:31.958622 kernel: SELinux: Initializing. Sep 4 17:18:31.958630 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Sep 4 17:18:31.958637 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Sep 4 17:18:31.958644 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1. Sep 4 17:18:31.958651 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1. Sep 4 17:18:31.958658 kernel: rcu: Hierarchical SRCU implementation. Sep 4 17:18:31.958666 kernel: rcu: Max phase no-delay instances is 400. Sep 4 17:18:31.958673 kernel: Platform MSI: ITS@0x8080000 domain created Sep 4 17:18:31.958681 kernel: PCI/MSI: ITS@0x8080000 domain created Sep 4 17:18:31.958688 kernel: Remapping and enabling EFI services. Sep 4 17:18:31.958701 kernel: smp: Bringing up secondary CPUs ... Sep 4 17:18:31.958708 kernel: Detected PIPT I-cache on CPU1 Sep 4 17:18:31.958715 kernel: GICv3: CPU1: found redistributor 1 region 0:0x00000000080c0000 Sep 4 17:18:31.958722 kernel: GICv3: CPU1: using allocated LPI pending table @0x0000000040110000 Sep 4 17:18:31.958729 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Sep 4 17:18:31.958736 kernel: CPU1: Booted secondary processor 0x0000000001 [0x413fd0c1] Sep 4 17:18:31.958743 kernel: Detected PIPT I-cache on CPU2 Sep 4 17:18:31.958750 kernel: GICv3: CPU2: found redistributor 2 region 0:0x00000000080e0000 Sep 4 17:18:31.958760 kernel: GICv3: CPU2: using allocated LPI pending table @0x0000000040120000 Sep 4 17:18:31.958767 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Sep 4 17:18:31.958779 kernel: CPU2: Booted secondary processor 0x0000000002 [0x413fd0c1] Sep 4 17:18:31.958788 kernel: Detected PIPT I-cache on CPU3 Sep 4 17:18:31.958795 kernel: GICv3: CPU3: found redistributor 3 region 0:0x0000000008100000 Sep 4 17:18:31.958803 kernel: GICv3: CPU3: using allocated LPI pending table @0x0000000040130000 Sep 4 17:18:31.958810 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Sep 4 17:18:31.958818 kernel: CPU3: Booted secondary processor 0x0000000003 [0x413fd0c1] Sep 4 17:18:31.958825 kernel: smp: Brought up 1 node, 4 CPUs Sep 4 17:18:31.958842 kernel: SMP: Total of 4 processors activated. Sep 4 17:18:31.958864 kernel: CPU features: detected: 32-bit EL0 Support Sep 4 17:18:31.958879 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence Sep 4 17:18:31.958887 kernel: CPU features: detected: Common not Private translations Sep 4 17:18:31.958895 kernel: CPU features: detected: CRC32 instructions Sep 4 17:18:31.958903 kernel: CPU features: detected: Enhanced Virtualization Traps Sep 4 17:18:31.958910 kernel: CPU features: detected: RCpc load-acquire (LDAPR) Sep 4 17:18:31.958918 kernel: CPU features: detected: LSE atomic instructions Sep 4 17:18:31.958927 kernel: CPU features: detected: Privileged Access Never Sep 4 17:18:31.958935 kernel: CPU features: detected: RAS Extension Support Sep 4 17:18:31.958943 kernel: CPU features: detected: Speculative Store Bypassing Safe (SSBS) Sep 4 17:18:31.958951 kernel: CPU: All CPU(s) started at EL1 Sep 4 17:18:31.958958 kernel: alternatives: applying system-wide alternatives Sep 4 17:18:31.958965 kernel: devtmpfs: initialized Sep 4 17:18:31.958973 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Sep 4 17:18:31.958980 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Sep 4 17:18:31.958988 kernel: pinctrl core: initialized pinctrl subsystem Sep 4 17:18:31.958997 kernel: SMBIOS 3.0.0 present. Sep 4 17:18:31.959005 kernel: DMI: QEMU KVM Virtual Machine, BIOS edk2-20230524-3.fc38 05/24/2023 Sep 4 17:18:31.959012 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Sep 4 17:18:31.959019 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Sep 4 17:18:31.959027 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Sep 4 17:18:31.959035 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Sep 4 17:18:31.959042 kernel: audit: initializing netlink subsys (disabled) Sep 4 17:18:31.959050 kernel: audit: type=2000 audit(0.024:1): state=initialized audit_enabled=0 res=1 Sep 4 17:18:31.959057 kernel: thermal_sys: Registered thermal governor 'step_wise' Sep 4 17:18:31.959066 kernel: cpuidle: using governor menu Sep 4 17:18:31.959074 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Sep 4 17:18:31.959081 kernel: ASID allocator initialised with 32768 entries Sep 4 17:18:31.959088 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Sep 4 17:18:31.959096 kernel: Serial: AMBA PL011 UART driver Sep 4 17:18:31.959104 kernel: Modules: 2G module region forced by RANDOMIZE_MODULE_REGION_FULL Sep 4 17:18:31.959111 kernel: Modules: 0 pages in range for non-PLT usage Sep 4 17:18:31.959118 kernel: Modules: 509120 pages in range for PLT usage Sep 4 17:18:31.959126 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Sep 4 17:18:31.959134 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page Sep 4 17:18:31.959142 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages Sep 4 17:18:31.959149 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page Sep 4 17:18:31.959157 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Sep 4 17:18:31.959164 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page Sep 4 17:18:31.959171 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages Sep 4 17:18:31.959178 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page Sep 4 17:18:31.959186 kernel: ACPI: Added _OSI(Module Device) Sep 4 17:18:31.959193 kernel: ACPI: Added _OSI(Processor Device) Sep 4 17:18:31.959202 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Sep 4 17:18:31.959209 kernel: ACPI: Added _OSI(Processor Aggregator Device) Sep 4 17:18:31.959216 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Sep 4 17:18:31.959223 kernel: ACPI: Interpreter enabled Sep 4 17:18:31.959230 kernel: ACPI: Using GIC for interrupt routing Sep 4 17:18:31.959238 kernel: ACPI: MCFG table detected, 1 entries Sep 4 17:18:31.959245 kernel: ARMH0011:00: ttyAMA0 at MMIO 0x9000000 (irq = 12, base_baud = 0) is a SBSA Sep 4 17:18:31.959252 kernel: printk: console [ttyAMA0] enabled Sep 4 17:18:31.959259 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Sep 4 17:18:31.959408 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Sep 4 17:18:31.959484 kernel: acpi PNP0A08:00: _OSC: platform does not support [LTR] Sep 4 17:18:31.959592 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME AER PCIeCapability] Sep 4 17:18:31.959672 kernel: acpi PNP0A08:00: ECAM area [mem 0x4010000000-0x401fffffff] reserved by PNP0C02:00 Sep 4 17:18:31.959738 kernel: acpi PNP0A08:00: ECAM at [mem 0x4010000000-0x401fffffff] for [bus 00-ff] Sep 4 17:18:31.959747 kernel: ACPI: Remapped I/O 0x000000003eff0000 to [io 0x0000-0xffff window] Sep 4 17:18:31.959756 kernel: PCI host bridge to bus 0000:00 Sep 4 17:18:31.959843 kernel: pci_bus 0000:00: root bus resource [mem 0x10000000-0x3efeffff window] Sep 4 17:18:31.959909 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0xffff window] Sep 4 17:18:31.959973 kernel: pci_bus 0000:00: root bus resource [mem 0x8000000000-0xffffffffff window] Sep 4 17:18:31.960033 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Sep 4 17:18:31.960125 kernel: pci 0000:00:00.0: [1b36:0008] type 00 class 0x060000 Sep 4 17:18:31.960214 kernel: pci 0000:00:01.0: [1af4:1005] type 00 class 0x00ff00 Sep 4 17:18:31.960283 kernel: pci 0000:00:01.0: reg 0x10: [io 0x0000-0x001f] Sep 4 17:18:31.960362 kernel: pci 0000:00:01.0: reg 0x14: [mem 0x10000000-0x10000fff] Sep 4 17:18:31.960429 kernel: pci 0000:00:01.0: reg 0x20: [mem 0x8000000000-0x8000003fff 64bit pref] Sep 4 17:18:31.960493 kernel: pci 0000:00:01.0: BAR 4: assigned [mem 0x8000000000-0x8000003fff 64bit pref] Sep 4 17:18:31.960558 kernel: pci 0000:00:01.0: BAR 1: assigned [mem 0x10000000-0x10000fff] Sep 4 17:18:31.960623 kernel: pci 0000:00:01.0: BAR 0: assigned [io 0x1000-0x101f] Sep 4 17:18:31.960686 kernel: pci_bus 0000:00: resource 4 [mem 0x10000000-0x3efeffff window] Sep 4 17:18:31.960746 kernel: pci_bus 0000:00: resource 5 [io 0x0000-0xffff window] Sep 4 17:18:31.960809 kernel: pci_bus 0000:00: resource 6 [mem 0x8000000000-0xffffffffff window] Sep 4 17:18:31.960819 kernel: ACPI: PCI: Interrupt link GSI0 configured for IRQ 35 Sep 4 17:18:31.960835 kernel: ACPI: PCI: Interrupt link GSI1 configured for IRQ 36 Sep 4 17:18:31.960844 kernel: ACPI: PCI: Interrupt link GSI2 configured for IRQ 37 Sep 4 17:18:31.960851 kernel: ACPI: PCI: Interrupt link GSI3 configured for IRQ 38 Sep 4 17:18:31.960858 kernel: iommu: Default domain type: Translated Sep 4 17:18:31.960866 kernel: iommu: DMA domain TLB invalidation policy: strict mode Sep 4 17:18:31.960873 kernel: efivars: Registered efivars operations Sep 4 17:18:31.960883 kernel: vgaarb: loaded Sep 4 17:18:31.960890 kernel: clocksource: Switched to clocksource arch_sys_counter Sep 4 17:18:31.960898 kernel: VFS: Disk quotas dquot_6.6.0 Sep 4 17:18:31.960906 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Sep 4 17:18:31.960913 kernel: pnp: PnP ACPI init Sep 4 17:18:31.960996 kernel: system 00:00: [mem 0x4010000000-0x401fffffff window] could not be reserved Sep 4 17:18:31.961007 kernel: pnp: PnP ACPI: found 1 devices Sep 4 17:18:31.961014 kernel: NET: Registered PF_INET protocol family Sep 4 17:18:31.961024 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Sep 4 17:18:31.961032 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Sep 4 17:18:31.961039 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Sep 4 17:18:31.961047 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Sep 4 17:18:31.961054 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Sep 4 17:18:31.961061 kernel: TCP: Hash tables configured (established 32768 bind 32768) Sep 4 17:18:31.961069 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Sep 4 17:18:31.961076 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Sep 4 17:18:31.961083 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Sep 4 17:18:31.961092 kernel: PCI: CLS 0 bytes, default 64 Sep 4 17:18:31.961099 kernel: kvm [1]: HYP mode not available Sep 4 17:18:31.961107 kernel: Initialise system trusted keyrings Sep 4 17:18:31.961114 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Sep 4 17:18:31.961121 kernel: Key type asymmetric registered Sep 4 17:18:31.961129 kernel: Asymmetric key parser 'x509' registered Sep 4 17:18:31.961136 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Sep 4 17:18:31.961143 kernel: io scheduler mq-deadline registered Sep 4 17:18:31.961151 kernel: io scheduler kyber registered Sep 4 17:18:31.961160 kernel: io scheduler bfq registered Sep 4 17:18:31.961168 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0C:00/input/input0 Sep 4 17:18:31.961175 kernel: ACPI: button: Power Button [PWRB] Sep 4 17:18:31.961183 kernel: ACPI: \_SB_.PCI0.GSI1: Enabled at IRQ 36 Sep 4 17:18:31.961252 kernel: virtio-pci 0000:00:01.0: enabling device (0005 -> 0007) Sep 4 17:18:31.961262 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Sep 4 17:18:31.961269 kernel: thunder_xcv, ver 1.0 Sep 4 17:18:31.961277 kernel: thunder_bgx, ver 1.0 Sep 4 17:18:31.961284 kernel: nicpf, ver 1.0 Sep 4 17:18:31.961299 kernel: nicvf, ver 1.0 Sep 4 17:18:31.961375 kernel: rtc-efi rtc-efi.0: registered as rtc0 Sep 4 17:18:31.961437 kernel: rtc-efi rtc-efi.0: setting system clock to 2024-09-04T17:18:31 UTC (1725470311) Sep 4 17:18:31.961447 kernel: hid: raw HID events driver (C) Jiri Kosina Sep 4 17:18:31.961455 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 7 counters available Sep 4 17:18:31.961462 kernel: watchdog: Delayed init of the lockup detector failed: -19 Sep 4 17:18:31.961469 kernel: watchdog: Hard watchdog permanently disabled Sep 4 17:18:31.961477 kernel: NET: Registered PF_INET6 protocol family Sep 4 17:18:31.961486 kernel: Segment Routing with IPv6 Sep 4 17:18:31.961494 kernel: In-situ OAM (IOAM) with IPv6 Sep 4 17:18:31.961501 kernel: NET: Registered PF_PACKET protocol family Sep 4 17:18:31.961508 kernel: Key type dns_resolver registered Sep 4 17:18:31.961515 kernel: registered taskstats version 1 Sep 4 17:18:31.961523 kernel: Loading compiled-in X.509 certificates Sep 4 17:18:31.961530 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.48-flatcar: 1f5b9f288f9cae6ec9698678cdc0f614482066f7' Sep 4 17:18:31.961538 kernel: Key type .fscrypt registered Sep 4 17:18:31.961545 kernel: Key type fscrypt-provisioning registered Sep 4 17:18:31.961554 kernel: ima: No TPM chip found, activating TPM-bypass! Sep 4 17:18:31.961561 kernel: ima: Allocated hash algorithm: sha1 Sep 4 17:18:31.961568 kernel: ima: No architecture policies found Sep 4 17:18:31.961575 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Sep 4 17:18:31.961583 kernel: clk: Disabling unused clocks Sep 4 17:18:31.961590 kernel: Freeing unused kernel memory: 39040K Sep 4 17:18:31.961597 kernel: Run /init as init process Sep 4 17:18:31.961604 kernel: with arguments: Sep 4 17:18:31.961612 kernel: /init Sep 4 17:18:31.961621 kernel: with environment: Sep 4 17:18:31.961628 kernel: HOME=/ Sep 4 17:18:31.961635 kernel: TERM=linux Sep 4 17:18:31.961642 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Sep 4 17:18:31.961651 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Sep 4 17:18:31.961660 systemd[1]: Detected virtualization kvm. Sep 4 17:18:31.961668 systemd[1]: Detected architecture arm64. Sep 4 17:18:31.961676 systemd[1]: Running in initrd. Sep 4 17:18:31.961686 systemd[1]: No hostname configured, using default hostname. Sep 4 17:18:31.961693 systemd[1]: Hostname set to . Sep 4 17:18:31.961701 systemd[1]: Initializing machine ID from VM UUID. Sep 4 17:18:31.961709 systemd[1]: Queued start job for default target initrd.target. Sep 4 17:18:31.961717 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 4 17:18:31.961725 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 4 17:18:31.961734 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Sep 4 17:18:31.961742 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 4 17:18:31.961751 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Sep 4 17:18:31.961760 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Sep 4 17:18:31.961769 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Sep 4 17:18:31.961777 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Sep 4 17:18:31.961785 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 4 17:18:31.961793 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 4 17:18:31.961802 systemd[1]: Reached target paths.target - Path Units. Sep 4 17:18:31.961810 systemd[1]: Reached target slices.target - Slice Units. Sep 4 17:18:31.961818 systemd[1]: Reached target swap.target - Swaps. Sep 4 17:18:31.961826 systemd[1]: Reached target timers.target - Timer Units. Sep 4 17:18:31.961845 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Sep 4 17:18:31.961863 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 4 17:18:31.961872 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Sep 4 17:18:31.961880 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Sep 4 17:18:31.961888 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 4 17:18:31.961898 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 4 17:18:31.961906 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 4 17:18:31.961914 systemd[1]: Reached target sockets.target - Socket Units. Sep 4 17:18:31.961922 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Sep 4 17:18:31.961930 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 4 17:18:31.961938 systemd[1]: Finished network-cleanup.service - Network Cleanup. Sep 4 17:18:31.961946 systemd[1]: Starting systemd-fsck-usr.service... Sep 4 17:18:31.961954 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 4 17:18:31.961962 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 4 17:18:31.961971 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 4 17:18:31.961979 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Sep 4 17:18:31.961987 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 4 17:18:31.961995 systemd[1]: Finished systemd-fsck-usr.service. Sep 4 17:18:31.962004 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Sep 4 17:18:31.962031 systemd-journald[238]: Collecting audit messages is disabled. Sep 4 17:18:31.962051 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Sep 4 17:18:31.962059 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 4 17:18:31.962069 kernel: Bridge firewalling registered Sep 4 17:18:31.962077 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 4 17:18:31.962085 systemd-journald[238]: Journal started Sep 4 17:18:31.962103 systemd-journald[238]: Runtime Journal (/run/log/journal/20a90041a00f41b08b39e29c8e1bd6b4) is 5.9M, max 47.3M, 41.4M free. Sep 4 17:18:31.944009 systemd-modules-load[239]: Inserted module 'overlay' Sep 4 17:18:31.964395 systemd[1]: Started systemd-journald.service - Journal Service. Sep 4 17:18:31.960084 systemd-modules-load[239]: Inserted module 'br_netfilter' Sep 4 17:18:31.966285 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 4 17:18:31.967605 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 4 17:18:31.980066 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 4 17:18:31.983033 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 4 17:18:31.984594 systemd[1]: Starting systemd-tmpfiles-setup.service - Create Volatile Files and Directories... Sep 4 17:18:31.986070 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 4 17:18:31.988422 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 4 17:18:32.003985 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Sep 4 17:18:32.005340 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 4 17:18:32.007177 systemd[1]: Finished systemd-tmpfiles-setup.service - Create Volatile Files and Directories. Sep 4 17:18:32.012628 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 4 17:18:32.014736 dracut-cmdline[275]: dracut-dracut-053 Sep 4 17:18:32.015592 dracut-cmdline[275]: Using kernel command line parameters: rd.driver.pre=btrfs BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected acpi=force verity.usrhash=7913866621ae0af53522ae1b4ff4e1e453dd69d966d437a439147039341ecbbc Sep 4 17:18:32.041305 systemd-resolved[286]: Positive Trust Anchors: Sep 4 17:18:32.041324 systemd-resolved[286]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 4 17:18:32.041355 systemd-resolved[286]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa corp home internal intranet lan local private test Sep 4 17:18:32.046015 systemd-resolved[286]: Defaulting to hostname 'linux'. Sep 4 17:18:32.046995 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 4 17:18:32.050554 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 4 17:18:32.087870 kernel: SCSI subsystem initialized Sep 4 17:18:32.092852 kernel: Loading iSCSI transport class v2.0-870. Sep 4 17:18:32.099853 kernel: iscsi: registered transport (tcp) Sep 4 17:18:32.113859 kernel: iscsi: registered transport (qla4xxx) Sep 4 17:18:32.113876 kernel: QLogic iSCSI HBA Driver Sep 4 17:18:32.163250 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Sep 4 17:18:32.179016 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Sep 4 17:18:32.197116 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Sep 4 17:18:32.197171 kernel: device-mapper: uevent: version 1.0.3 Sep 4 17:18:32.198413 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Sep 4 17:18:32.245879 kernel: raid6: neonx8 gen() 15767 MB/s Sep 4 17:18:32.262863 kernel: raid6: neonx4 gen() 15670 MB/s Sep 4 17:18:32.279864 kernel: raid6: neonx2 gen() 13239 MB/s Sep 4 17:18:32.296855 kernel: raid6: neonx1 gen() 10479 MB/s Sep 4 17:18:32.313860 kernel: raid6: int64x8 gen() 6962 MB/s Sep 4 17:18:32.330852 kernel: raid6: int64x4 gen() 7327 MB/s Sep 4 17:18:32.347852 kernel: raid6: int64x2 gen() 6128 MB/s Sep 4 17:18:32.364850 kernel: raid6: int64x1 gen() 5041 MB/s Sep 4 17:18:32.364866 kernel: raid6: using algorithm neonx8 gen() 15767 MB/s Sep 4 17:18:32.381854 kernel: raid6: .... xor() 11915 MB/s, rmw enabled Sep 4 17:18:32.381876 kernel: raid6: using neon recovery algorithm Sep 4 17:18:32.386856 kernel: xor: measuring software checksum speed Sep 4 17:18:32.387846 kernel: 8regs : 19883 MB/sec Sep 4 17:18:32.388974 kernel: 32regs : 19735 MB/sec Sep 4 17:18:32.388986 kernel: arm64_neon : 27197 MB/sec Sep 4 17:18:32.388995 kernel: xor: using function: arm64_neon (27197 MB/sec) Sep 4 17:18:32.443867 kernel: Btrfs loaded, zoned=no, fsverity=no Sep 4 17:18:32.456550 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Sep 4 17:18:32.467976 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 4 17:18:32.482232 systemd-udevd[461]: Using default interface naming scheme 'v255'. Sep 4 17:18:32.485403 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 4 17:18:32.498063 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Sep 4 17:18:32.510863 dracut-pre-trigger[469]: rd.md=0: removing MD RAID activation Sep 4 17:18:32.541628 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Sep 4 17:18:32.553007 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 4 17:18:32.597057 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 4 17:18:32.606016 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Sep 4 17:18:32.628565 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Sep 4 17:18:32.630598 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Sep 4 17:18:32.631886 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 4 17:18:32.634096 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 4 17:18:32.643917 kernel: virtio_blk virtio1: 1/0/0 default/read/poll queues Sep 4 17:18:32.644114 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Sep 4 17:18:32.644035 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Sep 4 17:18:32.658186 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Sep 4 17:18:32.661455 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Sep 4 17:18:32.661477 kernel: GPT:9289727 != 19775487 Sep 4 17:18:32.661486 kernel: GPT:Alternate GPT header not at the end of the disk. Sep 4 17:18:32.662927 kernel: GPT:9289727 != 19775487 Sep 4 17:18:32.662964 kernel: GPT: Use GNU Parted to correct GPT errors. Sep 4 17:18:32.667680 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 4 17:18:32.669533 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Sep 4 17:18:32.669661 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 4 17:18:32.673209 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 4 17:18:32.674441 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 4 17:18:32.674782 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 4 17:18:32.677295 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Sep 4 17:18:32.687527 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 4 17:18:32.691473 kernel: BTRFS: device fsid 2be47701-3393-455e-86fc-33755ceb9c20 devid 1 transid 35 /dev/vda3 scanned by (udev-worker) (517) Sep 4 17:18:32.691498 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/vda6 scanned by (udev-worker) (525) Sep 4 17:18:32.702738 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Sep 4 17:18:32.704224 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 4 17:18:32.712658 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Sep 4 17:18:32.716850 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Sep 4 17:18:32.717774 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Sep 4 17:18:32.723271 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Sep 4 17:18:32.740019 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Sep 4 17:18:32.742151 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Sep 4 17:18:32.746286 disk-uuid[553]: Primary Header is updated. Sep 4 17:18:32.746286 disk-uuid[553]: Secondary Entries is updated. Sep 4 17:18:32.746286 disk-uuid[553]: Secondary Header is updated. Sep 4 17:18:32.749858 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 4 17:18:32.767118 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 4 17:18:33.768806 disk-uuid[554]: The operation has completed successfully. Sep 4 17:18:33.769776 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Sep 4 17:18:33.800465 systemd[1]: disk-uuid.service: Deactivated successfully. Sep 4 17:18:33.800567 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Sep 4 17:18:33.820019 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Sep 4 17:18:33.824914 sh[575]: Success Sep 4 17:18:33.840044 kernel: device-mapper: verity: sha256 using implementation "sha256-ce" Sep 4 17:18:33.882045 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Sep 4 17:18:33.896133 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Sep 4 17:18:33.896976 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Sep 4 17:18:33.908532 kernel: BTRFS info (device dm-0): first mount of filesystem 2be47701-3393-455e-86fc-33755ceb9c20 Sep 4 17:18:33.908585 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm Sep 4 17:18:33.908597 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Sep 4 17:18:33.909332 kernel: BTRFS info (device dm-0): disabling log replay at mount time Sep 4 17:18:33.910844 kernel: BTRFS info (device dm-0): using free space tree Sep 4 17:18:33.916751 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Sep 4 17:18:33.917778 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Sep 4 17:18:33.926003 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Sep 4 17:18:33.927519 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Sep 4 17:18:33.940353 kernel: BTRFS info (device vda6): first mount of filesystem 26eaee0d-fa47-45db-8665-f2efa4a46ac0 Sep 4 17:18:33.940399 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm Sep 4 17:18:33.940411 kernel: BTRFS info (device vda6): using free space tree Sep 4 17:18:33.944865 kernel: BTRFS info (device vda6): auto enabling async discard Sep 4 17:18:33.953225 systemd[1]: mnt-oem.mount: Deactivated successfully. Sep 4 17:18:33.954903 kernel: BTRFS info (device vda6): last unmount of filesystem 26eaee0d-fa47-45db-8665-f2efa4a46ac0 Sep 4 17:18:33.961968 systemd[1]: Finished ignition-setup.service - Ignition (setup). Sep 4 17:18:33.972044 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Sep 4 17:18:34.045095 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 4 17:18:34.065059 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 4 17:18:34.089688 systemd-networkd[760]: lo: Link UP Sep 4 17:18:34.089699 systemd-networkd[760]: lo: Gained carrier Sep 4 17:18:34.090381 systemd-networkd[760]: Enumeration completed Sep 4 17:18:34.090485 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 4 17:18:34.090845 systemd-networkd[760]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 4 17:18:34.090848 systemd-networkd[760]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 4 17:18:34.091842 systemd[1]: Reached target network.target - Network. Sep 4 17:18:34.094269 systemd-networkd[760]: eth0: Link UP Sep 4 17:18:34.094273 systemd-networkd[760]: eth0: Gained carrier Sep 4 17:18:34.094291 systemd-networkd[760]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 4 17:18:34.103927 ignition[676]: Ignition 2.18.0 Sep 4 17:18:34.103938 ignition[676]: Stage: fetch-offline Sep 4 17:18:34.103986 ignition[676]: no configs at "/usr/lib/ignition/base.d" Sep 4 17:18:34.103995 ignition[676]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 4 17:18:34.104087 ignition[676]: parsed url from cmdline: "" Sep 4 17:18:34.104091 ignition[676]: no config URL provided Sep 4 17:18:34.104095 ignition[676]: reading system config file "/usr/lib/ignition/user.ign" Sep 4 17:18:34.104102 ignition[676]: no config at "/usr/lib/ignition/user.ign" Sep 4 17:18:34.104133 ignition[676]: op(1): [started] loading QEMU firmware config module Sep 4 17:18:34.104145 ignition[676]: op(1): executing: "modprobe" "qemu_fw_cfg" Sep 4 17:18:34.110014 systemd-networkd[760]: eth0: DHCPv4 address 10.0.0.60/16, gateway 10.0.0.1 acquired from 10.0.0.1 Sep 4 17:18:34.121036 ignition[676]: op(1): [finished] loading QEMU firmware config module Sep 4 17:18:34.160625 ignition[676]: parsing config with SHA512: 5fcbef9437ea4022d8ee3d2827e2d3dbaf24e47a5213da0228c0fd90647d5ed03479b574b8eac93d91f1bc1e8e412e85877010281a6863879047cea1acc28ef0 Sep 4 17:18:34.164773 unknown[676]: fetched base config from "system" Sep 4 17:18:34.164782 unknown[676]: fetched user config from "qemu" Sep 4 17:18:34.165173 ignition[676]: fetch-offline: fetch-offline passed Sep 4 17:18:34.165229 ignition[676]: Ignition finished successfully Sep 4 17:18:34.167299 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Sep 4 17:18:34.168929 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Sep 4 17:18:34.183063 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Sep 4 17:18:34.195975 ignition[772]: Ignition 2.18.0 Sep 4 17:18:34.195984 ignition[772]: Stage: kargs Sep 4 17:18:34.196147 ignition[772]: no configs at "/usr/lib/ignition/base.d" Sep 4 17:18:34.196156 ignition[772]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 4 17:18:34.197019 ignition[772]: kargs: kargs passed Sep 4 17:18:34.198997 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Sep 4 17:18:34.197064 ignition[772]: Ignition finished successfully Sep 4 17:18:34.207002 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Sep 4 17:18:34.216652 ignition[781]: Ignition 2.18.0 Sep 4 17:18:34.216660 ignition[781]: Stage: disks Sep 4 17:18:34.216807 ignition[781]: no configs at "/usr/lib/ignition/base.d" Sep 4 17:18:34.216816 ignition[781]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 4 17:18:34.219668 systemd[1]: Finished ignition-disks.service - Ignition (disks). Sep 4 17:18:34.217627 ignition[781]: disks: disks passed Sep 4 17:18:34.220795 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Sep 4 17:18:34.217669 ignition[781]: Ignition finished successfully Sep 4 17:18:34.222548 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Sep 4 17:18:34.224435 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 4 17:18:34.225795 systemd[1]: Reached target sysinit.target - System Initialization. Sep 4 17:18:34.227810 systemd[1]: Reached target basic.target - Basic System. Sep 4 17:18:34.241029 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Sep 4 17:18:34.260934 systemd-fsck[792]: ROOT: clean, 14/553520 files, 52654/553472 blocks Sep 4 17:18:34.268647 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Sep 4 17:18:34.280995 systemd[1]: Mounting sysroot.mount - /sysroot... Sep 4 17:18:34.350863 kernel: EXT4-fs (vda9): mounted filesystem f2f4f3ba-c5a3-49c0-ace4-444935e9934b r/w with ordered data mode. Quota mode: none. Sep 4 17:18:34.351499 systemd[1]: Mounted sysroot.mount - /sysroot. Sep 4 17:18:34.352961 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Sep 4 17:18:34.364937 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 4 17:18:34.375956 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Sep 4 17:18:34.376718 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Sep 4 17:18:34.376761 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Sep 4 17:18:34.376787 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Sep 4 17:18:34.381498 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Sep 4 17:18:34.387369 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/vda6 scanned by mount (800) Sep 4 17:18:34.384343 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Sep 4 17:18:34.390899 kernel: BTRFS info (device vda6): first mount of filesystem 26eaee0d-fa47-45db-8665-f2efa4a46ac0 Sep 4 17:18:34.390966 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm Sep 4 17:18:34.391012 kernel: BTRFS info (device vda6): using free space tree Sep 4 17:18:34.392853 kernel: BTRFS info (device vda6): auto enabling async discard Sep 4 17:18:34.394597 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 4 17:18:34.481886 initrd-setup-root[824]: cut: /sysroot/etc/passwd: No such file or directory Sep 4 17:18:34.486380 initrd-setup-root[831]: cut: /sysroot/etc/group: No such file or directory Sep 4 17:18:34.489273 initrd-setup-root[838]: cut: /sysroot/etc/shadow: No such file or directory Sep 4 17:18:34.492429 initrd-setup-root[845]: cut: /sysroot/etc/gshadow: No such file or directory Sep 4 17:18:34.571695 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Sep 4 17:18:34.579027 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Sep 4 17:18:34.581267 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Sep 4 17:18:34.586846 kernel: BTRFS info (device vda6): last unmount of filesystem 26eaee0d-fa47-45db-8665-f2efa4a46ac0 Sep 4 17:18:34.602441 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Sep 4 17:18:34.605751 ignition[913]: INFO : Ignition 2.18.0 Sep 4 17:18:34.605751 ignition[913]: INFO : Stage: mount Sep 4 17:18:34.607338 ignition[913]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 4 17:18:34.607338 ignition[913]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 4 17:18:34.607338 ignition[913]: INFO : mount: mount passed Sep 4 17:18:34.607338 ignition[913]: INFO : Ignition finished successfully Sep 4 17:18:34.608293 systemd[1]: Finished ignition-mount.service - Ignition (mount). Sep 4 17:18:34.614960 systemd[1]: Starting ignition-files.service - Ignition (files)... Sep 4 17:18:34.907764 systemd[1]: sysroot-oem.mount: Deactivated successfully. Sep 4 17:18:34.917026 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Sep 4 17:18:34.927445 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/vda6 scanned by mount (928) Sep 4 17:18:34.927485 kernel: BTRFS info (device vda6): first mount of filesystem 26eaee0d-fa47-45db-8665-f2efa4a46ac0 Sep 4 17:18:34.927496 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm Sep 4 17:18:34.928844 kernel: BTRFS info (device vda6): using free space tree Sep 4 17:18:34.933841 kernel: BTRFS info (device vda6): auto enabling async discard Sep 4 17:18:34.934919 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Sep 4 17:18:34.957982 ignition[945]: INFO : Ignition 2.18.0 Sep 4 17:18:34.957982 ignition[945]: INFO : Stage: files Sep 4 17:18:34.959617 ignition[945]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 4 17:18:34.959617 ignition[945]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 4 17:18:34.961764 ignition[945]: DEBUG : files: compiled without relabeling support, skipping Sep 4 17:18:34.962967 ignition[945]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Sep 4 17:18:34.962967 ignition[945]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Sep 4 17:18:34.965865 ignition[945]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Sep 4 17:18:34.965865 ignition[945]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Sep 4 17:18:34.965865 ignition[945]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Sep 4 17:18:34.965726 unknown[945]: wrote ssh authorized keys file for user: core Sep 4 17:18:34.971591 ignition[945]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Sep 4 17:18:34.971591 ignition[945]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-arm64.tar.gz: attempt #1 Sep 4 17:18:35.044190 ignition[945]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Sep 4 17:18:35.097227 ignition[945]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Sep 4 17:18:35.097227 ignition[945]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Sep 4 17:18:35.101363 ignition[945]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Sep 4 17:18:35.101363 ignition[945]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Sep 4 17:18:35.101363 ignition[945]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Sep 4 17:18:35.101363 ignition[945]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 4 17:18:35.101363 ignition[945]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 4 17:18:35.101363 ignition[945]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 4 17:18:35.101363 ignition[945]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 4 17:18:35.101363 ignition[945]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Sep 4 17:18:35.101363 ignition[945]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Sep 4 17:18:35.101363 ignition[945]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.29.2-arm64.raw" Sep 4 17:18:35.101363 ignition[945]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.29.2-arm64.raw" Sep 4 17:18:35.101363 ignition[945]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.29.2-arm64.raw" Sep 4 17:18:35.101363 ignition[945]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.29.2-arm64.raw: attempt #1 Sep 4 17:18:35.190181 systemd-networkd[760]: eth0: Gained IPv6LL Sep 4 17:18:35.430671 ignition[945]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Sep 4 17:18:35.997841 ignition[945]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.29.2-arm64.raw" Sep 4 17:18:35.997841 ignition[945]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Sep 4 17:18:36.001401 ignition[945]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 4 17:18:36.001401 ignition[945]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 4 17:18:36.001401 ignition[945]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Sep 4 17:18:36.001401 ignition[945]: INFO : files: op(d): [started] processing unit "coreos-metadata.service" Sep 4 17:18:36.001401 ignition[945]: INFO : files: op(d): op(e): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Sep 4 17:18:36.001401 ignition[945]: INFO : files: op(d): op(e): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Sep 4 17:18:36.001401 ignition[945]: INFO : files: op(d): [finished] processing unit "coreos-metadata.service" Sep 4 17:18:36.001401 ignition[945]: INFO : files: op(f): [started] setting preset to disabled for "coreos-metadata.service" Sep 4 17:18:36.027227 ignition[945]: INFO : files: op(f): op(10): [started] removing enablement symlink(s) for "coreos-metadata.service" Sep 4 17:18:36.030934 ignition[945]: INFO : files: op(f): op(10): [finished] removing enablement symlink(s) for "coreos-metadata.service" Sep 4 17:18:36.033428 ignition[945]: INFO : files: op(f): [finished] setting preset to disabled for "coreos-metadata.service" Sep 4 17:18:36.033428 ignition[945]: INFO : files: op(11): [started] setting preset to enabled for "prepare-helm.service" Sep 4 17:18:36.033428 ignition[945]: INFO : files: op(11): [finished] setting preset to enabled for "prepare-helm.service" Sep 4 17:18:36.033428 ignition[945]: INFO : files: createResultFile: createFiles: op(12): [started] writing file "/sysroot/etc/.ignition-result.json" Sep 4 17:18:36.033428 ignition[945]: INFO : files: createResultFile: createFiles: op(12): [finished] writing file "/sysroot/etc/.ignition-result.json" Sep 4 17:18:36.033428 ignition[945]: INFO : files: files passed Sep 4 17:18:36.033428 ignition[945]: INFO : Ignition finished successfully Sep 4 17:18:36.033790 systemd[1]: Finished ignition-files.service - Ignition (files). Sep 4 17:18:36.040007 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Sep 4 17:18:36.045007 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Sep 4 17:18:36.048108 systemd[1]: ignition-quench.service: Deactivated successfully. Sep 4 17:18:36.048186 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Sep 4 17:18:36.053046 initrd-setup-root-after-ignition[973]: grep: /sysroot/oem/oem-release: No such file or directory Sep 4 17:18:36.054791 initrd-setup-root-after-ignition[975]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 4 17:18:36.054791 initrd-setup-root-after-ignition[975]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Sep 4 17:18:36.057886 initrd-setup-root-after-ignition[979]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 4 17:18:36.059101 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 4 17:18:36.062297 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Sep 4 17:18:36.074568 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Sep 4 17:18:36.094934 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Sep 4 17:18:36.095064 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Sep 4 17:18:36.097343 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Sep 4 17:18:36.098961 systemd[1]: Reached target initrd.target - Initrd Default Target. Sep 4 17:18:36.100457 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Sep 4 17:18:36.101200 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Sep 4 17:18:36.117821 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 4 17:18:36.134579 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Sep 4 17:18:36.143683 systemd[1]: Stopped target network.target - Network. Sep 4 17:18:36.144758 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Sep 4 17:18:36.146968 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 4 17:18:36.149006 systemd[1]: Stopped target timers.target - Timer Units. Sep 4 17:18:36.150666 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Sep 4 17:18:36.150782 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Sep 4 17:18:36.153540 systemd[1]: Stopped target initrd.target - Initrd Default Target. Sep 4 17:18:36.155447 systemd[1]: Stopped target basic.target - Basic System. Sep 4 17:18:36.156806 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Sep 4 17:18:36.158203 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Sep 4 17:18:36.159791 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Sep 4 17:18:36.161692 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Sep 4 17:18:36.163456 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Sep 4 17:18:36.164966 systemd[1]: Stopped target sysinit.target - System Initialization. Sep 4 17:18:36.167030 systemd[1]: Stopped target local-fs.target - Local File Systems. Sep 4 17:18:36.168582 systemd[1]: Stopped target swap.target - Swaps. Sep 4 17:18:36.169958 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Sep 4 17:18:36.170079 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Sep 4 17:18:36.172497 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Sep 4 17:18:36.174393 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 4 17:18:36.176161 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Sep 4 17:18:36.177000 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 4 17:18:36.178232 systemd[1]: dracut-initqueue.service: Deactivated successfully. Sep 4 17:18:36.178353 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Sep 4 17:18:36.180859 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Sep 4 17:18:36.180971 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Sep 4 17:18:36.182917 systemd[1]: Stopped target paths.target - Path Units. Sep 4 17:18:36.184559 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Sep 4 17:18:36.185924 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 4 17:18:36.187528 systemd[1]: Stopped target slices.target - Slice Units. Sep 4 17:18:36.189017 systemd[1]: Stopped target sockets.target - Socket Units. Sep 4 17:18:36.190564 systemd[1]: iscsid.socket: Deactivated successfully. Sep 4 17:18:36.190653 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Sep 4 17:18:36.192602 systemd[1]: iscsiuio.socket: Deactivated successfully. Sep 4 17:18:36.192676 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Sep 4 17:18:36.194032 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Sep 4 17:18:36.194148 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Sep 4 17:18:36.195664 systemd[1]: ignition-files.service: Deactivated successfully. Sep 4 17:18:36.195772 systemd[1]: Stopped ignition-files.service - Ignition (files). Sep 4 17:18:36.207028 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Sep 4 17:18:36.208780 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Sep 4 17:18:36.209849 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Sep 4 17:18:36.211851 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Sep 4 17:18:36.213375 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Sep 4 17:18:36.213500 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Sep 4 17:18:36.218528 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Sep 4 17:18:36.218639 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Sep 4 17:18:36.219161 systemd-networkd[760]: eth0: DHCPv6 lease lost Sep 4 17:18:36.225195 ignition[1000]: INFO : Ignition 2.18.0 Sep 4 17:18:36.225195 ignition[1000]: INFO : Stage: umount Sep 4 17:18:36.225195 ignition[1000]: INFO : no configs at "/usr/lib/ignition/base.d" Sep 4 17:18:36.225195 ignition[1000]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Sep 4 17:18:36.225195 ignition[1000]: INFO : umount: umount passed Sep 4 17:18:36.225195 ignition[1000]: INFO : Ignition finished successfully Sep 4 17:18:36.225685 systemd[1]: sysroot-boot.mount: Deactivated successfully. Sep 4 17:18:36.226544 systemd[1]: systemd-networkd.service: Deactivated successfully. Sep 4 17:18:36.226643 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Sep 4 17:18:36.230151 systemd[1]: systemd-resolved.service: Deactivated successfully. Sep 4 17:18:36.230344 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Sep 4 17:18:36.234415 systemd[1]: ignition-mount.service: Deactivated successfully. Sep 4 17:18:36.234502 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Sep 4 17:18:36.237525 systemd[1]: initrd-cleanup.service: Deactivated successfully. Sep 4 17:18:36.237642 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Sep 4 17:18:36.243333 systemd[1]: systemd-networkd.socket: Deactivated successfully. Sep 4 17:18:36.243372 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Sep 4 17:18:36.244752 systemd[1]: ignition-disks.service: Deactivated successfully. Sep 4 17:18:36.244801 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Sep 4 17:18:36.246317 systemd[1]: ignition-kargs.service: Deactivated successfully. Sep 4 17:18:36.246361 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Sep 4 17:18:36.247750 systemd[1]: ignition-setup.service: Deactivated successfully. Sep 4 17:18:36.247790 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Sep 4 17:18:36.249691 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Sep 4 17:18:36.249737 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Sep 4 17:18:36.263008 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Sep 4 17:18:36.263923 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Sep 4 17:18:36.263988 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Sep 4 17:18:36.265797 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 4 17:18:36.265855 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Sep 4 17:18:36.268039 systemd[1]: systemd-modules-load.service: Deactivated successfully. Sep 4 17:18:36.268084 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Sep 4 17:18:36.269930 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Sep 4 17:18:36.269971 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create Volatile Files and Directories. Sep 4 17:18:36.272055 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 4 17:18:36.282411 systemd[1]: network-cleanup.service: Deactivated successfully. Sep 4 17:18:36.283251 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Sep 4 17:18:36.287534 systemd[1]: systemd-udevd.service: Deactivated successfully. Sep 4 17:18:36.288513 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 4 17:18:36.289945 systemd[1]: sysroot-boot.service: Deactivated successfully. Sep 4 17:18:36.290029 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Sep 4 17:18:36.292294 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Sep 4 17:18:36.292369 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Sep 4 17:18:36.293610 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Sep 4 17:18:36.293646 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Sep 4 17:18:36.295141 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Sep 4 17:18:36.295189 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Sep 4 17:18:36.297732 systemd[1]: dracut-cmdline.service: Deactivated successfully. Sep 4 17:18:36.297777 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Sep 4 17:18:36.300400 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Sep 4 17:18:36.300447 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Sep 4 17:18:36.302772 systemd[1]: initrd-setup-root.service: Deactivated successfully. Sep 4 17:18:36.302816 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Sep 4 17:18:36.315007 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Sep 4 17:18:36.316075 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Sep 4 17:18:36.316136 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 4 17:18:36.318003 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Sep 4 17:18:36.318044 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 4 17:18:36.319896 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Sep 4 17:18:36.319944 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Sep 4 17:18:36.322180 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 4 17:18:36.322223 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Sep 4 17:18:36.324429 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Sep 4 17:18:36.324535 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Sep 4 17:18:36.326783 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Sep 4 17:18:36.329129 systemd[1]: Starting initrd-switch-root.service - Switch Root... Sep 4 17:18:36.339442 systemd[1]: Switching root. Sep 4 17:18:36.362479 systemd-journald[238]: Journal stopped Sep 4 17:18:37.103737 systemd-journald[238]: Received SIGTERM from PID 1 (systemd). Sep 4 17:18:37.103794 kernel: SELinux: policy capability network_peer_controls=1 Sep 4 17:18:37.103808 kernel: SELinux: policy capability open_perms=1 Sep 4 17:18:37.103818 kernel: SELinux: policy capability extended_socket_class=1 Sep 4 17:18:37.103863 kernel: SELinux: policy capability always_check_network=0 Sep 4 17:18:37.103876 kernel: SELinux: policy capability cgroup_seclabel=1 Sep 4 17:18:37.103886 kernel: SELinux: policy capability nnp_nosuid_transition=1 Sep 4 17:18:37.103896 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Sep 4 17:18:37.103910 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Sep 4 17:18:37.103923 kernel: audit: type=1403 audit(1725470316.505:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Sep 4 17:18:37.103934 systemd[1]: Successfully loaded SELinux policy in 36.376ms. Sep 4 17:18:37.103956 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 9.037ms. Sep 4 17:18:37.103968 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Sep 4 17:18:37.103978 systemd[1]: Detected virtualization kvm. Sep 4 17:18:37.103989 systemd[1]: Detected architecture arm64. Sep 4 17:18:37.103999 systemd[1]: Detected first boot. Sep 4 17:18:37.104010 systemd[1]: Initializing machine ID from VM UUID. Sep 4 17:18:37.104021 zram_generator::config[1044]: No configuration found. Sep 4 17:18:37.104034 systemd[1]: Populated /etc with preset unit settings. Sep 4 17:18:37.104046 systemd[1]: initrd-switch-root.service: Deactivated successfully. Sep 4 17:18:37.104057 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Sep 4 17:18:37.104067 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Sep 4 17:18:37.104078 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Sep 4 17:18:37.104089 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Sep 4 17:18:37.104099 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Sep 4 17:18:37.104109 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Sep 4 17:18:37.104120 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Sep 4 17:18:37.104130 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Sep 4 17:18:37.104143 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Sep 4 17:18:37.104153 systemd[1]: Created slice user.slice - User and Session Slice. Sep 4 17:18:37.104164 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Sep 4 17:18:37.104175 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Sep 4 17:18:37.104186 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Sep 4 17:18:37.104200 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Sep 4 17:18:37.104211 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Sep 4 17:18:37.104224 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Sep 4 17:18:37.104235 systemd[1]: Expecting device dev-ttyAMA0.device - /dev/ttyAMA0... Sep 4 17:18:37.104246 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Sep 4 17:18:37.104257 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Sep 4 17:18:37.104286 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Sep 4 17:18:37.104301 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Sep 4 17:18:37.104313 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Sep 4 17:18:37.104324 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Sep 4 17:18:37.104335 systemd[1]: Reached target remote-fs.target - Remote File Systems. Sep 4 17:18:37.104345 systemd[1]: Reached target slices.target - Slice Units. Sep 4 17:18:37.104358 systemd[1]: Reached target swap.target - Swaps. Sep 4 17:18:37.104369 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Sep 4 17:18:37.104381 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Sep 4 17:18:37.104391 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Sep 4 17:18:37.104402 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Sep 4 17:18:37.104413 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Sep 4 17:18:37.104425 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Sep 4 17:18:37.104436 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Sep 4 17:18:37.104447 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Sep 4 17:18:37.104458 systemd[1]: Mounting media.mount - External Media Directory... Sep 4 17:18:37.104468 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Sep 4 17:18:37.104479 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Sep 4 17:18:37.104489 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Sep 4 17:18:37.104500 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Sep 4 17:18:37.104510 systemd[1]: Reached target machines.target - Containers. Sep 4 17:18:37.104521 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Sep 4 17:18:37.104532 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 4 17:18:37.104543 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Sep 4 17:18:37.104558 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Sep 4 17:18:37.104570 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 4 17:18:37.104580 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 4 17:18:37.104601 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 4 17:18:37.104611 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Sep 4 17:18:37.104621 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 4 17:18:37.104631 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Sep 4 17:18:37.104642 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Sep 4 17:18:37.104654 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Sep 4 17:18:37.104668 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Sep 4 17:18:37.104679 systemd[1]: Stopped systemd-fsck-usr.service. Sep 4 17:18:37.104689 systemd[1]: Starting systemd-journald.service - Journal Service... Sep 4 17:18:37.104699 kernel: loop: module loaded Sep 4 17:18:37.104708 kernel: fuse: init (API version 7.39) Sep 4 17:18:37.104719 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Sep 4 17:18:37.104730 kernel: ACPI: bus type drm_connector registered Sep 4 17:18:37.104741 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Sep 4 17:18:37.104752 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Sep 4 17:18:37.104762 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Sep 4 17:18:37.104772 systemd[1]: verity-setup.service: Deactivated successfully. Sep 4 17:18:37.104783 systemd[1]: Stopped verity-setup.service. Sep 4 17:18:37.104793 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Sep 4 17:18:37.104804 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Sep 4 17:18:37.104814 systemd[1]: Mounted media.mount - External Media Directory. Sep 4 17:18:37.104825 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Sep 4 17:18:37.104882 systemd-journald[1103]: Collecting audit messages is disabled. Sep 4 17:18:37.104907 systemd-journald[1103]: Journal started Sep 4 17:18:37.104933 systemd-journald[1103]: Runtime Journal (/run/log/journal/20a90041a00f41b08b39e29c8e1bd6b4) is 5.9M, max 47.3M, 41.4M free. Sep 4 17:18:36.882109 systemd[1]: Queued start job for default target multi-user.target. Sep 4 17:18:36.907046 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Sep 4 17:18:36.907449 systemd[1]: systemd-journald.service: Deactivated successfully. Sep 4 17:18:37.108669 systemd[1]: Started systemd-journald.service - Journal Service. Sep 4 17:18:37.109201 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Sep 4 17:18:37.110380 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Sep 4 17:18:37.111559 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Sep 4 17:18:37.113142 systemd[1]: modprobe@configfs.service: Deactivated successfully. Sep 4 17:18:37.113930 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Sep 4 17:18:37.115501 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 4 17:18:37.115650 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 4 17:18:37.117090 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 4 17:18:37.117224 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 4 17:18:37.118342 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 4 17:18:37.119884 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 4 17:18:37.121254 systemd[1]: modprobe@fuse.service: Deactivated successfully. Sep 4 17:18:37.121409 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Sep 4 17:18:37.123171 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 4 17:18:37.123332 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 4 17:18:37.125050 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Sep 4 17:18:37.126386 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Sep 4 17:18:37.128079 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Sep 4 17:18:37.140518 systemd[1]: Reached target network-pre.target - Preparation for Network. Sep 4 17:18:37.151427 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Sep 4 17:18:37.153623 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Sep 4 17:18:37.154882 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Sep 4 17:18:37.154931 systemd[1]: Reached target local-fs.target - Local File Systems. Sep 4 17:18:37.156798 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Sep 4 17:18:37.159892 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Sep 4 17:18:37.164233 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Sep 4 17:18:37.165369 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 4 17:18:37.170137 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Sep 4 17:18:37.171920 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Sep 4 17:18:37.173101 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 4 17:18:37.174159 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Sep 4 17:18:37.176143 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 4 17:18:37.177319 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Sep 4 17:18:37.182586 systemd-journald[1103]: Time spent on flushing to /var/log/journal/20a90041a00f41b08b39e29c8e1bd6b4 is 16.161ms for 853 entries. Sep 4 17:18:37.182586 systemd-journald[1103]: System Journal (/var/log/journal/20a90041a00f41b08b39e29c8e1bd6b4) is 8.0M, max 195.6M, 187.6M free. Sep 4 17:18:37.230487 systemd-journald[1103]: Received client request to flush runtime journal. Sep 4 17:18:37.230543 kernel: loop0: detected capacity change from 0 to 113672 Sep 4 17:18:37.230557 kernel: block loop0: the capability attribute has been deprecated. Sep 4 17:18:37.183028 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Sep 4 17:18:37.188228 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Sep 4 17:18:37.192426 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Sep 4 17:18:37.193905 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Sep 4 17:18:37.195014 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Sep 4 17:18:37.196010 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Sep 4 17:18:37.197235 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Sep 4 17:18:37.203992 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Sep 4 17:18:37.222471 udevadm[1164]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Sep 4 17:18:37.230971 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Sep 4 17:18:37.234132 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Sep 4 17:18:37.237701 systemd-tmpfiles[1155]: ACLs are not supported, ignoring. Sep 4 17:18:37.246506 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Sep 4 17:18:37.237719 systemd-tmpfiles[1155]: ACLs are not supported, ignoring. Sep 4 17:18:37.246436 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Sep 4 17:18:37.249152 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Sep 4 17:18:37.250574 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Sep 4 17:18:37.253206 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Sep 4 17:18:37.257011 systemd[1]: Starting systemd-sysusers.service - Create System Users... Sep 4 17:18:37.271963 kernel: loop1: detected capacity change from 0 to 59688 Sep 4 17:18:37.297505 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Sep 4 17:18:37.298971 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Sep 4 17:18:37.301605 systemd[1]: Finished systemd-sysusers.service - Create System Users. Sep 4 17:18:37.312171 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Sep 4 17:18:37.322119 kernel: loop2: detected capacity change from 0 to 194512 Sep 4 17:18:37.326367 systemd-tmpfiles[1181]: ACLs are not supported, ignoring. Sep 4 17:18:37.326384 systemd-tmpfiles[1181]: ACLs are not supported, ignoring. Sep 4 17:18:37.330305 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Sep 4 17:18:37.354863 kernel: loop3: detected capacity change from 0 to 113672 Sep 4 17:18:37.358847 kernel: loop4: detected capacity change from 0 to 59688 Sep 4 17:18:37.363856 kernel: loop5: detected capacity change from 0 to 194512 Sep 4 17:18:37.368339 (sd-merge)[1185]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. Sep 4 17:18:37.368736 (sd-merge)[1185]: Merged extensions into '/usr'. Sep 4 17:18:37.374574 systemd[1]: Reloading requested from client PID 1154 ('systemd-sysext') (unit systemd-sysext.service)... Sep 4 17:18:37.374732 systemd[1]: Reloading... Sep 4 17:18:37.440879 zram_generator::config[1210]: No configuration found. Sep 4 17:18:37.535552 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 4 17:18:37.538265 ldconfig[1149]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Sep 4 17:18:37.573497 systemd[1]: Reloading finished in 196 ms. Sep 4 17:18:37.608530 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Sep 4 17:18:37.609967 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Sep 4 17:18:37.619998 systemd[1]: Starting ensure-sysext.service... Sep 4 17:18:37.621826 systemd[1]: Starting systemd-tmpfiles-setup.service - Create Volatile Files and Directories... Sep 4 17:18:37.629108 systemd[1]: Reloading requested from client PID 1243 ('systemctl') (unit ensure-sysext.service)... Sep 4 17:18:37.629122 systemd[1]: Reloading... Sep 4 17:18:37.645992 systemd-tmpfiles[1244]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Sep 4 17:18:37.646579 systemd-tmpfiles[1244]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Sep 4 17:18:37.650706 systemd-tmpfiles[1244]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Sep 4 17:18:37.651092 systemd-tmpfiles[1244]: ACLs are not supported, ignoring. Sep 4 17:18:37.651204 systemd-tmpfiles[1244]: ACLs are not supported, ignoring. Sep 4 17:18:37.653394 systemd-tmpfiles[1244]: Detected autofs mount point /boot during canonicalization of boot. Sep 4 17:18:37.653505 systemd-tmpfiles[1244]: Skipping /boot Sep 4 17:18:37.666185 systemd-tmpfiles[1244]: Detected autofs mount point /boot during canonicalization of boot. Sep 4 17:18:37.666326 systemd-tmpfiles[1244]: Skipping /boot Sep 4 17:18:37.683068 zram_generator::config[1272]: No configuration found. Sep 4 17:18:37.772865 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 4 17:18:37.810940 systemd[1]: Reloading finished in 181 ms. Sep 4 17:18:37.824067 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Sep 4 17:18:37.837314 systemd[1]: Finished systemd-tmpfiles-setup.service - Create Volatile Files and Directories. Sep 4 17:18:37.845346 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Sep 4 17:18:37.847877 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Sep 4 17:18:37.849811 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Sep 4 17:18:37.854116 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Sep 4 17:18:37.859098 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Sep 4 17:18:37.864158 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Sep 4 17:18:37.867616 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 4 17:18:37.868743 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 4 17:18:37.874091 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 4 17:18:37.876198 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 4 17:18:37.877172 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 4 17:18:37.895186 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Sep 4 17:18:37.899872 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Sep 4 17:18:37.901805 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 4 17:18:37.901960 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 4 17:18:37.906411 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 4 17:18:37.906601 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 4 17:18:37.909243 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 4 17:18:37.909388 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 4 17:18:37.910014 systemd-udevd[1311]: Using default interface naming scheme 'v255'. Sep 4 17:18:37.918803 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 4 17:18:37.923182 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 4 17:18:37.926217 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 4 17:18:37.929197 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 4 17:18:37.930393 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 4 17:18:37.935106 systemd[1]: Starting systemd-update-done.service - Update is Completed... Sep 4 17:18:37.936525 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Sep 4 17:18:37.939280 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Sep 4 17:18:37.944898 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Sep 4 17:18:37.946601 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 4 17:18:37.946745 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 4 17:18:37.948778 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 4 17:18:37.949111 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 4 17:18:37.951098 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 4 17:18:37.951229 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 4 17:18:37.966973 systemd[1]: Started systemd-userdbd.service - User Database Manager. Sep 4 17:18:37.973823 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Sep 4 17:18:37.975453 augenrules[1364]: No rules Sep 4 17:18:37.982303 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Sep 4 17:18:37.985094 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Sep 4 17:18:37.989133 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Sep 4 17:18:37.993121 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Sep 4 17:18:37.994020 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Sep 4 17:18:37.995939 systemd[1]: Starting systemd-networkd.service - Network Configuration... Sep 4 17:18:37.997288 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 4 17:18:37.999854 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Sep 4 17:18:38.001386 systemd[1]: Finished systemd-update-done.service - Update is Completed. Sep 4 17:18:38.003411 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 4 17:18:38.003942 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Sep 4 17:18:38.005552 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 4 17:18:38.005690 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Sep 4 17:18:38.008495 kernel: BTRFS info: devid 1 device path /dev/mapper/usr changed to /dev/dm-0 scanned by (udev-worker) (1348) Sep 4 17:18:38.008343 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 4 17:18:38.008500 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Sep 4 17:18:38.018887 systemd[1]: Finished ensure-sysext.service. Sep 4 17:18:38.024495 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 4 17:18:38.024686 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Sep 4 17:18:38.033607 systemd[1]: Condition check resulted in dev-ttyAMA0.device - /dev/ttyAMA0 being skipped. Sep 4 17:18:38.034620 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 4 17:18:38.034700 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Sep 4 17:18:38.034944 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 35 scanned by (udev-worker) (1346) Sep 4 17:18:38.037059 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Sep 4 17:18:38.072529 systemd-resolved[1309]: Positive Trust Anchors: Sep 4 17:18:38.072549 systemd-resolved[1309]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 4 17:18:38.072581 systemd-resolved[1309]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa corp home internal intranet lan local private test Sep 4 17:18:38.090169 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Sep 4 17:18:38.091604 systemd-resolved[1309]: Defaulting to hostname 'linux'. Sep 4 17:18:38.103115 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Sep 4 17:18:38.106090 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Sep 4 17:18:38.107307 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Sep 4 17:18:38.110093 systemd-networkd[1377]: lo: Link UP Sep 4 17:18:38.110100 systemd-networkd[1377]: lo: Gained carrier Sep 4 17:18:38.110970 systemd-networkd[1377]: Enumeration completed Sep 4 17:18:38.113155 systemd[1]: Started systemd-networkd.service - Network Configuration. Sep 4 17:18:38.113566 systemd-networkd[1377]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 4 17:18:38.113575 systemd-networkd[1377]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 4 17:18:38.114311 systemd-networkd[1377]: eth0: Link UP Sep 4 17:18:38.114320 systemd-networkd[1377]: eth0: Gained carrier Sep 4 17:18:38.114333 systemd-networkd[1377]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Sep 4 17:18:38.119688 systemd[1]: Reached target network.target - Network. Sep 4 17:18:38.128123 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Sep 4 17:18:38.132192 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Sep 4 17:18:38.133602 systemd[1]: Reached target time-set.target - System Time Set. Sep 4 17:18:38.138955 systemd-networkd[1377]: eth0: DHCPv4 address 10.0.0.60/16, gateway 10.0.0.1 acquired from 10.0.0.1 Sep 4 17:18:38.139641 systemd-timesyncd[1390]: Network configuration changed, trying to establish connection. Sep 4 17:18:38.140452 systemd-timesyncd[1390]: Contacted time server 10.0.0.1:123 (10.0.0.1). Sep 4 17:18:38.140499 systemd-timesyncd[1390]: Initial clock synchronization to Wed 2024-09-04 17:18:38.224606 UTC. Sep 4 17:18:38.146153 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Sep 4 17:18:38.148933 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Sep 4 17:18:38.161204 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Sep 4 17:18:38.177086 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Sep 4 17:18:38.191852 lvm[1404]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Sep 4 17:18:38.207883 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Sep 4 17:18:38.221467 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Sep 4 17:18:38.223000 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Sep 4 17:18:38.223808 systemd[1]: Reached target sysinit.target - System Initialization. Sep 4 17:18:38.224670 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Sep 4 17:18:38.225905 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Sep 4 17:18:38.226970 systemd[1]: Started logrotate.timer - Daily rotation of log files. Sep 4 17:18:38.227893 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Sep 4 17:18:38.228775 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Sep 4 17:18:38.230020 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Sep 4 17:18:38.230057 systemd[1]: Reached target paths.target - Path Units. Sep 4 17:18:38.230697 systemd[1]: Reached target timers.target - Timer Units. Sep 4 17:18:38.232658 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Sep 4 17:18:38.235133 systemd[1]: Starting docker.socket - Docker Socket for the API... Sep 4 17:18:38.247090 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Sep 4 17:18:38.249456 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Sep 4 17:18:38.251056 systemd[1]: Listening on docker.socket - Docker Socket for the API. Sep 4 17:18:38.251980 systemd[1]: Reached target sockets.target - Socket Units. Sep 4 17:18:38.252908 systemd[1]: Reached target basic.target - Basic System. Sep 4 17:18:38.253876 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Sep 4 17:18:38.253912 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Sep 4 17:18:38.254901 systemd[1]: Starting containerd.service - containerd container runtime... Sep 4 17:18:38.256958 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Sep 4 17:18:38.259986 lvm[1411]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Sep 4 17:18:38.260994 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Sep 4 17:18:38.262967 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Sep 4 17:18:38.264025 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Sep 4 17:18:38.265100 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Sep 4 17:18:38.271554 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Sep 4 17:18:38.276097 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Sep 4 17:18:38.280520 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Sep 4 17:18:38.282767 jq[1414]: false Sep 4 17:18:38.293077 systemd[1]: Starting systemd-logind.service - User Login Management... Sep 4 17:18:38.301648 extend-filesystems[1415]: Found loop3 Sep 4 17:18:38.301648 extend-filesystems[1415]: Found loop4 Sep 4 17:18:38.301648 extend-filesystems[1415]: Found loop5 Sep 4 17:18:38.301648 extend-filesystems[1415]: Found vda Sep 4 17:18:38.301648 extend-filesystems[1415]: Found vda1 Sep 4 17:18:38.301648 extend-filesystems[1415]: Found vda2 Sep 4 17:18:38.301648 extend-filesystems[1415]: Found vda3 Sep 4 17:18:38.301648 extend-filesystems[1415]: Found usr Sep 4 17:18:38.301648 extend-filesystems[1415]: Found vda4 Sep 4 17:18:38.301648 extend-filesystems[1415]: Found vda6 Sep 4 17:18:38.301648 extend-filesystems[1415]: Found vda7 Sep 4 17:18:38.301648 extend-filesystems[1415]: Found vda9 Sep 4 17:18:38.301648 extend-filesystems[1415]: Checking size of /dev/vda9 Sep 4 17:18:38.301056 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Sep 4 17:18:38.301557 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Sep 4 17:18:38.302222 systemd[1]: Starting update-engine.service - Update Engine... Sep 4 17:18:38.335264 jq[1431]: true Sep 4 17:18:38.307026 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Sep 4 17:18:38.335439 dbus-daemon[1413]: [system] SELinux support is enabled Sep 4 17:18:38.313903 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Sep 4 17:18:38.318643 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Sep 4 17:18:38.318802 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Sep 4 17:18:38.319070 systemd[1]: motdgen.service: Deactivated successfully. Sep 4 17:18:38.319201 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Sep 4 17:18:38.321568 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Sep 4 17:18:38.321704 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Sep 4 17:18:38.335960 systemd[1]: Started dbus.service - D-Bus System Message Bus. Sep 4 17:18:38.340219 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Sep 4 17:18:38.340242 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Sep 4 17:18:38.342592 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Sep 4 17:18:38.352537 tar[1435]: linux-arm64/helm Sep 4 17:18:38.342617 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Sep 4 17:18:38.361671 extend-filesystems[1415]: Resized partition /dev/vda9 Sep 4 17:18:38.363202 (ntainerd)[1448]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Sep 4 17:18:38.366898 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 35 scanned by (udev-worker) (1365) Sep 4 17:18:38.371649 extend-filesystems[1450]: resize2fs 1.47.0 (5-Feb-2023) Sep 4 17:18:38.372804 jq[1436]: true Sep 4 17:18:38.372969 update_engine[1430]: I0904 17:18:38.372028 1430 main.cc:92] Flatcar Update Engine starting Sep 4 17:18:38.376254 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Sep 4 17:18:38.376329 update_engine[1430]: I0904 17:18:38.375442 1430 update_check_scheduler.cc:74] Next update check in 5m51s Sep 4 17:18:38.376886 systemd[1]: Started update-engine.service - Update Engine. Sep 4 17:18:38.386622 systemd-logind[1423]: Watching system buttons on /dev/input/event0 (Power Button) Sep 4 17:18:38.386785 systemd[1]: Started locksmithd.service - Cluster reboot manager. Sep 4 17:18:38.390188 systemd-logind[1423]: New seat seat0. Sep 4 17:18:38.395394 systemd[1]: Started systemd-logind.service - User Login Management. Sep 4 17:18:38.422839 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Sep 4 17:18:38.447456 extend-filesystems[1450]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Sep 4 17:18:38.447456 extend-filesystems[1450]: old_desc_blocks = 1, new_desc_blocks = 1 Sep 4 17:18:38.447456 extend-filesystems[1450]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Sep 4 17:18:38.458532 extend-filesystems[1415]: Resized filesystem in /dev/vda9 Sep 4 17:18:38.450227 systemd[1]: extend-filesystems.service: Deactivated successfully. Sep 4 17:18:38.450412 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Sep 4 17:18:38.459813 locksmithd[1453]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Sep 4 17:18:38.460299 bash[1467]: Updated "/home/core/.ssh/authorized_keys" Sep 4 17:18:38.462912 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Sep 4 17:18:38.464744 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Sep 4 17:18:38.580110 containerd[1448]: time="2024-09-04T17:18:38.579963080Z" level=info msg="starting containerd" revision=1fbfc07f8d28210e62bdbcbf7b950bac8028afbf version=v1.7.17 Sep 4 17:18:38.611903 containerd[1448]: time="2024-09-04T17:18:38.611855080Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Sep 4 17:18:38.611903 containerd[1448]: time="2024-09-04T17:18:38.611901640Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Sep 4 17:18:38.614235 containerd[1448]: time="2024-09-04T17:18:38.613273000Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.48-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Sep 4 17:18:38.614235 containerd[1448]: time="2024-09-04T17:18:38.613315120Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Sep 4 17:18:38.614235 containerd[1448]: time="2024-09-04T17:18:38.613568800Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Sep 4 17:18:38.614235 containerd[1448]: time="2024-09-04T17:18:38.613587320Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Sep 4 17:18:38.614235 containerd[1448]: time="2024-09-04T17:18:38.613662160Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Sep 4 17:18:38.614235 containerd[1448]: time="2024-09-04T17:18:38.613706640Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Sep 4 17:18:38.614235 containerd[1448]: time="2024-09-04T17:18:38.613718000Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Sep 4 17:18:38.614235 containerd[1448]: time="2024-09-04T17:18:38.613775240Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Sep 4 17:18:38.614235 containerd[1448]: time="2024-09-04T17:18:38.613989600Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Sep 4 17:18:38.614235 containerd[1448]: time="2024-09-04T17:18:38.614007800Z" level=warning msg="failed to load plugin io.containerd.snapshotter.v1.devmapper" error="devmapper not configured" Sep 4 17:18:38.614235 containerd[1448]: time="2024-09-04T17:18:38.614017200Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Sep 4 17:18:38.614488 containerd[1448]: time="2024-09-04T17:18:38.614105360Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Sep 4 17:18:38.614488 containerd[1448]: time="2024-09-04T17:18:38.614117520Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Sep 4 17:18:38.614488 containerd[1448]: time="2024-09-04T17:18:38.614170000Z" level=warning msg="could not use snapshotter devmapper in metadata plugin" error="devmapper not configured" Sep 4 17:18:38.614488 containerd[1448]: time="2024-09-04T17:18:38.614181680Z" level=info msg="metadata content store policy set" policy=shared Sep 4 17:18:38.619403 containerd[1448]: time="2024-09-04T17:18:38.619371600Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Sep 4 17:18:38.619568 containerd[1448]: time="2024-09-04T17:18:38.619550960Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Sep 4 17:18:38.619691 containerd[1448]: time="2024-09-04T17:18:38.619673880Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Sep 4 17:18:38.619781 containerd[1448]: time="2024-09-04T17:18:38.619765960Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Sep 4 17:18:38.619921 containerd[1448]: time="2024-09-04T17:18:38.619904800Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Sep 4 17:18:38.620044 containerd[1448]: time="2024-09-04T17:18:38.620029120Z" level=info msg="NRI interface is disabled by configuration." Sep 4 17:18:38.620112 containerd[1448]: time="2024-09-04T17:18:38.620098560Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Sep 4 17:18:38.620394 containerd[1448]: time="2024-09-04T17:18:38.620326600Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Sep 4 17:18:38.620467 containerd[1448]: time="2024-09-04T17:18:38.620454280Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Sep 4 17:18:38.620569 containerd[1448]: time="2024-09-04T17:18:38.620506840Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Sep 4 17:18:38.620628 containerd[1448]: time="2024-09-04T17:18:38.620616200Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Sep 4 17:18:38.620756 containerd[1448]: time="2024-09-04T17:18:38.620739560Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Sep 4 17:18:38.620901 containerd[1448]: time="2024-09-04T17:18:38.620817880Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Sep 4 17:18:38.620968 containerd[1448]: time="2024-09-04T17:18:38.620954520Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Sep 4 17:18:38.621085 containerd[1448]: time="2024-09-04T17:18:38.621068080Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Sep 4 17:18:38.621231 containerd[1448]: time="2024-09-04T17:18:38.621132280Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Sep 4 17:18:38.621231 containerd[1448]: time="2024-09-04T17:18:38.621152640Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Sep 4 17:18:38.621231 containerd[1448]: time="2024-09-04T17:18:38.621165880Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Sep 4 17:18:38.621231 containerd[1448]: time="2024-09-04T17:18:38.621178160Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Sep 4 17:18:38.621618 containerd[1448]: time="2024-09-04T17:18:38.621504120Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Sep 4 17:18:38.622282 containerd[1448]: time="2024-09-04T17:18:38.622073120Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Sep 4 17:18:38.622282 containerd[1448]: time="2024-09-04T17:18:38.622117240Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Sep 4 17:18:38.622282 containerd[1448]: time="2024-09-04T17:18:38.622131760Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Sep 4 17:18:38.622282 containerd[1448]: time="2024-09-04T17:18:38.622156360Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Sep 4 17:18:38.622561 containerd[1448]: time="2024-09-04T17:18:38.622539480Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Sep 4 17:18:38.622775 containerd[1448]: time="2024-09-04T17:18:38.622755800Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Sep 4 17:18:38.622878 containerd[1448]: time="2024-09-04T17:18:38.622863040Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Sep 4 17:18:38.623897 containerd[1448]: time="2024-09-04T17:18:38.622973400Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Sep 4 17:18:38.623897 containerd[1448]: time="2024-09-04T17:18:38.622997160Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Sep 4 17:18:38.623897 containerd[1448]: time="2024-09-04T17:18:38.623017360Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Sep 4 17:18:38.623897 containerd[1448]: time="2024-09-04T17:18:38.623029360Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Sep 4 17:18:38.623897 containerd[1448]: time="2024-09-04T17:18:38.623050840Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Sep 4 17:18:38.623897 containerd[1448]: time="2024-09-04T17:18:38.623064240Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Sep 4 17:18:38.623897 containerd[1448]: time="2024-09-04T17:18:38.623201240Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Sep 4 17:18:38.623897 containerd[1448]: time="2024-09-04T17:18:38.623218680Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Sep 4 17:18:38.623897 containerd[1448]: time="2024-09-04T17:18:38.623230400Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Sep 4 17:18:38.623897 containerd[1448]: time="2024-09-04T17:18:38.623246640Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Sep 4 17:18:38.623897 containerd[1448]: time="2024-09-04T17:18:38.623259880Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Sep 4 17:18:38.623897 containerd[1448]: time="2024-09-04T17:18:38.623288760Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Sep 4 17:18:38.623897 containerd[1448]: time="2024-09-04T17:18:38.623302880Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Sep 4 17:18:38.623897 containerd[1448]: time="2024-09-04T17:18:38.623314040Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Sep 4 17:18:38.624170 containerd[1448]: time="2024-09-04T17:18:38.623694680Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Sep 4 17:18:38.624170 containerd[1448]: time="2024-09-04T17:18:38.623753480Z" level=info msg="Connect containerd service" Sep 4 17:18:38.624170 containerd[1448]: time="2024-09-04T17:18:38.623785080Z" level=info msg="using legacy CRI server" Sep 4 17:18:38.624170 containerd[1448]: time="2024-09-04T17:18:38.623792600Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Sep 4 17:18:38.624579 containerd[1448]: time="2024-09-04T17:18:38.624550880Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Sep 4 17:18:38.625790 containerd[1448]: time="2024-09-04T17:18:38.625758800Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 4 17:18:38.626076 containerd[1448]: time="2024-09-04T17:18:38.625965160Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Sep 4 17:18:38.626076 containerd[1448]: time="2024-09-04T17:18:38.625993440Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Sep 4 17:18:38.626076 containerd[1448]: time="2024-09-04T17:18:38.626004000Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Sep 4 17:18:38.626076 containerd[1448]: time="2024-09-04T17:18:38.626017920Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Sep 4 17:18:38.627361 containerd[1448]: time="2024-09-04T17:18:38.626681360Z" level=info msg="Start subscribing containerd event" Sep 4 17:18:38.627361 containerd[1448]: time="2024-09-04T17:18:38.627137680Z" level=info msg="Start recovering state" Sep 4 17:18:38.627361 containerd[1448]: time="2024-09-04T17:18:38.627201640Z" level=info msg="Start event monitor" Sep 4 17:18:38.627361 containerd[1448]: time="2024-09-04T17:18:38.627212920Z" level=info msg="Start snapshots syncer" Sep 4 17:18:38.627361 containerd[1448]: time="2024-09-04T17:18:38.627221440Z" level=info msg="Start cni network conf syncer for default" Sep 4 17:18:38.627361 containerd[1448]: time="2024-09-04T17:18:38.627229560Z" level=info msg="Start streaming server" Sep 4 17:18:38.627860 containerd[1448]: time="2024-09-04T17:18:38.627825680Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Sep 4 17:18:38.628035 containerd[1448]: time="2024-09-04T17:18:38.628019200Z" level=info msg=serving... address=/run/containerd/containerd.sock Sep 4 17:18:38.628299 systemd[1]: Started containerd.service - containerd container runtime. Sep 4 17:18:38.629739 containerd[1448]: time="2024-09-04T17:18:38.629708320Z" level=info msg="containerd successfully booted in 0.050905s" Sep 4 17:18:38.750247 tar[1435]: linux-arm64/LICENSE Sep 4 17:18:38.750247 tar[1435]: linux-arm64/README.md Sep 4 17:18:38.761256 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Sep 4 17:18:39.464260 sshd_keygen[1432]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Sep 4 17:18:39.484921 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Sep 4 17:18:39.491116 systemd[1]: Starting issuegen.service - Generate /run/issue... Sep 4 17:18:39.498742 systemd[1]: issuegen.service: Deactivated successfully. Sep 4 17:18:39.498966 systemd[1]: Finished issuegen.service - Generate /run/issue. Sep 4 17:18:39.502306 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Sep 4 17:18:39.514624 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Sep 4 17:18:39.518003 systemd[1]: Started getty@tty1.service - Getty on tty1. Sep 4 17:18:39.520626 systemd[1]: Started serial-getty@ttyAMA0.service - Serial Getty on ttyAMA0. Sep 4 17:18:39.521917 systemd[1]: Reached target getty.target - Login Prompts. Sep 4 17:18:39.863337 systemd-networkd[1377]: eth0: Gained IPv6LL Sep 4 17:18:39.870597 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Sep 4 17:18:39.872275 systemd[1]: Reached target network-online.target - Network is Online. Sep 4 17:18:39.892134 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Sep 4 17:18:39.894327 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 4 17:18:39.896290 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Sep 4 17:18:39.912169 systemd[1]: coreos-metadata.service: Deactivated successfully. Sep 4 17:18:39.912357 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Sep 4 17:18:39.913904 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Sep 4 17:18:39.917198 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Sep 4 17:18:40.426525 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 4 17:18:40.429210 systemd[1]: Reached target multi-user.target - Multi-User System. Sep 4 17:18:40.430979 (kubelet)[1523]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 4 17:18:40.433926 systemd[1]: Startup finished in 563ms (kernel) + 4.777s (initrd) + 3.966s (userspace) = 9.307s. Sep 4 17:18:41.109856 kubelet[1523]: E0904 17:18:41.109751 1523 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 4 17:18:41.113078 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 4 17:18:41.113236 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 4 17:18:44.851561 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Sep 4 17:18:44.852669 systemd[1]: Started sshd@0-10.0.0.60:22-10.0.0.1:51042.service - OpenSSH per-connection server daemon (10.0.0.1:51042). Sep 4 17:18:44.924377 sshd[1539]: Accepted publickey for core from 10.0.0.1 port 51042 ssh2: RSA SHA256:TcdII3DD+/vh6fGiZDuqtLwdsO9LHnvXRMQO7IdpdiA Sep 4 17:18:44.926280 sshd[1539]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 4 17:18:44.939540 systemd-logind[1423]: New session 1 of user core. Sep 4 17:18:44.940927 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Sep 4 17:18:44.951207 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Sep 4 17:18:44.962556 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Sep 4 17:18:44.965004 systemd[1]: Starting user@500.service - User Manager for UID 500... Sep 4 17:18:44.974115 (systemd)[1543]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Sep 4 17:18:45.062874 systemd[1543]: Queued start job for default target default.target. Sep 4 17:18:45.073798 systemd[1543]: Created slice app.slice - User Application Slice. Sep 4 17:18:45.073828 systemd[1543]: Reached target paths.target - Paths. Sep 4 17:18:45.073859 systemd[1543]: Reached target timers.target - Timers. Sep 4 17:18:45.075098 systemd[1543]: Starting dbus.socket - D-Bus User Message Bus Socket... Sep 4 17:18:45.085347 systemd[1543]: Listening on dbus.socket - D-Bus User Message Bus Socket. Sep 4 17:18:45.085503 systemd[1543]: Reached target sockets.target - Sockets. Sep 4 17:18:45.085525 systemd[1543]: Reached target basic.target - Basic System. Sep 4 17:18:45.085571 systemd[1543]: Reached target default.target - Main User Target. Sep 4 17:18:45.085599 systemd[1543]: Startup finished in 105ms. Sep 4 17:18:45.085781 systemd[1]: Started user@500.service - User Manager for UID 500. Sep 4 17:18:45.087229 systemd[1]: Started session-1.scope - Session 1 of User core. Sep 4 17:18:45.152410 systemd[1]: Started sshd@1-10.0.0.60:22-10.0.0.1:51046.service - OpenSSH per-connection server daemon (10.0.0.1:51046). Sep 4 17:18:45.191777 sshd[1554]: Accepted publickey for core from 10.0.0.1 port 51046 ssh2: RSA SHA256:TcdII3DD+/vh6fGiZDuqtLwdsO9LHnvXRMQO7IdpdiA Sep 4 17:18:45.193429 sshd[1554]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 4 17:18:45.198225 systemd-logind[1423]: New session 2 of user core. Sep 4 17:18:45.210053 systemd[1]: Started session-2.scope - Session 2 of User core. Sep 4 17:18:45.264145 sshd[1554]: pam_unix(sshd:session): session closed for user core Sep 4 17:18:45.281545 systemd[1]: sshd@1-10.0.0.60:22-10.0.0.1:51046.service: Deactivated successfully. Sep 4 17:18:45.283093 systemd[1]: session-2.scope: Deactivated successfully. Sep 4 17:18:45.284493 systemd-logind[1423]: Session 2 logged out. Waiting for processes to exit. Sep 4 17:18:45.285731 systemd[1]: Started sshd@2-10.0.0.60:22-10.0.0.1:51058.service - OpenSSH per-connection server daemon (10.0.0.1:51058). Sep 4 17:18:45.287241 systemd-logind[1423]: Removed session 2. Sep 4 17:18:45.327942 sshd[1561]: Accepted publickey for core from 10.0.0.1 port 51058 ssh2: RSA SHA256:TcdII3DD+/vh6fGiZDuqtLwdsO9LHnvXRMQO7IdpdiA Sep 4 17:18:45.329360 sshd[1561]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 4 17:18:45.333335 systemd-logind[1423]: New session 3 of user core. Sep 4 17:18:45.345036 systemd[1]: Started session-3.scope - Session 3 of User core. Sep 4 17:18:45.398826 sshd[1561]: pam_unix(sshd:session): session closed for user core Sep 4 17:18:45.415312 systemd[1]: sshd@2-10.0.0.60:22-10.0.0.1:51058.service: Deactivated successfully. Sep 4 17:18:45.416864 systemd[1]: session-3.scope: Deactivated successfully. Sep 4 17:18:45.418217 systemd-logind[1423]: Session 3 logged out. Waiting for processes to exit. Sep 4 17:18:45.431200 systemd[1]: Started sshd@3-10.0.0.60:22-10.0.0.1:51062.service - OpenSSH per-connection server daemon (10.0.0.1:51062). Sep 4 17:18:45.439606 systemd-logind[1423]: Removed session 3. Sep 4 17:18:45.469856 sshd[1568]: Accepted publickey for core from 10.0.0.1 port 51062 ssh2: RSA SHA256:TcdII3DD+/vh6fGiZDuqtLwdsO9LHnvXRMQO7IdpdiA Sep 4 17:18:45.471120 sshd[1568]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 4 17:18:45.475075 systemd-logind[1423]: New session 4 of user core. Sep 4 17:18:45.485002 systemd[1]: Started session-4.scope - Session 4 of User core. Sep 4 17:18:45.539894 sshd[1568]: pam_unix(sshd:session): session closed for user core Sep 4 17:18:45.557416 systemd[1]: sshd@3-10.0.0.60:22-10.0.0.1:51062.service: Deactivated successfully. Sep 4 17:18:45.558968 systemd[1]: session-4.scope: Deactivated successfully. Sep 4 17:18:45.561900 systemd-logind[1423]: Session 4 logged out. Waiting for processes to exit. Sep 4 17:18:45.571526 systemd[1]: Started sshd@4-10.0.0.60:22-10.0.0.1:51066.service - OpenSSH per-connection server daemon (10.0.0.1:51066). Sep 4 17:18:45.572845 systemd-logind[1423]: Removed session 4. Sep 4 17:18:45.609314 sshd[1575]: Accepted publickey for core from 10.0.0.1 port 51066 ssh2: RSA SHA256:TcdII3DD+/vh6fGiZDuqtLwdsO9LHnvXRMQO7IdpdiA Sep 4 17:18:45.610645 sshd[1575]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 4 17:18:45.615798 systemd-logind[1423]: New session 5 of user core. Sep 4 17:18:45.631055 systemd[1]: Started session-5.scope - Session 5 of User core. Sep 4 17:18:45.714940 sudo[1578]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Sep 4 17:18:45.715204 sudo[1578]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) Sep 4 17:18:45.730811 sudo[1578]: pam_unix(sudo:session): session closed for user root Sep 4 17:18:45.733142 sshd[1575]: pam_unix(sshd:session): session closed for user core Sep 4 17:18:45.745141 systemd[1]: sshd@4-10.0.0.60:22-10.0.0.1:51066.service: Deactivated successfully. Sep 4 17:18:45.747256 systemd[1]: session-5.scope: Deactivated successfully. Sep 4 17:18:45.750123 systemd-logind[1423]: Session 5 logged out. Waiting for processes to exit. Sep 4 17:18:45.751779 systemd[1]: Started sshd@5-10.0.0.60:22-10.0.0.1:51068.service - OpenSSH per-connection server daemon (10.0.0.1:51068). Sep 4 17:18:45.752562 systemd-logind[1423]: Removed session 5. Sep 4 17:18:45.794287 sshd[1583]: Accepted publickey for core from 10.0.0.1 port 51068 ssh2: RSA SHA256:TcdII3DD+/vh6fGiZDuqtLwdsO9LHnvXRMQO7IdpdiA Sep 4 17:18:45.795894 sshd[1583]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 4 17:18:45.800446 systemd-logind[1423]: New session 6 of user core. Sep 4 17:18:45.812057 systemd[1]: Started session-6.scope - Session 6 of User core. Sep 4 17:18:45.868782 sudo[1587]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Sep 4 17:18:45.869062 sudo[1587]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) Sep 4 17:18:45.874142 sudo[1587]: pam_unix(sudo:session): session closed for user root Sep 4 17:18:45.879011 sudo[1586]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Sep 4 17:18:45.879257 sudo[1586]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) Sep 4 17:18:45.898488 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Sep 4 17:18:45.901908 auditctl[1590]: No rules Sep 4 17:18:45.904172 systemd[1]: audit-rules.service: Deactivated successfully. Sep 4 17:18:45.904385 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Sep 4 17:18:45.909992 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Sep 4 17:18:45.939479 augenrules[1608]: No rules Sep 4 17:18:45.940517 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Sep 4 17:18:45.942207 sudo[1586]: pam_unix(sudo:session): session closed for user root Sep 4 17:18:45.943995 sshd[1583]: pam_unix(sshd:session): session closed for user core Sep 4 17:18:45.953421 systemd[1]: sshd@5-10.0.0.60:22-10.0.0.1:51068.service: Deactivated successfully. Sep 4 17:18:45.955168 systemd[1]: session-6.scope: Deactivated successfully. Sep 4 17:18:45.957022 systemd-logind[1423]: Session 6 logged out. Waiting for processes to exit. Sep 4 17:18:45.959127 systemd[1]: Started sshd@6-10.0.0.60:22-10.0.0.1:51078.service - OpenSSH per-connection server daemon (10.0.0.1:51078). Sep 4 17:18:45.960828 systemd-logind[1423]: Removed session 6. Sep 4 17:18:46.005885 sshd[1616]: Accepted publickey for core from 10.0.0.1 port 51078 ssh2: RSA SHA256:TcdII3DD+/vh6fGiZDuqtLwdsO9LHnvXRMQO7IdpdiA Sep 4 17:18:46.006114 sshd[1616]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 4 17:18:46.010770 systemd-logind[1423]: New session 7 of user core. Sep 4 17:18:46.019992 systemd[1]: Started session-7.scope - Session 7 of User core. Sep 4 17:18:46.074502 sudo[1619]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Sep 4 17:18:46.075316 sudo[1619]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) Sep 4 17:18:46.203119 systemd[1]: Starting docker.service - Docker Application Container Engine... Sep 4 17:18:46.203209 (dockerd)[1630]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Sep 4 17:18:46.441473 dockerd[1630]: time="2024-09-04T17:18:46.440901236Z" level=info msg="Starting up" Sep 4 17:18:46.524227 dockerd[1630]: time="2024-09-04T17:18:46.524126165Z" level=info msg="Loading containers: start." Sep 4 17:18:46.630858 kernel: Initializing XFRM netlink socket Sep 4 17:18:46.702363 systemd-networkd[1377]: docker0: Link UP Sep 4 17:18:46.728636 dockerd[1630]: time="2024-09-04T17:18:46.728534838Z" level=info msg="Loading containers: done." Sep 4 17:18:46.783680 dockerd[1630]: time="2024-09-04T17:18:46.783616818Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Sep 4 17:18:46.784011 dockerd[1630]: time="2024-09-04T17:18:46.783879540Z" level=info msg="Docker daemon" commit=fca702de7f71362c8d103073c7e4a1d0a467fadd graphdriver=overlay2 version=24.0.9 Sep 4 17:18:46.784071 dockerd[1630]: time="2024-09-04T17:18:46.784035703Z" level=info msg="Daemon has completed initialization" Sep 4 17:18:46.809550 dockerd[1630]: time="2024-09-04T17:18:46.809492268Z" level=info msg="API listen on /run/docker.sock" Sep 4 17:18:46.809732 systemd[1]: Started docker.service - Docker Application Container Engine. Sep 4 17:18:47.406028 containerd[1448]: time="2024-09-04T17:18:47.405983056Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.29.8\"" Sep 4 17:18:47.509051 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck2019594939-merged.mount: Deactivated successfully. Sep 4 17:18:48.496910 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1580160601.mount: Deactivated successfully. Sep 4 17:18:50.408235 containerd[1448]: time="2024-09-04T17:18:50.408116588Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.29.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:18:50.411365 containerd[1448]: time="2024-09-04T17:18:50.411004411Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.29.8: active requests=0, bytes read=32283564" Sep 4 17:18:50.412197 containerd[1448]: time="2024-09-04T17:18:50.412163735Z" level=info msg="ImageCreate event name:\"sha256:6b88c4d45de58e9ed0353538f5b2ae206a8582fcb53e67d0505abbe3a567fbae\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:18:50.415269 containerd[1448]: time="2024-09-04T17:18:50.415233840Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:6f72fa926c9b05e10629fe1a092fd28dcd65b4fdfd0cc7bd55f85a57a6ba1fa5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:18:50.416872 containerd[1448]: time="2024-09-04T17:18:50.416661893Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.29.8\" with image id \"sha256:6b88c4d45de58e9ed0353538f5b2ae206a8582fcb53e67d0505abbe3a567fbae\", repo tag \"registry.k8s.io/kube-apiserver:v1.29.8\", repo digest \"registry.k8s.io/kube-apiserver@sha256:6f72fa926c9b05e10629fe1a092fd28dcd65b4fdfd0cc7bd55f85a57a6ba1fa5\", size \"32280362\" in 3.010629231s" Sep 4 17:18:50.416872 containerd[1448]: time="2024-09-04T17:18:50.416708927Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.29.8\" returns image reference \"sha256:6b88c4d45de58e9ed0353538f5b2ae206a8582fcb53e67d0505abbe3a567fbae\"" Sep 4 17:18:50.436229 containerd[1448]: time="2024-09-04T17:18:50.436182902Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.29.8\"" Sep 4 17:18:51.366569 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Sep 4 17:18:51.377091 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 4 17:18:51.471701 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 4 17:18:51.475518 (kubelet)[1843]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 4 17:18:51.541452 kubelet[1843]: E0904 17:18:51.541363 1843 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 4 17:18:51.544971 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 4 17:18:51.545117 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 4 17:18:52.738778 containerd[1448]: time="2024-09-04T17:18:52.738727449Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.29.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:18:52.739912 containerd[1448]: time="2024-09-04T17:18:52.739883752Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.29.8: active requests=0, bytes read=29368212" Sep 4 17:18:52.743221 containerd[1448]: time="2024-09-04T17:18:52.743179497Z" level=info msg="ImageCreate event name:\"sha256:bddc5fa0c49f499b7ec60c114671fcbb0436c22300448964f77acb6c13f0ffed\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:18:52.746428 containerd[1448]: time="2024-09-04T17:18:52.746390887Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:6f27d63ded20614c68554b477cd7a78eda78a498a92bfe8935cf964ca5b74d0b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:18:52.747942 containerd[1448]: time="2024-09-04T17:18:52.747895995Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.29.8\" with image id \"sha256:bddc5fa0c49f499b7ec60c114671fcbb0436c22300448964f77acb6c13f0ffed\", repo tag \"registry.k8s.io/kube-controller-manager:v1.29.8\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:6f27d63ded20614c68554b477cd7a78eda78a498a92bfe8935cf964ca5b74d0b\", size \"30855477\" in 2.311669915s" Sep 4 17:18:52.747942 containerd[1448]: time="2024-09-04T17:18:52.747939476Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.29.8\" returns image reference \"sha256:bddc5fa0c49f499b7ec60c114671fcbb0436c22300448964f77acb6c13f0ffed\"" Sep 4 17:18:52.768895 containerd[1448]: time="2024-09-04T17:18:52.768854663Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.29.8\"" Sep 4 17:18:54.284149 containerd[1448]: time="2024-09-04T17:18:54.284102819Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.29.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:18:54.285187 containerd[1448]: time="2024-09-04T17:18:54.284589910Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.29.8: active requests=0, bytes read=15751075" Sep 4 17:18:54.285850 containerd[1448]: time="2024-09-04T17:18:54.285629505Z" level=info msg="ImageCreate event name:\"sha256:db329f69447ed4eb4b489d7c357c7723493b3a72946edb35a6c16973d5f257d4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:18:54.288636 containerd[1448]: time="2024-09-04T17:18:54.288599199Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:da74a66675d95e39ec25da5e70729da746d0fa0b15ee0da872ac980519bc28bd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:18:54.290020 containerd[1448]: time="2024-09-04T17:18:54.289894918Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.29.8\" with image id \"sha256:db329f69447ed4eb4b489d7c357c7723493b3a72946edb35a6c16973d5f257d4\", repo tag \"registry.k8s.io/kube-scheduler:v1.29.8\", repo digest \"registry.k8s.io/kube-scheduler@sha256:da74a66675d95e39ec25da5e70729da746d0fa0b15ee0da872ac980519bc28bd\", size \"17238358\" in 1.520993012s" Sep 4 17:18:54.290020 containerd[1448]: time="2024-09-04T17:18:54.289938900Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.29.8\" returns image reference \"sha256:db329f69447ed4eb4b489d7c357c7723493b3a72946edb35a6c16973d5f257d4\"" Sep 4 17:18:54.310488 containerd[1448]: time="2024-09-04T17:18:54.310445639Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.29.8\"" Sep 4 17:18:55.521811 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3798662956.mount: Deactivated successfully. Sep 4 17:18:55.731446 containerd[1448]: time="2024-09-04T17:18:55.731394034Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.29.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:18:55.732873 containerd[1448]: time="2024-09-04T17:18:55.732677788Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.29.8: active requests=0, bytes read=25251885" Sep 4 17:18:55.733522 containerd[1448]: time="2024-09-04T17:18:55.733484030Z" level=info msg="ImageCreate event name:\"sha256:61223b17dfa4bd3d116a0b714c4f2cc2e3d83853942dfb8578f50cc8e91eb399\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:18:55.735429 containerd[1448]: time="2024-09-04T17:18:55.735365406Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:559a093080f70ca863922f5e4bb90d6926d52653a91edb5b72c685ebb65f1858\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:18:55.736117 containerd[1448]: time="2024-09-04T17:18:55.736064674Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.29.8\" with image id \"sha256:61223b17dfa4bd3d116a0b714c4f2cc2e3d83853942dfb8578f50cc8e91eb399\", repo tag \"registry.k8s.io/kube-proxy:v1.29.8\", repo digest \"registry.k8s.io/kube-proxy@sha256:559a093080f70ca863922f5e4bb90d6926d52653a91edb5b72c685ebb65f1858\", size \"25250902\" in 1.425574334s" Sep 4 17:18:55.736173 containerd[1448]: time="2024-09-04T17:18:55.736129195Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.29.8\" returns image reference \"sha256:61223b17dfa4bd3d116a0b714c4f2cc2e3d83853942dfb8578f50cc8e91eb399\"" Sep 4 17:18:55.755308 containerd[1448]: time="2024-09-04T17:18:55.755270726Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\"" Sep 4 17:18:56.412472 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4172976414.mount: Deactivated successfully. Sep 4 17:18:57.417464 containerd[1448]: time="2024-09-04T17:18:57.417412728Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:18:57.418551 containerd[1448]: time="2024-09-04T17:18:57.418510572Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.1: active requests=0, bytes read=16485383" Sep 4 17:18:57.419271 containerd[1448]: time="2024-09-04T17:18:57.419219807Z" level=info msg="ImageCreate event name:\"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:18:57.422423 containerd[1448]: time="2024-09-04T17:18:57.422385778Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:18:57.423926 containerd[1448]: time="2024-09-04T17:18:57.423885404Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.1\" with image id \"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\", size \"16482581\" in 1.668575193s" Sep 4 17:18:57.423958 containerd[1448]: time="2024-09-04T17:18:57.423928685Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\" returns image reference \"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\"" Sep 4 17:18:57.444066 containerd[1448]: time="2024-09-04T17:18:57.444011386Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\"" Sep 4 17:18:57.950487 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2142096931.mount: Deactivated successfully. Sep 4 17:18:57.954755 containerd[1448]: time="2024-09-04T17:18:57.954706255Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:18:57.955511 containerd[1448]: time="2024-09-04T17:18:57.955395431Z" level=info msg="stop pulling image registry.k8s.io/pause:3.9: active requests=0, bytes read=268823" Sep 4 17:18:57.956185 containerd[1448]: time="2024-09-04T17:18:57.956147546Z" level=info msg="ImageCreate event name:\"sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:18:57.959244 containerd[1448]: time="2024-09-04T17:18:57.959195245Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:18:57.959870 containerd[1448]: time="2024-09-04T17:18:57.959763585Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.9\" with image id \"sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e\", repo tag \"registry.k8s.io/pause:3.9\", repo digest \"registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097\", size \"268051\" in 515.71112ms" Sep 4 17:18:57.959870 containerd[1448]: time="2024-09-04T17:18:57.959793053Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\" returns image reference \"sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e\"" Sep 4 17:18:57.981190 containerd[1448]: time="2024-09-04T17:18:57.981153609Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.10-0\"" Sep 4 17:18:58.706605 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1432394134.mount: Deactivated successfully. Sep 4 17:19:00.820536 containerd[1448]: time="2024-09-04T17:19:00.820449899Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.10-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:19:00.822338 containerd[1448]: time="2024-09-04T17:19:00.822292873Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.10-0: active requests=0, bytes read=65200788" Sep 4 17:19:00.823382 containerd[1448]: time="2024-09-04T17:19:00.823348906Z" level=info msg="ImageCreate event name:\"sha256:79f8d13ae8b8839cadfb2f83416935f5184206d386028e2d1263577f0ab3620b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:19:00.827102 containerd[1448]: time="2024-09-04T17:19:00.826472777Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:22f892d7672adc0b9c86df67792afdb8b2dc08880f49f669eaaa59c47d7908c2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:19:00.827800 containerd[1448]: time="2024-09-04T17:19:00.827752313Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.10-0\" with image id \"sha256:79f8d13ae8b8839cadfb2f83416935f5184206d386028e2d1263577f0ab3620b\", repo tag \"registry.k8s.io/etcd:3.5.10-0\", repo digest \"registry.k8s.io/etcd@sha256:22f892d7672adc0b9c86df67792afdb8b2dc08880f49f669eaaa59c47d7908c2\", size \"65198393\" in 2.846409007s" Sep 4 17:19:00.827800 containerd[1448]: time="2024-09-04T17:19:00.827791298Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.10-0\" returns image reference \"sha256:79f8d13ae8b8839cadfb2f83416935f5184206d386028e2d1263577f0ab3620b\"" Sep 4 17:19:01.636151 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Sep 4 17:19:01.646035 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 4 17:19:01.739295 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 4 17:19:01.743777 (kubelet)[2071]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Sep 4 17:19:01.781445 kubelet[2071]: E0904 17:19:01.781385 2071 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 4 17:19:01.784276 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 4 17:19:01.784417 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 4 17:19:07.094093 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 4 17:19:07.106097 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 4 17:19:07.121730 systemd[1]: Reloading requested from client PID 2086 ('systemctl') (unit session-7.scope)... Sep 4 17:19:07.121751 systemd[1]: Reloading... Sep 4 17:19:07.180865 zram_generator::config[2126]: No configuration found. Sep 4 17:19:07.266505 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 4 17:19:07.321088 systemd[1]: Reloading finished in 199 ms. Sep 4 17:19:07.360429 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Sep 4 17:19:07.360508 systemd[1]: kubelet.service: Failed with result 'signal'. Sep 4 17:19:07.360778 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 4 17:19:07.363431 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 4 17:19:07.465649 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 4 17:19:07.472113 (kubelet)[2169]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 4 17:19:07.517903 kubelet[2169]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 4 17:19:07.517903 kubelet[2169]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Sep 4 17:19:07.517903 kubelet[2169]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 4 17:19:07.518269 kubelet[2169]: I0904 17:19:07.517940 2169 server.go:204] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 4 17:19:08.272069 kubelet[2169]: I0904 17:19:08.271156 2169 server.go:487] "Kubelet version" kubeletVersion="v1.29.2" Sep 4 17:19:08.272069 kubelet[2169]: I0904 17:19:08.271193 2169 server.go:489] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 4 17:19:08.272069 kubelet[2169]: I0904 17:19:08.271517 2169 server.go:919] "Client rotation is on, will bootstrap in background" Sep 4 17:19:08.303190 kubelet[2169]: E0904 17:19:08.303156 2169 certificate_manager.go:562] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://10.0.0.60:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 10.0.0.60:6443: connect: connection refused Sep 4 17:19:08.303369 kubelet[2169]: I0904 17:19:08.303217 2169 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 4 17:19:08.313892 kubelet[2169]: I0904 17:19:08.313855 2169 server.go:745] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 4 17:19:08.314109 kubelet[2169]: I0904 17:19:08.314086 2169 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 4 17:19:08.314300 kubelet[2169]: I0904 17:19:08.314275 2169 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Sep 4 17:19:08.314300 kubelet[2169]: I0904 17:19:08.314298 2169 topology_manager.go:138] "Creating topology manager with none policy" Sep 4 17:19:08.314420 kubelet[2169]: I0904 17:19:08.314307 2169 container_manager_linux.go:301] "Creating device plugin manager" Sep 4 17:19:08.314862 kubelet[2169]: I0904 17:19:08.314827 2169 state_mem.go:36] "Initialized new in-memory state store" Sep 4 17:19:08.317289 kubelet[2169]: I0904 17:19:08.317268 2169 kubelet.go:396] "Attempting to sync node with API server" Sep 4 17:19:08.317332 kubelet[2169]: I0904 17:19:08.317297 2169 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 4 17:19:08.317332 kubelet[2169]: I0904 17:19:08.317319 2169 kubelet.go:312] "Adding apiserver pod source" Sep 4 17:19:08.317332 kubelet[2169]: I0904 17:19:08.317330 2169 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 4 17:19:08.317974 kubelet[2169]: W0904 17:19:08.317900 2169 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.Node: Get "https://10.0.0.60:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.60:6443: connect: connection refused Sep 4 17:19:08.317974 kubelet[2169]: E0904 17:19:08.317969 2169 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.0.0.60:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.60:6443: connect: connection refused Sep 4 17:19:08.319157 kubelet[2169]: W0904 17:19:08.319087 2169 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.Service: Get "https://10.0.0.60:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.60:6443: connect: connection refused Sep 4 17:19:08.319157 kubelet[2169]: E0904 17:19:08.319136 2169 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.0.0.60:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.60:6443: connect: connection refused Sep 4 17:19:08.327403 kubelet[2169]: I0904 17:19:08.327097 2169 kuberuntime_manager.go:258] "Container runtime initialized" containerRuntime="containerd" version="v1.7.17" apiVersion="v1" Sep 4 17:19:08.331505 kubelet[2169]: I0904 17:19:08.331484 2169 kubelet.go:809] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 4 17:19:08.337275 kubelet[2169]: W0904 17:19:08.337256 2169 probe.go:268] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Sep 4 17:19:08.342424 kubelet[2169]: I0904 17:19:08.342286 2169 server.go:1256] "Started kubelet" Sep 4 17:19:08.342424 kubelet[2169]: I0904 17:19:08.342366 2169 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 4 17:19:08.342570 kubelet[2169]: I0904 17:19:08.342367 2169 server.go:162] "Starting to listen" address="0.0.0.0" port=10250 Sep 4 17:19:08.345708 kubelet[2169]: I0904 17:19:08.345665 2169 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 4 17:19:08.347120 kubelet[2169]: I0904 17:19:08.347079 2169 server.go:233] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 4 17:19:08.348809 kubelet[2169]: I0904 17:19:08.348533 2169 volume_manager.go:291] "Starting Kubelet Volume Manager" Sep 4 17:19:08.348809 kubelet[2169]: I0904 17:19:08.348659 2169 desired_state_of_world_populator.go:151] "Desired state populator starts to run" Sep 4 17:19:08.353263 kubelet[2169]: I0904 17:19:08.353024 2169 reconciler_new.go:29] "Reconciler: start to sync state" Sep 4 17:19:08.362744 kubelet[2169]: E0904 17:19:08.362176 2169 event.go:355] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.60:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.60:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.17f21a23cf68da53 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2024-09-04 17:19:08.342254163 +0000 UTC m=+0.866510115,LastTimestamp:2024-09-04 17:19:08.342254163 +0000 UTC m=+0.866510115,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Sep 4 17:19:08.362744 kubelet[2169]: W0904 17:19:08.362283 2169 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.CSIDriver: Get "https://10.0.0.60:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.60:6443: connect: connection refused Sep 4 17:19:08.362744 kubelet[2169]: E0904 17:19:08.362323 2169 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.0.0.60:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.60:6443: connect: connection refused Sep 4 17:19:08.366241 kubelet[2169]: E0904 17:19:08.366080 2169 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.60:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.60:6443: connect: connection refused" interval="200ms" Sep 4 17:19:08.366675 kubelet[2169]: I0904 17:19:08.366641 2169 server.go:461] "Adding debug handlers to kubelet server" Sep 4 17:19:08.367606 kubelet[2169]: I0904 17:19:08.367507 2169 factory.go:221] Registration of the containerd container factory successfully Sep 4 17:19:08.367606 kubelet[2169]: I0904 17:19:08.367550 2169 factory.go:221] Registration of the systemd container factory successfully Sep 4 17:19:08.367752 kubelet[2169]: I0904 17:19:08.367677 2169 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 4 17:19:08.374175 kubelet[2169]: I0904 17:19:08.374141 2169 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 4 17:19:08.375203 kubelet[2169]: I0904 17:19:08.375179 2169 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 4 17:19:08.375203 kubelet[2169]: I0904 17:19:08.375201 2169 status_manager.go:217] "Starting to sync pod status with apiserver" Sep 4 17:19:08.375267 kubelet[2169]: I0904 17:19:08.375220 2169 kubelet.go:2329] "Starting kubelet main sync loop" Sep 4 17:19:08.375293 kubelet[2169]: E0904 17:19:08.375273 2169 kubelet.go:2353] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 4 17:19:08.375536 kubelet[2169]: E0904 17:19:08.375504 2169 kubelet.go:1462] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 4 17:19:08.381819 kubelet[2169]: W0904 17:19:08.381769 2169 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.RuntimeClass: Get "https://10.0.0.60:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.60:6443: connect: connection refused Sep 4 17:19:08.381819 kubelet[2169]: E0904 17:19:08.381818 2169 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.0.0.60:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.60:6443: connect: connection refused Sep 4 17:19:08.382669 kubelet[2169]: I0904 17:19:08.382641 2169 cpu_manager.go:214] "Starting CPU manager" policy="none" Sep 4 17:19:08.382669 kubelet[2169]: I0904 17:19:08.382661 2169 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Sep 4 17:19:08.382760 kubelet[2169]: I0904 17:19:08.382739 2169 state_mem.go:36] "Initialized new in-memory state store" Sep 4 17:19:08.450214 kubelet[2169]: I0904 17:19:08.450173 2169 kubelet_node_status.go:73] "Attempting to register node" node="localhost" Sep 4 17:19:08.450633 kubelet[2169]: E0904 17:19:08.450597 2169 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.60:6443/api/v1/nodes\": dial tcp 10.0.0.60:6443: connect: connection refused" node="localhost" Sep 4 17:19:08.476078 kubelet[2169]: E0904 17:19:08.476040 2169 kubelet.go:2353] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Sep 4 17:19:08.483166 kubelet[2169]: I0904 17:19:08.483128 2169 policy_none.go:49] "None policy: Start" Sep 4 17:19:08.484016 kubelet[2169]: I0904 17:19:08.483989 2169 memory_manager.go:170] "Starting memorymanager" policy="None" Sep 4 17:19:08.484111 kubelet[2169]: I0904 17:19:08.484095 2169 state_mem.go:35] "Initializing new in-memory state store" Sep 4 17:19:08.489990 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Sep 4 17:19:08.502779 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Sep 4 17:19:08.505727 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Sep 4 17:19:08.520840 kubelet[2169]: I0904 17:19:08.520766 2169 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 4 17:19:08.521133 kubelet[2169]: I0904 17:19:08.521092 2169 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 4 17:19:08.524205 kubelet[2169]: E0904 17:19:08.524063 2169 eviction_manager.go:282] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Sep 4 17:19:08.567320 kubelet[2169]: E0904 17:19:08.567270 2169 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.60:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.60:6443: connect: connection refused" interval="400ms" Sep 4 17:19:08.652490 kubelet[2169]: I0904 17:19:08.652460 2169 kubelet_node_status.go:73] "Attempting to register node" node="localhost" Sep 4 17:19:08.652870 kubelet[2169]: E0904 17:19:08.652821 2169 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.60:6443/api/v1/nodes\": dial tcp 10.0.0.60:6443: connect: connection refused" node="localhost" Sep 4 17:19:08.678930 kubelet[2169]: I0904 17:19:08.678895 2169 topology_manager.go:215] "Topology Admit Handler" podUID="8f64dda70f667de40727e508224c02f1" podNamespace="kube-system" podName="kube-apiserver-localhost" Sep 4 17:19:08.681743 kubelet[2169]: I0904 17:19:08.681720 2169 topology_manager.go:215] "Topology Admit Handler" podUID="7fa6213ac08f24a6b78f4cd3838d26c9" podNamespace="kube-system" podName="kube-controller-manager-localhost" Sep 4 17:19:08.682532 kubelet[2169]: I0904 17:19:08.682511 2169 topology_manager.go:215] "Topology Admit Handler" podUID="d9ddd765c3b0fcde29edfee4da9578f6" podNamespace="kube-system" podName="kube-scheduler-localhost" Sep 4 17:19:08.688306 systemd[1]: Created slice kubepods-burstable-pod7fa6213ac08f24a6b78f4cd3838d26c9.slice - libcontainer container kubepods-burstable-pod7fa6213ac08f24a6b78f4cd3838d26c9.slice. Sep 4 17:19:08.707451 systemd[1]: Created slice kubepods-burstable-pod8f64dda70f667de40727e508224c02f1.slice - libcontainer container kubepods-burstable-pod8f64dda70f667de40727e508224c02f1.slice. Sep 4 17:19:08.721678 systemd[1]: Created slice kubepods-burstable-podd9ddd765c3b0fcde29edfee4da9578f6.slice - libcontainer container kubepods-burstable-podd9ddd765c3b0fcde29edfee4da9578f6.slice. Sep 4 17:19:08.756960 kubelet[2169]: I0904 17:19:08.756925 2169 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/8f64dda70f667de40727e508224c02f1-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"8f64dda70f667de40727e508224c02f1\") " pod="kube-system/kube-apiserver-localhost" Sep 4 17:19:08.757030 kubelet[2169]: I0904 17:19:08.756972 2169 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/7fa6213ac08f24a6b78f4cd3838d26c9-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"7fa6213ac08f24a6b78f4cd3838d26c9\") " pod="kube-system/kube-controller-manager-localhost" Sep 4 17:19:08.757030 kubelet[2169]: I0904 17:19:08.756994 2169 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/7fa6213ac08f24a6b78f4cd3838d26c9-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"7fa6213ac08f24a6b78f4cd3838d26c9\") " pod="kube-system/kube-controller-manager-localhost" Sep 4 17:19:08.757030 kubelet[2169]: I0904 17:19:08.757014 2169 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/7fa6213ac08f24a6b78f4cd3838d26c9-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"7fa6213ac08f24a6b78f4cd3838d26c9\") " pod="kube-system/kube-controller-manager-localhost" Sep 4 17:19:08.757030 kubelet[2169]: I0904 17:19:08.757032 2169 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/8f64dda70f667de40727e508224c02f1-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"8f64dda70f667de40727e508224c02f1\") " pod="kube-system/kube-apiserver-localhost" Sep 4 17:19:08.757124 kubelet[2169]: I0904 17:19:08.757052 2169 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/7fa6213ac08f24a6b78f4cd3838d26c9-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"7fa6213ac08f24a6b78f4cd3838d26c9\") " pod="kube-system/kube-controller-manager-localhost" Sep 4 17:19:08.757124 kubelet[2169]: I0904 17:19:08.757069 2169 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/7fa6213ac08f24a6b78f4cd3838d26c9-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"7fa6213ac08f24a6b78f4cd3838d26c9\") " pod="kube-system/kube-controller-manager-localhost" Sep 4 17:19:08.757124 kubelet[2169]: I0904 17:19:08.757087 2169 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/d9ddd765c3b0fcde29edfee4da9578f6-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"d9ddd765c3b0fcde29edfee4da9578f6\") " pod="kube-system/kube-scheduler-localhost" Sep 4 17:19:08.757124 kubelet[2169]: I0904 17:19:08.757105 2169 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/8f64dda70f667de40727e508224c02f1-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"8f64dda70f667de40727e508224c02f1\") " pod="kube-system/kube-apiserver-localhost" Sep 4 17:19:08.967911 kubelet[2169]: E0904 17:19:08.967879 2169 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.60:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.60:6443: connect: connection refused" interval="800ms" Sep 4 17:19:09.007308 kubelet[2169]: E0904 17:19:09.007281 2169 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:09.009756 containerd[1448]: time="2024-09-04T17:19:09.009707129Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:7fa6213ac08f24a6b78f4cd3838d26c9,Namespace:kube-system,Attempt:0,}" Sep 4 17:19:09.019976 kubelet[2169]: E0904 17:19:09.019947 2169 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:09.022809 containerd[1448]: time="2024-09-04T17:19:09.022771873Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:8f64dda70f667de40727e508224c02f1,Namespace:kube-system,Attempt:0,}" Sep 4 17:19:09.024079 kubelet[2169]: E0904 17:19:09.024014 2169 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:09.024763 containerd[1448]: time="2024-09-04T17:19:09.024636831Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:d9ddd765c3b0fcde29edfee4da9578f6,Namespace:kube-system,Attempt:0,}" Sep 4 17:19:09.054790 kubelet[2169]: I0904 17:19:09.054741 2169 kubelet_node_status.go:73] "Attempting to register node" node="localhost" Sep 4 17:19:09.055179 kubelet[2169]: E0904 17:19:09.055102 2169 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.60:6443/api/v1/nodes\": dial tcp 10.0.0.60:6443: connect: connection refused" node="localhost" Sep 4 17:19:09.292078 kubelet[2169]: W0904 17:19:09.291932 2169 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.Service: Get "https://10.0.0.60:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.60:6443: connect: connection refused Sep 4 17:19:09.292078 kubelet[2169]: E0904 17:19:09.291994 2169 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.0.0.60:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.60:6443: connect: connection refused Sep 4 17:19:09.294275 kubelet[2169]: W0904 17:19:09.294234 2169 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.RuntimeClass: Get "https://10.0.0.60:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.60:6443: connect: connection refused Sep 4 17:19:09.294275 kubelet[2169]: E0904 17:19:09.294261 2169 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.0.0.60:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.60:6443: connect: connection refused Sep 4 17:19:09.549215 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount728568011.mount: Deactivated successfully. Sep 4 17:19:09.552216 containerd[1448]: time="2024-09-04T17:19:09.552153600Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 4 17:19:09.555298 containerd[1448]: time="2024-09-04T17:19:09.555236391Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 4 17:19:09.558168 containerd[1448]: time="2024-09-04T17:19:09.558122665Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=269175" Sep 4 17:19:09.558880 containerd[1448]: time="2024-09-04T17:19:09.558851764Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Sep 4 17:19:09.559696 containerd[1448]: time="2024-09-04T17:19:09.559664880Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 4 17:19:09.560725 containerd[1448]: time="2024-09-04T17:19:09.560625544Z" level=info msg="ImageCreate event name:\"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 4 17:19:09.561124 containerd[1448]: time="2024-09-04T17:19:09.561093714Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Sep 4 17:19:09.563964 containerd[1448]: time="2024-09-04T17:19:09.563926897Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Sep 4 17:19:09.566332 containerd[1448]: time="2024-09-04T17:19:09.566123919Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 543.232823ms" Sep 4 17:19:09.566853 containerd[1448]: time="2024-09-04T17:19:09.566803489Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 556.989299ms" Sep 4 17:19:09.571799 containerd[1448]: time="2024-09-04T17:19:09.571729393Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 546.998664ms" Sep 4 17:19:09.735911 containerd[1448]: time="2024-09-04T17:19:09.735342559Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 4 17:19:09.735911 containerd[1448]: time="2024-09-04T17:19:09.735883143Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:19:09.736438 containerd[1448]: time="2024-09-04T17:19:09.735906827Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 4 17:19:09.736438 containerd[1448]: time="2024-09-04T17:19:09.736060217Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:19:09.736536 containerd[1448]: time="2024-09-04T17:19:09.733074484Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 4 17:19:09.736708 containerd[1448]: time="2024-09-04T17:19:09.736660052Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:19:09.736708 containerd[1448]: time="2024-09-04T17:19:09.736693178Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 4 17:19:09.737239 containerd[1448]: time="2024-09-04T17:19:09.736704380Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:19:09.737781 containerd[1448]: time="2024-09-04T17:19:09.737722015Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 4 17:19:09.737781 containerd[1448]: time="2024-09-04T17:19:09.737764343Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:19:09.737917 containerd[1448]: time="2024-09-04T17:19:09.737778626Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 4 17:19:09.737917 containerd[1448]: time="2024-09-04T17:19:09.737790228Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:19:09.756046 systemd[1]: Started cri-containerd-f085814aec8706960310907f0ba8acc748759de6fea412cf034d7c3973a7ab8d.scope - libcontainer container f085814aec8706960310907f0ba8acc748759de6fea412cf034d7c3973a7ab8d. Sep 4 17:19:09.760027 systemd[1]: Started cri-containerd-98b14a6e8334e36093f7be5bd30fa8463c73ed612cddd6b5e766c60b4d290def.scope - libcontainer container 98b14a6e8334e36093f7be5bd30fa8463c73ed612cddd6b5e766c60b4d290def. Sep 4 17:19:09.761556 systemd[1]: Started cri-containerd-eef3eb1dbf35b02fb2cfa156fdfc100639e98f7ef3cadf56ec9ad02e5af3f566.scope - libcontainer container eef3eb1dbf35b02fb2cfa156fdfc100639e98f7ef3cadf56ec9ad02e5af3f566. Sep 4 17:19:09.768914 kubelet[2169]: E0904 17:19:09.768877 2169 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.60:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.60:6443: connect: connection refused" interval="1.6s" Sep 4 17:19:09.791372 containerd[1448]: time="2024-09-04T17:19:09.791231034Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:d9ddd765c3b0fcde29edfee4da9578f6,Namespace:kube-system,Attempt:0,} returns sandbox id \"f085814aec8706960310907f0ba8acc748759de6fea412cf034d7c3973a7ab8d\"" Sep 4 17:19:09.792172 kubelet[2169]: E0904 17:19:09.792137 2169 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:09.800298 containerd[1448]: time="2024-09-04T17:19:09.800187951Z" level=info msg="CreateContainer within sandbox \"f085814aec8706960310907f0ba8acc748759de6fea412cf034d7c3973a7ab8d\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Sep 4 17:19:09.802395 containerd[1448]: time="2024-09-04T17:19:09.802280992Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:8f64dda70f667de40727e508224c02f1,Namespace:kube-system,Attempt:0,} returns sandbox id \"eef3eb1dbf35b02fb2cfa156fdfc100639e98f7ef3cadf56ec9ad02e5af3f566\"" Sep 4 17:19:09.802641 containerd[1448]: time="2024-09-04T17:19:09.802614216Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:7fa6213ac08f24a6b78f4cd3838d26c9,Namespace:kube-system,Attempt:0,} returns sandbox id \"98b14a6e8334e36093f7be5bd30fa8463c73ed612cddd6b5e766c60b4d290def\"" Sep 4 17:19:09.803077 kubelet[2169]: E0904 17:19:09.803033 2169 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:09.803679 kubelet[2169]: E0904 17:19:09.803543 2169 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:09.805524 containerd[1448]: time="2024-09-04T17:19:09.805404351Z" level=info msg="CreateContainer within sandbox \"98b14a6e8334e36093f7be5bd30fa8463c73ed612cddd6b5e766c60b4d290def\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Sep 4 17:19:09.805665 containerd[1448]: time="2024-09-04T17:19:09.805634115Z" level=info msg="CreateContainer within sandbox \"eef3eb1dbf35b02fb2cfa156fdfc100639e98f7ef3cadf56ec9ad02e5af3f566\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Sep 4 17:19:09.812154 kubelet[2169]: W0904 17:19:09.812092 2169 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.CSIDriver: Get "https://10.0.0.60:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.60:6443: connect: connection refused Sep 4 17:19:09.812154 kubelet[2169]: E0904 17:19:09.812150 2169 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.0.0.60:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.60:6443: connect: connection refused Sep 4 17:19:09.817923 containerd[1448]: time="2024-09-04T17:19:09.817865420Z" level=info msg="CreateContainer within sandbox \"f085814aec8706960310907f0ba8acc748759de6fea412cf034d7c3973a7ab8d\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"becc936e81b23fe6947c44f7fdbe5ce1cd1f937b9d8d92beab73252f0983ea42\"" Sep 4 17:19:09.818491 containerd[1448]: time="2024-09-04T17:19:09.818461734Z" level=info msg="StartContainer for \"becc936e81b23fe6947c44f7fdbe5ce1cd1f937b9d8d92beab73252f0983ea42\"" Sep 4 17:19:09.823572 containerd[1448]: time="2024-09-04T17:19:09.823454411Z" level=info msg="CreateContainer within sandbox \"98b14a6e8334e36093f7be5bd30fa8463c73ed612cddd6b5e766c60b4d290def\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"7e86b2c07d527b3671e897899b5898c803a36c5e47ca6c626dd8c3f4927d9869\"" Sep 4 17:19:09.824111 containerd[1448]: time="2024-09-04T17:19:09.824086412Z" level=info msg="StartContainer for \"7e86b2c07d527b3671e897899b5898c803a36c5e47ca6c626dd8c3f4927d9869\"" Sep 4 17:19:09.826888 containerd[1448]: time="2024-09-04T17:19:09.826854143Z" level=info msg="CreateContainer within sandbox \"eef3eb1dbf35b02fb2cfa156fdfc100639e98f7ef3cadf56ec9ad02e5af3f566\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"7f9d79c7be6a34e39f1513c3a05e18bb331fda99e89a610cd0dfe8614ad48256\"" Sep 4 17:19:09.827274 containerd[1448]: time="2024-09-04T17:19:09.827216772Z" level=info msg="StartContainer for \"7f9d79c7be6a34e39f1513c3a05e18bb331fda99e89a610cd0dfe8614ad48256\"" Sep 4 17:19:09.843977 systemd[1]: Started cri-containerd-becc936e81b23fe6947c44f7fdbe5ce1cd1f937b9d8d92beab73252f0983ea42.scope - libcontainer container becc936e81b23fe6947c44f7fdbe5ce1cd1f937b9d8d92beab73252f0983ea42. Sep 4 17:19:09.847293 systemd[1]: Started cri-containerd-7e86b2c07d527b3671e897899b5898c803a36c5e47ca6c626dd8c3f4927d9869.scope - libcontainer container 7e86b2c07d527b3671e897899b5898c803a36c5e47ca6c626dd8c3f4927d9869. Sep 4 17:19:09.848052 kubelet[2169]: W0904 17:19:09.848003 2169 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.Node: Get "https://10.0.0.60:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.60:6443: connect: connection refused Sep 4 17:19:09.848126 kubelet[2169]: E0904 17:19:09.848062 2169 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.0.0.60:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.60:6443: connect: connection refused Sep 4 17:19:09.854424 systemd[1]: Started cri-containerd-7f9d79c7be6a34e39f1513c3a05e18bb331fda99e89a610cd0dfe8614ad48256.scope - libcontainer container 7f9d79c7be6a34e39f1513c3a05e18bb331fda99e89a610cd0dfe8614ad48256. Sep 4 17:19:09.856440 kubelet[2169]: I0904 17:19:09.856415 2169 kubelet_node_status.go:73] "Attempting to register node" node="localhost" Sep 4 17:19:09.856791 kubelet[2169]: E0904 17:19:09.856769 2169 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.60:6443/api/v1/nodes\": dial tcp 10.0.0.60:6443: connect: connection refused" node="localhost" Sep 4 17:19:09.882191 containerd[1448]: time="2024-09-04T17:19:09.882029280Z" level=info msg="StartContainer for \"becc936e81b23fe6947c44f7fdbe5ce1cd1f937b9d8d92beab73252f0983ea42\" returns successfully" Sep 4 17:19:09.886712 containerd[1448]: time="2024-09-04T17:19:09.886661808Z" level=info msg="StartContainer for \"7e86b2c07d527b3671e897899b5898c803a36c5e47ca6c626dd8c3f4927d9869\" returns successfully" Sep 4 17:19:09.890594 containerd[1448]: time="2024-09-04T17:19:09.890136475Z" level=info msg="StartContainer for \"7f9d79c7be6a34e39f1513c3a05e18bb331fda99e89a610cd0dfe8614ad48256\" returns successfully" Sep 4 17:19:10.389492 kubelet[2169]: E0904 17:19:10.389390 2169 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:10.390783 kubelet[2169]: E0904 17:19:10.390758 2169 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:10.391759 kubelet[2169]: E0904 17:19:10.391741 2169 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:11.394816 kubelet[2169]: E0904 17:19:11.394775 2169 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:11.458851 kubelet[2169]: I0904 17:19:11.458778 2169 kubelet_node_status.go:73] "Attempting to register node" node="localhost" Sep 4 17:19:11.517132 kubelet[2169]: E0904 17:19:11.517098 2169 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Sep 4 17:19:11.614256 kubelet[2169]: I0904 17:19:11.614082 2169 kubelet_node_status.go:76] "Successfully registered node" node="localhost" Sep 4 17:19:11.622321 kubelet[2169]: E0904 17:19:11.622284 2169 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 4 17:19:11.723194 kubelet[2169]: E0904 17:19:11.723078 2169 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 4 17:19:11.823701 kubelet[2169]: E0904 17:19:11.823655 2169 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 4 17:19:11.924156 kubelet[2169]: E0904 17:19:11.924123 2169 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 4 17:19:12.025221 kubelet[2169]: E0904 17:19:12.025111 2169 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 4 17:19:12.125645 kubelet[2169]: E0904 17:19:12.125604 2169 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 4 17:19:12.226558 kubelet[2169]: E0904 17:19:12.226516 2169 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 4 17:19:12.327391 kubelet[2169]: E0904 17:19:12.327273 2169 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 4 17:19:12.428094 kubelet[2169]: E0904 17:19:12.428044 2169 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 4 17:19:12.529127 kubelet[2169]: E0904 17:19:12.529083 2169 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 4 17:19:12.629821 kubelet[2169]: E0904 17:19:12.629712 2169 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 4 17:19:12.730318 kubelet[2169]: E0904 17:19:12.730248 2169 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Sep 4 17:19:13.321107 kubelet[2169]: I0904 17:19:13.321003 2169 apiserver.go:52] "Watching apiserver" Sep 4 17:19:13.353437 kubelet[2169]: I0904 17:19:13.353377 2169 desired_state_of_world_populator.go:159] "Finished populating initial desired state of world" Sep 4 17:19:14.098489 systemd[1]: Reloading requested from client PID 2444 ('systemctl') (unit session-7.scope)... Sep 4 17:19:14.098509 systemd[1]: Reloading... Sep 4 17:19:14.152880 zram_generator::config[2481]: No configuration found. Sep 4 17:19:14.340533 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 4 17:19:14.407672 systemd[1]: Reloading finished in 308 ms. Sep 4 17:19:14.439351 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Sep 4 17:19:14.452951 systemd[1]: kubelet.service: Deactivated successfully. Sep 4 17:19:14.453946 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Sep 4 17:19:14.454081 systemd[1]: kubelet.service: Consumed 1.277s CPU time, 115.0M memory peak, 0B memory swap peak. Sep 4 17:19:14.462135 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Sep 4 17:19:14.560745 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Sep 4 17:19:14.573216 (kubelet)[2523]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Sep 4 17:19:14.637926 kubelet[2523]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 4 17:19:14.638857 kubelet[2523]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Sep 4 17:19:14.638857 kubelet[2523]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 4 17:19:14.638857 kubelet[2523]: I0904 17:19:14.638279 2523 server.go:204] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 4 17:19:14.645585 kubelet[2523]: I0904 17:19:14.645553 2523 server.go:487] "Kubelet version" kubeletVersion="v1.29.2" Sep 4 17:19:14.646032 kubelet[2523]: I0904 17:19:14.645707 2523 server.go:489] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 4 17:19:14.646243 kubelet[2523]: I0904 17:19:14.646223 2523 server.go:919] "Client rotation is on, will bootstrap in background" Sep 4 17:19:14.651408 kubelet[2523]: I0904 17:19:14.651376 2523 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Sep 4 17:19:14.653978 kubelet[2523]: I0904 17:19:14.653817 2523 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 4 17:19:14.666262 kubelet[2523]: I0904 17:19:14.666024 2523 server.go:745] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 4 17:19:14.666262 kubelet[2523]: I0904 17:19:14.666217 2523 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 4 17:19:14.667070 kubelet[2523]: I0904 17:19:14.666394 2523 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Sep 4 17:19:14.667070 kubelet[2523]: I0904 17:19:14.666421 2523 topology_manager.go:138] "Creating topology manager with none policy" Sep 4 17:19:14.667070 kubelet[2523]: I0904 17:19:14.666429 2523 container_manager_linux.go:301] "Creating device plugin manager" Sep 4 17:19:14.667070 kubelet[2523]: I0904 17:19:14.666459 2523 state_mem.go:36] "Initialized new in-memory state store" Sep 4 17:19:14.667070 kubelet[2523]: I0904 17:19:14.666545 2523 kubelet.go:396] "Attempting to sync node with API server" Sep 4 17:19:14.667070 kubelet[2523]: I0904 17:19:14.666558 2523 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 4 17:19:14.667070 kubelet[2523]: I0904 17:19:14.666596 2523 kubelet.go:312] "Adding apiserver pod source" Sep 4 17:19:14.667329 kubelet[2523]: I0904 17:19:14.666610 2523 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 4 17:19:14.671301 kubelet[2523]: I0904 17:19:14.671141 2523 kuberuntime_manager.go:258] "Container runtime initialized" containerRuntime="containerd" version="v1.7.17" apiVersion="v1" Sep 4 17:19:14.672321 kubelet[2523]: I0904 17:19:14.672190 2523 kubelet.go:809] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 4 17:19:14.673645 kubelet[2523]: I0904 17:19:14.672687 2523 server.go:1256] "Started kubelet" Sep 4 17:19:14.673645 kubelet[2523]: I0904 17:19:14.673136 2523 server.go:162] "Starting to listen" address="0.0.0.0" port=10250 Sep 4 17:19:14.680388 kubelet[2523]: I0904 17:19:14.675505 2523 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 4 17:19:14.680388 kubelet[2523]: I0904 17:19:14.675672 2523 server.go:233] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 4 17:19:14.680388 kubelet[2523]: I0904 17:19:14.676267 2523 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 4 17:19:14.680388 kubelet[2523]: I0904 17:19:14.678091 2523 server.go:461] "Adding debug handlers to kubelet server" Sep 4 17:19:14.685518 kubelet[2523]: I0904 17:19:14.685479 2523 volume_manager.go:291] "Starting Kubelet Volume Manager" Sep 4 17:19:14.686226 kubelet[2523]: I0904 17:19:14.686204 2523 desired_state_of_world_populator.go:151] "Desired state populator starts to run" Sep 4 17:19:14.686467 kubelet[2523]: I0904 17:19:14.686452 2523 reconciler_new.go:29] "Reconciler: start to sync state" Sep 4 17:19:14.700793 kubelet[2523]: I0904 17:19:14.700753 2523 factory.go:221] Registration of the systemd container factory successfully Sep 4 17:19:14.701979 kubelet[2523]: I0904 17:19:14.701880 2523 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 4 17:19:14.703283 kubelet[2523]: I0904 17:19:14.703258 2523 factory.go:221] Registration of the containerd container factory successfully Sep 4 17:19:14.710483 kubelet[2523]: I0904 17:19:14.710017 2523 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 4 17:19:14.712424 kubelet[2523]: I0904 17:19:14.712392 2523 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 4 17:19:14.712897 kubelet[2523]: I0904 17:19:14.712552 2523 status_manager.go:217] "Starting to sync pod status with apiserver" Sep 4 17:19:14.712897 kubelet[2523]: I0904 17:19:14.712580 2523 kubelet.go:2329] "Starting kubelet main sync loop" Sep 4 17:19:14.712897 kubelet[2523]: E0904 17:19:14.712645 2523 kubelet.go:2353] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 4 17:19:14.715891 kubelet[2523]: E0904 17:19:14.715854 2523 kubelet.go:1462] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 4 17:19:14.748363 kubelet[2523]: I0904 17:19:14.747990 2523 cpu_manager.go:214] "Starting CPU manager" policy="none" Sep 4 17:19:14.748363 kubelet[2523]: I0904 17:19:14.748018 2523 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Sep 4 17:19:14.748363 kubelet[2523]: I0904 17:19:14.748039 2523 state_mem.go:36] "Initialized new in-memory state store" Sep 4 17:19:14.748363 kubelet[2523]: I0904 17:19:14.748183 2523 state_mem.go:88] "Updated default CPUSet" cpuSet="" Sep 4 17:19:14.748363 kubelet[2523]: I0904 17:19:14.748202 2523 state_mem.go:96] "Updated CPUSet assignments" assignments={} Sep 4 17:19:14.748363 kubelet[2523]: I0904 17:19:14.748210 2523 policy_none.go:49] "None policy: Start" Sep 4 17:19:14.749674 kubelet[2523]: I0904 17:19:14.748983 2523 memory_manager.go:170] "Starting memorymanager" policy="None" Sep 4 17:19:14.749674 kubelet[2523]: I0904 17:19:14.749008 2523 state_mem.go:35] "Initializing new in-memory state store" Sep 4 17:19:14.749674 kubelet[2523]: I0904 17:19:14.749198 2523 state_mem.go:75] "Updated machine memory state" Sep 4 17:19:14.755431 kubelet[2523]: I0904 17:19:14.755242 2523 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 4 17:19:14.756630 kubelet[2523]: I0904 17:19:14.756569 2523 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 4 17:19:14.790193 kubelet[2523]: I0904 17:19:14.789887 2523 kubelet_node_status.go:73] "Attempting to register node" node="localhost" Sep 4 17:19:14.796679 kubelet[2523]: I0904 17:19:14.796637 2523 kubelet_node_status.go:112] "Node was previously registered" node="localhost" Sep 4 17:19:14.797595 kubelet[2523]: I0904 17:19:14.796722 2523 kubelet_node_status.go:76] "Successfully registered node" node="localhost" Sep 4 17:19:14.813712 kubelet[2523]: I0904 17:19:14.813662 2523 topology_manager.go:215] "Topology Admit Handler" podUID="8f64dda70f667de40727e508224c02f1" podNamespace="kube-system" podName="kube-apiserver-localhost" Sep 4 17:19:14.813845 kubelet[2523]: I0904 17:19:14.813760 2523 topology_manager.go:215] "Topology Admit Handler" podUID="7fa6213ac08f24a6b78f4cd3838d26c9" podNamespace="kube-system" podName="kube-controller-manager-localhost" Sep 4 17:19:14.813845 kubelet[2523]: I0904 17:19:14.813816 2523 topology_manager.go:215] "Topology Admit Handler" podUID="d9ddd765c3b0fcde29edfee4da9578f6" podNamespace="kube-system" podName="kube-scheduler-localhost" Sep 4 17:19:14.888282 kubelet[2523]: I0904 17:19:14.888157 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/d9ddd765c3b0fcde29edfee4da9578f6-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"d9ddd765c3b0fcde29edfee4da9578f6\") " pod="kube-system/kube-scheduler-localhost" Sep 4 17:19:14.888282 kubelet[2523]: I0904 17:19:14.888206 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/8f64dda70f667de40727e508224c02f1-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"8f64dda70f667de40727e508224c02f1\") " pod="kube-system/kube-apiserver-localhost" Sep 4 17:19:14.888282 kubelet[2523]: I0904 17:19:14.888230 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/8f64dda70f667de40727e508224c02f1-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"8f64dda70f667de40727e508224c02f1\") " pod="kube-system/kube-apiserver-localhost" Sep 4 17:19:14.888282 kubelet[2523]: I0904 17:19:14.888260 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/7fa6213ac08f24a6b78f4cd3838d26c9-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"7fa6213ac08f24a6b78f4cd3838d26c9\") " pod="kube-system/kube-controller-manager-localhost" Sep 4 17:19:14.888481 kubelet[2523]: I0904 17:19:14.888302 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/7fa6213ac08f24a6b78f4cd3838d26c9-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"7fa6213ac08f24a6b78f4cd3838d26c9\") " pod="kube-system/kube-controller-manager-localhost" Sep 4 17:19:14.888481 kubelet[2523]: I0904 17:19:14.888353 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/7fa6213ac08f24a6b78f4cd3838d26c9-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"7fa6213ac08f24a6b78f4cd3838d26c9\") " pod="kube-system/kube-controller-manager-localhost" Sep 4 17:19:14.888481 kubelet[2523]: I0904 17:19:14.888379 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/8f64dda70f667de40727e508224c02f1-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"8f64dda70f667de40727e508224c02f1\") " pod="kube-system/kube-apiserver-localhost" Sep 4 17:19:14.888481 kubelet[2523]: I0904 17:19:14.888399 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/7fa6213ac08f24a6b78f4cd3838d26c9-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"7fa6213ac08f24a6b78f4cd3838d26c9\") " pod="kube-system/kube-controller-manager-localhost" Sep 4 17:19:14.888481 kubelet[2523]: I0904 17:19:14.888419 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/7fa6213ac08f24a6b78f4cd3838d26c9-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"7fa6213ac08f24a6b78f4cd3838d26c9\") " pod="kube-system/kube-controller-manager-localhost" Sep 4 17:19:15.122001 kubelet[2523]: E0904 17:19:15.121853 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:15.124204 kubelet[2523]: E0904 17:19:15.124127 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:15.124668 kubelet[2523]: E0904 17:19:15.124477 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:15.667144 kubelet[2523]: I0904 17:19:15.667049 2523 apiserver.go:52] "Watching apiserver" Sep 4 17:19:15.686842 kubelet[2523]: I0904 17:19:15.686792 2523 desired_state_of_world_populator.go:159] "Finished populating initial desired state of world" Sep 4 17:19:15.731861 kubelet[2523]: E0904 17:19:15.730647 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:15.731861 kubelet[2523]: E0904 17:19:15.730713 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:15.737305 kubelet[2523]: E0904 17:19:15.736752 2523 kubelet.go:1921] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Sep 4 17:19:15.737305 kubelet[2523]: E0904 17:19:15.737241 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:15.752074 kubelet[2523]: I0904 17:19:15.752029 2523 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=1.7519846860000001 podStartE2EDuration="1.751984686s" podCreationTimestamp="2024-09-04 17:19:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-09-04 17:19:15.751932641 +0000 UTC m=+1.172846404" watchObservedRunningTime="2024-09-04 17:19:15.751984686 +0000 UTC m=+1.172898449" Sep 4 17:19:15.766059 kubelet[2523]: I0904 17:19:15.765568 2523 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.765527326 podStartE2EDuration="1.765527326s" podCreationTimestamp="2024-09-04 17:19:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-09-04 17:19:15.765227058 +0000 UTC m=+1.186140821" watchObservedRunningTime="2024-09-04 17:19:15.765527326 +0000 UTC m=+1.186441089" Sep 4 17:19:15.781325 kubelet[2523]: I0904 17:19:15.781279 2523 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.7812357319999998 podStartE2EDuration="1.781235732s" podCreationTimestamp="2024-09-04 17:19:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-09-04 17:19:15.779870043 +0000 UTC m=+1.200783806" watchObservedRunningTime="2024-09-04 17:19:15.781235732 +0000 UTC m=+1.202149495" Sep 4 17:19:16.732454 kubelet[2523]: E0904 17:19:16.732411 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:18.849267 sudo[1619]: pam_unix(sudo:session): session closed for user root Sep 4 17:19:18.851943 sshd[1616]: pam_unix(sshd:session): session closed for user core Sep 4 17:19:18.855591 systemd[1]: sshd@6-10.0.0.60:22-10.0.0.1:51078.service: Deactivated successfully. Sep 4 17:19:18.859317 systemd[1]: session-7.scope: Deactivated successfully. Sep 4 17:19:18.859584 systemd[1]: session-7.scope: Consumed 8.348s CPU time, 140.9M memory peak, 0B memory swap peak. Sep 4 17:19:18.860214 systemd-logind[1423]: Session 7 logged out. Waiting for processes to exit. Sep 4 17:19:18.861167 systemd-logind[1423]: Removed session 7. Sep 4 17:19:18.895371 kubelet[2523]: E0904 17:19:18.895201 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:23.129484 update_engine[1430]: I0904 17:19:23.128844 1430 update_attempter.cc:509] Updating boot flags... Sep 4 17:19:23.154787 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 35 scanned by (udev-worker) (2621) Sep 4 17:19:23.195853 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 35 scanned by (udev-worker) (2623) Sep 4 17:19:23.229885 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 35 scanned by (udev-worker) (2623) Sep 4 17:19:23.954259 kubelet[2523]: E0904 17:19:23.953917 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:24.748498 kubelet[2523]: E0904 17:19:24.748460 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:24.769956 kubelet[2523]: E0904 17:19:24.769444 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:25.749581 kubelet[2523]: E0904 17:19:25.749556 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:27.702943 kubelet[2523]: I0904 17:19:27.702908 2523 kuberuntime_manager.go:1529] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Sep 4 17:19:27.703273 containerd[1448]: time="2024-09-04T17:19:27.703240085Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Sep 4 17:19:27.703535 kubelet[2523]: I0904 17:19:27.703502 2523 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Sep 4 17:19:28.683025 kubelet[2523]: I0904 17:19:28.682592 2523 topology_manager.go:215] "Topology Admit Handler" podUID="76d4075c-17f4-443c-b6cb-083612dbc953" podNamespace="kube-system" podName="kube-proxy-s2jgw" Sep 4 17:19:28.690102 systemd[1]: Created slice kubepods-besteffort-pod76d4075c_17f4_443c_b6cb_083612dbc953.slice - libcontainer container kubepods-besteffort-pod76d4075c_17f4_443c_b6cb_083612dbc953.slice. Sep 4 17:19:28.803418 kubelet[2523]: I0904 17:19:28.803050 2523 topology_manager.go:215] "Topology Admit Handler" podUID="d011b165-a246-4085-8ff2-d5d4b39a93c0" podNamespace="tigera-operator" podName="tigera-operator-5d56685c77-9zxbw" Sep 4 17:19:28.810243 systemd[1]: Created slice kubepods-besteffort-podd011b165_a246_4085_8ff2_d5d4b39a93c0.slice - libcontainer container kubepods-besteffort-podd011b165_a246_4085_8ff2_d5d4b39a93c0.slice. Sep 4 17:19:28.877512 kubelet[2523]: I0904 17:19:28.877468 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/76d4075c-17f4-443c-b6cb-083612dbc953-xtables-lock\") pod \"kube-proxy-s2jgw\" (UID: \"76d4075c-17f4-443c-b6cb-083612dbc953\") " pod="kube-system/kube-proxy-s2jgw" Sep 4 17:19:28.877512 kubelet[2523]: I0904 17:19:28.877513 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/76d4075c-17f4-443c-b6cb-083612dbc953-kube-proxy\") pod \"kube-proxy-s2jgw\" (UID: \"76d4075c-17f4-443c-b6cb-083612dbc953\") " pod="kube-system/kube-proxy-s2jgw" Sep 4 17:19:28.877670 kubelet[2523]: I0904 17:19:28.877540 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/76d4075c-17f4-443c-b6cb-083612dbc953-lib-modules\") pod \"kube-proxy-s2jgw\" (UID: \"76d4075c-17f4-443c-b6cb-083612dbc953\") " pod="kube-system/kube-proxy-s2jgw" Sep 4 17:19:28.877670 kubelet[2523]: I0904 17:19:28.877560 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsvmc\" (UniqueName: \"kubernetes.io/projected/76d4075c-17f4-443c-b6cb-083612dbc953-kube-api-access-nsvmc\") pod \"kube-proxy-s2jgw\" (UID: \"76d4075c-17f4-443c-b6cb-083612dbc953\") " pod="kube-system/kube-proxy-s2jgw" Sep 4 17:19:28.904116 kubelet[2523]: E0904 17:19:28.904090 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:28.978983 kubelet[2523]: I0904 17:19:28.978712 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b89wr\" (UniqueName: \"kubernetes.io/projected/d011b165-a246-4085-8ff2-d5d4b39a93c0-kube-api-access-b89wr\") pod \"tigera-operator-5d56685c77-9zxbw\" (UID: \"d011b165-a246-4085-8ff2-d5d4b39a93c0\") " pod="tigera-operator/tigera-operator-5d56685c77-9zxbw" Sep 4 17:19:28.978983 kubelet[2523]: I0904 17:19:28.978801 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/d011b165-a246-4085-8ff2-d5d4b39a93c0-var-lib-calico\") pod \"tigera-operator-5d56685c77-9zxbw\" (UID: \"d011b165-a246-4085-8ff2-d5d4b39a93c0\") " pod="tigera-operator/tigera-operator-5d56685c77-9zxbw" Sep 4 17:19:29.001653 kubelet[2523]: E0904 17:19:29.001334 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:29.002944 containerd[1448]: time="2024-09-04T17:19:29.002908134Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-s2jgw,Uid:76d4075c-17f4-443c-b6cb-083612dbc953,Namespace:kube-system,Attempt:0,}" Sep 4 17:19:29.025794 containerd[1448]: time="2024-09-04T17:19:29.025678505Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 4 17:19:29.025794 containerd[1448]: time="2024-09-04T17:19:29.025747468Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:19:29.025794 containerd[1448]: time="2024-09-04T17:19:29.025763149Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 4 17:19:29.025794 containerd[1448]: time="2024-09-04T17:19:29.025776190Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:19:29.046059 systemd[1]: Started cri-containerd-8850a28588de14de4072af153f45127316d47463b3962712004f3eeb4fec2875.scope - libcontainer container 8850a28588de14de4072af153f45127316d47463b3962712004f3eeb4fec2875. Sep 4 17:19:29.063076 containerd[1448]: time="2024-09-04T17:19:29.063037629Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-s2jgw,Uid:76d4075c-17f4-443c-b6cb-083612dbc953,Namespace:kube-system,Attempt:0,} returns sandbox id \"8850a28588de14de4072af153f45127316d47463b3962712004f3eeb4fec2875\"" Sep 4 17:19:29.063902 kubelet[2523]: E0904 17:19:29.063818 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:29.066175 containerd[1448]: time="2024-09-04T17:19:29.066137692Z" level=info msg="CreateContainer within sandbox \"8850a28588de14de4072af153f45127316d47463b3962712004f3eeb4fec2875\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Sep 4 17:19:29.085805 containerd[1448]: time="2024-09-04T17:19:29.085759958Z" level=info msg="CreateContainer within sandbox \"8850a28588de14de4072af153f45127316d47463b3962712004f3eeb4fec2875\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"28913251d2686f1aca4fba4e609967243a8ead31b95ae5e652d532363ae6e61b\"" Sep 4 17:19:29.086918 containerd[1448]: time="2024-09-04T17:19:29.086883210Z" level=info msg="StartContainer for \"28913251d2686f1aca4fba4e609967243a8ead31b95ae5e652d532363ae6e61b\"" Sep 4 17:19:29.109009 systemd[1]: Started cri-containerd-28913251d2686f1aca4fba4e609967243a8ead31b95ae5e652d532363ae6e61b.scope - libcontainer container 28913251d2686f1aca4fba4e609967243a8ead31b95ae5e652d532363ae6e61b. Sep 4 17:19:29.116470 containerd[1448]: time="2024-09-04T17:19:29.116431814Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-5d56685c77-9zxbw,Uid:d011b165-a246-4085-8ff2-d5d4b39a93c0,Namespace:tigera-operator,Attempt:0,}" Sep 4 17:19:29.140372 containerd[1448]: time="2024-09-04T17:19:29.139705168Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 4 17:19:29.140372 containerd[1448]: time="2024-09-04T17:19:29.139782091Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:19:29.140372 containerd[1448]: time="2024-09-04T17:19:29.139801572Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 4 17:19:29.140372 containerd[1448]: time="2024-09-04T17:19:29.139814493Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:19:29.151382 containerd[1448]: time="2024-09-04T17:19:29.151325344Z" level=info msg="StartContainer for \"28913251d2686f1aca4fba4e609967243a8ead31b95ae5e652d532363ae6e61b\" returns successfully" Sep 4 17:19:29.174632 systemd[1]: Started cri-containerd-1de1b7742878f43e682b93961fa2cd87868a94b033429ad8c3521938d7e134d6.scope - libcontainer container 1de1b7742878f43e682b93961fa2cd87868a94b033429ad8c3521938d7e134d6. Sep 4 17:19:29.213237 containerd[1448]: time="2024-09-04T17:19:29.213103955Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-5d56685c77-9zxbw,Uid:d011b165-a246-4085-8ff2-d5d4b39a93c0,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"1de1b7742878f43e682b93961fa2cd87868a94b033429ad8c3521938d7e134d6\"" Sep 4 17:19:29.228198 containerd[1448]: time="2024-09-04T17:19:29.227952560Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.34.3\"" Sep 4 17:19:29.758992 kubelet[2523]: E0904 17:19:29.758961 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:29.770933 kubelet[2523]: I0904 17:19:29.770895 2523 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-proxy-s2jgw" podStartSLOduration=1.770853375 podStartE2EDuration="1.770853375s" podCreationTimestamp="2024-09-04 17:19:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-09-04 17:19:29.770620324 +0000 UTC m=+15.191534127" watchObservedRunningTime="2024-09-04 17:19:29.770853375 +0000 UTC m=+15.191767138" Sep 4 17:19:30.351785 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3698350865.mount: Deactivated successfully. Sep 4 17:19:31.309512 containerd[1448]: time="2024-09-04T17:19:31.309457769Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.34.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:19:31.310374 containerd[1448]: time="2024-09-04T17:19:31.310326725Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.34.3: active requests=0, bytes read=19485903" Sep 4 17:19:31.311325 containerd[1448]: time="2024-09-04T17:19:31.311096638Z" level=info msg="ImageCreate event name:\"sha256:2fd8a2c22d96f6b41bf5709bd6ebbb915093532073f7039d03ab056b4e148f56\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:19:31.313354 containerd[1448]: time="2024-09-04T17:19:31.313312891Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:2cc4de6ad019ccc3abbd2615c159d0dcfb2ecdab90dc5805f08837d7c014d458\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:19:31.314295 containerd[1448]: time="2024-09-04T17:19:31.314244571Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.34.3\" with image id \"sha256:2fd8a2c22d96f6b41bf5709bd6ebbb915093532073f7039d03ab056b4e148f56\", repo tag \"quay.io/tigera/operator:v1.34.3\", repo digest \"quay.io/tigera/operator@sha256:2cc4de6ad019ccc3abbd2615c159d0dcfb2ecdab90dc5805f08837d7c014d458\", size \"19480102\" in 2.086251208s" Sep 4 17:19:31.314295 containerd[1448]: time="2024-09-04T17:19:31.314278652Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.34.3\" returns image reference \"sha256:2fd8a2c22d96f6b41bf5709bd6ebbb915093532073f7039d03ab056b4e148f56\"" Sep 4 17:19:31.316803 containerd[1448]: time="2024-09-04T17:19:31.316771517Z" level=info msg="CreateContainer within sandbox \"1de1b7742878f43e682b93961fa2cd87868a94b033429ad8c3521938d7e134d6\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Sep 4 17:19:31.324808 containerd[1448]: time="2024-09-04T17:19:31.324692731Z" level=info msg="CreateContainer within sandbox \"1de1b7742878f43e682b93961fa2cd87868a94b033429ad8c3521938d7e134d6\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"cecc60da53e87b798322452e9349216082bee6b643b005fb06ba05279c2e5ebc\"" Sep 4 17:19:31.325384 containerd[1448]: time="2024-09-04T17:19:31.325163431Z" level=info msg="StartContainer for \"cecc60da53e87b798322452e9349216082bee6b643b005fb06ba05279c2e5ebc\"" Sep 4 17:19:31.353009 systemd[1]: Started cri-containerd-cecc60da53e87b798322452e9349216082bee6b643b005fb06ba05279c2e5ebc.scope - libcontainer container cecc60da53e87b798322452e9349216082bee6b643b005fb06ba05279c2e5ebc. Sep 4 17:19:31.379851 containerd[1448]: time="2024-09-04T17:19:31.379642928Z" level=info msg="StartContainer for \"cecc60da53e87b798322452e9349216082bee6b643b005fb06ba05279c2e5ebc\" returns successfully" Sep 4 17:19:31.793887 kubelet[2523]: I0904 17:19:31.793848 2523 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="tigera-operator/tigera-operator-5d56685c77-9zxbw" podStartSLOduration=1.6927428610000002 podStartE2EDuration="3.793796747s" podCreationTimestamp="2024-09-04 17:19:28 +0000 UTC" firstStartedPulling="2024-09-04 17:19:29.214530621 +0000 UTC m=+14.635444344" lastFinishedPulling="2024-09-04 17:19:31.315584467 +0000 UTC m=+16.736498230" observedRunningTime="2024-09-04 17:19:31.793603339 +0000 UTC m=+17.214517062" watchObservedRunningTime="2024-09-04 17:19:31.793796747 +0000 UTC m=+17.214710510" Sep 4 17:19:35.401247 kubelet[2523]: I0904 17:19:35.401196 2523 topology_manager.go:215] "Topology Admit Handler" podUID="61e2b0bc-6e8f-437e-9d1c-dfc2e356162f" podNamespace="calico-system" podName="calico-typha-75f7c9c896-k4b7n" Sep 4 17:19:35.415273 systemd[1]: Created slice kubepods-besteffort-pod61e2b0bc_6e8f_437e_9d1c_dfc2e356162f.slice - libcontainer container kubepods-besteffort-pod61e2b0bc_6e8f_437e_9d1c_dfc2e356162f.slice. Sep 4 17:19:35.454053 kubelet[2523]: I0904 17:19:35.454014 2523 topology_manager.go:215] "Topology Admit Handler" podUID="1944397e-374a-4781-85a3-d9f5b73e16bd" podNamespace="calico-system" podName="calico-node-ktcqt" Sep 4 17:19:35.464309 systemd[1]: Created slice kubepods-besteffort-pod1944397e_374a_4781_85a3_d9f5b73e16bd.slice - libcontainer container kubepods-besteffort-pod1944397e_374a_4781_85a3_d9f5b73e16bd.slice. Sep 4 17:19:35.522121 kubelet[2523]: I0904 17:19:35.522085 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/61e2b0bc-6e8f-437e-9d1c-dfc2e356162f-tigera-ca-bundle\") pod \"calico-typha-75f7c9c896-k4b7n\" (UID: \"61e2b0bc-6e8f-437e-9d1c-dfc2e356162f\") " pod="calico-system/calico-typha-75f7c9c896-k4b7n" Sep 4 17:19:35.522121 kubelet[2523]: I0904 17:19:35.522151 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/61e2b0bc-6e8f-437e-9d1c-dfc2e356162f-typha-certs\") pod \"calico-typha-75f7c9c896-k4b7n\" (UID: \"61e2b0bc-6e8f-437e-9d1c-dfc2e356162f\") " pod="calico-system/calico-typha-75f7c9c896-k4b7n" Sep 4 17:19:35.522121 kubelet[2523]: I0904 17:19:35.522217 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjz8l\" (UniqueName: \"kubernetes.io/projected/61e2b0bc-6e8f-437e-9d1c-dfc2e356162f-kube-api-access-rjz8l\") pod \"calico-typha-75f7c9c896-k4b7n\" (UID: \"61e2b0bc-6e8f-437e-9d1c-dfc2e356162f\") " pod="calico-system/calico-typha-75f7c9c896-k4b7n" Sep 4 17:19:35.562536 kubelet[2523]: I0904 17:19:35.562491 2523 topology_manager.go:215] "Topology Admit Handler" podUID="2e2776ed-fa4e-45b8-bb3f-17de64345308" podNamespace="calico-system" podName="csi-node-driver-nsxnh" Sep 4 17:19:35.562787 kubelet[2523]: E0904 17:19:35.562767 2523 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-nsxnh" podUID="2e2776ed-fa4e-45b8-bb3f-17de64345308" Sep 4 17:19:35.622953 kubelet[2523]: I0904 17:19:35.622918 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/1944397e-374a-4781-85a3-d9f5b73e16bd-xtables-lock\") pod \"calico-node-ktcqt\" (UID: \"1944397e-374a-4781-85a3-d9f5b73e16bd\") " pod="calico-system/calico-node-ktcqt" Sep 4 17:19:35.622953 kubelet[2523]: I0904 17:19:35.622961 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/1944397e-374a-4781-85a3-d9f5b73e16bd-var-run-calico\") pod \"calico-node-ktcqt\" (UID: \"1944397e-374a-4781-85a3-d9f5b73e16bd\") " pod="calico-system/calico-node-ktcqt" Sep 4 17:19:35.623123 kubelet[2523]: I0904 17:19:35.622984 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/1944397e-374a-4781-85a3-d9f5b73e16bd-var-lib-calico\") pod \"calico-node-ktcqt\" (UID: \"1944397e-374a-4781-85a3-d9f5b73e16bd\") " pod="calico-system/calico-node-ktcqt" Sep 4 17:19:35.623123 kubelet[2523]: I0904 17:19:35.623004 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/1944397e-374a-4781-85a3-d9f5b73e16bd-cni-log-dir\") pod \"calico-node-ktcqt\" (UID: \"1944397e-374a-4781-85a3-d9f5b73e16bd\") " pod="calico-system/calico-node-ktcqt" Sep 4 17:19:35.623123 kubelet[2523]: I0904 17:19:35.623024 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1944397e-374a-4781-85a3-d9f5b73e16bd-lib-modules\") pod \"calico-node-ktcqt\" (UID: \"1944397e-374a-4781-85a3-d9f5b73e16bd\") " pod="calico-system/calico-node-ktcqt" Sep 4 17:19:35.623123 kubelet[2523]: I0904 17:19:35.623047 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/1944397e-374a-4781-85a3-d9f5b73e16bd-flexvol-driver-host\") pod \"calico-node-ktcqt\" (UID: \"1944397e-374a-4781-85a3-d9f5b73e16bd\") " pod="calico-system/calico-node-ktcqt" Sep 4 17:19:35.623123 kubelet[2523]: I0904 17:19:35.623082 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1944397e-374a-4781-85a3-d9f5b73e16bd-tigera-ca-bundle\") pod \"calico-node-ktcqt\" (UID: \"1944397e-374a-4781-85a3-d9f5b73e16bd\") " pod="calico-system/calico-node-ktcqt" Sep 4 17:19:35.623229 kubelet[2523]: I0904 17:19:35.623102 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/1944397e-374a-4781-85a3-d9f5b73e16bd-cni-net-dir\") pod \"calico-node-ktcqt\" (UID: \"1944397e-374a-4781-85a3-d9f5b73e16bd\") " pod="calico-system/calico-node-ktcqt" Sep 4 17:19:35.623229 kubelet[2523]: I0904 17:19:35.623122 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9n9m\" (UniqueName: \"kubernetes.io/projected/1944397e-374a-4781-85a3-d9f5b73e16bd-kube-api-access-s9n9m\") pod \"calico-node-ktcqt\" (UID: \"1944397e-374a-4781-85a3-d9f5b73e16bd\") " pod="calico-system/calico-node-ktcqt" Sep 4 17:19:35.623229 kubelet[2523]: I0904 17:19:35.623161 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/1944397e-374a-4781-85a3-d9f5b73e16bd-cni-bin-dir\") pod \"calico-node-ktcqt\" (UID: \"1944397e-374a-4781-85a3-d9f5b73e16bd\") " pod="calico-system/calico-node-ktcqt" Sep 4 17:19:35.623229 kubelet[2523]: I0904 17:19:35.623181 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/1944397e-374a-4781-85a3-d9f5b73e16bd-policysync\") pod \"calico-node-ktcqt\" (UID: \"1944397e-374a-4781-85a3-d9f5b73e16bd\") " pod="calico-system/calico-node-ktcqt" Sep 4 17:19:35.623229 kubelet[2523]: I0904 17:19:35.623201 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/1944397e-374a-4781-85a3-d9f5b73e16bd-node-certs\") pod \"calico-node-ktcqt\" (UID: \"1944397e-374a-4781-85a3-d9f5b73e16bd\") " pod="calico-system/calico-node-ktcqt" Sep 4 17:19:35.719345 kubelet[2523]: E0904 17:19:35.719234 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:35.725187 kubelet[2523]: I0904 17:19:35.724107 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2e2776ed-fa4e-45b8-bb3f-17de64345308-socket-dir\") pod \"csi-node-driver-nsxnh\" (UID: \"2e2776ed-fa4e-45b8-bb3f-17de64345308\") " pod="calico-system/csi-node-driver-nsxnh" Sep 4 17:19:35.725187 kubelet[2523]: I0904 17:19:35.724794 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phzfm\" (UniqueName: \"kubernetes.io/projected/2e2776ed-fa4e-45b8-bb3f-17de64345308-kube-api-access-phzfm\") pod \"csi-node-driver-nsxnh\" (UID: \"2e2776ed-fa4e-45b8-bb3f-17de64345308\") " pod="calico-system/csi-node-driver-nsxnh" Sep 4 17:19:35.725187 kubelet[2523]: I0904 17:19:35.724851 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2e2776ed-fa4e-45b8-bb3f-17de64345308-kubelet-dir\") pod \"csi-node-driver-nsxnh\" (UID: \"2e2776ed-fa4e-45b8-bb3f-17de64345308\") " pod="calico-system/csi-node-driver-nsxnh" Sep 4 17:19:35.725187 kubelet[2523]: I0904 17:19:35.724874 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2e2776ed-fa4e-45b8-bb3f-17de64345308-registration-dir\") pod \"csi-node-driver-nsxnh\" (UID: \"2e2776ed-fa4e-45b8-bb3f-17de64345308\") " pod="calico-system/csi-node-driver-nsxnh" Sep 4 17:19:35.725187 kubelet[2523]: I0904 17:19:35.724953 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/2e2776ed-fa4e-45b8-bb3f-17de64345308-varrun\") pod \"csi-node-driver-nsxnh\" (UID: \"2e2776ed-fa4e-45b8-bb3f-17de64345308\") " pod="calico-system/csi-node-driver-nsxnh" Sep 4 17:19:35.725532 containerd[1448]: time="2024-09-04T17:19:35.725489333Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-75f7c9c896-k4b7n,Uid:61e2b0bc-6e8f-437e-9d1c-dfc2e356162f,Namespace:calico-system,Attempt:0,}" Sep 4 17:19:35.726790 kubelet[2523]: E0904 17:19:35.726515 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.726790 kubelet[2523]: W0904 17:19:35.726552 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.726790 kubelet[2523]: E0904 17:19:35.726570 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.727243 kubelet[2523]: E0904 17:19:35.726815 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.727243 kubelet[2523]: W0904 17:19:35.726826 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.727243 kubelet[2523]: E0904 17:19:35.726859 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.728088 kubelet[2523]: E0904 17:19:35.728072 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.728088 kubelet[2523]: W0904 17:19:35.728088 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.728161 kubelet[2523]: E0904 17:19:35.728109 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.729650 kubelet[2523]: E0904 17:19:35.729546 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.729650 kubelet[2523]: W0904 17:19:35.729630 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.730513 kubelet[2523]: E0904 17:19:35.730412 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.731075 kubelet[2523]: E0904 17:19:35.731009 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.731075 kubelet[2523]: W0904 17:19:35.731023 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.731258 kubelet[2523]: E0904 17:19:35.731191 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.731658 kubelet[2523]: E0904 17:19:35.731501 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.731658 kubelet[2523]: W0904 17:19:35.731514 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.731959 kubelet[2523]: E0904 17:19:35.731940 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.732243 kubelet[2523]: E0904 17:19:35.732178 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.732243 kubelet[2523]: W0904 17:19:35.732190 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.732673 kubelet[2523]: E0904 17:19:35.732458 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.732889 kubelet[2523]: E0904 17:19:35.732871 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.733067 kubelet[2523]: W0904 17:19:35.733052 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.733182 kubelet[2523]: E0904 17:19:35.733162 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.733613 kubelet[2523]: E0904 17:19:35.733552 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.733613 kubelet[2523]: W0904 17:19:35.733592 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.733958 kubelet[2523]: E0904 17:19:35.733907 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.734686 kubelet[2523]: E0904 17:19:35.734553 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.734686 kubelet[2523]: W0904 17:19:35.734568 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.734807 kubelet[2523]: E0904 17:19:35.734794 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.735131 kubelet[2523]: E0904 17:19:35.735057 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.735131 kubelet[2523]: W0904 17:19:35.735069 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.735411 kubelet[2523]: E0904 17:19:35.735382 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.736240 kubelet[2523]: E0904 17:19:35.736087 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.736240 kubelet[2523]: W0904 17:19:35.736104 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.736633 kubelet[2523]: E0904 17:19:35.736363 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.737211 kubelet[2523]: E0904 17:19:35.736916 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.737211 kubelet[2523]: W0904 17:19:35.736932 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.737354 kubelet[2523]: E0904 17:19:35.737327 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.738825 kubelet[2523]: E0904 17:19:35.737855 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.738825 kubelet[2523]: W0904 17:19:35.737870 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.739491 kubelet[2523]: E0904 17:19:35.739464 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.739674 kubelet[2523]: E0904 17:19:35.739663 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.739870 kubelet[2523]: W0904 17:19:35.739823 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.740096 kubelet[2523]: E0904 17:19:35.740083 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.740288 kubelet[2523]: E0904 17:19:35.740276 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.740366 kubelet[2523]: W0904 17:19:35.740355 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.740496 kubelet[2523]: E0904 17:19:35.740486 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.744880 kubelet[2523]: E0904 17:19:35.744326 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.744986 kubelet[2523]: W0904 17:19:35.744969 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.745106 kubelet[2523]: E0904 17:19:35.745096 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.745705 kubelet[2523]: E0904 17:19:35.745688 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.745794 kubelet[2523]: W0904 17:19:35.745782 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.747945 kubelet[2523]: E0904 17:19:35.747864 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.748108 kubelet[2523]: E0904 17:19:35.748087 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.748108 kubelet[2523]: W0904 17:19:35.748106 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.748218 kubelet[2523]: E0904 17:19:35.748191 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.748288 kubelet[2523]: E0904 17:19:35.748273 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.748288 kubelet[2523]: W0904 17:19:35.748286 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.748386 kubelet[2523]: E0904 17:19:35.748366 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.749512 kubelet[2523]: E0904 17:19:35.748444 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.749512 kubelet[2523]: W0904 17:19:35.748453 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.749512 kubelet[2523]: E0904 17:19:35.748568 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.749512 kubelet[2523]: E0904 17:19:35.748640 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.749512 kubelet[2523]: W0904 17:19:35.748646 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.749512 kubelet[2523]: E0904 17:19:35.748709 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.749512 kubelet[2523]: E0904 17:19:35.748894 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.749512 kubelet[2523]: W0904 17:19:35.748902 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.749512 kubelet[2523]: E0904 17:19:35.748973 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.749512 kubelet[2523]: E0904 17:19:35.749045 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.749728 kubelet[2523]: W0904 17:19:35.749054 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.749728 kubelet[2523]: E0904 17:19:35.749121 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.749949 kubelet[2523]: E0904 17:19:35.749924 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.749949 kubelet[2523]: W0904 17:19:35.749942 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.750074 kubelet[2523]: E0904 17:19:35.750039 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.751098 kubelet[2523]: E0904 17:19:35.750631 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.751098 kubelet[2523]: W0904 17:19:35.750645 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.751098 kubelet[2523]: E0904 17:19:35.751057 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.752717 kubelet[2523]: E0904 17:19:35.751643 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.752717 kubelet[2523]: W0904 17:19:35.751661 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.752717 kubelet[2523]: E0904 17:19:35.751916 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.752717 kubelet[2523]: E0904 17:19:35.752149 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.752717 kubelet[2523]: W0904 17:19:35.752160 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.752717 kubelet[2523]: E0904 17:19:35.752237 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.752717 kubelet[2523]: E0904 17:19:35.752309 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.752717 kubelet[2523]: W0904 17:19:35.752315 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.752717 kubelet[2523]: E0904 17:19:35.752433 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.752717 kubelet[2523]: E0904 17:19:35.752508 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.753071 kubelet[2523]: W0904 17:19:35.752515 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.753071 kubelet[2523]: E0904 17:19:35.752617 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.753071 kubelet[2523]: E0904 17:19:35.752697 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.753071 kubelet[2523]: W0904 17:19:35.752705 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.753071 kubelet[2523]: E0904 17:19:35.752768 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.753071 kubelet[2523]: E0904 17:19:35.752918 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.753071 kubelet[2523]: W0904 17:19:35.752925 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.753071 kubelet[2523]: E0904 17:19:35.752976 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.753633 kubelet[2523]: E0904 17:19:35.753100 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.753633 kubelet[2523]: W0904 17:19:35.753107 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.753633 kubelet[2523]: E0904 17:19:35.753118 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.753633 kubelet[2523]: E0904 17:19:35.753278 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.753633 kubelet[2523]: W0904 17:19:35.753285 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.753633 kubelet[2523]: E0904 17:19:35.753301 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.753633 kubelet[2523]: E0904 17:19:35.753463 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.753633 kubelet[2523]: W0904 17:19:35.753478 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.753633 kubelet[2523]: E0904 17:19:35.753489 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.758058 containerd[1448]: time="2024-09-04T17:19:35.755944416Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 4 17:19:35.758058 containerd[1448]: time="2024-09-04T17:19:35.755996258Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:19:35.758058 containerd[1448]: time="2024-09-04T17:19:35.756012098Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 4 17:19:35.758058 containerd[1448]: time="2024-09-04T17:19:35.756022419Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:19:35.768006 kubelet[2523]: E0904 17:19:35.767904 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:35.769963 containerd[1448]: time="2024-09-04T17:19:35.769914433Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-ktcqt,Uid:1944397e-374a-4781-85a3-d9f5b73e16bd,Namespace:calico-system,Attempt:0,}" Sep 4 17:19:35.791009 systemd[1]: Started cri-containerd-a29ffde78a22902103a0e892eede4dc001867c0a5a4b6f6f7549760295dfcb8f.scope - libcontainer container a29ffde78a22902103a0e892eede4dc001867c0a5a4b6f6f7549760295dfcb8f. Sep 4 17:19:35.822096 containerd[1448]: time="2024-09-04T17:19:35.821912282Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 4 17:19:35.822096 containerd[1448]: time="2024-09-04T17:19:35.822004845Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:19:35.822096 containerd[1448]: time="2024-09-04T17:19:35.822019246Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 4 17:19:35.822096 containerd[1448]: time="2024-09-04T17:19:35.822028886Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:19:35.830628 kubelet[2523]: E0904 17:19:35.830047 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.830628 kubelet[2523]: W0904 17:19:35.830067 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.830628 kubelet[2523]: E0904 17:19:35.830089 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.830628 kubelet[2523]: E0904 17:19:35.830307 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.830628 kubelet[2523]: W0904 17:19:35.830314 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.830628 kubelet[2523]: E0904 17:19:35.830326 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.830628 kubelet[2523]: E0904 17:19:35.830522 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.830628 kubelet[2523]: W0904 17:19:35.830530 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.830628 kubelet[2523]: E0904 17:19:35.830544 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.830981 kubelet[2523]: E0904 17:19:35.830710 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.830981 kubelet[2523]: W0904 17:19:35.830716 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.830981 kubelet[2523]: E0904 17:19:35.830726 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.832568 kubelet[2523]: E0904 17:19:35.831377 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.832568 kubelet[2523]: W0904 17:19:35.831404 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.832568 kubelet[2523]: E0904 17:19:35.831426 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.832568 kubelet[2523]: E0904 17:19:35.831765 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.832568 kubelet[2523]: W0904 17:19:35.831774 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.832568 kubelet[2523]: E0904 17:19:35.831858 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.832568 kubelet[2523]: E0904 17:19:35.831937 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.832568 kubelet[2523]: W0904 17:19:35.831944 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.832568 kubelet[2523]: E0904 17:19:35.832014 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.833047 kubelet[2523]: E0904 17:19:35.833027 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.833047 kubelet[2523]: W0904 17:19:35.833042 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.833698 kubelet[2523]: E0904 17:19:35.833137 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.833909 kubelet[2523]: E0904 17:19:35.833891 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.833909 kubelet[2523]: W0904 17:19:35.833908 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.834048 kubelet[2523]: E0904 17:19:35.834023 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.834469 kubelet[2523]: E0904 17:19:35.834435 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.834469 kubelet[2523]: W0904 17:19:35.834448 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.835941 kubelet[2523]: E0904 17:19:35.834541 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.835941 kubelet[2523]: E0904 17:19:35.834696 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.835941 kubelet[2523]: W0904 17:19:35.834717 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.835941 kubelet[2523]: E0904 17:19:35.834774 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.835941 kubelet[2523]: E0904 17:19:35.834935 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.835941 kubelet[2523]: W0904 17:19:35.834943 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.835941 kubelet[2523]: E0904 17:19:35.835032 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.835941 kubelet[2523]: E0904 17:19:35.835142 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.835941 kubelet[2523]: W0904 17:19:35.835150 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.835941 kubelet[2523]: E0904 17:19:35.835196 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.836261 kubelet[2523]: E0904 17:19:35.835318 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.836261 kubelet[2523]: W0904 17:19:35.835324 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.836261 kubelet[2523]: E0904 17:19:35.835355 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.836261 kubelet[2523]: E0904 17:19:35.835456 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.836261 kubelet[2523]: W0904 17:19:35.835463 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.836261 kubelet[2523]: E0904 17:19:35.835484 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.836261 kubelet[2523]: E0904 17:19:35.835625 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.836261 kubelet[2523]: W0904 17:19:35.835632 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.836261 kubelet[2523]: E0904 17:19:35.835642 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.836261 kubelet[2523]: E0904 17:19:35.835888 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.837987 kubelet[2523]: W0904 17:19:35.835899 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.837987 kubelet[2523]: E0904 17:19:35.835918 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.837987 kubelet[2523]: E0904 17:19:35.836488 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.837987 kubelet[2523]: W0904 17:19:35.836505 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.837987 kubelet[2523]: E0904 17:19:35.836528 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.837987 kubelet[2523]: E0904 17:19:35.836943 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.837987 kubelet[2523]: W0904 17:19:35.836955 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.837987 kubelet[2523]: E0904 17:19:35.837215 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.838156 kubelet[2523]: E0904 17:19:35.838094 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.838156 kubelet[2523]: W0904 17:19:35.838109 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.838197 kubelet[2523]: E0904 17:19:35.838184 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.838886 kubelet[2523]: E0904 17:19:35.838315 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.838886 kubelet[2523]: W0904 17:19:35.838325 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.838886 kubelet[2523]: E0904 17:19:35.838360 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.838886 kubelet[2523]: E0904 17:19:35.838488 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.838886 kubelet[2523]: W0904 17:19:35.838496 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.838886 kubelet[2523]: E0904 17:19:35.838509 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.838886 kubelet[2523]: E0904 17:19:35.838690 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.838886 kubelet[2523]: W0904 17:19:35.838701 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.838886 kubelet[2523]: E0904 17:19:35.838715 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.839550 kubelet[2523]: E0904 17:19:35.838978 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.839550 kubelet[2523]: W0904 17:19:35.838988 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.839550 kubelet[2523]: E0904 17:19:35.839008 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.839550 kubelet[2523]: E0904 17:19:35.839290 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.839550 kubelet[2523]: W0904 17:19:35.839302 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.839550 kubelet[2523]: E0904 17:19:35.839314 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.845017 containerd[1448]: time="2024-09-04T17:19:35.842269046Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-75f7c9c896-k4b7n,Uid:61e2b0bc-6e8f-437e-9d1c-dfc2e356162f,Namespace:calico-system,Attempt:0,} returns sandbox id \"a29ffde78a22902103a0e892eede4dc001867c0a5a4b6f6f7549760295dfcb8f\"" Sep 4 17:19:35.845181 kubelet[2523]: E0904 17:19:35.845094 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:35.848106 containerd[1448]: time="2024-09-04T17:19:35.848066012Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.28.1\"" Sep 4 17:19:35.856508 systemd[1]: Started cri-containerd-916e13ec9143780cd728eb1baa330349e3be3655cc4e66b693d9828ba4733c87.scope - libcontainer container 916e13ec9143780cd728eb1baa330349e3be3655cc4e66b693d9828ba4733c87. Sep 4 17:19:35.859687 kubelet[2523]: E0904 17:19:35.859661 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:35.859687 kubelet[2523]: W0904 17:19:35.859681 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:35.859857 kubelet[2523]: E0904 17:19:35.859702 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:35.882256 containerd[1448]: time="2024-09-04T17:19:35.882212946Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-ktcqt,Uid:1944397e-374a-4781-85a3-d9f5b73e16bd,Namespace:calico-system,Attempt:0,} returns sandbox id \"916e13ec9143780cd728eb1baa330349e3be3655cc4e66b693d9828ba4733c87\"" Sep 4 17:19:35.883251 kubelet[2523]: E0904 17:19:35.883221 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:37.449065 containerd[1448]: time="2024-09-04T17:19:37.448995530Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.28.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:19:37.450036 containerd[1448]: time="2024-09-04T17:19:37.449962082Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.28.1: active requests=0, bytes read=27474479" Sep 4 17:19:37.451283 containerd[1448]: time="2024-09-04T17:19:37.450880792Z" level=info msg="ImageCreate event name:\"sha256:c1d0081df1580fc17ebf95ca7499d2e1af1b1ab8c75835172213221419018924\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:19:37.453555 containerd[1448]: time="2024-09-04T17:19:37.453161227Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:d97114d8e1e5186f1180fc8ef5f1309e0a8bf97efce35e0a0223d057d78d95fb\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:19:37.454072 containerd[1448]: time="2024-09-04T17:19:37.454043656Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.28.1\" with image id \"sha256:c1d0081df1580fc17ebf95ca7499d2e1af1b1ab8c75835172213221419018924\", repo tag \"ghcr.io/flatcar/calico/typha:v3.28.1\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:d97114d8e1e5186f1180fc8ef5f1309e0a8bf97efce35e0a0223d057d78d95fb\", size \"28841990\" in 1.605933643s" Sep 4 17:19:37.454227 containerd[1448]: time="2024-09-04T17:19:37.454151620Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.28.1\" returns image reference \"sha256:c1d0081df1580fc17ebf95ca7499d2e1af1b1ab8c75835172213221419018924\"" Sep 4 17:19:37.456058 containerd[1448]: time="2024-09-04T17:19:37.456015361Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.28.1\"" Sep 4 17:19:37.461641 containerd[1448]: time="2024-09-04T17:19:37.461591664Z" level=info msg="CreateContainer within sandbox \"a29ffde78a22902103a0e892eede4dc001867c0a5a4b6f6f7549760295dfcb8f\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Sep 4 17:19:37.477779 containerd[1448]: time="2024-09-04T17:19:37.477722274Z" level=info msg="CreateContainer within sandbox \"a29ffde78a22902103a0e892eede4dc001867c0a5a4b6f6f7549760295dfcb8f\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"125b41db0511427002299a947b1f0cd4f0022627ef8a2c35aa3c989673ce8934\"" Sep 4 17:19:37.479403 containerd[1448]: time="2024-09-04T17:19:37.478379376Z" level=info msg="StartContainer for \"125b41db0511427002299a947b1f0cd4f0022627ef8a2c35aa3c989673ce8934\"" Sep 4 17:19:37.504066 systemd[1]: Started cri-containerd-125b41db0511427002299a947b1f0cd4f0022627ef8a2c35aa3c989673ce8934.scope - libcontainer container 125b41db0511427002299a947b1f0cd4f0022627ef8a2c35aa3c989673ce8934. Sep 4 17:19:37.539461 containerd[1448]: time="2024-09-04T17:19:37.537923572Z" level=info msg="StartContainer for \"125b41db0511427002299a947b1f0cd4f0022627ef8a2c35aa3c989673ce8934\" returns successfully" Sep 4 17:19:37.713021 kubelet[2523]: E0904 17:19:37.712901 2523 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-nsxnh" podUID="2e2776ed-fa4e-45b8-bb3f-17de64345308" Sep 4 17:19:37.821550 kubelet[2523]: E0904 17:19:37.821488 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:37.847856 kubelet[2523]: E0904 17:19:37.845966 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:37.847856 kubelet[2523]: W0904 17:19:37.845987 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:37.847856 kubelet[2523]: E0904 17:19:37.846008 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:37.847856 kubelet[2523]: E0904 17:19:37.846542 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:37.847856 kubelet[2523]: W0904 17:19:37.846556 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:37.847856 kubelet[2523]: E0904 17:19:37.846572 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:37.847856 kubelet[2523]: E0904 17:19:37.846794 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:37.847856 kubelet[2523]: W0904 17:19:37.846805 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:37.847856 kubelet[2523]: E0904 17:19:37.846816 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:37.847856 kubelet[2523]: E0904 17:19:37.847153 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:37.849131 kubelet[2523]: W0904 17:19:37.847168 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:37.849131 kubelet[2523]: E0904 17:19:37.847180 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:37.849131 kubelet[2523]: E0904 17:19:37.847459 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:37.849131 kubelet[2523]: W0904 17:19:37.847469 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:37.849131 kubelet[2523]: E0904 17:19:37.847482 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:37.849131 kubelet[2523]: E0904 17:19:37.848059 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:37.849131 kubelet[2523]: W0904 17:19:37.848072 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:37.849131 kubelet[2523]: E0904 17:19:37.848086 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:37.849131 kubelet[2523]: E0904 17:19:37.848332 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:37.849131 kubelet[2523]: W0904 17:19:37.848341 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:37.849374 kubelet[2523]: E0904 17:19:37.848354 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:37.849374 kubelet[2523]: E0904 17:19:37.848516 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:37.849374 kubelet[2523]: W0904 17:19:37.848524 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:37.849374 kubelet[2523]: E0904 17:19:37.848540 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:37.849374 kubelet[2523]: E0904 17:19:37.848747 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:37.849374 kubelet[2523]: W0904 17:19:37.848756 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:37.849374 kubelet[2523]: E0904 17:19:37.848767 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:37.849374 kubelet[2523]: E0904 17:19:37.848926 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:37.849374 kubelet[2523]: W0904 17:19:37.848934 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:37.849374 kubelet[2523]: E0904 17:19:37.848944 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:37.849628 kubelet[2523]: E0904 17:19:37.849118 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:37.849628 kubelet[2523]: W0904 17:19:37.849126 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:37.849628 kubelet[2523]: E0904 17:19:37.849138 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:37.849628 kubelet[2523]: E0904 17:19:37.849280 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:37.849628 kubelet[2523]: W0904 17:19:37.849287 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:37.849628 kubelet[2523]: E0904 17:19:37.849296 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:37.849628 kubelet[2523]: E0904 17:19:37.849436 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:37.849628 kubelet[2523]: W0904 17:19:37.849443 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:37.849628 kubelet[2523]: E0904 17:19:37.849454 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:37.849628 kubelet[2523]: E0904 17:19:37.849606 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:37.849885 kubelet[2523]: W0904 17:19:37.849612 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:37.849885 kubelet[2523]: E0904 17:19:37.849621 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:37.849885 kubelet[2523]: E0904 17:19:37.849751 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:37.849885 kubelet[2523]: W0904 17:19:37.849758 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:37.849885 kubelet[2523]: E0904 17:19:37.849767 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:37.850187 kubelet[2523]: I0904 17:19:37.850134 2523 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="calico-system/calico-typha-75f7c9c896-k4b7n" podStartSLOduration=1.242494806 podStartE2EDuration="2.850097947s" podCreationTimestamp="2024-09-04 17:19:35 +0000 UTC" firstStartedPulling="2024-09-04 17:19:35.846822888 +0000 UTC m=+21.267736611" lastFinishedPulling="2024-09-04 17:19:37.454425989 +0000 UTC m=+22.875339752" observedRunningTime="2024-09-04 17:19:37.845159345 +0000 UTC m=+23.266073108" watchObservedRunningTime="2024-09-04 17:19:37.850097947 +0000 UTC m=+23.271011710" Sep 4 17:19:37.945323 kubelet[2523]: E0904 17:19:37.945272 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:37.945323 kubelet[2523]: W0904 17:19:37.945297 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:37.945323 kubelet[2523]: E0904 17:19:37.945319 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:37.945560 kubelet[2523]: E0904 17:19:37.945541 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:37.945560 kubelet[2523]: W0904 17:19:37.945557 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:37.945626 kubelet[2523]: E0904 17:19:37.945578 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:37.945860 kubelet[2523]: E0904 17:19:37.945812 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:37.945860 kubelet[2523]: W0904 17:19:37.945825 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:37.945860 kubelet[2523]: E0904 17:19:37.945862 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:37.946127 kubelet[2523]: E0904 17:19:37.946096 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:37.946127 kubelet[2523]: W0904 17:19:37.946109 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:37.946127 kubelet[2523]: E0904 17:19:37.946125 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:37.946356 kubelet[2523]: E0904 17:19:37.946330 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:37.946356 kubelet[2523]: W0904 17:19:37.946341 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:37.946356 kubelet[2523]: E0904 17:19:37.946357 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:37.946545 kubelet[2523]: E0904 17:19:37.946510 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:37.946545 kubelet[2523]: W0904 17:19:37.946522 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:37.946545 kubelet[2523]: E0904 17:19:37.946544 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:37.946946 kubelet[2523]: E0904 17:19:37.946910 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:37.946946 kubelet[2523]: W0904 17:19:37.946932 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:37.947033 kubelet[2523]: E0904 17:19:37.946954 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:37.948718 kubelet[2523]: E0904 17:19:37.948685 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:37.948718 kubelet[2523]: W0904 17:19:37.948701 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:37.948850 kubelet[2523]: E0904 17:19:37.948743 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:37.949023 kubelet[2523]: E0904 17:19:37.948988 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:37.949023 kubelet[2523]: W0904 17:19:37.949003 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:37.949100 kubelet[2523]: E0904 17:19:37.949047 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:37.949205 kubelet[2523]: E0904 17:19:37.949178 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:37.949205 kubelet[2523]: W0904 17:19:37.949190 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:37.949264 kubelet[2523]: E0904 17:19:37.949227 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:37.949400 kubelet[2523]: E0904 17:19:37.949365 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:37.949400 kubelet[2523]: W0904 17:19:37.949380 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:37.949400 kubelet[2523]: E0904 17:19:37.949398 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:37.949641 kubelet[2523]: E0904 17:19:37.949609 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:37.949641 kubelet[2523]: W0904 17:19:37.949622 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:37.949641 kubelet[2523]: E0904 17:19:37.949638 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:37.949842 kubelet[2523]: E0904 17:19:37.949807 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:37.949842 kubelet[2523]: W0904 17:19:37.949821 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:37.949946 kubelet[2523]: E0904 17:19:37.949851 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:37.950036 kubelet[2523]: E0904 17:19:37.950012 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:37.950036 kubelet[2523]: W0904 17:19:37.950024 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:37.950036 kubelet[2523]: E0904 17:19:37.950040 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:37.950237 kubelet[2523]: E0904 17:19:37.950215 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:37.950237 kubelet[2523]: W0904 17:19:37.950225 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:37.950237 kubelet[2523]: E0904 17:19:37.950239 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:37.950439 kubelet[2523]: E0904 17:19:37.950413 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:37.950439 kubelet[2523]: W0904 17:19:37.950424 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:37.950439 kubelet[2523]: E0904 17:19:37.950440 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:37.950773 kubelet[2523]: E0904 17:19:37.950734 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:37.950773 kubelet[2523]: W0904 17:19:37.950754 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:37.950773 kubelet[2523]: E0904 17:19:37.950775 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:37.951018 kubelet[2523]: E0904 17:19:37.950984 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:37.951018 kubelet[2523]: W0904 17:19:37.950997 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:37.951018 kubelet[2523]: E0904 17:19:37.951010 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:38.745596 containerd[1448]: time="2024-09-04T17:19:38.745538648Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.28.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:19:38.749453 containerd[1448]: time="2024-09-04T17:19:38.749278406Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.28.1: active requests=0, bytes read=4916957" Sep 4 17:19:38.750362 containerd[1448]: time="2024-09-04T17:19:38.750321239Z" level=info msg="ImageCreate event name:\"sha256:20b54f73684933653d4a4b8b63c59211e3c828f94251ecf4d1bff2a334ff4ba0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:19:38.766749 containerd[1448]: time="2024-09-04T17:19:38.766693036Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:7938ad0cb2b49a32937962cc40dd826ad5858999c603bdf5fbf2092a4d50cf01\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:19:38.769236 containerd[1448]: time="2024-09-04T17:19:38.769182315Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.28.1\" with image id \"sha256:20b54f73684933653d4a4b8b63c59211e3c828f94251ecf4d1bff2a334ff4ba0\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.28.1\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:7938ad0cb2b49a32937962cc40dd826ad5858999c603bdf5fbf2092a4d50cf01\", size \"6284436\" in 1.313126513s" Sep 4 17:19:38.769236 containerd[1448]: time="2024-09-04T17:19:38.769232437Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.28.1\" returns image reference \"sha256:20b54f73684933653d4a4b8b63c59211e3c828f94251ecf4d1bff2a334ff4ba0\"" Sep 4 17:19:38.776023 containerd[1448]: time="2024-09-04T17:19:38.775787964Z" level=info msg="CreateContainer within sandbox \"916e13ec9143780cd728eb1baa330349e3be3655cc4e66b693d9828ba4733c87\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Sep 4 17:19:38.822772 kubelet[2523]: I0904 17:19:38.822735 2523 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 4 17:19:38.823422 kubelet[2523]: E0904 17:19:38.823399 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:38.856577 containerd[1448]: time="2024-09-04T17:19:38.856513957Z" level=info msg="CreateContainer within sandbox \"916e13ec9143780cd728eb1baa330349e3be3655cc4e66b693d9828ba4733c87\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"034aa31196b517a7af057164791bb9a69161c9392d76d9d99b3c6ca4c8fe90bc\"" Sep 4 17:19:38.857001 kubelet[2523]: E0904 17:19:38.856905 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:38.857001 kubelet[2523]: W0904 17:19:38.856930 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:38.857001 kubelet[2523]: E0904 17:19:38.856954 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:38.857578 containerd[1448]: time="2024-09-04T17:19:38.857504028Z" level=info msg="StartContainer for \"034aa31196b517a7af057164791bb9a69161c9392d76d9d99b3c6ca4c8fe90bc\"" Sep 4 17:19:38.858114 kubelet[2523]: E0904 17:19:38.858093 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:38.858114 kubelet[2523]: W0904 17:19:38.858110 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:38.858212 kubelet[2523]: E0904 17:19:38.858127 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:38.858435 kubelet[2523]: E0904 17:19:38.858318 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:38.858435 kubelet[2523]: W0904 17:19:38.858327 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:38.858435 kubelet[2523]: E0904 17:19:38.858338 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:38.859531 kubelet[2523]: E0904 17:19:38.858698 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:38.859531 kubelet[2523]: W0904 17:19:38.858709 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:38.859531 kubelet[2523]: E0904 17:19:38.858722 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:38.859531 kubelet[2523]: E0904 17:19:38.859150 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:38.859531 kubelet[2523]: W0904 17:19:38.859167 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:38.859531 kubelet[2523]: E0904 17:19:38.859202 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:38.859531 kubelet[2523]: E0904 17:19:38.859404 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:38.859531 kubelet[2523]: W0904 17:19:38.859414 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:38.859531 kubelet[2523]: E0904 17:19:38.859427 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:38.859825 kubelet[2523]: E0904 17:19:38.859602 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:38.859825 kubelet[2523]: W0904 17:19:38.859610 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:38.859825 kubelet[2523]: E0904 17:19:38.859620 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:38.859825 kubelet[2523]: E0904 17:19:38.859771 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:38.859825 kubelet[2523]: W0904 17:19:38.859778 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:38.859825 kubelet[2523]: E0904 17:19:38.859790 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:38.861002 kubelet[2523]: E0904 17:19:38.859954 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:38.861002 kubelet[2523]: W0904 17:19:38.859962 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:38.861002 kubelet[2523]: E0904 17:19:38.859972 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:38.861002 kubelet[2523]: E0904 17:19:38.860096 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:38.861002 kubelet[2523]: W0904 17:19:38.860102 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:38.861002 kubelet[2523]: E0904 17:19:38.860112 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:38.861002 kubelet[2523]: E0904 17:19:38.860305 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:38.861002 kubelet[2523]: W0904 17:19:38.860314 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:38.861002 kubelet[2523]: E0904 17:19:38.860326 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:38.861002 kubelet[2523]: E0904 17:19:38.860485 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:38.861271 kubelet[2523]: W0904 17:19:38.860493 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:38.861271 kubelet[2523]: E0904 17:19:38.860504 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:38.861271 kubelet[2523]: E0904 17:19:38.860701 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:38.861271 kubelet[2523]: W0904 17:19:38.860711 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:38.861271 kubelet[2523]: E0904 17:19:38.860724 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:38.861688 kubelet[2523]: E0904 17:19:38.861513 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:38.861688 kubelet[2523]: W0904 17:19:38.861525 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:38.861688 kubelet[2523]: E0904 17:19:38.861556 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:38.861688 kubelet[2523]: E0904 17:19:38.861752 2523 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Sep 4 17:19:38.861688 kubelet[2523]: W0904 17:19:38.861761 2523 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Sep 4 17:19:38.861688 kubelet[2523]: E0904 17:19:38.861789 2523 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Sep 4 17:19:38.888108 systemd[1]: Started cri-containerd-034aa31196b517a7af057164791bb9a69161c9392d76d9d99b3c6ca4c8fe90bc.scope - libcontainer container 034aa31196b517a7af057164791bb9a69161c9392d76d9d99b3c6ca4c8fe90bc. Sep 4 17:19:38.917157 containerd[1448]: time="2024-09-04T17:19:38.917107713Z" level=info msg="StartContainer for \"034aa31196b517a7af057164791bb9a69161c9392d76d9d99b3c6ca4c8fe90bc\" returns successfully" Sep 4 17:19:38.952448 systemd[1]: cri-containerd-034aa31196b517a7af057164791bb9a69161c9392d76d9d99b3c6ca4c8fe90bc.scope: Deactivated successfully. Sep 4 17:19:38.982522 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-034aa31196b517a7af057164791bb9a69161c9392d76d9d99b3c6ca4c8fe90bc-rootfs.mount: Deactivated successfully. Sep 4 17:19:39.000826 containerd[1448]: time="2024-09-04T17:19:38.992136045Z" level=info msg="shim disconnected" id=034aa31196b517a7af057164791bb9a69161c9392d76d9d99b3c6ca4c8fe90bc namespace=k8s.io Sep 4 17:19:39.000826 containerd[1448]: time="2024-09-04T17:19:39.000742157Z" level=warning msg="cleaning up after shim disconnected" id=034aa31196b517a7af057164791bb9a69161c9392d76d9d99b3c6ca4c8fe90bc namespace=k8s.io Sep 4 17:19:39.000826 containerd[1448]: time="2024-09-04T17:19:39.000759438Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 4 17:19:39.713726 kubelet[2523]: E0904 17:19:39.713679 2523 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-nsxnh" podUID="2e2776ed-fa4e-45b8-bb3f-17de64345308" Sep 4 17:19:39.826834 kubelet[2523]: E0904 17:19:39.826788 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:39.828087 containerd[1448]: time="2024-09-04T17:19:39.827864678Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.28.1\"" Sep 4 17:19:41.713149 kubelet[2523]: E0904 17:19:41.713093 2523 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-nsxnh" podUID="2e2776ed-fa4e-45b8-bb3f-17de64345308" Sep 4 17:19:43.227099 containerd[1448]: time="2024-09-04T17:19:43.227020415Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.28.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:19:43.228117 containerd[1448]: time="2024-09-04T17:19:43.227644232Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.28.1: active requests=0, bytes read=86859887" Sep 4 17:19:43.229870 containerd[1448]: time="2024-09-04T17:19:43.228688900Z" level=info msg="ImageCreate event name:\"sha256:6123e515001d9cafdf3dbe8f8dc8b5ae1c56165013052b8cbc7d27f3395cfd85\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:19:43.231985 containerd[1448]: time="2024-09-04T17:19:43.231932226Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:1cf32b2159ec9f938e747b82b9b7c74e26e17eb220e002a6a1bd6b5b1266e1fa\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:19:43.232608 containerd[1448]: time="2024-09-04T17:19:43.232574803Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.28.1\" with image id \"sha256:6123e515001d9cafdf3dbe8f8dc8b5ae1c56165013052b8cbc7d27f3395cfd85\", repo tag \"ghcr.io/flatcar/calico/cni:v3.28.1\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:1cf32b2159ec9f938e747b82b9b7c74e26e17eb220e002a6a1bd6b5b1266e1fa\", size \"88227406\" in 3.404675523s" Sep 4 17:19:43.232608 containerd[1448]: time="2024-09-04T17:19:43.232604363Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.28.1\" returns image reference \"sha256:6123e515001d9cafdf3dbe8f8dc8b5ae1c56165013052b8cbc7d27f3395cfd85\"" Sep 4 17:19:43.236225 containerd[1448]: time="2024-09-04T17:19:43.236185058Z" level=info msg="CreateContainer within sandbox \"916e13ec9143780cd728eb1baa330349e3be3655cc4e66b693d9828ba4733c87\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Sep 4 17:19:43.247993 containerd[1448]: time="2024-09-04T17:19:43.247811367Z" level=info msg="CreateContainer within sandbox \"916e13ec9143780cd728eb1baa330349e3be3655cc4e66b693d9828ba4733c87\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"e8e657473e736905ec1edcfedeacafe9f24fef9b7e1b7d3ba374c90ec5c1f1cb\"" Sep 4 17:19:43.248573 containerd[1448]: time="2024-09-04T17:19:43.248543386Z" level=info msg="StartContainer for \"e8e657473e736905ec1edcfedeacafe9f24fef9b7e1b7d3ba374c90ec5c1f1cb\"" Sep 4 17:19:43.278031 systemd[1]: Started cri-containerd-e8e657473e736905ec1edcfedeacafe9f24fef9b7e1b7d3ba374c90ec5c1f1cb.scope - libcontainer container e8e657473e736905ec1edcfedeacafe9f24fef9b7e1b7d3ba374c90ec5c1f1cb. Sep 4 17:19:43.304912 containerd[1448]: time="2024-09-04T17:19:43.304867081Z" level=info msg="StartContainer for \"e8e657473e736905ec1edcfedeacafe9f24fef9b7e1b7d3ba374c90ec5c1f1cb\" returns successfully" Sep 4 17:19:43.713106 kubelet[2523]: E0904 17:19:43.713067 2523 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-nsxnh" podUID="2e2776ed-fa4e-45b8-bb3f-17de64345308" Sep 4 17:19:43.851318 kubelet[2523]: E0904 17:19:43.850755 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:43.863093 systemd[1]: cri-containerd-e8e657473e736905ec1edcfedeacafe9f24fef9b7e1b7d3ba374c90ec5c1f1cb.scope: Deactivated successfully. Sep 4 17:19:43.886469 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e8e657473e736905ec1edcfedeacafe9f24fef9b7e1b7d3ba374c90ec5c1f1cb-rootfs.mount: Deactivated successfully. Sep 4 17:19:43.936971 kubelet[2523]: I0904 17:19:43.936581 2523 kubelet_node_status.go:497] "Fast updating node status as it just became ready" Sep 4 17:19:43.964383 kubelet[2523]: I0904 17:19:43.964217 2523 topology_manager.go:215] "Topology Admit Handler" podUID="5962684c-02f2-4d5e-87f8-a103e98d7b84" podNamespace="kube-system" podName="coredns-76f75df574-tjcbc" Sep 4 17:19:43.967017 kubelet[2523]: I0904 17:19:43.966953 2523 topology_manager.go:215] "Topology Admit Handler" podUID="eccc4d8b-807c-4895-8200-d4b15b8caceb" podNamespace="kube-system" podName="coredns-76f75df574-kldz2" Sep 4 17:19:43.967605 kubelet[2523]: I0904 17:19:43.967511 2523 topology_manager.go:215] "Topology Admit Handler" podUID="05f65067-932e-4656-a9e0-3f22a268ebd5" podNamespace="calico-system" podName="calico-kube-controllers-7ff458757c-hfq6x" Sep 4 17:19:43.974735 systemd[1]: Created slice kubepods-burstable-pod5962684c_02f2_4d5e_87f8_a103e98d7b84.slice - libcontainer container kubepods-burstable-pod5962684c_02f2_4d5e_87f8_a103e98d7b84.slice. Sep 4 17:19:43.982697 systemd[1]: Created slice kubepods-burstable-podeccc4d8b_807c_4895_8200_d4b15b8caceb.slice - libcontainer container kubepods-burstable-podeccc4d8b_807c_4895_8200_d4b15b8caceb.slice. Sep 4 17:19:43.988083 systemd[1]: Created slice kubepods-besteffort-pod05f65067_932e_4656_a9e0_3f22a268ebd5.slice - libcontainer container kubepods-besteffort-pod05f65067_932e_4656_a9e0_3f22a268ebd5.slice. Sep 4 17:19:44.014652 kubelet[2523]: I0904 17:19:44.014600 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eccc4d8b-807c-4895-8200-d4b15b8caceb-config-volume\") pod \"coredns-76f75df574-kldz2\" (UID: \"eccc4d8b-807c-4895-8200-d4b15b8caceb\") " pod="kube-system/coredns-76f75df574-kldz2" Sep 4 17:19:44.014652 kubelet[2523]: I0904 17:19:44.014650 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ckss\" (UniqueName: \"kubernetes.io/projected/eccc4d8b-807c-4895-8200-d4b15b8caceb-kube-api-access-9ckss\") pod \"coredns-76f75df574-kldz2\" (UID: \"eccc4d8b-807c-4895-8200-d4b15b8caceb\") " pod="kube-system/coredns-76f75df574-kldz2" Sep 4 17:19:44.014825 kubelet[2523]: I0904 17:19:44.014679 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zh789\" (UniqueName: \"kubernetes.io/projected/05f65067-932e-4656-a9e0-3f22a268ebd5-kube-api-access-zh789\") pod \"calico-kube-controllers-7ff458757c-hfq6x\" (UID: \"05f65067-932e-4656-a9e0-3f22a268ebd5\") " pod="calico-system/calico-kube-controllers-7ff458757c-hfq6x" Sep 4 17:19:44.014825 kubelet[2523]: I0904 17:19:44.014799 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6pgn\" (UniqueName: \"kubernetes.io/projected/5962684c-02f2-4d5e-87f8-a103e98d7b84-kube-api-access-s6pgn\") pod \"coredns-76f75df574-tjcbc\" (UID: \"5962684c-02f2-4d5e-87f8-a103e98d7b84\") " pod="kube-system/coredns-76f75df574-tjcbc" Sep 4 17:19:44.014901 kubelet[2523]: I0904 17:19:44.014853 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05f65067-932e-4656-a9e0-3f22a268ebd5-tigera-ca-bundle\") pod \"calico-kube-controllers-7ff458757c-hfq6x\" (UID: \"05f65067-932e-4656-a9e0-3f22a268ebd5\") " pod="calico-system/calico-kube-controllers-7ff458757c-hfq6x" Sep 4 17:19:44.014901 kubelet[2523]: I0904 17:19:44.014878 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5962684c-02f2-4d5e-87f8-a103e98d7b84-config-volume\") pod \"coredns-76f75df574-tjcbc\" (UID: \"5962684c-02f2-4d5e-87f8-a103e98d7b84\") " pod="kube-system/coredns-76f75df574-tjcbc" Sep 4 17:19:44.019518 containerd[1448]: time="2024-09-04T17:19:44.019459226Z" level=info msg="shim disconnected" id=e8e657473e736905ec1edcfedeacafe9f24fef9b7e1b7d3ba374c90ec5c1f1cb namespace=k8s.io Sep 4 17:19:44.019518 containerd[1448]: time="2024-09-04T17:19:44.019516467Z" level=warning msg="cleaning up after shim disconnected" id=e8e657473e736905ec1edcfedeacafe9f24fef9b7e1b7d3ba374c90ec5c1f1cb namespace=k8s.io Sep 4 17:19:44.019704 containerd[1448]: time="2024-09-04T17:19:44.019526347Z" level=info msg="cleaning up dead shim" namespace=k8s.io Sep 4 17:19:44.279641 kubelet[2523]: E0904 17:19:44.279241 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:44.280482 containerd[1448]: time="2024-09-04T17:19:44.280374251Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-76f75df574-tjcbc,Uid:5962684c-02f2-4d5e-87f8-a103e98d7b84,Namespace:kube-system,Attempt:0,}" Sep 4 17:19:44.286002 kubelet[2523]: E0904 17:19:44.285979 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:44.286452 containerd[1448]: time="2024-09-04T17:19:44.286397165Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-76f75df574-kldz2,Uid:eccc4d8b-807c-4895-8200-d4b15b8caceb,Namespace:kube-system,Attempt:0,}" Sep 4 17:19:44.292013 containerd[1448]: time="2024-09-04T17:19:44.291946868Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7ff458757c-hfq6x,Uid:05f65067-932e-4656-a9e0-3f22a268ebd5,Namespace:calico-system,Attempt:0,}" Sep 4 17:19:44.695119 containerd[1448]: time="2024-09-04T17:19:44.695063707Z" level=error msg="Failed to destroy network for sandbox \"53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:19:44.695817 containerd[1448]: time="2024-09-04T17:19:44.695770245Z" level=error msg="encountered an error cleaning up failed sandbox \"53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:19:44.695977 containerd[1448]: time="2024-09-04T17:19:44.695941690Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-76f75df574-kldz2,Uid:eccc4d8b-807c-4895-8200-d4b15b8caceb,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:19:44.696667 containerd[1448]: time="2024-09-04T17:19:44.696637188Z" level=error msg="Failed to destroy network for sandbox \"0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:19:44.697083 containerd[1448]: time="2024-09-04T17:19:44.697051678Z" level=error msg="encountered an error cleaning up failed sandbox \"0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:19:44.697138 containerd[1448]: time="2024-09-04T17:19:44.697114280Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7ff458757c-hfq6x,Uid:05f65067-932e-4656-a9e0-3f22a268ebd5,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:19:44.699565 containerd[1448]: time="2024-09-04T17:19:44.699500981Z" level=error msg="Failed to destroy network for sandbox \"652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:19:44.700333 containerd[1448]: time="2024-09-04T17:19:44.700299522Z" level=error msg="encountered an error cleaning up failed sandbox \"652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:19:44.700380 containerd[1448]: time="2024-09-04T17:19:44.700363723Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-76f75df574-tjcbc,Uid:5962684c-02f2-4d5e-87f8-a103e98d7b84,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:19:44.701012 kubelet[2523]: E0904 17:19:44.700979 2523 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:19:44.701070 kubelet[2523]: E0904 17:19:44.701057 2523 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-7ff458757c-hfq6x" Sep 4 17:19:44.701096 kubelet[2523]: E0904 17:19:44.701078 2523 kuberuntime_manager.go:1172] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-7ff458757c-hfq6x" Sep 4 17:19:44.701096 kubelet[2523]: E0904 17:19:44.701078 2523 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:19:44.701158 kubelet[2523]: E0904 17:19:44.701116 2523 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-76f75df574-tjcbc" Sep 4 17:19:44.701158 kubelet[2523]: E0904 17:19:44.701131 2523 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-7ff458757c-hfq6x_calico-system(05f65067-932e-4656-a9e0-3f22a268ebd5)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-7ff458757c-hfq6x_calico-system(05f65067-932e-4656-a9e0-3f22a268ebd5)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-7ff458757c-hfq6x" podUID="05f65067-932e-4656-a9e0-3f22a268ebd5" Sep 4 17:19:44.701158 kubelet[2523]: E0904 17:19:44.701134 2523 kuberuntime_manager.go:1172] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-76f75df574-tjcbc" Sep 4 17:19:44.701256 kubelet[2523]: E0904 17:19:44.701181 2523 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-76f75df574-tjcbc_kube-system(5962684c-02f2-4d5e-87f8-a103e98d7b84)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-76f75df574-tjcbc_kube-system(5962684c-02f2-4d5e-87f8-a103e98d7b84)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-76f75df574-tjcbc" podUID="5962684c-02f2-4d5e-87f8-a103e98d7b84" Sep 4 17:19:44.701256 kubelet[2523]: E0904 17:19:44.701218 2523 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:19:44.701315 kubelet[2523]: E0904 17:19:44.701261 2523 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-76f75df574-kldz2" Sep 4 17:19:44.701315 kubelet[2523]: E0904 17:19:44.701280 2523 kuberuntime_manager.go:1172] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-76f75df574-kldz2" Sep 4 17:19:44.701360 kubelet[2523]: E0904 17:19:44.701322 2523 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-76f75df574-kldz2_kube-system(eccc4d8b-807c-4895-8200-d4b15b8caceb)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-76f75df574-kldz2_kube-system(eccc4d8b-807c-4895-8200-d4b15b8caceb)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-76f75df574-kldz2" podUID="eccc4d8b-807c-4895-8200-d4b15b8caceb" Sep 4 17:19:44.857185 kubelet[2523]: E0904 17:19:44.857003 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:44.857994 containerd[1448]: time="2024-09-04T17:19:44.857615325Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.28.1\"" Sep 4 17:19:44.859894 kubelet[2523]: I0904 17:19:44.859871 2523 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" Sep 4 17:19:44.862197 containerd[1448]: time="2024-09-04T17:19:44.860912449Z" level=info msg="StopPodSandbox for \"0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be\"" Sep 4 17:19:44.862197 containerd[1448]: time="2024-09-04T17:19:44.861228817Z" level=info msg="Ensure that sandbox 0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be in task-service has been cleanup successfully" Sep 4 17:19:44.864745 kubelet[2523]: I0904 17:19:44.863973 2523 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" Sep 4 17:19:44.867852 containerd[1448]: time="2024-09-04T17:19:44.864995234Z" level=info msg="StopPodSandbox for \"53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2\"" Sep 4 17:19:44.867945 kubelet[2523]: I0904 17:19:44.866981 2523 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" Sep 4 17:19:44.868224 containerd[1448]: time="2024-09-04T17:19:44.868197956Z" level=info msg="Ensure that sandbox 53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2 in task-service has been cleanup successfully" Sep 4 17:19:44.868401 containerd[1448]: time="2024-09-04T17:19:44.867395576Z" level=info msg="StopPodSandbox for \"652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2\"" Sep 4 17:19:44.868600 containerd[1448]: time="2024-09-04T17:19:44.868582846Z" level=info msg="Ensure that sandbox 652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2 in task-service has been cleanup successfully" Sep 4 17:19:44.894055 containerd[1448]: time="2024-09-04T17:19:44.893952218Z" level=error msg="StopPodSandbox for \"0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be\" failed" error="failed to destroy network for sandbox \"0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:19:44.904373 kubelet[2523]: E0904 17:19:44.904328 2523 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" Sep 4 17:19:44.904482 kubelet[2523]: E0904 17:19:44.904440 2523 kuberuntime_manager.go:1381] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be"} Sep 4 17:19:44.904482 kubelet[2523]: E0904 17:19:44.904477 2523 kuberuntime_manager.go:1081] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"05f65067-932e-4656-a9e0-3f22a268ebd5\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Sep 4 17:19:44.904574 kubelet[2523]: E0904 17:19:44.904506 2523 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"05f65067-932e-4656-a9e0-3f22a268ebd5\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-7ff458757c-hfq6x" podUID="05f65067-932e-4656-a9e0-3f22a268ebd5" Sep 4 17:19:44.910841 containerd[1448]: time="2024-09-04T17:19:44.910768090Z" level=error msg="StopPodSandbox for \"652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2\" failed" error="failed to destroy network for sandbox \"652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:19:44.911269 kubelet[2523]: E0904 17:19:44.911227 2523 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" Sep 4 17:19:44.911329 kubelet[2523]: E0904 17:19:44.911276 2523 kuberuntime_manager.go:1381] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2"} Sep 4 17:19:44.911329 kubelet[2523]: E0904 17:19:44.911313 2523 kuberuntime_manager.go:1081] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"5962684c-02f2-4d5e-87f8-a103e98d7b84\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Sep 4 17:19:44.911401 kubelet[2523]: E0904 17:19:44.911352 2523 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"5962684c-02f2-4d5e-87f8-a103e98d7b84\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-76f75df574-tjcbc" podUID="5962684c-02f2-4d5e-87f8-a103e98d7b84" Sep 4 17:19:44.915763 containerd[1448]: time="2024-09-04T17:19:44.915694817Z" level=error msg="StopPodSandbox for \"53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2\" failed" error="failed to destroy network for sandbox \"53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:19:44.916025 kubelet[2523]: E0904 17:19:44.915990 2523 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" Sep 4 17:19:44.916083 kubelet[2523]: E0904 17:19:44.916037 2523 kuberuntime_manager.go:1381] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2"} Sep 4 17:19:44.916083 kubelet[2523]: E0904 17:19:44.916071 2523 kuberuntime_manager.go:1081] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"eccc4d8b-807c-4895-8200-d4b15b8caceb\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Sep 4 17:19:44.916157 kubelet[2523]: E0904 17:19:44.916097 2523 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"eccc4d8b-807c-4895-8200-d4b15b8caceb\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-76f75df574-kldz2" podUID="eccc4d8b-807c-4895-8200-d4b15b8caceb" Sep 4 17:19:45.244079 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2-shm.mount: Deactivated successfully. Sep 4 17:19:45.244176 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2-shm.mount: Deactivated successfully. Sep 4 17:19:45.719885 systemd[1]: Created slice kubepods-besteffort-pod2e2776ed_fa4e_45b8_bb3f_17de64345308.slice - libcontainer container kubepods-besteffort-pod2e2776ed_fa4e_45b8_bb3f_17de64345308.slice. Sep 4 17:19:45.723118 containerd[1448]: time="2024-09-04T17:19:45.723045039Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-nsxnh,Uid:2e2776ed-fa4e-45b8-bb3f-17de64345308,Namespace:calico-system,Attempt:0,}" Sep 4 17:19:45.825531 containerd[1448]: time="2024-09-04T17:19:45.824931218Z" level=error msg="Failed to destroy network for sandbox \"ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:19:45.826192 containerd[1448]: time="2024-09-04T17:19:45.825886761Z" level=error msg="encountered an error cleaning up failed sandbox \"ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:19:45.826192 containerd[1448]: time="2024-09-04T17:19:45.825941203Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-nsxnh,Uid:2e2776ed-fa4e-45b8-bb3f-17de64345308,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:19:45.826583 kubelet[2523]: E0904 17:19:45.826408 2523 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:19:45.826583 kubelet[2523]: E0904 17:19:45.826466 2523 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-nsxnh" Sep 4 17:19:45.826583 kubelet[2523]: E0904 17:19:45.826489 2523 kuberuntime_manager.go:1172] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-nsxnh" Sep 4 17:19:45.826754 kubelet[2523]: E0904 17:19:45.826546 2523 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-nsxnh_calico-system(2e2776ed-fa4e-45b8-bb3f-17de64345308)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-nsxnh_calico-system(2e2776ed-fa4e-45b8-bb3f-17de64345308)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-nsxnh" podUID="2e2776ed-fa4e-45b8-bb3f-17de64345308" Sep 4 17:19:45.828394 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954-shm.mount: Deactivated successfully. Sep 4 17:19:45.883245 kubelet[2523]: I0904 17:19:45.883033 2523 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" Sep 4 17:19:45.884664 containerd[1448]: time="2024-09-04T17:19:45.884604624Z" level=info msg="StopPodSandbox for \"ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954\"" Sep 4 17:19:45.884890 containerd[1448]: time="2024-09-04T17:19:45.884866151Z" level=info msg="Ensure that sandbox ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954 in task-service has been cleanup successfully" Sep 4 17:19:45.933438 containerd[1448]: time="2024-09-04T17:19:45.933370799Z" level=error msg="StopPodSandbox for \"ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954\" failed" error="failed to destroy network for sandbox \"ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Sep 4 17:19:45.933674 kubelet[2523]: E0904 17:19:45.933642 2523 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" Sep 4 17:19:45.933740 kubelet[2523]: E0904 17:19:45.933690 2523 kuberuntime_manager.go:1381] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954"} Sep 4 17:19:45.933740 kubelet[2523]: E0904 17:19:45.933737 2523 kuberuntime_manager.go:1081] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"2e2776ed-fa4e-45b8-bb3f-17de64345308\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Sep 4 17:19:45.933817 kubelet[2523]: E0904 17:19:45.933769 2523 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"2e2776ed-fa4e-45b8-bb3f-17de64345308\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-nsxnh" podUID="2e2776ed-fa4e-45b8-bb3f-17de64345308" Sep 4 17:19:47.616364 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1520993936.mount: Deactivated successfully. Sep 4 17:19:47.820654 containerd[1448]: time="2024-09-04T17:19:47.820292273Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.28.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:19:47.821163 containerd[1448]: time="2024-09-04T17:19:47.821126413Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.28.1: active requests=0, bytes read=113057300" Sep 4 17:19:47.822639 containerd[1448]: time="2024-09-04T17:19:47.822593127Z" level=info msg="ImageCreate event name:\"sha256:373272045e41e00ebf8da7ce9fc6b26d326fb8b3e665d9f78bb121976f83b1dc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:19:47.829536 containerd[1448]: time="2024-09-04T17:19:47.829487049Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:47908d8b3046dadd6fbea273ac5b0b9bb803cc7b58b9114c50bf7591767d2744\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:19:47.830190 containerd[1448]: time="2024-09-04T17:19:47.830061263Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.28.1\" with image id \"sha256:373272045e41e00ebf8da7ce9fc6b26d326fb8b3e665d9f78bb121976f83b1dc\", repo tag \"ghcr.io/flatcar/calico/node:v3.28.1\", repo digest \"ghcr.io/flatcar/calico/node@sha256:47908d8b3046dadd6fbea273ac5b0b9bb803cc7b58b9114c50bf7591767d2744\", size \"113057162\" in 2.972374377s" Sep 4 17:19:47.830190 containerd[1448]: time="2024-09-04T17:19:47.830097944Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.28.1\" returns image reference \"sha256:373272045e41e00ebf8da7ce9fc6b26d326fb8b3e665d9f78bb121976f83b1dc\"" Sep 4 17:19:47.843118 containerd[1448]: time="2024-09-04T17:19:47.843080369Z" level=info msg="CreateContainer within sandbox \"916e13ec9143780cd728eb1baa330349e3be3655cc4e66b693d9828ba4733c87\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Sep 4 17:19:47.879001 containerd[1448]: time="2024-09-04T17:19:47.878859329Z" level=info msg="CreateContainer within sandbox \"916e13ec9143780cd728eb1baa330349e3be3655cc4e66b693d9828ba4733c87\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"a1b2773ed403863a3bd712653eb0f93ae671045ba82092331391930d0e46e62f\"" Sep 4 17:19:47.880868 containerd[1448]: time="2024-09-04T17:19:47.879812112Z" level=info msg="StartContainer for \"a1b2773ed403863a3bd712653eb0f93ae671045ba82092331391930d0e46e62f\"" Sep 4 17:19:47.933057 systemd[1]: Started cri-containerd-a1b2773ed403863a3bd712653eb0f93ae671045ba82092331391930d0e46e62f.scope - libcontainer container a1b2773ed403863a3bd712653eb0f93ae671045ba82092331391930d0e46e62f. Sep 4 17:19:48.029193 containerd[1448]: time="2024-09-04T17:19:48.029019439Z" level=info msg="StartContainer for \"a1b2773ed403863a3bd712653eb0f93ae671045ba82092331391930d0e46e62f\" returns successfully" Sep 4 17:19:48.319930 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Sep 4 17:19:48.320070 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Sep 4 17:19:48.895808 kubelet[2523]: E0904 17:19:48.895764 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:48.907760 kubelet[2523]: I0904 17:19:48.907700 2523 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="calico-system/calico-node-ktcqt" podStartSLOduration=1.96088268 podStartE2EDuration="13.907655634s" podCreationTimestamp="2024-09-04 17:19:35 +0000 UTC" firstStartedPulling="2024-09-04 17:19:35.883641237 +0000 UTC m=+21.304555000" lastFinishedPulling="2024-09-04 17:19:47.830414191 +0000 UTC m=+33.251327954" observedRunningTime="2024-09-04 17:19:48.907627073 +0000 UTC m=+34.328540876" watchObservedRunningTime="2024-09-04 17:19:48.907655634 +0000 UTC m=+34.328569397" Sep 4 17:19:49.085690 systemd[1]: Started sshd@7-10.0.0.60:22-10.0.0.1:60546.service - OpenSSH per-connection server daemon (10.0.0.1:60546). Sep 4 17:19:49.133107 sshd[3651]: Accepted publickey for core from 10.0.0.1 port 60546 ssh2: RSA SHA256:TcdII3DD+/vh6fGiZDuqtLwdsO9LHnvXRMQO7IdpdiA Sep 4 17:19:49.134694 sshd[3651]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 4 17:19:49.139070 systemd-logind[1423]: New session 8 of user core. Sep 4 17:19:49.145042 systemd[1]: Started session-8.scope - Session 8 of User core. Sep 4 17:19:49.274380 sshd[3651]: pam_unix(sshd:session): session closed for user core Sep 4 17:19:49.276946 systemd[1]: sshd@7-10.0.0.60:22-10.0.0.1:60546.service: Deactivated successfully. Sep 4 17:19:49.278749 systemd[1]: session-8.scope: Deactivated successfully. Sep 4 17:19:49.280282 systemd-logind[1423]: Session 8 logged out. Waiting for processes to exit. Sep 4 17:19:49.281526 systemd-logind[1423]: Removed session 8. Sep 4 17:19:49.902315 kubelet[2523]: E0904 17:19:49.901967 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:54.287087 systemd[1]: Started sshd@8-10.0.0.60:22-10.0.0.1:60604.service - OpenSSH per-connection server daemon (10.0.0.1:60604). Sep 4 17:19:54.333326 sshd[3891]: Accepted publickey for core from 10.0.0.1 port 60604 ssh2: RSA SHA256:TcdII3DD+/vh6fGiZDuqtLwdsO9LHnvXRMQO7IdpdiA Sep 4 17:19:54.334914 sshd[3891]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 4 17:19:54.339282 systemd-logind[1423]: New session 9 of user core. Sep 4 17:19:54.358084 systemd[1]: Started session-9.scope - Session 9 of User core. Sep 4 17:19:54.495432 sshd[3891]: pam_unix(sshd:session): session closed for user core Sep 4 17:19:54.498510 systemd[1]: sshd@8-10.0.0.60:22-10.0.0.1:60604.service: Deactivated successfully. Sep 4 17:19:54.501736 systemd[1]: session-9.scope: Deactivated successfully. Sep 4 17:19:54.504605 systemd-logind[1423]: Session 9 logged out. Waiting for processes to exit. Sep 4 17:19:54.506210 systemd-logind[1423]: Removed session 9. Sep 4 17:19:56.714517 containerd[1448]: time="2024-09-04T17:19:56.714363170Z" level=info msg="StopPodSandbox for \"53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2\"" Sep 4 17:19:56.716738 containerd[1448]: time="2024-09-04T17:19:56.714984101Z" level=info msg="StopPodSandbox for \"652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2\"" Sep 4 17:19:56.930562 containerd[1448]: 2024-09-04 17:19:56.832 [INFO][3990] k8s.go 608: Cleaning up netns ContainerID="53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" Sep 4 17:19:56.930562 containerd[1448]: 2024-09-04 17:19:56.832 [INFO][3990] dataplane_linux.go 530: Deleting workload's device in netns. ContainerID="53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" iface="eth0" netns="/var/run/netns/cni-b51b0873-9172-919d-85fa-87d19878f27f" Sep 4 17:19:56.930562 containerd[1448]: 2024-09-04 17:19:56.833 [INFO][3990] dataplane_linux.go 541: Entered netns, deleting veth. ContainerID="53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" iface="eth0" netns="/var/run/netns/cni-b51b0873-9172-919d-85fa-87d19878f27f" Sep 4 17:19:56.930562 containerd[1448]: 2024-09-04 17:19:56.833 [INFO][3990] dataplane_linux.go 568: Workload's veth was already gone. Nothing to do. ContainerID="53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" iface="eth0" netns="/var/run/netns/cni-b51b0873-9172-919d-85fa-87d19878f27f" Sep 4 17:19:56.930562 containerd[1448]: 2024-09-04 17:19:56.833 [INFO][3990] k8s.go 615: Releasing IP address(es) ContainerID="53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" Sep 4 17:19:56.930562 containerd[1448]: 2024-09-04 17:19:56.834 [INFO][3990] utils.go 188: Calico CNI releasing IP address ContainerID="53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" Sep 4 17:19:56.930562 containerd[1448]: 2024-09-04 17:19:56.905 [INFO][4005] ipam_plugin.go 417: Releasing address using handleID ContainerID="53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" HandleID="k8s-pod-network.53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" Workload="localhost-k8s-coredns--76f75df574--kldz2-eth0" Sep 4 17:19:56.930562 containerd[1448]: 2024-09-04 17:19:56.906 [INFO][4005] ipam_plugin.go 358: About to acquire host-wide IPAM lock. Sep 4 17:19:56.930562 containerd[1448]: 2024-09-04 17:19:56.906 [INFO][4005] ipam_plugin.go 373: Acquired host-wide IPAM lock. Sep 4 17:19:56.930562 containerd[1448]: 2024-09-04 17:19:56.919 [WARNING][4005] ipam_plugin.go 434: Asked to release address but it doesn't exist. Ignoring ContainerID="53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" HandleID="k8s-pod-network.53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" Workload="localhost-k8s-coredns--76f75df574--kldz2-eth0" Sep 4 17:19:56.930562 containerd[1448]: 2024-09-04 17:19:56.919 [INFO][4005] ipam_plugin.go 445: Releasing address using workloadID ContainerID="53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" HandleID="k8s-pod-network.53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" Workload="localhost-k8s-coredns--76f75df574--kldz2-eth0" Sep 4 17:19:56.930562 containerd[1448]: 2024-09-04 17:19:56.921 [INFO][4005] ipam_plugin.go 379: Released host-wide IPAM lock. Sep 4 17:19:56.930562 containerd[1448]: 2024-09-04 17:19:56.928 [INFO][3990] k8s.go 621: Teardown processing complete. ContainerID="53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" Sep 4 17:19:56.931643 containerd[1448]: time="2024-09-04T17:19:56.931603445Z" level=info msg="TearDown network for sandbox \"53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2\" successfully" Sep 4 17:19:56.932389 containerd[1448]: time="2024-09-04T17:19:56.932354660Z" level=info msg="StopPodSandbox for \"53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2\" returns successfully" Sep 4 17:19:56.933159 kubelet[2523]: E0904 17:19:56.933133 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:56.934767 containerd[1448]: time="2024-09-04T17:19:56.933609563Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-76f75df574-kldz2,Uid:eccc4d8b-807c-4895-8200-d4b15b8caceb,Namespace:kube-system,Attempt:1,}" Sep 4 17:19:56.935273 systemd[1]: run-netns-cni\x2db51b0873\x2d9172\x2d919d\x2d85fa\x2d87d19878f27f.mount: Deactivated successfully. Sep 4 17:19:56.942280 containerd[1448]: 2024-09-04 17:19:56.883 [INFO][3991] k8s.go 608: Cleaning up netns ContainerID="652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" Sep 4 17:19:56.942280 containerd[1448]: 2024-09-04 17:19:56.884 [INFO][3991] dataplane_linux.go 530: Deleting workload's device in netns. ContainerID="652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" iface="eth0" netns="/var/run/netns/cni-9d7dcc78-711a-38bd-5554-57c3dfd44479" Sep 4 17:19:56.942280 containerd[1448]: 2024-09-04 17:19:56.884 [INFO][3991] dataplane_linux.go 541: Entered netns, deleting veth. ContainerID="652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" iface="eth0" netns="/var/run/netns/cni-9d7dcc78-711a-38bd-5554-57c3dfd44479" Sep 4 17:19:56.942280 containerd[1448]: 2024-09-04 17:19:56.884 [INFO][3991] dataplane_linux.go 568: Workload's veth was already gone. Nothing to do. ContainerID="652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" iface="eth0" netns="/var/run/netns/cni-9d7dcc78-711a-38bd-5554-57c3dfd44479" Sep 4 17:19:56.942280 containerd[1448]: 2024-09-04 17:19:56.884 [INFO][3991] k8s.go 615: Releasing IP address(es) ContainerID="652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" Sep 4 17:19:56.942280 containerd[1448]: 2024-09-04 17:19:56.884 [INFO][3991] utils.go 188: Calico CNI releasing IP address ContainerID="652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" Sep 4 17:19:56.942280 containerd[1448]: 2024-09-04 17:19:56.914 [INFO][4012] ipam_plugin.go 417: Releasing address using handleID ContainerID="652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" HandleID="k8s-pod-network.652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" Workload="localhost-k8s-coredns--76f75df574--tjcbc-eth0" Sep 4 17:19:56.942280 containerd[1448]: 2024-09-04 17:19:56.914 [INFO][4012] ipam_plugin.go 358: About to acquire host-wide IPAM lock. Sep 4 17:19:56.942280 containerd[1448]: 2024-09-04 17:19:56.921 [INFO][4012] ipam_plugin.go 373: Acquired host-wide IPAM lock. Sep 4 17:19:56.942280 containerd[1448]: 2024-09-04 17:19:56.933 [WARNING][4012] ipam_plugin.go 434: Asked to release address but it doesn't exist. Ignoring ContainerID="652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" HandleID="k8s-pod-network.652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" Workload="localhost-k8s-coredns--76f75df574--tjcbc-eth0" Sep 4 17:19:56.942280 containerd[1448]: 2024-09-04 17:19:56.933 [INFO][4012] ipam_plugin.go 445: Releasing address using workloadID ContainerID="652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" HandleID="k8s-pod-network.652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" Workload="localhost-k8s-coredns--76f75df574--tjcbc-eth0" Sep 4 17:19:56.942280 containerd[1448]: 2024-09-04 17:19:56.937 [INFO][4012] ipam_plugin.go 379: Released host-wide IPAM lock. Sep 4 17:19:56.942280 containerd[1448]: 2024-09-04 17:19:56.940 [INFO][3991] k8s.go 621: Teardown processing complete. ContainerID="652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" Sep 4 17:19:56.943440 containerd[1448]: time="2024-09-04T17:19:56.942431851Z" level=info msg="TearDown network for sandbox \"652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2\" successfully" Sep 4 17:19:56.943440 containerd[1448]: time="2024-09-04T17:19:56.942459331Z" level=info msg="StopPodSandbox for \"652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2\" returns successfully" Sep 4 17:19:56.943611 kubelet[2523]: E0904 17:19:56.943389 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:56.945586 containerd[1448]: time="2024-09-04T17:19:56.944330527Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-76f75df574-tjcbc,Uid:5962684c-02f2-4d5e-87f8-a103e98d7b84,Namespace:kube-system,Attempt:1,}" Sep 4 17:19:56.945445 systemd[1]: run-netns-cni\x2d9d7dcc78\x2d711a\x2d38bd\x2d5554\x2d57c3dfd44479.mount: Deactivated successfully. Sep 4 17:19:57.086382 systemd-networkd[1377]: calif3acbe904d3: Link UP Sep 4 17:19:57.087462 systemd-networkd[1377]: calif3acbe904d3: Gained carrier Sep 4 17:19:57.102807 containerd[1448]: 2024-09-04 17:19:56.988 [INFO][4022] utils.go 100: File /var/lib/calico/mtu does not exist Sep 4 17:19:57.102807 containerd[1448]: 2024-09-04 17:19:57.002 [INFO][4022] plugin.go 326: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--76f75df574--tjcbc-eth0 coredns-76f75df574- kube-system 5962684c-02f2-4d5e-87f8-a103e98d7b84 804 0 2024-09-04 17:19:28 +0000 UTC map[k8s-app:kube-dns pod-template-hash:76f75df574 projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-76f75df574-tjcbc eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] calif3acbe904d3 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="ec3c18800d5f45f726125a0fd3000e067d945a8331f40df4276d6f2732459772" Namespace="kube-system" Pod="coredns-76f75df574-tjcbc" WorkloadEndpoint="localhost-k8s-coredns--76f75df574--tjcbc-" Sep 4 17:19:57.102807 containerd[1448]: 2024-09-04 17:19:57.003 [INFO][4022] k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="ec3c18800d5f45f726125a0fd3000e067d945a8331f40df4276d6f2732459772" Namespace="kube-system" Pod="coredns-76f75df574-tjcbc" WorkloadEndpoint="localhost-k8s-coredns--76f75df574--tjcbc-eth0" Sep 4 17:19:57.102807 containerd[1448]: 2024-09-04 17:19:57.033 [INFO][4050] ipam_plugin.go 230: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="ec3c18800d5f45f726125a0fd3000e067d945a8331f40df4276d6f2732459772" HandleID="k8s-pod-network.ec3c18800d5f45f726125a0fd3000e067d945a8331f40df4276d6f2732459772" Workload="localhost-k8s-coredns--76f75df574--tjcbc-eth0" Sep 4 17:19:57.102807 containerd[1448]: 2024-09-04 17:19:57.047 [INFO][4050] ipam_plugin.go 270: Auto assigning IP ContainerID="ec3c18800d5f45f726125a0fd3000e067d945a8331f40df4276d6f2732459772" HandleID="k8s-pod-network.ec3c18800d5f45f726125a0fd3000e067d945a8331f40df4276d6f2732459772" Workload="localhost-k8s-coredns--76f75df574--tjcbc-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000374c30), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-76f75df574-tjcbc", "timestamp":"2024-09-04 17:19:57.033114877 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 4 17:19:57.102807 containerd[1448]: 2024-09-04 17:19:57.047 [INFO][4050] ipam_plugin.go 358: About to acquire host-wide IPAM lock. Sep 4 17:19:57.102807 containerd[1448]: 2024-09-04 17:19:57.047 [INFO][4050] ipam_plugin.go 373: Acquired host-wide IPAM lock. Sep 4 17:19:57.102807 containerd[1448]: 2024-09-04 17:19:57.047 [INFO][4050] ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 4 17:19:57.102807 containerd[1448]: 2024-09-04 17:19:57.049 [INFO][4050] ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.ec3c18800d5f45f726125a0fd3000e067d945a8331f40df4276d6f2732459772" host="localhost" Sep 4 17:19:57.102807 containerd[1448]: 2024-09-04 17:19:57.056 [INFO][4050] ipam.go 372: Looking up existing affinities for host host="localhost" Sep 4 17:19:57.102807 containerd[1448]: 2024-09-04 17:19:57.061 [INFO][4050] ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Sep 4 17:19:57.102807 containerd[1448]: 2024-09-04 17:19:57.063 [INFO][4050] ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 4 17:19:57.102807 containerd[1448]: 2024-09-04 17:19:57.065 [INFO][4050] ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 4 17:19:57.102807 containerd[1448]: 2024-09-04 17:19:57.065 [INFO][4050] ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.ec3c18800d5f45f726125a0fd3000e067d945a8331f40df4276d6f2732459772" host="localhost" Sep 4 17:19:57.102807 containerd[1448]: 2024-09-04 17:19:57.067 [INFO][4050] ipam.go 1685: Creating new handle: k8s-pod-network.ec3c18800d5f45f726125a0fd3000e067d945a8331f40df4276d6f2732459772 Sep 4 17:19:57.102807 containerd[1448]: 2024-09-04 17:19:57.071 [INFO][4050] ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.ec3c18800d5f45f726125a0fd3000e067d945a8331f40df4276d6f2732459772" host="localhost" Sep 4 17:19:57.102807 containerd[1448]: 2024-09-04 17:19:57.076 [INFO][4050] ipam.go 1216: Successfully claimed IPs: [192.168.88.129/26] block=192.168.88.128/26 handle="k8s-pod-network.ec3c18800d5f45f726125a0fd3000e067d945a8331f40df4276d6f2732459772" host="localhost" Sep 4 17:19:57.102807 containerd[1448]: 2024-09-04 17:19:57.076 [INFO][4050] ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.129/26] handle="k8s-pod-network.ec3c18800d5f45f726125a0fd3000e067d945a8331f40df4276d6f2732459772" host="localhost" Sep 4 17:19:57.102807 containerd[1448]: 2024-09-04 17:19:57.076 [INFO][4050] ipam_plugin.go 379: Released host-wide IPAM lock. Sep 4 17:19:57.102807 containerd[1448]: 2024-09-04 17:19:57.076 [INFO][4050] ipam_plugin.go 288: Calico CNI IPAM assigned addresses IPv4=[192.168.88.129/26] IPv6=[] ContainerID="ec3c18800d5f45f726125a0fd3000e067d945a8331f40df4276d6f2732459772" HandleID="k8s-pod-network.ec3c18800d5f45f726125a0fd3000e067d945a8331f40df4276d6f2732459772" Workload="localhost-k8s-coredns--76f75df574--tjcbc-eth0" Sep 4 17:19:57.103655 containerd[1448]: 2024-09-04 17:19:57.079 [INFO][4022] k8s.go 386: Populated endpoint ContainerID="ec3c18800d5f45f726125a0fd3000e067d945a8331f40df4276d6f2732459772" Namespace="kube-system" Pod="coredns-76f75df574-tjcbc" WorkloadEndpoint="localhost-k8s-coredns--76f75df574--tjcbc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--76f75df574--tjcbc-eth0", GenerateName:"coredns-76f75df574-", Namespace:"kube-system", SelfLink:"", UID:"5962684c-02f2-4d5e-87f8-a103e98d7b84", ResourceVersion:"804", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 19, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"76f75df574", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-76f75df574-tjcbc", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calif3acbe904d3", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:19:57.103655 containerd[1448]: 2024-09-04 17:19:57.080 [INFO][4022] k8s.go 387: Calico CNI using IPs: [192.168.88.129/32] ContainerID="ec3c18800d5f45f726125a0fd3000e067d945a8331f40df4276d6f2732459772" Namespace="kube-system" Pod="coredns-76f75df574-tjcbc" WorkloadEndpoint="localhost-k8s-coredns--76f75df574--tjcbc-eth0" Sep 4 17:19:57.103655 containerd[1448]: 2024-09-04 17:19:57.080 [INFO][4022] dataplane_linux.go 68: Setting the host side veth name to calif3acbe904d3 ContainerID="ec3c18800d5f45f726125a0fd3000e067d945a8331f40df4276d6f2732459772" Namespace="kube-system" Pod="coredns-76f75df574-tjcbc" WorkloadEndpoint="localhost-k8s-coredns--76f75df574--tjcbc-eth0" Sep 4 17:19:57.103655 containerd[1448]: 2024-09-04 17:19:57.088 [INFO][4022] dataplane_linux.go 479: Disabling IPv4 forwarding ContainerID="ec3c18800d5f45f726125a0fd3000e067d945a8331f40df4276d6f2732459772" Namespace="kube-system" Pod="coredns-76f75df574-tjcbc" WorkloadEndpoint="localhost-k8s-coredns--76f75df574--tjcbc-eth0" Sep 4 17:19:57.103655 containerd[1448]: 2024-09-04 17:19:57.088 [INFO][4022] k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="ec3c18800d5f45f726125a0fd3000e067d945a8331f40df4276d6f2732459772" Namespace="kube-system" Pod="coredns-76f75df574-tjcbc" WorkloadEndpoint="localhost-k8s-coredns--76f75df574--tjcbc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--76f75df574--tjcbc-eth0", GenerateName:"coredns-76f75df574-", Namespace:"kube-system", SelfLink:"", UID:"5962684c-02f2-4d5e-87f8-a103e98d7b84", ResourceVersion:"804", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 19, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"76f75df574", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"ec3c18800d5f45f726125a0fd3000e067d945a8331f40df4276d6f2732459772", Pod:"coredns-76f75df574-tjcbc", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calif3acbe904d3", MAC:"76:ca:ae:b6:e9:7b", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:19:57.103655 containerd[1448]: 2024-09-04 17:19:57.100 [INFO][4022] k8s.go 500: Wrote updated endpoint to datastore ContainerID="ec3c18800d5f45f726125a0fd3000e067d945a8331f40df4276d6f2732459772" Namespace="kube-system" Pod="coredns-76f75df574-tjcbc" WorkloadEndpoint="localhost-k8s-coredns--76f75df574--tjcbc-eth0" Sep 4 17:19:57.133009 systemd-networkd[1377]: calic733e3622ac: Link UP Sep 4 17:19:57.133244 systemd-networkd[1377]: calic733e3622ac: Gained carrier Sep 4 17:19:57.141807 containerd[1448]: time="2024-09-04T17:19:57.141454011Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 4 17:19:57.141807 containerd[1448]: time="2024-09-04T17:19:57.141516572Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:19:57.141807 containerd[1448]: time="2024-09-04T17:19:57.141533732Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 4 17:19:57.141807 containerd[1448]: time="2024-09-04T17:19:57.141543613Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:19:57.154739 containerd[1448]: 2024-09-04 17:19:56.989 [INFO][4028] utils.go 100: File /var/lib/calico/mtu does not exist Sep 4 17:19:57.154739 containerd[1448]: 2024-09-04 17:19:57.008 [INFO][4028] plugin.go 326: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--76f75df574--kldz2-eth0 coredns-76f75df574- kube-system eccc4d8b-807c-4895-8200-d4b15b8caceb 803 0 2024-09-04 17:19:28 +0000 UTC map[k8s-app:kube-dns pod-template-hash:76f75df574 projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-76f75df574-kldz2 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] calic733e3622ac [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="5a252404a16d8b1dae1655fd759d3c98aedc4b9bb2ad64aa9f11d3d710c53b06" Namespace="kube-system" Pod="coredns-76f75df574-kldz2" WorkloadEndpoint="localhost-k8s-coredns--76f75df574--kldz2-" Sep 4 17:19:57.154739 containerd[1448]: 2024-09-04 17:19:57.008 [INFO][4028] k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="5a252404a16d8b1dae1655fd759d3c98aedc4b9bb2ad64aa9f11d3d710c53b06" Namespace="kube-system" Pod="coredns-76f75df574-kldz2" WorkloadEndpoint="localhost-k8s-coredns--76f75df574--kldz2-eth0" Sep 4 17:19:57.154739 containerd[1448]: 2024-09-04 17:19:57.037 [INFO][4055] ipam_plugin.go 230: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="5a252404a16d8b1dae1655fd759d3c98aedc4b9bb2ad64aa9f11d3d710c53b06" HandleID="k8s-pod-network.5a252404a16d8b1dae1655fd759d3c98aedc4b9bb2ad64aa9f11d3d710c53b06" Workload="localhost-k8s-coredns--76f75df574--kldz2-eth0" Sep 4 17:19:57.154739 containerd[1448]: 2024-09-04 17:19:57.051 [INFO][4055] ipam_plugin.go 270: Auto assigning IP ContainerID="5a252404a16d8b1dae1655fd759d3c98aedc4b9bb2ad64aa9f11d3d710c53b06" HandleID="k8s-pod-network.5a252404a16d8b1dae1655fd759d3c98aedc4b9bb2ad64aa9f11d3d710c53b06" Workload="localhost-k8s-coredns--76f75df574--kldz2-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40003001f0), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-76f75df574-kldz2", "timestamp":"2024-09-04 17:19:57.037487479 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 4 17:19:57.154739 containerd[1448]: 2024-09-04 17:19:57.051 [INFO][4055] ipam_plugin.go 358: About to acquire host-wide IPAM lock. Sep 4 17:19:57.154739 containerd[1448]: 2024-09-04 17:19:57.076 [INFO][4055] ipam_plugin.go 373: Acquired host-wide IPAM lock. Sep 4 17:19:57.154739 containerd[1448]: 2024-09-04 17:19:57.076 [INFO][4055] ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 4 17:19:57.154739 containerd[1448]: 2024-09-04 17:19:57.079 [INFO][4055] ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.5a252404a16d8b1dae1655fd759d3c98aedc4b9bb2ad64aa9f11d3d710c53b06" host="localhost" Sep 4 17:19:57.154739 containerd[1448]: 2024-09-04 17:19:57.088 [INFO][4055] ipam.go 372: Looking up existing affinities for host host="localhost" Sep 4 17:19:57.154739 containerd[1448]: 2024-09-04 17:19:57.094 [INFO][4055] ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Sep 4 17:19:57.154739 containerd[1448]: 2024-09-04 17:19:57.097 [INFO][4055] ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 4 17:19:57.154739 containerd[1448]: 2024-09-04 17:19:57.106 [INFO][4055] ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 4 17:19:57.154739 containerd[1448]: 2024-09-04 17:19:57.107 [INFO][4055] ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.5a252404a16d8b1dae1655fd759d3c98aedc4b9bb2ad64aa9f11d3d710c53b06" host="localhost" Sep 4 17:19:57.154739 containerd[1448]: 2024-09-04 17:19:57.110 [INFO][4055] ipam.go 1685: Creating new handle: k8s-pod-network.5a252404a16d8b1dae1655fd759d3c98aedc4b9bb2ad64aa9f11d3d710c53b06 Sep 4 17:19:57.154739 containerd[1448]: 2024-09-04 17:19:57.115 [INFO][4055] ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.5a252404a16d8b1dae1655fd759d3c98aedc4b9bb2ad64aa9f11d3d710c53b06" host="localhost" Sep 4 17:19:57.154739 containerd[1448]: 2024-09-04 17:19:57.120 [INFO][4055] ipam.go 1216: Successfully claimed IPs: [192.168.88.130/26] block=192.168.88.128/26 handle="k8s-pod-network.5a252404a16d8b1dae1655fd759d3c98aedc4b9bb2ad64aa9f11d3d710c53b06" host="localhost" Sep 4 17:19:57.154739 containerd[1448]: 2024-09-04 17:19:57.120 [INFO][4055] ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.130/26] handle="k8s-pod-network.5a252404a16d8b1dae1655fd759d3c98aedc4b9bb2ad64aa9f11d3d710c53b06" host="localhost" Sep 4 17:19:57.154739 containerd[1448]: 2024-09-04 17:19:57.121 [INFO][4055] ipam_plugin.go 379: Released host-wide IPAM lock. Sep 4 17:19:57.154739 containerd[1448]: 2024-09-04 17:19:57.121 [INFO][4055] ipam_plugin.go 288: Calico CNI IPAM assigned addresses IPv4=[192.168.88.130/26] IPv6=[] ContainerID="5a252404a16d8b1dae1655fd759d3c98aedc4b9bb2ad64aa9f11d3d710c53b06" HandleID="k8s-pod-network.5a252404a16d8b1dae1655fd759d3c98aedc4b9bb2ad64aa9f11d3d710c53b06" Workload="localhost-k8s-coredns--76f75df574--kldz2-eth0" Sep 4 17:19:57.155376 containerd[1448]: 2024-09-04 17:19:57.125 [INFO][4028] k8s.go 386: Populated endpoint ContainerID="5a252404a16d8b1dae1655fd759d3c98aedc4b9bb2ad64aa9f11d3d710c53b06" Namespace="kube-system" Pod="coredns-76f75df574-kldz2" WorkloadEndpoint="localhost-k8s-coredns--76f75df574--kldz2-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--76f75df574--kldz2-eth0", GenerateName:"coredns-76f75df574-", Namespace:"kube-system", SelfLink:"", UID:"eccc4d8b-807c-4895-8200-d4b15b8caceb", ResourceVersion:"803", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 19, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"76f75df574", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-76f75df574-kldz2", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calic733e3622ac", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:19:57.155376 containerd[1448]: 2024-09-04 17:19:57.125 [INFO][4028] k8s.go 387: Calico CNI using IPs: [192.168.88.130/32] ContainerID="5a252404a16d8b1dae1655fd759d3c98aedc4b9bb2ad64aa9f11d3d710c53b06" Namespace="kube-system" Pod="coredns-76f75df574-kldz2" WorkloadEndpoint="localhost-k8s-coredns--76f75df574--kldz2-eth0" Sep 4 17:19:57.155376 containerd[1448]: 2024-09-04 17:19:57.126 [INFO][4028] dataplane_linux.go 68: Setting the host side veth name to calic733e3622ac ContainerID="5a252404a16d8b1dae1655fd759d3c98aedc4b9bb2ad64aa9f11d3d710c53b06" Namespace="kube-system" Pod="coredns-76f75df574-kldz2" WorkloadEndpoint="localhost-k8s-coredns--76f75df574--kldz2-eth0" Sep 4 17:19:57.155376 containerd[1448]: 2024-09-04 17:19:57.133 [INFO][4028] dataplane_linux.go 479: Disabling IPv4 forwarding ContainerID="5a252404a16d8b1dae1655fd759d3c98aedc4b9bb2ad64aa9f11d3d710c53b06" Namespace="kube-system" Pod="coredns-76f75df574-kldz2" WorkloadEndpoint="localhost-k8s-coredns--76f75df574--kldz2-eth0" Sep 4 17:19:57.155376 containerd[1448]: 2024-09-04 17:19:57.134 [INFO][4028] k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="5a252404a16d8b1dae1655fd759d3c98aedc4b9bb2ad64aa9f11d3d710c53b06" Namespace="kube-system" Pod="coredns-76f75df574-kldz2" WorkloadEndpoint="localhost-k8s-coredns--76f75df574--kldz2-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--76f75df574--kldz2-eth0", GenerateName:"coredns-76f75df574-", Namespace:"kube-system", SelfLink:"", UID:"eccc4d8b-807c-4895-8200-d4b15b8caceb", ResourceVersion:"803", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 19, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"76f75df574", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"5a252404a16d8b1dae1655fd759d3c98aedc4b9bb2ad64aa9f11d3d710c53b06", Pod:"coredns-76f75df574-kldz2", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calic733e3622ac", MAC:"6e:b4:68:c9:39:2d", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:19:57.155376 containerd[1448]: 2024-09-04 17:19:57.146 [INFO][4028] k8s.go 500: Wrote updated endpoint to datastore ContainerID="5a252404a16d8b1dae1655fd759d3c98aedc4b9bb2ad64aa9f11d3d710c53b06" Namespace="kube-system" Pod="coredns-76f75df574-kldz2" WorkloadEndpoint="localhost-k8s-coredns--76f75df574--kldz2-eth0" Sep 4 17:19:57.172104 systemd[1]: Started cri-containerd-ec3c18800d5f45f726125a0fd3000e067d945a8331f40df4276d6f2732459772.scope - libcontainer container ec3c18800d5f45f726125a0fd3000e067d945a8331f40df4276d6f2732459772. Sep 4 17:19:57.187572 systemd-resolved[1309]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 4 17:19:57.189640 containerd[1448]: time="2024-09-04T17:19:57.189404142Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 4 17:19:57.190250 containerd[1448]: time="2024-09-04T17:19:57.189584505Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:19:57.190250 containerd[1448]: time="2024-09-04T17:19:57.190101115Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 4 17:19:57.190250 containerd[1448]: time="2024-09-04T17:19:57.190126235Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:19:57.219455 systemd[1]: Started cri-containerd-5a252404a16d8b1dae1655fd759d3c98aedc4b9bb2ad64aa9f11d3d710c53b06.scope - libcontainer container 5a252404a16d8b1dae1655fd759d3c98aedc4b9bb2ad64aa9f11d3d710c53b06. Sep 4 17:19:57.220079 containerd[1448]: time="2024-09-04T17:19:57.219719585Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-76f75df574-tjcbc,Uid:5962684c-02f2-4d5e-87f8-a103e98d7b84,Namespace:kube-system,Attempt:1,} returns sandbox id \"ec3c18800d5f45f726125a0fd3000e067d945a8331f40df4276d6f2732459772\"" Sep 4 17:19:57.223251 kubelet[2523]: E0904 17:19:57.222315 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:57.228975 containerd[1448]: time="2024-09-04T17:19:57.228815155Z" level=info msg="CreateContainer within sandbox \"ec3c18800d5f45f726125a0fd3000e067d945a8331f40df4276d6f2732459772\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 4 17:19:57.246251 systemd-resolved[1309]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 4 17:19:57.252597 containerd[1448]: time="2024-09-04T17:19:57.252538155Z" level=info msg="CreateContainer within sandbox \"ec3c18800d5f45f726125a0fd3000e067d945a8331f40df4276d6f2732459772\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"8557619dcad2077b2a5d751089af35e97e23aa91a7aeb1ad249dca8e1a781d8a\"" Sep 4 17:19:57.256174 containerd[1448]: time="2024-09-04T17:19:57.256127822Z" level=info msg="StartContainer for \"8557619dcad2077b2a5d751089af35e97e23aa91a7aeb1ad249dca8e1a781d8a\"" Sep 4 17:19:57.277589 containerd[1448]: time="2024-09-04T17:19:57.277507220Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-76f75df574-kldz2,Uid:eccc4d8b-807c-4895-8200-d4b15b8caceb,Namespace:kube-system,Attempt:1,} returns sandbox id \"5a252404a16d8b1dae1655fd759d3c98aedc4b9bb2ad64aa9f11d3d710c53b06\"" Sep 4 17:19:57.278594 kubelet[2523]: E0904 17:19:57.278564 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:57.283480 containerd[1448]: time="2024-09-04T17:19:57.283419769Z" level=info msg="CreateContainer within sandbox \"5a252404a16d8b1dae1655fd759d3c98aedc4b9bb2ad64aa9f11d3d710c53b06\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 4 17:19:57.299651 containerd[1448]: time="2024-09-04T17:19:57.299537349Z" level=info msg="CreateContainer within sandbox \"5a252404a16d8b1dae1655fd759d3c98aedc4b9bb2ad64aa9f11d3d710c53b06\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"e6087049b85516cdcebd90ae5e8bb3557aaf0e86f616f0a859734ac8350cee7a\"" Sep 4 17:19:57.300203 containerd[1448]: time="2024-09-04T17:19:57.300147400Z" level=info msg="StartContainer for \"e6087049b85516cdcebd90ae5e8bb3557aaf0e86f616f0a859734ac8350cee7a\"" Sep 4 17:19:57.303096 systemd[1]: Started cri-containerd-8557619dcad2077b2a5d751089af35e97e23aa91a7aeb1ad249dca8e1a781d8a.scope - libcontainer container 8557619dcad2077b2a5d751089af35e97e23aa91a7aeb1ad249dca8e1a781d8a. Sep 4 17:19:57.338911 systemd[1]: Started cri-containerd-e6087049b85516cdcebd90ae5e8bb3557aaf0e86f616f0a859734ac8350cee7a.scope - libcontainer container e6087049b85516cdcebd90ae5e8bb3557aaf0e86f616f0a859734ac8350cee7a. Sep 4 17:19:57.348061 containerd[1448]: time="2024-09-04T17:19:57.347933448Z" level=info msg="StartContainer for \"8557619dcad2077b2a5d751089af35e97e23aa91a7aeb1ad249dca8e1a781d8a\" returns successfully" Sep 4 17:19:57.374060 containerd[1448]: time="2024-09-04T17:19:57.374007013Z" level=info msg="StartContainer for \"e6087049b85516cdcebd90ae5e8bb3557aaf0e86f616f0a859734ac8350cee7a\" returns successfully" Sep 4 17:19:57.934853 kubelet[2523]: E0904 17:19:57.934806 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:57.950378 kubelet[2523]: E0904 17:19:57.950310 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:57.951789 kubelet[2523]: I0904 17:19:57.951701 2523 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/coredns-76f75df574-kldz2" podStartSLOduration=29.951651429 podStartE2EDuration="29.951651429s" podCreationTimestamp="2024-09-04 17:19:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-09-04 17:19:57.950807333 +0000 UTC m=+43.371721096" watchObservedRunningTime="2024-09-04 17:19:57.951651429 +0000 UTC m=+43.372565232" Sep 4 17:19:57.992390 kubelet[2523]: I0904 17:19:57.992330 2523 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/coredns-76f75df574-tjcbc" podStartSLOduration=29.992280584 podStartE2EDuration="29.992280584s" podCreationTimestamp="2024-09-04 17:19:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-09-04 17:19:57.991228124 +0000 UTC m=+43.412141927" watchObservedRunningTime="2024-09-04 17:19:57.992280584 +0000 UTC m=+43.413194307" Sep 4 17:19:58.774141 systemd-networkd[1377]: calif3acbe904d3: Gained IPv6LL Sep 4 17:19:58.902048 systemd-networkd[1377]: calic733e3622ac: Gained IPv6LL Sep 4 17:19:58.950119 kubelet[2523]: E0904 17:19:58.950027 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:58.950458 kubelet[2523]: E0904 17:19:58.950371 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:19:59.525184 systemd[1]: Started sshd@9-10.0.0.60:22-10.0.0.1:60622.service - OpenSSH per-connection server daemon (10.0.0.1:60622). Sep 4 17:19:59.569011 sshd[4334]: Accepted publickey for core from 10.0.0.1 port 60622 ssh2: RSA SHA256:TcdII3DD+/vh6fGiZDuqtLwdsO9LHnvXRMQO7IdpdiA Sep 4 17:19:59.570906 sshd[4334]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 4 17:19:59.582325 systemd-logind[1423]: New session 10 of user core. Sep 4 17:19:59.591065 systemd[1]: Started session-10.scope - Session 10 of User core. Sep 4 17:19:59.715935 containerd[1448]: time="2024-09-04T17:19:59.715889251Z" level=info msg="StopPodSandbox for \"0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be\"" Sep 4 17:19:59.744814 sshd[4334]: pam_unix(sshd:session): session closed for user core Sep 4 17:19:59.752706 systemd[1]: sshd@9-10.0.0.60:22-10.0.0.1:60622.service: Deactivated successfully. Sep 4 17:19:59.756126 systemd[1]: session-10.scope: Deactivated successfully. Sep 4 17:19:59.759168 systemd-logind[1423]: Session 10 logged out. Waiting for processes to exit. Sep 4 17:19:59.765192 systemd[1]: Started sshd@10-10.0.0.60:22-10.0.0.1:60628.service - OpenSSH per-connection server daemon (10.0.0.1:60628). Sep 4 17:19:59.768869 systemd-logind[1423]: Removed session 10. Sep 4 17:19:59.828248 sshd[4376]: Accepted publickey for core from 10.0.0.1 port 60628 ssh2: RSA SHA256:TcdII3DD+/vh6fGiZDuqtLwdsO9LHnvXRMQO7IdpdiA Sep 4 17:19:59.831502 sshd[4376]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 4 17:19:59.834852 containerd[1448]: 2024-09-04 17:19:59.785 [INFO][4366] k8s.go 608: Cleaning up netns ContainerID="0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" Sep 4 17:19:59.834852 containerd[1448]: 2024-09-04 17:19:59.785 [INFO][4366] dataplane_linux.go 530: Deleting workload's device in netns. ContainerID="0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" iface="eth0" netns="/var/run/netns/cni-c3d31b22-915d-400e-32b7-460fe6fe1315" Sep 4 17:19:59.834852 containerd[1448]: 2024-09-04 17:19:59.785 [INFO][4366] dataplane_linux.go 541: Entered netns, deleting veth. ContainerID="0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" iface="eth0" netns="/var/run/netns/cni-c3d31b22-915d-400e-32b7-460fe6fe1315" Sep 4 17:19:59.834852 containerd[1448]: 2024-09-04 17:19:59.786 [INFO][4366] dataplane_linux.go 568: Workload's veth was already gone. Nothing to do. ContainerID="0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" iface="eth0" netns="/var/run/netns/cni-c3d31b22-915d-400e-32b7-460fe6fe1315" Sep 4 17:19:59.834852 containerd[1448]: 2024-09-04 17:19:59.786 [INFO][4366] k8s.go 615: Releasing IP address(es) ContainerID="0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" Sep 4 17:19:59.834852 containerd[1448]: 2024-09-04 17:19:59.786 [INFO][4366] utils.go 188: Calico CNI releasing IP address ContainerID="0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" Sep 4 17:19:59.834852 containerd[1448]: 2024-09-04 17:19:59.819 [INFO][4377] ipam_plugin.go 417: Releasing address using handleID ContainerID="0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" HandleID="k8s-pod-network.0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" Workload="localhost-k8s-calico--kube--controllers--7ff458757c--hfq6x-eth0" Sep 4 17:19:59.834852 containerd[1448]: 2024-09-04 17:19:59.819 [INFO][4377] ipam_plugin.go 358: About to acquire host-wide IPAM lock. Sep 4 17:19:59.834852 containerd[1448]: 2024-09-04 17:19:59.819 [INFO][4377] ipam_plugin.go 373: Acquired host-wide IPAM lock. Sep 4 17:19:59.834852 containerd[1448]: 2024-09-04 17:19:59.828 [WARNING][4377] ipam_plugin.go 434: Asked to release address but it doesn't exist. Ignoring ContainerID="0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" HandleID="k8s-pod-network.0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" Workload="localhost-k8s-calico--kube--controllers--7ff458757c--hfq6x-eth0" Sep 4 17:19:59.834852 containerd[1448]: 2024-09-04 17:19:59.828 [INFO][4377] ipam_plugin.go 445: Releasing address using workloadID ContainerID="0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" HandleID="k8s-pod-network.0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" Workload="localhost-k8s-calico--kube--controllers--7ff458757c--hfq6x-eth0" Sep 4 17:19:59.834852 containerd[1448]: 2024-09-04 17:19:59.831 [INFO][4377] ipam_plugin.go 379: Released host-wide IPAM lock. Sep 4 17:19:59.834852 containerd[1448]: 2024-09-04 17:19:59.833 [INFO][4366] k8s.go 621: Teardown processing complete. ContainerID="0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" Sep 4 17:19:59.835285 containerd[1448]: time="2024-09-04T17:19:59.835019387Z" level=info msg="TearDown network for sandbox \"0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be\" successfully" Sep 4 17:19:59.835285 containerd[1448]: time="2024-09-04T17:19:59.835049427Z" level=info msg="StopPodSandbox for \"0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be\" returns successfully" Sep 4 17:19:59.837015 systemd[1]: run-netns-cni\x2dc3d31b22\x2d915d\x2d400e\x2d32b7\x2d460fe6fe1315.mount: Deactivated successfully. Sep 4 17:19:59.838147 containerd[1448]: time="2024-09-04T17:19:59.838106562Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7ff458757c-hfq6x,Uid:05f65067-932e-4656-a9e0-3f22a268ebd5,Namespace:calico-system,Attempt:1,}" Sep 4 17:19:59.840081 systemd-logind[1423]: New session 11 of user core. Sep 4 17:19:59.856053 systemd[1]: Started session-11.scope - Session 11 of User core. Sep 4 17:19:59.952570 kubelet[2523]: E0904 17:19:59.952515 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:20:00.008102 systemd-networkd[1377]: cali20d0faf09d3: Link UP Sep 4 17:20:00.008522 systemd-networkd[1377]: cali20d0faf09d3: Gained carrier Sep 4 17:20:00.025370 containerd[1448]: 2024-09-04 17:19:59.906 [INFO][4387] utils.go 100: File /var/lib/calico/mtu does not exist Sep 4 17:20:00.025370 containerd[1448]: 2024-09-04 17:19:59.919 [INFO][4387] plugin.go 326: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--kube--controllers--7ff458757c--hfq6x-eth0 calico-kube-controllers-7ff458757c- calico-system 05f65067-932e-4656-a9e0-3f22a268ebd5 848 0 2024-09-04 17:19:35 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:7ff458757c projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s localhost calico-kube-controllers-7ff458757c-hfq6x eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali20d0faf09d3 [] []}} ContainerID="06f8886531fe2b285112a3dcf4561038d7e1714c54aed5c12a11f5a53ff1691b" Namespace="calico-system" Pod="calico-kube-controllers-7ff458757c-hfq6x" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7ff458757c--hfq6x-" Sep 4 17:20:00.025370 containerd[1448]: 2024-09-04 17:19:59.919 [INFO][4387] k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="06f8886531fe2b285112a3dcf4561038d7e1714c54aed5c12a11f5a53ff1691b" Namespace="calico-system" Pod="calico-kube-controllers-7ff458757c-hfq6x" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7ff458757c--hfq6x-eth0" Sep 4 17:20:00.025370 containerd[1448]: 2024-09-04 17:19:59.951 [INFO][4405] ipam_plugin.go 230: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="06f8886531fe2b285112a3dcf4561038d7e1714c54aed5c12a11f5a53ff1691b" HandleID="k8s-pod-network.06f8886531fe2b285112a3dcf4561038d7e1714c54aed5c12a11f5a53ff1691b" Workload="localhost-k8s-calico--kube--controllers--7ff458757c--hfq6x-eth0" Sep 4 17:20:00.025370 containerd[1448]: 2024-09-04 17:19:59.968 [INFO][4405] ipam_plugin.go 270: Auto assigning IP ContainerID="06f8886531fe2b285112a3dcf4561038d7e1714c54aed5c12a11f5a53ff1691b" HandleID="k8s-pod-network.06f8886531fe2b285112a3dcf4561038d7e1714c54aed5c12a11f5a53ff1691b" Workload="localhost-k8s-calico--kube--controllers--7ff458757c--hfq6x-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x400004ddb0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"calico-kube-controllers-7ff458757c-hfq6x", "timestamp":"2024-09-04 17:19:59.951438914 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 4 17:20:00.025370 containerd[1448]: 2024-09-04 17:19:59.968 [INFO][4405] ipam_plugin.go 358: About to acquire host-wide IPAM lock. Sep 4 17:20:00.025370 containerd[1448]: 2024-09-04 17:19:59.968 [INFO][4405] ipam_plugin.go 373: Acquired host-wide IPAM lock. Sep 4 17:20:00.025370 containerd[1448]: 2024-09-04 17:19:59.968 [INFO][4405] ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 4 17:20:00.025370 containerd[1448]: 2024-09-04 17:19:59.970 [INFO][4405] ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.06f8886531fe2b285112a3dcf4561038d7e1714c54aed5c12a11f5a53ff1691b" host="localhost" Sep 4 17:20:00.025370 containerd[1448]: 2024-09-04 17:19:59.977 [INFO][4405] ipam.go 372: Looking up existing affinities for host host="localhost" Sep 4 17:20:00.025370 containerd[1448]: 2024-09-04 17:19:59.982 [INFO][4405] ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Sep 4 17:20:00.025370 containerd[1448]: 2024-09-04 17:19:59.984 [INFO][4405] ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 4 17:20:00.025370 containerd[1448]: 2024-09-04 17:19:59.988 [INFO][4405] ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 4 17:20:00.025370 containerd[1448]: 2024-09-04 17:19:59.988 [INFO][4405] ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.06f8886531fe2b285112a3dcf4561038d7e1714c54aed5c12a11f5a53ff1691b" host="localhost" Sep 4 17:20:00.025370 containerd[1448]: 2024-09-04 17:19:59.990 [INFO][4405] ipam.go 1685: Creating new handle: k8s-pod-network.06f8886531fe2b285112a3dcf4561038d7e1714c54aed5c12a11f5a53ff1691b Sep 4 17:20:00.025370 containerd[1448]: 2024-09-04 17:19:59.993 [INFO][4405] ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.06f8886531fe2b285112a3dcf4561038d7e1714c54aed5c12a11f5a53ff1691b" host="localhost" Sep 4 17:20:00.025370 containerd[1448]: 2024-09-04 17:20:00.000 [INFO][4405] ipam.go 1216: Successfully claimed IPs: [192.168.88.131/26] block=192.168.88.128/26 handle="k8s-pod-network.06f8886531fe2b285112a3dcf4561038d7e1714c54aed5c12a11f5a53ff1691b" host="localhost" Sep 4 17:20:00.025370 containerd[1448]: 2024-09-04 17:20:00.000 [INFO][4405] ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.131/26] handle="k8s-pod-network.06f8886531fe2b285112a3dcf4561038d7e1714c54aed5c12a11f5a53ff1691b" host="localhost" Sep 4 17:20:00.025370 containerd[1448]: 2024-09-04 17:20:00.000 [INFO][4405] ipam_plugin.go 379: Released host-wide IPAM lock. Sep 4 17:20:00.025370 containerd[1448]: 2024-09-04 17:20:00.000 [INFO][4405] ipam_plugin.go 288: Calico CNI IPAM assigned addresses IPv4=[192.168.88.131/26] IPv6=[] ContainerID="06f8886531fe2b285112a3dcf4561038d7e1714c54aed5c12a11f5a53ff1691b" HandleID="k8s-pod-network.06f8886531fe2b285112a3dcf4561038d7e1714c54aed5c12a11f5a53ff1691b" Workload="localhost-k8s-calico--kube--controllers--7ff458757c--hfq6x-eth0" Sep 4 17:20:00.026027 containerd[1448]: 2024-09-04 17:20:00.005 [INFO][4387] k8s.go 386: Populated endpoint ContainerID="06f8886531fe2b285112a3dcf4561038d7e1714c54aed5c12a11f5a53ff1691b" Namespace="calico-system" Pod="calico-kube-controllers-7ff458757c-hfq6x" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7ff458757c--hfq6x-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--7ff458757c--hfq6x-eth0", GenerateName:"calico-kube-controllers-7ff458757c-", Namespace:"calico-system", SelfLink:"", UID:"05f65067-932e-4656-a9e0-3f22a268ebd5", ResourceVersion:"848", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 19, 35, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"7ff458757c", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-kube-controllers-7ff458757c-hfq6x", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali20d0faf09d3", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:20:00.026027 containerd[1448]: 2024-09-04 17:20:00.005 [INFO][4387] k8s.go 387: Calico CNI using IPs: [192.168.88.131/32] ContainerID="06f8886531fe2b285112a3dcf4561038d7e1714c54aed5c12a11f5a53ff1691b" Namespace="calico-system" Pod="calico-kube-controllers-7ff458757c-hfq6x" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7ff458757c--hfq6x-eth0" Sep 4 17:20:00.026027 containerd[1448]: 2024-09-04 17:20:00.006 [INFO][4387] dataplane_linux.go 68: Setting the host side veth name to cali20d0faf09d3 ContainerID="06f8886531fe2b285112a3dcf4561038d7e1714c54aed5c12a11f5a53ff1691b" Namespace="calico-system" Pod="calico-kube-controllers-7ff458757c-hfq6x" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7ff458757c--hfq6x-eth0" Sep 4 17:20:00.026027 containerd[1448]: 2024-09-04 17:20:00.008 [INFO][4387] dataplane_linux.go 479: Disabling IPv4 forwarding ContainerID="06f8886531fe2b285112a3dcf4561038d7e1714c54aed5c12a11f5a53ff1691b" Namespace="calico-system" Pod="calico-kube-controllers-7ff458757c-hfq6x" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7ff458757c--hfq6x-eth0" Sep 4 17:20:00.026027 containerd[1448]: 2024-09-04 17:20:00.010 [INFO][4387] k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="06f8886531fe2b285112a3dcf4561038d7e1714c54aed5c12a11f5a53ff1691b" Namespace="calico-system" Pod="calico-kube-controllers-7ff458757c-hfq6x" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7ff458757c--hfq6x-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--7ff458757c--hfq6x-eth0", GenerateName:"calico-kube-controllers-7ff458757c-", Namespace:"calico-system", SelfLink:"", UID:"05f65067-932e-4656-a9e0-3f22a268ebd5", ResourceVersion:"848", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 19, 35, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"7ff458757c", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"06f8886531fe2b285112a3dcf4561038d7e1714c54aed5c12a11f5a53ff1691b", Pod:"calico-kube-controllers-7ff458757c-hfq6x", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali20d0faf09d3", MAC:"12:7b:1e:4f:76:54", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:20:00.026027 containerd[1448]: 2024-09-04 17:20:00.023 [INFO][4387] k8s.go 500: Wrote updated endpoint to datastore ContainerID="06f8886531fe2b285112a3dcf4561038d7e1714c54aed5c12a11f5a53ff1691b" Namespace="calico-system" Pod="calico-kube-controllers-7ff458757c-hfq6x" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--7ff458757c--hfq6x-eth0" Sep 4 17:20:00.053383 containerd[1448]: time="2024-09-04T17:20:00.052868717Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 4 17:20:00.053597 containerd[1448]: time="2024-09-04T17:20:00.053423927Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:20:00.053597 containerd[1448]: time="2024-09-04T17:20:00.053449488Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 4 17:20:00.053597 containerd[1448]: time="2024-09-04T17:20:00.053460248Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:20:00.079052 systemd[1]: Started cri-containerd-06f8886531fe2b285112a3dcf4561038d7e1714c54aed5c12a11f5a53ff1691b.scope - libcontainer container 06f8886531fe2b285112a3dcf4561038d7e1714c54aed5c12a11f5a53ff1691b. Sep 4 17:20:00.097007 systemd-resolved[1309]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 4 17:20:00.109456 sshd[4376]: pam_unix(sshd:session): session closed for user core Sep 4 17:20:00.120248 systemd[1]: sshd@10-10.0.0.60:22-10.0.0.1:60628.service: Deactivated successfully. Sep 4 17:20:00.124324 systemd[1]: session-11.scope: Deactivated successfully. Sep 4 17:20:00.128747 systemd-logind[1423]: Session 11 logged out. Waiting for processes to exit. Sep 4 17:20:00.139383 systemd[1]: Started sshd@11-10.0.0.60:22-10.0.0.1:60644.service - OpenSSH per-connection server daemon (10.0.0.1:60644). Sep 4 17:20:00.141992 containerd[1448]: time="2024-09-04T17:20:00.141924768Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-7ff458757c-hfq6x,Uid:05f65067-932e-4656-a9e0-3f22a268ebd5,Namespace:calico-system,Attempt:1,} returns sandbox id \"06f8886531fe2b285112a3dcf4561038d7e1714c54aed5c12a11f5a53ff1691b\"" Sep 4 17:20:00.142482 systemd-logind[1423]: Removed session 11. Sep 4 17:20:00.144573 containerd[1448]: time="2024-09-04T17:20:00.144536334Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.28.1\"" Sep 4 17:20:00.186088 sshd[4463]: Accepted publickey for core from 10.0.0.1 port 60644 ssh2: RSA SHA256:TcdII3DD+/vh6fGiZDuqtLwdsO9LHnvXRMQO7IdpdiA Sep 4 17:20:00.187918 sshd[4463]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 4 17:20:00.193726 systemd-logind[1423]: New session 12 of user core. Sep 4 17:20:00.199056 systemd[1]: Started session-12.scope - Session 12 of User core. Sep 4 17:20:00.373799 sshd[4463]: pam_unix(sshd:session): session closed for user core Sep 4 17:20:00.380323 systemd[1]: sshd@11-10.0.0.60:22-10.0.0.1:60644.service: Deactivated successfully. Sep 4 17:20:00.383773 systemd[1]: session-12.scope: Deactivated successfully. Sep 4 17:20:00.387458 systemd-logind[1423]: Session 12 logged out. Waiting for processes to exit. Sep 4 17:20:00.388315 systemd-logind[1423]: Removed session 12. Sep 4 17:20:00.589053 kubelet[2523]: I0904 17:20:00.588464 2523 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 4 17:20:00.589945 kubelet[2523]: E0904 17:20:00.589922 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:20:00.714270 containerd[1448]: time="2024-09-04T17:20:00.714180618Z" level=info msg="StopPodSandbox for \"ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954\"" Sep 4 17:20:00.807510 containerd[1448]: 2024-09-04 17:20:00.762 [INFO][4525] k8s.go 608: Cleaning up netns ContainerID="ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" Sep 4 17:20:00.807510 containerd[1448]: 2024-09-04 17:20:00.762 [INFO][4525] dataplane_linux.go 530: Deleting workload's device in netns. ContainerID="ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" iface="eth0" netns="/var/run/netns/cni-167084cc-d756-76a3-5b08-6cac904456d7" Sep 4 17:20:00.807510 containerd[1448]: 2024-09-04 17:20:00.762 [INFO][4525] dataplane_linux.go 541: Entered netns, deleting veth. ContainerID="ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" iface="eth0" netns="/var/run/netns/cni-167084cc-d756-76a3-5b08-6cac904456d7" Sep 4 17:20:00.807510 containerd[1448]: 2024-09-04 17:20:00.762 [INFO][4525] dataplane_linux.go 568: Workload's veth was already gone. Nothing to do. ContainerID="ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" iface="eth0" netns="/var/run/netns/cni-167084cc-d756-76a3-5b08-6cac904456d7" Sep 4 17:20:00.807510 containerd[1448]: 2024-09-04 17:20:00.762 [INFO][4525] k8s.go 615: Releasing IP address(es) ContainerID="ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" Sep 4 17:20:00.807510 containerd[1448]: 2024-09-04 17:20:00.762 [INFO][4525] utils.go 188: Calico CNI releasing IP address ContainerID="ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" Sep 4 17:20:00.807510 containerd[1448]: 2024-09-04 17:20:00.792 [INFO][4533] ipam_plugin.go 417: Releasing address using handleID ContainerID="ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" HandleID="k8s-pod-network.ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" Workload="localhost-k8s-csi--node--driver--nsxnh-eth0" Sep 4 17:20:00.807510 containerd[1448]: 2024-09-04 17:20:00.792 [INFO][4533] ipam_plugin.go 358: About to acquire host-wide IPAM lock. Sep 4 17:20:00.807510 containerd[1448]: 2024-09-04 17:20:00.792 [INFO][4533] ipam_plugin.go 373: Acquired host-wide IPAM lock. Sep 4 17:20:00.807510 containerd[1448]: 2024-09-04 17:20:00.802 [WARNING][4533] ipam_plugin.go 434: Asked to release address but it doesn't exist. Ignoring ContainerID="ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" HandleID="k8s-pod-network.ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" Workload="localhost-k8s-csi--node--driver--nsxnh-eth0" Sep 4 17:20:00.807510 containerd[1448]: 2024-09-04 17:20:00.802 [INFO][4533] ipam_plugin.go 445: Releasing address using workloadID ContainerID="ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" HandleID="k8s-pod-network.ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" Workload="localhost-k8s-csi--node--driver--nsxnh-eth0" Sep 4 17:20:00.807510 containerd[1448]: 2024-09-04 17:20:00.804 [INFO][4533] ipam_plugin.go 379: Released host-wide IPAM lock. Sep 4 17:20:00.807510 containerd[1448]: 2024-09-04 17:20:00.806 [INFO][4525] k8s.go 621: Teardown processing complete. ContainerID="ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" Sep 4 17:20:00.808530 containerd[1448]: time="2024-09-04T17:20:00.808391399Z" level=info msg="TearDown network for sandbox \"ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954\" successfully" Sep 4 17:20:00.808530 containerd[1448]: time="2024-09-04T17:20:00.808427480Z" level=info msg="StopPodSandbox for \"ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954\" returns successfully" Sep 4 17:20:00.809305 containerd[1448]: time="2024-09-04T17:20:00.809268535Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-nsxnh,Uid:2e2776ed-fa4e-45b8-bb3f-17de64345308,Namespace:calico-system,Attempt:1,}" Sep 4 17:20:00.840463 systemd[1]: run-netns-cni\x2d167084cc\x2dd756\x2d76a3\x2d5b08\x2d6cac904456d7.mount: Deactivated successfully. Sep 4 17:20:00.939459 systemd-networkd[1377]: cali39027bddf0e: Link UP Sep 4 17:20:00.939790 systemd-networkd[1377]: cali39027bddf0e: Gained carrier Sep 4 17:20:00.953881 containerd[1448]: 2024-09-04 17:20:00.851 [INFO][4541] utils.go 100: File /var/lib/calico/mtu does not exist Sep 4 17:20:00.953881 containerd[1448]: 2024-09-04 17:20:00.866 [INFO][4541] plugin.go 326: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-csi--node--driver--nsxnh-eth0 csi-node-driver- calico-system 2e2776ed-fa4e-45b8-bb3f-17de64345308 895 0 2024-09-04 17:19:35 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:78cd84fb8c k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:default] map[] [] [] []} {k8s localhost csi-node-driver-nsxnh eth0 default [] [] [kns.calico-system ksa.calico-system.default] cali39027bddf0e [] []}} ContainerID="3b0cf9c39426749bc59147d0533e7ec0f9828bcf644887bfd3fb814581142291" Namespace="calico-system" Pod="csi-node-driver-nsxnh" WorkloadEndpoint="localhost-k8s-csi--node--driver--nsxnh-" Sep 4 17:20:00.953881 containerd[1448]: 2024-09-04 17:20:00.866 [INFO][4541] k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="3b0cf9c39426749bc59147d0533e7ec0f9828bcf644887bfd3fb814581142291" Namespace="calico-system" Pod="csi-node-driver-nsxnh" WorkloadEndpoint="localhost-k8s-csi--node--driver--nsxnh-eth0" Sep 4 17:20:00.953881 containerd[1448]: 2024-09-04 17:20:00.893 [INFO][4554] ipam_plugin.go 230: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="3b0cf9c39426749bc59147d0533e7ec0f9828bcf644887bfd3fb814581142291" HandleID="k8s-pod-network.3b0cf9c39426749bc59147d0533e7ec0f9828bcf644887bfd3fb814581142291" Workload="localhost-k8s-csi--node--driver--nsxnh-eth0" Sep 4 17:20:00.953881 containerd[1448]: 2024-09-04 17:20:00.904 [INFO][4554] ipam_plugin.go 270: Auto assigning IP ContainerID="3b0cf9c39426749bc59147d0533e7ec0f9828bcf644887bfd3fb814581142291" HandleID="k8s-pod-network.3b0cf9c39426749bc59147d0533e7ec0f9828bcf644887bfd3fb814581142291" Workload="localhost-k8s-csi--node--driver--nsxnh-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002f2c40), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"csi-node-driver-nsxnh", "timestamp":"2024-09-04 17:20:00.893446499 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 4 17:20:00.953881 containerd[1448]: 2024-09-04 17:20:00.904 [INFO][4554] ipam_plugin.go 358: About to acquire host-wide IPAM lock. Sep 4 17:20:00.953881 containerd[1448]: 2024-09-04 17:20:00.904 [INFO][4554] ipam_plugin.go 373: Acquired host-wide IPAM lock. Sep 4 17:20:00.953881 containerd[1448]: 2024-09-04 17:20:00.904 [INFO][4554] ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 4 17:20:00.953881 containerd[1448]: 2024-09-04 17:20:00.907 [INFO][4554] ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.3b0cf9c39426749bc59147d0533e7ec0f9828bcf644887bfd3fb814581142291" host="localhost" Sep 4 17:20:00.953881 containerd[1448]: 2024-09-04 17:20:00.911 [INFO][4554] ipam.go 372: Looking up existing affinities for host host="localhost" Sep 4 17:20:00.953881 containerd[1448]: 2024-09-04 17:20:00.916 [INFO][4554] ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Sep 4 17:20:00.953881 containerd[1448]: 2024-09-04 17:20:00.918 [INFO][4554] ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 4 17:20:00.953881 containerd[1448]: 2024-09-04 17:20:00.921 [INFO][4554] ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 4 17:20:00.953881 containerd[1448]: 2024-09-04 17:20:00.921 [INFO][4554] ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.3b0cf9c39426749bc59147d0533e7ec0f9828bcf644887bfd3fb814581142291" host="localhost" Sep 4 17:20:00.953881 containerd[1448]: 2024-09-04 17:20:00.923 [INFO][4554] ipam.go 1685: Creating new handle: k8s-pod-network.3b0cf9c39426749bc59147d0533e7ec0f9828bcf644887bfd3fb814581142291 Sep 4 17:20:00.953881 containerd[1448]: 2024-09-04 17:20:00.928 [INFO][4554] ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.3b0cf9c39426749bc59147d0533e7ec0f9828bcf644887bfd3fb814581142291" host="localhost" Sep 4 17:20:00.953881 containerd[1448]: 2024-09-04 17:20:00.935 [INFO][4554] ipam.go 1216: Successfully claimed IPs: [192.168.88.132/26] block=192.168.88.128/26 handle="k8s-pod-network.3b0cf9c39426749bc59147d0533e7ec0f9828bcf644887bfd3fb814581142291" host="localhost" Sep 4 17:20:00.953881 containerd[1448]: 2024-09-04 17:20:00.935 [INFO][4554] ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.132/26] handle="k8s-pod-network.3b0cf9c39426749bc59147d0533e7ec0f9828bcf644887bfd3fb814581142291" host="localhost" Sep 4 17:20:00.953881 containerd[1448]: 2024-09-04 17:20:00.935 [INFO][4554] ipam_plugin.go 379: Released host-wide IPAM lock. Sep 4 17:20:00.953881 containerd[1448]: 2024-09-04 17:20:00.935 [INFO][4554] ipam_plugin.go 288: Calico CNI IPAM assigned addresses IPv4=[192.168.88.132/26] IPv6=[] ContainerID="3b0cf9c39426749bc59147d0533e7ec0f9828bcf644887bfd3fb814581142291" HandleID="k8s-pod-network.3b0cf9c39426749bc59147d0533e7ec0f9828bcf644887bfd3fb814581142291" Workload="localhost-k8s-csi--node--driver--nsxnh-eth0" Sep 4 17:20:00.954439 containerd[1448]: 2024-09-04 17:20:00.938 [INFO][4541] k8s.go 386: Populated endpoint ContainerID="3b0cf9c39426749bc59147d0533e7ec0f9828bcf644887bfd3fb814581142291" Namespace="calico-system" Pod="csi-node-driver-nsxnh" WorkloadEndpoint="localhost-k8s-csi--node--driver--nsxnh-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--nsxnh-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"2e2776ed-fa4e-45b8-bb3f-17de64345308", ResourceVersion:"895", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 19, 35, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"78cd84fb8c", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"default"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"csi-node-driver-nsxnh", Endpoint:"eth0", ServiceAccountName:"default", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.default"}, InterfaceName:"cali39027bddf0e", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:20:00.954439 containerd[1448]: 2024-09-04 17:20:00.938 [INFO][4541] k8s.go 387: Calico CNI using IPs: [192.168.88.132/32] ContainerID="3b0cf9c39426749bc59147d0533e7ec0f9828bcf644887bfd3fb814581142291" Namespace="calico-system" Pod="csi-node-driver-nsxnh" WorkloadEndpoint="localhost-k8s-csi--node--driver--nsxnh-eth0" Sep 4 17:20:00.954439 containerd[1448]: 2024-09-04 17:20:00.938 [INFO][4541] dataplane_linux.go 68: Setting the host side veth name to cali39027bddf0e ContainerID="3b0cf9c39426749bc59147d0533e7ec0f9828bcf644887bfd3fb814581142291" Namespace="calico-system" Pod="csi-node-driver-nsxnh" WorkloadEndpoint="localhost-k8s-csi--node--driver--nsxnh-eth0" Sep 4 17:20:00.954439 containerd[1448]: 2024-09-04 17:20:00.939 [INFO][4541] dataplane_linux.go 479: Disabling IPv4 forwarding ContainerID="3b0cf9c39426749bc59147d0533e7ec0f9828bcf644887bfd3fb814581142291" Namespace="calico-system" Pod="csi-node-driver-nsxnh" WorkloadEndpoint="localhost-k8s-csi--node--driver--nsxnh-eth0" Sep 4 17:20:00.954439 containerd[1448]: 2024-09-04 17:20:00.940 [INFO][4541] k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="3b0cf9c39426749bc59147d0533e7ec0f9828bcf644887bfd3fb814581142291" Namespace="calico-system" Pod="csi-node-driver-nsxnh" WorkloadEndpoint="localhost-k8s-csi--node--driver--nsxnh-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--nsxnh-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"2e2776ed-fa4e-45b8-bb3f-17de64345308", ResourceVersion:"895", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 19, 35, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"78cd84fb8c", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"default"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"3b0cf9c39426749bc59147d0533e7ec0f9828bcf644887bfd3fb814581142291", Pod:"csi-node-driver-nsxnh", Endpoint:"eth0", ServiceAccountName:"default", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.default"}, InterfaceName:"cali39027bddf0e", MAC:"de:ae:47:f2:47:6d", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:20:00.954439 containerd[1448]: 2024-09-04 17:20:00.951 [INFO][4541] k8s.go 500: Wrote updated endpoint to datastore ContainerID="3b0cf9c39426749bc59147d0533e7ec0f9828bcf644887bfd3fb814581142291" Namespace="calico-system" Pod="csi-node-driver-nsxnh" WorkloadEndpoint="localhost-k8s-csi--node--driver--nsxnh-eth0" Sep 4 17:20:00.956612 kubelet[2523]: E0904 17:20:00.956443 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:20:00.976820 containerd[1448]: time="2024-09-04T17:20:00.975946234Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 4 17:20:00.976820 containerd[1448]: time="2024-09-04T17:20:00.976442802Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:20:00.976820 containerd[1448]: time="2024-09-04T17:20:00.976460723Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 4 17:20:00.976820 containerd[1448]: time="2024-09-04T17:20:00.976470483Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:20:01.000130 systemd[1]: Started cri-containerd-3b0cf9c39426749bc59147d0533e7ec0f9828bcf644887bfd3fb814581142291.scope - libcontainer container 3b0cf9c39426749bc59147d0533e7ec0f9828bcf644887bfd3fb814581142291. Sep 4 17:20:01.011805 systemd-resolved[1309]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 4 17:20:01.024999 containerd[1448]: time="2024-09-04T17:20:01.024946651Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-nsxnh,Uid:2e2776ed-fa4e-45b8-bb3f-17de64345308,Namespace:calico-system,Attempt:1,} returns sandbox id \"3b0cf9c39426749bc59147d0533e7ec0f9828bcf644887bfd3fb814581142291\"" Sep 4 17:20:01.079035 systemd-networkd[1377]: cali20d0faf09d3: Gained IPv6LL Sep 4 17:20:01.174276 kernel: bpftool[4629]: memfd_create() called without MFD_EXEC or MFD_NOEXEC_SEAL set Sep 4 17:20:01.357760 systemd-networkd[1377]: vxlan.calico: Link UP Sep 4 17:20:01.357767 systemd-networkd[1377]: vxlan.calico: Gained carrier Sep 4 17:20:01.941712 containerd[1448]: time="2024-09-04T17:20:01.941653319Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.28.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:20:01.942271 containerd[1448]: time="2024-09-04T17:20:01.942173128Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.28.1: active requests=0, bytes read=31361753" Sep 4 17:20:01.943321 containerd[1448]: time="2024-09-04T17:20:01.943286588Z" level=info msg="ImageCreate event name:\"sha256:dde0e0aa888dfe01de8f2b6b4879c4391e01cc95a7a8a608194d8ed663fe6a39\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:20:01.945502 containerd[1448]: time="2024-09-04T17:20:01.945449145Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:9a7338f7187d4d2352fe49eedee44b191ac92557a2e71aa3de3527ed85c1641b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:20:01.946514 containerd[1448]: time="2024-09-04T17:20:01.946473443Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.28.1\" with image id \"sha256:dde0e0aa888dfe01de8f2b6b4879c4391e01cc95a7a8a608194d8ed663fe6a39\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.28.1\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:9a7338f7187d4d2352fe49eedee44b191ac92557a2e71aa3de3527ed85c1641b\", size \"32729240\" in 1.801876188s" Sep 4 17:20:01.946514 containerd[1448]: time="2024-09-04T17:20:01.946514124Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.28.1\" returns image reference \"sha256:dde0e0aa888dfe01de8f2b6b4879c4391e01cc95a7a8a608194d8ed663fe6a39\"" Sep 4 17:20:01.947616 containerd[1448]: time="2024-09-04T17:20:01.947571822Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.28.1\"" Sep 4 17:20:01.959563 containerd[1448]: time="2024-09-04T17:20:01.959141783Z" level=info msg="CreateContainer within sandbox \"06f8886531fe2b285112a3dcf4561038d7e1714c54aed5c12a11f5a53ff1691b\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Sep 4 17:20:01.983818 containerd[1448]: time="2024-09-04T17:20:01.983611487Z" level=info msg="CreateContainer within sandbox \"06f8886531fe2b285112a3dcf4561038d7e1714c54aed5c12a11f5a53ff1691b\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"eb381b157f01514333d7c767cf87991cb85d403f673469a17f2fa7712fc1fbe3\"" Sep 4 17:20:01.984255 containerd[1448]: time="2024-09-04T17:20:01.984222738Z" level=info msg="StartContainer for \"eb381b157f01514333d7c767cf87991cb85d403f673469a17f2fa7712fc1fbe3\"" Sep 4 17:20:02.010054 systemd[1]: Started cri-containerd-eb381b157f01514333d7c767cf87991cb85d403f673469a17f2fa7712fc1fbe3.scope - libcontainer container eb381b157f01514333d7c767cf87991cb85d403f673469a17f2fa7712fc1fbe3. Sep 4 17:20:02.049826 containerd[1448]: time="2024-09-04T17:20:02.049769223Z" level=info msg="StartContainer for \"eb381b157f01514333d7c767cf87991cb85d403f673469a17f2fa7712fc1fbe3\" returns successfully" Sep 4 17:20:02.869981 systemd-networkd[1377]: cali39027bddf0e: Gained IPv6LL Sep 4 17:20:02.871016 systemd-networkd[1377]: vxlan.calico: Gained IPv6LL Sep 4 17:20:02.983680 kubelet[2523]: I0904 17:20:02.983622 2523 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-7ff458757c-hfq6x" podStartSLOduration=26.180631696 podStartE2EDuration="27.983444941s" podCreationTimestamp="2024-09-04 17:19:35 +0000 UTC" firstStartedPulling="2024-09-04 17:20:00.143998204 +0000 UTC m=+45.564911927" lastFinishedPulling="2024-09-04 17:20:01.946811409 +0000 UTC m=+47.367725172" observedRunningTime="2024-09-04 17:20:02.9833633 +0000 UTC m=+48.404277063" watchObservedRunningTime="2024-09-04 17:20:02.983444941 +0000 UTC m=+48.404358664" Sep 4 17:20:03.039367 containerd[1448]: time="2024-09-04T17:20:03.038006265Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.28.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:20:03.039367 containerd[1448]: time="2024-09-04T17:20:03.038779718Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.28.1: active requests=0, bytes read=7211060" Sep 4 17:20:03.039776 containerd[1448]: time="2024-09-04T17:20:03.039688893Z" level=info msg="ImageCreate event name:\"sha256:dd6cf4bf9b3656f9dd9713f21ac1be96858f750a9a3bf340983fb7072f4eda2a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:20:03.042513 containerd[1448]: time="2024-09-04T17:20:03.042473740Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:01e16d03dd0c29a8e1e302455eb15c2d0326c49cbaca4bbe8dc0e2d5308c5add\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:20:03.043288 containerd[1448]: time="2024-09-04T17:20:03.043208793Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.28.1\" with image id \"sha256:dd6cf4bf9b3656f9dd9713f21ac1be96858f750a9a3bf340983fb7072f4eda2a\", repo tag \"ghcr.io/flatcar/calico/csi:v3.28.1\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:01e16d03dd0c29a8e1e302455eb15c2d0326c49cbaca4bbe8dc0e2d5308c5add\", size \"8578579\" in 1.09559237s" Sep 4 17:20:03.043288 containerd[1448]: time="2024-09-04T17:20:03.043248873Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.28.1\" returns image reference \"sha256:dd6cf4bf9b3656f9dd9713f21ac1be96858f750a9a3bf340983fb7072f4eda2a\"" Sep 4 17:20:03.045559 containerd[1448]: time="2024-09-04T17:20:03.044846860Z" level=info msg="CreateContainer within sandbox \"3b0cf9c39426749bc59147d0533e7ec0f9828bcf644887bfd3fb814581142291\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Sep 4 17:20:03.063373 containerd[1448]: time="2024-09-04T17:20:03.063321251Z" level=info msg="CreateContainer within sandbox \"3b0cf9c39426749bc59147d0533e7ec0f9828bcf644887bfd3fb814581142291\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"9c0b7ad759bada06bcf1cf25f5d69799103b34ac4f670a1b3a92a8da45ea0b74\"" Sep 4 17:20:03.065203 containerd[1448]: time="2024-09-04T17:20:03.064586473Z" level=info msg="StartContainer for \"9c0b7ad759bada06bcf1cf25f5d69799103b34ac4f670a1b3a92a8da45ea0b74\"" Sep 4 17:20:03.099052 systemd[1]: Started cri-containerd-9c0b7ad759bada06bcf1cf25f5d69799103b34ac4f670a1b3a92a8da45ea0b74.scope - libcontainer container 9c0b7ad759bada06bcf1cf25f5d69799103b34ac4f670a1b3a92a8da45ea0b74. Sep 4 17:20:03.124025 containerd[1448]: time="2024-09-04T17:20:03.123894952Z" level=info msg="StartContainer for \"9c0b7ad759bada06bcf1cf25f5d69799103b34ac4f670a1b3a92a8da45ea0b74\" returns successfully" Sep 4 17:20:03.125774 containerd[1448]: time="2024-09-04T17:20:03.125738863Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.28.1\"" Sep 4 17:20:04.649149 containerd[1448]: time="2024-09-04T17:20:04.649074778Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.28.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:20:04.650011 containerd[1448]: time="2024-09-04T17:20:04.649960313Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.28.1: active requests=0, bytes read=12116870" Sep 4 17:20:04.651110 containerd[1448]: time="2024-09-04T17:20:04.651071012Z" level=info msg="ImageCreate event name:\"sha256:4df800f2dc90e056e3dc95be5afe5cd399ce8785c6817ddeaf07b498cb85207a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:20:04.654147 containerd[1448]: time="2024-09-04T17:20:04.654089622Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.28.1\" with image id \"sha256:4df800f2dc90e056e3dc95be5afe5cd399ce8785c6817ddeaf07b498cb85207a\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.28.1\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:682cc97e4580d25b7314032c008a552bb05182fac34eba82cc389113c7767076\", size \"13484341\" in 1.528307678s" Sep 4 17:20:04.654147 containerd[1448]: time="2024-09-04T17:20:04.654146583Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.28.1\" returns image reference \"sha256:4df800f2dc90e056e3dc95be5afe5cd399ce8785c6817ddeaf07b498cb85207a\"" Sep 4 17:20:04.654300 containerd[1448]: time="2024-09-04T17:20:04.654086342Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:682cc97e4580d25b7314032c008a552bb05182fac34eba82cc389113c7767076\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:20:04.656175 containerd[1448]: time="2024-09-04T17:20:04.656133496Z" level=info msg="CreateContainer within sandbox \"3b0cf9c39426749bc59147d0533e7ec0f9828bcf644887bfd3fb814581142291\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Sep 4 17:20:04.672870 containerd[1448]: time="2024-09-04T17:20:04.672791213Z" level=info msg="CreateContainer within sandbox \"3b0cf9c39426749bc59147d0533e7ec0f9828bcf644887bfd3fb814581142291\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"6f73a3aec54184d2d5a73cc660abe9f9a184ba04719a48e59a8141ce75e456e0\"" Sep 4 17:20:04.673566 containerd[1448]: time="2024-09-04T17:20:04.673497464Z" level=info msg="StartContainer for \"6f73a3aec54184d2d5a73cc660abe9f9a184ba04719a48e59a8141ce75e456e0\"" Sep 4 17:20:04.720084 systemd[1]: Started cri-containerd-6f73a3aec54184d2d5a73cc660abe9f9a184ba04719a48e59a8141ce75e456e0.scope - libcontainer container 6f73a3aec54184d2d5a73cc660abe9f9a184ba04719a48e59a8141ce75e456e0. Sep 4 17:20:04.755411 containerd[1448]: time="2024-09-04T17:20:04.755348384Z" level=info msg="StartContainer for \"6f73a3aec54184d2d5a73cc660abe9f9a184ba04719a48e59a8141ce75e456e0\" returns successfully" Sep 4 17:20:04.986116 kubelet[2523]: I0904 17:20:04.985880 2523 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="calico-system/csi-node-driver-nsxnh" podStartSLOduration=26.357681301 podStartE2EDuration="29.985807734s" podCreationTimestamp="2024-09-04 17:19:35 +0000 UTC" firstStartedPulling="2024-09-04 17:20:01.026316435 +0000 UTC m=+46.447230158" lastFinishedPulling="2024-09-04 17:20:04.654442868 +0000 UTC m=+50.075356591" observedRunningTime="2024-09-04 17:20:04.984842558 +0000 UTC m=+50.405756321" watchObservedRunningTime="2024-09-04 17:20:04.985807734 +0000 UTC m=+50.406721497" Sep 4 17:20:05.388869 systemd[1]: Started sshd@12-10.0.0.60:22-10.0.0.1:48148.service - OpenSSH per-connection server daemon (10.0.0.1:48148). Sep 4 17:20:05.455684 sshd[4898]: Accepted publickey for core from 10.0.0.1 port 48148 ssh2: RSA SHA256:TcdII3DD+/vh6fGiZDuqtLwdsO9LHnvXRMQO7IdpdiA Sep 4 17:20:05.457610 sshd[4898]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 4 17:20:05.461755 systemd-logind[1423]: New session 13 of user core. Sep 4 17:20:05.467058 systemd[1]: Started session-13.scope - Session 13 of User core. Sep 4 17:20:05.633706 sshd[4898]: pam_unix(sshd:session): session closed for user core Sep 4 17:20:05.646672 systemd[1]: sshd@12-10.0.0.60:22-10.0.0.1:48148.service: Deactivated successfully. Sep 4 17:20:05.654998 systemd[1]: session-13.scope: Deactivated successfully. Sep 4 17:20:05.656010 systemd-logind[1423]: Session 13 logged out. Waiting for processes to exit. Sep 4 17:20:05.667458 systemd[1]: Started sshd@13-10.0.0.60:22-10.0.0.1:48160.service - OpenSSH per-connection server daemon (10.0.0.1:48160). Sep 4 17:20:05.670240 systemd-logind[1423]: Removed session 13. Sep 4 17:20:05.710500 sshd[4918]: Accepted publickey for core from 10.0.0.1 port 48160 ssh2: RSA SHA256:TcdII3DD+/vh6fGiZDuqtLwdsO9LHnvXRMQO7IdpdiA Sep 4 17:20:05.713229 sshd[4918]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 4 17:20:05.718751 systemd-logind[1423]: New session 14 of user core. Sep 4 17:20:05.726650 systemd[1]: Started session-14.scope - Session 14 of User core. Sep 4 17:20:05.827696 kubelet[2523]: I0904 17:20:05.827596 2523 csi_plugin.go:99] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Sep 4 17:20:05.835868 kubelet[2523]: I0904 17:20:05.835816 2523 csi_plugin.go:112] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Sep 4 17:20:06.030108 sshd[4918]: pam_unix(sshd:session): session closed for user core Sep 4 17:20:06.035908 systemd[1]: sshd@13-10.0.0.60:22-10.0.0.1:48160.service: Deactivated successfully. Sep 4 17:20:06.038114 systemd[1]: session-14.scope: Deactivated successfully. Sep 4 17:20:06.039221 systemd-logind[1423]: Session 14 logged out. Waiting for processes to exit. Sep 4 17:20:06.048985 systemd[1]: Started sshd@14-10.0.0.60:22-10.0.0.1:48176.service - OpenSSH per-connection server daemon (10.0.0.1:48176). Sep 4 17:20:06.051687 systemd-logind[1423]: Removed session 14. Sep 4 17:20:06.092365 sshd[4931]: Accepted publickey for core from 10.0.0.1 port 48176 ssh2: RSA SHA256:TcdII3DD+/vh6fGiZDuqtLwdsO9LHnvXRMQO7IdpdiA Sep 4 17:20:06.093943 sshd[4931]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 4 17:20:06.103403 systemd-logind[1423]: New session 15 of user core. Sep 4 17:20:06.109063 systemd[1]: Started session-15.scope - Session 15 of User core. Sep 4 17:20:07.697981 sshd[4931]: pam_unix(sshd:session): session closed for user core Sep 4 17:20:07.708392 systemd[1]: sshd@14-10.0.0.60:22-10.0.0.1:48176.service: Deactivated successfully. Sep 4 17:20:07.712305 systemd[1]: session-15.scope: Deactivated successfully. Sep 4 17:20:07.715195 systemd-logind[1423]: Session 15 logged out. Waiting for processes to exit. Sep 4 17:20:07.727388 systemd[1]: Started sshd@15-10.0.0.60:22-10.0.0.1:48228.service - OpenSSH per-connection server daemon (10.0.0.1:48228). Sep 4 17:20:07.729444 systemd-logind[1423]: Removed session 15. Sep 4 17:20:07.780247 sshd[4958]: Accepted publickey for core from 10.0.0.1 port 48228 ssh2: RSA SHA256:TcdII3DD+/vh6fGiZDuqtLwdsO9LHnvXRMQO7IdpdiA Sep 4 17:20:07.782189 sshd[4958]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 4 17:20:07.787427 systemd-logind[1423]: New session 16 of user core. Sep 4 17:20:07.793034 systemd[1]: Started session-16.scope - Session 16 of User core. Sep 4 17:20:08.133596 sshd[4958]: pam_unix(sshd:session): session closed for user core Sep 4 17:20:08.142896 systemd[1]: sshd@15-10.0.0.60:22-10.0.0.1:48228.service: Deactivated successfully. Sep 4 17:20:08.145646 systemd[1]: session-16.scope: Deactivated successfully. Sep 4 17:20:08.148248 systemd-logind[1423]: Session 16 logged out. Waiting for processes to exit. Sep 4 17:20:08.158467 systemd[1]: Started sshd@16-10.0.0.60:22-10.0.0.1:48244.service - OpenSSH per-connection server daemon (10.0.0.1:48244). Sep 4 17:20:08.159950 systemd-logind[1423]: Removed session 16. Sep 4 17:20:08.199965 sshd[4971]: Accepted publickey for core from 10.0.0.1 port 48244 ssh2: RSA SHA256:TcdII3DD+/vh6fGiZDuqtLwdsO9LHnvXRMQO7IdpdiA Sep 4 17:20:08.201339 sshd[4971]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 4 17:20:08.206618 systemd-logind[1423]: New session 17 of user core. Sep 4 17:20:08.217076 systemd[1]: Started session-17.scope - Session 17 of User core. Sep 4 17:20:08.374537 sshd[4971]: pam_unix(sshd:session): session closed for user core Sep 4 17:20:08.377903 systemd[1]: sshd@16-10.0.0.60:22-10.0.0.1:48244.service: Deactivated successfully. Sep 4 17:20:08.379946 systemd[1]: session-17.scope: Deactivated successfully. Sep 4 17:20:08.381547 systemd-logind[1423]: Session 17 logged out. Waiting for processes to exit. Sep 4 17:20:08.383307 systemd-logind[1423]: Removed session 17. Sep 4 17:20:13.385622 systemd[1]: Started sshd@17-10.0.0.60:22-10.0.0.1:59798.service - OpenSSH per-connection server daemon (10.0.0.1:59798). Sep 4 17:20:13.426075 sshd[5002]: Accepted publickey for core from 10.0.0.1 port 59798 ssh2: RSA SHA256:TcdII3DD+/vh6fGiZDuqtLwdsO9LHnvXRMQO7IdpdiA Sep 4 17:20:13.427540 sshd[5002]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 4 17:20:13.431219 systemd-logind[1423]: New session 18 of user core. Sep 4 17:20:13.442054 systemd[1]: Started session-18.scope - Session 18 of User core. Sep 4 17:20:13.589387 sshd[5002]: pam_unix(sshd:session): session closed for user core Sep 4 17:20:13.592351 systemd[1]: sshd@17-10.0.0.60:22-10.0.0.1:59798.service: Deactivated successfully. Sep 4 17:20:13.595164 systemd-logind[1423]: Session 18 logged out. Waiting for processes to exit. Sep 4 17:20:13.595467 systemd[1]: session-18.scope: Deactivated successfully. Sep 4 17:20:13.596693 systemd-logind[1423]: Removed session 18. Sep 4 17:20:14.717700 containerd[1448]: time="2024-09-04T17:20:14.716843821Z" level=info msg="StopPodSandbox for \"0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be\"" Sep 4 17:20:14.813017 containerd[1448]: 2024-09-04 17:20:14.770 [WARNING][5030] k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--7ff458757c--hfq6x-eth0", GenerateName:"calico-kube-controllers-7ff458757c-", Namespace:"calico-system", SelfLink:"", UID:"05f65067-932e-4656-a9e0-3f22a268ebd5", ResourceVersion:"919", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 19, 35, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"7ff458757c", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"06f8886531fe2b285112a3dcf4561038d7e1714c54aed5c12a11f5a53ff1691b", Pod:"calico-kube-controllers-7ff458757c-hfq6x", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali20d0faf09d3", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:20:14.813017 containerd[1448]: 2024-09-04 17:20:14.770 [INFO][5030] k8s.go 608: Cleaning up netns ContainerID="0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" Sep 4 17:20:14.813017 containerd[1448]: 2024-09-04 17:20:14.770 [INFO][5030] dataplane_linux.go 526: CleanUpNamespace called with no netns name, ignoring. ContainerID="0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" iface="eth0" netns="" Sep 4 17:20:14.813017 containerd[1448]: 2024-09-04 17:20:14.770 [INFO][5030] k8s.go 615: Releasing IP address(es) ContainerID="0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" Sep 4 17:20:14.813017 containerd[1448]: 2024-09-04 17:20:14.770 [INFO][5030] utils.go 188: Calico CNI releasing IP address ContainerID="0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" Sep 4 17:20:14.813017 containerd[1448]: 2024-09-04 17:20:14.798 [INFO][5040] ipam_plugin.go 417: Releasing address using handleID ContainerID="0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" HandleID="k8s-pod-network.0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" Workload="localhost-k8s-calico--kube--controllers--7ff458757c--hfq6x-eth0" Sep 4 17:20:14.813017 containerd[1448]: 2024-09-04 17:20:14.798 [INFO][5040] ipam_plugin.go 358: About to acquire host-wide IPAM lock. Sep 4 17:20:14.813017 containerd[1448]: 2024-09-04 17:20:14.798 [INFO][5040] ipam_plugin.go 373: Acquired host-wide IPAM lock. Sep 4 17:20:14.813017 containerd[1448]: 2024-09-04 17:20:14.807 [WARNING][5040] ipam_plugin.go 434: Asked to release address but it doesn't exist. Ignoring ContainerID="0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" HandleID="k8s-pod-network.0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" Workload="localhost-k8s-calico--kube--controllers--7ff458757c--hfq6x-eth0" Sep 4 17:20:14.813017 containerd[1448]: 2024-09-04 17:20:14.807 [INFO][5040] ipam_plugin.go 445: Releasing address using workloadID ContainerID="0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" HandleID="k8s-pod-network.0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" Workload="localhost-k8s-calico--kube--controllers--7ff458757c--hfq6x-eth0" Sep 4 17:20:14.813017 containerd[1448]: 2024-09-04 17:20:14.809 [INFO][5040] ipam_plugin.go 379: Released host-wide IPAM lock. Sep 4 17:20:14.813017 containerd[1448]: 2024-09-04 17:20:14.811 [INFO][5030] k8s.go 621: Teardown processing complete. ContainerID="0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" Sep 4 17:20:14.813478 containerd[1448]: time="2024-09-04T17:20:14.813057771Z" level=info msg="TearDown network for sandbox \"0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be\" successfully" Sep 4 17:20:14.813478 containerd[1448]: time="2024-09-04T17:20:14.813089612Z" level=info msg="StopPodSandbox for \"0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be\" returns successfully" Sep 4 17:20:14.813824 containerd[1448]: time="2024-09-04T17:20:14.813798100Z" level=info msg="RemovePodSandbox for \"0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be\"" Sep 4 17:20:14.822213 containerd[1448]: time="2024-09-04T17:20:14.818159274Z" level=info msg="Forcibly stopping sandbox \"0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be\"" Sep 4 17:20:14.893695 containerd[1448]: 2024-09-04 17:20:14.858 [WARNING][5064] k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--7ff458757c--hfq6x-eth0", GenerateName:"calico-kube-controllers-7ff458757c-", Namespace:"calico-system", SelfLink:"", UID:"05f65067-932e-4656-a9e0-3f22a268ebd5", ResourceVersion:"919", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 19, 35, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"7ff458757c", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"06f8886531fe2b285112a3dcf4561038d7e1714c54aed5c12a11f5a53ff1691b", Pod:"calico-kube-controllers-7ff458757c-hfq6x", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali20d0faf09d3", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:20:14.893695 containerd[1448]: 2024-09-04 17:20:14.858 [INFO][5064] k8s.go 608: Cleaning up netns ContainerID="0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" Sep 4 17:20:14.893695 containerd[1448]: 2024-09-04 17:20:14.858 [INFO][5064] dataplane_linux.go 526: CleanUpNamespace called with no netns name, ignoring. ContainerID="0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" iface="eth0" netns="" Sep 4 17:20:14.893695 containerd[1448]: 2024-09-04 17:20:14.858 [INFO][5064] k8s.go 615: Releasing IP address(es) ContainerID="0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" Sep 4 17:20:14.893695 containerd[1448]: 2024-09-04 17:20:14.858 [INFO][5064] utils.go 188: Calico CNI releasing IP address ContainerID="0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" Sep 4 17:20:14.893695 containerd[1448]: 2024-09-04 17:20:14.880 [INFO][5072] ipam_plugin.go 417: Releasing address using handleID ContainerID="0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" HandleID="k8s-pod-network.0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" Workload="localhost-k8s-calico--kube--controllers--7ff458757c--hfq6x-eth0" Sep 4 17:20:14.893695 containerd[1448]: 2024-09-04 17:20:14.880 [INFO][5072] ipam_plugin.go 358: About to acquire host-wide IPAM lock. Sep 4 17:20:14.893695 containerd[1448]: 2024-09-04 17:20:14.880 [INFO][5072] ipam_plugin.go 373: Acquired host-wide IPAM lock. Sep 4 17:20:14.893695 containerd[1448]: 2024-09-04 17:20:14.888 [WARNING][5072] ipam_plugin.go 434: Asked to release address but it doesn't exist. Ignoring ContainerID="0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" HandleID="k8s-pod-network.0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" Workload="localhost-k8s-calico--kube--controllers--7ff458757c--hfq6x-eth0" Sep 4 17:20:14.893695 containerd[1448]: 2024-09-04 17:20:14.888 [INFO][5072] ipam_plugin.go 445: Releasing address using workloadID ContainerID="0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" HandleID="k8s-pod-network.0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" Workload="localhost-k8s-calico--kube--controllers--7ff458757c--hfq6x-eth0" Sep 4 17:20:14.893695 containerd[1448]: 2024-09-04 17:20:14.890 [INFO][5072] ipam_plugin.go 379: Released host-wide IPAM lock. Sep 4 17:20:14.893695 containerd[1448]: 2024-09-04 17:20:14.891 [INFO][5064] k8s.go 621: Teardown processing complete. ContainerID="0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be" Sep 4 17:20:14.894125 containerd[1448]: time="2024-09-04T17:20:14.893728649Z" level=info msg="TearDown network for sandbox \"0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be\" successfully" Sep 4 17:20:14.902870 containerd[1448]: time="2024-09-04T17:20:14.901381744Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Sep 4 17:20:14.902870 containerd[1448]: time="2024-09-04T17:20:14.901472665Z" level=info msg="RemovePodSandbox \"0b174ff464b0e3b831d09bd016718a1f95fc962219217104e5a174e6085373be\" returns successfully" Sep 4 17:20:14.902870 containerd[1448]: time="2024-09-04T17:20:14.902105273Z" level=info msg="StopPodSandbox for \"652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2\"" Sep 4 17:20:14.978549 containerd[1448]: 2024-09-04 17:20:14.944 [WARNING][5094] k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--76f75df574--tjcbc-eth0", GenerateName:"coredns-76f75df574-", Namespace:"kube-system", SelfLink:"", UID:"5962684c-02f2-4d5e-87f8-a103e98d7b84", ResourceVersion:"835", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 19, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"76f75df574", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"ec3c18800d5f45f726125a0fd3000e067d945a8331f40df4276d6f2732459772", Pod:"coredns-76f75df574-tjcbc", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calif3acbe904d3", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:20:14.978549 containerd[1448]: 2024-09-04 17:20:14.944 [INFO][5094] k8s.go 608: Cleaning up netns ContainerID="652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" Sep 4 17:20:14.978549 containerd[1448]: 2024-09-04 17:20:14.944 [INFO][5094] dataplane_linux.go 526: CleanUpNamespace called with no netns name, ignoring. ContainerID="652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" iface="eth0" netns="" Sep 4 17:20:14.978549 containerd[1448]: 2024-09-04 17:20:14.944 [INFO][5094] k8s.go 615: Releasing IP address(es) ContainerID="652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" Sep 4 17:20:14.978549 containerd[1448]: 2024-09-04 17:20:14.944 [INFO][5094] utils.go 188: Calico CNI releasing IP address ContainerID="652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" Sep 4 17:20:14.978549 containerd[1448]: 2024-09-04 17:20:14.964 [INFO][5101] ipam_plugin.go 417: Releasing address using handleID ContainerID="652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" HandleID="k8s-pod-network.652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" Workload="localhost-k8s-coredns--76f75df574--tjcbc-eth0" Sep 4 17:20:14.978549 containerd[1448]: 2024-09-04 17:20:14.964 [INFO][5101] ipam_plugin.go 358: About to acquire host-wide IPAM lock. Sep 4 17:20:14.978549 containerd[1448]: 2024-09-04 17:20:14.964 [INFO][5101] ipam_plugin.go 373: Acquired host-wide IPAM lock. Sep 4 17:20:14.978549 containerd[1448]: 2024-09-04 17:20:14.972 [WARNING][5101] ipam_plugin.go 434: Asked to release address but it doesn't exist. Ignoring ContainerID="652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" HandleID="k8s-pod-network.652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" Workload="localhost-k8s-coredns--76f75df574--tjcbc-eth0" Sep 4 17:20:14.978549 containerd[1448]: 2024-09-04 17:20:14.972 [INFO][5101] ipam_plugin.go 445: Releasing address using workloadID ContainerID="652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" HandleID="k8s-pod-network.652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" Workload="localhost-k8s-coredns--76f75df574--tjcbc-eth0" Sep 4 17:20:14.978549 containerd[1448]: 2024-09-04 17:20:14.974 [INFO][5101] ipam_plugin.go 379: Released host-wide IPAM lock. Sep 4 17:20:14.978549 containerd[1448]: 2024-09-04 17:20:14.975 [INFO][5094] k8s.go 621: Teardown processing complete. ContainerID="652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" Sep 4 17:20:14.978549 containerd[1448]: time="2024-09-04T17:20:14.978072812Z" level=info msg="TearDown network for sandbox \"652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2\" successfully" Sep 4 17:20:14.978549 containerd[1448]: time="2024-09-04T17:20:14.978102573Z" level=info msg="StopPodSandbox for \"652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2\" returns successfully" Sep 4 17:20:14.979546 containerd[1448]: time="2024-09-04T17:20:14.979241387Z" level=info msg="RemovePodSandbox for \"652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2\"" Sep 4 17:20:14.979546 containerd[1448]: time="2024-09-04T17:20:14.979273947Z" level=info msg="Forcibly stopping sandbox \"652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2\"" Sep 4 17:20:15.052700 containerd[1448]: 2024-09-04 17:20:15.019 [WARNING][5124] k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--76f75df574--tjcbc-eth0", GenerateName:"coredns-76f75df574-", Namespace:"kube-system", SelfLink:"", UID:"5962684c-02f2-4d5e-87f8-a103e98d7b84", ResourceVersion:"835", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 19, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"76f75df574", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"ec3c18800d5f45f726125a0fd3000e067d945a8331f40df4276d6f2732459772", Pod:"coredns-76f75df574-tjcbc", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calif3acbe904d3", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:20:15.052700 containerd[1448]: 2024-09-04 17:20:15.019 [INFO][5124] k8s.go 608: Cleaning up netns ContainerID="652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" Sep 4 17:20:15.052700 containerd[1448]: 2024-09-04 17:20:15.019 [INFO][5124] dataplane_linux.go 526: CleanUpNamespace called with no netns name, ignoring. ContainerID="652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" iface="eth0" netns="" Sep 4 17:20:15.052700 containerd[1448]: 2024-09-04 17:20:15.019 [INFO][5124] k8s.go 615: Releasing IP address(es) ContainerID="652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" Sep 4 17:20:15.052700 containerd[1448]: 2024-09-04 17:20:15.019 [INFO][5124] utils.go 188: Calico CNI releasing IP address ContainerID="652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" Sep 4 17:20:15.052700 containerd[1448]: 2024-09-04 17:20:15.038 [INFO][5133] ipam_plugin.go 417: Releasing address using handleID ContainerID="652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" HandleID="k8s-pod-network.652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" Workload="localhost-k8s-coredns--76f75df574--tjcbc-eth0" Sep 4 17:20:15.052700 containerd[1448]: 2024-09-04 17:20:15.038 [INFO][5133] ipam_plugin.go 358: About to acquire host-wide IPAM lock. Sep 4 17:20:15.052700 containerd[1448]: 2024-09-04 17:20:15.038 [INFO][5133] ipam_plugin.go 373: Acquired host-wide IPAM lock. Sep 4 17:20:15.052700 containerd[1448]: 2024-09-04 17:20:15.047 [WARNING][5133] ipam_plugin.go 434: Asked to release address but it doesn't exist. Ignoring ContainerID="652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" HandleID="k8s-pod-network.652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" Workload="localhost-k8s-coredns--76f75df574--tjcbc-eth0" Sep 4 17:20:15.052700 containerd[1448]: 2024-09-04 17:20:15.047 [INFO][5133] ipam_plugin.go 445: Releasing address using workloadID ContainerID="652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" HandleID="k8s-pod-network.652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" Workload="localhost-k8s-coredns--76f75df574--tjcbc-eth0" Sep 4 17:20:15.052700 containerd[1448]: 2024-09-04 17:20:15.048 [INFO][5133] ipam_plugin.go 379: Released host-wide IPAM lock. Sep 4 17:20:15.052700 containerd[1448]: 2024-09-04 17:20:15.051 [INFO][5124] k8s.go 621: Teardown processing complete. ContainerID="652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2" Sep 4 17:20:15.053135 containerd[1448]: time="2024-09-04T17:20:15.052740997Z" level=info msg="TearDown network for sandbox \"652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2\" successfully" Sep 4 17:20:15.055534 containerd[1448]: time="2024-09-04T17:20:15.055478769Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Sep 4 17:20:15.055634 containerd[1448]: time="2024-09-04T17:20:15.055556688Z" level=info msg="RemovePodSandbox \"652eb11d3720d2beb121f071fdbd3009e21ec138cb288d94a77f0908638e86e2\" returns successfully" Sep 4 17:20:15.056121 containerd[1448]: time="2024-09-04T17:20:15.056080883Z" level=info msg="StopPodSandbox for \"53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2\"" Sep 4 17:20:15.143735 containerd[1448]: 2024-09-04 17:20:15.102 [WARNING][5155] k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--76f75df574--kldz2-eth0", GenerateName:"coredns-76f75df574-", Namespace:"kube-system", SelfLink:"", UID:"eccc4d8b-807c-4895-8200-d4b15b8caceb", ResourceVersion:"830", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 19, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"76f75df574", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"5a252404a16d8b1dae1655fd759d3c98aedc4b9bb2ad64aa9f11d3d710c53b06", Pod:"coredns-76f75df574-kldz2", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calic733e3622ac", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:20:15.143735 containerd[1448]: 2024-09-04 17:20:15.102 [INFO][5155] k8s.go 608: Cleaning up netns ContainerID="53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" Sep 4 17:20:15.143735 containerd[1448]: 2024-09-04 17:20:15.102 [INFO][5155] dataplane_linux.go 526: CleanUpNamespace called with no netns name, ignoring. ContainerID="53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" iface="eth0" netns="" Sep 4 17:20:15.143735 containerd[1448]: 2024-09-04 17:20:15.103 [INFO][5155] k8s.go 615: Releasing IP address(es) ContainerID="53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" Sep 4 17:20:15.143735 containerd[1448]: 2024-09-04 17:20:15.103 [INFO][5155] utils.go 188: Calico CNI releasing IP address ContainerID="53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" Sep 4 17:20:15.143735 containerd[1448]: 2024-09-04 17:20:15.129 [INFO][5163] ipam_plugin.go 417: Releasing address using handleID ContainerID="53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" HandleID="k8s-pod-network.53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" Workload="localhost-k8s-coredns--76f75df574--kldz2-eth0" Sep 4 17:20:15.143735 containerd[1448]: 2024-09-04 17:20:15.129 [INFO][5163] ipam_plugin.go 358: About to acquire host-wide IPAM lock. Sep 4 17:20:15.143735 containerd[1448]: 2024-09-04 17:20:15.129 [INFO][5163] ipam_plugin.go 373: Acquired host-wide IPAM lock. Sep 4 17:20:15.143735 containerd[1448]: 2024-09-04 17:20:15.138 [WARNING][5163] ipam_plugin.go 434: Asked to release address but it doesn't exist. Ignoring ContainerID="53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" HandleID="k8s-pod-network.53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" Workload="localhost-k8s-coredns--76f75df574--kldz2-eth0" Sep 4 17:20:15.143735 containerd[1448]: 2024-09-04 17:20:15.138 [INFO][5163] ipam_plugin.go 445: Releasing address using workloadID ContainerID="53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" HandleID="k8s-pod-network.53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" Workload="localhost-k8s-coredns--76f75df574--kldz2-eth0" Sep 4 17:20:15.143735 containerd[1448]: 2024-09-04 17:20:15.140 [INFO][5163] ipam_plugin.go 379: Released host-wide IPAM lock. Sep 4 17:20:15.143735 containerd[1448]: 2024-09-04 17:20:15.142 [INFO][5155] k8s.go 621: Teardown processing complete. ContainerID="53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" Sep 4 17:20:15.144213 containerd[1448]: time="2024-09-04T17:20:15.143780057Z" level=info msg="TearDown network for sandbox \"53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2\" successfully" Sep 4 17:20:15.144213 containerd[1448]: time="2024-09-04T17:20:15.143809857Z" level=info msg="StopPodSandbox for \"53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2\" returns successfully" Sep 4 17:20:15.144808 containerd[1448]: time="2024-09-04T17:20:15.144485210Z" level=info msg="RemovePodSandbox for \"53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2\"" Sep 4 17:20:15.144808 containerd[1448]: time="2024-09-04T17:20:15.144524569Z" level=info msg="Forcibly stopping sandbox \"53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2\"" Sep 4 17:20:15.212709 containerd[1448]: 2024-09-04 17:20:15.180 [WARNING][5186] k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--76f75df574--kldz2-eth0", GenerateName:"coredns-76f75df574-", Namespace:"kube-system", SelfLink:"", UID:"eccc4d8b-807c-4895-8200-d4b15b8caceb", ResourceVersion:"830", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 19, 28, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"76f75df574", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"5a252404a16d8b1dae1655fd759d3c98aedc4b9bb2ad64aa9f11d3d710c53b06", Pod:"coredns-76f75df574-kldz2", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calic733e3622ac", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:20:15.212709 containerd[1448]: 2024-09-04 17:20:15.181 [INFO][5186] k8s.go 608: Cleaning up netns ContainerID="53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" Sep 4 17:20:15.212709 containerd[1448]: 2024-09-04 17:20:15.181 [INFO][5186] dataplane_linux.go 526: CleanUpNamespace called with no netns name, ignoring. ContainerID="53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" iface="eth0" netns="" Sep 4 17:20:15.212709 containerd[1448]: 2024-09-04 17:20:15.181 [INFO][5186] k8s.go 615: Releasing IP address(es) ContainerID="53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" Sep 4 17:20:15.212709 containerd[1448]: 2024-09-04 17:20:15.181 [INFO][5186] utils.go 188: Calico CNI releasing IP address ContainerID="53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" Sep 4 17:20:15.212709 containerd[1448]: 2024-09-04 17:20:15.200 [INFO][5193] ipam_plugin.go 417: Releasing address using handleID ContainerID="53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" HandleID="k8s-pod-network.53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" Workload="localhost-k8s-coredns--76f75df574--kldz2-eth0" Sep 4 17:20:15.212709 containerd[1448]: 2024-09-04 17:20:15.200 [INFO][5193] ipam_plugin.go 358: About to acquire host-wide IPAM lock. Sep 4 17:20:15.212709 containerd[1448]: 2024-09-04 17:20:15.200 [INFO][5193] ipam_plugin.go 373: Acquired host-wide IPAM lock. Sep 4 17:20:15.212709 containerd[1448]: 2024-09-04 17:20:15.208 [WARNING][5193] ipam_plugin.go 434: Asked to release address but it doesn't exist. Ignoring ContainerID="53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" HandleID="k8s-pod-network.53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" Workload="localhost-k8s-coredns--76f75df574--kldz2-eth0" Sep 4 17:20:15.212709 containerd[1448]: 2024-09-04 17:20:15.208 [INFO][5193] ipam_plugin.go 445: Releasing address using workloadID ContainerID="53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" HandleID="k8s-pod-network.53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" Workload="localhost-k8s-coredns--76f75df574--kldz2-eth0" Sep 4 17:20:15.212709 containerd[1448]: 2024-09-04 17:20:15.209 [INFO][5193] ipam_plugin.go 379: Released host-wide IPAM lock. Sep 4 17:20:15.212709 containerd[1448]: 2024-09-04 17:20:15.211 [INFO][5186] k8s.go 621: Teardown processing complete. ContainerID="53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2" Sep 4 17:20:15.213257 containerd[1448]: time="2024-09-04T17:20:15.212748304Z" level=info msg="TearDown network for sandbox \"53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2\" successfully" Sep 4 17:20:15.215324 containerd[1448]: time="2024-09-04T17:20:15.215289518Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Sep 4 17:20:15.215409 containerd[1448]: time="2024-09-04T17:20:15.215354758Z" level=info msg="RemovePodSandbox \"53da25b4a51d4221062cc47ab15b6af8305d915192a5e15b8af5bf8bc60a4ce2\" returns successfully" Sep 4 17:20:15.216204 containerd[1448]: time="2024-09-04T17:20:15.215900232Z" level=info msg="StopPodSandbox for \"ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954\"" Sep 4 17:20:15.289492 containerd[1448]: 2024-09-04 17:20:15.255 [WARNING][5216] k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--nsxnh-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"2e2776ed-fa4e-45b8-bb3f-17de64345308", ResourceVersion:"941", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 19, 35, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"78cd84fb8c", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"default"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"3b0cf9c39426749bc59147d0533e7ec0f9828bcf644887bfd3fb814581142291", Pod:"csi-node-driver-nsxnh", Endpoint:"eth0", ServiceAccountName:"default", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.default"}, InterfaceName:"cali39027bddf0e", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:20:15.289492 containerd[1448]: 2024-09-04 17:20:15.255 [INFO][5216] k8s.go 608: Cleaning up netns ContainerID="ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" Sep 4 17:20:15.289492 containerd[1448]: 2024-09-04 17:20:15.255 [INFO][5216] dataplane_linux.go 526: CleanUpNamespace called with no netns name, ignoring. ContainerID="ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" iface="eth0" netns="" Sep 4 17:20:15.289492 containerd[1448]: 2024-09-04 17:20:15.255 [INFO][5216] k8s.go 615: Releasing IP address(es) ContainerID="ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" Sep 4 17:20:15.289492 containerd[1448]: 2024-09-04 17:20:15.255 [INFO][5216] utils.go 188: Calico CNI releasing IP address ContainerID="ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" Sep 4 17:20:15.289492 containerd[1448]: 2024-09-04 17:20:15.275 [INFO][5225] ipam_plugin.go 417: Releasing address using handleID ContainerID="ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" HandleID="k8s-pod-network.ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" Workload="localhost-k8s-csi--node--driver--nsxnh-eth0" Sep 4 17:20:15.289492 containerd[1448]: 2024-09-04 17:20:15.275 [INFO][5225] ipam_plugin.go 358: About to acquire host-wide IPAM lock. Sep 4 17:20:15.289492 containerd[1448]: 2024-09-04 17:20:15.275 [INFO][5225] ipam_plugin.go 373: Acquired host-wide IPAM lock. Sep 4 17:20:15.289492 containerd[1448]: 2024-09-04 17:20:15.284 [WARNING][5225] ipam_plugin.go 434: Asked to release address but it doesn't exist. Ignoring ContainerID="ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" HandleID="k8s-pod-network.ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" Workload="localhost-k8s-csi--node--driver--nsxnh-eth0" Sep 4 17:20:15.289492 containerd[1448]: 2024-09-04 17:20:15.284 [INFO][5225] ipam_plugin.go 445: Releasing address using workloadID ContainerID="ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" HandleID="k8s-pod-network.ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" Workload="localhost-k8s-csi--node--driver--nsxnh-eth0" Sep 4 17:20:15.289492 containerd[1448]: 2024-09-04 17:20:15.286 [INFO][5225] ipam_plugin.go 379: Released host-wide IPAM lock. Sep 4 17:20:15.289492 containerd[1448]: 2024-09-04 17:20:15.287 [INFO][5216] k8s.go 621: Teardown processing complete. ContainerID="ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" Sep 4 17:20:15.289492 containerd[1448]: time="2024-09-04T17:20:15.289465632Z" level=info msg="TearDown network for sandbox \"ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954\" successfully" Sep 4 17:20:15.289492 containerd[1448]: time="2024-09-04T17:20:15.289491392Z" level=info msg="StopPodSandbox for \"ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954\" returns successfully" Sep 4 17:20:15.291360 containerd[1448]: time="2024-09-04T17:20:15.291312173Z" level=info msg="RemovePodSandbox for \"ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954\"" Sep 4 17:20:15.291415 containerd[1448]: time="2024-09-04T17:20:15.291354852Z" level=info msg="Forcibly stopping sandbox \"ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954\"" Sep 4 17:20:15.360345 containerd[1448]: 2024-09-04 17:20:15.327 [WARNING][5248] k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--nsxnh-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"2e2776ed-fa4e-45b8-bb3f-17de64345308", ResourceVersion:"941", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 19, 35, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"78cd84fb8c", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"default"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"3b0cf9c39426749bc59147d0533e7ec0f9828bcf644887bfd3fb814581142291", Pod:"csi-node-driver-nsxnh", Endpoint:"eth0", ServiceAccountName:"default", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.default"}, InterfaceName:"cali39027bddf0e", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:20:15.360345 containerd[1448]: 2024-09-04 17:20:15.328 [INFO][5248] k8s.go 608: Cleaning up netns ContainerID="ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" Sep 4 17:20:15.360345 containerd[1448]: 2024-09-04 17:20:15.328 [INFO][5248] dataplane_linux.go 526: CleanUpNamespace called with no netns name, ignoring. ContainerID="ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" iface="eth0" netns="" Sep 4 17:20:15.360345 containerd[1448]: 2024-09-04 17:20:15.328 [INFO][5248] k8s.go 615: Releasing IP address(es) ContainerID="ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" Sep 4 17:20:15.360345 containerd[1448]: 2024-09-04 17:20:15.328 [INFO][5248] utils.go 188: Calico CNI releasing IP address ContainerID="ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" Sep 4 17:20:15.360345 containerd[1448]: 2024-09-04 17:20:15.347 [INFO][5255] ipam_plugin.go 417: Releasing address using handleID ContainerID="ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" HandleID="k8s-pod-network.ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" Workload="localhost-k8s-csi--node--driver--nsxnh-eth0" Sep 4 17:20:15.360345 containerd[1448]: 2024-09-04 17:20:15.347 [INFO][5255] ipam_plugin.go 358: About to acquire host-wide IPAM lock. Sep 4 17:20:15.360345 containerd[1448]: 2024-09-04 17:20:15.347 [INFO][5255] ipam_plugin.go 373: Acquired host-wide IPAM lock. Sep 4 17:20:15.360345 containerd[1448]: 2024-09-04 17:20:15.355 [WARNING][5255] ipam_plugin.go 434: Asked to release address but it doesn't exist. Ignoring ContainerID="ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" HandleID="k8s-pod-network.ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" Workload="localhost-k8s-csi--node--driver--nsxnh-eth0" Sep 4 17:20:15.360345 containerd[1448]: 2024-09-04 17:20:15.355 [INFO][5255] ipam_plugin.go 445: Releasing address using workloadID ContainerID="ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" HandleID="k8s-pod-network.ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" Workload="localhost-k8s-csi--node--driver--nsxnh-eth0" Sep 4 17:20:15.360345 containerd[1448]: 2024-09-04 17:20:15.357 [INFO][5255] ipam_plugin.go 379: Released host-wide IPAM lock. Sep 4 17:20:15.360345 containerd[1448]: 2024-09-04 17:20:15.358 [INFO][5248] k8s.go 621: Teardown processing complete. ContainerID="ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954" Sep 4 17:20:15.360824 containerd[1448]: time="2024-09-04T17:20:15.360377699Z" level=info msg="TearDown network for sandbox \"ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954\" successfully" Sep 4 17:20:15.363078 containerd[1448]: time="2024-09-04T17:20:15.363040392Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Sep 4 17:20:15.363130 containerd[1448]: time="2024-09-04T17:20:15.363111991Z" level=info msg="RemovePodSandbox \"ac3f024268d301249cfb08eb06683d6f076f416f37dcc90bcf46532ccd57e954\" returns successfully" Sep 4 17:20:17.798295 kubelet[2523]: E0904 17:20:17.796817 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:20:18.605192 systemd[1]: Started sshd@18-10.0.0.60:22-10.0.0.1:59824.service - OpenSSH per-connection server daemon (10.0.0.1:59824). Sep 4 17:20:18.676458 sshd[5286]: Accepted publickey for core from 10.0.0.1 port 59824 ssh2: RSA SHA256:TcdII3DD+/vh6fGiZDuqtLwdsO9LHnvXRMQO7IdpdiA Sep 4 17:20:18.677925 sshd[5286]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 4 17:20:18.685281 systemd-logind[1423]: New session 19 of user core. Sep 4 17:20:18.691099 systemd[1]: Started session-19.scope - Session 19 of User core. Sep 4 17:20:18.905305 sshd[5286]: pam_unix(sshd:session): session closed for user core Sep 4 17:20:18.909126 systemd[1]: sshd@18-10.0.0.60:22-10.0.0.1:59824.service: Deactivated successfully. Sep 4 17:20:18.911591 systemd[1]: session-19.scope: Deactivated successfully. Sep 4 17:20:18.916148 systemd-logind[1423]: Session 19 logged out. Waiting for processes to exit. Sep 4 17:20:18.917858 systemd-logind[1423]: Removed session 19. Sep 4 17:20:19.340608 kubelet[2523]: I0904 17:20:19.340487 2523 topology_manager.go:215] "Topology Admit Handler" podUID="7d1242be-0ebc-42c3-b5ab-dab07a53f9b5" podNamespace="calico-apiserver" podName="calico-apiserver-864697bc77-knr8n" Sep 4 17:20:19.350479 systemd[1]: Created slice kubepods-besteffort-pod7d1242be_0ebc_42c3_b5ab_dab07a53f9b5.slice - libcontainer container kubepods-besteffort-pod7d1242be_0ebc_42c3_b5ab_dab07a53f9b5.slice. Sep 4 17:20:19.360093 kubelet[2523]: I0904 17:20:19.358138 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjr4c\" (UniqueName: \"kubernetes.io/projected/7d1242be-0ebc-42c3-b5ab-dab07a53f9b5-kube-api-access-rjr4c\") pod \"calico-apiserver-864697bc77-knr8n\" (UID: \"7d1242be-0ebc-42c3-b5ab-dab07a53f9b5\") " pod="calico-apiserver/calico-apiserver-864697bc77-knr8n" Sep 4 17:20:19.360093 kubelet[2523]: I0904 17:20:19.358187 2523 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/7d1242be-0ebc-42c3-b5ab-dab07a53f9b5-calico-apiserver-certs\") pod \"calico-apiserver-864697bc77-knr8n\" (UID: \"7d1242be-0ebc-42c3-b5ab-dab07a53f9b5\") " pod="calico-apiserver/calico-apiserver-864697bc77-knr8n" Sep 4 17:20:19.459478 kubelet[2523]: E0904 17:20:19.459281 2523 secret.go:194] Couldn't get secret calico-apiserver/calico-apiserver-certs: secret "calico-apiserver-certs" not found Sep 4 17:20:19.466617 kubelet[2523]: E0904 17:20:19.466586 2523 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7d1242be-0ebc-42c3-b5ab-dab07a53f9b5-calico-apiserver-certs podName:7d1242be-0ebc-42c3-b5ab-dab07a53f9b5 nodeName:}" failed. No retries permitted until 2024-09-04 17:20:19.959342614 +0000 UTC m=+65.380256337 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "calico-apiserver-certs" (UniqueName: "kubernetes.io/secret/7d1242be-0ebc-42c3-b5ab-dab07a53f9b5-calico-apiserver-certs") pod "calico-apiserver-864697bc77-knr8n" (UID: "7d1242be-0ebc-42c3-b5ab-dab07a53f9b5") : secret "calico-apiserver-certs" not found Sep 4 17:20:19.962276 kubelet[2523]: E0904 17:20:19.962224 2523 secret.go:194] Couldn't get secret calico-apiserver/calico-apiserver-certs: secret "calico-apiserver-certs" not found Sep 4 17:20:19.962422 kubelet[2523]: E0904 17:20:19.962302 2523 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7d1242be-0ebc-42c3-b5ab-dab07a53f9b5-calico-apiserver-certs podName:7d1242be-0ebc-42c3-b5ab-dab07a53f9b5 nodeName:}" failed. No retries permitted until 2024-09-04 17:20:20.962286872 +0000 UTC m=+66.383200635 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "calico-apiserver-certs" (UniqueName: "kubernetes.io/secret/7d1242be-0ebc-42c3-b5ab-dab07a53f9b5-calico-apiserver-certs") pod "calico-apiserver-864697bc77-knr8n" (UID: "7d1242be-0ebc-42c3-b5ab-dab07a53f9b5") : secret "calico-apiserver-certs" not found Sep 4 17:20:21.155050 containerd[1448]: time="2024-09-04T17:20:21.154990471Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-864697bc77-knr8n,Uid:7d1242be-0ebc-42c3-b5ab-dab07a53f9b5,Namespace:calico-apiserver,Attempt:0,}" Sep 4 17:20:21.298947 systemd-networkd[1377]: cali24f025c9d45: Link UP Sep 4 17:20:21.299184 systemd-networkd[1377]: cali24f025c9d45: Gained carrier Sep 4 17:20:21.317573 containerd[1448]: 2024-09-04 17:20:21.207 [INFO][5306] plugin.go 326: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--864697bc77--knr8n-eth0 calico-apiserver-864697bc77- calico-apiserver 7d1242be-0ebc-42c3-b5ab-dab07a53f9b5 1092 0 2024-09-04 17:20:19 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:864697bc77 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-864697bc77-knr8n eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali24f025c9d45 [] []}} ContainerID="6173e45bf7c0306e0a96e575d501c216370625706f6571aa7e398bd5f475f339" Namespace="calico-apiserver" Pod="calico-apiserver-864697bc77-knr8n" WorkloadEndpoint="localhost-k8s-calico--apiserver--864697bc77--knr8n-" Sep 4 17:20:21.317573 containerd[1448]: 2024-09-04 17:20:21.207 [INFO][5306] k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="6173e45bf7c0306e0a96e575d501c216370625706f6571aa7e398bd5f475f339" Namespace="calico-apiserver" Pod="calico-apiserver-864697bc77-knr8n" WorkloadEndpoint="localhost-k8s-calico--apiserver--864697bc77--knr8n-eth0" Sep 4 17:20:21.317573 containerd[1448]: 2024-09-04 17:20:21.231 [INFO][5318] ipam_plugin.go 230: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="6173e45bf7c0306e0a96e575d501c216370625706f6571aa7e398bd5f475f339" HandleID="k8s-pod-network.6173e45bf7c0306e0a96e575d501c216370625706f6571aa7e398bd5f475f339" Workload="localhost-k8s-calico--apiserver--864697bc77--knr8n-eth0" Sep 4 17:20:21.317573 containerd[1448]: 2024-09-04 17:20:21.243 [INFO][5318] ipam_plugin.go 270: Auto assigning IP ContainerID="6173e45bf7c0306e0a96e575d501c216370625706f6571aa7e398bd5f475f339" HandleID="k8s-pod-network.6173e45bf7c0306e0a96e575d501c216370625706f6571aa7e398bd5f475f339" Workload="localhost-k8s-calico--apiserver--864697bc77--knr8n-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002e4790), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-864697bc77-knr8n", "timestamp":"2024-09-04 17:20:21.231001921 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Sep 4 17:20:21.317573 containerd[1448]: 2024-09-04 17:20:21.244 [INFO][5318] ipam_plugin.go 358: About to acquire host-wide IPAM lock. Sep 4 17:20:21.317573 containerd[1448]: 2024-09-04 17:20:21.244 [INFO][5318] ipam_plugin.go 373: Acquired host-wide IPAM lock. Sep 4 17:20:21.317573 containerd[1448]: 2024-09-04 17:20:21.244 [INFO][5318] ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Sep 4 17:20:21.317573 containerd[1448]: 2024-09-04 17:20:21.252 [INFO][5318] ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.6173e45bf7c0306e0a96e575d501c216370625706f6571aa7e398bd5f475f339" host="localhost" Sep 4 17:20:21.317573 containerd[1448]: 2024-09-04 17:20:21.256 [INFO][5318] ipam.go 372: Looking up existing affinities for host host="localhost" Sep 4 17:20:21.317573 containerd[1448]: 2024-09-04 17:20:21.260 [INFO][5318] ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Sep 4 17:20:21.317573 containerd[1448]: 2024-09-04 17:20:21.266 [INFO][5318] ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Sep 4 17:20:21.317573 containerd[1448]: 2024-09-04 17:20:21.269 [INFO][5318] ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Sep 4 17:20:21.317573 containerd[1448]: 2024-09-04 17:20:21.269 [INFO][5318] ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.6173e45bf7c0306e0a96e575d501c216370625706f6571aa7e398bd5f475f339" host="localhost" Sep 4 17:20:21.317573 containerd[1448]: 2024-09-04 17:20:21.272 [INFO][5318] ipam.go 1685: Creating new handle: k8s-pod-network.6173e45bf7c0306e0a96e575d501c216370625706f6571aa7e398bd5f475f339 Sep 4 17:20:21.317573 containerd[1448]: 2024-09-04 17:20:21.278 [INFO][5318] ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.6173e45bf7c0306e0a96e575d501c216370625706f6571aa7e398bd5f475f339" host="localhost" Sep 4 17:20:21.317573 containerd[1448]: 2024-09-04 17:20:21.284 [INFO][5318] ipam.go 1216: Successfully claimed IPs: [192.168.88.133/26] block=192.168.88.128/26 handle="k8s-pod-network.6173e45bf7c0306e0a96e575d501c216370625706f6571aa7e398bd5f475f339" host="localhost" Sep 4 17:20:21.317573 containerd[1448]: 2024-09-04 17:20:21.285 [INFO][5318] ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.133/26] handle="k8s-pod-network.6173e45bf7c0306e0a96e575d501c216370625706f6571aa7e398bd5f475f339" host="localhost" Sep 4 17:20:21.317573 containerd[1448]: 2024-09-04 17:20:21.285 [INFO][5318] ipam_plugin.go 379: Released host-wide IPAM lock. Sep 4 17:20:21.317573 containerd[1448]: 2024-09-04 17:20:21.285 [INFO][5318] ipam_plugin.go 288: Calico CNI IPAM assigned addresses IPv4=[192.168.88.133/26] IPv6=[] ContainerID="6173e45bf7c0306e0a96e575d501c216370625706f6571aa7e398bd5f475f339" HandleID="k8s-pod-network.6173e45bf7c0306e0a96e575d501c216370625706f6571aa7e398bd5f475f339" Workload="localhost-k8s-calico--apiserver--864697bc77--knr8n-eth0" Sep 4 17:20:21.318122 containerd[1448]: 2024-09-04 17:20:21.289 [INFO][5306] k8s.go 386: Populated endpoint ContainerID="6173e45bf7c0306e0a96e575d501c216370625706f6571aa7e398bd5f475f339" Namespace="calico-apiserver" Pod="calico-apiserver-864697bc77-knr8n" WorkloadEndpoint="localhost-k8s-calico--apiserver--864697bc77--knr8n-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--864697bc77--knr8n-eth0", GenerateName:"calico-apiserver-864697bc77-", Namespace:"calico-apiserver", SelfLink:"", UID:"7d1242be-0ebc-42c3-b5ab-dab07a53f9b5", ResourceVersion:"1092", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 20, 19, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"864697bc77", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-864697bc77-knr8n", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali24f025c9d45", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:20:21.318122 containerd[1448]: 2024-09-04 17:20:21.289 [INFO][5306] k8s.go 387: Calico CNI using IPs: [192.168.88.133/32] ContainerID="6173e45bf7c0306e0a96e575d501c216370625706f6571aa7e398bd5f475f339" Namespace="calico-apiserver" Pod="calico-apiserver-864697bc77-knr8n" WorkloadEndpoint="localhost-k8s-calico--apiserver--864697bc77--knr8n-eth0" Sep 4 17:20:21.318122 containerd[1448]: 2024-09-04 17:20:21.289 [INFO][5306] dataplane_linux.go 68: Setting the host side veth name to cali24f025c9d45 ContainerID="6173e45bf7c0306e0a96e575d501c216370625706f6571aa7e398bd5f475f339" Namespace="calico-apiserver" Pod="calico-apiserver-864697bc77-knr8n" WorkloadEndpoint="localhost-k8s-calico--apiserver--864697bc77--knr8n-eth0" Sep 4 17:20:21.318122 containerd[1448]: 2024-09-04 17:20:21.296 [INFO][5306] dataplane_linux.go 479: Disabling IPv4 forwarding ContainerID="6173e45bf7c0306e0a96e575d501c216370625706f6571aa7e398bd5f475f339" Namespace="calico-apiserver" Pod="calico-apiserver-864697bc77-knr8n" WorkloadEndpoint="localhost-k8s-calico--apiserver--864697bc77--knr8n-eth0" Sep 4 17:20:21.318122 containerd[1448]: 2024-09-04 17:20:21.303 [INFO][5306] k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="6173e45bf7c0306e0a96e575d501c216370625706f6571aa7e398bd5f475f339" Namespace="calico-apiserver" Pod="calico-apiserver-864697bc77-knr8n" WorkloadEndpoint="localhost-k8s-calico--apiserver--864697bc77--knr8n-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--864697bc77--knr8n-eth0", GenerateName:"calico-apiserver-864697bc77-", Namespace:"calico-apiserver", SelfLink:"", UID:"7d1242be-0ebc-42c3-b5ab-dab07a53f9b5", ResourceVersion:"1092", Generation:0, CreationTimestamp:time.Date(2024, time.September, 4, 17, 20, 19, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"864697bc77", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"6173e45bf7c0306e0a96e575d501c216370625706f6571aa7e398bd5f475f339", Pod:"calico-apiserver-864697bc77-knr8n", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali24f025c9d45", MAC:"62:c4:ae:24:cc:c7", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Sep 4 17:20:21.318122 containerd[1448]: 2024-09-04 17:20:21.315 [INFO][5306] k8s.go 500: Wrote updated endpoint to datastore ContainerID="6173e45bf7c0306e0a96e575d501c216370625706f6571aa7e398bd5f475f339" Namespace="calico-apiserver" Pod="calico-apiserver-864697bc77-knr8n" WorkloadEndpoint="localhost-k8s-calico--apiserver--864697bc77--knr8n-eth0" Sep 4 17:20:21.341012 containerd[1448]: time="2024-09-04T17:20:21.340496867Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 4 17:20:21.341012 containerd[1448]: time="2024-09-04T17:20:21.340981144Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:20:21.341326 containerd[1448]: time="2024-09-04T17:20:21.341186383Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 4 17:20:21.341326 containerd[1448]: time="2024-09-04T17:20:21.341208022Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 4 17:20:21.364006 systemd[1]: Started cri-containerd-6173e45bf7c0306e0a96e575d501c216370625706f6571aa7e398bd5f475f339.scope - libcontainer container 6173e45bf7c0306e0a96e575d501c216370625706f6571aa7e398bd5f475f339. Sep 4 17:20:21.377543 systemd-resolved[1309]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Sep 4 17:20:21.402162 containerd[1448]: time="2024-09-04T17:20:21.402121014Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-864697bc77-knr8n,Uid:7d1242be-0ebc-42c3-b5ab-dab07a53f9b5,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"6173e45bf7c0306e0a96e575d501c216370625706f6571aa7e398bd5f475f339\"" Sep 4 17:20:21.404008 containerd[1448]: time="2024-09-04T17:20:21.403742683Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.28.1\"" Sep 4 17:20:22.581997 systemd-networkd[1377]: cali24f025c9d45: Gained IPv6LL Sep 4 17:20:23.213400 containerd[1448]: time="2024-09-04T17:20:23.213347592Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.28.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:20:23.214463 containerd[1448]: time="2024-09-04T17:20:23.214425866Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.28.1: active requests=0, bytes read=37849884" Sep 4 17:20:23.215276 containerd[1448]: time="2024-09-04T17:20:23.215247061Z" level=info msg="ImageCreate event name:\"sha256:913d8e601c95ebd056c4c949f148ec565327fa2c94a6c34bb4fcfbd9063a58ec\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:20:23.218161 containerd[1448]: time="2024-09-04T17:20:23.218096645Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:b4ee1aa27bdeddc34dd200145eb033b716cf598570206c96693a35a317ab4f1e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Sep 4 17:20:23.218815 containerd[1448]: time="2024-09-04T17:20:23.218780521Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.28.1\" with image id \"sha256:913d8e601c95ebd056c4c949f148ec565327fa2c94a6c34bb4fcfbd9063a58ec\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.28.1\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:b4ee1aa27bdeddc34dd200145eb033b716cf598570206c96693a35a317ab4f1e\", size \"39217419\" in 1.815003678s" Sep 4 17:20:23.218815 containerd[1448]: time="2024-09-04T17:20:23.218814721Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.28.1\" returns image reference \"sha256:913d8e601c95ebd056c4c949f148ec565327fa2c94a6c34bb4fcfbd9063a58ec\"" Sep 4 17:20:23.228032 containerd[1448]: time="2024-09-04T17:20:23.227976309Z" level=info msg="CreateContainer within sandbox \"6173e45bf7c0306e0a96e575d501c216370625706f6571aa7e398bd5f475f339\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Sep 4 17:20:23.239303 containerd[1448]: time="2024-09-04T17:20:23.239244206Z" level=info msg="CreateContainer within sandbox \"6173e45bf7c0306e0a96e575d501c216370625706f6571aa7e398bd5f475f339\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"9e7947db6bdf140683fd0298e974afeee798aac804c4cd0b602d526b53eabb0b\"" Sep 4 17:20:23.239947 containerd[1448]: time="2024-09-04T17:20:23.239749843Z" level=info msg="StartContainer for \"9e7947db6bdf140683fd0298e974afeee798aac804c4cd0b602d526b53eabb0b\"" Sep 4 17:20:23.284021 systemd[1]: Started cri-containerd-9e7947db6bdf140683fd0298e974afeee798aac804c4cd0b602d526b53eabb0b.scope - libcontainer container 9e7947db6bdf140683fd0298e974afeee798aac804c4cd0b602d526b53eabb0b. Sep 4 17:20:23.326299 containerd[1448]: time="2024-09-04T17:20:23.326180395Z" level=info msg="StartContainer for \"9e7947db6bdf140683fd0298e974afeee798aac804c4cd0b602d526b53eabb0b\" returns successfully" Sep 4 17:20:23.920049 systemd[1]: Started sshd@19-10.0.0.60:22-10.0.0.1:39892.service - OpenSSH per-connection server daemon (10.0.0.1:39892). Sep 4 17:20:23.972765 sshd[5446]: Accepted publickey for core from 10.0.0.1 port 39892 ssh2: RSA SHA256:TcdII3DD+/vh6fGiZDuqtLwdsO9LHnvXRMQO7IdpdiA Sep 4 17:20:23.973990 sshd[5446]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 4 17:20:23.979928 systemd-logind[1423]: New session 20 of user core. Sep 4 17:20:23.984992 systemd[1]: Started session-20.scope - Session 20 of User core. Sep 4 17:20:24.053017 kubelet[2523]: I0904 17:20:24.052976 2523 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-864697bc77-knr8n" podStartSLOduration=3.237326444 podStartE2EDuration="5.052930679s" podCreationTimestamp="2024-09-04 17:20:19 +0000 UTC" firstStartedPulling="2024-09-04 17:20:21.403428565 +0000 UTC m=+66.824342288" lastFinishedPulling="2024-09-04 17:20:23.21903276 +0000 UTC m=+68.639946523" observedRunningTime="2024-09-04 17:20:24.04114238 +0000 UTC m=+69.462056103" watchObservedRunningTime="2024-09-04 17:20:24.052930679 +0000 UTC m=+69.473844402" Sep 4 17:20:24.153243 sshd[5446]: pam_unix(sshd:session): session closed for user core Sep 4 17:20:24.157057 systemd[1]: sshd@19-10.0.0.60:22-10.0.0.1:39892.service: Deactivated successfully. Sep 4 17:20:24.159196 systemd[1]: session-20.scope: Deactivated successfully. Sep 4 17:20:24.159988 systemd-logind[1423]: Session 20 logged out. Waiting for processes to exit. Sep 4 17:20:24.160993 systemd-logind[1423]: Removed session 20. Sep 4 17:20:27.713801 kubelet[2523]: E0904 17:20:27.713764 2523 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Sep 4 17:20:29.167553 systemd[1]: Started sshd@20-10.0.0.60:22-10.0.0.1:39896.service - OpenSSH per-connection server daemon (10.0.0.1:39896). Sep 4 17:20:29.235019 sshd[5489]: Accepted publickey for core from 10.0.0.1 port 39896 ssh2: RSA SHA256:TcdII3DD+/vh6fGiZDuqtLwdsO9LHnvXRMQO7IdpdiA Sep 4 17:20:29.236319 sshd[5489]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 4 17:20:29.240491 systemd-logind[1423]: New session 21 of user core. Sep 4 17:20:29.248024 systemd[1]: Started session-21.scope - Session 21 of User core. Sep 4 17:20:29.414817 sshd[5489]: pam_unix(sshd:session): session closed for user core Sep 4 17:20:29.419669 systemd[1]: sshd@20-10.0.0.60:22-10.0.0.1:39896.service: Deactivated successfully. Sep 4 17:20:29.421696 systemd[1]: session-21.scope: Deactivated successfully. Sep 4 17:20:29.422354 systemd-logind[1423]: Session 21 logged out. Waiting for processes to exit. Sep 4 17:20:29.423102 systemd-logind[1423]: Removed session 21.