Mar 20 21:14:23.892821 kernel: Booting Linux on physical CPU 0x0000000000 [0x413fd0c1] Mar 20 21:14:23.892841 kernel: Linux version 6.6.83-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 14.2.1_p20241221 p7) 14.2.1 20241221, GNU ld (Gentoo 2.44 p1) 2.44.0) #1 SMP PREEMPT Thu Mar 20 19:37:53 -00 2025 Mar 20 21:14:23.892851 kernel: KASLR enabled Mar 20 21:14:23.892856 kernel: efi: EFI v2.7 by EDK II Mar 20 21:14:23.892862 kernel: efi: SMBIOS 3.0=0xdced0000 MEMATTR=0xdbbae018 ACPI 2.0=0xd9b43018 RNG=0xd9b43a18 MEMRESERVE=0xd9b40498 Mar 20 21:14:23.892867 kernel: random: crng init done Mar 20 21:14:23.892873 kernel: secureboot: Secure boot disabled Mar 20 21:14:23.892879 kernel: ACPI: Early table checksum verification disabled Mar 20 21:14:23.892885 kernel: ACPI: RSDP 0x00000000D9B43018 000024 (v02 BOCHS ) Mar 20 21:14:23.892892 kernel: ACPI: XSDT 0x00000000D9B43F18 000064 (v01 BOCHS BXPC 00000001 01000013) Mar 20 21:14:23.892898 kernel: ACPI: FACP 0x00000000D9B43B18 000114 (v06 BOCHS BXPC 00000001 BXPC 00000001) Mar 20 21:14:23.892903 kernel: ACPI: DSDT 0x00000000D9B41018 0014A2 (v02 BOCHS BXPC 00000001 BXPC 00000001) Mar 20 21:14:23.892909 kernel: ACPI: APIC 0x00000000D9B43C98 0001A8 (v04 BOCHS BXPC 00000001 BXPC 00000001) Mar 20 21:14:23.892914 kernel: ACPI: PPTT 0x00000000D9B43098 00009C (v02 BOCHS BXPC 00000001 BXPC 00000001) Mar 20 21:14:23.892921 kernel: ACPI: GTDT 0x00000000D9B43818 000060 (v02 BOCHS BXPC 00000001 BXPC 00000001) Mar 20 21:14:23.892928 kernel: ACPI: MCFG 0x00000000D9B43A98 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 20 21:14:23.892934 kernel: ACPI: SPCR 0x00000000D9B43918 000050 (v02 BOCHS BXPC 00000001 BXPC 00000001) Mar 20 21:14:23.892940 kernel: ACPI: DBG2 0x00000000D9B43998 000057 (v00 BOCHS BXPC 00000001 BXPC 00000001) Mar 20 21:14:23.892946 kernel: ACPI: IORT 0x00000000D9B43198 000080 (v03 BOCHS BXPC 00000001 BXPC 00000001) Mar 20 21:14:23.892952 kernel: ACPI: SPCR: console: pl011,mmio,0x9000000,9600 Mar 20 21:14:23.892958 kernel: NUMA: Failed to initialise from firmware Mar 20 21:14:23.892964 kernel: NUMA: Faking a node at [mem 0x0000000040000000-0x00000000dcffffff] Mar 20 21:14:23.892970 kernel: NUMA: NODE_DATA [mem 0xdc958800-0xdc95dfff] Mar 20 21:14:23.892975 kernel: Zone ranges: Mar 20 21:14:23.892981 kernel: DMA [mem 0x0000000040000000-0x00000000dcffffff] Mar 20 21:14:23.892989 kernel: DMA32 empty Mar 20 21:14:23.892995 kernel: Normal empty Mar 20 21:14:23.893001 kernel: Movable zone start for each node Mar 20 21:14:23.893007 kernel: Early memory node ranges Mar 20 21:14:23.893013 kernel: node 0: [mem 0x0000000040000000-0x00000000d967ffff] Mar 20 21:14:23.893019 kernel: node 0: [mem 0x00000000d9680000-0x00000000d968ffff] Mar 20 21:14:23.893025 kernel: node 0: [mem 0x00000000d9690000-0x00000000d976ffff] Mar 20 21:14:23.893043 kernel: node 0: [mem 0x00000000d9770000-0x00000000d9b3ffff] Mar 20 21:14:23.893056 kernel: node 0: [mem 0x00000000d9b40000-0x00000000dce1ffff] Mar 20 21:14:23.893063 kernel: node 0: [mem 0x00000000dce20000-0x00000000dceaffff] Mar 20 21:14:23.893069 kernel: node 0: [mem 0x00000000dceb0000-0x00000000dcebffff] Mar 20 21:14:23.893075 kernel: node 0: [mem 0x00000000dcec0000-0x00000000dcfdffff] Mar 20 21:14:23.893083 kernel: node 0: [mem 0x00000000dcfe0000-0x00000000dcffffff] Mar 20 21:14:23.893089 kernel: Initmem setup node 0 [mem 0x0000000040000000-0x00000000dcffffff] Mar 20 21:14:23.893095 kernel: On node 0, zone DMA: 12288 pages in unavailable ranges Mar 20 21:14:23.893103 kernel: psci: probing for conduit method from ACPI. Mar 20 21:14:23.893110 kernel: psci: PSCIv1.1 detected in firmware. Mar 20 21:14:23.893116 kernel: psci: Using standard PSCI v0.2 function IDs Mar 20 21:14:23.893124 kernel: psci: Trusted OS migration not required Mar 20 21:14:23.893130 kernel: psci: SMC Calling Convention v1.1 Mar 20 21:14:23.893137 kernel: smccc: KVM: hypervisor services detected (0x00000000 0x00000000 0x00000000 0x00000003) Mar 20 21:14:23.893143 kernel: percpu: Embedded 31 pages/cpu s86696 r8192 d32088 u126976 Mar 20 21:14:23.893149 kernel: pcpu-alloc: s86696 r8192 d32088 u126976 alloc=31*4096 Mar 20 21:14:23.893156 kernel: pcpu-alloc: [0] 0 [0] 1 [0] 2 [0] 3 Mar 20 21:14:23.893162 kernel: Detected PIPT I-cache on CPU0 Mar 20 21:14:23.893169 kernel: CPU features: detected: GIC system register CPU interface Mar 20 21:14:23.893176 kernel: CPU features: detected: Hardware dirty bit management Mar 20 21:14:23.893182 kernel: CPU features: detected: Spectre-v4 Mar 20 21:14:23.893189 kernel: CPU features: detected: Spectre-BHB Mar 20 21:14:23.893196 kernel: CPU features: kernel page table isolation forced ON by KASLR Mar 20 21:14:23.893202 kernel: CPU features: detected: Kernel page table isolation (KPTI) Mar 20 21:14:23.893208 kernel: CPU features: detected: ARM erratum 1418040 Mar 20 21:14:23.893215 kernel: CPU features: detected: SSBS not fully self-synchronizing Mar 20 21:14:23.893221 kernel: alternatives: applying boot alternatives Mar 20 21:14:23.893228 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected acpi=force verity.usrhash=0beb08f475de014f6ab4e06127ed84e918521fd470084f537ae9409b262d0ed3 Mar 20 21:14:23.893235 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Mar 20 21:14:23.893241 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Mar 20 21:14:23.893248 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Mar 20 21:14:23.893254 kernel: Fallback order for Node 0: 0 Mar 20 21:14:23.893262 kernel: Built 1 zonelists, mobility grouping on. Total pages: 633024 Mar 20 21:14:23.893268 kernel: Policy zone: DMA Mar 20 21:14:23.893274 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Mar 20 21:14:23.893280 kernel: software IO TLB: area num 4. Mar 20 21:14:23.893287 kernel: software IO TLB: mapped [mem 0x00000000d2e00000-0x00000000d6e00000] (64MB) Mar 20 21:14:23.893293 kernel: Memory: 2387412K/2572288K available (10304K kernel code, 2186K rwdata, 8096K rodata, 38464K init, 897K bss, 184876K reserved, 0K cma-reserved) Mar 20 21:14:23.893300 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Mar 20 21:14:23.893306 kernel: rcu: Preemptible hierarchical RCU implementation. Mar 20 21:14:23.893313 kernel: rcu: RCU event tracing is enabled. Mar 20 21:14:23.893320 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Mar 20 21:14:23.893326 kernel: Trampoline variant of Tasks RCU enabled. Mar 20 21:14:23.893332 kernel: Tracing variant of Tasks RCU enabled. Mar 20 21:14:23.893340 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Mar 20 21:14:23.893346 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Mar 20 21:14:23.893353 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Mar 20 21:14:23.893359 kernel: GICv3: 256 SPIs implemented Mar 20 21:14:23.893365 kernel: GICv3: 0 Extended SPIs implemented Mar 20 21:14:23.893372 kernel: Root IRQ handler: gic_handle_irq Mar 20 21:14:23.893378 kernel: GICv3: GICv3 features: 16 PPIs, DirectLPI Mar 20 21:14:23.893384 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000080a0000 Mar 20 21:14:23.893390 kernel: ITS [mem 0x08080000-0x0809ffff] Mar 20 21:14:23.893397 kernel: ITS@0x0000000008080000: allocated 8192 Devices @400c0000 (indirect, esz 8, psz 64K, shr 1) Mar 20 21:14:23.893403 kernel: ITS@0x0000000008080000: allocated 8192 Interrupt Collections @400d0000 (flat, esz 8, psz 64K, shr 1) Mar 20 21:14:23.893411 kernel: GICv3: using LPI property table @0x00000000400f0000 Mar 20 21:14:23.893417 kernel: GICv3: CPU0: using allocated LPI pending table @0x0000000040100000 Mar 20 21:14:23.893424 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Mar 20 21:14:23.893430 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Mar 20 21:14:23.893437 kernel: arch_timer: cp15 timer(s) running at 25.00MHz (virt). Mar 20 21:14:23.893443 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns Mar 20 21:14:23.893450 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns Mar 20 21:14:23.893456 kernel: arm-pv: using stolen time PV Mar 20 21:14:23.893462 kernel: Console: colour dummy device 80x25 Mar 20 21:14:23.893469 kernel: ACPI: Core revision 20230628 Mar 20 21:14:23.893476 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000) Mar 20 21:14:23.893483 kernel: pid_max: default: 32768 minimum: 301 Mar 20 21:14:23.893490 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Mar 20 21:14:23.893496 kernel: landlock: Up and running. Mar 20 21:14:23.893511 kernel: SELinux: Initializing. Mar 20 21:14:23.893518 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Mar 20 21:14:23.893524 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Mar 20 21:14:23.893531 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Mar 20 21:14:23.893538 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. Mar 20 21:14:23.893544 kernel: rcu: Hierarchical SRCU implementation. Mar 20 21:14:23.893553 kernel: rcu: Max phase no-delay instances is 400. Mar 20 21:14:23.893559 kernel: Platform MSI: ITS@0x8080000 domain created Mar 20 21:14:23.893566 kernel: PCI/MSI: ITS@0x8080000 domain created Mar 20 21:14:23.893572 kernel: Remapping and enabling EFI services. Mar 20 21:14:23.893578 kernel: smp: Bringing up secondary CPUs ... Mar 20 21:14:23.893585 kernel: Detected PIPT I-cache on CPU1 Mar 20 21:14:23.893592 kernel: GICv3: CPU1: found redistributor 1 region 0:0x00000000080c0000 Mar 20 21:14:23.893598 kernel: GICv3: CPU1: using allocated LPI pending table @0x0000000040110000 Mar 20 21:14:23.893605 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Mar 20 21:14:23.893612 kernel: CPU1: Booted secondary processor 0x0000000001 [0x413fd0c1] Mar 20 21:14:23.893619 kernel: Detected PIPT I-cache on CPU2 Mar 20 21:14:23.893630 kernel: GICv3: CPU2: found redistributor 2 region 0:0x00000000080e0000 Mar 20 21:14:23.893638 kernel: GICv3: CPU2: using allocated LPI pending table @0x0000000040120000 Mar 20 21:14:23.893645 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Mar 20 21:14:23.893652 kernel: CPU2: Booted secondary processor 0x0000000002 [0x413fd0c1] Mar 20 21:14:23.893658 kernel: Detected PIPT I-cache on CPU3 Mar 20 21:14:23.893665 kernel: GICv3: CPU3: found redistributor 3 region 0:0x0000000008100000 Mar 20 21:14:23.893672 kernel: GICv3: CPU3: using allocated LPI pending table @0x0000000040130000 Mar 20 21:14:23.893680 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Mar 20 21:14:23.893687 kernel: CPU3: Booted secondary processor 0x0000000003 [0x413fd0c1] Mar 20 21:14:23.893694 kernel: smp: Brought up 1 node, 4 CPUs Mar 20 21:14:23.893701 kernel: SMP: Total of 4 processors activated. Mar 20 21:14:23.893708 kernel: CPU features: detected: 32-bit EL0 Support Mar 20 21:14:23.893714 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence Mar 20 21:14:23.893722 kernel: CPU features: detected: Common not Private translations Mar 20 21:14:23.893734 kernel: CPU features: detected: CRC32 instructions Mar 20 21:14:23.893742 kernel: CPU features: detected: Enhanced Virtualization Traps Mar 20 21:14:23.893749 kernel: CPU features: detected: RCpc load-acquire (LDAPR) Mar 20 21:14:23.893756 kernel: CPU features: detected: LSE atomic instructions Mar 20 21:14:23.893763 kernel: CPU features: detected: Privileged Access Never Mar 20 21:14:23.893770 kernel: CPU features: detected: RAS Extension Support Mar 20 21:14:23.893777 kernel: CPU features: detected: Speculative Store Bypassing Safe (SSBS) Mar 20 21:14:23.893783 kernel: CPU: All CPU(s) started at EL1 Mar 20 21:14:23.893791 kernel: alternatives: applying system-wide alternatives Mar 20 21:14:23.893798 kernel: devtmpfs: initialized Mar 20 21:14:23.893805 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Mar 20 21:14:23.893813 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Mar 20 21:14:23.893820 kernel: pinctrl core: initialized pinctrl subsystem Mar 20 21:14:23.893827 kernel: SMBIOS 3.0.0 present. Mar 20 21:14:23.893833 kernel: DMI: QEMU KVM Virtual Machine, BIOS unknown 02/02/2022 Mar 20 21:14:23.893840 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Mar 20 21:14:23.893847 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Mar 20 21:14:23.893854 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Mar 20 21:14:23.893861 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Mar 20 21:14:23.893869 kernel: audit: initializing netlink subsys (disabled) Mar 20 21:14:23.893876 kernel: audit: type=2000 audit(0.018:1): state=initialized audit_enabled=0 res=1 Mar 20 21:14:23.893883 kernel: thermal_sys: Registered thermal governor 'step_wise' Mar 20 21:14:23.893890 kernel: cpuidle: using governor menu Mar 20 21:14:23.893897 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Mar 20 21:14:23.893904 kernel: ASID allocator initialised with 32768 entries Mar 20 21:14:23.893910 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Mar 20 21:14:23.893917 kernel: Serial: AMBA PL011 UART driver Mar 20 21:14:23.893924 kernel: Modules: 2G module region forced by RANDOMIZE_MODULE_REGION_FULL Mar 20 21:14:23.893932 kernel: Modules: 0 pages in range for non-PLT usage Mar 20 21:14:23.893939 kernel: Modules: 509248 pages in range for PLT usage Mar 20 21:14:23.893946 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Mar 20 21:14:23.893953 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page Mar 20 21:14:23.893960 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages Mar 20 21:14:23.893968 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page Mar 20 21:14:23.893975 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Mar 20 21:14:23.893981 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page Mar 20 21:14:23.893988 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages Mar 20 21:14:23.894008 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page Mar 20 21:14:23.894016 kernel: ACPI: Added _OSI(Module Device) Mar 20 21:14:23.894024 kernel: ACPI: Added _OSI(Processor Device) Mar 20 21:14:23.894031 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Mar 20 21:14:23.894038 kernel: ACPI: Added _OSI(Processor Aggregator Device) Mar 20 21:14:23.894044 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Mar 20 21:14:23.894057 kernel: ACPI: Interpreter enabled Mar 20 21:14:23.894064 kernel: ACPI: Using GIC for interrupt routing Mar 20 21:14:23.894071 kernel: ACPI: MCFG table detected, 1 entries Mar 20 21:14:23.894078 kernel: ARMH0011:00: ttyAMA0 at MMIO 0x9000000 (irq = 12, base_baud = 0) is a SBSA Mar 20 21:14:23.894086 kernel: printk: console [ttyAMA0] enabled Mar 20 21:14:23.894093 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Mar 20 21:14:23.894219 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Mar 20 21:14:23.894290 kernel: acpi PNP0A08:00: _OSC: platform does not support [LTR] Mar 20 21:14:23.894358 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME AER PCIeCapability] Mar 20 21:14:23.894425 kernel: acpi PNP0A08:00: ECAM area [mem 0x4010000000-0x401fffffff] reserved by PNP0C02:00 Mar 20 21:14:23.894489 kernel: acpi PNP0A08:00: ECAM at [mem 0x4010000000-0x401fffffff] for [bus 00-ff] Mar 20 21:14:23.894501 kernel: ACPI: Remapped I/O 0x000000003eff0000 to [io 0x0000-0xffff window] Mar 20 21:14:23.894518 kernel: PCI host bridge to bus 0000:00 Mar 20 21:14:23.894598 kernel: pci_bus 0000:00: root bus resource [mem 0x10000000-0x3efeffff window] Mar 20 21:14:23.894658 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0xffff window] Mar 20 21:14:23.894716 kernel: pci_bus 0000:00: root bus resource [mem 0x8000000000-0xffffffffff window] Mar 20 21:14:23.894778 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Mar 20 21:14:23.894861 kernel: pci 0000:00:00.0: [1b36:0008] type 00 class 0x060000 Mar 20 21:14:23.894941 kernel: pci 0000:00:01.0: [1af4:1005] type 00 class 0x00ff00 Mar 20 21:14:23.895007 kernel: pci 0000:00:01.0: reg 0x10: [io 0x0000-0x001f] Mar 20 21:14:23.895084 kernel: pci 0000:00:01.0: reg 0x14: [mem 0x10000000-0x10000fff] Mar 20 21:14:23.895151 kernel: pci 0000:00:01.0: reg 0x20: [mem 0x8000000000-0x8000003fff 64bit pref] Mar 20 21:14:23.895219 kernel: pci 0000:00:01.0: BAR 4: assigned [mem 0x8000000000-0x8000003fff 64bit pref] Mar 20 21:14:23.895299 kernel: pci 0000:00:01.0: BAR 1: assigned [mem 0x10000000-0x10000fff] Mar 20 21:14:23.895364 kernel: pci 0000:00:01.0: BAR 0: assigned [io 0x1000-0x101f] Mar 20 21:14:23.895427 kernel: pci_bus 0000:00: resource 4 [mem 0x10000000-0x3efeffff window] Mar 20 21:14:23.895485 kernel: pci_bus 0000:00: resource 5 [io 0x0000-0xffff window] Mar 20 21:14:23.895553 kernel: pci_bus 0000:00: resource 6 [mem 0x8000000000-0xffffffffff window] Mar 20 21:14:23.895564 kernel: ACPI: PCI: Interrupt link GSI0 configured for IRQ 35 Mar 20 21:14:23.895571 kernel: ACPI: PCI: Interrupt link GSI1 configured for IRQ 36 Mar 20 21:14:23.895578 kernel: ACPI: PCI: Interrupt link GSI2 configured for IRQ 37 Mar 20 21:14:23.895585 kernel: ACPI: PCI: Interrupt link GSI3 configured for IRQ 38 Mar 20 21:14:23.895594 kernel: iommu: Default domain type: Translated Mar 20 21:14:23.895601 kernel: iommu: DMA domain TLB invalidation policy: strict mode Mar 20 21:14:23.895608 kernel: efivars: Registered efivars operations Mar 20 21:14:23.895614 kernel: vgaarb: loaded Mar 20 21:14:23.895621 kernel: clocksource: Switched to clocksource arch_sys_counter Mar 20 21:14:23.895628 kernel: VFS: Disk quotas dquot_6.6.0 Mar 20 21:14:23.895635 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Mar 20 21:14:23.895642 kernel: pnp: PnP ACPI init Mar 20 21:14:23.895712 kernel: system 00:00: [mem 0x4010000000-0x401fffffff window] could not be reserved Mar 20 21:14:23.895723 kernel: pnp: PnP ACPI: found 1 devices Mar 20 21:14:23.895730 kernel: NET: Registered PF_INET protocol family Mar 20 21:14:23.895737 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Mar 20 21:14:23.895744 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Mar 20 21:14:23.895751 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Mar 20 21:14:23.895758 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Mar 20 21:14:23.895765 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Mar 20 21:14:23.895772 kernel: TCP: Hash tables configured (established 32768 bind 32768) Mar 20 21:14:23.895779 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Mar 20 21:14:23.895787 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Mar 20 21:14:23.895794 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Mar 20 21:14:23.895801 kernel: PCI: CLS 0 bytes, default 64 Mar 20 21:14:23.895808 kernel: kvm [1]: HYP mode not available Mar 20 21:14:23.895815 kernel: Initialise system trusted keyrings Mar 20 21:14:23.895821 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Mar 20 21:14:23.895828 kernel: Key type asymmetric registered Mar 20 21:14:23.895835 kernel: Asymmetric key parser 'x509' registered Mar 20 21:14:23.895842 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Mar 20 21:14:23.895850 kernel: io scheduler mq-deadline registered Mar 20 21:14:23.895857 kernel: io scheduler kyber registered Mar 20 21:14:23.895863 kernel: io scheduler bfq registered Mar 20 21:14:23.895870 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0C:00/input/input0 Mar 20 21:14:23.895877 kernel: ACPI: button: Power Button [PWRB] Mar 20 21:14:23.895884 kernel: ACPI: \_SB_.PCI0.GSI1: Enabled at IRQ 36 Mar 20 21:14:23.895948 kernel: virtio-pci 0000:00:01.0: enabling device (0005 -> 0007) Mar 20 21:14:23.895957 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Mar 20 21:14:23.895964 kernel: thunder_xcv, ver 1.0 Mar 20 21:14:23.895972 kernel: thunder_bgx, ver 1.0 Mar 20 21:14:23.895979 kernel: nicpf, ver 1.0 Mar 20 21:14:23.895986 kernel: nicvf, ver 1.0 Mar 20 21:14:23.896069 kernel: rtc-efi rtc-efi.0: registered as rtc0 Mar 20 21:14:23.896135 kernel: rtc-efi rtc-efi.0: setting system clock to 2025-03-20T21:14:23 UTC (1742505263) Mar 20 21:14:23.896145 kernel: hid: raw HID events driver (C) Jiri Kosina Mar 20 21:14:23.896152 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 7 counters available Mar 20 21:14:23.896159 kernel: watchdog: Delayed init of the lockup detector failed: -19 Mar 20 21:14:23.896168 kernel: watchdog: Hard watchdog permanently disabled Mar 20 21:14:23.896175 kernel: NET: Registered PF_INET6 protocol family Mar 20 21:14:23.896182 kernel: Segment Routing with IPv6 Mar 20 21:14:23.896189 kernel: In-situ OAM (IOAM) with IPv6 Mar 20 21:14:23.896196 kernel: NET: Registered PF_PACKET protocol family Mar 20 21:14:23.896203 kernel: Key type dns_resolver registered Mar 20 21:14:23.896209 kernel: registered taskstats version 1 Mar 20 21:14:23.896216 kernel: Loading compiled-in X.509 certificates Mar 20 21:14:23.896223 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.83-flatcar: 3a6f52a6c751e8bbe3389ae978b265effe8f77af' Mar 20 21:14:23.896232 kernel: Key type .fscrypt registered Mar 20 21:14:23.896238 kernel: Key type fscrypt-provisioning registered Mar 20 21:14:23.896245 kernel: ima: No TPM chip found, activating TPM-bypass! Mar 20 21:14:23.896252 kernel: ima: Allocated hash algorithm: sha1 Mar 20 21:14:23.896259 kernel: ima: No architecture policies found Mar 20 21:14:23.896266 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Mar 20 21:14:23.896272 kernel: clk: Disabling unused clocks Mar 20 21:14:23.896279 kernel: Freeing unused kernel memory: 38464K Mar 20 21:14:23.896286 kernel: Run /init as init process Mar 20 21:14:23.896294 kernel: with arguments: Mar 20 21:14:23.896301 kernel: /init Mar 20 21:14:23.896307 kernel: with environment: Mar 20 21:14:23.896314 kernel: HOME=/ Mar 20 21:14:23.896321 kernel: TERM=linux Mar 20 21:14:23.896328 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Mar 20 21:14:23.896335 systemd[1]: Successfully made /usr/ read-only. Mar 20 21:14:23.896345 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Mar 20 21:14:23.896354 systemd[1]: Detected virtualization kvm. Mar 20 21:14:23.896361 systemd[1]: Detected architecture arm64. Mar 20 21:14:23.896369 systemd[1]: Running in initrd. Mar 20 21:14:23.896376 systemd[1]: No hostname configured, using default hostname. Mar 20 21:14:23.896384 systemd[1]: Hostname set to . Mar 20 21:14:23.896391 systemd[1]: Initializing machine ID from VM UUID. Mar 20 21:14:23.896398 systemd[1]: Queued start job for default target initrd.target. Mar 20 21:14:23.896406 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 20 21:14:23.896414 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 20 21:14:23.896422 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Mar 20 21:14:23.896430 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Mar 20 21:14:23.896438 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Mar 20 21:14:23.896446 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Mar 20 21:14:23.896454 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Mar 20 21:14:23.896463 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Mar 20 21:14:23.896471 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 20 21:14:23.896478 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Mar 20 21:14:23.896486 systemd[1]: Reached target paths.target - Path Units. Mar 20 21:14:23.896493 systemd[1]: Reached target slices.target - Slice Units. Mar 20 21:14:23.896507 systemd[1]: Reached target swap.target - Swaps. Mar 20 21:14:23.896516 systemd[1]: Reached target timers.target - Timer Units. Mar 20 21:14:23.896524 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Mar 20 21:14:23.896531 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Mar 20 21:14:23.896541 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Mar 20 21:14:23.896548 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Mar 20 21:14:23.896556 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Mar 20 21:14:23.896563 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Mar 20 21:14:23.896571 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Mar 20 21:14:23.896578 systemd[1]: Reached target sockets.target - Socket Units. Mar 20 21:14:23.896586 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Mar 20 21:14:23.896593 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Mar 20 21:14:23.896602 systemd[1]: Finished network-cleanup.service - Network Cleanup. Mar 20 21:14:23.896609 systemd[1]: Starting systemd-fsck-usr.service... Mar 20 21:14:23.896617 systemd[1]: Starting systemd-journald.service - Journal Service... Mar 20 21:14:23.896624 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Mar 20 21:14:23.896631 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 20 21:14:23.896639 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Mar 20 21:14:23.896646 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Mar 20 21:14:23.896655 systemd[1]: Finished systemd-fsck-usr.service. Mar 20 21:14:23.896663 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Mar 20 21:14:23.896671 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 20 21:14:23.896695 systemd-journald[236]: Collecting audit messages is disabled. Mar 20 21:14:23.896714 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Mar 20 21:14:23.896721 kernel: Bridge firewalling registered Mar 20 21:14:23.896729 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 20 21:14:23.896736 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Mar 20 21:14:23.896744 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Mar 20 21:14:23.896752 systemd-journald[236]: Journal started Mar 20 21:14:23.896771 systemd-journald[236]: Runtime Journal (/run/log/journal/f098d547e5774ed1984143b69d7eb72d) is 5.9M, max 47.3M, 41.4M free. Mar 20 21:14:23.872706 systemd-modules-load[237]: Inserted module 'overlay' Mar 20 21:14:23.891130 systemd-modules-load[237]: Inserted module 'br_netfilter' Mar 20 21:14:23.900566 systemd[1]: Started systemd-journald.service - Journal Service. Mar 20 21:14:23.903965 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 20 21:14:23.905569 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Mar 20 21:14:23.907773 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Mar 20 21:14:23.914743 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 20 21:14:23.918596 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Mar 20 21:14:23.921415 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 20 21:14:23.924720 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 20 21:14:23.926091 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 20 21:14:23.930303 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Mar 20 21:14:23.932869 dracut-cmdline[270]: dracut-dracut-053 Mar 20 21:14:23.934340 dracut-cmdline[270]: Using kernel command line parameters: rd.driver.pre=btrfs BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected acpi=force verity.usrhash=0beb08f475de014f6ab4e06127ed84e918521fd470084f537ae9409b262d0ed3 Mar 20 21:14:23.969730 systemd-resolved[283]: Positive Trust Anchors: Mar 20 21:14:23.969746 systemd-resolved[283]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 20 21:14:23.969782 systemd-resolved[283]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Mar 20 21:14:23.974663 systemd-resolved[283]: Defaulting to hostname 'linux'. Mar 20 21:14:23.975816 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Mar 20 21:14:23.978817 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Mar 20 21:14:24.003526 kernel: SCSI subsystem initialized Mar 20 21:14:24.006516 kernel: Loading iSCSI transport class v2.0-870. Mar 20 21:14:24.014529 kernel: iscsi: registered transport (tcp) Mar 20 21:14:24.027530 kernel: iscsi: registered transport (qla4xxx) Mar 20 21:14:24.027562 kernel: QLogic iSCSI HBA Driver Mar 20 21:14:24.066837 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Mar 20 21:14:24.068830 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Mar 20 21:14:24.092849 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Mar 20 21:14:24.092931 kernel: device-mapper: uevent: version 1.0.3 Mar 20 21:14:24.092958 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Mar 20 21:14:24.139540 kernel: raid6: neonx8 gen() 13992 MB/s Mar 20 21:14:24.156532 kernel: raid6: neonx4 gen() 15780 MB/s Mar 20 21:14:24.173529 kernel: raid6: neonx2 gen() 13182 MB/s Mar 20 21:14:24.190519 kernel: raid6: neonx1 gen() 10473 MB/s Mar 20 21:14:24.207527 kernel: raid6: int64x8 gen() 6786 MB/s Mar 20 21:14:24.224529 kernel: raid6: int64x4 gen() 7325 MB/s Mar 20 21:14:24.241532 kernel: raid6: int64x2 gen() 6087 MB/s Mar 20 21:14:24.258526 kernel: raid6: int64x1 gen() 5034 MB/s Mar 20 21:14:24.258555 kernel: raid6: using algorithm neonx4 gen() 15780 MB/s Mar 20 21:14:24.275532 kernel: raid6: .... xor() 12395 MB/s, rmw enabled Mar 20 21:14:24.275557 kernel: raid6: using neon recovery algorithm Mar 20 21:14:24.280516 kernel: xor: measuring software checksum speed Mar 20 21:14:24.280531 kernel: 8regs : 21624 MB/sec Mar 20 21:14:24.280540 kernel: 32regs : 20383 MB/sec Mar 20 21:14:24.281825 kernel: arm64_neon : 27936 MB/sec Mar 20 21:14:24.281858 kernel: xor: using function: arm64_neon (27936 MB/sec) Mar 20 21:14:24.333543 kernel: Btrfs loaded, zoned=no, fsverity=no Mar 20 21:14:24.343861 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Mar 20 21:14:24.346300 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 20 21:14:24.374395 systemd-udevd[459]: Using default interface naming scheme 'v255'. Mar 20 21:14:24.378077 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 20 21:14:24.386331 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Mar 20 21:14:24.413322 dracut-pre-trigger[466]: rd.md=0: removing MD RAID activation Mar 20 21:14:24.437088 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Mar 20 21:14:24.439540 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Mar 20 21:14:24.489669 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Mar 20 21:14:24.491316 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Mar 20 21:14:24.511436 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Mar 20 21:14:24.514136 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Mar 20 21:14:24.515291 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 20 21:14:24.517427 systemd[1]: Reached target remote-fs.target - Remote File Systems. Mar 20 21:14:24.520454 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Mar 20 21:14:24.538882 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Mar 20 21:14:24.542663 kernel: virtio_blk virtio1: 1/0/0 default/read/poll queues Mar 20 21:14:24.552375 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Mar 20 21:14:24.552474 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Mar 20 21:14:24.552485 kernel: GPT:9289727 != 19775487 Mar 20 21:14:24.552494 kernel: GPT:Alternate GPT header not at the end of the disk. Mar 20 21:14:24.552520 kernel: GPT:9289727 != 19775487 Mar 20 21:14:24.552531 kernel: GPT: Use GNU Parted to correct GPT errors. Mar 20 21:14:24.552539 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Mar 20 21:14:24.552894 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Mar 20 21:14:24.553000 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 20 21:14:24.555834 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 20 21:14:24.556747 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 20 21:14:24.556883 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 20 21:14:24.559068 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Mar 20 21:14:24.560663 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 20 21:14:24.572536 kernel: BTRFS: device fsid 892d57a1-84f1-442c-90df-b8383db1b8c3 devid 1 transid 38 /dev/vda3 scanned by (udev-worker) (526) Mar 20 21:14:24.574529 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/vda6 scanned by (udev-worker) (513) Mar 20 21:14:24.585546 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 20 21:14:24.593171 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Mar 20 21:14:24.604755 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Mar 20 21:14:24.611835 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Mar 20 21:14:24.617762 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Mar 20 21:14:24.618666 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Mar 20 21:14:24.620985 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Mar 20 21:14:24.623031 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 20 21:14:24.640327 disk-uuid[560]: Primary Header is updated. Mar 20 21:14:24.640327 disk-uuid[560]: Secondary Entries is updated. Mar 20 21:14:24.640327 disk-uuid[560]: Secondary Header is updated. Mar 20 21:14:24.647827 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Mar 20 21:14:24.648545 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 20 21:14:25.656526 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Mar 20 21:14:25.656612 disk-uuid[565]: The operation has completed successfully. Mar 20 21:14:25.681043 systemd[1]: disk-uuid.service: Deactivated successfully. Mar 20 21:14:25.681136 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Mar 20 21:14:25.707242 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Mar 20 21:14:25.723237 sh[582]: Success Mar 20 21:14:25.739305 kernel: device-mapper: verity: sha256 using implementation "sha256-ce" Mar 20 21:14:25.770001 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Mar 20 21:14:25.772370 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Mar 20 21:14:25.788472 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Mar 20 21:14:25.793033 kernel: BTRFS info (device dm-0): first mount of filesystem 892d57a1-84f1-442c-90df-b8383db1b8c3 Mar 20 21:14:25.793062 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm Mar 20 21:14:25.793072 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Mar 20 21:14:25.794843 kernel: BTRFS info (device dm-0): disabling log replay at mount time Mar 20 21:14:25.794860 kernel: BTRFS info (device dm-0): using free space tree Mar 20 21:14:25.798086 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Mar 20 21:14:25.799144 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Mar 20 21:14:25.799824 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Mar 20 21:14:25.802454 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Mar 20 21:14:25.824048 kernel: BTRFS info (device vda6): first mount of filesystem d2d05864-61d3-424d-8bc5-6b85db5f6d34 Mar 20 21:14:25.824096 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm Mar 20 21:14:25.824107 kernel: BTRFS info (device vda6): using free space tree Mar 20 21:14:25.826528 kernel: BTRFS info (device vda6): auto enabling async discard Mar 20 21:14:25.830740 kernel: BTRFS info (device vda6): last unmount of filesystem d2d05864-61d3-424d-8bc5-6b85db5f6d34 Mar 20 21:14:25.833065 systemd[1]: Finished ignition-setup.service - Ignition (setup). Mar 20 21:14:25.834670 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Mar 20 21:14:25.901546 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Mar 20 21:14:25.905925 systemd[1]: Starting systemd-networkd.service - Network Configuration... Mar 20 21:14:25.933209 ignition[670]: Ignition 2.20.0 Mar 20 21:14:25.933219 ignition[670]: Stage: fetch-offline Mar 20 21:14:25.933248 ignition[670]: no configs at "/usr/lib/ignition/base.d" Mar 20 21:14:25.933256 ignition[670]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 20 21:14:25.933406 ignition[670]: parsed url from cmdline: "" Mar 20 21:14:25.933409 ignition[670]: no config URL provided Mar 20 21:14:25.933413 ignition[670]: reading system config file "/usr/lib/ignition/user.ign" Mar 20 21:14:25.933420 ignition[670]: no config at "/usr/lib/ignition/user.ign" Mar 20 21:14:25.933442 ignition[670]: op(1): [started] loading QEMU firmware config module Mar 20 21:14:25.933451 ignition[670]: op(1): executing: "modprobe" "qemu_fw_cfg" Mar 20 21:14:25.940191 ignition[670]: op(1): [finished] loading QEMU firmware config module Mar 20 21:14:25.950112 systemd-networkd[772]: lo: Link UP Mar 20 21:14:25.950123 systemd-networkd[772]: lo: Gained carrier Mar 20 21:14:25.950932 systemd-networkd[772]: Enumeration completed Mar 20 21:14:25.951325 systemd-networkd[772]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 20 21:14:25.951329 systemd-networkd[772]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 20 21:14:25.951939 systemd-networkd[772]: eth0: Link UP Mar 20 21:14:25.951942 systemd-networkd[772]: eth0: Gained carrier Mar 20 21:14:25.951947 systemd-networkd[772]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 20 21:14:25.952548 systemd[1]: Started systemd-networkd.service - Network Configuration. Mar 20 21:14:25.953560 systemd[1]: Reached target network.target - Network. Mar 20 21:14:25.971542 systemd-networkd[772]: eth0: DHCPv4 address 10.0.0.54/16, gateway 10.0.0.1 acquired from 10.0.0.1 Mar 20 21:14:25.986615 ignition[670]: parsing config with SHA512: e66022d5fa955b23342c33f9d94628c130533b95f8883f010d2f887ad6eacbb78d05d4d889cad91fe2be0e31e1ae6c41ba9be3a710be0613eada5dcba598211f Mar 20 21:14:25.992871 unknown[670]: fetched base config from "system" Mar 20 21:14:25.992881 unknown[670]: fetched user config from "qemu" Mar 20 21:14:25.993265 ignition[670]: fetch-offline: fetch-offline passed Mar 20 21:14:25.993331 ignition[670]: Ignition finished successfully Mar 20 21:14:25.995293 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Mar 20 21:14:25.996575 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Mar 20 21:14:25.997255 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Mar 20 21:14:26.025280 ignition[780]: Ignition 2.20.0 Mar 20 21:14:26.025290 ignition[780]: Stage: kargs Mar 20 21:14:26.025441 ignition[780]: no configs at "/usr/lib/ignition/base.d" Mar 20 21:14:26.025450 ignition[780]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 20 21:14:26.026267 ignition[780]: kargs: kargs passed Mar 20 21:14:26.029050 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Mar 20 21:14:26.026310 ignition[780]: Ignition finished successfully Mar 20 21:14:26.030873 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Mar 20 21:14:26.058542 ignition[789]: Ignition 2.20.0 Mar 20 21:14:26.058552 ignition[789]: Stage: disks Mar 20 21:14:26.058686 ignition[789]: no configs at "/usr/lib/ignition/base.d" Mar 20 21:14:26.058695 ignition[789]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 20 21:14:26.059473 ignition[789]: disks: disks passed Mar 20 21:14:26.060841 systemd[1]: Finished ignition-disks.service - Ignition (disks). Mar 20 21:14:26.059534 ignition[789]: Ignition finished successfully Mar 20 21:14:26.061828 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Mar 20 21:14:26.062795 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Mar 20 21:14:26.064217 systemd[1]: Reached target local-fs.target - Local File Systems. Mar 20 21:14:26.065395 systemd[1]: Reached target sysinit.target - System Initialization. Mar 20 21:14:26.066783 systemd[1]: Reached target basic.target - Basic System. Mar 20 21:14:26.068888 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Mar 20 21:14:26.088673 systemd-fsck[800]: ROOT: clean, 14/553520 files, 52654/553472 blocks Mar 20 21:14:26.092072 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Mar 20 21:14:26.094388 systemd[1]: Mounting sysroot.mount - /sysroot... Mar 20 21:14:26.154397 systemd[1]: Mounted sysroot.mount - /sysroot. Mar 20 21:14:26.155730 kernel: EXT4-fs (vda9): mounted filesystem 78c526d9-91af-4481-a769-6d3064caa829 r/w with ordered data mode. Quota mode: none. Mar 20 21:14:26.155628 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Mar 20 21:14:26.157688 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Mar 20 21:14:26.159246 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Mar 20 21:14:26.160258 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Mar 20 21:14:26.160297 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Mar 20 21:14:26.160319 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Mar 20 21:14:26.174224 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Mar 20 21:14:26.177060 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Mar 20 21:14:26.179283 kernel: BTRFS: device label OEM devid 1 transid 16 /dev/vda6 scanned by mount (808) Mar 20 21:14:26.181555 kernel: BTRFS info (device vda6): first mount of filesystem d2d05864-61d3-424d-8bc5-6b85db5f6d34 Mar 20 21:14:26.181590 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm Mar 20 21:14:26.181606 kernel: BTRFS info (device vda6): using free space tree Mar 20 21:14:26.183552 kernel: BTRFS info (device vda6): auto enabling async discard Mar 20 21:14:26.184770 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Mar 20 21:14:26.221471 initrd-setup-root[832]: cut: /sysroot/etc/passwd: No such file or directory Mar 20 21:14:26.225417 initrd-setup-root[839]: cut: /sysroot/etc/group: No such file or directory Mar 20 21:14:26.228385 initrd-setup-root[846]: cut: /sysroot/etc/shadow: No such file or directory Mar 20 21:14:26.231333 initrd-setup-root[853]: cut: /sysroot/etc/gshadow: No such file or directory Mar 20 21:14:26.301444 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Mar 20 21:14:26.303177 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Mar 20 21:14:26.304528 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Mar 20 21:14:26.327552 kernel: BTRFS info (device vda6): last unmount of filesystem d2d05864-61d3-424d-8bc5-6b85db5f6d34 Mar 20 21:14:26.344733 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Mar 20 21:14:26.355604 ignition[922]: INFO : Ignition 2.20.0 Mar 20 21:14:26.355604 ignition[922]: INFO : Stage: mount Mar 20 21:14:26.355604 ignition[922]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 20 21:14:26.355604 ignition[922]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 20 21:14:26.355604 ignition[922]: INFO : mount: mount passed Mar 20 21:14:26.355604 ignition[922]: INFO : Ignition finished successfully Mar 20 21:14:26.356846 systemd[1]: Finished ignition-mount.service - Ignition (mount). Mar 20 21:14:26.359046 systemd[1]: Starting ignition-files.service - Ignition (files)... Mar 20 21:14:26.927904 systemd[1]: sysroot-oem.mount: Deactivated successfully. Mar 20 21:14:26.929367 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Mar 20 21:14:26.947727 kernel: BTRFS: device label OEM devid 1 transid 17 /dev/vda6 scanned by mount (935) Mar 20 21:14:26.947768 kernel: BTRFS info (device vda6): first mount of filesystem d2d05864-61d3-424d-8bc5-6b85db5f6d34 Mar 20 21:14:26.947778 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm Mar 20 21:14:26.948857 kernel: BTRFS info (device vda6): using free space tree Mar 20 21:14:26.950518 kernel: BTRFS info (device vda6): auto enabling async discard Mar 20 21:14:26.951809 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Mar 20 21:14:26.972963 ignition[953]: INFO : Ignition 2.20.0 Mar 20 21:14:26.972963 ignition[953]: INFO : Stage: files Mar 20 21:14:26.974525 ignition[953]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 20 21:14:26.974525 ignition[953]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 20 21:14:26.974525 ignition[953]: DEBUG : files: compiled without relabeling support, skipping Mar 20 21:14:26.977781 ignition[953]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Mar 20 21:14:26.977781 ignition[953]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Mar 20 21:14:26.977781 ignition[953]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Mar 20 21:14:26.977781 ignition[953]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Mar 20 21:14:26.977781 ignition[953]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Mar 20 21:14:26.976976 unknown[953]: wrote ssh authorized keys file for user: core Mar 20 21:14:26.985056 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Mar 20 21:14:26.985056 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-arm64.tar.gz: attempt #1 Mar 20 21:14:27.026281 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Mar 20 21:14:27.078779 systemd-networkd[772]: eth0: Gained IPv6LL Mar 20 21:14:27.151955 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Mar 20 21:14:27.154233 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Mar 20 21:14:27.154233 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Mar 20 21:14:27.154233 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Mar 20 21:14:27.154233 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Mar 20 21:14:27.154233 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Mar 20 21:14:27.154233 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Mar 20 21:14:27.154233 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Mar 20 21:14:27.154233 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Mar 20 21:14:27.154233 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Mar 20 21:14:27.168978 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Mar 20 21:14:27.168978 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.0-arm64.raw" Mar 20 21:14:27.168978 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.0-arm64.raw" Mar 20 21:14:27.168978 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.0-arm64.raw" Mar 20 21:14:27.168978 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.31.0-arm64.raw: attempt #1 Mar 20 21:14:27.553985 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Mar 20 21:14:28.205280 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.0-arm64.raw" Mar 20 21:14:28.205280 ignition[953]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Mar 20 21:14:28.209055 ignition[953]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Mar 20 21:14:28.209055 ignition[953]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Mar 20 21:14:28.209055 ignition[953]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Mar 20 21:14:28.209055 ignition[953]: INFO : files: op(d): [started] processing unit "coreos-metadata.service" Mar 20 21:14:28.209055 ignition[953]: INFO : files: op(d): op(e): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Mar 20 21:14:28.209055 ignition[953]: INFO : files: op(d): op(e): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Mar 20 21:14:28.209055 ignition[953]: INFO : files: op(d): [finished] processing unit "coreos-metadata.service" Mar 20 21:14:28.209055 ignition[953]: INFO : files: op(f): [started] setting preset to disabled for "coreos-metadata.service" Mar 20 21:14:28.222636 ignition[953]: INFO : files: op(f): op(10): [started] removing enablement symlink(s) for "coreos-metadata.service" Mar 20 21:14:28.224784 ignition[953]: INFO : files: op(f): op(10): [finished] removing enablement symlink(s) for "coreos-metadata.service" Mar 20 21:14:28.226341 ignition[953]: INFO : files: op(f): [finished] setting preset to disabled for "coreos-metadata.service" Mar 20 21:14:28.226341 ignition[953]: INFO : files: op(11): [started] setting preset to enabled for "prepare-helm.service" Mar 20 21:14:28.226341 ignition[953]: INFO : files: op(11): [finished] setting preset to enabled for "prepare-helm.service" Mar 20 21:14:28.226341 ignition[953]: INFO : files: createResultFile: createFiles: op(12): [started] writing file "/sysroot/etc/.ignition-result.json" Mar 20 21:14:28.226341 ignition[953]: INFO : files: createResultFile: createFiles: op(12): [finished] writing file "/sysroot/etc/.ignition-result.json" Mar 20 21:14:28.226341 ignition[953]: INFO : files: files passed Mar 20 21:14:28.226341 ignition[953]: INFO : Ignition finished successfully Mar 20 21:14:28.227945 systemd[1]: Finished ignition-files.service - Ignition (files). Mar 20 21:14:28.231135 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Mar 20 21:14:28.232851 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Mar 20 21:14:28.245155 systemd[1]: ignition-quench.service: Deactivated successfully. Mar 20 21:14:28.245454 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Mar 20 21:14:28.249768 initrd-setup-root-after-ignition[980]: grep: /sysroot/oem/oem-release: No such file or directory Mar 20 21:14:28.251160 initrd-setup-root-after-ignition[983]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Mar 20 21:14:28.251160 initrd-setup-root-after-ignition[983]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Mar 20 21:14:28.254947 initrd-setup-root-after-ignition[987]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Mar 20 21:14:28.252224 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Mar 20 21:14:28.253706 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Mar 20 21:14:28.256520 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Mar 20 21:14:28.286057 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Mar 20 21:14:28.286179 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Mar 20 21:14:28.288114 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Mar 20 21:14:28.288907 systemd[1]: Reached target initrd.target - Initrd Default Target. Mar 20 21:14:28.289686 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Mar 20 21:14:28.290357 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Mar 20 21:14:28.314353 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Mar 20 21:14:28.316416 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Mar 20 21:14:28.334849 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Mar 20 21:14:28.335820 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 20 21:14:28.337650 systemd[1]: Stopped target timers.target - Timer Units. Mar 20 21:14:28.339239 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Mar 20 21:14:28.339345 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Mar 20 21:14:28.341692 systemd[1]: Stopped target initrd.target - Initrd Default Target. Mar 20 21:14:28.343516 systemd[1]: Stopped target basic.target - Basic System. Mar 20 21:14:28.345042 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Mar 20 21:14:28.346550 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Mar 20 21:14:28.348337 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Mar 20 21:14:28.350169 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Mar 20 21:14:28.351840 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Mar 20 21:14:28.353539 systemd[1]: Stopped target sysinit.target - System Initialization. Mar 20 21:14:28.355307 systemd[1]: Stopped target local-fs.target - Local File Systems. Mar 20 21:14:28.356904 systemd[1]: Stopped target swap.target - Swaps. Mar 20 21:14:28.358289 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Mar 20 21:14:28.358398 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Mar 20 21:14:28.360647 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Mar 20 21:14:28.362369 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 20 21:14:28.364096 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Mar 20 21:14:28.367578 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 20 21:14:28.368576 systemd[1]: dracut-initqueue.service: Deactivated successfully. Mar 20 21:14:28.368684 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Mar 20 21:14:28.371346 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Mar 20 21:14:28.371451 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Mar 20 21:14:28.373389 systemd[1]: Stopped target paths.target - Path Units. Mar 20 21:14:28.374818 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Mar 20 21:14:28.378572 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 20 21:14:28.379621 systemd[1]: Stopped target slices.target - Slice Units. Mar 20 21:14:28.381542 systemd[1]: Stopped target sockets.target - Socket Units. Mar 20 21:14:28.383024 systemd[1]: iscsid.socket: Deactivated successfully. Mar 20 21:14:28.383107 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Mar 20 21:14:28.384407 systemd[1]: iscsiuio.socket: Deactivated successfully. Mar 20 21:14:28.384481 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Mar 20 21:14:28.385865 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Mar 20 21:14:28.385974 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Mar 20 21:14:28.387512 systemd[1]: ignition-files.service: Deactivated successfully. Mar 20 21:14:28.387613 systemd[1]: Stopped ignition-files.service - Ignition (files). Mar 20 21:14:28.389621 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Mar 20 21:14:28.390979 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Mar 20 21:14:28.392011 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Mar 20 21:14:28.392134 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Mar 20 21:14:28.393938 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Mar 20 21:14:28.394034 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Mar 20 21:14:28.406761 systemd[1]: initrd-cleanup.service: Deactivated successfully. Mar 20 21:14:28.406849 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Mar 20 21:14:28.414700 systemd[1]: sysroot-boot.mount: Deactivated successfully. Mar 20 21:14:28.417875 ignition[1007]: INFO : Ignition 2.20.0 Mar 20 21:14:28.419848 ignition[1007]: INFO : Stage: umount Mar 20 21:14:28.419848 ignition[1007]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 20 21:14:28.419848 ignition[1007]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 20 21:14:28.419848 ignition[1007]: INFO : umount: umount passed Mar 20 21:14:28.419848 ignition[1007]: INFO : Ignition finished successfully Mar 20 21:14:28.418981 systemd[1]: sysroot-boot.service: Deactivated successfully. Mar 20 21:14:28.419069 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Mar 20 21:14:28.421897 systemd[1]: ignition-mount.service: Deactivated successfully. Mar 20 21:14:28.422017 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Mar 20 21:14:28.423178 systemd[1]: Stopped target network.target - Network. Mar 20 21:14:28.425647 systemd[1]: ignition-disks.service: Deactivated successfully. Mar 20 21:14:28.425714 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Mar 20 21:14:28.426923 systemd[1]: ignition-kargs.service: Deactivated successfully. Mar 20 21:14:28.426967 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Mar 20 21:14:28.428222 systemd[1]: ignition-setup.service: Deactivated successfully. Mar 20 21:14:28.428265 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Mar 20 21:14:28.429522 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Mar 20 21:14:28.429566 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Mar 20 21:14:28.430813 systemd[1]: initrd-setup-root.service: Deactivated successfully. Mar 20 21:14:28.430853 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Mar 20 21:14:28.432234 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Mar 20 21:14:28.433451 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Mar 20 21:14:28.440892 systemd[1]: systemd-resolved.service: Deactivated successfully. Mar 20 21:14:28.441002 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Mar 20 21:14:28.443585 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Mar 20 21:14:28.443839 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Mar 20 21:14:28.443899 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 20 21:14:28.446129 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Mar 20 21:14:28.448290 systemd[1]: systemd-networkd.service: Deactivated successfully. Mar 20 21:14:28.448388 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Mar 20 21:14:28.451201 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Mar 20 21:14:28.451352 systemd[1]: systemd-networkd.socket: Deactivated successfully. Mar 20 21:14:28.451383 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Mar 20 21:14:28.453211 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Mar 20 21:14:28.453962 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Mar 20 21:14:28.454017 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Mar 20 21:14:28.455464 systemd[1]: systemd-sysctl.service: Deactivated successfully. Mar 20 21:14:28.455518 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Mar 20 21:14:28.457676 systemd[1]: systemd-modules-load.service: Deactivated successfully. Mar 20 21:14:28.457719 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Mar 20 21:14:28.459133 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 20 21:14:28.461526 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Mar 20 21:14:28.476034 systemd[1]: systemd-udevd.service: Deactivated successfully. Mar 20 21:14:28.476172 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 20 21:14:28.477810 systemd[1]: network-cleanup.service: Deactivated successfully. Mar 20 21:14:28.477966 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Mar 20 21:14:28.479608 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Mar 20 21:14:28.479663 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Mar 20 21:14:28.481041 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Mar 20 21:14:28.481070 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Mar 20 21:14:28.482294 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Mar 20 21:14:28.482338 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Mar 20 21:14:28.484352 systemd[1]: dracut-cmdline.service: Deactivated successfully. Mar 20 21:14:28.484415 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Mar 20 21:14:28.486401 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Mar 20 21:14:28.486447 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 20 21:14:28.489141 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Mar 20 21:14:28.490493 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Mar 20 21:14:28.490554 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 20 21:14:28.492936 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Mar 20 21:14:28.492976 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Mar 20 21:14:28.494516 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Mar 20 21:14:28.494558 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Mar 20 21:14:28.496122 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 20 21:14:28.496158 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 20 21:14:28.506258 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Mar 20 21:14:28.506378 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Mar 20 21:14:28.508058 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Mar 20 21:14:28.510220 systemd[1]: Starting initrd-switch-root.service - Switch Root... Mar 20 21:14:28.529077 systemd[1]: Switching root. Mar 20 21:14:28.551467 systemd-journald[236]: Journal stopped Mar 20 21:14:29.240306 systemd-journald[236]: Received SIGTERM from PID 1 (systemd). Mar 20 21:14:29.240362 kernel: SELinux: policy capability network_peer_controls=1 Mar 20 21:14:29.240374 kernel: SELinux: policy capability open_perms=1 Mar 20 21:14:29.240384 kernel: SELinux: policy capability extended_socket_class=1 Mar 20 21:14:29.240393 kernel: SELinux: policy capability always_check_network=0 Mar 20 21:14:29.240403 kernel: SELinux: policy capability cgroup_seclabel=1 Mar 20 21:14:29.240415 kernel: SELinux: policy capability nnp_nosuid_transition=1 Mar 20 21:14:29.240428 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Mar 20 21:14:29.240441 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Mar 20 21:14:29.240451 kernel: audit: type=1403 audit(1742505268.678:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Mar 20 21:14:29.240466 systemd[1]: Successfully loaded SELinux policy in 33.007ms. Mar 20 21:14:29.240483 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 9.364ms. Mar 20 21:14:29.240494 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Mar 20 21:14:29.240524 systemd[1]: Detected virtualization kvm. Mar 20 21:14:29.240536 systemd[1]: Detected architecture arm64. Mar 20 21:14:29.240548 systemd[1]: Detected first boot. Mar 20 21:14:29.240559 systemd[1]: Initializing machine ID from VM UUID. Mar 20 21:14:29.240570 zram_generator::config[1052]: No configuration found. Mar 20 21:14:29.240581 kernel: NET: Registered PF_VSOCK protocol family Mar 20 21:14:29.240590 systemd[1]: Populated /etc with preset unit settings. Mar 20 21:14:29.240601 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Mar 20 21:14:29.240612 systemd[1]: initrd-switch-root.service: Deactivated successfully. Mar 20 21:14:29.240623 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Mar 20 21:14:29.240635 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Mar 20 21:14:29.240646 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Mar 20 21:14:29.240657 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Mar 20 21:14:29.240667 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Mar 20 21:14:29.240678 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Mar 20 21:14:29.240690 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Mar 20 21:14:29.240705 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Mar 20 21:14:29.240716 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Mar 20 21:14:29.240728 systemd[1]: Created slice user.slice - User and Session Slice. Mar 20 21:14:29.240738 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 20 21:14:29.240749 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 20 21:14:29.240760 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Mar 20 21:14:29.240770 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Mar 20 21:14:29.240781 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Mar 20 21:14:29.240792 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Mar 20 21:14:29.240802 systemd[1]: Expecting device dev-ttyAMA0.device - /dev/ttyAMA0... Mar 20 21:14:29.240813 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 20 21:14:29.240832 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Mar 20 21:14:29.240844 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Mar 20 21:14:29.240854 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Mar 20 21:14:29.240866 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Mar 20 21:14:29.240876 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 20 21:14:29.240886 systemd[1]: Reached target remote-fs.target - Remote File Systems. Mar 20 21:14:29.240896 systemd[1]: Reached target slices.target - Slice Units. Mar 20 21:14:29.240906 systemd[1]: Reached target swap.target - Swaps. Mar 20 21:14:29.240918 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Mar 20 21:14:29.240929 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Mar 20 21:14:29.240940 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Mar 20 21:14:29.240950 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Mar 20 21:14:29.240961 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Mar 20 21:14:29.240972 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Mar 20 21:14:29.240983 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Mar 20 21:14:29.240993 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Mar 20 21:14:29.241004 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Mar 20 21:14:29.241017 systemd[1]: Mounting media.mount - External Media Directory... Mar 20 21:14:29.241027 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Mar 20 21:14:29.241037 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Mar 20 21:14:29.241047 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Mar 20 21:14:29.241058 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Mar 20 21:14:29.241068 systemd[1]: Reached target machines.target - Containers. Mar 20 21:14:29.241078 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Mar 20 21:14:29.241088 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 20 21:14:29.241100 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Mar 20 21:14:29.241111 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Mar 20 21:14:29.241121 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 20 21:14:29.241132 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Mar 20 21:14:29.241143 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 20 21:14:29.241154 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Mar 20 21:14:29.241165 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 20 21:14:29.241175 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Mar 20 21:14:29.241185 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Mar 20 21:14:29.241198 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Mar 20 21:14:29.241208 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Mar 20 21:14:29.241219 systemd[1]: Stopped systemd-fsck-usr.service. Mar 20 21:14:29.241229 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Mar 20 21:14:29.241240 kernel: fuse: init (API version 7.39) Mar 20 21:14:29.241250 systemd[1]: Starting systemd-journald.service - Journal Service... Mar 20 21:14:29.241260 kernel: loop: module loaded Mar 20 21:14:29.241269 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Mar 20 21:14:29.241279 kernel: ACPI: bus type drm_connector registered Mar 20 21:14:29.241291 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Mar 20 21:14:29.241302 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Mar 20 21:14:29.241313 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Mar 20 21:14:29.241324 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Mar 20 21:14:29.241353 systemd-journald[1124]: Collecting audit messages is disabled. Mar 20 21:14:29.241374 systemd[1]: verity-setup.service: Deactivated successfully. Mar 20 21:14:29.241385 systemd[1]: Stopped verity-setup.service. Mar 20 21:14:29.241395 systemd-journald[1124]: Journal started Mar 20 21:14:29.241416 systemd-journald[1124]: Runtime Journal (/run/log/journal/f098d547e5774ed1984143b69d7eb72d) is 5.9M, max 47.3M, 41.4M free. Mar 20 21:14:29.062305 systemd[1]: Queued start job for default target multi-user.target. Mar 20 21:14:29.079353 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Mar 20 21:14:29.079740 systemd[1]: systemd-journald.service: Deactivated successfully. Mar 20 21:14:29.246528 systemd[1]: Started systemd-journald.service - Journal Service. Mar 20 21:14:29.246987 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Mar 20 21:14:29.247878 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Mar 20 21:14:29.248853 systemd[1]: Mounted media.mount - External Media Directory. Mar 20 21:14:29.249780 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Mar 20 21:14:29.250703 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Mar 20 21:14:29.251673 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Mar 20 21:14:29.254538 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Mar 20 21:14:29.255652 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Mar 20 21:14:29.256871 systemd[1]: modprobe@configfs.service: Deactivated successfully. Mar 20 21:14:29.257049 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Mar 20 21:14:29.258219 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 20 21:14:29.258376 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 20 21:14:29.259526 systemd[1]: modprobe@drm.service: Deactivated successfully. Mar 20 21:14:29.259704 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Mar 20 21:14:29.260911 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 20 21:14:29.261063 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 20 21:14:29.262213 systemd[1]: modprobe@fuse.service: Deactivated successfully. Mar 20 21:14:29.262369 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Mar 20 21:14:29.263574 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 20 21:14:29.263738 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 20 21:14:29.264820 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Mar 20 21:14:29.265972 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Mar 20 21:14:29.267293 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Mar 20 21:14:29.268529 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Mar 20 21:14:29.281434 systemd[1]: Reached target network-pre.target - Preparation for Network. Mar 20 21:14:29.283939 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Mar 20 21:14:29.286021 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Mar 20 21:14:29.287123 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Mar 20 21:14:29.287157 systemd[1]: Reached target local-fs.target - Local File Systems. Mar 20 21:14:29.289039 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Mar 20 21:14:29.293597 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Mar 20 21:14:29.295698 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Mar 20 21:14:29.296993 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 20 21:14:29.298220 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Mar 20 21:14:29.300206 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Mar 20 21:14:29.301424 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 20 21:14:29.304675 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Mar 20 21:14:29.305615 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 20 21:14:29.308187 systemd-journald[1124]: Time spent on flushing to /var/log/journal/f098d547e5774ed1984143b69d7eb72d is 21.485ms for 865 entries. Mar 20 21:14:29.308187 systemd-journald[1124]: System Journal (/var/log/journal/f098d547e5774ed1984143b69d7eb72d) is 8M, max 195.6M, 187.6M free. Mar 20 21:14:29.333233 systemd-journald[1124]: Received client request to flush runtime journal. Mar 20 21:14:29.333268 kernel: loop0: detected capacity change from 0 to 126448 Mar 20 21:14:29.308311 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 20 21:14:29.311532 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Mar 20 21:14:29.315219 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Mar 20 21:14:29.320269 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Mar 20 21:14:29.321690 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Mar 20 21:14:29.323199 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Mar 20 21:14:29.324769 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Mar 20 21:14:29.334797 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Mar 20 21:14:29.343657 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Mar 20 21:14:29.344944 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Mar 20 21:14:29.348719 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Mar 20 21:14:29.350395 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 20 21:14:29.356666 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Mar 20 21:14:29.359716 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Mar 20 21:14:29.362267 udevadm[1178]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. Mar 20 21:14:29.365703 systemd-tmpfiles[1170]: ACLs are not supported, ignoring. Mar 20 21:14:29.365718 systemd-tmpfiles[1170]: ACLs are not supported, ignoring. Mar 20 21:14:29.376946 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Mar 20 21:14:29.380531 kernel: loop1: detected capacity change from 0 to 189592 Mar 20 21:14:29.382690 systemd[1]: Starting systemd-sysusers.service - Create System Users... Mar 20 21:14:29.388262 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Mar 20 21:14:29.411835 systemd[1]: Finished systemd-sysusers.service - Create System Users. Mar 20 21:14:29.414302 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Mar 20 21:14:29.414614 kernel: loop2: detected capacity change from 0 to 103832 Mar 20 21:14:29.437370 systemd-tmpfiles[1194]: ACLs are not supported, ignoring. Mar 20 21:14:29.437391 systemd-tmpfiles[1194]: ACLs are not supported, ignoring. Mar 20 21:14:29.442982 kernel: loop3: detected capacity change from 0 to 126448 Mar 20 21:14:29.440713 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 20 21:14:29.447529 kernel: loop4: detected capacity change from 0 to 189592 Mar 20 21:14:29.453523 kernel: loop5: detected capacity change from 0 to 103832 Mar 20 21:14:29.457072 (sd-merge)[1197]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. Mar 20 21:14:29.457436 (sd-merge)[1197]: Merged extensions into '/usr'. Mar 20 21:14:29.461649 systemd[1]: Reload requested from client PID 1169 ('systemd-sysext') (unit systemd-sysext.service)... Mar 20 21:14:29.461755 systemd[1]: Reloading... Mar 20 21:14:29.531641 zram_generator::config[1223]: No configuration found. Mar 20 21:14:29.573189 ldconfig[1164]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Mar 20 21:14:29.620070 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 20 21:14:29.669703 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Mar 20 21:14:29.669932 systemd[1]: Reloading finished in 207 ms. Mar 20 21:14:29.690228 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Mar 20 21:14:29.691703 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Mar 20 21:14:29.703813 systemd[1]: Starting ensure-sysext.service... Mar 20 21:14:29.705824 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Mar 20 21:14:29.723087 systemd[1]: Reload requested from client PID 1261 ('systemctl') (unit ensure-sysext.service)... Mar 20 21:14:29.723210 systemd[1]: Reloading... Mar 20 21:14:29.728608 systemd-tmpfiles[1262]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Mar 20 21:14:29.728820 systemd-tmpfiles[1262]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Mar 20 21:14:29.729447 systemd-tmpfiles[1262]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Mar 20 21:14:29.729700 systemd-tmpfiles[1262]: ACLs are not supported, ignoring. Mar 20 21:14:29.729748 systemd-tmpfiles[1262]: ACLs are not supported, ignoring. Mar 20 21:14:29.732298 systemd-tmpfiles[1262]: Detected autofs mount point /boot during canonicalization of boot. Mar 20 21:14:29.732312 systemd-tmpfiles[1262]: Skipping /boot Mar 20 21:14:29.741219 systemd-tmpfiles[1262]: Detected autofs mount point /boot during canonicalization of boot. Mar 20 21:14:29.741234 systemd-tmpfiles[1262]: Skipping /boot Mar 20 21:14:29.782543 zram_generator::config[1291]: No configuration found. Mar 20 21:14:29.864420 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 20 21:14:29.914354 systemd[1]: Reloading finished in 190 ms. Mar 20 21:14:29.926954 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Mar 20 21:14:29.933520 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 20 21:14:29.948129 systemd[1]: Starting audit-rules.service - Load Audit Rules... Mar 20 21:14:29.950134 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Mar 20 21:14:29.963369 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Mar 20 21:14:29.966601 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Mar 20 21:14:29.971262 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 20 21:14:29.979215 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Mar 20 21:14:29.987422 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Mar 20 21:14:29.993670 systemd[1]: Starting systemd-update-done.service - Update is Completed... Mar 20 21:14:29.997741 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Mar 20 21:14:30.002260 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 20 21:14:30.003463 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 20 21:14:30.008615 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 20 21:14:30.010688 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 20 21:14:30.012621 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 20 21:14:30.012789 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Mar 20 21:14:30.020242 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Mar 20 21:14:30.023405 systemd[1]: Finished systemd-update-done.service - Update is Completed. Mar 20 21:14:30.023639 systemd-udevd[1332]: Using default interface naming scheme 'v255'. Mar 20 21:14:30.025192 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 20 21:14:30.025347 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 20 21:14:30.027066 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 20 21:14:30.027208 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 20 21:14:30.030279 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 20 21:14:30.030479 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 20 21:14:30.034542 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Mar 20 21:14:30.040924 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 20 21:14:30.042186 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 20 21:14:30.047690 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 20 21:14:30.050837 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 20 21:14:30.051869 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 20 21:14:30.051982 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Mar 20 21:14:30.052061 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Mar 20 21:14:30.052803 augenrules[1374]: No rules Mar 20 21:14:30.054219 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 20 21:14:30.056657 systemd[1]: audit-rules.service: Deactivated successfully. Mar 20 21:14:30.058650 systemd[1]: Finished audit-rules.service - Load Audit Rules. Mar 20 21:14:30.060025 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 20 21:14:30.060166 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 20 21:14:30.061589 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 20 21:14:30.061905 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 20 21:14:30.070951 systemd[1]: Started systemd-userdbd.service - User Database Manager. Mar 20 21:14:30.077611 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 20 21:14:30.077771 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 20 21:14:30.088113 systemd[1]: Finished ensure-sysext.service. Mar 20 21:14:30.097419 systemd[1]: Condition check resulted in dev-ttyAMA0.device - /dev/ttyAMA0 being skipped. Mar 20 21:14:30.100185 systemd[1]: Starting audit-rules.service - Load Audit Rules... Mar 20 21:14:30.100994 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 20 21:14:30.102826 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 20 21:14:30.119614 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 38 scanned by (udev-worker) (1377) Mar 20 21:14:30.119767 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Mar 20 21:14:30.125632 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 20 21:14:30.126544 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 20 21:14:30.126589 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Mar 20 21:14:30.128732 systemd[1]: Starting systemd-networkd.service - Network Configuration... Mar 20 21:14:30.132605 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Mar 20 21:14:30.134584 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Mar 20 21:14:30.135060 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 20 21:14:30.135248 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 20 21:14:30.137967 systemd[1]: modprobe@drm.service: Deactivated successfully. Mar 20 21:14:30.138129 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Mar 20 21:14:30.142344 augenrules[1403]: /sbin/augenrules: No change Mar 20 21:14:30.151900 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 20 21:14:30.152074 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 20 21:14:30.159055 systemd-resolved[1330]: Positive Trust Anchors: Mar 20 21:14:30.159070 systemd-resolved[1330]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 20 21:14:30.159101 systemd-resolved[1330]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Mar 20 21:14:30.159767 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Mar 20 21:14:30.164215 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Mar 20 21:14:30.165233 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 20 21:14:30.165881 augenrules[1433]: No rules Mar 20 21:14:30.165287 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 20 21:14:30.167138 systemd-resolved[1330]: Defaulting to hostname 'linux'. Mar 20 21:14:30.167915 systemd[1]: audit-rules.service: Deactivated successfully. Mar 20 21:14:30.169564 systemd[1]: Finished audit-rules.service - Load Audit Rules. Mar 20 21:14:30.170728 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Mar 20 21:14:30.176028 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Mar 20 21:14:30.198417 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Mar 20 21:14:30.214003 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Mar 20 21:14:30.215150 systemd[1]: Reached target time-set.target - System Time Set. Mar 20 21:14:30.215682 systemd-networkd[1416]: lo: Link UP Mar 20 21:14:30.215692 systemd-networkd[1416]: lo: Gained carrier Mar 20 21:14:30.216481 systemd-networkd[1416]: Enumeration completed Mar 20 21:14:30.216583 systemd[1]: Started systemd-networkd.service - Network Configuration. Mar 20 21:14:30.217190 systemd-networkd[1416]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 20 21:14:30.217193 systemd-networkd[1416]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 20 21:14:30.217660 systemd[1]: Reached target network.target - Network. Mar 20 21:14:30.219430 systemd-networkd[1416]: eth0: Link UP Mar 20 21:14:30.219436 systemd-networkd[1416]: eth0: Gained carrier Mar 20 21:14:30.219450 systemd-networkd[1416]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 20 21:14:30.221119 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Mar 20 21:14:30.224137 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Mar 20 21:14:30.230585 systemd-networkd[1416]: eth0: DHCPv4 address 10.0.0.54/16, gateway 10.0.0.1 acquired from 10.0.0.1 Mar 20 21:14:30.231593 systemd-timesyncd[1418]: Network configuration changed, trying to establish connection. Mar 20 21:14:30.697350 systemd-timesyncd[1418]: Contacted time server 10.0.0.1:123 (10.0.0.1). Mar 20 21:14:30.697393 systemd-timesyncd[1418]: Initial clock synchronization to Thu 2025-03-20 21:14:30.697262 UTC. Mar 20 21:14:30.700142 systemd-resolved[1330]: Clock change detected. Flushing caches. Mar 20 21:14:30.703655 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 20 21:14:30.707283 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Mar 20 21:14:30.714258 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Mar 20 21:14:30.716655 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Mar 20 21:14:30.738297 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 20 21:14:30.745994 lvm[1454]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Mar 20 21:14:30.778299 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Mar 20 21:14:30.779404 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Mar 20 21:14:30.780194 systemd[1]: Reached target sysinit.target - System Initialization. Mar 20 21:14:30.781019 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Mar 20 21:14:30.781946 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Mar 20 21:14:30.782995 systemd[1]: Started logrotate.timer - Daily rotation of log files. Mar 20 21:14:30.783880 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Mar 20 21:14:30.784822 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Mar 20 21:14:30.785706 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Mar 20 21:14:30.785738 systemd[1]: Reached target paths.target - Path Units. Mar 20 21:14:30.786377 systemd[1]: Reached target timers.target - Timer Units. Mar 20 21:14:30.787827 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Mar 20 21:14:30.789759 systemd[1]: Starting docker.socket - Docker Socket for the API... Mar 20 21:14:30.792635 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Mar 20 21:14:30.793723 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Mar 20 21:14:30.794677 systemd[1]: Reached target ssh-access.target - SSH Access Available. Mar 20 21:14:30.798822 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Mar 20 21:14:30.799981 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Mar 20 21:14:30.801910 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Mar 20 21:14:30.803282 systemd[1]: Listening on docker.socket - Docker Socket for the API. Mar 20 21:14:30.804122 systemd[1]: Reached target sockets.target - Socket Units. Mar 20 21:14:30.804865 systemd[1]: Reached target basic.target - Basic System. Mar 20 21:14:30.805554 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Mar 20 21:14:30.805581 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Mar 20 21:14:30.806460 systemd[1]: Starting containerd.service - containerd container runtime... Mar 20 21:14:30.808091 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Mar 20 21:14:30.809035 lvm[1461]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Mar 20 21:14:30.810763 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Mar 20 21:14:30.812808 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Mar 20 21:14:30.813867 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Mar 20 21:14:30.816600 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Mar 20 21:14:30.818426 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Mar 20 21:14:30.821331 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Mar 20 21:14:30.823696 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Mar 20 21:14:30.827362 systemd[1]: Starting systemd-logind.service - User Login Management... Mar 20 21:14:30.827510 jq[1464]: false Mar 20 21:14:30.831094 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Mar 20 21:14:30.831522 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Mar 20 21:14:30.832295 systemd[1]: Starting update-engine.service - Update Engine... Mar 20 21:14:30.836331 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Mar 20 21:14:30.838152 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Mar 20 21:14:30.843211 extend-filesystems[1465]: Found loop3 Mar 20 21:14:30.844281 extend-filesystems[1465]: Found loop4 Mar 20 21:14:30.845180 extend-filesystems[1465]: Found loop5 Mar 20 21:14:30.845180 extend-filesystems[1465]: Found vda Mar 20 21:14:30.845180 extend-filesystems[1465]: Found vda1 Mar 20 21:14:30.845180 extend-filesystems[1465]: Found vda2 Mar 20 21:14:30.845180 extend-filesystems[1465]: Found vda3 Mar 20 21:14:30.845180 extend-filesystems[1465]: Found usr Mar 20 21:14:30.845180 extend-filesystems[1465]: Found vda4 Mar 20 21:14:30.845180 extend-filesystems[1465]: Found vda6 Mar 20 21:14:30.845180 extend-filesystems[1465]: Found vda7 Mar 20 21:14:30.845180 extend-filesystems[1465]: Found vda9 Mar 20 21:14:30.845180 extend-filesystems[1465]: Checking size of /dev/vda9 Mar 20 21:14:30.844282 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Mar 20 21:14:30.844452 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Mar 20 21:14:30.848710 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Mar 20 21:14:30.848910 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Mar 20 21:14:30.853710 systemd[1]: motdgen.service: Deactivated successfully. Mar 20 21:14:30.853906 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Mar 20 21:14:30.855802 dbus-daemon[1463]: [system] SELinux support is enabled Mar 20 21:14:30.858436 systemd[1]: Started dbus.service - D-Bus System Message Bus. Mar 20 21:14:30.863699 jq[1476]: true Mar 20 21:14:30.866980 (ntainerd)[1485]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Mar 20 21:14:30.872790 tar[1483]: linux-arm64/helm Mar 20 21:14:30.872578 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Mar 20 21:14:30.872607 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Mar 20 21:14:30.873584 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Mar 20 21:14:30.873608 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Mar 20 21:14:30.875267 extend-filesystems[1465]: Resized partition /dev/vda9 Mar 20 21:14:30.880245 extend-filesystems[1499]: resize2fs 1.47.2 (1-Jan-2025) Mar 20 21:14:30.889380 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Mar 20 21:14:30.896388 jq[1495]: true Mar 20 21:14:30.911274 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 38 scanned by (udev-worker) (1370) Mar 20 21:14:30.914293 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Mar 20 21:14:30.916287 update_engine[1475]: I20250320 21:14:30.916118 1475 main.cc:92] Flatcar Update Engine starting Mar 20 21:14:30.925551 update_engine[1475]: I20250320 21:14:30.922426 1475 update_check_scheduler.cc:74] Next update check in 10m59s Mar 20 21:14:30.922051 systemd[1]: Started update-engine.service - Update Engine. Mar 20 21:14:30.924455 systemd[1]: Started locksmithd.service - Cluster reboot manager. Mar 20 21:14:30.928518 extend-filesystems[1499]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Mar 20 21:14:30.928518 extend-filesystems[1499]: old_desc_blocks = 1, new_desc_blocks = 1 Mar 20 21:14:30.928518 extend-filesystems[1499]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Mar 20 21:14:30.935058 extend-filesystems[1465]: Resized filesystem in /dev/vda9 Mar 20 21:14:30.934092 systemd[1]: extend-filesystems.service: Deactivated successfully. Mar 20 21:14:30.934325 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Mar 20 21:14:30.968133 systemd-logind[1470]: Watching system buttons on /dev/input/event0 (Power Button) Mar 20 21:14:30.968674 systemd-logind[1470]: New seat seat0. Mar 20 21:14:30.969291 systemd[1]: Started systemd-logind.service - User Login Management. Mar 20 21:14:30.972773 bash[1519]: Updated "/home/core/.ssh/authorized_keys" Mar 20 21:14:30.974268 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Mar 20 21:14:30.977900 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Mar 20 21:14:31.022686 locksmithd[1507]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Mar 20 21:14:31.117592 containerd[1485]: time="2025-03-20T21:14:31Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Mar 20 21:14:31.118581 containerd[1485]: time="2025-03-20T21:14:31.118547676Z" level=info msg="starting containerd" revision=88aa2f531d6c2922003cc7929e51daf1c14caa0a version=v2.0.1 Mar 20 21:14:31.130354 containerd[1485]: time="2025-03-20T21:14:31.130319516Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="6.72µs" Mar 20 21:14:31.130354 containerd[1485]: time="2025-03-20T21:14:31.130349276Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Mar 20 21:14:31.130451 containerd[1485]: time="2025-03-20T21:14:31.130366836Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Mar 20 21:14:31.130532 containerd[1485]: time="2025-03-20T21:14:31.130507156Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Mar 20 21:14:31.130532 containerd[1485]: time="2025-03-20T21:14:31.130529356Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Mar 20 21:14:31.130576 containerd[1485]: time="2025-03-20T21:14:31.130553876Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Mar 20 21:14:31.130626 containerd[1485]: time="2025-03-20T21:14:31.130609756Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Mar 20 21:14:31.130656 containerd[1485]: time="2025-03-20T21:14:31.130625756Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Mar 20 21:14:31.130916 containerd[1485]: time="2025-03-20T21:14:31.130892396Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Mar 20 21:14:31.130916 containerd[1485]: time="2025-03-20T21:14:31.130915636Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Mar 20 21:14:31.130968 containerd[1485]: time="2025-03-20T21:14:31.130926996Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Mar 20 21:14:31.130968 containerd[1485]: time="2025-03-20T21:14:31.130935316Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Mar 20 21:14:31.131015 containerd[1485]: time="2025-03-20T21:14:31.131004876Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Mar 20 21:14:31.131226 containerd[1485]: time="2025-03-20T21:14:31.131197036Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Mar 20 21:14:31.131270 containerd[1485]: time="2025-03-20T21:14:31.131251636Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Mar 20 21:14:31.131270 containerd[1485]: time="2025-03-20T21:14:31.131266316Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Mar 20 21:14:31.131316 containerd[1485]: time="2025-03-20T21:14:31.131294596Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Mar 20 21:14:31.131599 containerd[1485]: time="2025-03-20T21:14:31.131534236Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Mar 20 21:14:31.131650 containerd[1485]: time="2025-03-20T21:14:31.131602556Z" level=info msg="metadata content store policy set" policy=shared Mar 20 21:14:31.136000 containerd[1485]: time="2025-03-20T21:14:31.135968396Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Mar 20 21:14:31.136056 containerd[1485]: time="2025-03-20T21:14:31.136028316Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Mar 20 21:14:31.136056 containerd[1485]: time="2025-03-20T21:14:31.136044276Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Mar 20 21:14:31.136101 containerd[1485]: time="2025-03-20T21:14:31.136056556Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Mar 20 21:14:31.136101 containerd[1485]: time="2025-03-20T21:14:31.136081956Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Mar 20 21:14:31.136101 containerd[1485]: time="2025-03-20T21:14:31.136095276Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Mar 20 21:14:31.136179 containerd[1485]: time="2025-03-20T21:14:31.136107116Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Mar 20 21:14:31.136179 containerd[1485]: time="2025-03-20T21:14:31.136119636Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Mar 20 21:14:31.136179 containerd[1485]: time="2025-03-20T21:14:31.136134076Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Mar 20 21:14:31.136179 containerd[1485]: time="2025-03-20T21:14:31.136144956Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Mar 20 21:14:31.136179 containerd[1485]: time="2025-03-20T21:14:31.136154436Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Mar 20 21:14:31.136179 containerd[1485]: time="2025-03-20T21:14:31.136166756Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Mar 20 21:14:31.136406 containerd[1485]: time="2025-03-20T21:14:31.136302156Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Mar 20 21:14:31.136406 containerd[1485]: time="2025-03-20T21:14:31.136323756Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Mar 20 21:14:31.136406 containerd[1485]: time="2025-03-20T21:14:31.136336116Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Mar 20 21:14:31.136406 containerd[1485]: time="2025-03-20T21:14:31.136347076Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Mar 20 21:14:31.136406 containerd[1485]: time="2025-03-20T21:14:31.136358836Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Mar 20 21:14:31.136406 containerd[1485]: time="2025-03-20T21:14:31.136368716Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Mar 20 21:14:31.136406 containerd[1485]: time="2025-03-20T21:14:31.136380356Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Mar 20 21:14:31.136406 containerd[1485]: time="2025-03-20T21:14:31.136390516Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Mar 20 21:14:31.136406 containerd[1485]: time="2025-03-20T21:14:31.136401316Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Mar 20 21:14:31.136594 containerd[1485]: time="2025-03-20T21:14:31.136413316Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Mar 20 21:14:31.136594 containerd[1485]: time="2025-03-20T21:14:31.136425036Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Mar 20 21:14:31.136724 containerd[1485]: time="2025-03-20T21:14:31.136681436Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Mar 20 21:14:31.136724 containerd[1485]: time="2025-03-20T21:14:31.136702756Z" level=info msg="Start snapshots syncer" Mar 20 21:14:31.136724 containerd[1485]: time="2025-03-20T21:14:31.136729756Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Mar 20 21:14:31.138042 containerd[1485]: time="2025-03-20T21:14:31.137994716Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Mar 20 21:14:31.138250 containerd[1485]: time="2025-03-20T21:14:31.138204916Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Mar 20 21:14:31.138738 containerd[1485]: time="2025-03-20T21:14:31.138706676Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Mar 20 21:14:31.138888 containerd[1485]: time="2025-03-20T21:14:31.138848356Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Mar 20 21:14:31.138888 containerd[1485]: time="2025-03-20T21:14:31.138879916Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Mar 20 21:14:31.138936 containerd[1485]: time="2025-03-20T21:14:31.138893236Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Mar 20 21:14:31.138936 containerd[1485]: time="2025-03-20T21:14:31.138905396Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Mar 20 21:14:31.138936 containerd[1485]: time="2025-03-20T21:14:31.138920116Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Mar 20 21:14:31.138936 containerd[1485]: time="2025-03-20T21:14:31.138931596Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Mar 20 21:14:31.138999 containerd[1485]: time="2025-03-20T21:14:31.138943276Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Mar 20 21:14:31.138999 containerd[1485]: time="2025-03-20T21:14:31.138971076Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Mar 20 21:14:31.138999 containerd[1485]: time="2025-03-20T21:14:31.138988876Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Mar 20 21:14:31.139138 containerd[1485]: time="2025-03-20T21:14:31.138999716Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Mar 20 21:14:31.139138 containerd[1485]: time="2025-03-20T21:14:31.139052596Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Mar 20 21:14:31.139138 containerd[1485]: time="2025-03-20T21:14:31.139067396Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Mar 20 21:14:31.139138 containerd[1485]: time="2025-03-20T21:14:31.139077636Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Mar 20 21:14:31.139138 containerd[1485]: time="2025-03-20T21:14:31.139088156Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Mar 20 21:14:31.139138 containerd[1485]: time="2025-03-20T21:14:31.139096796Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Mar 20 21:14:31.139138 containerd[1485]: time="2025-03-20T21:14:31.139111796Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Mar 20 21:14:31.139138 containerd[1485]: time="2025-03-20T21:14:31.139123116Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Mar 20 21:14:31.139304 containerd[1485]: time="2025-03-20T21:14:31.139199996Z" level=info msg="runtime interface created" Mar 20 21:14:31.139304 containerd[1485]: time="2025-03-20T21:14:31.139205756Z" level=info msg="created NRI interface" Mar 20 21:14:31.139304 containerd[1485]: time="2025-03-20T21:14:31.139213996Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Mar 20 21:14:31.139304 containerd[1485]: time="2025-03-20T21:14:31.139250836Z" level=info msg="Connect containerd service" Mar 20 21:14:31.139304 containerd[1485]: time="2025-03-20T21:14:31.139279676Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Mar 20 21:14:31.144495 containerd[1485]: time="2025-03-20T21:14:31.144453996Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Mar 20 21:14:31.253667 containerd[1485]: time="2025-03-20T21:14:31.253507756Z" level=info msg="Start subscribing containerd event" Mar 20 21:14:31.253667 containerd[1485]: time="2025-03-20T21:14:31.253602116Z" level=info msg="Start recovering state" Mar 20 21:14:31.253887 containerd[1485]: time="2025-03-20T21:14:31.253781076Z" level=info msg="Start event monitor" Mar 20 21:14:31.253887 containerd[1485]: time="2025-03-20T21:14:31.253827036Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Mar 20 21:14:31.253887 containerd[1485]: time="2025-03-20T21:14:31.253873316Z" level=info msg=serving... address=/run/containerd/containerd.sock Mar 20 21:14:31.254185 containerd[1485]: time="2025-03-20T21:14:31.253952716Z" level=info msg="Start cni network conf syncer for default" Mar 20 21:14:31.254185 containerd[1485]: time="2025-03-20T21:14:31.253968916Z" level=info msg="Start streaming server" Mar 20 21:14:31.254185 containerd[1485]: time="2025-03-20T21:14:31.253979556Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Mar 20 21:14:31.254185 containerd[1485]: time="2025-03-20T21:14:31.253987036Z" level=info msg="runtime interface starting up..." Mar 20 21:14:31.254185 containerd[1485]: time="2025-03-20T21:14:31.253992956Z" level=info msg="starting plugins..." Mar 20 21:14:31.254363 containerd[1485]: time="2025-03-20T21:14:31.254346476Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Mar 20 21:14:31.254572 containerd[1485]: time="2025-03-20T21:14:31.254558996Z" level=info msg="containerd successfully booted in 0.137470s" Mar 20 21:14:31.254652 systemd[1]: Started containerd.service - containerd container runtime. Mar 20 21:14:31.262561 sshd_keygen[1479]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Mar 20 21:14:31.266962 tar[1483]: linux-arm64/LICENSE Mar 20 21:14:31.267021 tar[1483]: linux-arm64/README.md Mar 20 21:14:31.281550 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Mar 20 21:14:31.285191 systemd[1]: Starting issuegen.service - Generate /run/issue... Mar 20 21:14:31.286634 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Mar 20 21:14:31.301651 systemd[1]: issuegen.service: Deactivated successfully. Mar 20 21:14:31.301842 systemd[1]: Finished issuegen.service - Generate /run/issue. Mar 20 21:14:31.304367 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Mar 20 21:14:31.325379 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Mar 20 21:14:31.327866 systemd[1]: Started getty@tty1.service - Getty on tty1. Mar 20 21:14:31.329863 systemd[1]: Started serial-getty@ttyAMA0.service - Serial Getty on ttyAMA0. Mar 20 21:14:31.331049 systemd[1]: Reached target getty.target - Login Prompts. Mar 20 21:14:31.957351 systemd-networkd[1416]: eth0: Gained IPv6LL Mar 20 21:14:31.959899 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Mar 20 21:14:31.961403 systemd[1]: Reached target network-online.target - Network is Online. Mar 20 21:14:31.963623 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Mar 20 21:14:31.965697 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 20 21:14:31.974134 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Mar 20 21:14:31.988626 systemd[1]: coreos-metadata.service: Deactivated successfully. Mar 20 21:14:31.989237 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Mar 20 21:14:31.990812 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Mar 20 21:14:31.992969 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Mar 20 21:14:32.442962 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 20 21:14:32.444182 systemd[1]: Reached target multi-user.target - Multi-User System. Mar 20 21:14:32.446077 (kubelet)[1588]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 20 21:14:32.449309 systemd[1]: Startup finished in 513ms (kernel) + 4.989s (initrd) + 3.341s (userspace) = 8.844s. Mar 20 21:14:32.869712 kubelet[1588]: E0320 21:14:32.869604 1588 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 20 21:14:32.872137 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 20 21:14:32.872300 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 20 21:14:32.872844 systemd[1]: kubelet.service: Consumed 783ms CPU time, 235.3M memory peak. Mar 20 21:14:37.150905 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Mar 20 21:14:37.152092 systemd[1]: Started sshd@0-10.0.0.54:22-10.0.0.1:53012.service - OpenSSH per-connection server daemon (10.0.0.1:53012). Mar 20 21:14:37.233753 sshd[1601]: Accepted publickey for core from 10.0.0.1 port 53012 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:14:37.237189 sshd-session[1601]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:14:37.246709 systemd-logind[1470]: New session 1 of user core. Mar 20 21:14:37.247663 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Mar 20 21:14:37.248581 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Mar 20 21:14:37.271717 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Mar 20 21:14:37.274828 systemd[1]: Starting user@500.service - User Manager for UID 500... Mar 20 21:14:37.287255 (systemd)[1605]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Mar 20 21:14:37.289373 systemd-logind[1470]: New session c1 of user core. Mar 20 21:14:37.397324 systemd[1605]: Queued start job for default target default.target. Mar 20 21:14:37.413270 systemd[1605]: Created slice app.slice - User Application Slice. Mar 20 21:14:37.413296 systemd[1605]: Reached target paths.target - Paths. Mar 20 21:14:37.413345 systemd[1605]: Reached target timers.target - Timers. Mar 20 21:14:37.414605 systemd[1605]: Starting dbus.socket - D-Bus User Message Bus Socket... Mar 20 21:14:37.423803 systemd[1605]: Listening on dbus.socket - D-Bus User Message Bus Socket. Mar 20 21:14:37.423870 systemd[1605]: Reached target sockets.target - Sockets. Mar 20 21:14:37.423908 systemd[1605]: Reached target basic.target - Basic System. Mar 20 21:14:37.423936 systemd[1605]: Reached target default.target - Main User Target. Mar 20 21:14:37.423962 systemd[1605]: Startup finished in 129ms. Mar 20 21:14:37.424194 systemd[1]: Started user@500.service - User Manager for UID 500. Mar 20 21:14:37.425631 systemd[1]: Started session-1.scope - Session 1 of User core. Mar 20 21:14:37.486349 systemd[1]: Started sshd@1-10.0.0.54:22-10.0.0.1:53028.service - OpenSSH per-connection server daemon (10.0.0.1:53028). Mar 20 21:14:37.536257 sshd[1616]: Accepted publickey for core from 10.0.0.1 port 53028 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:14:37.537396 sshd-session[1616]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:14:37.541689 systemd-logind[1470]: New session 2 of user core. Mar 20 21:14:37.554431 systemd[1]: Started session-2.scope - Session 2 of User core. Mar 20 21:14:37.604599 sshd[1618]: Connection closed by 10.0.0.1 port 53028 Mar 20 21:14:37.605004 sshd-session[1616]: pam_unix(sshd:session): session closed for user core Mar 20 21:14:37.618322 systemd[1]: Started sshd@2-10.0.0.54:22-10.0.0.1:53044.service - OpenSSH per-connection server daemon (10.0.0.1:53044). Mar 20 21:14:37.618724 systemd[1]: sshd@1-10.0.0.54:22-10.0.0.1:53028.service: Deactivated successfully. Mar 20 21:14:37.620074 systemd[1]: session-2.scope: Deactivated successfully. Mar 20 21:14:37.621351 systemd-logind[1470]: Session 2 logged out. Waiting for processes to exit. Mar 20 21:14:37.632064 systemd-logind[1470]: Removed session 2. Mar 20 21:14:37.671358 sshd[1621]: Accepted publickey for core from 10.0.0.1 port 53044 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:14:37.672316 sshd-session[1621]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:14:37.676544 systemd-logind[1470]: New session 3 of user core. Mar 20 21:14:37.686351 systemd[1]: Started session-3.scope - Session 3 of User core. Mar 20 21:14:37.734320 sshd[1626]: Connection closed by 10.0.0.1 port 53044 Mar 20 21:14:37.734864 sshd-session[1621]: pam_unix(sshd:session): session closed for user core Mar 20 21:14:37.745597 systemd[1]: sshd@2-10.0.0.54:22-10.0.0.1:53044.service: Deactivated successfully. Mar 20 21:14:37.746985 systemd[1]: session-3.scope: Deactivated successfully. Mar 20 21:14:37.747626 systemd-logind[1470]: Session 3 logged out. Waiting for processes to exit. Mar 20 21:14:37.749280 systemd[1]: Started sshd@3-10.0.0.54:22-10.0.0.1:53050.service - OpenSSH per-connection server daemon (10.0.0.1:53050). Mar 20 21:14:37.750028 systemd-logind[1470]: Removed session 3. Mar 20 21:14:37.803194 sshd[1631]: Accepted publickey for core from 10.0.0.1 port 53050 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:14:37.804698 sshd-session[1631]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:14:37.808900 systemd-logind[1470]: New session 4 of user core. Mar 20 21:14:37.818423 systemd[1]: Started session-4.scope - Session 4 of User core. Mar 20 21:14:37.869532 sshd[1634]: Connection closed by 10.0.0.1 port 53050 Mar 20 21:14:37.869408 sshd-session[1631]: pam_unix(sshd:session): session closed for user core Mar 20 21:14:37.883337 systemd[1]: sshd@3-10.0.0.54:22-10.0.0.1:53050.service: Deactivated successfully. Mar 20 21:14:37.884825 systemd[1]: session-4.scope: Deactivated successfully. Mar 20 21:14:37.885536 systemd-logind[1470]: Session 4 logged out. Waiting for processes to exit. Mar 20 21:14:37.887199 systemd[1]: Started sshd@4-10.0.0.54:22-10.0.0.1:53056.service - OpenSSH per-connection server daemon (10.0.0.1:53056). Mar 20 21:14:37.888020 systemd-logind[1470]: Removed session 4. Mar 20 21:14:37.938829 sshd[1639]: Accepted publickey for core from 10.0.0.1 port 53056 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:14:37.940020 sshd-session[1639]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:14:37.944278 systemd-logind[1470]: New session 5 of user core. Mar 20 21:14:37.954422 systemd[1]: Started session-5.scope - Session 5 of User core. Mar 20 21:14:38.015501 sudo[1643]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Mar 20 21:14:38.015784 sudo[1643]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 20 21:14:38.028074 sudo[1643]: pam_unix(sudo:session): session closed for user root Mar 20 21:14:38.031115 sshd[1642]: Connection closed by 10.0.0.1 port 53056 Mar 20 21:14:38.031582 sshd-session[1639]: pam_unix(sshd:session): session closed for user core Mar 20 21:14:38.042365 systemd[1]: sshd@4-10.0.0.54:22-10.0.0.1:53056.service: Deactivated successfully. Mar 20 21:14:38.044618 systemd[1]: session-5.scope: Deactivated successfully. Mar 20 21:14:38.045290 systemd-logind[1470]: Session 5 logged out. Waiting for processes to exit. Mar 20 21:14:38.047185 systemd[1]: Started sshd@5-10.0.0.54:22-10.0.0.1:53066.service - OpenSSH per-connection server daemon (10.0.0.1:53066). Mar 20 21:14:38.048008 systemd-logind[1470]: Removed session 5. Mar 20 21:14:38.100949 sshd[1648]: Accepted publickey for core from 10.0.0.1 port 53066 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:14:38.102207 sshd-session[1648]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:14:38.107791 systemd-logind[1470]: New session 6 of user core. Mar 20 21:14:38.115391 systemd[1]: Started session-6.scope - Session 6 of User core. Mar 20 21:14:38.167159 sudo[1653]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Mar 20 21:14:38.167479 sudo[1653]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 20 21:14:38.171029 sudo[1653]: pam_unix(sudo:session): session closed for user root Mar 20 21:14:38.175583 sudo[1652]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Mar 20 21:14:38.175858 sudo[1652]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 20 21:14:38.183929 systemd[1]: Starting audit-rules.service - Load Audit Rules... Mar 20 21:14:38.217115 augenrules[1675]: No rules Mar 20 21:14:38.218398 systemd[1]: audit-rules.service: Deactivated successfully. Mar 20 21:14:38.218707 systemd[1]: Finished audit-rules.service - Load Audit Rules. Mar 20 21:14:38.220003 sudo[1652]: pam_unix(sudo:session): session closed for user root Mar 20 21:14:38.221157 sshd[1651]: Connection closed by 10.0.0.1 port 53066 Mar 20 21:14:38.221503 sshd-session[1648]: pam_unix(sshd:session): session closed for user core Mar 20 21:14:38.234287 systemd[1]: sshd@5-10.0.0.54:22-10.0.0.1:53066.service: Deactivated successfully. Mar 20 21:14:38.235824 systemd[1]: session-6.scope: Deactivated successfully. Mar 20 21:14:38.238437 systemd-logind[1470]: Session 6 logged out. Waiting for processes to exit. Mar 20 21:14:38.239651 systemd[1]: Started sshd@6-10.0.0.54:22-10.0.0.1:53078.service - OpenSSH per-connection server daemon (10.0.0.1:53078). Mar 20 21:14:38.240459 systemd-logind[1470]: Removed session 6. Mar 20 21:14:38.289630 sshd[1683]: Accepted publickey for core from 10.0.0.1 port 53078 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:14:38.290785 sshd-session[1683]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:14:38.295166 systemd-logind[1470]: New session 7 of user core. Mar 20 21:14:38.307387 systemd[1]: Started session-7.scope - Session 7 of User core. Mar 20 21:14:38.357808 sudo[1687]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Mar 20 21:14:38.358094 sudo[1687]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 20 21:14:38.745698 systemd[1]: Starting docker.service - Docker Application Container Engine... Mar 20 21:14:38.759506 (dockerd)[1709]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Mar 20 21:14:39.008235 dockerd[1709]: time="2025-03-20T21:14:39.006355556Z" level=info msg="Starting up" Mar 20 21:14:39.010578 dockerd[1709]: time="2025-03-20T21:14:39.010544316Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Mar 20 21:14:39.106250 dockerd[1709]: time="2025-03-20T21:14:39.106145036Z" level=info msg="Loading containers: start." Mar 20 21:14:39.247268 kernel: Initializing XFRM netlink socket Mar 20 21:14:39.315064 systemd-networkd[1416]: docker0: Link UP Mar 20 21:14:39.380563 dockerd[1709]: time="2025-03-20T21:14:39.380448516Z" level=info msg="Loading containers: done." Mar 20 21:14:39.394305 dockerd[1709]: time="2025-03-20T21:14:39.394251236Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Mar 20 21:14:39.394453 dockerd[1709]: time="2025-03-20T21:14:39.394339876Z" level=info msg="Docker daemon" commit=c710b88579fcb5e0d53f96dcae976d79323b9166 containerd-snapshotter=false storage-driver=overlay2 version=27.4.1 Mar 20 21:14:39.394538 dockerd[1709]: time="2025-03-20T21:14:39.394505636Z" level=info msg="Daemon has completed initialization" Mar 20 21:14:39.425413 dockerd[1709]: time="2025-03-20T21:14:39.425354036Z" level=info msg="API listen on /run/docker.sock" Mar 20 21:14:39.425527 systemd[1]: Started docker.service - Docker Application Container Engine. Mar 20 21:14:40.226436 containerd[1485]: time="2025-03-20T21:14:40.226393996Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.7\"" Mar 20 21:14:40.887896 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount142290042.mount: Deactivated successfully. Mar 20 21:14:43.069517 containerd[1485]: time="2025-03-20T21:14:43.069466036Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.31.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:14:43.070395 containerd[1485]: time="2025-03-20T21:14:43.070346276Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.31.7: active requests=0, bytes read=25552768" Mar 20 21:14:43.072261 containerd[1485]: time="2025-03-20T21:14:43.071500996Z" level=info msg="ImageCreate event name:\"sha256:26ae5fde2308729bfda71fa20aa73cb5a1a4490f107f62dc7e1c4c49823cc084\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:14:43.073469 containerd[1485]: time="2025-03-20T21:14:43.073438316Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:22c19cc70fe5806d0a2cb28a6b6b33fd34e6f9e50616bdf6d53649bcfafbc277\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:14:43.074431 containerd[1485]: time="2025-03-20T21:14:43.074387116Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.31.7\" with image id \"sha256:26ae5fde2308729bfda71fa20aa73cb5a1a4490f107f62dc7e1c4c49823cc084\", repo tag \"registry.k8s.io/kube-apiserver:v1.31.7\", repo digest \"registry.k8s.io/kube-apiserver@sha256:22c19cc70fe5806d0a2cb28a6b6b33fd34e6f9e50616bdf6d53649bcfafbc277\", size \"25549566\" in 2.84794904s" Mar 20 21:14:43.074508 containerd[1485]: time="2025-03-20T21:14:43.074429916Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.7\" returns image reference \"sha256:26ae5fde2308729bfda71fa20aa73cb5a1a4490f107f62dc7e1c4c49823cc084\"" Mar 20 21:14:43.075310 containerd[1485]: time="2025-03-20T21:14:43.075199916Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.7\"" Mar 20 21:14:43.122743 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Mar 20 21:14:43.124198 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 20 21:14:43.235367 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 20 21:14:43.238603 (kubelet)[1977]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 20 21:14:43.327196 kubelet[1977]: E0320 21:14:43.325042 1977 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 20 21:14:43.328111 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 20 21:14:43.328276 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 20 21:14:43.328554 systemd[1]: kubelet.service: Consumed 186ms CPU time, 97.3M memory peak. Mar 20 21:14:44.790116 containerd[1485]: time="2025-03-20T21:14:44.790076436Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.31.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:14:44.791031 containerd[1485]: time="2025-03-20T21:14:44.790780676Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.31.7: active requests=0, bytes read=22458980" Mar 20 21:14:44.792191 containerd[1485]: time="2025-03-20T21:14:44.792161596Z" level=info msg="ImageCreate event name:\"sha256:3f2886c2c7c101461e78c37591f8beb12ac073f8dcf5e32c95da9e9689d0c1d3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:14:44.795460 containerd[1485]: time="2025-03-20T21:14:44.795419436Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:6abe7a0accecf29db6ebab18a10f844678ffed693d79e2e51a18a6f2b4530cbb\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:14:44.796252 containerd[1485]: time="2025-03-20T21:14:44.796203836Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.31.7\" with image id \"sha256:3f2886c2c7c101461e78c37591f8beb12ac073f8dcf5e32c95da9e9689d0c1d3\", repo tag \"registry.k8s.io/kube-controller-manager:v1.31.7\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:6abe7a0accecf29db6ebab18a10f844678ffed693d79e2e51a18a6f2b4530cbb\", size \"23899774\" in 1.72095144s" Mar 20 21:14:44.796369 containerd[1485]: time="2025-03-20T21:14:44.796352036Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.7\" returns image reference \"sha256:3f2886c2c7c101461e78c37591f8beb12ac073f8dcf5e32c95da9e9689d0c1d3\"" Mar 20 21:14:44.797552 containerd[1485]: time="2025-03-20T21:14:44.797526676Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.7\"" Mar 20 21:14:46.420034 containerd[1485]: time="2025-03-20T21:14:46.419974436Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.31.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:14:46.420627 containerd[1485]: time="2025-03-20T21:14:46.420573076Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.31.7: active requests=0, bytes read=17125831" Mar 20 21:14:46.421306 containerd[1485]: time="2025-03-20T21:14:46.421249556Z" level=info msg="ImageCreate event name:\"sha256:3dd474fdc8c0d007008dd47bafecdd344fbdace928731ae8b09f58f633f4a30f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:14:46.423683 containerd[1485]: time="2025-03-20T21:14:46.423654116Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:fb80249bcb77ee72b1c9fa5b70bc28a83ed107c9ca71957841ad91db379963bf\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:14:46.425380 containerd[1485]: time="2025-03-20T21:14:46.425167396Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.31.7\" with image id \"sha256:3dd474fdc8c0d007008dd47bafecdd344fbdace928731ae8b09f58f633f4a30f\", repo tag \"registry.k8s.io/kube-scheduler:v1.31.7\", repo digest \"registry.k8s.io/kube-scheduler@sha256:fb80249bcb77ee72b1c9fa5b70bc28a83ed107c9ca71957841ad91db379963bf\", size \"18566643\" in 1.62760856s" Mar 20 21:14:46.425380 containerd[1485]: time="2025-03-20T21:14:46.425204036Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.7\" returns image reference \"sha256:3dd474fdc8c0d007008dd47bafecdd344fbdace928731ae8b09f58f633f4a30f\"" Mar 20 21:14:46.425739 containerd[1485]: time="2025-03-20T21:14:46.425642316Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.7\"" Mar 20 21:14:47.493261 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1293095975.mount: Deactivated successfully. Mar 20 21:14:47.730160 containerd[1485]: time="2025-03-20T21:14:47.730114236Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.31.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:14:47.730998 containerd[1485]: time="2025-03-20T21:14:47.730816076Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.31.7: active requests=0, bytes read=26871917" Mar 20 21:14:47.731857 containerd[1485]: time="2025-03-20T21:14:47.731776436Z" level=info msg="ImageCreate event name:\"sha256:939054a0dc9c7c1596b061fc2380758139ce62751b44a0b21b3afc7abd7eb3ff\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:14:47.733717 containerd[1485]: time="2025-03-20T21:14:47.733661876Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:e5839270c96c3ad1bea1dce4935126d3281297527f3655408d2970aa4b5cf178\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:14:47.734317 containerd[1485]: time="2025-03-20T21:14:47.734166476Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.31.7\" with image id \"sha256:939054a0dc9c7c1596b061fc2380758139ce62751b44a0b21b3afc7abd7eb3ff\", repo tag \"registry.k8s.io/kube-proxy:v1.31.7\", repo digest \"registry.k8s.io/kube-proxy@sha256:e5839270c96c3ad1bea1dce4935126d3281297527f3655408d2970aa4b5cf178\", size \"26870934\" in 1.30847416s" Mar 20 21:14:47.734317 containerd[1485]: time="2025-03-20T21:14:47.734203876Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.7\" returns image reference \"sha256:939054a0dc9c7c1596b061fc2380758139ce62751b44a0b21b3afc7abd7eb3ff\"" Mar 20 21:14:47.734773 containerd[1485]: time="2025-03-20T21:14:47.734747836Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\"" Mar 20 21:14:48.303622 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2635030056.mount: Deactivated successfully. Mar 20 21:14:49.316701 containerd[1485]: time="2025-03-20T21:14:49.316635956Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:14:49.317730 containerd[1485]: time="2025-03-20T21:14:49.317204676Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.1: active requests=0, bytes read=16485383" Mar 20 21:14:49.318369 containerd[1485]: time="2025-03-20T21:14:49.318280356Z" level=info msg="ImageCreate event name:\"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:14:49.321278 containerd[1485]: time="2025-03-20T21:14:49.321117196Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:14:49.322984 containerd[1485]: time="2025-03-20T21:14:49.322951556Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.1\" with image id \"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\", size \"16482581\" in 1.58817168s" Mar 20 21:14:49.323059 containerd[1485]: time="2025-03-20T21:14:49.322988436Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\" returns image reference \"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\"" Mar 20 21:14:49.323574 containerd[1485]: time="2025-03-20T21:14:49.323409876Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Mar 20 21:14:49.738419 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3841287183.mount: Deactivated successfully. Mar 20 21:14:49.744633 containerd[1485]: time="2025-03-20T21:14:49.744589596Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 20 21:14:49.745423 containerd[1485]: time="2025-03-20T21:14:49.745315076Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=268705" Mar 20 21:14:49.746084 containerd[1485]: time="2025-03-20T21:14:49.746049756Z" level=info msg="ImageCreate event name:\"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 20 21:14:49.748242 containerd[1485]: time="2025-03-20T21:14:49.748196196Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 20 21:14:49.749197 containerd[1485]: time="2025-03-20T21:14:49.749157596Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"267933\" in 425.7194ms" Mar 20 21:14:49.749197 containerd[1485]: time="2025-03-20T21:14:49.749189876Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\"" Mar 20 21:14:49.749635 containerd[1485]: time="2025-03-20T21:14:49.749611716Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\"" Mar 20 21:14:50.243597 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount391953765.mount: Deactivated successfully. Mar 20 21:14:53.086924 containerd[1485]: time="2025-03-20T21:14:53.086773756Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.15-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:14:53.087820 containerd[1485]: time="2025-03-20T21:14:53.087753556Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.15-0: active requests=0, bytes read=66406427" Mar 20 21:14:53.090115 containerd[1485]: time="2025-03-20T21:14:53.088888716Z" level=info msg="ImageCreate event name:\"sha256:27e3830e1402783674d8b594038967deea9d51f0d91b34c93c8f39d2f68af7da\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:14:53.091291 containerd[1485]: time="2025-03-20T21:14:53.091264676Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:14:53.092388 containerd[1485]: time="2025-03-20T21:14:53.092351916Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.15-0\" with image id \"sha256:27e3830e1402783674d8b594038967deea9d51f0d91b34c93c8f39d2f68af7da\", repo tag \"registry.k8s.io/etcd:3.5.15-0\", repo digest \"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\", size \"66535646\" in 3.34270768s" Mar 20 21:14:53.092388 containerd[1485]: time="2025-03-20T21:14:53.092384716Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\" returns image reference \"sha256:27e3830e1402783674d8b594038967deea9d51f0d91b34c93c8f39d2f68af7da\"" Mar 20 21:14:53.578714 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Mar 20 21:14:53.580763 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 20 21:14:53.685810 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 20 21:14:53.688877 (kubelet)[2120]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 20 21:14:53.730810 kubelet[2120]: E0320 21:14:53.730745 2120 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 20 21:14:53.734213 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 20 21:14:53.734383 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 20 21:14:53.734779 systemd[1]: kubelet.service: Consumed 128ms CPU time, 97.1M memory peak. Mar 20 21:14:57.504718 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 20 21:14:57.504854 systemd[1]: kubelet.service: Consumed 128ms CPU time, 97.1M memory peak. Mar 20 21:14:57.506780 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 20 21:14:57.529854 systemd[1]: Reload requested from client PID 2149 ('systemctl') (unit session-7.scope)... Mar 20 21:14:57.529874 systemd[1]: Reloading... Mar 20 21:14:57.607279 zram_generator::config[2194]: No configuration found. Mar 20 21:14:57.688741 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 20 21:14:57.764163 systemd[1]: Reloading finished in 233 ms. Mar 20 21:14:57.805350 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 20 21:14:57.806978 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Mar 20 21:14:57.808629 systemd[1]: kubelet.service: Deactivated successfully. Mar 20 21:14:57.808866 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 20 21:14:57.808901 systemd[1]: kubelet.service: Consumed 86ms CPU time, 82.6M memory peak. Mar 20 21:14:57.810267 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 20 21:14:57.925186 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 20 21:14:57.928509 (kubelet)[2240]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Mar 20 21:14:57.962225 kubelet[2240]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 20 21:14:57.962225 kubelet[2240]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Mar 20 21:14:57.962225 kubelet[2240]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 20 21:14:57.962560 kubelet[2240]: I0320 21:14:57.962426 2240 server.go:206] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 20 21:14:58.876117 kubelet[2240]: I0320 21:14:58.876067 2240 server.go:486] "Kubelet version" kubeletVersion="v1.31.0" Mar 20 21:14:58.876117 kubelet[2240]: I0320 21:14:58.876104 2240 server.go:488] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 20 21:14:58.876425 kubelet[2240]: I0320 21:14:58.876392 2240 server.go:929] "Client rotation is on, will bootstrap in background" Mar 20 21:14:58.942153 kubelet[2240]: E0320 21:14:58.942107 2240 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.0.0.54:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.54:6443: connect: connection refused" logger="UnhandledError" Mar 20 21:14:58.942663 kubelet[2240]: I0320 21:14:58.942635 2240 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 20 21:14:58.952260 kubelet[2240]: I0320 21:14:58.952232 2240 server.go:1426] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 20 21:14:58.958296 kubelet[2240]: I0320 21:14:58.958270 2240 server.go:744] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Mar 20 21:14:58.959315 kubelet[2240]: I0320 21:14:58.959272 2240 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Mar 20 21:14:58.959480 kubelet[2240]: I0320 21:14:58.959449 2240 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 20 21:14:58.959662 kubelet[2240]: I0320 21:14:58.959478 2240 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 20 21:14:58.959773 kubelet[2240]: I0320 21:14:58.959760 2240 topology_manager.go:138] "Creating topology manager with none policy" Mar 20 21:14:58.959773 kubelet[2240]: I0320 21:14:58.959773 2240 container_manager_linux.go:300] "Creating device plugin manager" Mar 20 21:14:58.960066 kubelet[2240]: I0320 21:14:58.960045 2240 state_mem.go:36] "Initialized new in-memory state store" Mar 20 21:14:58.962375 kubelet[2240]: I0320 21:14:58.962338 2240 kubelet.go:408] "Attempting to sync node with API server" Mar 20 21:14:58.962375 kubelet[2240]: I0320 21:14:58.962374 2240 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 20 21:14:58.962651 kubelet[2240]: I0320 21:14:58.962407 2240 kubelet.go:314] "Adding apiserver pod source" Mar 20 21:14:58.962651 kubelet[2240]: I0320 21:14:58.962419 2240 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 20 21:14:58.965802 kubelet[2240]: W0320 21:14:58.965626 2240 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.54:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.54:6443: connect: connection refused Mar 20 21:14:58.965802 kubelet[2240]: E0320 21:14:58.965693 2240 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.0.0.54:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.54:6443: connect: connection refused" logger="UnhandledError" Mar 20 21:14:58.965802 kubelet[2240]: W0320 21:14:58.965724 2240 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.54:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.0.0.54:6443: connect: connection refused Mar 20 21:14:58.965802 kubelet[2240]: E0320 21:14:58.965775 2240 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.0.0.54:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.54:6443: connect: connection refused" logger="UnhandledError" Mar 20 21:14:58.966045 kubelet[2240]: I0320 21:14:58.966026 2240 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v2.0.1" apiVersion="v1" Mar 20 21:14:58.967968 kubelet[2240]: I0320 21:14:58.967943 2240 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Mar 20 21:14:58.969010 kubelet[2240]: W0320 21:14:58.968988 2240 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Mar 20 21:14:58.969934 kubelet[2240]: I0320 21:14:58.969914 2240 server.go:1269] "Started kubelet" Mar 20 21:14:58.971564 kubelet[2240]: I0320 21:14:58.971535 2240 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 20 21:14:58.972599 kubelet[2240]: I0320 21:14:58.972142 2240 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Mar 20 21:14:58.974243 kubelet[2240]: I0320 21:14:58.974142 2240 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Mar 20 21:14:58.974243 kubelet[2240]: I0320 21:14:58.974211 2240 server.go:460] "Adding debug handlers to kubelet server" Mar 20 21:14:58.975341 kubelet[2240]: I0320 21:14:58.975290 2240 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 20 21:14:58.976056 kubelet[2240]: I0320 21:14:58.975529 2240 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 20 21:14:58.976147 kubelet[2240]: I0320 21:14:58.976129 2240 volume_manager.go:289] "Starting Kubelet Volume Manager" Mar 20 21:14:58.976495 kubelet[2240]: I0320 21:14:58.976248 2240 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 20 21:14:58.976495 kubelet[2240]: I0320 21:14:58.976299 2240 reconciler.go:26] "Reconciler: start to sync state" Mar 20 21:14:58.976495 kubelet[2240]: W0320 21:14:58.976312 2240 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.54:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.54:6443: connect: connection refused Mar 20 21:14:58.976495 kubelet[2240]: E0320 21:14:58.976358 2240 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.0.0.54:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.54:6443: connect: connection refused" logger="UnhandledError" Mar 20 21:14:58.976635 kubelet[2240]: E0320 21:14:58.976517 2240 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.54:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.54:6443: connect: connection refused" interval="200ms" Mar 20 21:14:58.981907 kubelet[2240]: E0320 21:14:58.976949 2240 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 20 21:14:58.981907 kubelet[2240]: I0320 21:14:58.977311 2240 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Mar 20 21:14:58.986499 kubelet[2240]: E0320 21:14:58.974912 2240 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.54:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.54:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.182e9f5654560a24 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-03-20 21:14:58.969889316 +0000 UTC m=+1.038618881,LastTimestamp:2025-03-20 21:14:58.969889316 +0000 UTC m=+1.038618881,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Mar 20 21:14:58.986499 kubelet[2240]: E0320 21:14:58.986488 2240 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Mar 20 21:14:58.990179 kubelet[2240]: I0320 21:14:58.990150 2240 factory.go:221] Registration of the containerd container factory successfully Mar 20 21:14:58.990179 kubelet[2240]: I0320 21:14:58.990175 2240 factory.go:221] Registration of the systemd container factory successfully Mar 20 21:14:58.998810 kubelet[2240]: I0320 21:14:58.998768 2240 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Mar 20 21:14:59.000003 kubelet[2240]: I0320 21:14:58.999976 2240 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Mar 20 21:14:59.000003 kubelet[2240]: I0320 21:14:59.000003 2240 status_manager.go:217] "Starting to sync pod status with apiserver" Mar 20 21:14:59.000081 kubelet[2240]: I0320 21:14:59.000020 2240 kubelet.go:2321] "Starting kubelet main sync loop" Mar 20 21:14:59.000081 kubelet[2240]: E0320 21:14:59.000070 2240 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 20 21:14:59.000735 kubelet[2240]: W0320 21:14:59.000704 2240 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.54:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.54:6443: connect: connection refused Mar 20 21:14:59.000820 kubelet[2240]: E0320 21:14:59.000747 2240 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.0.0.54:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.54:6443: connect: connection refused" logger="UnhandledError" Mar 20 21:14:59.004696 kubelet[2240]: I0320 21:14:59.004673 2240 cpu_manager.go:214] "Starting CPU manager" policy="none" Mar 20 21:14:59.004696 kubelet[2240]: I0320 21:14:59.004691 2240 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Mar 20 21:14:59.004791 kubelet[2240]: I0320 21:14:59.004712 2240 state_mem.go:36] "Initialized new in-memory state store" Mar 20 21:14:59.078006 kubelet[2240]: E0320 21:14:59.077971 2240 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 20 21:14:59.100370 kubelet[2240]: E0320 21:14:59.100328 2240 kubelet.go:2345] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Mar 20 21:14:59.107457 kubelet[2240]: I0320 21:14:59.107425 2240 policy_none.go:49] "None policy: Start" Mar 20 21:14:59.108184 kubelet[2240]: I0320 21:14:59.108160 2240 memory_manager.go:170] "Starting memorymanager" policy="None" Mar 20 21:14:59.108184 kubelet[2240]: I0320 21:14:59.108189 2240 state_mem.go:35] "Initializing new in-memory state store" Mar 20 21:14:59.115328 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Mar 20 21:14:59.126330 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Mar 20 21:14:59.129971 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Mar 20 21:14:59.141816 kubelet[2240]: I0320 21:14:59.141203 2240 manager.go:510] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Mar 20 21:14:59.141816 kubelet[2240]: I0320 21:14:59.141437 2240 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 20 21:14:59.141816 kubelet[2240]: I0320 21:14:59.141448 2240 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 20 21:14:59.141816 kubelet[2240]: I0320 21:14:59.141719 2240 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 20 21:14:59.143635 kubelet[2240]: E0320 21:14:59.143603 2240 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Mar 20 21:14:59.177843 kubelet[2240]: E0320 21:14:59.177787 2240 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.54:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.54:6443: connect: connection refused" interval="400ms" Mar 20 21:14:59.244121 kubelet[2240]: I0320 21:14:59.243873 2240 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Mar 20 21:14:59.244448 kubelet[2240]: E0320 21:14:59.244286 2240 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.0.0.54:6443/api/v1/nodes\": dial tcp 10.0.0.54:6443: connect: connection refused" node="localhost" Mar 20 21:14:59.312147 systemd[1]: Created slice kubepods-burstable-pod2260474bc3e2e353aadeddb8b8d59f55.slice - libcontainer container kubepods-burstable-pod2260474bc3e2e353aadeddb8b8d59f55.slice. Mar 20 21:14:59.328456 systemd[1]: Created slice kubepods-burstable-pod60762308083b5ef6c837b1be48ec53d6.slice - libcontainer container kubepods-burstable-pod60762308083b5ef6c837b1be48ec53d6.slice. Mar 20 21:14:59.332977 systemd[1]: Created slice kubepods-burstable-pod6f32907a07e55aea05abdc5cd284a8d5.slice - libcontainer container kubepods-burstable-pod6f32907a07e55aea05abdc5cd284a8d5.slice. Mar 20 21:14:59.377805 kubelet[2240]: I0320 21:14:59.377688 2240 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/2260474bc3e2e353aadeddb8b8d59f55-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"2260474bc3e2e353aadeddb8b8d59f55\") " pod="kube-system/kube-apiserver-localhost" Mar 20 21:14:59.377805 kubelet[2240]: I0320 21:14:59.377723 2240 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/60762308083b5ef6c837b1be48ec53d6-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"60762308083b5ef6c837b1be48ec53d6\") " pod="kube-system/kube-controller-manager-localhost" Mar 20 21:14:59.377805 kubelet[2240]: I0320 21:14:59.377744 2240 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/60762308083b5ef6c837b1be48ec53d6-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"60762308083b5ef6c837b1be48ec53d6\") " pod="kube-system/kube-controller-manager-localhost" Mar 20 21:14:59.377805 kubelet[2240]: I0320 21:14:59.377762 2240 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/60762308083b5ef6c837b1be48ec53d6-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"60762308083b5ef6c837b1be48ec53d6\") " pod="kube-system/kube-controller-manager-localhost" Mar 20 21:14:59.377805 kubelet[2240]: I0320 21:14:59.377775 2240 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/6f32907a07e55aea05abdc5cd284a8d5-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"6f32907a07e55aea05abdc5cd284a8d5\") " pod="kube-system/kube-scheduler-localhost" Mar 20 21:14:59.377985 kubelet[2240]: I0320 21:14:59.377789 2240 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/2260474bc3e2e353aadeddb8b8d59f55-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"2260474bc3e2e353aadeddb8b8d59f55\") " pod="kube-system/kube-apiserver-localhost" Mar 20 21:14:59.377985 kubelet[2240]: I0320 21:14:59.377803 2240 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/2260474bc3e2e353aadeddb8b8d59f55-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"2260474bc3e2e353aadeddb8b8d59f55\") " pod="kube-system/kube-apiserver-localhost" Mar 20 21:14:59.377985 kubelet[2240]: I0320 21:14:59.377816 2240 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/60762308083b5ef6c837b1be48ec53d6-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"60762308083b5ef6c837b1be48ec53d6\") " pod="kube-system/kube-controller-manager-localhost" Mar 20 21:14:59.377985 kubelet[2240]: I0320 21:14:59.377832 2240 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/60762308083b5ef6c837b1be48ec53d6-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"60762308083b5ef6c837b1be48ec53d6\") " pod="kube-system/kube-controller-manager-localhost" Mar 20 21:14:59.446525 kubelet[2240]: I0320 21:14:59.446394 2240 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Mar 20 21:14:59.446739 kubelet[2240]: E0320 21:14:59.446688 2240 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.0.0.54:6443/api/v1/nodes\": dial tcp 10.0.0.54:6443: connect: connection refused" node="localhost" Mar 20 21:14:59.578542 kubelet[2240]: E0320 21:14:59.578498 2240 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.54:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.54:6443: connect: connection refused" interval="800ms" Mar 20 21:14:59.626479 containerd[1485]: time="2025-03-20T21:14:59.626430876Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:2260474bc3e2e353aadeddb8b8d59f55,Namespace:kube-system,Attempt:0,}" Mar 20 21:14:59.632007 containerd[1485]: time="2025-03-20T21:14:59.631927276Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:60762308083b5ef6c837b1be48ec53d6,Namespace:kube-system,Attempt:0,}" Mar 20 21:14:59.636532 containerd[1485]: time="2025-03-20T21:14:59.636495276Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:6f32907a07e55aea05abdc5cd284a8d5,Namespace:kube-system,Attempt:0,}" Mar 20 21:14:59.710169 containerd[1485]: time="2025-03-20T21:14:59.710107836Z" level=info msg="connecting to shim de3abc48bdc665541ddedcf492559b015e8e6d6c54ecab378e29eeb218a07158" address="unix:///run/containerd/s/26b911832b1a82a925579dc95402fc28cf933844fcc43812a83cdfa4bdbd1866" namespace=k8s.io protocol=ttrpc version=3 Mar 20 21:14:59.711635 containerd[1485]: time="2025-03-20T21:14:59.711503436Z" level=info msg="connecting to shim bf1cd9c76fb6584ba737a16fa998bca0cb27eb73a90f43c71cd107df56cdbfc1" address="unix:///run/containerd/s/4ebc6dbe3628c958b6caaded93b57e1ae33b8e238850597344e8087f23dccde3" namespace=k8s.io protocol=ttrpc version=3 Mar 20 21:14:59.713822 containerd[1485]: time="2025-03-20T21:14:59.713785556Z" level=info msg="connecting to shim c6ff0b709e5aa0311d65ae1f9ca839bb12942d73dd4e5c17066b05ef4f14c066" address="unix:///run/containerd/s/d0877fe1f0cb7e2ccb3b3e7818498ba48d16c15d8c660debef04be569c2600ec" namespace=k8s.io protocol=ttrpc version=3 Mar 20 21:14:59.738643 systemd[1]: Started cri-containerd-bf1cd9c76fb6584ba737a16fa998bca0cb27eb73a90f43c71cd107df56cdbfc1.scope - libcontainer container bf1cd9c76fb6584ba737a16fa998bca0cb27eb73a90f43c71cd107df56cdbfc1. Mar 20 21:14:59.743993 systemd[1]: Started cri-containerd-de3abc48bdc665541ddedcf492559b015e8e6d6c54ecab378e29eeb218a07158.scope - libcontainer container de3abc48bdc665541ddedcf492559b015e8e6d6c54ecab378e29eeb218a07158. Mar 20 21:14:59.747901 systemd[1]: Started cri-containerd-c6ff0b709e5aa0311d65ae1f9ca839bb12942d73dd4e5c17066b05ef4f14c066.scope - libcontainer container c6ff0b709e5aa0311d65ae1f9ca839bb12942d73dd4e5c17066b05ef4f14c066. Mar 20 21:14:59.788248 containerd[1485]: time="2025-03-20T21:14:59.787379116Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:2260474bc3e2e353aadeddb8b8d59f55,Namespace:kube-system,Attempt:0,} returns sandbox id \"bf1cd9c76fb6584ba737a16fa998bca0cb27eb73a90f43c71cd107df56cdbfc1\"" Mar 20 21:14:59.791996 containerd[1485]: time="2025-03-20T21:14:59.791921676Z" level=info msg="CreateContainer within sandbox \"bf1cd9c76fb6584ba737a16fa998bca0cb27eb73a90f43c71cd107df56cdbfc1\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Mar 20 21:14:59.792559 containerd[1485]: time="2025-03-20T21:14:59.792489876Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:60762308083b5ef6c837b1be48ec53d6,Namespace:kube-system,Attempt:0,} returns sandbox id \"c6ff0b709e5aa0311d65ae1f9ca839bb12942d73dd4e5c17066b05ef4f14c066\"" Mar 20 21:14:59.795388 containerd[1485]: time="2025-03-20T21:14:59.795349836Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:6f32907a07e55aea05abdc5cd284a8d5,Namespace:kube-system,Attempt:0,} returns sandbox id \"de3abc48bdc665541ddedcf492559b015e8e6d6c54ecab378e29eeb218a07158\"" Mar 20 21:14:59.795572 containerd[1485]: time="2025-03-20T21:14:59.795538796Z" level=info msg="CreateContainer within sandbox \"c6ff0b709e5aa0311d65ae1f9ca839bb12942d73dd4e5c17066b05ef4f14c066\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Mar 20 21:14:59.798044 containerd[1485]: time="2025-03-20T21:14:59.798018076Z" level=info msg="CreateContainer within sandbox \"de3abc48bdc665541ddedcf492559b015e8e6d6c54ecab378e29eeb218a07158\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Mar 20 21:14:59.801164 containerd[1485]: time="2025-03-20T21:14:59.801123436Z" level=info msg="Container 25aae4bb122020e866107740e1d02419fe399ef4709ba2625efe6aca7b3cc9f9: CDI devices from CRI Config.CDIDevices: []" Mar 20 21:14:59.806938 containerd[1485]: time="2025-03-20T21:14:59.806893036Z" level=info msg="Container 5b814b57126748bd8c6189166eba713cdae0bff67bbc0f7e18300b807337421d: CDI devices from CRI Config.CDIDevices: []" Mar 20 21:14:59.811051 containerd[1485]: time="2025-03-20T21:14:59.810998716Z" level=info msg="CreateContainer within sandbox \"bf1cd9c76fb6584ba737a16fa998bca0cb27eb73a90f43c71cd107df56cdbfc1\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"25aae4bb122020e866107740e1d02419fe399ef4709ba2625efe6aca7b3cc9f9\"" Mar 20 21:14:59.811659 containerd[1485]: time="2025-03-20T21:14:59.811601036Z" level=info msg="StartContainer for \"25aae4bb122020e866107740e1d02419fe399ef4709ba2625efe6aca7b3cc9f9\"" Mar 20 21:14:59.812810 containerd[1485]: time="2025-03-20T21:14:59.812779516Z" level=info msg="connecting to shim 25aae4bb122020e866107740e1d02419fe399ef4709ba2625efe6aca7b3cc9f9" address="unix:///run/containerd/s/4ebc6dbe3628c958b6caaded93b57e1ae33b8e238850597344e8087f23dccde3" protocol=ttrpc version=3 Mar 20 21:14:59.815741 containerd[1485]: time="2025-03-20T21:14:59.815639036Z" level=info msg="Container 9550a2a8546bdbccb1898a840c1920a481a7835b038827b866affb4dfc23e258: CDI devices from CRI Config.CDIDevices: []" Mar 20 21:14:59.821324 containerd[1485]: time="2025-03-20T21:14:59.821288156Z" level=info msg="CreateContainer within sandbox \"c6ff0b709e5aa0311d65ae1f9ca839bb12942d73dd4e5c17066b05ef4f14c066\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"5b814b57126748bd8c6189166eba713cdae0bff67bbc0f7e18300b807337421d\"" Mar 20 21:14:59.822093 containerd[1485]: time="2025-03-20T21:14:59.822038556Z" level=info msg="StartContainer for \"5b814b57126748bd8c6189166eba713cdae0bff67bbc0f7e18300b807337421d\"" Mar 20 21:14:59.824619 containerd[1485]: time="2025-03-20T21:14:59.824466116Z" level=info msg="connecting to shim 5b814b57126748bd8c6189166eba713cdae0bff67bbc0f7e18300b807337421d" address="unix:///run/containerd/s/d0877fe1f0cb7e2ccb3b3e7818498ba48d16c15d8c660debef04be569c2600ec" protocol=ttrpc version=3 Mar 20 21:14:59.824962 containerd[1485]: time="2025-03-20T21:14:59.824928716Z" level=info msg="CreateContainer within sandbox \"de3abc48bdc665541ddedcf492559b015e8e6d6c54ecab378e29eeb218a07158\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"9550a2a8546bdbccb1898a840c1920a481a7835b038827b866affb4dfc23e258\"" Mar 20 21:14:59.825588 containerd[1485]: time="2025-03-20T21:14:59.825557356Z" level=info msg="StartContainer for \"9550a2a8546bdbccb1898a840c1920a481a7835b038827b866affb4dfc23e258\"" Mar 20 21:14:59.826682 containerd[1485]: time="2025-03-20T21:14:59.826638716Z" level=info msg="connecting to shim 9550a2a8546bdbccb1898a840c1920a481a7835b038827b866affb4dfc23e258" address="unix:///run/containerd/s/26b911832b1a82a925579dc95402fc28cf933844fcc43812a83cdfa4bdbd1866" protocol=ttrpc version=3 Mar 20 21:14:59.836403 systemd[1]: Started cri-containerd-25aae4bb122020e866107740e1d02419fe399ef4709ba2625efe6aca7b3cc9f9.scope - libcontainer container 25aae4bb122020e866107740e1d02419fe399ef4709ba2625efe6aca7b3cc9f9. Mar 20 21:14:59.840317 systemd[1]: Started cri-containerd-5b814b57126748bd8c6189166eba713cdae0bff67bbc0f7e18300b807337421d.scope - libcontainer container 5b814b57126748bd8c6189166eba713cdae0bff67bbc0f7e18300b807337421d. Mar 20 21:14:59.843064 systemd[1]: Started cri-containerd-9550a2a8546bdbccb1898a840c1920a481a7835b038827b866affb4dfc23e258.scope - libcontainer container 9550a2a8546bdbccb1898a840c1920a481a7835b038827b866affb4dfc23e258. Mar 20 21:14:59.848213 kubelet[2240]: I0320 21:14:59.848174 2240 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Mar 20 21:14:59.849653 kubelet[2240]: E0320 21:14:59.848781 2240 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.0.0.54:6443/api/v1/nodes\": dial tcp 10.0.0.54:6443: connect: connection refused" node="localhost" Mar 20 21:14:59.890989 containerd[1485]: time="2025-03-20T21:14:59.886959956Z" level=info msg="StartContainer for \"25aae4bb122020e866107740e1d02419fe399ef4709ba2625efe6aca7b3cc9f9\" returns successfully" Mar 20 21:14:59.906128 containerd[1485]: time="2025-03-20T21:14:59.906098116Z" level=info msg="StartContainer for \"9550a2a8546bdbccb1898a840c1920a481a7835b038827b866affb4dfc23e258\" returns successfully" Mar 20 21:14:59.907261 containerd[1485]: time="2025-03-20T21:14:59.907238996Z" level=info msg="StartContainer for \"5b814b57126748bd8c6189166eba713cdae0bff67bbc0f7e18300b807337421d\" returns successfully" Mar 20 21:14:59.976949 kubelet[2240]: W0320 21:14:59.976875 2240 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.54:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.54:6443: connect: connection refused Mar 20 21:14:59.977320 kubelet[2240]: E0320 21:14:59.976955 2240 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.0.0.54:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.54:6443: connect: connection refused" logger="UnhandledError" Mar 20 21:15:00.651685 kubelet[2240]: I0320 21:15:00.651622 2240 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Mar 20 21:15:01.667628 kubelet[2240]: E0320 21:15:01.667581 2240 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Mar 20 21:15:01.871287 kubelet[2240]: I0320 21:15:01.871244 2240 kubelet_node_status.go:75] "Successfully registered node" node="localhost" Mar 20 21:15:01.871287 kubelet[2240]: E0320 21:15:01.871290 2240 kubelet_node_status.go:535] "Error updating node status, will retry" err="error getting node \"localhost\": node \"localhost\" not found" Mar 20 21:15:01.880978 kubelet[2240]: E0320 21:15:01.880942 2240 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 20 21:15:01.981347 kubelet[2240]: E0320 21:15:01.981234 2240 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 20 21:15:02.081388 kubelet[2240]: E0320 21:15:02.081347 2240 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 20 21:15:02.182026 kubelet[2240]: E0320 21:15:02.181968 2240 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 20 21:15:02.282622 kubelet[2240]: E0320 21:15:02.282517 2240 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 20 21:15:02.383078 kubelet[2240]: E0320 21:15:02.383033 2240 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 20 21:15:02.484078 kubelet[2240]: E0320 21:15:02.484040 2240 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 20 21:15:02.584720 kubelet[2240]: E0320 21:15:02.584607 2240 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 20 21:15:02.684906 kubelet[2240]: E0320 21:15:02.684858 2240 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 20 21:15:02.785713 kubelet[2240]: E0320 21:15:02.785678 2240 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 20 21:15:02.886746 kubelet[2240]: E0320 21:15:02.886450 2240 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 20 21:15:02.970614 kubelet[2240]: I0320 21:15:02.970369 2240 apiserver.go:52] "Watching apiserver" Mar 20 21:15:02.976612 kubelet[2240]: I0320 21:15:02.976559 2240 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 20 21:15:03.889142 systemd[1]: Reload requested from client PID 2510 ('systemctl') (unit session-7.scope)... Mar 20 21:15:03.889160 systemd[1]: Reloading... Mar 20 21:15:03.956321 zram_generator::config[2557]: No configuration found. Mar 20 21:15:04.035912 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 20 21:15:04.118616 systemd[1]: Reloading finished in 229 ms. Mar 20 21:15:04.136397 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Mar 20 21:15:04.146415 systemd[1]: kubelet.service: Deactivated successfully. Mar 20 21:15:04.146745 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 20 21:15:04.146880 systemd[1]: kubelet.service: Consumed 1.467s CPU time, 118.1M memory peak. Mar 20 21:15:04.148662 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 20 21:15:04.262160 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 20 21:15:04.265770 (kubelet)[2596]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Mar 20 21:15:04.298739 kubelet[2596]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 20 21:15:04.298739 kubelet[2596]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Mar 20 21:15:04.298739 kubelet[2596]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 20 21:15:04.299063 kubelet[2596]: I0320 21:15:04.298802 2596 server.go:206] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 20 21:15:04.305241 kubelet[2596]: I0320 21:15:04.304305 2596 server.go:486] "Kubelet version" kubeletVersion="v1.31.0" Mar 20 21:15:04.305241 kubelet[2596]: I0320 21:15:04.304328 2596 server.go:488] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 20 21:15:04.305241 kubelet[2596]: I0320 21:15:04.304532 2596 server.go:929] "Client rotation is on, will bootstrap in background" Mar 20 21:15:04.307371 kubelet[2596]: I0320 21:15:04.307347 2596 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Mar 20 21:15:04.309335 kubelet[2596]: I0320 21:15:04.309304 2596 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 20 21:15:04.312904 kubelet[2596]: I0320 21:15:04.312880 2596 server.go:1426] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 20 21:15:04.315680 kubelet[2596]: I0320 21:15:04.315647 2596 server.go:744] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Mar 20 21:15:04.315795 kubelet[2596]: I0320 21:15:04.315773 2596 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Mar 20 21:15:04.315901 kubelet[2596]: I0320 21:15:04.315871 2596 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 20 21:15:04.316060 kubelet[2596]: I0320 21:15:04.315894 2596 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 20 21:15:04.316136 kubelet[2596]: I0320 21:15:04.316065 2596 topology_manager.go:138] "Creating topology manager with none policy" Mar 20 21:15:04.316136 kubelet[2596]: I0320 21:15:04.316074 2596 container_manager_linux.go:300] "Creating device plugin manager" Mar 20 21:15:04.316136 kubelet[2596]: I0320 21:15:04.316102 2596 state_mem.go:36] "Initialized new in-memory state store" Mar 20 21:15:04.316213 kubelet[2596]: I0320 21:15:04.316192 2596 kubelet.go:408] "Attempting to sync node with API server" Mar 20 21:15:04.316213 kubelet[2596]: I0320 21:15:04.316202 2596 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 20 21:15:04.316278 kubelet[2596]: I0320 21:15:04.316235 2596 kubelet.go:314] "Adding apiserver pod source" Mar 20 21:15:04.316278 kubelet[2596]: I0320 21:15:04.316254 2596 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 20 21:15:04.317555 kubelet[2596]: I0320 21:15:04.317534 2596 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v2.0.1" apiVersion="v1" Mar 20 21:15:04.319045 kubelet[2596]: I0320 21:15:04.319014 2596 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Mar 20 21:15:04.323363 kubelet[2596]: I0320 21:15:04.323332 2596 server.go:1269] "Started kubelet" Mar 20 21:15:04.325499 kubelet[2596]: I0320 21:15:04.325073 2596 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 20 21:15:04.325499 kubelet[2596]: I0320 21:15:04.325334 2596 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 20 21:15:04.325499 kubelet[2596]: I0320 21:15:04.325369 2596 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 20 21:15:04.325499 kubelet[2596]: I0320 21:15:04.325423 2596 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Mar 20 21:15:04.327281 kubelet[2596]: I0320 21:15:04.326792 2596 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Mar 20 21:15:04.327281 kubelet[2596]: I0320 21:15:04.327021 2596 volume_manager.go:289] "Starting Kubelet Volume Manager" Mar 20 21:15:04.327281 kubelet[2596]: E0320 21:15:04.327119 2596 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 20 21:15:04.328066 kubelet[2596]: I0320 21:15:04.328036 2596 server.go:460] "Adding debug handlers to kubelet server" Mar 20 21:15:04.328294 kubelet[2596]: I0320 21:15:04.328277 2596 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 20 21:15:04.328490 kubelet[2596]: I0320 21:15:04.328478 2596 reconciler.go:26] "Reconciler: start to sync state" Mar 20 21:15:04.340214 kubelet[2596]: I0320 21:15:04.339467 2596 factory.go:221] Registration of the systemd container factory successfully Mar 20 21:15:04.340214 kubelet[2596]: I0320 21:15:04.339566 2596 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Mar 20 21:15:04.340642 kubelet[2596]: I0320 21:15:04.340326 2596 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Mar 20 21:15:04.343452 kubelet[2596]: I0320 21:15:04.343420 2596 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Mar 20 21:15:04.343452 kubelet[2596]: I0320 21:15:04.343452 2596 status_manager.go:217] "Starting to sync pod status with apiserver" Mar 20 21:15:04.343555 kubelet[2596]: I0320 21:15:04.343467 2596 kubelet.go:2321] "Starting kubelet main sync loop" Mar 20 21:15:04.344548 kubelet[2596]: E0320 21:15:04.344395 2596 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 20 21:15:04.344758 kubelet[2596]: I0320 21:15:04.344737 2596 factory.go:221] Registration of the containerd container factory successfully Mar 20 21:15:04.349096 kubelet[2596]: E0320 21:15:04.349068 2596 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Mar 20 21:15:04.373527 kubelet[2596]: I0320 21:15:04.373499 2596 cpu_manager.go:214] "Starting CPU manager" policy="none" Mar 20 21:15:04.373527 kubelet[2596]: I0320 21:15:04.373521 2596 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Mar 20 21:15:04.373662 kubelet[2596]: I0320 21:15:04.373542 2596 state_mem.go:36] "Initialized new in-memory state store" Mar 20 21:15:04.373706 kubelet[2596]: I0320 21:15:04.373688 2596 state_mem.go:88] "Updated default CPUSet" cpuSet="" Mar 20 21:15:04.373733 kubelet[2596]: I0320 21:15:04.373705 2596 state_mem.go:96] "Updated CPUSet assignments" assignments={} Mar 20 21:15:04.373733 kubelet[2596]: I0320 21:15:04.373722 2596 policy_none.go:49] "None policy: Start" Mar 20 21:15:04.374446 kubelet[2596]: I0320 21:15:04.374431 2596 memory_manager.go:170] "Starting memorymanager" policy="None" Mar 20 21:15:04.374487 kubelet[2596]: I0320 21:15:04.374454 2596 state_mem.go:35] "Initializing new in-memory state store" Mar 20 21:15:04.374633 kubelet[2596]: I0320 21:15:04.374617 2596 state_mem.go:75] "Updated machine memory state" Mar 20 21:15:04.378327 kubelet[2596]: I0320 21:15:04.378304 2596 manager.go:510] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Mar 20 21:15:04.378473 kubelet[2596]: I0320 21:15:04.378455 2596 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 20 21:15:04.378507 kubelet[2596]: I0320 21:15:04.378471 2596 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 20 21:15:04.378657 kubelet[2596]: I0320 21:15:04.378650 2596 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 20 21:15:04.480855 kubelet[2596]: I0320 21:15:04.480748 2596 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Mar 20 21:15:04.487377 kubelet[2596]: I0320 21:15:04.487332 2596 kubelet_node_status.go:111] "Node was previously registered" node="localhost" Mar 20 21:15:04.487481 kubelet[2596]: I0320 21:15:04.487457 2596 kubelet_node_status.go:75] "Successfully registered node" node="localhost" Mar 20 21:15:04.529747 kubelet[2596]: I0320 21:15:04.529706 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/60762308083b5ef6c837b1be48ec53d6-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"60762308083b5ef6c837b1be48ec53d6\") " pod="kube-system/kube-controller-manager-localhost" Mar 20 21:15:04.529871 kubelet[2596]: I0320 21:15:04.529751 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/60762308083b5ef6c837b1be48ec53d6-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"60762308083b5ef6c837b1be48ec53d6\") " pod="kube-system/kube-controller-manager-localhost" Mar 20 21:15:04.529871 kubelet[2596]: I0320 21:15:04.529799 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/2260474bc3e2e353aadeddb8b8d59f55-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"2260474bc3e2e353aadeddb8b8d59f55\") " pod="kube-system/kube-apiserver-localhost" Mar 20 21:15:04.529871 kubelet[2596]: I0320 21:15:04.529832 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/2260474bc3e2e353aadeddb8b8d59f55-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"2260474bc3e2e353aadeddb8b8d59f55\") " pod="kube-system/kube-apiserver-localhost" Mar 20 21:15:04.529871 kubelet[2596]: I0320 21:15:04.529860 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/60762308083b5ef6c837b1be48ec53d6-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"60762308083b5ef6c837b1be48ec53d6\") " pod="kube-system/kube-controller-manager-localhost" Mar 20 21:15:04.529964 kubelet[2596]: I0320 21:15:04.529891 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/60762308083b5ef6c837b1be48ec53d6-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"60762308083b5ef6c837b1be48ec53d6\") " pod="kube-system/kube-controller-manager-localhost" Mar 20 21:15:04.529964 kubelet[2596]: I0320 21:15:04.529937 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/60762308083b5ef6c837b1be48ec53d6-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"60762308083b5ef6c837b1be48ec53d6\") " pod="kube-system/kube-controller-manager-localhost" Mar 20 21:15:04.529964 kubelet[2596]: I0320 21:15:04.529956 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/6f32907a07e55aea05abdc5cd284a8d5-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"6f32907a07e55aea05abdc5cd284a8d5\") " pod="kube-system/kube-scheduler-localhost" Mar 20 21:15:04.530031 kubelet[2596]: I0320 21:15:04.529970 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/2260474bc3e2e353aadeddb8b8d59f55-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"2260474bc3e2e353aadeddb8b8d59f55\") " pod="kube-system/kube-apiserver-localhost" Mar 20 21:15:05.317119 kubelet[2596]: I0320 21:15:05.316891 2596 apiserver.go:52] "Watching apiserver" Mar 20 21:15:05.329428 kubelet[2596]: I0320 21:15:05.329381 2596 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 20 21:15:05.363902 kubelet[2596]: E0320 21:15:05.363856 2596 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Mar 20 21:15:05.365411 kubelet[2596]: I0320 21:15:05.365244 2596 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=1.365213789 podStartE2EDuration="1.365213789s" podCreationTimestamp="2025-03-20 21:15:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-20 21:15:05.357663624 +0000 UTC m=+1.089147719" watchObservedRunningTime="2025-03-20 21:15:05.365213789 +0000 UTC m=+1.096697884" Mar 20 21:15:05.373320 kubelet[2596]: I0320 21:15:05.373235 2596 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.373210784 podStartE2EDuration="1.373210784s" podCreationTimestamp="2025-03-20 21:15:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-20 21:15:05.373138979 +0000 UTC m=+1.104623074" watchObservedRunningTime="2025-03-20 21:15:05.373210784 +0000 UTC m=+1.104694879" Mar 20 21:15:05.373439 kubelet[2596]: I0320 21:15:05.373332 2596 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.3733271120000001 podStartE2EDuration="1.373327112s" podCreationTimestamp="2025-03-20 21:15:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-20 21:15:05.365369959 +0000 UTC m=+1.096854055" watchObservedRunningTime="2025-03-20 21:15:05.373327112 +0000 UTC m=+1.104811207" Mar 20 21:15:08.974208 sudo[1687]: pam_unix(sudo:session): session closed for user root Mar 20 21:15:08.975352 sshd[1686]: Connection closed by 10.0.0.1 port 53078 Mar 20 21:15:08.975882 sshd-session[1683]: pam_unix(sshd:session): session closed for user core Mar 20 21:15:08.979372 systemd[1]: sshd@6-10.0.0.54:22-10.0.0.1:53078.service: Deactivated successfully. Mar 20 21:15:08.981310 systemd[1]: session-7.scope: Deactivated successfully. Mar 20 21:15:08.981510 systemd[1]: session-7.scope: Consumed 6.067s CPU time, 221.3M memory peak. Mar 20 21:15:08.982485 systemd-logind[1470]: Session 7 logged out. Waiting for processes to exit. Mar 20 21:15:08.983302 systemd-logind[1470]: Removed session 7. Mar 20 21:15:09.925105 kubelet[2596]: I0320 21:15:09.925075 2596 kuberuntime_manager.go:1633] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Mar 20 21:15:09.925763 containerd[1485]: time="2025-03-20T21:15:09.925675368Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Mar 20 21:15:09.926033 kubelet[2596]: I0320 21:15:09.925886 2596 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Mar 20 21:15:10.787041 systemd[1]: Created slice kubepods-besteffort-pode2f53a88_062e_45eb_aa35_a3b9ad610268.slice - libcontainer container kubepods-besteffort-pode2f53a88_062e_45eb_aa35_a3b9ad610268.slice. Mar 20 21:15:10.873698 kubelet[2596]: I0320 21:15:10.873465 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/e2f53a88-062e-45eb-aa35-a3b9ad610268-kube-proxy\") pod \"kube-proxy-bnp4d\" (UID: \"e2f53a88-062e-45eb-aa35-a3b9ad610268\") " pod="kube-system/kube-proxy-bnp4d" Mar 20 21:15:10.873698 kubelet[2596]: I0320 21:15:10.873508 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/e2f53a88-062e-45eb-aa35-a3b9ad610268-xtables-lock\") pod \"kube-proxy-bnp4d\" (UID: \"e2f53a88-062e-45eb-aa35-a3b9ad610268\") " pod="kube-system/kube-proxy-bnp4d" Mar 20 21:15:10.873698 kubelet[2596]: I0320 21:15:10.873524 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e2f53a88-062e-45eb-aa35-a3b9ad610268-lib-modules\") pod \"kube-proxy-bnp4d\" (UID: \"e2f53a88-062e-45eb-aa35-a3b9ad610268\") " pod="kube-system/kube-proxy-bnp4d" Mar 20 21:15:10.873698 kubelet[2596]: I0320 21:15:10.873614 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zt4zt\" (UniqueName: \"kubernetes.io/projected/e2f53a88-062e-45eb-aa35-a3b9ad610268-kube-api-access-zt4zt\") pod \"kube-proxy-bnp4d\" (UID: \"e2f53a88-062e-45eb-aa35-a3b9ad610268\") " pod="kube-system/kube-proxy-bnp4d" Mar 20 21:15:10.898475 kubelet[2596]: W0320 21:15:10.898398 2596 reflector.go:561] object-"tigera-operator"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:localhost" cannot list resource "configmaps" in API group "" in the namespace "tigera-operator": no relationship found between node 'localhost' and this object Mar 20 21:15:10.898475 kubelet[2596]: E0320 21:15:10.898444 2596 reflector.go:158] "Unhandled Error" err="object-\"tigera-operator\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:localhost\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"tigera-operator\": no relationship found between node 'localhost' and this object" logger="UnhandledError" Mar 20 21:15:10.898619 kubelet[2596]: W0320 21:15:10.898494 2596 reflector.go:561] object-"tigera-operator"/"kubernetes-services-endpoint": failed to list *v1.ConfigMap: configmaps "kubernetes-services-endpoint" is forbidden: User "system:node:localhost" cannot list resource "configmaps" in API group "" in the namespace "tigera-operator": no relationship found between node 'localhost' and this object Mar 20 21:15:10.898619 kubelet[2596]: E0320 21:15:10.898523 2596 reflector.go:158] "Unhandled Error" err="object-\"tigera-operator\"/\"kubernetes-services-endpoint\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kubernetes-services-endpoint\" is forbidden: User \"system:node:localhost\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"tigera-operator\": no relationship found between node 'localhost' and this object" logger="UnhandledError" Mar 20 21:15:10.903245 systemd[1]: Created slice kubepods-besteffort-podc7bf5851_d5f9_42d3_a10c_fd0c6f514196.slice - libcontainer container kubepods-besteffort-podc7bf5851_d5f9_42d3_a10c_fd0c6f514196.slice. Mar 20 21:15:10.974598 kubelet[2596]: I0320 21:15:10.974557 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5czl\" (UniqueName: \"kubernetes.io/projected/c7bf5851-d5f9-42d3-a10c-fd0c6f514196-kube-api-access-z5czl\") pod \"tigera-operator-64ff5465b7-zfsvt\" (UID: \"c7bf5851-d5f9-42d3-a10c-fd0c6f514196\") " pod="tigera-operator/tigera-operator-64ff5465b7-zfsvt" Mar 20 21:15:10.975121 kubelet[2596]: I0320 21:15:10.974979 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/c7bf5851-d5f9-42d3-a10c-fd0c6f514196-var-lib-calico\") pod \"tigera-operator-64ff5465b7-zfsvt\" (UID: \"c7bf5851-d5f9-42d3-a10c-fd0c6f514196\") " pod="tigera-operator/tigera-operator-64ff5465b7-zfsvt" Mar 20 21:15:11.098174 containerd[1485]: time="2025-03-20T21:15:11.098068599Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-bnp4d,Uid:e2f53a88-062e-45eb-aa35-a3b9ad610268,Namespace:kube-system,Attempt:0,}" Mar 20 21:15:11.113849 containerd[1485]: time="2025-03-20T21:15:11.113814421Z" level=info msg="connecting to shim df3e91c579a1da06eee105f037b38b0efeb906f97b925fd79a928b9554540222" address="unix:///run/containerd/s/9e987967687be20fd18c9f275061dd934acbca205caac11d6e4e6b7e8f4c70f4" namespace=k8s.io protocol=ttrpc version=3 Mar 20 21:15:11.136548 systemd[1]: Started cri-containerd-df3e91c579a1da06eee105f037b38b0efeb906f97b925fd79a928b9554540222.scope - libcontainer container df3e91c579a1da06eee105f037b38b0efeb906f97b925fd79a928b9554540222. Mar 20 21:15:11.156481 containerd[1485]: time="2025-03-20T21:15:11.156441992Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-bnp4d,Uid:e2f53a88-062e-45eb-aa35-a3b9ad610268,Namespace:kube-system,Attempt:0,} returns sandbox id \"df3e91c579a1da06eee105f037b38b0efeb906f97b925fd79a928b9554540222\"" Mar 20 21:15:11.158899 containerd[1485]: time="2025-03-20T21:15:11.158803664Z" level=info msg="CreateContainer within sandbox \"df3e91c579a1da06eee105f037b38b0efeb906f97b925fd79a928b9554540222\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Mar 20 21:15:11.166250 containerd[1485]: time="2025-03-20T21:15:11.165830555Z" level=info msg="Container 1a460a45179149559dee9a177cd0a67ac841188bbfaeae6ff92d5c36ac2b1c34: CDI devices from CRI Config.CDIDevices: []" Mar 20 21:15:11.179680 containerd[1485]: time="2025-03-20T21:15:11.179638767Z" level=info msg="CreateContainer within sandbox \"df3e91c579a1da06eee105f037b38b0efeb906f97b925fd79a928b9554540222\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"1a460a45179149559dee9a177cd0a67ac841188bbfaeae6ff92d5c36ac2b1c34\"" Mar 20 21:15:11.180228 containerd[1485]: time="2025-03-20T21:15:11.180196113Z" level=info msg="StartContainer for \"1a460a45179149559dee9a177cd0a67ac841188bbfaeae6ff92d5c36ac2b1c34\"" Mar 20 21:15:11.181551 containerd[1485]: time="2025-03-20T21:15:11.181520776Z" level=info msg="connecting to shim 1a460a45179149559dee9a177cd0a67ac841188bbfaeae6ff92d5c36ac2b1c34" address="unix:///run/containerd/s/9e987967687be20fd18c9f275061dd934acbca205caac11d6e4e6b7e8f4c70f4" protocol=ttrpc version=3 Mar 20 21:15:11.201380 systemd[1]: Started cri-containerd-1a460a45179149559dee9a177cd0a67ac841188bbfaeae6ff92d5c36ac2b1c34.scope - libcontainer container 1a460a45179149559dee9a177cd0a67ac841188bbfaeae6ff92d5c36ac2b1c34. Mar 20 21:15:11.230744 containerd[1485]: time="2025-03-20T21:15:11.230713056Z" level=info msg="StartContainer for \"1a460a45179149559dee9a177cd0a67ac841188bbfaeae6ff92d5c36ac2b1c34\" returns successfully" Mar 20 21:15:12.083104 kubelet[2596]: E0320 21:15:12.083055 2596 projected.go:288] Couldn't get configMap tigera-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 20 21:15:12.083104 kubelet[2596]: E0320 21:15:12.083090 2596 projected.go:194] Error preparing data for projected volume kube-api-access-z5czl for pod tigera-operator/tigera-operator-64ff5465b7-zfsvt: failed to sync configmap cache: timed out waiting for the condition Mar 20 21:15:12.083496 kubelet[2596]: E0320 21:15:12.083144 2596 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c7bf5851-d5f9-42d3-a10c-fd0c6f514196-kube-api-access-z5czl podName:c7bf5851-d5f9-42d3-a10c-fd0c6f514196 nodeName:}" failed. No retries permitted until 2025-03-20 21:15:12.583125151 +0000 UTC m=+8.314609246 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-z5czl" (UniqueName: "kubernetes.io/projected/c7bf5851-d5f9-42d3-a10c-fd0c6f514196-kube-api-access-z5czl") pod "tigera-operator-64ff5465b7-zfsvt" (UID: "c7bf5851-d5f9-42d3-a10c-fd0c6f514196") : failed to sync configmap cache: timed out waiting for the condition Mar 20 21:15:12.706157 containerd[1485]: time="2025-03-20T21:15:12.706101945Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-64ff5465b7-zfsvt,Uid:c7bf5851-d5f9-42d3-a10c-fd0c6f514196,Namespace:tigera-operator,Attempt:0,}" Mar 20 21:15:12.720151 containerd[1485]: time="2025-03-20T21:15:12.720083523Z" level=info msg="connecting to shim f453c31569beedde97e494dc086be6eeb7aa67a59f8477aa2d4c59219d2c5e65" address="unix:///run/containerd/s/35f79ba93ef6e2a4ca0a51ec1d4653d05ca4e8ec2e06ef58026b035765c851ea" namespace=k8s.io protocol=ttrpc version=3 Mar 20 21:15:12.741459 systemd[1]: Started cri-containerd-f453c31569beedde97e494dc086be6eeb7aa67a59f8477aa2d4c59219d2c5e65.scope - libcontainer container f453c31569beedde97e494dc086be6eeb7aa67a59f8477aa2d4c59219d2c5e65. Mar 20 21:15:12.770125 containerd[1485]: time="2025-03-20T21:15:12.770073974Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-64ff5465b7-zfsvt,Uid:c7bf5851-d5f9-42d3-a10c-fd0c6f514196,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"f453c31569beedde97e494dc086be6eeb7aa67a59f8477aa2d4c59219d2c5e65\"" Mar 20 21:15:12.774756 containerd[1485]: time="2025-03-20T21:15:12.774537812Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.36.5\"" Mar 20 21:15:14.146239 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount269403705.mount: Deactivated successfully. Mar 20 21:15:15.195022 containerd[1485]: time="2025-03-20T21:15:15.194655835Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.36.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:15:15.195432 containerd[1485]: time="2025-03-20T21:15:15.195030889Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.36.5: active requests=0, bytes read=19271115" Mar 20 21:15:15.196036 containerd[1485]: time="2025-03-20T21:15:15.195983924Z" level=info msg="ImageCreate event name:\"sha256:a709184cc04589116e7266cb3575491ae8f2ac1c959975fea966447025f66eaa\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:15:15.197827 containerd[1485]: time="2025-03-20T21:15:15.197796830Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:3341fa9475c0325b86228c8726389f9bae9fd6c430c66fe5cd5dc39d7bb6ad4b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:15:15.198640 containerd[1485]: time="2025-03-20T21:15:15.198618420Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.36.5\" with image id \"sha256:a709184cc04589116e7266cb3575491ae8f2ac1c959975fea966447025f66eaa\", repo tag \"quay.io/tigera/operator:v1.36.5\", repo digest \"quay.io/tigera/operator@sha256:3341fa9475c0325b86228c8726389f9bae9fd6c430c66fe5cd5dc39d7bb6ad4b\", size \"19267110\" in 2.424043886s" Mar 20 21:15:15.198693 containerd[1485]: time="2025-03-20T21:15:15.198646781Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.36.5\" returns image reference \"sha256:a709184cc04589116e7266cb3575491ae8f2ac1c959975fea966447025f66eaa\"" Mar 20 21:15:15.202832 containerd[1485]: time="2025-03-20T21:15:15.202799812Z" level=info msg="CreateContainer within sandbox \"f453c31569beedde97e494dc086be6eeb7aa67a59f8477aa2d4c59219d2c5e65\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Mar 20 21:15:15.208210 containerd[1485]: time="2025-03-20T21:15:15.208173008Z" level=info msg="Container d23590733d5acc7c1adcc7593768f6eecd78985395d64a69d9264e684e357bf1: CDI devices from CRI Config.CDIDevices: []" Mar 20 21:15:15.213209 containerd[1485]: time="2025-03-20T21:15:15.213166910Z" level=info msg="CreateContainer within sandbox \"f453c31569beedde97e494dc086be6eeb7aa67a59f8477aa2d4c59219d2c5e65\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"d23590733d5acc7c1adcc7593768f6eecd78985395d64a69d9264e684e357bf1\"" Mar 20 21:15:15.213927 containerd[1485]: time="2025-03-20T21:15:15.213673328Z" level=info msg="StartContainer for \"d23590733d5acc7c1adcc7593768f6eecd78985395d64a69d9264e684e357bf1\"" Mar 20 21:15:15.214804 containerd[1485]: time="2025-03-20T21:15:15.214777969Z" level=info msg="connecting to shim d23590733d5acc7c1adcc7593768f6eecd78985395d64a69d9264e684e357bf1" address="unix:///run/containerd/s/35f79ba93ef6e2a4ca0a51ec1d4653d05ca4e8ec2e06ef58026b035765c851ea" protocol=ttrpc version=3 Mar 20 21:15:15.249408 systemd[1]: Started cri-containerd-d23590733d5acc7c1adcc7593768f6eecd78985395d64a69d9264e684e357bf1.scope - libcontainer container d23590733d5acc7c1adcc7593768f6eecd78985395d64a69d9264e684e357bf1. Mar 20 21:15:15.286968 containerd[1485]: time="2025-03-20T21:15:15.286930118Z" level=info msg="StartContainer for \"d23590733d5acc7c1adcc7593768f6eecd78985395d64a69d9264e684e357bf1\" returns successfully" Mar 20 21:15:15.359369 kubelet[2596]: I0320 21:15:15.359164 2596 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-bnp4d" podStartSLOduration=5.35914823 podStartE2EDuration="5.35914823s" podCreationTimestamp="2025-03-20 21:15:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-20 21:15:11.392325841 +0000 UTC m=+7.123809936" watchObservedRunningTime="2025-03-20 21:15:15.35914823 +0000 UTC m=+11.090632325" Mar 20 21:15:16.606669 update_engine[1475]: I20250320 21:15:16.606600 1475 update_attempter.cc:509] Updating boot flags... Mar 20 21:15:16.632372 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 38 scanned by (udev-worker) (2990) Mar 20 21:15:16.675497 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 38 scanned by (udev-worker) (2991) Mar 20 21:15:16.701294 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 38 scanned by (udev-worker) (2991) Mar 20 21:15:19.561364 kubelet[2596]: I0320 21:15:19.561272 2596 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-64ff5465b7-zfsvt" podStartSLOduration=7.130683626 podStartE2EDuration="9.561256057s" podCreationTimestamp="2025-03-20 21:15:10 +0000 UTC" firstStartedPulling="2025-03-20 21:15:12.771051498 +0000 UTC m=+8.502535593" lastFinishedPulling="2025-03-20 21:15:15.201623929 +0000 UTC m=+10.933108024" observedRunningTime="2025-03-20 21:15:15.396074816 +0000 UTC m=+11.127558871" watchObservedRunningTime="2025-03-20 21:15:19.561256057 +0000 UTC m=+15.292740152" Mar 20 21:15:19.571919 systemd[1]: Created slice kubepods-besteffort-pod56c99114_e719_44a0_a9da_5baa4b1740ed.slice - libcontainer container kubepods-besteffort-pod56c99114_e719_44a0_a9da_5baa4b1740ed.slice. Mar 20 21:15:19.630496 kubelet[2596]: I0320 21:15:19.630393 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/56c99114-e719-44a0-a9da-5baa4b1740ed-tigera-ca-bundle\") pod \"calico-typha-85bb97-2rdv6\" (UID: \"56c99114-e719-44a0-a9da-5baa4b1740ed\") " pod="calico-system/calico-typha-85bb97-2rdv6" Mar 20 21:15:19.630496 kubelet[2596]: I0320 21:15:19.630445 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rwq8\" (UniqueName: \"kubernetes.io/projected/56c99114-e719-44a0-a9da-5baa4b1740ed-kube-api-access-9rwq8\") pod \"calico-typha-85bb97-2rdv6\" (UID: \"56c99114-e719-44a0-a9da-5baa4b1740ed\") " pod="calico-system/calico-typha-85bb97-2rdv6" Mar 20 21:15:19.630496 kubelet[2596]: I0320 21:15:19.630467 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/56c99114-e719-44a0-a9da-5baa4b1740ed-typha-certs\") pod \"calico-typha-85bb97-2rdv6\" (UID: \"56c99114-e719-44a0-a9da-5baa4b1740ed\") " pod="calico-system/calico-typha-85bb97-2rdv6" Mar 20 21:15:19.757000 systemd[1]: Created slice kubepods-besteffort-podefdc3f69_b8c2_414c_97d9_7a8d451b0630.slice - libcontainer container kubepods-besteffort-podefdc3f69_b8c2_414c_97d9_7a8d451b0630.slice. Mar 20 21:15:19.832851 kubelet[2596]: I0320 21:15:19.832512 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/efdc3f69-b8c2-414c-97d9-7a8d451b0630-policysync\") pod \"calico-node-dpxdp\" (UID: \"efdc3f69-b8c2-414c-97d9-7a8d451b0630\") " pod="calico-system/calico-node-dpxdp" Mar 20 21:15:19.832851 kubelet[2596]: I0320 21:15:19.832552 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/efdc3f69-b8c2-414c-97d9-7a8d451b0630-tigera-ca-bundle\") pod \"calico-node-dpxdp\" (UID: \"efdc3f69-b8c2-414c-97d9-7a8d451b0630\") " pod="calico-system/calico-node-dpxdp" Mar 20 21:15:19.832851 kubelet[2596]: I0320 21:15:19.832571 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/efdc3f69-b8c2-414c-97d9-7a8d451b0630-node-certs\") pod \"calico-node-dpxdp\" (UID: \"efdc3f69-b8c2-414c-97d9-7a8d451b0630\") " pod="calico-system/calico-node-dpxdp" Mar 20 21:15:19.832851 kubelet[2596]: I0320 21:15:19.832587 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/efdc3f69-b8c2-414c-97d9-7a8d451b0630-xtables-lock\") pod \"calico-node-dpxdp\" (UID: \"efdc3f69-b8c2-414c-97d9-7a8d451b0630\") " pod="calico-system/calico-node-dpxdp" Mar 20 21:15:19.832851 kubelet[2596]: I0320 21:15:19.832602 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/efdc3f69-b8c2-414c-97d9-7a8d451b0630-var-lib-calico\") pod \"calico-node-dpxdp\" (UID: \"efdc3f69-b8c2-414c-97d9-7a8d451b0630\") " pod="calico-system/calico-node-dpxdp" Mar 20 21:15:19.833037 kubelet[2596]: I0320 21:15:19.832617 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/efdc3f69-b8c2-414c-97d9-7a8d451b0630-cni-bin-dir\") pod \"calico-node-dpxdp\" (UID: \"efdc3f69-b8c2-414c-97d9-7a8d451b0630\") " pod="calico-system/calico-node-dpxdp" Mar 20 21:15:19.833037 kubelet[2596]: I0320 21:15:19.832632 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/efdc3f69-b8c2-414c-97d9-7a8d451b0630-flexvol-driver-host\") pod \"calico-node-dpxdp\" (UID: \"efdc3f69-b8c2-414c-97d9-7a8d451b0630\") " pod="calico-system/calico-node-dpxdp" Mar 20 21:15:19.833037 kubelet[2596]: I0320 21:15:19.832647 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gjxn\" (UniqueName: \"kubernetes.io/projected/efdc3f69-b8c2-414c-97d9-7a8d451b0630-kube-api-access-5gjxn\") pod \"calico-node-dpxdp\" (UID: \"efdc3f69-b8c2-414c-97d9-7a8d451b0630\") " pod="calico-system/calico-node-dpxdp" Mar 20 21:15:19.833037 kubelet[2596]: I0320 21:15:19.832664 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/efdc3f69-b8c2-414c-97d9-7a8d451b0630-var-run-calico\") pod \"calico-node-dpxdp\" (UID: \"efdc3f69-b8c2-414c-97d9-7a8d451b0630\") " pod="calico-system/calico-node-dpxdp" Mar 20 21:15:19.833037 kubelet[2596]: I0320 21:15:19.832678 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/efdc3f69-b8c2-414c-97d9-7a8d451b0630-cni-log-dir\") pod \"calico-node-dpxdp\" (UID: \"efdc3f69-b8c2-414c-97d9-7a8d451b0630\") " pod="calico-system/calico-node-dpxdp" Mar 20 21:15:19.833143 kubelet[2596]: I0320 21:15:19.832693 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/efdc3f69-b8c2-414c-97d9-7a8d451b0630-lib-modules\") pod \"calico-node-dpxdp\" (UID: \"efdc3f69-b8c2-414c-97d9-7a8d451b0630\") " pod="calico-system/calico-node-dpxdp" Mar 20 21:15:19.833143 kubelet[2596]: I0320 21:15:19.832707 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/efdc3f69-b8c2-414c-97d9-7a8d451b0630-cni-net-dir\") pod \"calico-node-dpxdp\" (UID: \"efdc3f69-b8c2-414c-97d9-7a8d451b0630\") " pod="calico-system/calico-node-dpxdp" Mar 20 21:15:19.876137 containerd[1485]: time="2025-03-20T21:15:19.876093721Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-85bb97-2rdv6,Uid:56c99114-e719-44a0-a9da-5baa4b1740ed,Namespace:calico-system,Attempt:0,}" Mar 20 21:15:19.926772 containerd[1485]: time="2025-03-20T21:15:19.926719026Z" level=info msg="connecting to shim 6689fc16b6f9db21fc1f1cf9e5fe8313930d7154ca624118ca770e06c974ad09" address="unix:///run/containerd/s/a0c1b064d1d82f8877bd97ed45e902abc8f69aa6afca199d71211d499d806c73" namespace=k8s.io protocol=ttrpc version=3 Mar 20 21:15:19.942392 kubelet[2596]: E0320 21:15:19.942359 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:19.942392 kubelet[2596]: W0320 21:15:19.942384 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:19.942537 kubelet[2596]: E0320 21:15:19.942494 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:19.952969 kubelet[2596]: E0320 21:15:19.952464 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:19.952969 kubelet[2596]: W0320 21:15:19.952486 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:19.952969 kubelet[2596]: E0320 21:15:19.952505 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:19.963771 kubelet[2596]: E0320 21:15:19.961980 2596 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-7htv8" podUID="4e0ac6d0-32f4-4ad3-a9b3-9da162162a58" Mar 20 21:15:19.991405 systemd[1]: Started cri-containerd-6689fc16b6f9db21fc1f1cf9e5fe8313930d7154ca624118ca770e06c974ad09.scope - libcontainer container 6689fc16b6f9db21fc1f1cf9e5fe8313930d7154ca624118ca770e06c974ad09. Mar 20 21:15:20.024595 containerd[1485]: time="2025-03-20T21:15:20.024533657Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-85bb97-2rdv6,Uid:56c99114-e719-44a0-a9da-5baa4b1740ed,Namespace:calico-system,Attempt:0,} returns sandbox id \"6689fc16b6f9db21fc1f1cf9e5fe8313930d7154ca624118ca770e06c974ad09\"" Mar 20 21:15:20.031944 kubelet[2596]: E0320 21:15:20.031733 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.031944 kubelet[2596]: W0320 21:15:20.031752 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.031944 kubelet[2596]: E0320 21:15:20.031772 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.032303 kubelet[2596]: E0320 21:15:20.032279 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.032303 kubelet[2596]: W0320 21:15:20.032294 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.032303 kubelet[2596]: E0320 21:15:20.032305 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.033638 kubelet[2596]: E0320 21:15:20.033408 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.033638 kubelet[2596]: W0320 21:15:20.033434 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.033638 kubelet[2596]: E0320 21:15:20.033454 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.033815 kubelet[2596]: E0320 21:15:20.033729 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.033815 kubelet[2596]: W0320 21:15:20.033740 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.033815 kubelet[2596]: E0320 21:15:20.033751 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.034562 kubelet[2596]: E0320 21:15:20.034442 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.034562 kubelet[2596]: W0320 21:15:20.034459 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.034562 kubelet[2596]: E0320 21:15:20.034473 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.034562 kubelet[2596]: E0320 21:15:20.035267 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.034562 kubelet[2596]: W0320 21:15:20.035276 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.034562 kubelet[2596]: E0320 21:15:20.035287 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.034562 kubelet[2596]: E0320 21:15:20.035459 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.034562 kubelet[2596]: W0320 21:15:20.035467 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.034562 kubelet[2596]: E0320 21:15:20.035475 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.034562 kubelet[2596]: E0320 21:15:20.035648 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.037560 kubelet[2596]: W0320 21:15:20.035656 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.037560 kubelet[2596]: E0320 21:15:20.035665 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.037560 kubelet[2596]: E0320 21:15:20.037384 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.037560 kubelet[2596]: W0320 21:15:20.037410 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.037560 kubelet[2596]: E0320 21:15:20.037423 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.038436 kubelet[2596]: E0320 21:15:20.038334 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.038436 kubelet[2596]: W0320 21:15:20.038353 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.038436 kubelet[2596]: E0320 21:15:20.038367 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.039519 kubelet[2596]: E0320 21:15:20.039327 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.039519 kubelet[2596]: W0320 21:15:20.039348 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.039519 kubelet[2596]: E0320 21:15:20.039361 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.039689 kubelet[2596]: E0320 21:15:20.039573 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.039689 kubelet[2596]: W0320 21:15:20.039581 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.039689 kubelet[2596]: E0320 21:15:20.039590 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.042415 kubelet[2596]: E0320 21:15:20.040454 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.042415 kubelet[2596]: W0320 21:15:20.040472 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.042415 kubelet[2596]: E0320 21:15:20.040486 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.042593 kubelet[2596]: E0320 21:15:20.042510 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.042593 kubelet[2596]: W0320 21:15:20.042530 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.042593 kubelet[2596]: E0320 21:15:20.042544 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.042680 containerd[1485]: time="2025-03-20T21:15:20.042646735Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.29.2\"" Mar 20 21:15:20.043750 kubelet[2596]: E0320 21:15:20.043088 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.043750 kubelet[2596]: W0320 21:15:20.043098 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.043750 kubelet[2596]: E0320 21:15:20.043110 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.043750 kubelet[2596]: E0320 21:15:20.043582 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.043750 kubelet[2596]: W0320 21:15:20.043594 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.043750 kubelet[2596]: E0320 21:15:20.043605 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.044767 kubelet[2596]: E0320 21:15:20.044641 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.044767 kubelet[2596]: W0320 21:15:20.044664 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.044767 kubelet[2596]: E0320 21:15:20.044678 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.045074 kubelet[2596]: E0320 21:15:20.045044 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.045412 kubelet[2596]: W0320 21:15:20.045291 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.045412 kubelet[2596]: E0320 21:15:20.045313 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.046704 kubelet[2596]: E0320 21:15:20.046688 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.046774 kubelet[2596]: W0320 21:15:20.046761 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.046825 kubelet[2596]: E0320 21:15:20.046814 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.047067 kubelet[2596]: E0320 21:15:20.047054 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.047336 kubelet[2596]: W0320 21:15:20.047129 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.047336 kubelet[2596]: E0320 21:15:20.047146 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.047487 kubelet[2596]: E0320 21:15:20.047474 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.047547 kubelet[2596]: W0320 21:15:20.047535 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.047627 kubelet[2596]: E0320 21:15:20.047616 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.048172 kubelet[2596]: I0320 21:15:20.048153 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/4e0ac6d0-32f4-4ad3-a9b3-9da162162a58-socket-dir\") pod \"csi-node-driver-7htv8\" (UID: \"4e0ac6d0-32f4-4ad3-a9b3-9da162162a58\") " pod="calico-system/csi-node-driver-7htv8" Mar 20 21:15:20.048572 kubelet[2596]: E0320 21:15:20.048537 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.048572 kubelet[2596]: W0320 21:15:20.048550 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.048572 kubelet[2596]: E0320 21:15:20.048565 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.049162 kubelet[2596]: E0320 21:15:20.049101 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.049162 kubelet[2596]: W0320 21:15:20.049121 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.049162 kubelet[2596]: E0320 21:15:20.049138 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.049522 kubelet[2596]: E0320 21:15:20.049351 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.049522 kubelet[2596]: W0320 21:15:20.049360 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.049522 kubelet[2596]: E0320 21:15:20.049369 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.049522 kubelet[2596]: I0320 21:15:20.049392 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4e0ac6d0-32f4-4ad3-a9b3-9da162162a58-kubelet-dir\") pod \"csi-node-driver-7htv8\" (UID: \"4e0ac6d0-32f4-4ad3-a9b3-9da162162a58\") " pod="calico-system/csi-node-driver-7htv8" Mar 20 21:15:20.049882 kubelet[2596]: E0320 21:15:20.049600 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.049882 kubelet[2596]: W0320 21:15:20.049613 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.049882 kubelet[2596]: E0320 21:15:20.049623 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.049882 kubelet[2596]: I0320 21:15:20.049637 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/4e0ac6d0-32f4-4ad3-a9b3-9da162162a58-registration-dir\") pod \"csi-node-driver-7htv8\" (UID: \"4e0ac6d0-32f4-4ad3-a9b3-9da162162a58\") " pod="calico-system/csi-node-driver-7htv8" Mar 20 21:15:20.052188 kubelet[2596]: E0320 21:15:20.052054 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.052188 kubelet[2596]: W0320 21:15:20.052071 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.052188 kubelet[2596]: E0320 21:15:20.052091 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.052595 kubelet[2596]: E0320 21:15:20.052548 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.052830 kubelet[2596]: W0320 21:15:20.052731 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.052986 kubelet[2596]: E0320 21:15:20.052949 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.053189 kubelet[2596]: E0320 21:15:20.053174 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.053343 kubelet[2596]: W0320 21:15:20.053259 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.053488 kubelet[2596]: E0320 21:15:20.053463 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.053488 kubelet[2596]: E0320 21:15:20.053471 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.053541 kubelet[2596]: I0320 21:15:20.053498 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rps7\" (UniqueName: \"kubernetes.io/projected/4e0ac6d0-32f4-4ad3-a9b3-9da162162a58-kube-api-access-5rps7\") pod \"csi-node-driver-7htv8\" (UID: \"4e0ac6d0-32f4-4ad3-a9b3-9da162162a58\") " pod="calico-system/csi-node-driver-7htv8" Mar 20 21:15:20.053541 kubelet[2596]: W0320 21:15:20.053475 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.053541 kubelet[2596]: E0320 21:15:20.053528 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.053744 kubelet[2596]: E0320 21:15:20.053730 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.053780 kubelet[2596]: W0320 21:15:20.053744 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.053780 kubelet[2596]: E0320 21:15:20.053757 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.054070 kubelet[2596]: E0320 21:15:20.053926 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.054070 kubelet[2596]: W0320 21:15:20.053942 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.054070 kubelet[2596]: E0320 21:15:20.053955 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.054070 kubelet[2596]: I0320 21:15:20.053977 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/4e0ac6d0-32f4-4ad3-a9b3-9da162162a58-varrun\") pod \"csi-node-driver-7htv8\" (UID: \"4e0ac6d0-32f4-4ad3-a9b3-9da162162a58\") " pod="calico-system/csi-node-driver-7htv8" Mar 20 21:15:20.054665 kubelet[2596]: E0320 21:15:20.054510 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.054665 kubelet[2596]: W0320 21:15:20.054527 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.054665 kubelet[2596]: E0320 21:15:20.054551 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.055281 kubelet[2596]: E0320 21:15:20.055123 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.055281 kubelet[2596]: W0320 21:15:20.055137 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.055281 kubelet[2596]: E0320 21:15:20.055235 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.055389 kubelet[2596]: E0320 21:15:20.055349 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.055389 kubelet[2596]: W0320 21:15:20.055357 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.055389 kubelet[2596]: E0320 21:15:20.055369 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.056793 kubelet[2596]: E0320 21:15:20.056305 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.056793 kubelet[2596]: W0320 21:15:20.056327 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.056793 kubelet[2596]: E0320 21:15:20.056340 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.064781 containerd[1485]: time="2025-03-20T21:15:20.064724238Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-dpxdp,Uid:efdc3f69-b8c2-414c-97d9-7a8d451b0630,Namespace:calico-system,Attempt:0,}" Mar 20 21:15:20.088535 containerd[1485]: time="2025-03-20T21:15:20.088384422Z" level=info msg="connecting to shim 54ed274cf11ce43c3701d2bf81a57f8bf3ce02515cfbaa6de5aff9e86c50d2e6" address="unix:///run/containerd/s/6f52bec33ef527195a31a6634958f16dbfd178422935e01a514e23b3d3efee65" namespace=k8s.io protocol=ttrpc version=3 Mar 20 21:15:20.116388 systemd[1]: Started cri-containerd-54ed274cf11ce43c3701d2bf81a57f8bf3ce02515cfbaa6de5aff9e86c50d2e6.scope - libcontainer container 54ed274cf11ce43c3701d2bf81a57f8bf3ce02515cfbaa6de5aff9e86c50d2e6. Mar 20 21:15:20.154665 kubelet[2596]: E0320 21:15:20.154502 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.154665 kubelet[2596]: W0320 21:15:20.154527 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.154665 kubelet[2596]: E0320 21:15:20.154547 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.155186 kubelet[2596]: E0320 21:15:20.155051 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.155186 kubelet[2596]: W0320 21:15:20.155064 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.155186 kubelet[2596]: E0320 21:15:20.155085 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.155384 kubelet[2596]: E0320 21:15:20.155371 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.155466 kubelet[2596]: W0320 21:15:20.155451 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.155535 kubelet[2596]: E0320 21:15:20.155522 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.155729 kubelet[2596]: E0320 21:15:20.155708 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.155729 kubelet[2596]: W0320 21:15:20.155727 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.155807 kubelet[2596]: E0320 21:15:20.155745 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.155927 kubelet[2596]: E0320 21:15:20.155908 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.155927 kubelet[2596]: W0320 21:15:20.155919 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.156003 kubelet[2596]: E0320 21:15:20.155932 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.156075 kubelet[2596]: E0320 21:15:20.156065 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.156107 kubelet[2596]: W0320 21:15:20.156079 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.156107 kubelet[2596]: E0320 21:15:20.156093 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.156343 kubelet[2596]: E0320 21:15:20.156326 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.156343 kubelet[2596]: W0320 21:15:20.156342 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.156437 kubelet[2596]: E0320 21:15:20.156359 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.157060 kubelet[2596]: E0320 21:15:20.157030 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.157060 kubelet[2596]: W0320 21:15:20.157043 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.157060 kubelet[2596]: E0320 21:15:20.157058 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.157447 kubelet[2596]: E0320 21:15:20.157235 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.157447 kubelet[2596]: W0320 21:15:20.157243 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.157447 kubelet[2596]: E0320 21:15:20.157435 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.157447 kubelet[2596]: W0320 21:15:20.157445 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.157636 kubelet[2596]: E0320 21:15:20.157571 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.157636 kubelet[2596]: E0320 21:15:20.157616 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.157636 kubelet[2596]: W0320 21:15:20.157625 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.157636 kubelet[2596]: E0320 21:15:20.157635 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.157780 kubelet[2596]: E0320 21:15:20.157619 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.158061 kubelet[2596]: E0320 21:15:20.158030 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.158061 kubelet[2596]: W0320 21:15:20.158050 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.158061 kubelet[2596]: E0320 21:15:20.158066 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.158275 kubelet[2596]: E0320 21:15:20.158236 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.158275 kubelet[2596]: W0320 21:15:20.158252 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.158275 kubelet[2596]: E0320 21:15:20.158262 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.158487 kubelet[2596]: E0320 21:15:20.158474 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.158487 kubelet[2596]: W0320 21:15:20.158487 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.158807 kubelet[2596]: E0320 21:15:20.158501 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.158807 kubelet[2596]: E0320 21:15:20.158742 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.158807 kubelet[2596]: W0320 21:15:20.158752 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.158807 kubelet[2596]: E0320 21:15:20.158763 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.159311 kubelet[2596]: E0320 21:15:20.158957 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.159311 kubelet[2596]: W0320 21:15:20.158984 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.159311 kubelet[2596]: E0320 21:15:20.158993 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.159311 kubelet[2596]: E0320 21:15:20.159180 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.159311 kubelet[2596]: W0320 21:15:20.159188 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.159311 kubelet[2596]: E0320 21:15:20.159199 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.159697 kubelet[2596]: E0320 21:15:20.159682 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.159697 kubelet[2596]: W0320 21:15:20.159695 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.159842 kubelet[2596]: E0320 21:15:20.159729 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.159881 containerd[1485]: time="2025-03-20T21:15:20.159771546Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-dpxdp,Uid:efdc3f69-b8c2-414c-97d9-7a8d451b0630,Namespace:calico-system,Attempt:0,} returns sandbox id \"54ed274cf11ce43c3701d2bf81a57f8bf3ce02515cfbaa6de5aff9e86c50d2e6\"" Mar 20 21:15:20.159936 kubelet[2596]: E0320 21:15:20.159900 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.159936 kubelet[2596]: W0320 21:15:20.159909 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.160232 kubelet[2596]: E0320 21:15:20.159952 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.160232 kubelet[2596]: E0320 21:15:20.160100 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.160232 kubelet[2596]: W0320 21:15:20.160109 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.160232 kubelet[2596]: E0320 21:15:20.160135 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.160337 kubelet[2596]: E0320 21:15:20.160322 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.160337 kubelet[2596]: W0320 21:15:20.160331 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.160560 kubelet[2596]: E0320 21:15:20.160384 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.160560 kubelet[2596]: E0320 21:15:20.160514 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.160560 kubelet[2596]: W0320 21:15:20.160526 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.160560 kubelet[2596]: E0320 21:15:20.160543 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.160973 kubelet[2596]: E0320 21:15:20.160820 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.160973 kubelet[2596]: W0320 21:15:20.160833 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.160973 kubelet[2596]: E0320 21:15:20.160845 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.161186 kubelet[2596]: E0320 21:15:20.161171 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.161560 kubelet[2596]: W0320 21:15:20.161338 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.161560 kubelet[2596]: E0320 21:15:20.161358 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.161744 kubelet[2596]: E0320 21:15:20.161732 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.161888 kubelet[2596]: W0320 21:15:20.161838 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.162094 kubelet[2596]: E0320 21:15:20.161984 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:20.173708 kubelet[2596]: E0320 21:15:20.173667 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:20.173708 kubelet[2596]: W0320 21:15:20.173702 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:20.173708 kubelet[2596]: E0320 21:15:20.173721 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:21.351583 kubelet[2596]: E0320 21:15:21.351479 2596 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-7htv8" podUID="4e0ac6d0-32f4-4ad3-a9b3-9da162162a58" Mar 20 21:15:21.896614 containerd[1485]: time="2025-03-20T21:15:21.896553307Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.29.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:15:21.897080 containerd[1485]: time="2025-03-20T21:15:21.897036519Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.29.2: active requests=0, bytes read=28363957" Mar 20 21:15:21.897915 containerd[1485]: time="2025-03-20T21:15:21.897885140Z" level=info msg="ImageCreate event name:\"sha256:38a4e8457549414848315eae0d5ab8ecd6c51f4baaea849fe5edce714d81a999\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:15:21.899528 containerd[1485]: time="2025-03-20T21:15:21.899490179Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:9839fd34b4c1bad50beed72aec59c64893487a46eea57dc2d7d66c3041d7bcce\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:15:21.900295 containerd[1485]: time="2025-03-20T21:15:21.900257758Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.29.2\" with image id \"sha256:38a4e8457549414848315eae0d5ab8ecd6c51f4baaea849fe5edce714d81a999\", repo tag \"ghcr.io/flatcar/calico/typha:v3.29.2\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:9839fd34b4c1bad50beed72aec59c64893487a46eea57dc2d7d66c3041d7bcce\", size \"29733706\" in 1.857565662s" Mar 20 21:15:21.900344 containerd[1485]: time="2025-03-20T21:15:21.900302639Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.29.2\" returns image reference \"sha256:38a4e8457549414848315eae0d5ab8ecd6c51f4baaea849fe5edce714d81a999\"" Mar 20 21:15:21.904683 containerd[1485]: time="2025-03-20T21:15:21.904655027Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.2\"" Mar 20 21:15:21.917254 containerd[1485]: time="2025-03-20T21:15:21.917199497Z" level=info msg="CreateContainer within sandbox \"6689fc16b6f9db21fc1f1cf9e5fe8313930d7154ca624118ca770e06c974ad09\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Mar 20 21:15:21.923440 containerd[1485]: time="2025-03-20T21:15:21.923396211Z" level=info msg="Container cbc449a4607b3cecaadbd9ddb27b00d8ebc037a515d0affabb0b4b21de249ecd: CDI devices from CRI Config.CDIDevices: []" Mar 20 21:15:21.929123 containerd[1485]: time="2025-03-20T21:15:21.929058231Z" level=info msg="CreateContainer within sandbox \"6689fc16b6f9db21fc1f1cf9e5fe8313930d7154ca624118ca770e06c974ad09\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"cbc449a4607b3cecaadbd9ddb27b00d8ebc037a515d0affabb0b4b21de249ecd\"" Mar 20 21:15:21.929683 containerd[1485]: time="2025-03-20T21:15:21.929648125Z" level=info msg="StartContainer for \"cbc449a4607b3cecaadbd9ddb27b00d8ebc037a515d0affabb0b4b21de249ecd\"" Mar 20 21:15:21.930853 containerd[1485]: time="2025-03-20T21:15:21.930814794Z" level=info msg="connecting to shim cbc449a4607b3cecaadbd9ddb27b00d8ebc037a515d0affabb0b4b21de249ecd" address="unix:///run/containerd/s/a0c1b064d1d82f8877bd97ed45e902abc8f69aa6afca199d71211d499d806c73" protocol=ttrpc version=3 Mar 20 21:15:21.954369 systemd[1]: Started cri-containerd-cbc449a4607b3cecaadbd9ddb27b00d8ebc037a515d0affabb0b4b21de249ecd.scope - libcontainer container cbc449a4607b3cecaadbd9ddb27b00d8ebc037a515d0affabb0b4b21de249ecd. Mar 20 21:15:21.992249 containerd[1485]: time="2025-03-20T21:15:21.992197233Z" level=info msg="StartContainer for \"cbc449a4607b3cecaadbd9ddb27b00d8ebc037a515d0affabb0b4b21de249ecd\" returns successfully" Mar 20 21:15:22.456705 kubelet[2596]: I0320 21:15:22.456531 2596 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-85bb97-2rdv6" podStartSLOduration=1.584384496 podStartE2EDuration="3.456515736s" podCreationTimestamp="2025-03-20 21:15:19 +0000 UTC" firstStartedPulling="2025-03-20 21:15:20.03183169 +0000 UTC m=+15.763315785" lastFinishedPulling="2025-03-20 21:15:21.90396289 +0000 UTC m=+17.635447025" observedRunningTime="2025-03-20 21:15:22.456096087 +0000 UTC m=+18.187580182" watchObservedRunningTime="2025-03-20 21:15:22.456515736 +0000 UTC m=+18.187999791" Mar 20 21:15:22.462574 kubelet[2596]: E0320 21:15:22.462545 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:22.462574 kubelet[2596]: W0320 21:15:22.462564 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:22.462667 kubelet[2596]: E0320 21:15:22.462581 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:22.462786 kubelet[2596]: E0320 21:15:22.462766 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:22.462786 kubelet[2596]: W0320 21:15:22.462777 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:22.462786 kubelet[2596]: E0320 21:15:22.462786 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:22.462949 kubelet[2596]: E0320 21:15:22.462932 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:22.462949 kubelet[2596]: W0320 21:15:22.462943 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:22.463001 kubelet[2596]: E0320 21:15:22.462952 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:22.463108 kubelet[2596]: E0320 21:15:22.463098 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:22.463132 kubelet[2596]: W0320 21:15:22.463109 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:22.463132 kubelet[2596]: E0320 21:15:22.463117 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:22.463280 kubelet[2596]: E0320 21:15:22.463270 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:22.463308 kubelet[2596]: W0320 21:15:22.463280 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:22.463308 kubelet[2596]: E0320 21:15:22.463288 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:22.463467 kubelet[2596]: E0320 21:15:22.463449 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:22.463467 kubelet[2596]: W0320 21:15:22.463460 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:22.463509 kubelet[2596]: E0320 21:15:22.463467 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:22.463604 kubelet[2596]: E0320 21:15:22.463595 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:22.463629 kubelet[2596]: W0320 21:15:22.463603 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:22.463629 kubelet[2596]: E0320 21:15:22.463610 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:22.463756 kubelet[2596]: E0320 21:15:22.463746 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:22.463781 kubelet[2596]: W0320 21:15:22.463756 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:22.463781 kubelet[2596]: E0320 21:15:22.463763 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:22.463906 kubelet[2596]: E0320 21:15:22.463897 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:22.463930 kubelet[2596]: W0320 21:15:22.463906 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:22.463930 kubelet[2596]: E0320 21:15:22.463913 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:22.464048 kubelet[2596]: E0320 21:15:22.464039 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:22.464075 kubelet[2596]: W0320 21:15:22.464047 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:22.464075 kubelet[2596]: E0320 21:15:22.464058 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:22.464186 kubelet[2596]: E0320 21:15:22.464177 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:22.464209 kubelet[2596]: W0320 21:15:22.464186 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:22.464209 kubelet[2596]: E0320 21:15:22.464193 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:22.464352 kubelet[2596]: E0320 21:15:22.464342 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:22.464383 kubelet[2596]: W0320 21:15:22.464352 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:22.464383 kubelet[2596]: E0320 21:15:22.464360 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:22.464506 kubelet[2596]: E0320 21:15:22.464497 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:22.464529 kubelet[2596]: W0320 21:15:22.464506 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:22.464529 kubelet[2596]: E0320 21:15:22.464515 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:22.464649 kubelet[2596]: E0320 21:15:22.464637 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:22.464670 kubelet[2596]: W0320 21:15:22.464649 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:22.464670 kubelet[2596]: E0320 21:15:22.464658 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:22.464795 kubelet[2596]: E0320 21:15:22.464786 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:22.464909 kubelet[2596]: W0320 21:15:22.464796 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:22.464909 kubelet[2596]: E0320 21:15:22.464803 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:22.475143 kubelet[2596]: E0320 21:15:22.475120 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:22.475143 kubelet[2596]: W0320 21:15:22.475136 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:22.475229 kubelet[2596]: E0320 21:15:22.475150 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:22.475386 kubelet[2596]: E0320 21:15:22.475359 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:22.475386 kubelet[2596]: W0320 21:15:22.475375 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:22.475446 kubelet[2596]: E0320 21:15:22.475391 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:22.475586 kubelet[2596]: E0320 21:15:22.475565 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:22.475586 kubelet[2596]: W0320 21:15:22.475576 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:22.475635 kubelet[2596]: E0320 21:15:22.475589 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:22.475824 kubelet[2596]: E0320 21:15:22.475798 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:22.475824 kubelet[2596]: W0320 21:15:22.475810 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:22.475824 kubelet[2596]: E0320 21:15:22.475823 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:22.476012 kubelet[2596]: E0320 21:15:22.475991 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:22.476012 kubelet[2596]: W0320 21:15:22.476001 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:22.476064 kubelet[2596]: E0320 21:15:22.476013 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:22.476154 kubelet[2596]: E0320 21:15:22.476143 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:22.476154 kubelet[2596]: W0320 21:15:22.476152 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:22.476203 kubelet[2596]: E0320 21:15:22.476163 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:22.476376 kubelet[2596]: E0320 21:15:22.476357 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:22.476376 kubelet[2596]: W0320 21:15:22.476372 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:22.476427 kubelet[2596]: E0320 21:15:22.476386 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:22.476628 kubelet[2596]: E0320 21:15:22.476608 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:22.476628 kubelet[2596]: W0320 21:15:22.476626 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:22.476697 kubelet[2596]: E0320 21:15:22.476644 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:22.476846 kubelet[2596]: E0320 21:15:22.476821 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:22.476846 kubelet[2596]: W0320 21:15:22.476834 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:22.476898 kubelet[2596]: E0320 21:15:22.476848 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:22.477037 kubelet[2596]: E0320 21:15:22.477024 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:22.477063 kubelet[2596]: W0320 21:15:22.477036 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:22.477063 kubelet[2596]: E0320 21:15:22.477051 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:22.477265 kubelet[2596]: E0320 21:15:22.477252 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:22.477265 kubelet[2596]: W0320 21:15:22.477264 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:22.477318 kubelet[2596]: E0320 21:15:22.477277 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:22.477462 kubelet[2596]: E0320 21:15:22.477448 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:22.477462 kubelet[2596]: W0320 21:15:22.477461 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:22.477519 kubelet[2596]: E0320 21:15:22.477475 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:22.477663 kubelet[2596]: E0320 21:15:22.477651 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:22.477663 kubelet[2596]: W0320 21:15:22.477661 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:22.477717 kubelet[2596]: E0320 21:15:22.477673 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:22.477896 kubelet[2596]: E0320 21:15:22.477882 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:22.477921 kubelet[2596]: W0320 21:15:22.477895 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:22.477921 kubelet[2596]: E0320 21:15:22.477906 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:22.478094 kubelet[2596]: E0320 21:15:22.478081 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:22.478124 kubelet[2596]: W0320 21:15:22.478093 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:22.478124 kubelet[2596]: E0320 21:15:22.478108 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:22.478614 kubelet[2596]: E0320 21:15:22.478589 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:22.478614 kubelet[2596]: W0320 21:15:22.478606 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:22.478702 kubelet[2596]: E0320 21:15:22.478634 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:22.478860 kubelet[2596]: E0320 21:15:22.478846 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:22.478860 kubelet[2596]: W0320 21:15:22.478859 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:22.478916 kubelet[2596]: E0320 21:15:22.478870 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:22.479266 kubelet[2596]: E0320 21:15:22.479249 2596 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Mar 20 21:15:22.479266 kubelet[2596]: W0320 21:15:22.479264 2596 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Mar 20 21:15:22.479338 kubelet[2596]: E0320 21:15:22.479274 2596 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Mar 20 21:15:23.275889 containerd[1485]: time="2025-03-20T21:15:23.275840623Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:15:23.276823 containerd[1485]: time="2025-03-20T21:15:23.276678641Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.2: active requests=0, bytes read=5120152" Mar 20 21:15:23.277568 containerd[1485]: time="2025-03-20T21:15:23.277424417Z" level=info msg="ImageCreate event name:\"sha256:bf0e51f0111c4e6f7bc448c15934e73123805f3c5e66e455c7eb7392854e0921\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:15:23.279134 containerd[1485]: time="2025-03-20T21:15:23.279100373Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:51d9341a4a37e278a906f40ecc73f5076e768612c21621f1b1d4f2b2f0735a1d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:15:23.279930 containerd[1485]: time="2025-03-20T21:15:23.279886671Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.2\" with image id \"sha256:bf0e51f0111c4e6f7bc448c15934e73123805f3c5e66e455c7eb7392854e0921\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.2\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:51d9341a4a37e278a906f40ecc73f5076e768612c21621f1b1d4f2b2f0735a1d\", size \"6489869\" in 1.375195962s" Mar 20 21:15:23.279930 containerd[1485]: time="2025-03-20T21:15:23.279921671Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.2\" returns image reference \"sha256:bf0e51f0111c4e6f7bc448c15934e73123805f3c5e66e455c7eb7392854e0921\"" Mar 20 21:15:23.282145 containerd[1485]: time="2025-03-20T21:15:23.282054638Z" level=info msg="CreateContainer within sandbox \"54ed274cf11ce43c3701d2bf81a57f8bf3ce02515cfbaa6de5aff9e86c50d2e6\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Mar 20 21:15:23.291912 containerd[1485]: time="2025-03-20T21:15:23.291873371Z" level=info msg="Container a433c0863a65b06e47e0353cf2a1ab0b9f9a9e43e65e1ae62213764aa79b5564: CDI devices from CRI Config.CDIDevices: []" Mar 20 21:15:23.299957 containerd[1485]: time="2025-03-20T21:15:23.299913746Z" level=info msg="CreateContainer within sandbox \"54ed274cf11ce43c3701d2bf81a57f8bf3ce02515cfbaa6de5aff9e86c50d2e6\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"a433c0863a65b06e47e0353cf2a1ab0b9f9a9e43e65e1ae62213764aa79b5564\"" Mar 20 21:15:23.300434 containerd[1485]: time="2025-03-20T21:15:23.300388356Z" level=info msg="StartContainer for \"a433c0863a65b06e47e0353cf2a1ab0b9f9a9e43e65e1ae62213764aa79b5564\"" Mar 20 21:15:23.301846 containerd[1485]: time="2025-03-20T21:15:23.301750426Z" level=info msg="connecting to shim a433c0863a65b06e47e0353cf2a1ab0b9f9a9e43e65e1ae62213764aa79b5564" address="unix:///run/containerd/s/6f52bec33ef527195a31a6634958f16dbfd178422935e01a514e23b3d3efee65" protocol=ttrpc version=3 Mar 20 21:15:23.323395 systemd[1]: Started cri-containerd-a433c0863a65b06e47e0353cf2a1ab0b9f9a9e43e65e1ae62213764aa79b5564.scope - libcontainer container a433c0863a65b06e47e0353cf2a1ab0b9f9a9e43e65e1ae62213764aa79b5564. Mar 20 21:15:23.345513 kubelet[2596]: E0320 21:15:23.345458 2596 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-7htv8" podUID="4e0ac6d0-32f4-4ad3-a9b3-9da162162a58" Mar 20 21:15:23.353684 containerd[1485]: time="2025-03-20T21:15:23.353650035Z" level=info msg="StartContainer for \"a433c0863a65b06e47e0353cf2a1ab0b9f9a9e43e65e1ae62213764aa79b5564\" returns successfully" Mar 20 21:15:23.376372 systemd[1]: cri-containerd-a433c0863a65b06e47e0353cf2a1ab0b9f9a9e43e65e1ae62213764aa79b5564.scope: Deactivated successfully. Mar 20 21:15:23.412178 containerd[1485]: time="2025-03-20T21:15:23.412129626Z" level=info msg="received exit event container_id:\"a433c0863a65b06e47e0353cf2a1ab0b9f9a9e43e65e1ae62213764aa79b5564\" id:\"a433c0863a65b06e47e0353cf2a1ab0b9f9a9e43e65e1ae62213764aa79b5564\" pid:3267 exited_at:{seconds:1742505323 nanos:402429296}" Mar 20 21:15:23.412514 containerd[1485]: time="2025-03-20T21:15:23.412323951Z" level=info msg="TaskExit event in podsandbox handler container_id:\"a433c0863a65b06e47e0353cf2a1ab0b9f9a9e43e65e1ae62213764aa79b5564\" id:\"a433c0863a65b06e47e0353cf2a1ab0b9f9a9e43e65e1ae62213764aa79b5564\" pid:3267 exited_at:{seconds:1742505323 nanos:402429296}" Mar 20 21:15:23.438357 kubelet[2596]: I0320 21:15:23.438027 2596 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 20 21:15:23.458528 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a433c0863a65b06e47e0353cf2a1ab0b9f9a9e43e65e1ae62213764aa79b5564-rootfs.mount: Deactivated successfully. Mar 20 21:15:24.442835 containerd[1485]: time="2025-03-20T21:15:24.442570035Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.2\"" Mar 20 21:15:25.344445 kubelet[2596]: E0320 21:15:25.344369 2596 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-7htv8" podUID="4e0ac6d0-32f4-4ad3-a9b3-9da162162a58" Mar 20 21:15:27.284070 containerd[1485]: time="2025-03-20T21:15:27.284022209Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.29.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:15:27.285394 containerd[1485]: time="2025-03-20T21:15:27.285342911Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.29.2: active requests=0, bytes read=91227396" Mar 20 21:15:27.286380 containerd[1485]: time="2025-03-20T21:15:27.286350048Z" level=info msg="ImageCreate event name:\"sha256:57c2b1dcdc0045be5220c7237f900bce5f47c006714073859cf102b0eaa65290\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:15:27.288015 containerd[1485]: time="2025-03-20T21:15:27.287981835Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:890e1db6ae363695cfc23ffae4d612cc85cdd99d759bd539af6683969d0c3c25\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:15:27.288952 containerd[1485]: time="2025-03-20T21:15:27.288910731Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.29.2\" with image id \"sha256:57c2b1dcdc0045be5220c7237f900bce5f47c006714073859cf102b0eaa65290\", repo tag \"ghcr.io/flatcar/calico/cni:v3.29.2\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:890e1db6ae363695cfc23ffae4d612cc85cdd99d759bd539af6683969d0c3c25\", size \"92597153\" in 2.846301135s" Mar 20 21:15:27.288952 containerd[1485]: time="2025-03-20T21:15:27.288949811Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.2\" returns image reference \"sha256:57c2b1dcdc0045be5220c7237f900bce5f47c006714073859cf102b0eaa65290\"" Mar 20 21:15:27.290847 containerd[1485]: time="2025-03-20T21:15:27.290795002Z" level=info msg="CreateContainer within sandbox \"54ed274cf11ce43c3701d2bf81a57f8bf3ce02515cfbaa6de5aff9e86c50d2e6\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Mar 20 21:15:27.304369 containerd[1485]: time="2025-03-20T21:15:27.304334910Z" level=info msg="Container 7182e74c7e9d381b5702394a0ea0272b5db6b919096a92491e2e24db6654a75a: CDI devices from CRI Config.CDIDevices: []" Mar 20 21:15:27.311875 containerd[1485]: time="2025-03-20T21:15:27.311779795Z" level=info msg="CreateContainer within sandbox \"54ed274cf11ce43c3701d2bf81a57f8bf3ce02515cfbaa6de5aff9e86c50d2e6\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"7182e74c7e9d381b5702394a0ea0272b5db6b919096a92491e2e24db6654a75a\"" Mar 20 21:15:27.312255 containerd[1485]: time="2025-03-20T21:15:27.312140441Z" level=info msg="StartContainer for \"7182e74c7e9d381b5702394a0ea0272b5db6b919096a92491e2e24db6654a75a\"" Mar 20 21:15:27.313886 containerd[1485]: time="2025-03-20T21:15:27.313745748Z" level=info msg="connecting to shim 7182e74c7e9d381b5702394a0ea0272b5db6b919096a92491e2e24db6654a75a" address="unix:///run/containerd/s/6f52bec33ef527195a31a6634958f16dbfd178422935e01a514e23b3d3efee65" protocol=ttrpc version=3 Mar 20 21:15:27.341370 systemd[1]: Started cri-containerd-7182e74c7e9d381b5702394a0ea0272b5db6b919096a92491e2e24db6654a75a.scope - libcontainer container 7182e74c7e9d381b5702394a0ea0272b5db6b919096a92491e2e24db6654a75a. Mar 20 21:15:27.344156 kubelet[2596]: E0320 21:15:27.344118 2596 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-7htv8" podUID="4e0ac6d0-32f4-4ad3-a9b3-9da162162a58" Mar 20 21:15:27.376256 containerd[1485]: time="2025-03-20T21:15:27.376146076Z" level=info msg="StartContainer for \"7182e74c7e9d381b5702394a0ea0272b5db6b919096a92491e2e24db6654a75a\" returns successfully" Mar 20 21:15:28.011823 systemd[1]: cri-containerd-7182e74c7e9d381b5702394a0ea0272b5db6b919096a92491e2e24db6654a75a.scope: Deactivated successfully. Mar 20 21:15:28.012081 systemd[1]: cri-containerd-7182e74c7e9d381b5702394a0ea0272b5db6b919096a92491e2e24db6654a75a.scope: Consumed 439ms CPU time, 159.9M memory peak, 4K read from disk, 150.3M written to disk. Mar 20 21:15:28.014622 containerd[1485]: time="2025-03-20T21:15:28.014206982Z" level=info msg="TaskExit event in podsandbox handler container_id:\"7182e74c7e9d381b5702394a0ea0272b5db6b919096a92491e2e24db6654a75a\" id:\"7182e74c7e9d381b5702394a0ea0272b5db6b919096a92491e2e24db6654a75a\" pid:3329 exited_at:{seconds:1742505328 nanos:13916257}" Mar 20 21:15:28.014622 containerd[1485]: time="2025-03-20T21:15:28.014256142Z" level=info msg="received exit event container_id:\"7182e74c7e9d381b5702394a0ea0272b5db6b919096a92491e2e24db6654a75a\" id:\"7182e74c7e9d381b5702394a0ea0272b5db6b919096a92491e2e24db6654a75a\" pid:3329 exited_at:{seconds:1742505328 nanos:13916257}" Mar 20 21:15:28.039992 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-7182e74c7e9d381b5702394a0ea0272b5db6b919096a92491e2e24db6654a75a-rootfs.mount: Deactivated successfully. Mar 20 21:15:28.074473 kubelet[2596]: I0320 21:15:28.074443 2596 kubelet_node_status.go:488] "Fast updating node status as it just became ready" Mar 20 21:15:28.118611 systemd[1]: Created slice kubepods-burstable-pod87c040b4_ae2a_4fa1_b2e7_b2f5f4cd4424.slice - libcontainer container kubepods-burstable-pod87c040b4_ae2a_4fa1_b2e7_b2f5f4cd4424.slice. Mar 20 21:15:28.125180 systemd[1]: Created slice kubepods-besteffort-pod80afa69e_2493_484d_9af0_def2e6594d53.slice - libcontainer container kubepods-besteffort-pod80afa69e_2493_484d_9af0_def2e6594d53.slice. Mar 20 21:15:28.131391 systemd[1]: Created slice kubepods-burstable-pode1b822dc_3680_44ce_8c5d_15c8549a9319.slice - libcontainer container kubepods-burstable-pode1b822dc_3680_44ce_8c5d_15c8549a9319.slice. Mar 20 21:15:28.138164 systemd[1]: Created slice kubepods-besteffort-pod42908acd_daa0_4843_a9d2_569c24a392b7.slice - libcontainer container kubepods-besteffort-pod42908acd_daa0_4843_a9d2_569c24a392b7.slice. Mar 20 21:15:28.146420 systemd[1]: Created slice kubepods-besteffort-pod428dbe74_ae1a_409c_8f83_1014b5c9a996.slice - libcontainer container kubepods-besteffort-pod428dbe74_ae1a_409c_8f83_1014b5c9a996.slice. Mar 20 21:15:28.266049 kubelet[2596]: I0320 21:15:28.265868 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bftfs\" (UniqueName: \"kubernetes.io/projected/42908acd-daa0-4843-a9d2-569c24a392b7-kube-api-access-bftfs\") pod \"calico-apiserver-8646c76b8-xnq47\" (UID: \"42908acd-daa0-4843-a9d2-569c24a392b7\") " pod="calico-apiserver/calico-apiserver-8646c76b8-xnq47" Mar 20 21:15:28.266049 kubelet[2596]: I0320 21:15:28.265915 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/428dbe74-ae1a-409c-8f83-1014b5c9a996-tigera-ca-bundle\") pod \"calico-kube-controllers-5ddfd69446-sh2ld\" (UID: \"428dbe74-ae1a-409c-8f83-1014b5c9a996\") " pod="calico-system/calico-kube-controllers-5ddfd69446-sh2ld" Mar 20 21:15:28.266049 kubelet[2596]: I0320 21:15:28.265933 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vp6km\" (UniqueName: \"kubernetes.io/projected/80afa69e-2493-484d-9af0-def2e6594d53-kube-api-access-vp6km\") pod \"calico-apiserver-8646c76b8-kbdmr\" (UID: \"80afa69e-2493-484d-9af0-def2e6594d53\") " pod="calico-apiserver/calico-apiserver-8646c76b8-kbdmr" Mar 20 21:15:28.266049 kubelet[2596]: I0320 21:15:28.265954 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87c040b4-ae2a-4fa1-b2e7-b2f5f4cd4424-config-volume\") pod \"coredns-6f6b679f8f-qxfz8\" (UID: \"87c040b4-ae2a-4fa1-b2e7-b2f5f4cd4424\") " pod="kube-system/coredns-6f6b679f8f-qxfz8" Mar 20 21:15:28.266049 kubelet[2596]: I0320 21:15:28.265974 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/42908acd-daa0-4843-a9d2-569c24a392b7-calico-apiserver-certs\") pod \"calico-apiserver-8646c76b8-xnq47\" (UID: \"42908acd-daa0-4843-a9d2-569c24a392b7\") " pod="calico-apiserver/calico-apiserver-8646c76b8-xnq47" Mar 20 21:15:28.266375 kubelet[2596]: I0320 21:15:28.265991 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e1b822dc-3680-44ce-8c5d-15c8549a9319-config-volume\") pod \"coredns-6f6b679f8f-mbg5n\" (UID: \"e1b822dc-3680-44ce-8c5d-15c8549a9319\") " pod="kube-system/coredns-6f6b679f8f-mbg5n" Mar 20 21:15:28.266375 kubelet[2596]: I0320 21:15:28.266008 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/80afa69e-2493-484d-9af0-def2e6594d53-calico-apiserver-certs\") pod \"calico-apiserver-8646c76b8-kbdmr\" (UID: \"80afa69e-2493-484d-9af0-def2e6594d53\") " pod="calico-apiserver/calico-apiserver-8646c76b8-kbdmr" Mar 20 21:15:28.266375 kubelet[2596]: I0320 21:15:28.266048 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wnmd\" (UniqueName: \"kubernetes.io/projected/e1b822dc-3680-44ce-8c5d-15c8549a9319-kube-api-access-9wnmd\") pod \"coredns-6f6b679f8f-mbg5n\" (UID: \"e1b822dc-3680-44ce-8c5d-15c8549a9319\") " pod="kube-system/coredns-6f6b679f8f-mbg5n" Mar 20 21:15:28.266375 kubelet[2596]: I0320 21:15:28.266065 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgwb8\" (UniqueName: \"kubernetes.io/projected/87c040b4-ae2a-4fa1-b2e7-b2f5f4cd4424-kube-api-access-qgwb8\") pod \"coredns-6f6b679f8f-qxfz8\" (UID: \"87c040b4-ae2a-4fa1-b2e7-b2f5f4cd4424\") " pod="kube-system/coredns-6f6b679f8f-qxfz8" Mar 20 21:15:28.266375 kubelet[2596]: I0320 21:15:28.266086 2596 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qz2f6\" (UniqueName: \"kubernetes.io/projected/428dbe74-ae1a-409c-8f83-1014b5c9a996-kube-api-access-qz2f6\") pod \"calico-kube-controllers-5ddfd69446-sh2ld\" (UID: \"428dbe74-ae1a-409c-8f83-1014b5c9a996\") " pod="calico-system/calico-kube-controllers-5ddfd69446-sh2ld" Mar 20 21:15:28.423891 containerd[1485]: time="2025-03-20T21:15:28.423839473Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-qxfz8,Uid:87c040b4-ae2a-4fa1-b2e7-b2f5f4cd4424,Namespace:kube-system,Attempt:0,}" Mar 20 21:15:28.427424 containerd[1485]: time="2025-03-20T21:15:28.427387409Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-8646c76b8-kbdmr,Uid:80afa69e-2493-484d-9af0-def2e6594d53,Namespace:calico-apiserver,Attempt:0,}" Mar 20 21:15:28.436675 containerd[1485]: time="2025-03-20T21:15:28.436647275Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-mbg5n,Uid:e1b822dc-3680-44ce-8c5d-15c8549a9319,Namespace:kube-system,Attempt:0,}" Mar 20 21:15:28.443762 containerd[1485]: time="2025-03-20T21:15:28.443652025Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-8646c76b8-xnq47,Uid:42908acd-daa0-4843-a9d2-569c24a392b7,Namespace:calico-apiserver,Attempt:0,}" Mar 20 21:15:28.453855 containerd[1485]: time="2025-03-20T21:15:28.453654423Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5ddfd69446-sh2ld,Uid:428dbe74-ae1a-409c-8f83-1014b5c9a996,Namespace:calico-system,Attempt:0,}" Mar 20 21:15:28.492003 containerd[1485]: time="2025-03-20T21:15:28.487779640Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.2\"" Mar 20 21:15:28.795442 containerd[1485]: time="2025-03-20T21:15:28.795392285Z" level=error msg="Failed to destroy network for sandbox \"cc984ae93bb0d28e88b87f3cc85678cc341623a1f024563683401e920b9c316b\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 20 21:15:28.796362 containerd[1485]: time="2025-03-20T21:15:28.796317139Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-8646c76b8-xnq47,Uid:42908acd-daa0-4843-a9d2-569c24a392b7,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"cc984ae93bb0d28e88b87f3cc85678cc341623a1f024563683401e920b9c316b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 20 21:15:28.798537 containerd[1485]: time="2025-03-20T21:15:28.798502694Z" level=error msg="Failed to destroy network for sandbox \"f3a6720e582362b2333dd479322860676ced33e468a3a414bc0cacc0dc7e9dbb\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 20 21:15:28.799360 containerd[1485]: time="2025-03-20T21:15:28.799323867Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-8646c76b8-kbdmr,Uid:80afa69e-2493-484d-9af0-def2e6594d53,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"f3a6720e582362b2333dd479322860676ced33e468a3a414bc0cacc0dc7e9dbb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 20 21:15:28.802510 containerd[1485]: time="2025-03-20T21:15:28.802286073Z" level=error msg="Failed to destroy network for sandbox \"84f9e604e1a3c560838c2661effcc34f5f3a1bd1572f54f80cf6e329bcae4807\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 20 21:15:28.803968 kubelet[2596]: E0320 21:15:28.803585 2596 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f3a6720e582362b2333dd479322860676ced33e468a3a414bc0cacc0dc7e9dbb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 20 21:15:28.803968 kubelet[2596]: E0320 21:15:28.803673 2596 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f3a6720e582362b2333dd479322860676ced33e468a3a414bc0cacc0dc7e9dbb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-8646c76b8-kbdmr" Mar 20 21:15:28.803968 kubelet[2596]: E0320 21:15:28.803692 2596 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f3a6720e582362b2333dd479322860676ced33e468a3a414bc0cacc0dc7e9dbb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-8646c76b8-kbdmr" Mar 20 21:15:28.805580 kubelet[2596]: E0320 21:15:28.803738 2596 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-8646c76b8-kbdmr_calico-apiserver(80afa69e-2493-484d-9af0-def2e6594d53)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-8646c76b8-kbdmr_calico-apiserver(80afa69e-2493-484d-9af0-def2e6594d53)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"f3a6720e582362b2333dd479322860676ced33e468a3a414bc0cacc0dc7e9dbb\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-8646c76b8-kbdmr" podUID="80afa69e-2493-484d-9af0-def2e6594d53" Mar 20 21:15:28.805580 kubelet[2596]: E0320 21:15:28.803938 2596 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"cc984ae93bb0d28e88b87f3cc85678cc341623a1f024563683401e920b9c316b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 20 21:15:28.805580 kubelet[2596]: E0320 21:15:28.803989 2596 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"cc984ae93bb0d28e88b87f3cc85678cc341623a1f024563683401e920b9c316b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-8646c76b8-xnq47" Mar 20 21:15:28.805677 kubelet[2596]: E0320 21:15:28.804030 2596 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"cc984ae93bb0d28e88b87f3cc85678cc341623a1f024563683401e920b9c316b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-8646c76b8-xnq47" Mar 20 21:15:28.805677 kubelet[2596]: E0320 21:15:28.804071 2596 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-8646c76b8-xnq47_calico-apiserver(42908acd-daa0-4843-a9d2-569c24a392b7)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-8646c76b8-xnq47_calico-apiserver(42908acd-daa0-4843-a9d2-569c24a392b7)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"cc984ae93bb0d28e88b87f3cc85678cc341623a1f024563683401e920b9c316b\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-8646c76b8-xnq47" podUID="42908acd-daa0-4843-a9d2-569c24a392b7" Mar 20 21:15:28.806441 containerd[1485]: time="2025-03-20T21:15:28.806399458Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-qxfz8,Uid:87c040b4-ae2a-4fa1-b2e7-b2f5f4cd4424,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"84f9e604e1a3c560838c2661effcc34f5f3a1bd1572f54f80cf6e329bcae4807\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 20 21:15:28.807980 containerd[1485]: time="2025-03-20T21:15:28.807879081Z" level=error msg="Failed to destroy network for sandbox \"91eec9eb1ea2dd6f1d82abf0ba3a30d56b8e62340cb73b34758a9223875aeb8c\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 20 21:15:28.808115 kubelet[2596]: E0320 21:15:28.807903 2596 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"84f9e604e1a3c560838c2661effcc34f5f3a1bd1572f54f80cf6e329bcae4807\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 20 21:15:28.808158 kubelet[2596]: E0320 21:15:28.808137 2596 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"84f9e604e1a3c560838c2661effcc34f5f3a1bd1572f54f80cf6e329bcae4807\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-qxfz8" Mar 20 21:15:28.808191 kubelet[2596]: E0320 21:15:28.808159 2596 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"84f9e604e1a3c560838c2661effcc34f5f3a1bd1572f54f80cf6e329bcae4807\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-qxfz8" Mar 20 21:15:28.808373 kubelet[2596]: E0320 21:15:28.808208 2596 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-6f6b679f8f-qxfz8_kube-system(87c040b4-ae2a-4fa1-b2e7-b2f5f4cd4424)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-6f6b679f8f-qxfz8_kube-system(87c040b4-ae2a-4fa1-b2e7-b2f5f4cd4424)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"84f9e604e1a3c560838c2661effcc34f5f3a1bd1572f54f80cf6e329bcae4807\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-6f6b679f8f-qxfz8" podUID="87c040b4-ae2a-4fa1-b2e7-b2f5f4cd4424" Mar 20 21:15:28.808914 containerd[1485]: time="2025-03-20T21:15:28.808882097Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-mbg5n,Uid:e1b822dc-3680-44ce-8c5d-15c8549a9319,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"91eec9eb1ea2dd6f1d82abf0ba3a30d56b8e62340cb73b34758a9223875aeb8c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 20 21:15:28.809150 kubelet[2596]: E0320 21:15:28.809119 2596 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"91eec9eb1ea2dd6f1d82abf0ba3a30d56b8e62340cb73b34758a9223875aeb8c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 20 21:15:28.809224 kubelet[2596]: E0320 21:15:28.809160 2596 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"91eec9eb1ea2dd6f1d82abf0ba3a30d56b8e62340cb73b34758a9223875aeb8c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-mbg5n" Mar 20 21:15:28.809224 kubelet[2596]: E0320 21:15:28.809177 2596 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"91eec9eb1ea2dd6f1d82abf0ba3a30d56b8e62340cb73b34758a9223875aeb8c\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-mbg5n" Mar 20 21:15:28.809363 kubelet[2596]: E0320 21:15:28.809207 2596 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-6f6b679f8f-mbg5n_kube-system(e1b822dc-3680-44ce-8c5d-15c8549a9319)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-6f6b679f8f-mbg5n_kube-system(e1b822dc-3680-44ce-8c5d-15c8549a9319)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"91eec9eb1ea2dd6f1d82abf0ba3a30d56b8e62340cb73b34758a9223875aeb8c\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-6f6b679f8f-mbg5n" podUID="e1b822dc-3680-44ce-8c5d-15c8549a9319" Mar 20 21:15:28.812798 containerd[1485]: time="2025-03-20T21:15:28.812749678Z" level=error msg="Failed to destroy network for sandbox \"a134c180c667bce353ed4642356f5e2a3bc6f2a78dcb3d0eb75b48b7b0ff1d3a\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 20 21:15:28.814491 containerd[1485]: time="2025-03-20T21:15:28.814458865Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5ddfd69446-sh2ld,Uid:428dbe74-ae1a-409c-8f83-1014b5c9a996,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"a134c180c667bce353ed4642356f5e2a3bc6f2a78dcb3d0eb75b48b7b0ff1d3a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 20 21:15:28.814690 kubelet[2596]: E0320 21:15:28.814649 2596 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a134c180c667bce353ed4642356f5e2a3bc6f2a78dcb3d0eb75b48b7b0ff1d3a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 20 21:15:28.814729 kubelet[2596]: E0320 21:15:28.814697 2596 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a134c180c667bce353ed4642356f5e2a3bc6f2a78dcb3d0eb75b48b7b0ff1d3a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5ddfd69446-sh2ld" Mar 20 21:15:28.814729 kubelet[2596]: E0320 21:15:28.814714 2596 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a134c180c667bce353ed4642356f5e2a3bc6f2a78dcb3d0eb75b48b7b0ff1d3a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-5ddfd69446-sh2ld" Mar 20 21:15:28.814774 kubelet[2596]: E0320 21:15:28.814748 2596 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-5ddfd69446-sh2ld_calico-system(428dbe74-ae1a-409c-8f83-1014b5c9a996)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-5ddfd69446-sh2ld_calico-system(428dbe74-ae1a-409c-8f83-1014b5c9a996)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"a134c180c667bce353ed4642356f5e2a3bc6f2a78dcb3d0eb75b48b7b0ff1d3a\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-5ddfd69446-sh2ld" podUID="428dbe74-ae1a-409c-8f83-1014b5c9a996" Mar 20 21:15:29.348995 systemd[1]: Created slice kubepods-besteffort-pod4e0ac6d0_32f4_4ad3_a9b3_9da162162a58.slice - libcontainer container kubepods-besteffort-pod4e0ac6d0_32f4_4ad3_a9b3_9da162162a58.slice. Mar 20 21:15:29.351176 containerd[1485]: time="2025-03-20T21:15:29.351088332Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-7htv8,Uid:4e0ac6d0-32f4-4ad3-a9b3-9da162162a58,Namespace:calico-system,Attempt:0,}" Mar 20 21:15:29.371959 systemd[1]: run-netns-cni\x2dca129dfc\x2dc993\x2d290f\x2d7431\x2dcc5385851d4f.mount: Deactivated successfully. Mar 20 21:15:29.395308 containerd[1485]: time="2025-03-20T21:15:29.395261424Z" level=error msg="Failed to destroy network for sandbox \"531f116d719ab2b9b18db37278bb3ea4097039530c37a179cf081ab1c00b1473\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 20 21:15:29.397791 systemd[1]: run-netns-cni\x2d3d42eda3\x2d9262\x2de7f0\x2d760d\x2d34637c32afb9.mount: Deactivated successfully. Mar 20 21:15:29.397928 containerd[1485]: time="2025-03-20T21:15:29.397819742Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-7htv8,Uid:4e0ac6d0-32f4-4ad3-a9b3-9da162162a58,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"531f116d719ab2b9b18db37278bb3ea4097039530c37a179cf081ab1c00b1473\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 20 21:15:29.398651 kubelet[2596]: E0320 21:15:29.398071 2596 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"531f116d719ab2b9b18db37278bb3ea4097039530c37a179cf081ab1c00b1473\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Mar 20 21:15:29.398651 kubelet[2596]: E0320 21:15:29.398130 2596 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"531f116d719ab2b9b18db37278bb3ea4097039530c37a179cf081ab1c00b1473\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-7htv8" Mar 20 21:15:29.398651 kubelet[2596]: E0320 21:15:29.398163 2596 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"531f116d719ab2b9b18db37278bb3ea4097039530c37a179cf081ab1c00b1473\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-7htv8" Mar 20 21:15:29.398807 kubelet[2596]: E0320 21:15:29.398213 2596 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-7htv8_calico-system(4e0ac6d0-32f4-4ad3-a9b3-9da162162a58)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-7htv8_calico-system(4e0ac6d0-32f4-4ad3-a9b3-9da162162a58)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"531f116d719ab2b9b18db37278bb3ea4097039530c37a179cf081ab1c00b1473\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-7htv8" podUID="4e0ac6d0-32f4-4ad3-a9b3-9da162162a58" Mar 20 21:15:32.330965 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount525178462.mount: Deactivated successfully. Mar 20 21:15:32.453865 containerd[1485]: time="2025-03-20T21:15:32.453822137Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.29.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:15:32.454946 containerd[1485]: time="2025-03-20T21:15:32.454523586Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.29.2: active requests=0, bytes read=137086024" Mar 20 21:15:32.455557 containerd[1485]: time="2025-03-20T21:15:32.455505758Z" level=info msg="ImageCreate event name:\"sha256:8fd1983cc851d15f05a37eb3ff85b0cde86869beec7630d2940c86fc7b98d0c1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:15:32.457245 containerd[1485]: time="2025-03-20T21:15:32.457178058Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:d9a21be37fe591ee5ab5a2e3dc26408ea165a44a55705102ffaa002de9908b32\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:15:32.457640 containerd[1485]: time="2025-03-20T21:15:32.457605623Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.29.2\" with image id \"sha256:8fd1983cc851d15f05a37eb3ff85b0cde86869beec7630d2940c86fc7b98d0c1\", repo tag \"ghcr.io/flatcar/calico/node:v3.29.2\", repo digest \"ghcr.io/flatcar/calico/node@sha256:d9a21be37fe591ee5ab5a2e3dc26408ea165a44a55705102ffaa002de9908b32\", size \"137085886\" in 3.969788023s" Mar 20 21:15:32.457686 containerd[1485]: time="2025-03-20T21:15:32.457639703Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.2\" returns image reference \"sha256:8fd1983cc851d15f05a37eb3ff85b0cde86869beec7630d2940c86fc7b98d0c1\"" Mar 20 21:15:32.474744 containerd[1485]: time="2025-03-20T21:15:32.474701431Z" level=info msg="CreateContainer within sandbox \"54ed274cf11ce43c3701d2bf81a57f8bf3ce02515cfbaa6de5aff9e86c50d2e6\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Mar 20 21:15:32.483236 containerd[1485]: time="2025-03-20T21:15:32.482661328Z" level=info msg="Container 3298a708f055861f6f20ce36a7ee251ccfc70f10f7107e13a9514eaec3f5beb1: CDI devices from CRI Config.CDIDevices: []" Mar 20 21:15:32.508004 containerd[1485]: time="2025-03-20T21:15:32.507946756Z" level=info msg="CreateContainer within sandbox \"54ed274cf11ce43c3701d2bf81a57f8bf3ce02515cfbaa6de5aff9e86c50d2e6\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"3298a708f055861f6f20ce36a7ee251ccfc70f10f7107e13a9514eaec3f5beb1\"" Mar 20 21:15:32.509000 containerd[1485]: time="2025-03-20T21:15:32.508761365Z" level=info msg="StartContainer for \"3298a708f055861f6f20ce36a7ee251ccfc70f10f7107e13a9514eaec3f5beb1\"" Mar 20 21:15:32.512369 containerd[1485]: time="2025-03-20T21:15:32.512345049Z" level=info msg="connecting to shim 3298a708f055861f6f20ce36a7ee251ccfc70f10f7107e13a9514eaec3f5beb1" address="unix:///run/containerd/s/6f52bec33ef527195a31a6634958f16dbfd178422935e01a514e23b3d3efee65" protocol=ttrpc version=3 Mar 20 21:15:32.542399 systemd[1]: Started cri-containerd-3298a708f055861f6f20ce36a7ee251ccfc70f10f7107e13a9514eaec3f5beb1.scope - libcontainer container 3298a708f055861f6f20ce36a7ee251ccfc70f10f7107e13a9514eaec3f5beb1. Mar 20 21:15:32.601779 containerd[1485]: time="2025-03-20T21:15:32.599726432Z" level=info msg="StartContainer for \"3298a708f055861f6f20ce36a7ee251ccfc70f10f7107e13a9514eaec3f5beb1\" returns successfully" Mar 20 21:15:32.743254 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Mar 20 21:15:32.743361 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Mar 20 21:15:32.965476 systemd[1]: Started sshd@7-10.0.0.54:22-10.0.0.1:37526.service - OpenSSH per-connection server daemon (10.0.0.1:37526). Mar 20 21:15:33.031537 sshd[3659]: Accepted publickey for core from 10.0.0.1 port 37526 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:15:33.031976 sshd-session[3659]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:15:33.035540 systemd-logind[1470]: New session 8 of user core. Mar 20 21:15:33.048360 systemd[1]: Started session-8.scope - Session 8 of User core. Mar 20 21:15:33.175378 sshd[3662]: Connection closed by 10.0.0.1 port 37526 Mar 20 21:15:33.174754 sshd-session[3659]: pam_unix(sshd:session): session closed for user core Mar 20 21:15:33.179826 systemd[1]: sshd@7-10.0.0.54:22-10.0.0.1:37526.service: Deactivated successfully. Mar 20 21:15:33.182630 systemd[1]: session-8.scope: Deactivated successfully. Mar 20 21:15:33.183650 systemd-logind[1470]: Session 8 logged out. Waiting for processes to exit. Mar 20 21:15:33.184950 systemd-logind[1470]: Removed session 8. Mar 20 21:15:33.528656 kubelet[2596]: I0320 21:15:33.528571 2596 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-dpxdp" podStartSLOduration=2.2320495830000002 podStartE2EDuration="14.528555371s" podCreationTimestamp="2025-03-20 21:15:19 +0000 UTC" firstStartedPulling="2025-03-20 21:15:20.162027846 +0000 UTC m=+15.893511901" lastFinishedPulling="2025-03-20 21:15:32.458533594 +0000 UTC m=+28.190017689" observedRunningTime="2025-03-20 21:15:33.527302597 +0000 UTC m=+29.258786692" watchObservedRunningTime="2025-03-20 21:15:33.528555371 +0000 UTC m=+29.260039466" Mar 20 21:15:34.514839 kubelet[2596]: I0320 21:15:34.514802 2596 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 20 21:15:34.761020 containerd[1485]: time="2025-03-20T21:15:34.760974566Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3298a708f055861f6f20ce36a7ee251ccfc70f10f7107e13a9514eaec3f5beb1\" id:\"afe45aa9dbcfcb35a7b60466b2b3cffa83f861b63b5ea7ead4165dd002cb7825\" pid:3787 exit_status:1 exited_at:{seconds:1742505334 nanos:760694243}" Mar 20 21:15:34.832349 containerd[1485]: time="2025-03-20T21:15:34.832190688Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3298a708f055861f6f20ce36a7ee251ccfc70f10f7107e13a9514eaec3f5beb1\" id:\"fc420a3ee81f0f419bfae6633592b7a73b5472bcb78161097a800af1ca645aeb\" pid:3812 exit_status:1 exited_at:{seconds:1742505334 nanos:831849284}" Mar 20 21:15:35.565371 containerd[1485]: time="2025-03-20T21:15:35.565331470Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3298a708f055861f6f20ce36a7ee251ccfc70f10f7107e13a9514eaec3f5beb1\" id:\"3cb078fdda45fa5b9134f5dc472297ec4554c8b7f959979b1f210d7c166c11b9\" pid:3861 exit_status:1 exited_at:{seconds:1742505335 nanos:565076828}" Mar 20 21:15:35.665864 kubelet[2596]: I0320 21:15:35.665824 2596 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 20 21:15:36.269416 kernel: bpftool[3927]: memfd_create() called without MFD_EXEC or MFD_NOEXEC_SEAL set Mar 20 21:15:36.432844 systemd-networkd[1416]: vxlan.calico: Link UP Mar 20 21:15:36.432852 systemd-networkd[1416]: vxlan.calico: Gained carrier Mar 20 21:15:37.877405 systemd-networkd[1416]: vxlan.calico: Gained IPv6LL Mar 20 21:15:38.188508 systemd[1]: Started sshd@8-10.0.0.54:22-10.0.0.1:37542.service - OpenSSH per-connection server daemon (10.0.0.1:37542). Mar 20 21:15:38.256690 sshd[4013]: Accepted publickey for core from 10.0.0.1 port 37542 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:15:38.258246 sshd-session[4013]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:15:38.262736 systemd-logind[1470]: New session 9 of user core. Mar 20 21:15:38.273443 systemd[1]: Started session-9.scope - Session 9 of User core. Mar 20 21:15:38.414058 sshd[4016]: Connection closed by 10.0.0.1 port 37542 Mar 20 21:15:38.414432 sshd-session[4013]: pam_unix(sshd:session): session closed for user core Mar 20 21:15:38.417884 systemd[1]: sshd@8-10.0.0.54:22-10.0.0.1:37542.service: Deactivated successfully. Mar 20 21:15:38.421875 systemd[1]: session-9.scope: Deactivated successfully. Mar 20 21:15:38.422831 systemd-logind[1470]: Session 9 logged out. Waiting for processes to exit. Mar 20 21:15:38.423829 systemd-logind[1470]: Removed session 9. Mar 20 21:15:39.344994 containerd[1485]: time="2025-03-20T21:15:39.344886650Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-qxfz8,Uid:87c040b4-ae2a-4fa1-b2e7-b2f5f4cd4424,Namespace:kube-system,Attempt:0,}" Mar 20 21:15:39.573508 systemd-networkd[1416]: cali4f695bc4646: Link UP Mar 20 21:15:39.573671 systemd-networkd[1416]: cali4f695bc4646: Gained carrier Mar 20 21:15:39.587359 containerd[1485]: 2025-03-20 21:15:39.405 [INFO][4029] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--6f6b679f8f--qxfz8-eth0 coredns-6f6b679f8f- kube-system 87c040b4-ae2a-4fa1-b2e7-b2f5f4cd4424 697 0 2025-03-20 21:15:10 +0000 UTC map[k8s-app:kube-dns pod-template-hash:6f6b679f8f projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-6f6b679f8f-qxfz8 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali4f695bc4646 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="1bd1d7dec319bbb163863bfbdf9bbafd3719c64a370206bd48fddfc8374df856" Namespace="kube-system" Pod="coredns-6f6b679f8f-qxfz8" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--qxfz8-" Mar 20 21:15:39.587359 containerd[1485]: 2025-03-20 21:15:39.405 [INFO][4029] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="1bd1d7dec319bbb163863bfbdf9bbafd3719c64a370206bd48fddfc8374df856" Namespace="kube-system" Pod="coredns-6f6b679f8f-qxfz8" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--qxfz8-eth0" Mar 20 21:15:39.587359 containerd[1485]: 2025-03-20 21:15:39.515 [INFO][4044] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="1bd1d7dec319bbb163863bfbdf9bbafd3719c64a370206bd48fddfc8374df856" HandleID="k8s-pod-network.1bd1d7dec319bbb163863bfbdf9bbafd3719c64a370206bd48fddfc8374df856" Workload="localhost-k8s-coredns--6f6b679f8f--qxfz8-eth0" Mar 20 21:15:39.587852 containerd[1485]: 2025-03-20 21:15:39.528 [INFO][4044] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="1bd1d7dec319bbb163863bfbdf9bbafd3719c64a370206bd48fddfc8374df856" HandleID="k8s-pod-network.1bd1d7dec319bbb163863bfbdf9bbafd3719c64a370206bd48fddfc8374df856" Workload="localhost-k8s-coredns--6f6b679f8f--qxfz8-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000277200), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-6f6b679f8f-qxfz8", "timestamp":"2025-03-20 21:15:39.515694732 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Mar 20 21:15:39.587852 containerd[1485]: 2025-03-20 21:15:39.528 [INFO][4044] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Mar 20 21:15:39.587852 containerd[1485]: 2025-03-20 21:15:39.528 [INFO][4044] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Mar 20 21:15:39.587852 containerd[1485]: 2025-03-20 21:15:39.529 [INFO][4044] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Mar 20 21:15:39.587852 containerd[1485]: 2025-03-20 21:15:39.533 [INFO][4044] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.1bd1d7dec319bbb163863bfbdf9bbafd3719c64a370206bd48fddfc8374df856" host="localhost" Mar 20 21:15:39.587852 containerd[1485]: 2025-03-20 21:15:39.539 [INFO][4044] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" Mar 20 21:15:39.587852 containerd[1485]: 2025-03-20 21:15:39.542 [INFO][4044] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Mar 20 21:15:39.587852 containerd[1485]: 2025-03-20 21:15:39.544 [INFO][4044] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Mar 20 21:15:39.587852 containerd[1485]: 2025-03-20 21:15:39.547 [INFO][4044] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Mar 20 21:15:39.587852 containerd[1485]: 2025-03-20 21:15:39.547 [INFO][4044] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.1bd1d7dec319bbb163863bfbdf9bbafd3719c64a370206bd48fddfc8374df856" host="localhost" Mar 20 21:15:39.588078 containerd[1485]: 2025-03-20 21:15:39.549 [INFO][4044] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.1bd1d7dec319bbb163863bfbdf9bbafd3719c64a370206bd48fddfc8374df856 Mar 20 21:15:39.588078 containerd[1485]: 2025-03-20 21:15:39.558 [INFO][4044] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.1bd1d7dec319bbb163863bfbdf9bbafd3719c64a370206bd48fddfc8374df856" host="localhost" Mar 20 21:15:39.588078 containerd[1485]: 2025-03-20 21:15:39.563 [INFO][4044] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.129/26] block=192.168.88.128/26 handle="k8s-pod-network.1bd1d7dec319bbb163863bfbdf9bbafd3719c64a370206bd48fddfc8374df856" host="localhost" Mar 20 21:15:39.588078 containerd[1485]: 2025-03-20 21:15:39.563 [INFO][4044] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.129/26] handle="k8s-pod-network.1bd1d7dec319bbb163863bfbdf9bbafd3719c64a370206bd48fddfc8374df856" host="localhost" Mar 20 21:15:39.588078 containerd[1485]: 2025-03-20 21:15:39.563 [INFO][4044] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Mar 20 21:15:39.588078 containerd[1485]: 2025-03-20 21:15:39.563 [INFO][4044] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.129/26] IPv6=[] ContainerID="1bd1d7dec319bbb163863bfbdf9bbafd3719c64a370206bd48fddfc8374df856" HandleID="k8s-pod-network.1bd1d7dec319bbb163863bfbdf9bbafd3719c64a370206bd48fddfc8374df856" Workload="localhost-k8s-coredns--6f6b679f8f--qxfz8-eth0" Mar 20 21:15:39.588193 containerd[1485]: 2025-03-20 21:15:39.566 [INFO][4029] cni-plugin/k8s.go 386: Populated endpoint ContainerID="1bd1d7dec319bbb163863bfbdf9bbafd3719c64a370206bd48fddfc8374df856" Namespace="kube-system" Pod="coredns-6f6b679f8f-qxfz8" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--qxfz8-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--6f6b679f8f--qxfz8-eth0", GenerateName:"coredns-6f6b679f8f-", Namespace:"kube-system", SelfLink:"", UID:"87c040b4-ae2a-4fa1-b2e7-b2f5f4cd4424", ResourceVersion:"697", Generation:0, CreationTimestamp:time.Date(2025, time.March, 20, 21, 15, 10, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"6f6b679f8f", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-6f6b679f8f-qxfz8", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali4f695bc4646", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Mar 20 21:15:39.588286 containerd[1485]: 2025-03-20 21:15:39.566 [INFO][4029] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.129/32] ContainerID="1bd1d7dec319bbb163863bfbdf9bbafd3719c64a370206bd48fddfc8374df856" Namespace="kube-system" Pod="coredns-6f6b679f8f-qxfz8" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--qxfz8-eth0" Mar 20 21:15:39.588286 containerd[1485]: 2025-03-20 21:15:39.569 [INFO][4029] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali4f695bc4646 ContainerID="1bd1d7dec319bbb163863bfbdf9bbafd3719c64a370206bd48fddfc8374df856" Namespace="kube-system" Pod="coredns-6f6b679f8f-qxfz8" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--qxfz8-eth0" Mar 20 21:15:39.588286 containerd[1485]: 2025-03-20 21:15:39.574 [INFO][4029] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="1bd1d7dec319bbb163863bfbdf9bbafd3719c64a370206bd48fddfc8374df856" Namespace="kube-system" Pod="coredns-6f6b679f8f-qxfz8" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--qxfz8-eth0" Mar 20 21:15:39.588365 containerd[1485]: 2025-03-20 21:15:39.574 [INFO][4029] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="1bd1d7dec319bbb163863bfbdf9bbafd3719c64a370206bd48fddfc8374df856" Namespace="kube-system" Pod="coredns-6f6b679f8f-qxfz8" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--qxfz8-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--6f6b679f8f--qxfz8-eth0", GenerateName:"coredns-6f6b679f8f-", Namespace:"kube-system", SelfLink:"", UID:"87c040b4-ae2a-4fa1-b2e7-b2f5f4cd4424", ResourceVersion:"697", Generation:0, CreationTimestamp:time.Date(2025, time.March, 20, 21, 15, 10, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"6f6b679f8f", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"1bd1d7dec319bbb163863bfbdf9bbafd3719c64a370206bd48fddfc8374df856", Pod:"coredns-6f6b679f8f-qxfz8", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali4f695bc4646", MAC:"12:d3:68:76:ed:ee", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Mar 20 21:15:39.588365 containerd[1485]: 2025-03-20 21:15:39.582 [INFO][4029] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="1bd1d7dec319bbb163863bfbdf9bbafd3719c64a370206bd48fddfc8374df856" Namespace="kube-system" Pod="coredns-6f6b679f8f-qxfz8" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--qxfz8-eth0" Mar 20 21:15:39.617264 containerd[1485]: time="2025-03-20T21:15:39.616715155Z" level=info msg="connecting to shim 1bd1d7dec319bbb163863bfbdf9bbafd3719c64a370206bd48fddfc8374df856" address="unix:///run/containerd/s/f59317837e1192d48c3c4b7831b2ecef5f1b5c72b0bdb5c152f99274c1cdad5d" namespace=k8s.io protocol=ttrpc version=3 Mar 20 21:15:39.642453 systemd[1]: Started cri-containerd-1bd1d7dec319bbb163863bfbdf9bbafd3719c64a370206bd48fddfc8374df856.scope - libcontainer container 1bd1d7dec319bbb163863bfbdf9bbafd3719c64a370206bd48fddfc8374df856. Mar 20 21:15:39.656408 systemd-resolved[1330]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Mar 20 21:15:39.716710 containerd[1485]: time="2025-03-20T21:15:39.716665168Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-qxfz8,Uid:87c040b4-ae2a-4fa1-b2e7-b2f5f4cd4424,Namespace:kube-system,Attempt:0,} returns sandbox id \"1bd1d7dec319bbb163863bfbdf9bbafd3719c64a370206bd48fddfc8374df856\"" Mar 20 21:15:39.719423 containerd[1485]: time="2025-03-20T21:15:39.719384910Z" level=info msg="CreateContainer within sandbox \"1bd1d7dec319bbb163863bfbdf9bbafd3719c64a370206bd48fddfc8374df856\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Mar 20 21:15:39.747422 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3868115902.mount: Deactivated successfully. Mar 20 21:15:39.749511 containerd[1485]: time="2025-03-20T21:15:39.749459622Z" level=info msg="Container ccae1f9e695f85944ad81c141aca5b1552d328ef319db34f02107209796ff9ae: CDI devices from CRI Config.CDIDevices: []" Mar 20 21:15:39.770898 containerd[1485]: time="2025-03-20T21:15:39.770834868Z" level=info msg="CreateContainer within sandbox \"1bd1d7dec319bbb163863bfbdf9bbafd3719c64a370206bd48fddfc8374df856\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"ccae1f9e695f85944ad81c141aca5b1552d328ef319db34f02107209796ff9ae\"" Mar 20 21:15:39.771738 containerd[1485]: time="2025-03-20T21:15:39.771711315Z" level=info msg="StartContainer for \"ccae1f9e695f85944ad81c141aca5b1552d328ef319db34f02107209796ff9ae\"" Mar 20 21:15:39.773846 containerd[1485]: time="2025-03-20T21:15:39.773822691Z" level=info msg="connecting to shim ccae1f9e695f85944ad81c141aca5b1552d328ef319db34f02107209796ff9ae" address="unix:///run/containerd/s/f59317837e1192d48c3c4b7831b2ecef5f1b5c72b0bdb5c152f99274c1cdad5d" protocol=ttrpc version=3 Mar 20 21:15:39.799385 systemd[1]: Started cri-containerd-ccae1f9e695f85944ad81c141aca5b1552d328ef319db34f02107209796ff9ae.scope - libcontainer container ccae1f9e695f85944ad81c141aca5b1552d328ef319db34f02107209796ff9ae. Mar 20 21:15:39.843534 containerd[1485]: time="2025-03-20T21:15:39.843480231Z" level=info msg="StartContainer for \"ccae1f9e695f85944ad81c141aca5b1552d328ef319db34f02107209796ff9ae\" returns successfully" Mar 20 21:15:40.345031 containerd[1485]: time="2025-03-20T21:15:40.344990509Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5ddfd69446-sh2ld,Uid:428dbe74-ae1a-409c-8f83-1014b5c9a996,Namespace:calico-system,Attempt:0,}" Mar 20 21:15:40.464046 systemd-networkd[1416]: cali2c3802e6845: Link UP Mar 20 21:15:40.464568 systemd-networkd[1416]: cali2c3802e6845: Gained carrier Mar 20 21:15:40.475873 containerd[1485]: 2025-03-20 21:15:40.383 [INFO][4146] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--kube--controllers--5ddfd69446--sh2ld-eth0 calico-kube-controllers-5ddfd69446- calico-system 428dbe74-ae1a-409c-8f83-1014b5c9a996 705 0 2025-03-20 21:15:20 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:5ddfd69446 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s localhost calico-kube-controllers-5ddfd69446-sh2ld eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali2c3802e6845 [] []}} ContainerID="7ec20bb10b38c16b2e9a5db7fba57e5985e572fb6c605731af4bee45d12ddcd6" Namespace="calico-system" Pod="calico-kube-controllers-5ddfd69446-sh2ld" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5ddfd69446--sh2ld-" Mar 20 21:15:40.475873 containerd[1485]: 2025-03-20 21:15:40.383 [INFO][4146] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="7ec20bb10b38c16b2e9a5db7fba57e5985e572fb6c605731af4bee45d12ddcd6" Namespace="calico-system" Pod="calico-kube-controllers-5ddfd69446-sh2ld" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5ddfd69446--sh2ld-eth0" Mar 20 21:15:40.475873 containerd[1485]: 2025-03-20 21:15:40.422 [INFO][4159] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="7ec20bb10b38c16b2e9a5db7fba57e5985e572fb6c605731af4bee45d12ddcd6" HandleID="k8s-pod-network.7ec20bb10b38c16b2e9a5db7fba57e5985e572fb6c605731af4bee45d12ddcd6" Workload="localhost-k8s-calico--kube--controllers--5ddfd69446--sh2ld-eth0" Mar 20 21:15:40.475873 containerd[1485]: 2025-03-20 21:15:40.433 [INFO][4159] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="7ec20bb10b38c16b2e9a5db7fba57e5985e572fb6c605731af4bee45d12ddcd6" HandleID="k8s-pod-network.7ec20bb10b38c16b2e9a5db7fba57e5985e572fb6c605731af4bee45d12ddcd6" Workload="localhost-k8s-calico--kube--controllers--5ddfd69446--sh2ld-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002d92e0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"calico-kube-controllers-5ddfd69446-sh2ld", "timestamp":"2025-03-20 21:15:40.422719073 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Mar 20 21:15:40.475873 containerd[1485]: 2025-03-20 21:15:40.433 [INFO][4159] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Mar 20 21:15:40.475873 containerd[1485]: 2025-03-20 21:15:40.433 [INFO][4159] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Mar 20 21:15:40.475873 containerd[1485]: 2025-03-20 21:15:40.433 [INFO][4159] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Mar 20 21:15:40.475873 containerd[1485]: 2025-03-20 21:15:40.435 [INFO][4159] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.7ec20bb10b38c16b2e9a5db7fba57e5985e572fb6c605731af4bee45d12ddcd6" host="localhost" Mar 20 21:15:40.475873 containerd[1485]: 2025-03-20 21:15:40.439 [INFO][4159] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" Mar 20 21:15:40.475873 containerd[1485]: 2025-03-20 21:15:40.446 [INFO][4159] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Mar 20 21:15:40.475873 containerd[1485]: 2025-03-20 21:15:40.448 [INFO][4159] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Mar 20 21:15:40.475873 containerd[1485]: 2025-03-20 21:15:40.450 [INFO][4159] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Mar 20 21:15:40.475873 containerd[1485]: 2025-03-20 21:15:40.450 [INFO][4159] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.7ec20bb10b38c16b2e9a5db7fba57e5985e572fb6c605731af4bee45d12ddcd6" host="localhost" Mar 20 21:15:40.475873 containerd[1485]: 2025-03-20 21:15:40.451 [INFO][4159] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.7ec20bb10b38c16b2e9a5db7fba57e5985e572fb6c605731af4bee45d12ddcd6 Mar 20 21:15:40.475873 containerd[1485]: 2025-03-20 21:15:40.455 [INFO][4159] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.7ec20bb10b38c16b2e9a5db7fba57e5985e572fb6c605731af4bee45d12ddcd6" host="localhost" Mar 20 21:15:40.475873 containerd[1485]: 2025-03-20 21:15:40.460 [INFO][4159] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.130/26] block=192.168.88.128/26 handle="k8s-pod-network.7ec20bb10b38c16b2e9a5db7fba57e5985e572fb6c605731af4bee45d12ddcd6" host="localhost" Mar 20 21:15:40.475873 containerd[1485]: 2025-03-20 21:15:40.460 [INFO][4159] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.130/26] handle="k8s-pod-network.7ec20bb10b38c16b2e9a5db7fba57e5985e572fb6c605731af4bee45d12ddcd6" host="localhost" Mar 20 21:15:40.475873 containerd[1485]: 2025-03-20 21:15:40.460 [INFO][4159] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Mar 20 21:15:40.475873 containerd[1485]: 2025-03-20 21:15:40.460 [INFO][4159] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.130/26] IPv6=[] ContainerID="7ec20bb10b38c16b2e9a5db7fba57e5985e572fb6c605731af4bee45d12ddcd6" HandleID="k8s-pod-network.7ec20bb10b38c16b2e9a5db7fba57e5985e572fb6c605731af4bee45d12ddcd6" Workload="localhost-k8s-calico--kube--controllers--5ddfd69446--sh2ld-eth0" Mar 20 21:15:40.476384 containerd[1485]: 2025-03-20 21:15:40.462 [INFO][4146] cni-plugin/k8s.go 386: Populated endpoint ContainerID="7ec20bb10b38c16b2e9a5db7fba57e5985e572fb6c605731af4bee45d12ddcd6" Namespace="calico-system" Pod="calico-kube-controllers-5ddfd69446-sh2ld" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5ddfd69446--sh2ld-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--5ddfd69446--sh2ld-eth0", GenerateName:"calico-kube-controllers-5ddfd69446-", Namespace:"calico-system", SelfLink:"", UID:"428dbe74-ae1a-409c-8f83-1014b5c9a996", ResourceVersion:"705", Generation:0, CreationTimestamp:time.Date(2025, time.March, 20, 21, 15, 20, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"5ddfd69446", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-kube-controllers-5ddfd69446-sh2ld", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali2c3802e6845", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Mar 20 21:15:40.476384 containerd[1485]: 2025-03-20 21:15:40.462 [INFO][4146] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.130/32] ContainerID="7ec20bb10b38c16b2e9a5db7fba57e5985e572fb6c605731af4bee45d12ddcd6" Namespace="calico-system" Pod="calico-kube-controllers-5ddfd69446-sh2ld" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5ddfd69446--sh2ld-eth0" Mar 20 21:15:40.476384 containerd[1485]: 2025-03-20 21:15:40.462 [INFO][4146] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali2c3802e6845 ContainerID="7ec20bb10b38c16b2e9a5db7fba57e5985e572fb6c605731af4bee45d12ddcd6" Namespace="calico-system" Pod="calico-kube-controllers-5ddfd69446-sh2ld" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5ddfd69446--sh2ld-eth0" Mar 20 21:15:40.476384 containerd[1485]: 2025-03-20 21:15:40.464 [INFO][4146] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="7ec20bb10b38c16b2e9a5db7fba57e5985e572fb6c605731af4bee45d12ddcd6" Namespace="calico-system" Pod="calico-kube-controllers-5ddfd69446-sh2ld" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5ddfd69446--sh2ld-eth0" Mar 20 21:15:40.476384 containerd[1485]: 2025-03-20 21:15:40.465 [INFO][4146] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="7ec20bb10b38c16b2e9a5db7fba57e5985e572fb6c605731af4bee45d12ddcd6" Namespace="calico-system" Pod="calico-kube-controllers-5ddfd69446-sh2ld" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5ddfd69446--sh2ld-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--5ddfd69446--sh2ld-eth0", GenerateName:"calico-kube-controllers-5ddfd69446-", Namespace:"calico-system", SelfLink:"", UID:"428dbe74-ae1a-409c-8f83-1014b5c9a996", ResourceVersion:"705", Generation:0, CreationTimestamp:time.Date(2025, time.March, 20, 21, 15, 20, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"5ddfd69446", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"7ec20bb10b38c16b2e9a5db7fba57e5985e572fb6c605731af4bee45d12ddcd6", Pod:"calico-kube-controllers-5ddfd69446-sh2ld", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali2c3802e6845", MAC:"96:c9:45:9d:ca:de", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Mar 20 21:15:40.476384 containerd[1485]: 2025-03-20 21:15:40.473 [INFO][4146] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="7ec20bb10b38c16b2e9a5db7fba57e5985e572fb6c605731af4bee45d12ddcd6" Namespace="calico-system" Pod="calico-kube-controllers-5ddfd69446-sh2ld" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--5ddfd69446--sh2ld-eth0" Mar 20 21:15:40.512438 containerd[1485]: time="2025-03-20T21:15:40.512393844Z" level=info msg="connecting to shim 7ec20bb10b38c16b2e9a5db7fba57e5985e572fb6c605731af4bee45d12ddcd6" address="unix:///run/containerd/s/027c9d8956e1fa6502c4d92bdd205debca2c7dbdbafb3d0e268f3b25ac9d7e21" namespace=k8s.io protocol=ttrpc version=3 Mar 20 21:15:40.539398 systemd[1]: Started cri-containerd-7ec20bb10b38c16b2e9a5db7fba57e5985e572fb6c605731af4bee45d12ddcd6.scope - libcontainer container 7ec20bb10b38c16b2e9a5db7fba57e5985e572fb6c605731af4bee45d12ddcd6. Mar 20 21:15:40.551351 systemd-resolved[1330]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Mar 20 21:15:40.556190 kubelet[2596]: I0320 21:15:40.555307 2596 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-6f6b679f8f-qxfz8" podStartSLOduration=30.555289516 podStartE2EDuration="30.555289516s" podCreationTimestamp="2025-03-20 21:15:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-20 21:15:40.553461543 +0000 UTC m=+36.284945638" watchObservedRunningTime="2025-03-20 21:15:40.555289516 +0000 UTC m=+36.286773611" Mar 20 21:15:40.585376 containerd[1485]: time="2025-03-20T21:15:40.585326854Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-5ddfd69446-sh2ld,Uid:428dbe74-ae1a-409c-8f83-1014b5c9a996,Namespace:calico-system,Attempt:0,} returns sandbox id \"7ec20bb10b38c16b2e9a5db7fba57e5985e572fb6c605731af4bee45d12ddcd6\"" Mar 20 21:15:40.587597 containerd[1485]: time="2025-03-20T21:15:40.587543430Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.29.2\"" Mar 20 21:15:41.333343 systemd-networkd[1416]: cali4f695bc4646: Gained IPv6LL Mar 20 21:15:41.345118 containerd[1485]: time="2025-03-20T21:15:41.345050655Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-8646c76b8-kbdmr,Uid:80afa69e-2493-484d-9af0-def2e6594d53,Namespace:calico-apiserver,Attempt:0,}" Mar 20 21:15:41.479120 systemd-networkd[1416]: cali57168d9c8d7: Link UP Mar 20 21:15:41.479357 systemd-networkd[1416]: cali57168d9c8d7: Gained carrier Mar 20 21:15:41.506409 containerd[1485]: 2025-03-20 21:15:41.381 [INFO][4233] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--8646c76b8--kbdmr-eth0 calico-apiserver-8646c76b8- calico-apiserver 80afa69e-2493-484d-9af0-def2e6594d53 702 0 2025-03-20 21:15:18 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:8646c76b8 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-8646c76b8-kbdmr eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali57168d9c8d7 [] []}} ContainerID="bb60bacaad0d905abce4377ae3f4cdab4d18db3d6b4c674d817fdcffeb4d4b17" Namespace="calico-apiserver" Pod="calico-apiserver-8646c76b8-kbdmr" WorkloadEndpoint="localhost-k8s-calico--apiserver--8646c76b8--kbdmr-" Mar 20 21:15:41.506409 containerd[1485]: 2025-03-20 21:15:41.381 [INFO][4233] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="bb60bacaad0d905abce4377ae3f4cdab4d18db3d6b4c674d817fdcffeb4d4b17" Namespace="calico-apiserver" Pod="calico-apiserver-8646c76b8-kbdmr" WorkloadEndpoint="localhost-k8s-calico--apiserver--8646c76b8--kbdmr-eth0" Mar 20 21:15:41.506409 containerd[1485]: 2025-03-20 21:15:41.410 [INFO][4246] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="bb60bacaad0d905abce4377ae3f4cdab4d18db3d6b4c674d817fdcffeb4d4b17" HandleID="k8s-pod-network.bb60bacaad0d905abce4377ae3f4cdab4d18db3d6b4c674d817fdcffeb4d4b17" Workload="localhost-k8s-calico--apiserver--8646c76b8--kbdmr-eth0" Mar 20 21:15:41.506409 containerd[1485]: 2025-03-20 21:15:41.428 [INFO][4246] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="bb60bacaad0d905abce4377ae3f4cdab4d18db3d6b4c674d817fdcffeb4d4b17" HandleID="k8s-pod-network.bb60bacaad0d905abce4377ae3f4cdab4d18db3d6b4c674d817fdcffeb4d4b17" Workload="localhost-k8s-calico--apiserver--8646c76b8--kbdmr-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000313610), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-8646c76b8-kbdmr", "timestamp":"2025-03-20 21:15:41.410261418 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Mar 20 21:15:41.506409 containerd[1485]: 2025-03-20 21:15:41.428 [INFO][4246] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Mar 20 21:15:41.506409 containerd[1485]: 2025-03-20 21:15:41.428 [INFO][4246] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Mar 20 21:15:41.506409 containerd[1485]: 2025-03-20 21:15:41.428 [INFO][4246] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Mar 20 21:15:41.506409 containerd[1485]: 2025-03-20 21:15:41.432 [INFO][4246] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.bb60bacaad0d905abce4377ae3f4cdab4d18db3d6b4c674d817fdcffeb4d4b17" host="localhost" Mar 20 21:15:41.506409 containerd[1485]: 2025-03-20 21:15:41.437 [INFO][4246] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" Mar 20 21:15:41.506409 containerd[1485]: 2025-03-20 21:15:41.441 [INFO][4246] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Mar 20 21:15:41.506409 containerd[1485]: 2025-03-20 21:15:41.443 [INFO][4246] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Mar 20 21:15:41.506409 containerd[1485]: 2025-03-20 21:15:41.446 [INFO][4246] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Mar 20 21:15:41.506409 containerd[1485]: 2025-03-20 21:15:41.446 [INFO][4246] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.bb60bacaad0d905abce4377ae3f4cdab4d18db3d6b4c674d817fdcffeb4d4b17" host="localhost" Mar 20 21:15:41.506409 containerd[1485]: 2025-03-20 21:15:41.447 [INFO][4246] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.bb60bacaad0d905abce4377ae3f4cdab4d18db3d6b4c674d817fdcffeb4d4b17 Mar 20 21:15:41.506409 containerd[1485]: 2025-03-20 21:15:41.456 [INFO][4246] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.bb60bacaad0d905abce4377ae3f4cdab4d18db3d6b4c674d817fdcffeb4d4b17" host="localhost" Mar 20 21:15:41.506409 containerd[1485]: 2025-03-20 21:15:41.475 [INFO][4246] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.131/26] block=192.168.88.128/26 handle="k8s-pod-network.bb60bacaad0d905abce4377ae3f4cdab4d18db3d6b4c674d817fdcffeb4d4b17" host="localhost" Mar 20 21:15:41.506409 containerd[1485]: 2025-03-20 21:15:41.475 [INFO][4246] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.131/26] handle="k8s-pod-network.bb60bacaad0d905abce4377ae3f4cdab4d18db3d6b4c674d817fdcffeb4d4b17" host="localhost" Mar 20 21:15:41.506409 containerd[1485]: 2025-03-20 21:15:41.475 [INFO][4246] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Mar 20 21:15:41.506409 containerd[1485]: 2025-03-20 21:15:41.475 [INFO][4246] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.131/26] IPv6=[] ContainerID="bb60bacaad0d905abce4377ae3f4cdab4d18db3d6b4c674d817fdcffeb4d4b17" HandleID="k8s-pod-network.bb60bacaad0d905abce4377ae3f4cdab4d18db3d6b4c674d817fdcffeb4d4b17" Workload="localhost-k8s-calico--apiserver--8646c76b8--kbdmr-eth0" Mar 20 21:15:41.506987 containerd[1485]: 2025-03-20 21:15:41.477 [INFO][4233] cni-plugin/k8s.go 386: Populated endpoint ContainerID="bb60bacaad0d905abce4377ae3f4cdab4d18db3d6b4c674d817fdcffeb4d4b17" Namespace="calico-apiserver" Pod="calico-apiserver-8646c76b8-kbdmr" WorkloadEndpoint="localhost-k8s-calico--apiserver--8646c76b8--kbdmr-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--8646c76b8--kbdmr-eth0", GenerateName:"calico-apiserver-8646c76b8-", Namespace:"calico-apiserver", SelfLink:"", UID:"80afa69e-2493-484d-9af0-def2e6594d53", ResourceVersion:"702", Generation:0, CreationTimestamp:time.Date(2025, time.March, 20, 21, 15, 18, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"8646c76b8", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-8646c76b8-kbdmr", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali57168d9c8d7", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Mar 20 21:15:41.506987 containerd[1485]: 2025-03-20 21:15:41.477 [INFO][4233] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.131/32] ContainerID="bb60bacaad0d905abce4377ae3f4cdab4d18db3d6b4c674d817fdcffeb4d4b17" Namespace="calico-apiserver" Pod="calico-apiserver-8646c76b8-kbdmr" WorkloadEndpoint="localhost-k8s-calico--apiserver--8646c76b8--kbdmr-eth0" Mar 20 21:15:41.506987 containerd[1485]: 2025-03-20 21:15:41.477 [INFO][4233] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali57168d9c8d7 ContainerID="bb60bacaad0d905abce4377ae3f4cdab4d18db3d6b4c674d817fdcffeb4d4b17" Namespace="calico-apiserver" Pod="calico-apiserver-8646c76b8-kbdmr" WorkloadEndpoint="localhost-k8s-calico--apiserver--8646c76b8--kbdmr-eth0" Mar 20 21:15:41.506987 containerd[1485]: 2025-03-20 21:15:41.479 [INFO][4233] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="bb60bacaad0d905abce4377ae3f4cdab4d18db3d6b4c674d817fdcffeb4d4b17" Namespace="calico-apiserver" Pod="calico-apiserver-8646c76b8-kbdmr" WorkloadEndpoint="localhost-k8s-calico--apiserver--8646c76b8--kbdmr-eth0" Mar 20 21:15:41.506987 containerd[1485]: 2025-03-20 21:15:41.480 [INFO][4233] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="bb60bacaad0d905abce4377ae3f4cdab4d18db3d6b4c674d817fdcffeb4d4b17" Namespace="calico-apiserver" Pod="calico-apiserver-8646c76b8-kbdmr" WorkloadEndpoint="localhost-k8s-calico--apiserver--8646c76b8--kbdmr-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--8646c76b8--kbdmr-eth0", GenerateName:"calico-apiserver-8646c76b8-", Namespace:"calico-apiserver", SelfLink:"", UID:"80afa69e-2493-484d-9af0-def2e6594d53", ResourceVersion:"702", Generation:0, CreationTimestamp:time.Date(2025, time.March, 20, 21, 15, 18, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"8646c76b8", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"bb60bacaad0d905abce4377ae3f4cdab4d18db3d6b4c674d817fdcffeb4d4b17", Pod:"calico-apiserver-8646c76b8-kbdmr", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali57168d9c8d7", MAC:"1e:9d:60:0c:96:27", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Mar 20 21:15:41.506987 containerd[1485]: 2025-03-20 21:15:41.504 [INFO][4233] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="bb60bacaad0d905abce4377ae3f4cdab4d18db3d6b4c674d817fdcffeb4d4b17" Namespace="calico-apiserver" Pod="calico-apiserver-8646c76b8-kbdmr" WorkloadEndpoint="localhost-k8s-calico--apiserver--8646c76b8--kbdmr-eth0" Mar 20 21:15:41.546400 containerd[1485]: time="2025-03-20T21:15:41.546191064Z" level=info msg="connecting to shim bb60bacaad0d905abce4377ae3f4cdab4d18db3d6b4c674d817fdcffeb4d4b17" address="unix:///run/containerd/s/a1aa89c8fa4955b52fc9c67dc3d6b9fdfe1af4d6ebdd6064071db1bc6c71bd3c" namespace=k8s.io protocol=ttrpc version=3 Mar 20 21:15:41.583413 systemd[1]: Started cri-containerd-bb60bacaad0d905abce4377ae3f4cdab4d18db3d6b4c674d817fdcffeb4d4b17.scope - libcontainer container bb60bacaad0d905abce4377ae3f4cdab4d18db3d6b4c674d817fdcffeb4d4b17. Mar 20 21:15:41.597507 systemd-resolved[1330]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Mar 20 21:15:41.623603 containerd[1485]: time="2025-03-20T21:15:41.623560910Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-8646c76b8-kbdmr,Uid:80afa69e-2493-484d-9af0-def2e6594d53,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"bb60bacaad0d905abce4377ae3f4cdab4d18db3d6b4c674d817fdcffeb4d4b17\"" Mar 20 21:15:41.781343 systemd-networkd[1416]: cali2c3802e6845: Gained IPv6LL Mar 20 21:15:41.968998 containerd[1485]: time="2025-03-20T21:15:41.968864260Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.29.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:15:41.969562 containerd[1485]: time="2025-03-20T21:15:41.969509825Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.29.2: active requests=0, bytes read=32560257" Mar 20 21:15:41.970196 containerd[1485]: time="2025-03-20T21:15:41.970165589Z" level=info msg="ImageCreate event name:\"sha256:39a6e91a11a792441d34dccf5e11416a0fd297782f169fdb871a5558ad50b229\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:15:41.972112 containerd[1485]: time="2025-03-20T21:15:41.972058282Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:6d1f392b747f912366ec5c60ee1130952c2c07e8ce24c53480187daa0e3364aa\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:15:41.972828 containerd[1485]: time="2025-03-20T21:15:41.972594086Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.29.2\" with image id \"sha256:39a6e91a11a792441d34dccf5e11416a0fd297782f169fdb871a5558ad50b229\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.29.2\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:6d1f392b747f912366ec5c60ee1130952c2c07e8ce24c53480187daa0e3364aa\", size \"33929982\" in 1.385011136s" Mar 20 21:15:41.972828 containerd[1485]: time="2025-03-20T21:15:41.972630886Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.29.2\" returns image reference \"sha256:39a6e91a11a792441d34dccf5e11416a0fd297782f169fdb871a5558ad50b229\"" Mar 20 21:15:41.973979 containerd[1485]: time="2025-03-20T21:15:41.973949055Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.2\"" Mar 20 21:15:41.983933 containerd[1485]: time="2025-03-20T21:15:41.983863282Z" level=info msg="CreateContainer within sandbox \"7ec20bb10b38c16b2e9a5db7fba57e5985e572fb6c605731af4bee45d12ddcd6\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Mar 20 21:15:41.990259 containerd[1485]: time="2025-03-20T21:15:41.989965604Z" level=info msg="Container c14e33d2f5b60700e68dcfd113314dd88a94ba5b5df327532409a6162dc44321: CDI devices from CRI Config.CDIDevices: []" Mar 20 21:15:41.995666 containerd[1485]: time="2025-03-20T21:15:41.995632202Z" level=info msg="CreateContainer within sandbox \"7ec20bb10b38c16b2e9a5db7fba57e5985e572fb6c605731af4bee45d12ddcd6\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"c14e33d2f5b60700e68dcfd113314dd88a94ba5b5df327532409a6162dc44321\"" Mar 20 21:15:41.996085 containerd[1485]: time="2025-03-20T21:15:41.996060445Z" level=info msg="StartContainer for \"c14e33d2f5b60700e68dcfd113314dd88a94ba5b5df327532409a6162dc44321\"" Mar 20 21:15:41.997087 containerd[1485]: time="2025-03-20T21:15:41.997043212Z" level=info msg="connecting to shim c14e33d2f5b60700e68dcfd113314dd88a94ba5b5df327532409a6162dc44321" address="unix:///run/containerd/s/027c9d8956e1fa6502c4d92bdd205debca2c7dbdbafb3d0e268f3b25ac9d7e21" protocol=ttrpc version=3 Mar 20 21:15:42.020450 systemd[1]: Started cri-containerd-c14e33d2f5b60700e68dcfd113314dd88a94ba5b5df327532409a6162dc44321.scope - libcontainer container c14e33d2f5b60700e68dcfd113314dd88a94ba5b5df327532409a6162dc44321. Mar 20 21:15:42.087549 containerd[1485]: time="2025-03-20T21:15:42.087496151Z" level=info msg="StartContainer for \"c14e33d2f5b60700e68dcfd113314dd88a94ba5b5df327532409a6162dc44321\" returns successfully" Mar 20 21:15:42.561449 kubelet[2596]: I0320 21:15:42.561156 2596 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-5ddfd69446-sh2ld" podStartSLOduration=21.174386026 podStartE2EDuration="22.561139133s" podCreationTimestamp="2025-03-20 21:15:20 +0000 UTC" firstStartedPulling="2025-03-20 21:15:40.586548103 +0000 UTC m=+36.318032198" lastFinishedPulling="2025-03-20 21:15:41.97330117 +0000 UTC m=+37.704785305" observedRunningTime="2025-03-20 21:15:42.560129487 +0000 UTC m=+38.291613622" watchObservedRunningTime="2025-03-20 21:15:42.561139133 +0000 UTC m=+38.292623268" Mar 20 21:15:43.344494 containerd[1485]: time="2025-03-20T21:15:43.344454474Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-mbg5n,Uid:e1b822dc-3680-44ce-8c5d-15c8549a9319,Namespace:kube-system,Attempt:0,}" Mar 20 21:15:43.345295 containerd[1485]: time="2025-03-20T21:15:43.345107798Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-8646c76b8-xnq47,Uid:42908acd-daa0-4843-a9d2-569c24a392b7,Namespace:calico-apiserver,Attempt:0,}" Mar 20 21:15:43.431884 systemd[1]: Started sshd@9-10.0.0.54:22-10.0.0.1:45582.service - OpenSSH per-connection server daemon (10.0.0.1:45582). Mar 20 21:15:43.509388 systemd-networkd[1416]: cali57168d9c8d7: Gained IPv6LL Mar 20 21:15:43.520004 sshd[4409]: Accepted publickey for core from 10.0.0.1 port 45582 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:15:43.525417 sshd-session[4409]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:15:43.526360 systemd-networkd[1416]: califf0ac8946a0: Link UP Mar 20 21:15:43.527412 systemd-networkd[1416]: califf0ac8946a0: Gained carrier Mar 20 21:15:43.536134 systemd-logind[1470]: New session 10 of user core. Mar 20 21:15:43.542165 systemd[1]: Started session-10.scope - Session 10 of User core. Mar 20 21:15:43.551926 containerd[1485]: 2025-03-20 21:15:43.397 [INFO][4366] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--8646c76b8--xnq47-eth0 calico-apiserver-8646c76b8- calico-apiserver 42908acd-daa0-4843-a9d2-569c24a392b7 704 0 2025-03-20 21:15:18 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:8646c76b8 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-8646c76b8-xnq47 eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] califf0ac8946a0 [] []}} ContainerID="fdf3da8083f926e45241c057e9ea9678fb84411bc46b23959743ff765df8f1a5" Namespace="calico-apiserver" Pod="calico-apiserver-8646c76b8-xnq47" WorkloadEndpoint="localhost-k8s-calico--apiserver--8646c76b8--xnq47-" Mar 20 21:15:43.551926 containerd[1485]: 2025-03-20 21:15:43.398 [INFO][4366] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="fdf3da8083f926e45241c057e9ea9678fb84411bc46b23959743ff765df8f1a5" Namespace="calico-apiserver" Pod="calico-apiserver-8646c76b8-xnq47" WorkloadEndpoint="localhost-k8s-calico--apiserver--8646c76b8--xnq47-eth0" Mar 20 21:15:43.551926 containerd[1485]: 2025-03-20 21:15:43.448 [INFO][4397] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="fdf3da8083f926e45241c057e9ea9678fb84411bc46b23959743ff765df8f1a5" HandleID="k8s-pod-network.fdf3da8083f926e45241c057e9ea9678fb84411bc46b23959743ff765df8f1a5" Workload="localhost-k8s-calico--apiserver--8646c76b8--xnq47-eth0" Mar 20 21:15:43.551926 containerd[1485]: 2025-03-20 21:15:43.467 [INFO][4397] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="fdf3da8083f926e45241c057e9ea9678fb84411bc46b23959743ff765df8f1a5" HandleID="k8s-pod-network.fdf3da8083f926e45241c057e9ea9678fb84411bc46b23959743ff765df8f1a5" Workload="localhost-k8s-calico--apiserver--8646c76b8--xnq47-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002e76e0), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-8646c76b8-xnq47", "timestamp":"2025-03-20 21:15:43.448767778 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Mar 20 21:15:43.551926 containerd[1485]: 2025-03-20 21:15:43.467 [INFO][4397] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Mar 20 21:15:43.551926 containerd[1485]: 2025-03-20 21:15:43.467 [INFO][4397] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Mar 20 21:15:43.551926 containerd[1485]: 2025-03-20 21:15:43.467 [INFO][4397] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Mar 20 21:15:43.551926 containerd[1485]: 2025-03-20 21:15:43.471 [INFO][4397] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.fdf3da8083f926e45241c057e9ea9678fb84411bc46b23959743ff765df8f1a5" host="localhost" Mar 20 21:15:43.551926 containerd[1485]: 2025-03-20 21:15:43.476 [INFO][4397] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" Mar 20 21:15:43.551926 containerd[1485]: 2025-03-20 21:15:43.485 [INFO][4397] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Mar 20 21:15:43.551926 containerd[1485]: 2025-03-20 21:15:43.487 [INFO][4397] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Mar 20 21:15:43.551926 containerd[1485]: 2025-03-20 21:15:43.490 [INFO][4397] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Mar 20 21:15:43.551926 containerd[1485]: 2025-03-20 21:15:43.490 [INFO][4397] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.fdf3da8083f926e45241c057e9ea9678fb84411bc46b23959743ff765df8f1a5" host="localhost" Mar 20 21:15:43.551926 containerd[1485]: 2025-03-20 21:15:43.491 [INFO][4397] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.fdf3da8083f926e45241c057e9ea9678fb84411bc46b23959743ff765df8f1a5 Mar 20 21:15:43.551926 containerd[1485]: 2025-03-20 21:15:43.496 [INFO][4397] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.fdf3da8083f926e45241c057e9ea9678fb84411bc46b23959743ff765df8f1a5" host="localhost" Mar 20 21:15:43.551926 containerd[1485]: 2025-03-20 21:15:43.503 [INFO][4397] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.132/26] block=192.168.88.128/26 handle="k8s-pod-network.fdf3da8083f926e45241c057e9ea9678fb84411bc46b23959743ff765df8f1a5" host="localhost" Mar 20 21:15:43.551926 containerd[1485]: 2025-03-20 21:15:43.503 [INFO][4397] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.132/26] handle="k8s-pod-network.fdf3da8083f926e45241c057e9ea9678fb84411bc46b23959743ff765df8f1a5" host="localhost" Mar 20 21:15:43.551926 containerd[1485]: 2025-03-20 21:15:43.503 [INFO][4397] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Mar 20 21:15:43.551926 containerd[1485]: 2025-03-20 21:15:43.503 [INFO][4397] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.132/26] IPv6=[] ContainerID="fdf3da8083f926e45241c057e9ea9678fb84411bc46b23959743ff765df8f1a5" HandleID="k8s-pod-network.fdf3da8083f926e45241c057e9ea9678fb84411bc46b23959743ff765df8f1a5" Workload="localhost-k8s-calico--apiserver--8646c76b8--xnq47-eth0" Mar 20 21:15:43.552466 containerd[1485]: 2025-03-20 21:15:43.511 [INFO][4366] cni-plugin/k8s.go 386: Populated endpoint ContainerID="fdf3da8083f926e45241c057e9ea9678fb84411bc46b23959743ff765df8f1a5" Namespace="calico-apiserver" Pod="calico-apiserver-8646c76b8-xnq47" WorkloadEndpoint="localhost-k8s-calico--apiserver--8646c76b8--xnq47-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--8646c76b8--xnq47-eth0", GenerateName:"calico-apiserver-8646c76b8-", Namespace:"calico-apiserver", SelfLink:"", UID:"42908acd-daa0-4843-a9d2-569c24a392b7", ResourceVersion:"704", Generation:0, CreationTimestamp:time.Date(2025, time.March, 20, 21, 15, 18, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"8646c76b8", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-8646c76b8-xnq47", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"califf0ac8946a0", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Mar 20 21:15:43.552466 containerd[1485]: 2025-03-20 21:15:43.512 [INFO][4366] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.132/32] ContainerID="fdf3da8083f926e45241c057e9ea9678fb84411bc46b23959743ff765df8f1a5" Namespace="calico-apiserver" Pod="calico-apiserver-8646c76b8-xnq47" WorkloadEndpoint="localhost-k8s-calico--apiserver--8646c76b8--xnq47-eth0" Mar 20 21:15:43.552466 containerd[1485]: 2025-03-20 21:15:43.512 [INFO][4366] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to califf0ac8946a0 ContainerID="fdf3da8083f926e45241c057e9ea9678fb84411bc46b23959743ff765df8f1a5" Namespace="calico-apiserver" Pod="calico-apiserver-8646c76b8-xnq47" WorkloadEndpoint="localhost-k8s-calico--apiserver--8646c76b8--xnq47-eth0" Mar 20 21:15:43.552466 containerd[1485]: 2025-03-20 21:15:43.528 [INFO][4366] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="fdf3da8083f926e45241c057e9ea9678fb84411bc46b23959743ff765df8f1a5" Namespace="calico-apiserver" Pod="calico-apiserver-8646c76b8-xnq47" WorkloadEndpoint="localhost-k8s-calico--apiserver--8646c76b8--xnq47-eth0" Mar 20 21:15:43.552466 containerd[1485]: 2025-03-20 21:15:43.529 [INFO][4366] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="fdf3da8083f926e45241c057e9ea9678fb84411bc46b23959743ff765df8f1a5" Namespace="calico-apiserver" Pod="calico-apiserver-8646c76b8-xnq47" WorkloadEndpoint="localhost-k8s-calico--apiserver--8646c76b8--xnq47-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--8646c76b8--xnq47-eth0", GenerateName:"calico-apiserver-8646c76b8-", Namespace:"calico-apiserver", SelfLink:"", UID:"42908acd-daa0-4843-a9d2-569c24a392b7", ResourceVersion:"704", Generation:0, CreationTimestamp:time.Date(2025, time.March, 20, 21, 15, 18, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"8646c76b8", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"fdf3da8083f926e45241c057e9ea9678fb84411bc46b23959743ff765df8f1a5", Pod:"calico-apiserver-8646c76b8-xnq47", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"califf0ac8946a0", MAC:"3e:98:4b:ff:af:ea", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Mar 20 21:15:43.552466 containerd[1485]: 2025-03-20 21:15:43.545 [INFO][4366] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="fdf3da8083f926e45241c057e9ea9678fb84411bc46b23959743ff765df8f1a5" Namespace="calico-apiserver" Pod="calico-apiserver-8646c76b8-xnq47" WorkloadEndpoint="localhost-k8s-calico--apiserver--8646c76b8--xnq47-eth0" Mar 20 21:15:43.558273 kubelet[2596]: I0320 21:15:43.557389 2596 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 20 21:15:43.609006 containerd[1485]: time="2025-03-20T21:15:43.608387053Z" level=info msg="connecting to shim fdf3da8083f926e45241c057e9ea9678fb84411bc46b23959743ff765df8f1a5" address="unix:///run/containerd/s/c249e1229128cbe39c88ed9cd98c8b569b1212095b5ef7b7b0840820bba2ec25" namespace=k8s.io protocol=ttrpc version=3 Mar 20 21:15:43.633168 systemd-networkd[1416]: cali9e5c282701d: Link UP Mar 20 21:15:43.633339 systemd-networkd[1416]: cali9e5c282701d: Gained carrier Mar 20 21:15:43.648422 containerd[1485]: 2025-03-20 21:15:43.402 [INFO][4363] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--6f6b679f8f--mbg5n-eth0 coredns-6f6b679f8f- kube-system e1b822dc-3680-44ce-8c5d-15c8549a9319 703 0 2025-03-20 21:15:10 +0000 UTC map[k8s-app:kube-dns pod-template-hash:6f6b679f8f projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-6f6b679f8f-mbg5n eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali9e5c282701d [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="42bb7f7cb8701cdfea0cb1630e5d4e3ee1e28fa6aca4dab6ff1fc6e9e3affaef" Namespace="kube-system" Pod="coredns-6f6b679f8f-mbg5n" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--mbg5n-" Mar 20 21:15:43.648422 containerd[1485]: 2025-03-20 21:15:43.402 [INFO][4363] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="42bb7f7cb8701cdfea0cb1630e5d4e3ee1e28fa6aca4dab6ff1fc6e9e3affaef" Namespace="kube-system" Pod="coredns-6f6b679f8f-mbg5n" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--mbg5n-eth0" Mar 20 21:15:43.648422 containerd[1485]: 2025-03-20 21:15:43.448 [INFO][4395] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="42bb7f7cb8701cdfea0cb1630e5d4e3ee1e28fa6aca4dab6ff1fc6e9e3affaef" HandleID="k8s-pod-network.42bb7f7cb8701cdfea0cb1630e5d4e3ee1e28fa6aca4dab6ff1fc6e9e3affaef" Workload="localhost-k8s-coredns--6f6b679f8f--mbg5n-eth0" Mar 20 21:15:43.648422 containerd[1485]: 2025-03-20 21:15:43.466 [INFO][4395] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="42bb7f7cb8701cdfea0cb1630e5d4e3ee1e28fa6aca4dab6ff1fc6e9e3affaef" HandleID="k8s-pod-network.42bb7f7cb8701cdfea0cb1630e5d4e3ee1e28fa6aca4dab6ff1fc6e9e3affaef" Workload="localhost-k8s-coredns--6f6b679f8f--mbg5n-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40003e00b0), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-6f6b679f8f-mbg5n", "timestamp":"2025-03-20 21:15:43.448674818 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Mar 20 21:15:43.648422 containerd[1485]: 2025-03-20 21:15:43.467 [INFO][4395] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Mar 20 21:15:43.648422 containerd[1485]: 2025-03-20 21:15:43.505 [INFO][4395] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Mar 20 21:15:43.648422 containerd[1485]: 2025-03-20 21:15:43.505 [INFO][4395] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Mar 20 21:15:43.648422 containerd[1485]: 2025-03-20 21:15:43.573 [INFO][4395] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.42bb7f7cb8701cdfea0cb1630e5d4e3ee1e28fa6aca4dab6ff1fc6e9e3affaef" host="localhost" Mar 20 21:15:43.648422 containerd[1485]: 2025-03-20 21:15:43.582 [INFO][4395] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" Mar 20 21:15:43.648422 containerd[1485]: 2025-03-20 21:15:43.599 [INFO][4395] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Mar 20 21:15:43.648422 containerd[1485]: 2025-03-20 21:15:43.606 [INFO][4395] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Mar 20 21:15:43.648422 containerd[1485]: 2025-03-20 21:15:43.610 [INFO][4395] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Mar 20 21:15:43.648422 containerd[1485]: 2025-03-20 21:15:43.610 [INFO][4395] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.42bb7f7cb8701cdfea0cb1630e5d4e3ee1e28fa6aca4dab6ff1fc6e9e3affaef" host="localhost" Mar 20 21:15:43.648422 containerd[1485]: 2025-03-20 21:15:43.613 [INFO][4395] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.42bb7f7cb8701cdfea0cb1630e5d4e3ee1e28fa6aca4dab6ff1fc6e9e3affaef Mar 20 21:15:43.648422 containerd[1485]: 2025-03-20 21:15:43.621 [INFO][4395] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.42bb7f7cb8701cdfea0cb1630e5d4e3ee1e28fa6aca4dab6ff1fc6e9e3affaef" host="localhost" Mar 20 21:15:43.648422 containerd[1485]: 2025-03-20 21:15:43.627 [INFO][4395] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.133/26] block=192.168.88.128/26 handle="k8s-pod-network.42bb7f7cb8701cdfea0cb1630e5d4e3ee1e28fa6aca4dab6ff1fc6e9e3affaef" host="localhost" Mar 20 21:15:43.648422 containerd[1485]: 2025-03-20 21:15:43.627 [INFO][4395] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.133/26] handle="k8s-pod-network.42bb7f7cb8701cdfea0cb1630e5d4e3ee1e28fa6aca4dab6ff1fc6e9e3affaef" host="localhost" Mar 20 21:15:43.648422 containerd[1485]: 2025-03-20 21:15:43.627 [INFO][4395] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Mar 20 21:15:43.648422 containerd[1485]: 2025-03-20 21:15:43.627 [INFO][4395] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.133/26] IPv6=[] ContainerID="42bb7f7cb8701cdfea0cb1630e5d4e3ee1e28fa6aca4dab6ff1fc6e9e3affaef" HandleID="k8s-pod-network.42bb7f7cb8701cdfea0cb1630e5d4e3ee1e28fa6aca4dab6ff1fc6e9e3affaef" Workload="localhost-k8s-coredns--6f6b679f8f--mbg5n-eth0" Mar 20 21:15:43.648944 containerd[1485]: 2025-03-20 21:15:43.630 [INFO][4363] cni-plugin/k8s.go 386: Populated endpoint ContainerID="42bb7f7cb8701cdfea0cb1630e5d4e3ee1e28fa6aca4dab6ff1fc6e9e3affaef" Namespace="kube-system" Pod="coredns-6f6b679f8f-mbg5n" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--mbg5n-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--6f6b679f8f--mbg5n-eth0", GenerateName:"coredns-6f6b679f8f-", Namespace:"kube-system", SelfLink:"", UID:"e1b822dc-3680-44ce-8c5d-15c8549a9319", ResourceVersion:"703", Generation:0, CreationTimestamp:time.Date(2025, time.March, 20, 21, 15, 10, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"6f6b679f8f", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-6f6b679f8f-mbg5n", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali9e5c282701d", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Mar 20 21:15:43.648944 containerd[1485]: 2025-03-20 21:15:43.631 [INFO][4363] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.133/32] ContainerID="42bb7f7cb8701cdfea0cb1630e5d4e3ee1e28fa6aca4dab6ff1fc6e9e3affaef" Namespace="kube-system" Pod="coredns-6f6b679f8f-mbg5n" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--mbg5n-eth0" Mar 20 21:15:43.648944 containerd[1485]: 2025-03-20 21:15:43.631 [INFO][4363] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali9e5c282701d ContainerID="42bb7f7cb8701cdfea0cb1630e5d4e3ee1e28fa6aca4dab6ff1fc6e9e3affaef" Namespace="kube-system" Pod="coredns-6f6b679f8f-mbg5n" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--mbg5n-eth0" Mar 20 21:15:43.648944 containerd[1485]: 2025-03-20 21:15:43.632 [INFO][4363] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="42bb7f7cb8701cdfea0cb1630e5d4e3ee1e28fa6aca4dab6ff1fc6e9e3affaef" Namespace="kube-system" Pod="coredns-6f6b679f8f-mbg5n" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--mbg5n-eth0" Mar 20 21:15:43.648944 containerd[1485]: 2025-03-20 21:15:43.632 [INFO][4363] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="42bb7f7cb8701cdfea0cb1630e5d4e3ee1e28fa6aca4dab6ff1fc6e9e3affaef" Namespace="kube-system" Pod="coredns-6f6b679f8f-mbg5n" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--mbg5n-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--6f6b679f8f--mbg5n-eth0", GenerateName:"coredns-6f6b679f8f-", Namespace:"kube-system", SelfLink:"", UID:"e1b822dc-3680-44ce-8c5d-15c8549a9319", ResourceVersion:"703", Generation:0, CreationTimestamp:time.Date(2025, time.March, 20, 21, 15, 10, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"6f6b679f8f", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"42bb7f7cb8701cdfea0cb1630e5d4e3ee1e28fa6aca4dab6ff1fc6e9e3affaef", Pod:"coredns-6f6b679f8f-mbg5n", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali9e5c282701d", MAC:"26:45:69:82:0b:81", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Mar 20 21:15:43.648944 containerd[1485]: 2025-03-20 21:15:43.646 [INFO][4363] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="42bb7f7cb8701cdfea0cb1630e5d4e3ee1e28fa6aca4dab6ff1fc6e9e3affaef" Namespace="kube-system" Pod="coredns-6f6b679f8f-mbg5n" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--mbg5n-eth0" Mar 20 21:15:43.672493 systemd[1]: Started cri-containerd-fdf3da8083f926e45241c057e9ea9678fb84411bc46b23959743ff765df8f1a5.scope - libcontainer container fdf3da8083f926e45241c057e9ea9678fb84411bc46b23959743ff765df8f1a5. Mar 20 21:15:43.688535 containerd[1485]: time="2025-03-20T21:15:43.687709648Z" level=info msg="connecting to shim 42bb7f7cb8701cdfea0cb1630e5d4e3ee1e28fa6aca4dab6ff1fc6e9e3affaef" address="unix:///run/containerd/s/29bbeb68a7d545ceda2f4a943177828a74e21428ebd464059d26a1f798a6be39" namespace=k8s.io protocol=ttrpc version=3 Mar 20 21:15:43.717300 systemd-resolved[1330]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Mar 20 21:15:43.721458 systemd[1]: Started cri-containerd-42bb7f7cb8701cdfea0cb1630e5d4e3ee1e28fa6aca4dab6ff1fc6e9e3affaef.scope - libcontainer container 42bb7f7cb8701cdfea0cb1630e5d4e3ee1e28fa6aca4dab6ff1fc6e9e3affaef. Mar 20 21:15:43.749922 systemd-resolved[1330]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Mar 20 21:15:43.775232 containerd[1485]: time="2025-03-20T21:15:43.775153651Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-8646c76b8-xnq47,Uid:42908acd-daa0-4843-a9d2-569c24a392b7,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"fdf3da8083f926e45241c057e9ea9678fb84411bc46b23959743ff765df8f1a5\"" Mar 20 21:15:43.781510 sshd[4421]: Connection closed by 10.0.0.1 port 45582 Mar 20 21:15:43.782611 sshd-session[4409]: pam_unix(sshd:session): session closed for user core Mar 20 21:15:43.785199 containerd[1485]: time="2025-03-20T21:15:43.785087790Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-mbg5n,Uid:e1b822dc-3680-44ce-8c5d-15c8549a9319,Namespace:kube-system,Attempt:0,} returns sandbox id \"42bb7f7cb8701cdfea0cb1630e5d4e3ee1e28fa6aca4dab6ff1fc6e9e3affaef\"" Mar 20 21:15:43.788165 containerd[1485]: time="2025-03-20T21:15:43.788139808Z" level=info msg="CreateContainer within sandbox \"42bb7f7cb8701cdfea0cb1630e5d4e3ee1e28fa6aca4dab6ff1fc6e9e3affaef\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Mar 20 21:15:43.795829 systemd[1]: Started sshd@10-10.0.0.54:22-10.0.0.1:45588.service - OpenSSH per-connection server daemon (10.0.0.1:45588). Mar 20 21:15:43.797107 systemd[1]: sshd@9-10.0.0.54:22-10.0.0.1:45582.service: Deactivated successfully. Mar 20 21:15:43.804288 containerd[1485]: time="2025-03-20T21:15:43.804240785Z" level=info msg="Container a1989d9231126907f6d0d25a2e6a98668241bee10dd196648d27bc57a2dcde5b: CDI devices from CRI Config.CDIDevices: []" Mar 20 21:15:43.804775 systemd[1]: session-10.scope: Deactivated successfully. Mar 20 21:15:43.807461 systemd-logind[1470]: Session 10 logged out. Waiting for processes to exit. Mar 20 21:15:43.810677 containerd[1485]: time="2025-03-20T21:15:43.810542342Z" level=info msg="CreateContainer within sandbox \"42bb7f7cb8701cdfea0cb1630e5d4e3ee1e28fa6aca4dab6ff1fc6e9e3affaef\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"a1989d9231126907f6d0d25a2e6a98668241bee10dd196648d27bc57a2dcde5b\"" Mar 20 21:15:43.812012 containerd[1485]: time="2025-03-20T21:15:43.811638589Z" level=info msg="StartContainer for \"a1989d9231126907f6d0d25a2e6a98668241bee10dd196648d27bc57a2dcde5b\"" Mar 20 21:15:43.812635 systemd-logind[1470]: Removed session 10. Mar 20 21:15:43.813289 containerd[1485]: time="2025-03-20T21:15:43.812751516Z" level=info msg="connecting to shim a1989d9231126907f6d0d25a2e6a98668241bee10dd196648d27bc57a2dcde5b" address="unix:///run/containerd/s/29bbeb68a7d545ceda2f4a943177828a74e21428ebd464059d26a1f798a6be39" protocol=ttrpc version=3 Mar 20 21:15:43.832616 systemd[1]: Started cri-containerd-a1989d9231126907f6d0d25a2e6a98668241bee10dd196648d27bc57a2dcde5b.scope - libcontainer container a1989d9231126907f6d0d25a2e6a98668241bee10dd196648d27bc57a2dcde5b. Mar 20 21:15:43.855072 sshd[4550]: Accepted publickey for core from 10.0.0.1 port 45588 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:15:43.857074 sshd-session[4550]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:15:43.863247 systemd-logind[1470]: New session 11 of user core. Mar 20 21:15:43.868348 systemd[1]: Started session-11.scope - Session 11 of User core. Mar 20 21:15:43.869047 containerd[1485]: time="2025-03-20T21:15:43.869004692Z" level=info msg="StartContainer for \"a1989d9231126907f6d0d25a2e6a98668241bee10dd196648d27bc57a2dcde5b\" returns successfully" Mar 20 21:15:44.047963 containerd[1485]: time="2025-03-20T21:15:44.047910625Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.29.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:15:44.048643 containerd[1485]: time="2025-03-20T21:15:44.048369107Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.29.2: active requests=0, bytes read=40253267" Mar 20 21:15:44.049411 containerd[1485]: time="2025-03-20T21:15:44.049351873Z" level=info msg="ImageCreate event name:\"sha256:15defb01cf01d9d97dc594b25d63dee89192c67a6c991b6a78d49fa834325f4e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:15:44.051509 containerd[1485]: time="2025-03-20T21:15:44.051472045Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:3623f5b60fad0da3387a8649371b53171a4b1226f4d989d2acad9145dc0ef56f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:15:44.052329 containerd[1485]: time="2025-03-20T21:15:44.052189209Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.29.2\" with image id \"sha256:15defb01cf01d9d97dc594b25d63dee89192c67a6c991b6a78d49fa834325f4e\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.29.2\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:3623f5b60fad0da3387a8649371b53171a4b1226f4d989d2acad9145dc0ef56f\", size \"41623040\" in 2.078206794s" Mar 20 21:15:44.052329 containerd[1485]: time="2025-03-20T21:15:44.052242849Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.2\" returns image reference \"sha256:15defb01cf01d9d97dc594b25d63dee89192c67a6c991b6a78d49fa834325f4e\"" Mar 20 21:15:44.054302 containerd[1485]: time="2025-03-20T21:15:44.054083259Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.2\"" Mar 20 21:15:44.055635 containerd[1485]: time="2025-03-20T21:15:44.055239906Z" level=info msg="CreateContainer within sandbox \"bb60bacaad0d905abce4377ae3f4cdab4d18db3d6b4c674d817fdcffeb4d4b17\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Mar 20 21:15:44.060743 containerd[1485]: time="2025-03-20T21:15:44.060703777Z" level=info msg="Container ad4699d874faf1ba9a0794922797f614c74e415fff8508d0b1f006e22f7e35a0: CDI devices from CRI Config.CDIDevices: []" Mar 20 21:15:44.071529 containerd[1485]: time="2025-03-20T21:15:44.071487557Z" level=info msg="CreateContainer within sandbox \"bb60bacaad0d905abce4377ae3f4cdab4d18db3d6b4c674d817fdcffeb4d4b17\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"ad4699d874faf1ba9a0794922797f614c74e415fff8508d0b1f006e22f7e35a0\"" Mar 20 21:15:44.072158 containerd[1485]: time="2025-03-20T21:15:44.072137641Z" level=info msg="StartContainer for \"ad4699d874faf1ba9a0794922797f614c74e415fff8508d0b1f006e22f7e35a0\"" Mar 20 21:15:44.074286 containerd[1485]: time="2025-03-20T21:15:44.074258053Z" level=info msg="connecting to shim ad4699d874faf1ba9a0794922797f614c74e415fff8508d0b1f006e22f7e35a0" address="unix:///run/containerd/s/a1aa89c8fa4955b52fc9c67dc3d6b9fdfe1af4d6ebdd6064071db1bc6c71bd3c" protocol=ttrpc version=3 Mar 20 21:15:44.100444 systemd[1]: Started cri-containerd-ad4699d874faf1ba9a0794922797f614c74e415fff8508d0b1f006e22f7e35a0.scope - libcontainer container ad4699d874faf1ba9a0794922797f614c74e415fff8508d0b1f006e22f7e35a0. Mar 20 21:15:44.142701 containerd[1485]: time="2025-03-20T21:15:44.141084387Z" level=info msg="StartContainer for \"ad4699d874faf1ba9a0794922797f614c74e415fff8508d0b1f006e22f7e35a0\" returns successfully" Mar 20 21:15:44.149720 sshd[4587]: Connection closed by 10.0.0.1 port 45588 Mar 20 21:15:44.153341 sshd-session[4550]: pam_unix(sshd:session): session closed for user core Mar 20 21:15:44.165771 systemd[1]: sshd@10-10.0.0.54:22-10.0.0.1:45588.service: Deactivated successfully. Mar 20 21:15:44.167959 systemd[1]: session-11.scope: Deactivated successfully. Mar 20 21:15:44.172325 systemd-logind[1470]: Session 11 logged out. Waiting for processes to exit. Mar 20 21:15:44.174464 systemd[1]: Started sshd@11-10.0.0.54:22-10.0.0.1:45592.service - OpenSSH per-connection server daemon (10.0.0.1:45592). Mar 20 21:15:44.175341 systemd-logind[1470]: Removed session 11. Mar 20 21:15:44.244987 sshd[4633]: Accepted publickey for core from 10.0.0.1 port 45592 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:15:44.246840 sshd-session[4633]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:15:44.252040 systemd-logind[1470]: New session 12 of user core. Mar 20 21:15:44.256411 systemd[1]: Started session-12.scope - Session 12 of User core. Mar 20 21:15:44.308545 containerd[1485]: time="2025-03-20T21:15:44.308457646Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.29.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:15:44.310029 containerd[1485]: time="2025-03-20T21:15:44.309981095Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.29.2: active requests=0, bytes read=77" Mar 20 21:15:44.312106 containerd[1485]: time="2025-03-20T21:15:44.312069626Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.29.2\" with image id \"sha256:15defb01cf01d9d97dc594b25d63dee89192c67a6c991b6a78d49fa834325f4e\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.29.2\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:3623f5b60fad0da3387a8649371b53171a4b1226f4d989d2acad9145dc0ef56f\", size \"41623040\" in 257.956326ms" Mar 20 21:15:44.312185 containerd[1485]: time="2025-03-20T21:15:44.312116027Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.2\" returns image reference \"sha256:15defb01cf01d9d97dc594b25d63dee89192c67a6c991b6a78d49fa834325f4e\"" Mar 20 21:15:44.314666 containerd[1485]: time="2025-03-20T21:15:44.314633041Z" level=info msg="CreateContainer within sandbox \"fdf3da8083f926e45241c057e9ea9678fb84411bc46b23959743ff765df8f1a5\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Mar 20 21:15:44.323762 containerd[1485]: time="2025-03-20T21:15:44.323724372Z" level=info msg="Container d6f8fec12a92c681515876f0f48394e95454d678fa466ccd8cb28302d9e49834: CDI devices from CRI Config.CDIDevices: []" Mar 20 21:15:44.334910 containerd[1485]: time="2025-03-20T21:15:44.334863194Z" level=info msg="CreateContainer within sandbox \"fdf3da8083f926e45241c057e9ea9678fb84411bc46b23959743ff765df8f1a5\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"d6f8fec12a92c681515876f0f48394e95454d678fa466ccd8cb28302d9e49834\"" Mar 20 21:15:44.335507 containerd[1485]: time="2025-03-20T21:15:44.335478718Z" level=info msg="StartContainer for \"d6f8fec12a92c681515876f0f48394e95454d678fa466ccd8cb28302d9e49834\"" Mar 20 21:15:44.336696 containerd[1485]: time="2025-03-20T21:15:44.336644244Z" level=info msg="connecting to shim d6f8fec12a92c681515876f0f48394e95454d678fa466ccd8cb28302d9e49834" address="unix:///run/containerd/s/c249e1229128cbe39c88ed9cd98c8b569b1212095b5ef7b7b0840820bba2ec25" protocol=ttrpc version=3 Mar 20 21:15:44.374145 systemd[1]: Started cri-containerd-d6f8fec12a92c681515876f0f48394e95454d678fa466ccd8cb28302d9e49834.scope - libcontainer container d6f8fec12a92c681515876f0f48394e95454d678fa466ccd8cb28302d9e49834. Mar 20 21:15:44.511668 containerd[1485]: time="2025-03-20T21:15:44.511493025Z" level=info msg="StartContainer for \"d6f8fec12a92c681515876f0f48394e95454d678fa466ccd8cb28302d9e49834\" returns successfully" Mar 20 21:15:44.570645 sshd[4640]: Connection closed by 10.0.0.1 port 45592 Mar 20 21:15:44.572165 sshd-session[4633]: pam_unix(sshd:session): session closed for user core Mar 20 21:15:44.578020 systemd[1]: sshd@11-10.0.0.54:22-10.0.0.1:45592.service: Deactivated successfully. Mar 20 21:15:44.579788 systemd[1]: session-12.scope: Deactivated successfully. Mar 20 21:15:44.583761 systemd-logind[1470]: Session 12 logged out. Waiting for processes to exit. Mar 20 21:15:44.584937 kubelet[2596]: I0320 21:15:44.584282 2596 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-8646c76b8-kbdmr" podStartSLOduration=24.164758478 podStartE2EDuration="26.584260313s" podCreationTimestamp="2025-03-20 21:15:18 +0000 UTC" firstStartedPulling="2025-03-20 21:15:41.633991981 +0000 UTC m=+37.365476076" lastFinishedPulling="2025-03-20 21:15:44.053493816 +0000 UTC m=+39.784977911" observedRunningTime="2025-03-20 21:15:44.577103433 +0000 UTC m=+40.308587528" watchObservedRunningTime="2025-03-20 21:15:44.584260313 +0000 UTC m=+40.315744528" Mar 20 21:15:44.586733 systemd-logind[1470]: Removed session 12. Mar 20 21:15:44.594828 kubelet[2596]: I0320 21:15:44.594757 2596 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-6f6b679f8f-mbg5n" podStartSLOduration=34.594738371 podStartE2EDuration="34.594738371s" podCreationTimestamp="2025-03-20 21:15:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-20 21:15:44.593689566 +0000 UTC m=+40.325173661" watchObservedRunningTime="2025-03-20 21:15:44.594738371 +0000 UTC m=+40.326222466" Mar 20 21:15:44.612131 kubelet[2596]: I0320 21:15:44.611905 2596 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-8646c76b8-xnq47" podStartSLOduration=26.076047779 podStartE2EDuration="26.611885108s" podCreationTimestamp="2025-03-20 21:15:18 +0000 UTC" firstStartedPulling="2025-03-20 21:15:43.776901301 +0000 UTC m=+39.508385396" lastFinishedPulling="2025-03-20 21:15:44.31273867 +0000 UTC m=+40.044222725" observedRunningTime="2025-03-20 21:15:44.610761221 +0000 UTC m=+40.342245276" watchObservedRunningTime="2025-03-20 21:15:44.611885108 +0000 UTC m=+40.343369163" Mar 20 21:15:44.789354 systemd-networkd[1416]: cali9e5c282701d: Gained IPv6LL Mar 20 21:15:45.346248 containerd[1485]: time="2025-03-20T21:15:45.345913943Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-7htv8,Uid:4e0ac6d0-32f4-4ad3-a9b3-9da162162a58,Namespace:calico-system,Attempt:0,}" Mar 20 21:15:45.489661 systemd-networkd[1416]: calif32e9f21078: Link UP Mar 20 21:15:45.490063 systemd-networkd[1416]: calif32e9f21078: Gained carrier Mar 20 21:15:45.493451 systemd-networkd[1416]: califf0ac8946a0: Gained IPv6LL Mar 20 21:15:45.505078 containerd[1485]: 2025-03-20 21:15:45.392 [INFO][4697] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-csi--node--driver--7htv8-eth0 csi-node-driver- calico-system 4e0ac6d0-32f4-4ad3-a9b3-9da162162a58 622 0 2025-03-20 21:15:19 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:568c96974f k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s localhost csi-node-driver-7htv8 eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] calif32e9f21078 [] []}} ContainerID="56354cbb177d440d3d5b9900021ffbe9a0248a43096667251a459d7483cb59a9" Namespace="calico-system" Pod="csi-node-driver-7htv8" WorkloadEndpoint="localhost-k8s-csi--node--driver--7htv8-" Mar 20 21:15:45.505078 containerd[1485]: 2025-03-20 21:15:45.392 [INFO][4697] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="56354cbb177d440d3d5b9900021ffbe9a0248a43096667251a459d7483cb59a9" Namespace="calico-system" Pod="csi-node-driver-7htv8" WorkloadEndpoint="localhost-k8s-csi--node--driver--7htv8-eth0" Mar 20 21:15:45.505078 containerd[1485]: 2025-03-20 21:15:45.428 [INFO][4712] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="56354cbb177d440d3d5b9900021ffbe9a0248a43096667251a459d7483cb59a9" HandleID="k8s-pod-network.56354cbb177d440d3d5b9900021ffbe9a0248a43096667251a459d7483cb59a9" Workload="localhost-k8s-csi--node--driver--7htv8-eth0" Mar 20 21:15:45.505078 containerd[1485]: 2025-03-20 21:15:45.442 [INFO][4712] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="56354cbb177d440d3d5b9900021ffbe9a0248a43096667251a459d7483cb59a9" HandleID="k8s-pod-network.56354cbb177d440d3d5b9900021ffbe9a0248a43096667251a459d7483cb59a9" Workload="localhost-k8s-csi--node--driver--7htv8-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002dfa20), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"csi-node-driver-7htv8", "timestamp":"2025-03-20 21:15:45.428319736 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Mar 20 21:15:45.505078 containerd[1485]: 2025-03-20 21:15:45.442 [INFO][4712] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Mar 20 21:15:45.505078 containerd[1485]: 2025-03-20 21:15:45.442 [INFO][4712] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Mar 20 21:15:45.505078 containerd[1485]: 2025-03-20 21:15:45.442 [INFO][4712] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Mar 20 21:15:45.505078 containerd[1485]: 2025-03-20 21:15:45.445 [INFO][4712] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.56354cbb177d440d3d5b9900021ffbe9a0248a43096667251a459d7483cb59a9" host="localhost" Mar 20 21:15:45.505078 containerd[1485]: 2025-03-20 21:15:45.449 [INFO][4712] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" Mar 20 21:15:45.505078 containerd[1485]: 2025-03-20 21:15:45.454 [INFO][4712] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Mar 20 21:15:45.505078 containerd[1485]: 2025-03-20 21:15:45.460 [INFO][4712] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Mar 20 21:15:45.505078 containerd[1485]: 2025-03-20 21:15:45.462 [INFO][4712] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Mar 20 21:15:45.505078 containerd[1485]: 2025-03-20 21:15:45.462 [INFO][4712] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.56354cbb177d440d3d5b9900021ffbe9a0248a43096667251a459d7483cb59a9" host="localhost" Mar 20 21:15:45.505078 containerd[1485]: 2025-03-20 21:15:45.464 [INFO][4712] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.56354cbb177d440d3d5b9900021ffbe9a0248a43096667251a459d7483cb59a9 Mar 20 21:15:45.505078 containerd[1485]: 2025-03-20 21:15:45.467 [INFO][4712] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.56354cbb177d440d3d5b9900021ffbe9a0248a43096667251a459d7483cb59a9" host="localhost" Mar 20 21:15:45.505078 containerd[1485]: 2025-03-20 21:15:45.476 [INFO][4712] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.134/26] block=192.168.88.128/26 handle="k8s-pod-network.56354cbb177d440d3d5b9900021ffbe9a0248a43096667251a459d7483cb59a9" host="localhost" Mar 20 21:15:45.505078 containerd[1485]: 2025-03-20 21:15:45.476 [INFO][4712] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.134/26] handle="k8s-pod-network.56354cbb177d440d3d5b9900021ffbe9a0248a43096667251a459d7483cb59a9" host="localhost" Mar 20 21:15:45.505078 containerd[1485]: 2025-03-20 21:15:45.476 [INFO][4712] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Mar 20 21:15:45.505078 containerd[1485]: 2025-03-20 21:15:45.476 [INFO][4712] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.134/26] IPv6=[] ContainerID="56354cbb177d440d3d5b9900021ffbe9a0248a43096667251a459d7483cb59a9" HandleID="k8s-pod-network.56354cbb177d440d3d5b9900021ffbe9a0248a43096667251a459d7483cb59a9" Workload="localhost-k8s-csi--node--driver--7htv8-eth0" Mar 20 21:15:45.506357 containerd[1485]: 2025-03-20 21:15:45.480 [INFO][4697] cni-plugin/k8s.go 386: Populated endpoint ContainerID="56354cbb177d440d3d5b9900021ffbe9a0248a43096667251a459d7483cb59a9" Namespace="calico-system" Pod="csi-node-driver-7htv8" WorkloadEndpoint="localhost-k8s-csi--node--driver--7htv8-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--7htv8-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"4e0ac6d0-32f4-4ad3-a9b3-9da162162a58", ResourceVersion:"622", Generation:0, CreationTimestamp:time.Date(2025, time.March, 20, 21, 15, 19, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"568c96974f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"csi-node-driver-7htv8", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calif32e9f21078", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Mar 20 21:15:45.506357 containerd[1485]: 2025-03-20 21:15:45.480 [INFO][4697] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.134/32] ContainerID="56354cbb177d440d3d5b9900021ffbe9a0248a43096667251a459d7483cb59a9" Namespace="calico-system" Pod="csi-node-driver-7htv8" WorkloadEndpoint="localhost-k8s-csi--node--driver--7htv8-eth0" Mar 20 21:15:45.506357 containerd[1485]: 2025-03-20 21:15:45.480 [INFO][4697] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calif32e9f21078 ContainerID="56354cbb177d440d3d5b9900021ffbe9a0248a43096667251a459d7483cb59a9" Namespace="calico-system" Pod="csi-node-driver-7htv8" WorkloadEndpoint="localhost-k8s-csi--node--driver--7htv8-eth0" Mar 20 21:15:45.506357 containerd[1485]: 2025-03-20 21:15:45.489 [INFO][4697] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="56354cbb177d440d3d5b9900021ffbe9a0248a43096667251a459d7483cb59a9" Namespace="calico-system" Pod="csi-node-driver-7htv8" WorkloadEndpoint="localhost-k8s-csi--node--driver--7htv8-eth0" Mar 20 21:15:45.506357 containerd[1485]: 2025-03-20 21:15:45.491 [INFO][4697] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="56354cbb177d440d3d5b9900021ffbe9a0248a43096667251a459d7483cb59a9" Namespace="calico-system" Pod="csi-node-driver-7htv8" WorkloadEndpoint="localhost-k8s-csi--node--driver--7htv8-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--7htv8-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"4e0ac6d0-32f4-4ad3-a9b3-9da162162a58", ResourceVersion:"622", Generation:0, CreationTimestamp:time.Date(2025, time.March, 20, 21, 15, 19, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"568c96974f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"56354cbb177d440d3d5b9900021ffbe9a0248a43096667251a459d7483cb59a9", Pod:"csi-node-driver-7htv8", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calif32e9f21078", MAC:"66:b5:21:6e:8b:20", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Mar 20 21:15:45.506357 containerd[1485]: 2025-03-20 21:15:45.502 [INFO][4697] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="56354cbb177d440d3d5b9900021ffbe9a0248a43096667251a459d7483cb59a9" Namespace="calico-system" Pod="csi-node-driver-7htv8" WorkloadEndpoint="localhost-k8s-csi--node--driver--7htv8-eth0" Mar 20 21:15:45.525716 containerd[1485]: time="2025-03-20T21:15:45.525668608Z" level=info msg="connecting to shim 56354cbb177d440d3d5b9900021ffbe9a0248a43096667251a459d7483cb59a9" address="unix:///run/containerd/s/373d716ba405c86405018d5350f34c6f6f5c307cc52dc39148aed67dd752b826" namespace=k8s.io protocol=ttrpc version=3 Mar 20 21:15:45.551387 systemd[1]: Started cri-containerd-56354cbb177d440d3d5b9900021ffbe9a0248a43096667251a459d7483cb59a9.scope - libcontainer container 56354cbb177d440d3d5b9900021ffbe9a0248a43096667251a459d7483cb59a9. Mar 20 21:15:45.566451 systemd-resolved[1330]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Mar 20 21:15:45.582773 kubelet[2596]: I0320 21:15:45.582718 2596 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 20 21:15:45.583541 kubelet[2596]: I0320 21:15:45.582732 2596 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 20 21:15:45.583658 containerd[1485]: time="2025-03-20T21:15:45.583164511Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-7htv8,Uid:4e0ac6d0-32f4-4ad3-a9b3-9da162162a58,Namespace:calico-system,Attempt:0,} returns sandbox id \"56354cbb177d440d3d5b9900021ffbe9a0248a43096667251a459d7483cb59a9\"" Mar 20 21:15:45.585356 containerd[1485]: time="2025-03-20T21:15:45.585142641Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.2\"" Mar 20 21:15:46.437529 containerd[1485]: time="2025-03-20T21:15:46.437474500Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.29.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:15:46.438007 containerd[1485]: time="2025-03-20T21:15:46.437945422Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.29.2: active requests=0, bytes read=7473801" Mar 20 21:15:46.438869 containerd[1485]: time="2025-03-20T21:15:46.438836387Z" level=info msg="ImageCreate event name:\"sha256:f39063099e467ddd9d84500bfd4d97c404bb5f706a2161afc8979f4a94b8ad0b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:15:46.440637 containerd[1485]: time="2025-03-20T21:15:46.440601315Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:214b4eef7008808bda55ad3cc1d4a3cd8df9e0e8094dff213fa3241104eb892c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:15:46.441311 containerd[1485]: time="2025-03-20T21:15:46.441281719Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.29.2\" with image id \"sha256:f39063099e467ddd9d84500bfd4d97c404bb5f706a2161afc8979f4a94b8ad0b\", repo tag \"ghcr.io/flatcar/calico/csi:v3.29.2\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:214b4eef7008808bda55ad3cc1d4a3cd8df9e0e8094dff213fa3241104eb892c\", size \"8843558\" in 855.915357ms" Mar 20 21:15:46.441341 containerd[1485]: time="2025-03-20T21:15:46.441314599Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.2\" returns image reference \"sha256:f39063099e467ddd9d84500bfd4d97c404bb5f706a2161afc8979f4a94b8ad0b\"" Mar 20 21:15:46.444063 containerd[1485]: time="2025-03-20T21:15:46.444034012Z" level=info msg="CreateContainer within sandbox \"56354cbb177d440d3d5b9900021ffbe9a0248a43096667251a459d7483cb59a9\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Mar 20 21:15:46.454079 containerd[1485]: time="2025-03-20T21:15:46.451527409Z" level=info msg="Container e0bc6deef9ced06603e2f6be06d9d8c24696117fa84278eff0ade88ffa7b8cdc: CDI devices from CRI Config.CDIDevices: []" Mar 20 21:15:46.460759 containerd[1485]: time="2025-03-20T21:15:46.460714254Z" level=info msg="CreateContainer within sandbox \"56354cbb177d440d3d5b9900021ffbe9a0248a43096667251a459d7483cb59a9\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"e0bc6deef9ced06603e2f6be06d9d8c24696117fa84278eff0ade88ffa7b8cdc\"" Mar 20 21:15:46.461156 containerd[1485]: time="2025-03-20T21:15:46.461131257Z" level=info msg="StartContainer for \"e0bc6deef9ced06603e2f6be06d9d8c24696117fa84278eff0ade88ffa7b8cdc\"" Mar 20 21:15:46.463423 containerd[1485]: time="2025-03-20T21:15:46.463396108Z" level=info msg="connecting to shim e0bc6deef9ced06603e2f6be06d9d8c24696117fa84278eff0ade88ffa7b8cdc" address="unix:///run/containerd/s/373d716ba405c86405018d5350f34c6f6f5c307cc52dc39148aed67dd752b826" protocol=ttrpc version=3 Mar 20 21:15:46.484367 systemd[1]: Started cri-containerd-e0bc6deef9ced06603e2f6be06d9d8c24696117fa84278eff0ade88ffa7b8cdc.scope - libcontainer container e0bc6deef9ced06603e2f6be06d9d8c24696117fa84278eff0ade88ffa7b8cdc. Mar 20 21:15:46.516749 containerd[1485]: time="2025-03-20T21:15:46.516615090Z" level=info msg="StartContainer for \"e0bc6deef9ced06603e2f6be06d9d8c24696117fa84278eff0ade88ffa7b8cdc\" returns successfully" Mar 20 21:15:46.517975 containerd[1485]: time="2025-03-20T21:15:46.517944057Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.2\"" Mar 20 21:15:47.408311 containerd[1485]: time="2025-03-20T21:15:47.408265799Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:15:47.409407 containerd[1485]: time="2025-03-20T21:15:47.408724042Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.29.2: active requests=0, bytes read=13121717" Mar 20 21:15:47.409694 containerd[1485]: time="2025-03-20T21:15:47.409665886Z" level=info msg="ImageCreate event name:\"sha256:5b766f5f5d1b2ccc7c16f12d59c6c17c490ae33a8973c1fa7b2bcf3b8aa5098a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:15:47.412333 containerd[1485]: time="2025-03-20T21:15:47.412279618Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:54ef0afa50feb3f691782e8d6df9a7f27d127a3af9bbcbd0bcdadac98e8be8e3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 20 21:15:47.412968 containerd[1485]: time="2025-03-20T21:15:47.412813980Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.2\" with image id \"sha256:5b766f5f5d1b2ccc7c16f12d59c6c17c490ae33a8973c1fa7b2bcf3b8aa5098a\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.2\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:54ef0afa50feb3f691782e8d6df9a7f27d127a3af9bbcbd0bcdadac98e8be8e3\", size \"14491426\" in 894.832763ms" Mar 20 21:15:47.412968 containerd[1485]: time="2025-03-20T21:15:47.412849061Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.2\" returns image reference \"sha256:5b766f5f5d1b2ccc7c16f12d59c6c17c490ae33a8973c1fa7b2bcf3b8aa5098a\"" Mar 20 21:15:47.415831 containerd[1485]: time="2025-03-20T21:15:47.415800554Z" level=info msg="CreateContainer within sandbox \"56354cbb177d440d3d5b9900021ffbe9a0248a43096667251a459d7483cb59a9\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Mar 20 21:15:47.427029 containerd[1485]: time="2025-03-20T21:15:47.425958801Z" level=info msg="Container 3cb9010637ec40861c9b311402478726839133a624aeb14954fcd935061b3e9a: CDI devices from CRI Config.CDIDevices: []" Mar 20 21:15:47.428828 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3869690322.mount: Deactivated successfully. Mar 20 21:15:47.439939 containerd[1485]: time="2025-03-20T21:15:47.439906946Z" level=info msg="CreateContainer within sandbox \"56354cbb177d440d3d5b9900021ffbe9a0248a43096667251a459d7483cb59a9\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"3cb9010637ec40861c9b311402478726839133a624aeb14954fcd935061b3e9a\"" Mar 20 21:15:47.440617 containerd[1485]: time="2025-03-20T21:15:47.440514828Z" level=info msg="StartContainer for \"3cb9010637ec40861c9b311402478726839133a624aeb14954fcd935061b3e9a\"" Mar 20 21:15:47.442424 containerd[1485]: time="2025-03-20T21:15:47.442382997Z" level=info msg="connecting to shim 3cb9010637ec40861c9b311402478726839133a624aeb14954fcd935061b3e9a" address="unix:///run/containerd/s/373d716ba405c86405018d5350f34c6f6f5c307cc52dc39148aed67dd752b826" protocol=ttrpc version=3 Mar 20 21:15:47.468358 systemd[1]: Started cri-containerd-3cb9010637ec40861c9b311402478726839133a624aeb14954fcd935061b3e9a.scope - libcontainer container 3cb9010637ec40861c9b311402478726839133a624aeb14954fcd935061b3e9a. Mar 20 21:15:47.478365 systemd-networkd[1416]: calif32e9f21078: Gained IPv6LL Mar 20 21:15:47.502612 containerd[1485]: time="2025-03-20T21:15:47.502568555Z" level=info msg="StartContainer for \"3cb9010637ec40861c9b311402478726839133a624aeb14954fcd935061b3e9a\" returns successfully" Mar 20 21:15:47.601124 kubelet[2596]: I0320 21:15:47.601062 2596 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-7htv8" podStartSLOduration=26.772464226 podStartE2EDuration="28.60104465s" podCreationTimestamp="2025-03-20 21:15:19 +0000 UTC" firstStartedPulling="2025-03-20 21:15:45.58491764 +0000 UTC m=+41.316401735" lastFinishedPulling="2025-03-20 21:15:47.413498064 +0000 UTC m=+43.144982159" observedRunningTime="2025-03-20 21:15:47.600384807 +0000 UTC m=+43.331868902" watchObservedRunningTime="2025-03-20 21:15:47.60104465 +0000 UTC m=+43.332528745" Mar 20 21:15:48.059071 kubelet[2596]: I0320 21:15:48.058965 2596 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 20 21:15:48.417449 kubelet[2596]: I0320 21:15:48.417361 2596 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Mar 20 21:15:48.426787 kubelet[2596]: I0320 21:15:48.426746 2596 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Mar 20 21:15:49.586100 systemd[1]: Started sshd@12-10.0.0.54:22-10.0.0.1:45594.service - OpenSSH per-connection server daemon (10.0.0.1:45594). Mar 20 21:15:49.649279 sshd[4862]: Accepted publickey for core from 10.0.0.1 port 45594 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:15:49.654086 sshd-session[4862]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:15:49.658606 systemd-logind[1470]: New session 13 of user core. Mar 20 21:15:49.670416 systemd[1]: Started session-13.scope - Session 13 of User core. Mar 20 21:15:49.890545 sshd[4864]: Connection closed by 10.0.0.1 port 45594 Mar 20 21:15:49.890640 sshd-session[4862]: pam_unix(sshd:session): session closed for user core Mar 20 21:15:49.895400 systemd[1]: sshd@12-10.0.0.54:22-10.0.0.1:45594.service: Deactivated successfully. Mar 20 21:15:49.897993 systemd[1]: session-13.scope: Deactivated successfully. Mar 20 21:15:49.898761 systemd-logind[1470]: Session 13 logged out. Waiting for processes to exit. Mar 20 21:15:49.900719 systemd-logind[1470]: Removed session 13. Mar 20 21:15:51.675049 kubelet[2596]: I0320 21:15:51.674998 2596 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 20 21:15:51.707396 containerd[1485]: time="2025-03-20T21:15:51.707327702Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c14e33d2f5b60700e68dcfd113314dd88a94ba5b5df327532409a6162dc44321\" id:\"30dca0ea409e8bbe44626bc6daee7a9bbe6afc1e160e07572579cd173cfd6b25\" pid:4893 exited_at:{seconds:1742505351 nanos:706969181}" Mar 20 21:15:51.742056 containerd[1485]: time="2025-03-20T21:15:51.741694865Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c14e33d2f5b60700e68dcfd113314dd88a94ba5b5df327532409a6162dc44321\" id:\"65947fa3d324771fe16fd99b0c85fa6714f421abc3f105d0617dbc8fd51a18d7\" pid:4915 exited_at:{seconds:1742505351 nanos:741371464}" Mar 20 21:15:54.903868 systemd[1]: Started sshd@13-10.0.0.54:22-10.0.0.1:53506.service - OpenSSH per-connection server daemon (10.0.0.1:53506). Mar 20 21:15:54.956763 sshd[4928]: Accepted publickey for core from 10.0.0.1 port 53506 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:15:54.958004 sshd-session[4928]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:15:54.962411 systemd-logind[1470]: New session 14 of user core. Mar 20 21:15:54.973362 systemd[1]: Started session-14.scope - Session 14 of User core. Mar 20 21:15:55.102492 sshd[4930]: Connection closed by 10.0.0.1 port 53506 Mar 20 21:15:55.102817 sshd-session[4928]: pam_unix(sshd:session): session closed for user core Mar 20 21:15:55.106005 systemd[1]: sshd@13-10.0.0.54:22-10.0.0.1:53506.service: Deactivated successfully. Mar 20 21:15:55.107819 systemd[1]: session-14.scope: Deactivated successfully. Mar 20 21:15:55.108384 systemd-logind[1470]: Session 14 logged out. Waiting for processes to exit. Mar 20 21:15:55.109269 systemd-logind[1470]: Removed session 14. Mar 20 21:16:00.114451 systemd[1]: Started sshd@14-10.0.0.54:22-10.0.0.1:53508.service - OpenSSH per-connection server daemon (10.0.0.1:53508). Mar 20 21:16:00.160671 sshd[4953]: Accepted publickey for core from 10.0.0.1 port 53508 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:16:00.161820 sshd-session[4953]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:16:00.165695 systemd-logind[1470]: New session 15 of user core. Mar 20 21:16:00.171422 systemd[1]: Started session-15.scope - Session 15 of User core. Mar 20 21:16:00.321485 sshd[4955]: Connection closed by 10.0.0.1 port 53508 Mar 20 21:16:00.322164 sshd-session[4953]: pam_unix(sshd:session): session closed for user core Mar 20 21:16:00.326479 systemd[1]: sshd@14-10.0.0.54:22-10.0.0.1:53508.service: Deactivated successfully. Mar 20 21:16:00.328181 systemd[1]: session-15.scope: Deactivated successfully. Mar 20 21:16:00.328813 systemd-logind[1470]: Session 15 logged out. Waiting for processes to exit. Mar 20 21:16:00.329592 systemd-logind[1470]: Removed session 15. Mar 20 21:16:04.737382 containerd[1485]: time="2025-03-20T21:16:04.737320272Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3298a708f055861f6f20ce36a7ee251ccfc70f10f7107e13a9514eaec3f5beb1\" id:\"d16645aab25d144b6ad73b6870e15ccbd63a4f6fd6f5498b9bb2429e7ffede7a\" pid:4981 exited_at:{seconds:1742505364 nanos:736727111}" Mar 20 21:16:05.342876 systemd[1]: Started sshd@15-10.0.0.54:22-10.0.0.1:45422.service - OpenSSH per-connection server daemon (10.0.0.1:45422). Mar 20 21:16:05.414939 sshd[4994]: Accepted publickey for core from 10.0.0.1 port 45422 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:16:05.416476 sshd-session[4994]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:16:05.422788 systemd-logind[1470]: New session 16 of user core. Mar 20 21:16:05.429398 systemd[1]: Started session-16.scope - Session 16 of User core. Mar 20 21:16:05.598577 sshd[4996]: Connection closed by 10.0.0.1 port 45422 Mar 20 21:16:05.598890 sshd-session[4994]: pam_unix(sshd:session): session closed for user core Mar 20 21:16:05.605145 systemd[1]: sshd@15-10.0.0.54:22-10.0.0.1:45422.service: Deactivated successfully. Mar 20 21:16:05.608045 systemd[1]: session-16.scope: Deactivated successfully. Mar 20 21:16:05.609731 systemd-logind[1470]: Session 16 logged out. Waiting for processes to exit. Mar 20 21:16:05.612567 systemd-logind[1470]: Removed session 16. Mar 20 21:16:10.612865 systemd[1]: Started sshd@16-10.0.0.54:22-10.0.0.1:45424.service - OpenSSH per-connection server daemon (10.0.0.1:45424). Mar 20 21:16:10.666940 sshd[5014]: Accepted publickey for core from 10.0.0.1 port 45424 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:16:10.668556 sshd-session[5014]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:16:10.672319 systemd-logind[1470]: New session 17 of user core. Mar 20 21:16:10.683353 systemd[1]: Started session-17.scope - Session 17 of User core. Mar 20 21:16:10.848035 sshd[5016]: Connection closed by 10.0.0.1 port 45424 Mar 20 21:16:10.848521 sshd-session[5014]: pam_unix(sshd:session): session closed for user core Mar 20 21:16:10.851521 systemd[1]: sshd@16-10.0.0.54:22-10.0.0.1:45424.service: Deactivated successfully. Mar 20 21:16:10.854279 systemd[1]: session-17.scope: Deactivated successfully. Mar 20 21:16:10.855788 systemd-logind[1470]: Session 17 logged out. Waiting for processes to exit. Mar 20 21:16:10.856922 systemd-logind[1470]: Removed session 17. Mar 20 21:16:15.860594 systemd[1]: Started sshd@17-10.0.0.54:22-10.0.0.1:49990.service - OpenSSH per-connection server daemon (10.0.0.1:49990). Mar 20 21:16:15.908378 sshd[5032]: Accepted publickey for core from 10.0.0.1 port 49990 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:16:15.909669 sshd-session[5032]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:16:15.913678 systemd-logind[1470]: New session 18 of user core. Mar 20 21:16:15.921355 systemd[1]: Started session-18.scope - Session 18 of User core. Mar 20 21:16:16.044247 sshd[5034]: Connection closed by 10.0.0.1 port 49990 Mar 20 21:16:16.043924 sshd-session[5032]: pam_unix(sshd:session): session closed for user core Mar 20 21:16:16.048137 systemd[1]: sshd@17-10.0.0.54:22-10.0.0.1:49990.service: Deactivated successfully. Mar 20 21:16:16.049969 systemd[1]: session-18.scope: Deactivated successfully. Mar 20 21:16:16.050659 systemd-logind[1470]: Session 18 logged out. Waiting for processes to exit. Mar 20 21:16:16.051953 systemd-logind[1470]: Removed session 18. Mar 20 21:16:17.639399 kubelet[2596]: I0320 21:16:17.639275 2596 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 20 21:16:21.055810 systemd[1]: Started sshd@18-10.0.0.54:22-10.0.0.1:49994.service - OpenSSH per-connection server daemon (10.0.0.1:49994). Mar 20 21:16:21.111718 sshd[5055]: Accepted publickey for core from 10.0.0.1 port 49994 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:16:21.112833 sshd-session[5055]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:16:21.116675 systemd-logind[1470]: New session 19 of user core. Mar 20 21:16:21.127350 systemd[1]: Started session-19.scope - Session 19 of User core. Mar 20 21:16:21.278530 sshd[5057]: Connection closed by 10.0.0.1 port 49994 Mar 20 21:16:21.279037 sshd-session[5055]: pam_unix(sshd:session): session closed for user core Mar 20 21:16:21.282298 systemd[1]: sshd@18-10.0.0.54:22-10.0.0.1:49994.service: Deactivated successfully. Mar 20 21:16:21.284073 systemd[1]: session-19.scope: Deactivated successfully. Mar 20 21:16:21.284718 systemd-logind[1470]: Session 19 logged out. Waiting for processes to exit. Mar 20 21:16:21.285506 systemd-logind[1470]: Removed session 19. Mar 20 21:16:21.730357 containerd[1485]: time="2025-03-20T21:16:21.730308218Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c14e33d2f5b60700e68dcfd113314dd88a94ba5b5df327532409a6162dc44321\" id:\"e3030975bdbbc692da9fc15a1b36e47d056c1773ab1ea0422f5c8f1ae749e5a6\" pid:5081 exited_at:{seconds:1742505381 nanos:730073004}" Mar 20 21:16:26.294063 systemd[1]: Started sshd@19-10.0.0.54:22-10.0.0.1:52708.service - OpenSSH per-connection server daemon (10.0.0.1:52708). Mar 20 21:16:26.354029 sshd[5095]: Accepted publickey for core from 10.0.0.1 port 52708 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:16:26.355371 sshd-session[5095]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:16:26.359227 systemd-logind[1470]: New session 20 of user core. Mar 20 21:16:26.366376 systemd[1]: Started session-20.scope - Session 20 of User core. Mar 20 21:16:26.496107 sshd[5097]: Connection closed by 10.0.0.1 port 52708 Mar 20 21:16:26.496814 sshd-session[5095]: pam_unix(sshd:session): session closed for user core Mar 20 21:16:26.499247 systemd[1]: session-20.scope: Deactivated successfully. Mar 20 21:16:26.500468 systemd[1]: sshd@19-10.0.0.54:22-10.0.0.1:52708.service: Deactivated successfully. Mar 20 21:16:26.502389 systemd-logind[1470]: Session 20 logged out. Waiting for processes to exit. Mar 20 21:16:26.503041 systemd-logind[1470]: Removed session 20. Mar 20 21:16:31.507709 systemd[1]: Started sshd@20-10.0.0.54:22-10.0.0.1:52724.service - OpenSSH per-connection server daemon (10.0.0.1:52724). Mar 20 21:16:31.563854 sshd[5110]: Accepted publickey for core from 10.0.0.1 port 52724 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:16:31.564970 sshd-session[5110]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:16:31.569295 systemd-logind[1470]: New session 21 of user core. Mar 20 21:16:31.575426 systemd[1]: Started session-21.scope - Session 21 of User core. Mar 20 21:16:31.729924 sshd[5112]: Connection closed by 10.0.0.1 port 52724 Mar 20 21:16:31.730606 sshd-session[5110]: pam_unix(sshd:session): session closed for user core Mar 20 21:16:31.733999 systemd[1]: sshd@20-10.0.0.54:22-10.0.0.1:52724.service: Deactivated successfully. Mar 20 21:16:31.735683 systemd[1]: session-21.scope: Deactivated successfully. Mar 20 21:16:31.737703 systemd-logind[1470]: Session 21 logged out. Waiting for processes to exit. Mar 20 21:16:31.738711 systemd-logind[1470]: Removed session 21. Mar 20 21:16:34.728683 containerd[1485]: time="2025-03-20T21:16:34.728627124Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3298a708f055861f6f20ce36a7ee251ccfc70f10f7107e13a9514eaec3f5beb1\" id:\"48f84139dfc1a2017f6c605b22cb399344dfd8eb985bb340563a57c1392e7b69\" pid:5135 exited_at:{seconds:1742505394 nanos:728317351}" Mar 20 21:16:36.741468 systemd[1]: Started sshd@21-10.0.0.54:22-10.0.0.1:57074.service - OpenSSH per-connection server daemon (10.0.0.1:57074). Mar 20 21:16:36.794691 sshd[5148]: Accepted publickey for core from 10.0.0.1 port 57074 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:16:36.796055 sshd-session[5148]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:16:36.800115 systemd-logind[1470]: New session 22 of user core. Mar 20 21:16:36.806379 systemd[1]: Started session-22.scope - Session 22 of User core. Mar 20 21:16:36.959396 sshd[5150]: Connection closed by 10.0.0.1 port 57074 Mar 20 21:16:36.959732 sshd-session[5148]: pam_unix(sshd:session): session closed for user core Mar 20 21:16:36.962963 systemd[1]: sshd@21-10.0.0.54:22-10.0.0.1:57074.service: Deactivated successfully. Mar 20 21:16:36.964758 systemd[1]: session-22.scope: Deactivated successfully. Mar 20 21:16:36.965421 systemd-logind[1470]: Session 22 logged out. Waiting for processes to exit. Mar 20 21:16:36.966523 systemd-logind[1470]: Removed session 22. Mar 20 21:16:41.979705 systemd[1]: Started sshd@22-10.0.0.54:22-10.0.0.1:57078.service - OpenSSH per-connection server daemon (10.0.0.1:57078). Mar 20 21:16:42.035253 sshd[5166]: Accepted publickey for core from 10.0.0.1 port 57078 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:16:42.036523 sshd-session[5166]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:16:42.040084 systemd-logind[1470]: New session 23 of user core. Mar 20 21:16:42.054429 systemd[1]: Started session-23.scope - Session 23 of User core. Mar 20 21:16:42.249049 sshd[5168]: Connection closed by 10.0.0.1 port 57078 Mar 20 21:16:42.248891 sshd-session[5166]: pam_unix(sshd:session): session closed for user core Mar 20 21:16:42.256713 systemd[1]: sshd@22-10.0.0.54:22-10.0.0.1:57078.service: Deactivated successfully. Mar 20 21:16:42.258498 systemd[1]: session-23.scope: Deactivated successfully. Mar 20 21:16:42.259074 systemd-logind[1470]: Session 23 logged out. Waiting for processes to exit. Mar 20 21:16:42.259894 systemd-logind[1470]: Removed session 23. Mar 20 21:16:47.259512 systemd[1]: Started sshd@23-10.0.0.54:22-10.0.0.1:34540.service - OpenSSH per-connection server daemon (10.0.0.1:34540). Mar 20 21:16:47.307719 sshd[5181]: Accepted publickey for core from 10.0.0.1 port 34540 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:16:47.308919 sshd-session[5181]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:16:47.312514 systemd-logind[1470]: New session 24 of user core. Mar 20 21:16:47.319353 systemd[1]: Started session-24.scope - Session 24 of User core. Mar 20 21:16:47.466266 sshd[5183]: Connection closed by 10.0.0.1 port 34540 Mar 20 21:16:47.466598 sshd-session[5181]: pam_unix(sshd:session): session closed for user core Mar 20 21:16:47.470110 systemd[1]: sshd@23-10.0.0.54:22-10.0.0.1:34540.service: Deactivated successfully. Mar 20 21:16:47.471749 systemd[1]: session-24.scope: Deactivated successfully. Mar 20 21:16:47.472388 systemd-logind[1470]: Session 24 logged out. Waiting for processes to exit. Mar 20 21:16:47.473207 systemd-logind[1470]: Removed session 24. Mar 20 21:16:47.818462 containerd[1485]: time="2025-03-20T21:16:47.818418247Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c14e33d2f5b60700e68dcfd113314dd88a94ba5b5df327532409a6162dc44321\" id:\"c25113e767f64f8a3509282aef054d8031d6fe07e656ec8921ede858a6ca0a57\" pid:5208 exited_at:{seconds:1742505407 nanos:818145278}" Mar 20 21:16:51.706930 containerd[1485]: time="2025-03-20T21:16:51.706878907Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c14e33d2f5b60700e68dcfd113314dd88a94ba5b5df327532409a6162dc44321\" id:\"70acbebaa4ed677075e8474219638886b5df864747c9b478e1db4d2e087ae8db\" pid:5230 exited_at:{seconds:1742505411 nanos:706670701}" Mar 20 21:16:52.477561 systemd[1]: Started sshd@24-10.0.0.54:22-10.0.0.1:51394.service - OpenSSH per-connection server daemon (10.0.0.1:51394). Mar 20 21:16:52.531324 sshd[5241]: Accepted publickey for core from 10.0.0.1 port 51394 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:16:52.532573 sshd-session[5241]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:16:52.537299 systemd-logind[1470]: New session 25 of user core. Mar 20 21:16:52.549368 systemd[1]: Started session-25.scope - Session 25 of User core. Mar 20 21:16:52.691813 sshd[5243]: Connection closed by 10.0.0.1 port 51394 Mar 20 21:16:52.692449 sshd-session[5241]: pam_unix(sshd:session): session closed for user core Mar 20 21:16:52.695751 systemd[1]: sshd@24-10.0.0.54:22-10.0.0.1:51394.service: Deactivated successfully. Mar 20 21:16:52.697671 systemd[1]: session-25.scope: Deactivated successfully. Mar 20 21:16:52.698389 systemd-logind[1470]: Session 25 logged out. Waiting for processes to exit. Mar 20 21:16:52.699621 systemd-logind[1470]: Removed session 25. Mar 20 21:16:57.708414 systemd[1]: Started sshd@25-10.0.0.54:22-10.0.0.1:51410.service - OpenSSH per-connection server daemon (10.0.0.1:51410). Mar 20 21:16:57.772149 sshd[5262]: Accepted publickey for core from 10.0.0.1 port 51410 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:16:57.773508 sshd-session[5262]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:16:57.778200 systemd-logind[1470]: New session 26 of user core. Mar 20 21:16:57.786403 systemd[1]: Started session-26.scope - Session 26 of User core. Mar 20 21:16:57.924002 sshd[5264]: Connection closed by 10.0.0.1 port 51410 Mar 20 21:16:57.924537 sshd-session[5262]: pam_unix(sshd:session): session closed for user core Mar 20 21:16:57.928048 systemd[1]: sshd@25-10.0.0.54:22-10.0.0.1:51410.service: Deactivated successfully. Mar 20 21:16:57.929841 systemd[1]: session-26.scope: Deactivated successfully. Mar 20 21:16:57.930533 systemd-logind[1470]: Session 26 logged out. Waiting for processes to exit. Mar 20 21:16:57.931319 systemd-logind[1470]: Removed session 26. Mar 20 21:17:02.938938 systemd[1]: Started sshd@26-10.0.0.54:22-10.0.0.1:41756.service - OpenSSH per-connection server daemon (10.0.0.1:41756). Mar 20 21:17:02.983225 sshd[5277]: Accepted publickey for core from 10.0.0.1 port 41756 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:17:02.984325 sshd-session[5277]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:17:02.988292 systemd-logind[1470]: New session 27 of user core. Mar 20 21:17:02.998377 systemd[1]: Started session-27.scope - Session 27 of User core. Mar 20 21:17:03.110129 sshd[5279]: Connection closed by 10.0.0.1 port 41756 Mar 20 21:17:03.110463 sshd-session[5277]: pam_unix(sshd:session): session closed for user core Mar 20 21:17:03.113727 systemd[1]: sshd@26-10.0.0.54:22-10.0.0.1:41756.service: Deactivated successfully. Mar 20 21:17:03.115588 systemd[1]: session-27.scope: Deactivated successfully. Mar 20 21:17:03.116377 systemd-logind[1470]: Session 27 logged out. Waiting for processes to exit. Mar 20 21:17:03.117161 systemd-logind[1470]: Removed session 27. Mar 20 21:17:04.733240 containerd[1485]: time="2025-03-20T21:17:04.733183898Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3298a708f055861f6f20ce36a7ee251ccfc70f10f7107e13a9514eaec3f5beb1\" id:\"7c59c6f9b0b513000b9906fa60a359149c1f9170a60200a5e356e2c0ff03c49c\" pid:5305 exited_at:{seconds:1742505424 nanos:732847851}" Mar 20 21:17:08.121609 systemd[1]: Started sshd@27-10.0.0.54:22-10.0.0.1:41766.service - OpenSSH per-connection server daemon (10.0.0.1:41766). Mar 20 21:17:08.181020 sshd[5320]: Accepted publickey for core from 10.0.0.1 port 41766 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:17:08.182092 sshd-session[5320]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:17:08.185818 systemd-logind[1470]: New session 28 of user core. Mar 20 21:17:08.198356 systemd[1]: Started session-28.scope - Session 28 of User core. Mar 20 21:17:08.318595 sshd[5322]: Connection closed by 10.0.0.1 port 41766 Mar 20 21:17:08.318956 sshd-session[5320]: pam_unix(sshd:session): session closed for user core Mar 20 21:17:08.322315 systemd[1]: sshd@27-10.0.0.54:22-10.0.0.1:41766.service: Deactivated successfully. Mar 20 21:17:08.324290 systemd[1]: session-28.scope: Deactivated successfully. Mar 20 21:17:08.325003 systemd-logind[1470]: Session 28 logged out. Waiting for processes to exit. Mar 20 21:17:08.325971 systemd-logind[1470]: Removed session 28. Mar 20 21:17:13.333466 systemd[1]: Started sshd@28-10.0.0.54:22-10.0.0.1:60144.service - OpenSSH per-connection server daemon (10.0.0.1:60144). Mar 20 21:17:13.387285 sshd[5355]: Accepted publickey for core from 10.0.0.1 port 60144 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:17:13.388338 sshd-session[5355]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:17:13.391914 systemd-logind[1470]: New session 29 of user core. Mar 20 21:17:13.410348 systemd[1]: Started session-29.scope - Session 29 of User core. Mar 20 21:17:13.534097 sshd[5358]: Connection closed by 10.0.0.1 port 60144 Mar 20 21:17:13.534457 sshd-session[5355]: pam_unix(sshd:session): session closed for user core Mar 20 21:17:13.537633 systemd[1]: sshd@28-10.0.0.54:22-10.0.0.1:60144.service: Deactivated successfully. Mar 20 21:17:13.539465 systemd[1]: session-29.scope: Deactivated successfully. Mar 20 21:17:13.540067 systemd-logind[1470]: Session 29 logged out. Waiting for processes to exit. Mar 20 21:17:13.540822 systemd-logind[1470]: Removed session 29. Mar 20 21:17:18.545718 systemd[1]: Started sshd@29-10.0.0.54:22-10.0.0.1:60158.service - OpenSSH per-connection server daemon (10.0.0.1:60158). Mar 20 21:17:18.601913 sshd[5373]: Accepted publickey for core from 10.0.0.1 port 60158 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:17:18.603194 sshd-session[5373]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:17:18.606877 systemd-logind[1470]: New session 30 of user core. Mar 20 21:17:18.614375 systemd[1]: Started session-30.scope - Session 30 of User core. Mar 20 21:17:18.781769 sshd[5375]: Connection closed by 10.0.0.1 port 60158 Mar 20 21:17:18.782368 sshd-session[5373]: pam_unix(sshd:session): session closed for user core Mar 20 21:17:18.785626 systemd[1]: sshd@29-10.0.0.54:22-10.0.0.1:60158.service: Deactivated successfully. Mar 20 21:17:18.787818 systemd[1]: session-30.scope: Deactivated successfully. Mar 20 21:17:18.788442 systemd-logind[1470]: Session 30 logged out. Waiting for processes to exit. Mar 20 21:17:18.789342 systemd-logind[1470]: Removed session 30. Mar 20 21:17:21.705557 containerd[1485]: time="2025-03-20T21:17:21.705506987Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c14e33d2f5b60700e68dcfd113314dd88a94ba5b5df327532409a6162dc44321\" id:\"f3319f074b02aa89d5ca43399d35272eb330c591a8e69eb422d66d245cfb1167\" pid:5399 exited_at:{seconds:1742505441 nanos:705254943}" Mar 20 21:17:23.793729 systemd[1]: Started sshd@30-10.0.0.54:22-10.0.0.1:46370.service - OpenSSH per-connection server daemon (10.0.0.1:46370). Mar 20 21:17:23.850322 sshd[5410]: Accepted publickey for core from 10.0.0.1 port 46370 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:17:23.851469 sshd-session[5410]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:17:23.855655 systemd-logind[1470]: New session 31 of user core. Mar 20 21:17:23.867360 systemd[1]: Started session-31.scope - Session 31 of User core. Mar 20 21:17:23.993240 sshd[5412]: Connection closed by 10.0.0.1 port 46370 Mar 20 21:17:23.993590 sshd-session[5410]: pam_unix(sshd:session): session closed for user core Mar 20 21:17:23.996166 systemd[1]: sshd@30-10.0.0.54:22-10.0.0.1:46370.service: Deactivated successfully. Mar 20 21:17:23.997944 systemd[1]: session-31.scope: Deactivated successfully. Mar 20 21:17:23.999601 systemd-logind[1470]: Session 31 logged out. Waiting for processes to exit. Mar 20 21:17:24.000545 systemd-logind[1470]: Removed session 31. Mar 20 21:17:29.004573 systemd[1]: Started sshd@31-10.0.0.54:22-10.0.0.1:46376.service - OpenSSH per-connection server daemon (10.0.0.1:46376). Mar 20 21:17:29.062184 sshd[5425]: Accepted publickey for core from 10.0.0.1 port 46376 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:17:29.063524 sshd-session[5425]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:17:29.068106 systemd-logind[1470]: New session 32 of user core. Mar 20 21:17:29.075376 systemd[1]: Started session-32.scope - Session 32 of User core. Mar 20 21:17:29.195530 sshd[5427]: Connection closed by 10.0.0.1 port 46376 Mar 20 21:17:29.195856 sshd-session[5425]: pam_unix(sshd:session): session closed for user core Mar 20 21:17:29.199041 systemd[1]: sshd@31-10.0.0.54:22-10.0.0.1:46376.service: Deactivated successfully. Mar 20 21:17:29.202589 systemd[1]: session-32.scope: Deactivated successfully. Mar 20 21:17:29.205560 systemd-logind[1470]: Session 32 logged out. Waiting for processes to exit. Mar 20 21:17:29.206607 systemd-logind[1470]: Removed session 32. Mar 20 21:17:34.210668 systemd[1]: Started sshd@32-10.0.0.54:22-10.0.0.1:39372.service - OpenSSH per-connection server daemon (10.0.0.1:39372). Mar 20 21:17:34.261358 sshd[5442]: Accepted publickey for core from 10.0.0.1 port 39372 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:17:34.262543 sshd-session[5442]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:17:34.266550 systemd-logind[1470]: New session 33 of user core. Mar 20 21:17:34.273419 systemd[1]: Started session-33.scope - Session 33 of User core. Mar 20 21:17:34.401088 sshd[5444]: Connection closed by 10.0.0.1 port 39372 Mar 20 21:17:34.401936 sshd-session[5442]: pam_unix(sshd:session): session closed for user core Mar 20 21:17:34.406658 systemd[1]: sshd@32-10.0.0.54:22-10.0.0.1:39372.service: Deactivated successfully. Mar 20 21:17:34.408410 systemd[1]: session-33.scope: Deactivated successfully. Mar 20 21:17:34.409889 systemd-logind[1470]: Session 33 logged out. Waiting for processes to exit. Mar 20 21:17:34.411454 systemd-logind[1470]: Removed session 33. Mar 20 21:17:34.727901 containerd[1485]: time="2025-03-20T21:17:34.727844404Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3298a708f055861f6f20ce36a7ee251ccfc70f10f7107e13a9514eaec3f5beb1\" id:\"77c987f0c3e797a73082300cf1c4a732b0b7889c3fee33377df3849276469232\" pid:5469 exited_at:{seconds:1742505454 nanos:727313596}" Mar 20 21:17:39.412505 systemd[1]: Started sshd@33-10.0.0.54:22-10.0.0.1:39384.service - OpenSSH per-connection server daemon (10.0.0.1:39384). Mar 20 21:17:39.473737 sshd[5483]: Accepted publickey for core from 10.0.0.1 port 39384 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:17:39.475111 sshd-session[5483]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:17:39.479278 systemd-logind[1470]: New session 34 of user core. Mar 20 21:17:39.485577 systemd[1]: Started session-34.scope - Session 34 of User core. Mar 20 21:17:39.620661 sshd[5485]: Connection closed by 10.0.0.1 port 39384 Mar 20 21:17:39.621064 sshd-session[5483]: pam_unix(sshd:session): session closed for user core Mar 20 21:17:39.624511 systemd[1]: sshd@33-10.0.0.54:22-10.0.0.1:39384.service: Deactivated successfully. Mar 20 21:17:39.626613 systemd[1]: session-34.scope: Deactivated successfully. Mar 20 21:17:39.627151 systemd-logind[1470]: Session 34 logged out. Waiting for processes to exit. Mar 20 21:17:39.627912 systemd-logind[1470]: Removed session 34. Mar 20 21:17:44.639642 systemd[1]: Started sshd@34-10.0.0.54:22-10.0.0.1:58888.service - OpenSSH per-connection server daemon (10.0.0.1:58888). Mar 20 21:17:44.692659 sshd[5501]: Accepted publickey for core from 10.0.0.1 port 58888 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:17:44.693786 sshd-session[5501]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:17:44.698447 systemd-logind[1470]: New session 35 of user core. Mar 20 21:17:44.707447 systemd[1]: Started session-35.scope - Session 35 of User core. Mar 20 21:17:44.849889 sshd[5503]: Connection closed by 10.0.0.1 port 58888 Mar 20 21:17:44.850234 sshd-session[5501]: pam_unix(sshd:session): session closed for user core Mar 20 21:17:44.853495 systemd[1]: sshd@34-10.0.0.54:22-10.0.0.1:58888.service: Deactivated successfully. Mar 20 21:17:44.855458 systemd[1]: session-35.scope: Deactivated successfully. Mar 20 21:17:44.856143 systemd-logind[1470]: Session 35 logged out. Waiting for processes to exit. Mar 20 21:17:44.856841 systemd-logind[1470]: Removed session 35. Mar 20 21:17:47.818384 containerd[1485]: time="2025-03-20T21:17:47.818340305Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c14e33d2f5b60700e68dcfd113314dd88a94ba5b5df327532409a6162dc44321\" id:\"ade70b92b797a4fda9506bdcde4daf8b4e2433caab16cff3b2295d243e859083\" pid:5527 exited_at:{seconds:1742505467 nanos:818006821}" Mar 20 21:17:49.861537 systemd[1]: Started sshd@35-10.0.0.54:22-10.0.0.1:58896.service - OpenSSH per-connection server daemon (10.0.0.1:58896). Mar 20 21:17:49.916941 sshd[5538]: Accepted publickey for core from 10.0.0.1 port 58896 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:17:49.918050 sshd-session[5538]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:17:49.921817 systemd-logind[1470]: New session 36 of user core. Mar 20 21:17:49.932360 systemd[1]: Started session-36.scope - Session 36 of User core. Mar 20 21:17:50.043102 sshd[5540]: Connection closed by 10.0.0.1 port 58896 Mar 20 21:17:50.043442 sshd-session[5538]: pam_unix(sshd:session): session closed for user core Mar 20 21:17:50.047353 systemd[1]: sshd@35-10.0.0.54:22-10.0.0.1:58896.service: Deactivated successfully. Mar 20 21:17:50.049139 systemd[1]: session-36.scope: Deactivated successfully. Mar 20 21:17:50.049835 systemd-logind[1470]: Session 36 logged out. Waiting for processes to exit. Mar 20 21:17:50.050550 systemd-logind[1470]: Removed session 36. Mar 20 21:17:51.706588 containerd[1485]: time="2025-03-20T21:17:51.706542930Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c14e33d2f5b60700e68dcfd113314dd88a94ba5b5df327532409a6162dc44321\" id:\"44366e98c4dc5e4347b505ed5cb647564057f2f76d02c43201b3365edd848242\" pid:5565 exited_at:{seconds:1742505471 nanos:706339887}" Mar 20 21:17:55.054546 systemd[1]: Started sshd@36-10.0.0.54:22-10.0.0.1:38330.service - OpenSSH per-connection server daemon (10.0.0.1:38330). Mar 20 21:17:55.110011 sshd[5576]: Accepted publickey for core from 10.0.0.1 port 38330 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:17:55.111183 sshd-session[5576]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:17:55.115294 systemd-logind[1470]: New session 37 of user core. Mar 20 21:17:55.127388 systemd[1]: Started session-37.scope - Session 37 of User core. Mar 20 21:17:55.241186 sshd[5578]: Connection closed by 10.0.0.1 port 38330 Mar 20 21:17:55.241532 sshd-session[5576]: pam_unix(sshd:session): session closed for user core Mar 20 21:17:55.244790 systemd[1]: sshd@36-10.0.0.54:22-10.0.0.1:38330.service: Deactivated successfully. Mar 20 21:17:55.246657 systemd[1]: session-37.scope: Deactivated successfully. Mar 20 21:17:55.247323 systemd-logind[1470]: Session 37 logged out. Waiting for processes to exit. Mar 20 21:17:55.248259 systemd-logind[1470]: Removed session 37. Mar 20 21:18:00.252581 systemd[1]: Started sshd@37-10.0.0.54:22-10.0.0.1:38346.service - OpenSSH per-connection server daemon (10.0.0.1:38346). Mar 20 21:18:00.306483 sshd[5592]: Accepted publickey for core from 10.0.0.1 port 38346 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:18:00.307674 sshd-session[5592]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:18:00.311941 systemd-logind[1470]: New session 38 of user core. Mar 20 21:18:00.321352 systemd[1]: Started session-38.scope - Session 38 of User core. Mar 20 21:18:00.431932 sshd[5594]: Connection closed by 10.0.0.1 port 38346 Mar 20 21:18:00.432471 sshd-session[5592]: pam_unix(sshd:session): session closed for user core Mar 20 21:18:00.436090 systemd[1]: sshd@37-10.0.0.54:22-10.0.0.1:38346.service: Deactivated successfully. Mar 20 21:18:00.438102 systemd[1]: session-38.scope: Deactivated successfully. Mar 20 21:18:00.438847 systemd-logind[1470]: Session 38 logged out. Waiting for processes to exit. Mar 20 21:18:00.439682 systemd-logind[1470]: Removed session 38. Mar 20 21:18:04.733715 containerd[1485]: time="2025-03-20T21:18:04.733658052Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3298a708f055861f6f20ce36a7ee251ccfc70f10f7107e13a9514eaec3f5beb1\" id:\"8b38f91de0af845f3d25496ac1e26f64c7062a9a2130d06948aa630887e15f72\" pid:5621 exited_at:{seconds:1742505484 nanos:733333089}" Mar 20 21:18:05.443747 systemd[1]: Started sshd@38-10.0.0.54:22-10.0.0.1:40116.service - OpenSSH per-connection server daemon (10.0.0.1:40116). Mar 20 21:18:05.500312 sshd[5635]: Accepted publickey for core from 10.0.0.1 port 40116 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:18:05.501587 sshd-session[5635]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:18:05.505813 systemd-logind[1470]: New session 39 of user core. Mar 20 21:18:05.521388 systemd[1]: Started session-39.scope - Session 39 of User core. Mar 20 21:18:05.641777 sshd[5637]: Connection closed by 10.0.0.1 port 40116 Mar 20 21:18:05.642133 sshd-session[5635]: pam_unix(sshd:session): session closed for user core Mar 20 21:18:05.645725 systemd[1]: sshd@38-10.0.0.54:22-10.0.0.1:40116.service: Deactivated successfully. Mar 20 21:18:05.647934 systemd[1]: session-39.scope: Deactivated successfully. Mar 20 21:18:05.648877 systemd-logind[1470]: Session 39 logged out. Waiting for processes to exit. Mar 20 21:18:05.649790 systemd-logind[1470]: Removed session 39. Mar 20 21:18:10.654277 systemd[1]: Started sshd@39-10.0.0.54:22-10.0.0.1:40130.service - OpenSSH per-connection server daemon (10.0.0.1:40130). Mar 20 21:18:10.706717 sshd[5651]: Accepted publickey for core from 10.0.0.1 port 40130 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:18:10.707883 sshd-session[5651]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:18:10.712312 systemd-logind[1470]: New session 40 of user core. Mar 20 21:18:10.722369 systemd[1]: Started session-40.scope - Session 40 of User core. Mar 20 21:18:10.830486 sshd[5653]: Connection closed by 10.0.0.1 port 40130 Mar 20 21:18:10.830800 sshd-session[5651]: pam_unix(sshd:session): session closed for user core Mar 20 21:18:10.833869 systemd[1]: sshd@39-10.0.0.54:22-10.0.0.1:40130.service: Deactivated successfully. Mar 20 21:18:10.835670 systemd[1]: session-40.scope: Deactivated successfully. Mar 20 21:18:10.836297 systemd-logind[1470]: Session 40 logged out. Waiting for processes to exit. Mar 20 21:18:10.837146 systemd-logind[1470]: Removed session 40. Mar 20 21:18:15.846720 systemd[1]: Started sshd@40-10.0.0.54:22-10.0.0.1:51496.service - OpenSSH per-connection server daemon (10.0.0.1:51496). Mar 20 21:18:15.903285 sshd[5670]: Accepted publickey for core from 10.0.0.1 port 51496 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:18:15.904718 sshd-session[5670]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:18:15.909179 systemd-logind[1470]: New session 41 of user core. Mar 20 21:18:15.921402 systemd[1]: Started session-41.scope - Session 41 of User core. Mar 20 21:18:16.049508 sshd[5672]: Connection closed by 10.0.0.1 port 51496 Mar 20 21:18:16.049873 sshd-session[5670]: pam_unix(sshd:session): session closed for user core Mar 20 21:18:16.052406 systemd[1]: sshd@40-10.0.0.54:22-10.0.0.1:51496.service: Deactivated successfully. Mar 20 21:18:16.054138 systemd[1]: session-41.scope: Deactivated successfully. Mar 20 21:18:16.055535 systemd-logind[1470]: Session 41 logged out. Waiting for processes to exit. Mar 20 21:18:16.056453 systemd-logind[1470]: Removed session 41. Mar 20 21:18:21.061618 systemd[1]: Started sshd@41-10.0.0.54:22-10.0.0.1:51502.service - OpenSSH per-connection server daemon (10.0.0.1:51502). Mar 20 21:18:21.112541 sshd[5692]: Accepted publickey for core from 10.0.0.1 port 51502 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:18:21.113767 sshd-session[5692]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:18:21.117478 systemd-logind[1470]: New session 42 of user core. Mar 20 21:18:21.131413 systemd[1]: Started session-42.scope - Session 42 of User core. Mar 20 21:18:21.248855 sshd[5694]: Connection closed by 10.0.0.1 port 51502 Mar 20 21:18:21.249163 sshd-session[5692]: pam_unix(sshd:session): session closed for user core Mar 20 21:18:21.252275 systemd[1]: sshd@41-10.0.0.54:22-10.0.0.1:51502.service: Deactivated successfully. Mar 20 21:18:21.253981 systemd[1]: session-42.scope: Deactivated successfully. Mar 20 21:18:21.254575 systemd-logind[1470]: Session 42 logged out. Waiting for processes to exit. Mar 20 21:18:21.255354 systemd-logind[1470]: Removed session 42. Mar 20 21:18:21.705050 containerd[1485]: time="2025-03-20T21:18:21.704989913Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c14e33d2f5b60700e68dcfd113314dd88a94ba5b5df327532409a6162dc44321\" id:\"09bb7096b7d74e9a9aae86d4eda4a8fac46fd352977b93788a53686a67621be0\" pid:5718 exited_at:{seconds:1742505501 nanos:704792142}" Mar 20 21:18:26.261116 systemd[1]: Started sshd@42-10.0.0.54:22-10.0.0.1:43552.service - OpenSSH per-connection server daemon (10.0.0.1:43552). Mar 20 21:18:26.314152 sshd[5731]: Accepted publickey for core from 10.0.0.1 port 43552 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:18:26.315525 sshd-session[5731]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:18:26.320122 systemd-logind[1470]: New session 43 of user core. Mar 20 21:18:26.330375 systemd[1]: Started session-43.scope - Session 43 of User core. Mar 20 21:18:26.446903 sshd[5733]: Connection closed by 10.0.0.1 port 43552 Mar 20 21:18:26.447253 sshd-session[5731]: pam_unix(sshd:session): session closed for user core Mar 20 21:18:26.450503 systemd[1]: sshd@42-10.0.0.54:22-10.0.0.1:43552.service: Deactivated successfully. Mar 20 21:18:26.452909 systemd[1]: session-43.scope: Deactivated successfully. Mar 20 21:18:26.453587 systemd-logind[1470]: Session 43 logged out. Waiting for processes to exit. Mar 20 21:18:26.454428 systemd-logind[1470]: Removed session 43. Mar 20 21:18:31.458547 systemd[1]: Started sshd@43-10.0.0.54:22-10.0.0.1:43554.service - OpenSSH per-connection server daemon (10.0.0.1:43554). Mar 20 21:18:31.509588 sshd[5746]: Accepted publickey for core from 10.0.0.1 port 43554 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:18:31.510741 sshd-session[5746]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:18:31.514187 systemd-logind[1470]: New session 44 of user core. Mar 20 21:18:31.522405 systemd[1]: Started session-44.scope - Session 44 of User core. Mar 20 21:18:31.635781 sshd[5748]: Connection closed by 10.0.0.1 port 43554 Mar 20 21:18:31.636294 sshd-session[5746]: pam_unix(sshd:session): session closed for user core Mar 20 21:18:31.639646 systemd[1]: sshd@43-10.0.0.54:22-10.0.0.1:43554.service: Deactivated successfully. Mar 20 21:18:31.641386 systemd[1]: session-44.scope: Deactivated successfully. Mar 20 21:18:31.641964 systemd-logind[1470]: Session 44 logged out. Waiting for processes to exit. Mar 20 21:18:31.642658 systemd-logind[1470]: Removed session 44. Mar 20 21:18:34.732480 containerd[1485]: time="2025-03-20T21:18:34.732434409Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3298a708f055861f6f20ce36a7ee251ccfc70f10f7107e13a9514eaec3f5beb1\" id:\"d16ed9e830e90d31a47985b2a2428eb97ab71655530a6c644eb3174dbae4d254\" pid:5772 exited_at:{seconds:1742505514 nanos:731973426}" Mar 20 21:18:36.647560 systemd[1]: Started sshd@44-10.0.0.54:22-10.0.0.1:58048.service - OpenSSH per-connection server daemon (10.0.0.1:58048). Mar 20 21:18:36.703635 sshd[5786]: Accepted publickey for core from 10.0.0.1 port 58048 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:18:36.704771 sshd-session[5786]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:18:36.708944 systemd-logind[1470]: New session 45 of user core. Mar 20 21:18:36.715344 systemd[1]: Started session-45.scope - Session 45 of User core. Mar 20 21:18:36.822174 sshd[5788]: Connection closed by 10.0.0.1 port 58048 Mar 20 21:18:36.822653 sshd-session[5786]: pam_unix(sshd:session): session closed for user core Mar 20 21:18:36.825251 systemd[1]: sshd@44-10.0.0.54:22-10.0.0.1:58048.service: Deactivated successfully. Mar 20 21:18:36.827007 systemd[1]: session-45.scope: Deactivated successfully. Mar 20 21:18:36.828295 systemd-logind[1470]: Session 45 logged out. Waiting for processes to exit. Mar 20 21:18:36.829160 systemd-logind[1470]: Removed session 45. Mar 20 21:18:41.839448 systemd[1]: Started sshd@45-10.0.0.54:22-10.0.0.1:58058.service - OpenSSH per-connection server daemon (10.0.0.1:58058). Mar 20 21:18:41.887023 sshd[5803]: Accepted publickey for core from 10.0.0.1 port 58058 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:18:41.888253 sshd-session[5803]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:18:41.892738 systemd-logind[1470]: New session 46 of user core. Mar 20 21:18:41.899366 systemd[1]: Started session-46.scope - Session 46 of User core. Mar 20 21:18:42.010256 sshd[5805]: Connection closed by 10.0.0.1 port 58058 Mar 20 21:18:42.010407 sshd-session[5803]: pam_unix(sshd:session): session closed for user core Mar 20 21:18:42.013314 systemd[1]: sshd@45-10.0.0.54:22-10.0.0.1:58058.service: Deactivated successfully. Mar 20 21:18:42.015205 systemd[1]: session-46.scope: Deactivated successfully. Mar 20 21:18:42.016657 systemd-logind[1470]: Session 46 logged out. Waiting for processes to exit. Mar 20 21:18:42.017627 systemd-logind[1470]: Removed session 46. Mar 20 21:18:47.026516 systemd[1]: Started sshd@46-10.0.0.54:22-10.0.0.1:51468.service - OpenSSH per-connection server daemon (10.0.0.1:51468). Mar 20 21:18:47.078971 sshd[5835]: Accepted publickey for core from 10.0.0.1 port 51468 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:18:47.080120 sshd-session[5835]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:18:47.083787 systemd-logind[1470]: New session 47 of user core. Mar 20 21:18:47.091420 systemd[1]: Started session-47.scope - Session 47 of User core. Mar 20 21:18:47.198996 sshd[5837]: Connection closed by 10.0.0.1 port 51468 Mar 20 21:18:47.199579 sshd-session[5835]: pam_unix(sshd:session): session closed for user core Mar 20 21:18:47.210928 systemd[1]: sshd@46-10.0.0.54:22-10.0.0.1:51468.service: Deactivated successfully. Mar 20 21:18:47.213178 systemd[1]: session-47.scope: Deactivated successfully. Mar 20 21:18:47.215449 systemd-logind[1470]: Session 47 logged out. Waiting for processes to exit. Mar 20 21:18:47.216814 systemd[1]: Started sshd@47-10.0.0.54:22-10.0.0.1:51478.service - OpenSSH per-connection server daemon (10.0.0.1:51478). Mar 20 21:18:47.217866 systemd-logind[1470]: Removed session 47. Mar 20 21:18:47.268736 sshd[5849]: Accepted publickey for core from 10.0.0.1 port 51478 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:18:47.269863 sshd-session[5849]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:18:47.274648 systemd-logind[1470]: New session 48 of user core. Mar 20 21:18:47.289355 systemd[1]: Started session-48.scope - Session 48 of User core. Mar 20 21:18:47.488322 sshd[5852]: Connection closed by 10.0.0.1 port 51478 Mar 20 21:18:47.489103 sshd-session[5849]: pam_unix(sshd:session): session closed for user core Mar 20 21:18:47.503456 systemd[1]: sshd@47-10.0.0.54:22-10.0.0.1:51478.service: Deactivated successfully. Mar 20 21:18:47.505002 systemd[1]: session-48.scope: Deactivated successfully. Mar 20 21:18:47.505672 systemd-logind[1470]: Session 48 logged out. Waiting for processes to exit. Mar 20 21:18:47.507463 systemd[1]: Started sshd@48-10.0.0.54:22-10.0.0.1:51488.service - OpenSSH per-connection server daemon (10.0.0.1:51488). Mar 20 21:18:47.508832 systemd-logind[1470]: Removed session 48. Mar 20 21:18:47.569474 sshd[5862]: Accepted publickey for core from 10.0.0.1 port 51488 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:18:47.570709 sshd-session[5862]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:18:47.574413 systemd-logind[1470]: New session 49 of user core. Mar 20 21:18:47.584341 systemd[1]: Started session-49.scope - Session 49 of User core. Mar 20 21:18:47.813549 containerd[1485]: time="2025-03-20T21:18:47.813505511Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c14e33d2f5b60700e68dcfd113314dd88a94ba5b5df327532409a6162dc44321\" id:\"c1e5ebe1aabdff8f11c89834d8261f0c0b15da979c6ec5e35896d704407d3f49\" pid:5884 exited_at:{seconds:1742505527 nanos:813253541}" Mar 20 21:18:49.054483 sshd[5865]: Connection closed by 10.0.0.1 port 51488 Mar 20 21:18:49.055161 sshd-session[5862]: pam_unix(sshd:session): session closed for user core Mar 20 21:18:49.065920 systemd[1]: sshd@48-10.0.0.54:22-10.0.0.1:51488.service: Deactivated successfully. Mar 20 21:18:49.069607 systemd[1]: session-49.scope: Deactivated successfully. Mar 20 21:18:49.069843 systemd[1]: session-49.scope: Consumed 507ms CPU time, 68.1M memory peak. Mar 20 21:18:49.070444 systemd-logind[1470]: Session 49 logged out. Waiting for processes to exit. Mar 20 21:18:49.074822 systemd[1]: Started sshd@49-10.0.0.54:22-10.0.0.1:51498.service - OpenSSH per-connection server daemon (10.0.0.1:51498). Mar 20 21:18:49.076468 systemd-logind[1470]: Removed session 49. Mar 20 21:18:49.131707 sshd[5907]: Accepted publickey for core from 10.0.0.1 port 51498 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:18:49.133061 sshd-session[5907]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:18:49.137716 systemd-logind[1470]: New session 50 of user core. Mar 20 21:18:49.151446 systemd[1]: Started session-50.scope - Session 50 of User core. Mar 20 21:18:49.443269 sshd[5910]: Connection closed by 10.0.0.1 port 51498 Mar 20 21:18:49.443342 sshd-session[5907]: pam_unix(sshd:session): session closed for user core Mar 20 21:18:49.455291 systemd[1]: sshd@49-10.0.0.54:22-10.0.0.1:51498.service: Deactivated successfully. Mar 20 21:18:49.457498 systemd[1]: session-50.scope: Deactivated successfully. Mar 20 21:18:49.458243 systemd-logind[1470]: Session 50 logged out. Waiting for processes to exit. Mar 20 21:18:49.460739 systemd[1]: Started sshd@50-10.0.0.54:22-10.0.0.1:51514.service - OpenSSH per-connection server daemon (10.0.0.1:51514). Mar 20 21:18:49.461793 systemd-logind[1470]: Removed session 50. Mar 20 21:18:49.514273 sshd[5920]: Accepted publickey for core from 10.0.0.1 port 51514 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:18:49.515491 sshd-session[5920]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:18:49.519974 systemd-logind[1470]: New session 51 of user core. Mar 20 21:18:49.529375 systemd[1]: Started session-51.scope - Session 51 of User core. Mar 20 21:18:49.691993 sshd[5923]: Connection closed by 10.0.0.1 port 51514 Mar 20 21:18:49.692684 sshd-session[5920]: pam_unix(sshd:session): session closed for user core Mar 20 21:18:49.696295 systemd[1]: sshd@50-10.0.0.54:22-10.0.0.1:51514.service: Deactivated successfully. Mar 20 21:18:49.698202 systemd[1]: session-51.scope: Deactivated successfully. Mar 20 21:18:49.698869 systemd-logind[1470]: Session 51 logged out. Waiting for processes to exit. Mar 20 21:18:49.699820 systemd-logind[1470]: Removed session 51. Mar 20 21:18:51.707985 containerd[1485]: time="2025-03-20T21:18:51.707943741Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c14e33d2f5b60700e68dcfd113314dd88a94ba5b5df327532409a6162dc44321\" id:\"3a57466339d30796a46c26e49295a8d2ac6f14be1494398b79fd441a44459b34\" pid:5949 exited_at:{seconds:1742505531 nanos:707483962}" Mar 20 21:18:54.705807 systemd[1]: Started sshd@51-10.0.0.54:22-10.0.0.1:34900.service - OpenSSH per-connection server daemon (10.0.0.1:34900). Mar 20 21:18:54.757943 sshd[5959]: Accepted publickey for core from 10.0.0.1 port 34900 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:18:54.759376 sshd-session[5959]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:18:54.763520 systemd-logind[1470]: New session 52 of user core. Mar 20 21:18:54.773436 systemd[1]: Started session-52.scope - Session 52 of User core. Mar 20 21:18:54.889612 sshd[5961]: Connection closed by 10.0.0.1 port 34900 Mar 20 21:18:54.890154 sshd-session[5959]: pam_unix(sshd:session): session closed for user core Mar 20 21:18:54.893800 systemd[1]: sshd@51-10.0.0.54:22-10.0.0.1:34900.service: Deactivated successfully. Mar 20 21:18:54.896492 systemd[1]: session-52.scope: Deactivated successfully. Mar 20 21:18:54.897435 systemd-logind[1470]: Session 52 logged out. Waiting for processes to exit. Mar 20 21:18:54.898514 systemd-logind[1470]: Removed session 52. Mar 20 21:18:59.906458 systemd[1]: Started sshd@52-10.0.0.54:22-10.0.0.1:34902.service - OpenSSH per-connection server daemon (10.0.0.1:34902). Mar 20 21:18:59.961078 sshd[5974]: Accepted publickey for core from 10.0.0.1 port 34902 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:18:59.962277 sshd-session[5974]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:18:59.965953 systemd-logind[1470]: New session 53 of user core. Mar 20 21:18:59.974367 systemd[1]: Started session-53.scope - Session 53 of User core. Mar 20 21:19:00.089019 sshd[5976]: Connection closed by 10.0.0.1 port 34902 Mar 20 21:19:00.089406 sshd-session[5974]: pam_unix(sshd:session): session closed for user core Mar 20 21:19:00.092829 systemd[1]: sshd@52-10.0.0.54:22-10.0.0.1:34902.service: Deactivated successfully. Mar 20 21:19:00.095420 systemd[1]: session-53.scope: Deactivated successfully. Mar 20 21:19:00.096435 systemd-logind[1470]: Session 53 logged out. Waiting for processes to exit. Mar 20 21:19:00.097799 systemd-logind[1470]: Removed session 53. Mar 20 21:19:04.733908 containerd[1485]: time="2025-03-20T21:19:04.733850685Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3298a708f055861f6f20ce36a7ee251ccfc70f10f7107e13a9514eaec3f5beb1\" id:\"6eeb7d62f943d501d331cbaba4d8a352a9b5d16fddf7ae571cecccacdcd42ed0\" pid:6002 exited_at:{seconds:1742505544 nanos:733386188}" Mar 20 21:19:05.105289 systemd[1]: Started sshd@53-10.0.0.54:22-10.0.0.1:34328.service - OpenSSH per-connection server daemon (10.0.0.1:34328). Mar 20 21:19:05.159825 sshd[6016]: Accepted publickey for core from 10.0.0.1 port 34328 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:19:05.161043 sshd-session[6016]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:19:05.165082 systemd-logind[1470]: New session 54 of user core. Mar 20 21:19:05.176389 systemd[1]: Started session-54.scope - Session 54 of User core. Mar 20 21:19:05.298209 sshd[6018]: Connection closed by 10.0.0.1 port 34328 Mar 20 21:19:05.298773 sshd-session[6016]: pam_unix(sshd:session): session closed for user core Mar 20 21:19:05.301838 systemd[1]: sshd@53-10.0.0.54:22-10.0.0.1:34328.service: Deactivated successfully. Mar 20 21:19:05.303612 systemd[1]: session-54.scope: Deactivated successfully. Mar 20 21:19:05.304886 systemd-logind[1470]: Session 54 logged out. Waiting for processes to exit. Mar 20 21:19:05.305730 systemd-logind[1470]: Removed session 54. Mar 20 21:19:10.315924 systemd[1]: Started sshd@54-10.0.0.54:22-10.0.0.1:34342.service - OpenSSH per-connection server daemon (10.0.0.1:34342). Mar 20 21:19:10.359898 sshd[6032]: Accepted publickey for core from 10.0.0.1 port 34342 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:19:10.361091 sshd-session[6032]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:19:10.364790 systemd-logind[1470]: New session 55 of user core. Mar 20 21:19:10.374430 systemd[1]: Started session-55.scope - Session 55 of User core. Mar 20 21:19:10.479773 sshd[6034]: Connection closed by 10.0.0.1 port 34342 Mar 20 21:19:10.480291 sshd-session[6032]: pam_unix(sshd:session): session closed for user core Mar 20 21:19:10.482941 systemd[1]: sshd@54-10.0.0.54:22-10.0.0.1:34342.service: Deactivated successfully. Mar 20 21:19:10.484592 systemd[1]: session-55.scope: Deactivated successfully. Mar 20 21:19:10.485870 systemd-logind[1470]: Session 55 logged out. Waiting for processes to exit. Mar 20 21:19:10.486895 systemd-logind[1470]: Removed session 55. Mar 20 21:19:15.491690 systemd[1]: Started sshd@55-10.0.0.54:22-10.0.0.1:57934.service - OpenSSH per-connection server daemon (10.0.0.1:57934). Mar 20 21:19:15.546656 sshd[6051]: Accepted publickey for core from 10.0.0.1 port 57934 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:19:15.547951 sshd-session[6051]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:19:15.552636 systemd-logind[1470]: New session 56 of user core. Mar 20 21:19:15.568442 systemd[1]: Started session-56.scope - Session 56 of User core. Mar 20 21:19:15.679155 sshd[6053]: Connection closed by 10.0.0.1 port 57934 Mar 20 21:19:15.679720 sshd-session[6051]: pam_unix(sshd:session): session closed for user core Mar 20 21:19:15.682990 systemd[1]: sshd@55-10.0.0.54:22-10.0.0.1:57934.service: Deactivated successfully. Mar 20 21:19:15.684830 systemd[1]: session-56.scope: Deactivated successfully. Mar 20 21:19:15.686013 systemd-logind[1470]: Session 56 logged out. Waiting for processes to exit. Mar 20 21:19:15.686964 systemd-logind[1470]: Removed session 56. Mar 20 21:19:20.690775 systemd[1]: Started sshd@56-10.0.0.54:22-10.0.0.1:57936.service - OpenSSH per-connection server daemon (10.0.0.1:57936). Mar 20 21:19:20.747072 sshd[6067]: Accepted publickey for core from 10.0.0.1 port 57936 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:19:20.748437 sshd-session[6067]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:19:20.752678 systemd-logind[1470]: New session 57 of user core. Mar 20 21:19:20.761390 systemd[1]: Started session-57.scope - Session 57 of User core. Mar 20 21:19:20.880349 sshd[6069]: Connection closed by 10.0.0.1 port 57936 Mar 20 21:19:20.880908 sshd-session[6067]: pam_unix(sshd:session): session closed for user core Mar 20 21:19:20.884408 systemd[1]: sshd@56-10.0.0.54:22-10.0.0.1:57936.service: Deactivated successfully. Mar 20 21:19:20.886310 systemd[1]: session-57.scope: Deactivated successfully. Mar 20 21:19:20.886957 systemd-logind[1470]: Session 57 logged out. Waiting for processes to exit. Mar 20 21:19:20.887908 systemd-logind[1470]: Removed session 57. Mar 20 21:19:21.706775 containerd[1485]: time="2025-03-20T21:19:21.706732337Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c14e33d2f5b60700e68dcfd113314dd88a94ba5b5df327532409a6162dc44321\" id:\"28c8958d875b645c5534214161aeaaa783b94d9971406bea1d3039fa9b0dfd16\" pid:6094 exited_at:{seconds:1742505561 nanos:706552092}" Mar 20 21:19:25.896853 systemd[1]: Started sshd@57-10.0.0.54:22-10.0.0.1:43280.service - OpenSSH per-connection server daemon (10.0.0.1:43280). Mar 20 21:19:25.955574 sshd[6105]: Accepted publickey for core from 10.0.0.1 port 43280 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:19:25.956965 sshd-session[6105]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:19:25.961045 systemd-logind[1470]: New session 58 of user core. Mar 20 21:19:25.973415 systemd[1]: Started session-58.scope - Session 58 of User core. Mar 20 21:19:26.088098 sshd[6107]: Connection closed by 10.0.0.1 port 43280 Mar 20 21:19:26.088560 sshd-session[6105]: pam_unix(sshd:session): session closed for user core Mar 20 21:19:26.091986 systemd[1]: sshd@57-10.0.0.54:22-10.0.0.1:43280.service: Deactivated successfully. Mar 20 21:19:26.094057 systemd[1]: session-58.scope: Deactivated successfully. Mar 20 21:19:26.095890 systemd-logind[1470]: Session 58 logged out. Waiting for processes to exit. Mar 20 21:19:26.096867 systemd-logind[1470]: Removed session 58. Mar 20 21:19:31.103812 systemd[1]: Started sshd@58-10.0.0.54:22-10.0.0.1:43290.service - OpenSSH per-connection server daemon (10.0.0.1:43290). Mar 20 21:19:31.157807 sshd[6122]: Accepted publickey for core from 10.0.0.1 port 43290 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:19:31.159575 sshd-session[6122]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:19:31.164257 systemd-logind[1470]: New session 59 of user core. Mar 20 21:19:31.173390 systemd[1]: Started session-59.scope - Session 59 of User core. Mar 20 21:19:31.284297 sshd[6124]: Connection closed by 10.0.0.1 port 43290 Mar 20 21:19:31.284850 sshd-session[6122]: pam_unix(sshd:session): session closed for user core Mar 20 21:19:31.287606 systemd[1]: session-59.scope: Deactivated successfully. Mar 20 21:19:31.288972 systemd[1]: sshd@58-10.0.0.54:22-10.0.0.1:43290.service: Deactivated successfully. Mar 20 21:19:31.291118 systemd-logind[1470]: Session 59 logged out. Waiting for processes to exit. Mar 20 21:19:31.292019 systemd-logind[1470]: Removed session 59. Mar 20 21:19:34.726859 containerd[1485]: time="2025-03-20T21:19:34.726788103Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3298a708f055861f6f20ce36a7ee251ccfc70f10f7107e13a9514eaec3f5beb1\" id:\"0acf8e225fd3e74b3dfd9ff8ad971ecc3f433468998de7c0ed508f8879a95ae3\" pid:6149 exited_at:{seconds:1742505574 nanos:726467614}" Mar 20 21:19:36.297289 systemd[1]: Started sshd@59-10.0.0.54:22-10.0.0.1:39378.service - OpenSSH per-connection server daemon (10.0.0.1:39378). Mar 20 21:19:36.353510 sshd[6163]: Accepted publickey for core from 10.0.0.1 port 39378 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:19:36.354610 sshd-session[6163]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:19:36.359178 systemd-logind[1470]: New session 60 of user core. Mar 20 21:19:36.365365 systemd[1]: Started session-60.scope - Session 60 of User core. Mar 20 21:19:36.487262 sshd[6165]: Connection closed by 10.0.0.1 port 39378 Mar 20 21:19:36.487828 sshd-session[6163]: pam_unix(sshd:session): session closed for user core Mar 20 21:19:36.491250 systemd[1]: sshd@59-10.0.0.54:22-10.0.0.1:39378.service: Deactivated successfully. Mar 20 21:19:36.494899 systemd[1]: session-60.scope: Deactivated successfully. Mar 20 21:19:36.495566 systemd-logind[1470]: Session 60 logged out. Waiting for processes to exit. Mar 20 21:19:36.496472 systemd-logind[1470]: Removed session 60. Mar 20 21:19:41.499725 systemd[1]: Started sshd@60-10.0.0.54:22-10.0.0.1:39380.service - OpenSSH per-connection server daemon (10.0.0.1:39380). Mar 20 21:19:41.560860 sshd[6182]: Accepted publickey for core from 10.0.0.1 port 39380 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:19:41.562115 sshd-session[6182]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:19:41.566682 systemd-logind[1470]: New session 61 of user core. Mar 20 21:19:41.572362 systemd[1]: Started session-61.scope - Session 61 of User core. Mar 20 21:19:41.716048 sshd[6186]: Connection closed by 10.0.0.1 port 39380 Mar 20 21:19:41.716495 sshd-session[6182]: pam_unix(sshd:session): session closed for user core Mar 20 21:19:41.719918 systemd[1]: sshd@60-10.0.0.54:22-10.0.0.1:39380.service: Deactivated successfully. Mar 20 21:19:41.722296 systemd[1]: session-61.scope: Deactivated successfully. Mar 20 21:19:41.722884 systemd-logind[1470]: Session 61 logged out. Waiting for processes to exit. Mar 20 21:19:41.723685 systemd-logind[1470]: Removed session 61. Mar 20 21:19:46.728147 systemd[1]: Started sshd@61-10.0.0.54:22-10.0.0.1:49986.service - OpenSSH per-connection server daemon (10.0.0.1:49986). Mar 20 21:19:46.785290 sshd[6200]: Accepted publickey for core from 10.0.0.1 port 49986 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:19:46.786545 sshd-session[6200]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:19:46.791425 systemd-logind[1470]: New session 62 of user core. Mar 20 21:19:46.797358 systemd[1]: Started session-62.scope - Session 62 of User core. Mar 20 21:19:46.933452 sshd[6203]: Connection closed by 10.0.0.1 port 49986 Mar 20 21:19:46.933961 sshd-session[6200]: pam_unix(sshd:session): session closed for user core Mar 20 21:19:46.937492 systemd[1]: sshd@61-10.0.0.54:22-10.0.0.1:49986.service: Deactivated successfully. Mar 20 21:19:46.939723 systemd[1]: session-62.scope: Deactivated successfully. Mar 20 21:19:46.940513 systemd-logind[1470]: Session 62 logged out. Waiting for processes to exit. Mar 20 21:19:46.941538 systemd-logind[1470]: Removed session 62. Mar 20 21:19:47.811830 containerd[1485]: time="2025-03-20T21:19:47.811788587Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c14e33d2f5b60700e68dcfd113314dd88a94ba5b5df327532409a6162dc44321\" id:\"56d8082fd484ca4644fc33a80ad901dcaaa117f679fb9c0abc8d53888f65b84a\" pid:6228 exited_at:{seconds:1742505587 nanos:811552701}" Mar 20 21:19:51.708699 containerd[1485]: time="2025-03-20T21:19:51.708658866Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c14e33d2f5b60700e68dcfd113314dd88a94ba5b5df327532409a6162dc44321\" id:\"91738dc4283179faff909c7e07328f95960a76eacc1f4b712fda83fc9ed4a976\" pid:6250 exited_at:{seconds:1742505591 nanos:708168534}" Mar 20 21:19:51.945699 systemd[1]: Started sshd@62-10.0.0.54:22-10.0.0.1:49990.service - OpenSSH per-connection server daemon (10.0.0.1:49990). Mar 20 21:19:52.002706 sshd[6261]: Accepted publickey for core from 10.0.0.1 port 49990 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:19:52.004305 sshd-session[6261]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:19:52.008985 systemd-logind[1470]: New session 63 of user core. Mar 20 21:19:52.017367 systemd[1]: Started session-63.scope - Session 63 of User core. Mar 20 21:19:52.168461 sshd[6263]: Connection closed by 10.0.0.1 port 49990 Mar 20 21:19:52.168838 sshd-session[6261]: pam_unix(sshd:session): session closed for user core Mar 20 21:19:52.171539 systemd[1]: sshd@62-10.0.0.54:22-10.0.0.1:49990.service: Deactivated successfully. Mar 20 21:19:52.173233 systemd[1]: session-63.scope: Deactivated successfully. Mar 20 21:19:52.175446 systemd-logind[1470]: Session 63 logged out. Waiting for processes to exit. Mar 20 21:19:52.176500 systemd-logind[1470]: Removed session 63. Mar 20 21:19:57.180489 systemd[1]: Started sshd@63-10.0.0.54:22-10.0.0.1:36418.service - OpenSSH per-connection server daemon (10.0.0.1:36418). Mar 20 21:19:57.235972 sshd[6276]: Accepted publickey for core from 10.0.0.1 port 36418 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:19:57.237066 sshd-session[6276]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:19:57.241128 systemd-logind[1470]: New session 64 of user core. Mar 20 21:19:57.254347 systemd[1]: Started session-64.scope - Session 64 of User core. Mar 20 21:19:57.372577 sshd[6278]: Connection closed by 10.0.0.1 port 36418 Mar 20 21:19:57.373101 sshd-session[6276]: pam_unix(sshd:session): session closed for user core Mar 20 21:19:57.376455 systemd[1]: sshd@63-10.0.0.54:22-10.0.0.1:36418.service: Deactivated successfully. Mar 20 21:19:57.378478 systemd[1]: session-64.scope: Deactivated successfully. Mar 20 21:19:57.379084 systemd-logind[1470]: Session 64 logged out. Waiting for processes to exit. Mar 20 21:19:57.380101 systemd-logind[1470]: Removed session 64. Mar 20 21:19:59.793103 containerd[1485]: time="2025-03-20T21:19:59.787834942Z" level=warning msg="container event discarded" container=bf1cd9c76fb6584ba737a16fa998bca0cb27eb73a90f43c71cd107df56cdbfc1 type=CONTAINER_CREATED_EVENT Mar 20 21:19:59.809280 containerd[1485]: time="2025-03-20T21:19:59.809246336Z" level=warning msg="container event discarded" container=bf1cd9c76fb6584ba737a16fa998bca0cb27eb73a90f43c71cd107df56cdbfc1 type=CONTAINER_STARTED_EVENT Mar 20 21:19:59.809280 containerd[1485]: time="2025-03-20T21:19:59.809276376Z" level=warning msg="container event discarded" container=c6ff0b709e5aa0311d65ae1f9ca839bb12942d73dd4e5c17066b05ef4f14c066 type=CONTAINER_CREATED_EVENT Mar 20 21:19:59.809375 containerd[1485]: time="2025-03-20T21:19:59.809285977Z" level=warning msg="container event discarded" container=c6ff0b709e5aa0311d65ae1f9ca839bb12942d73dd4e5c17066b05ef4f14c066 type=CONTAINER_STARTED_EVENT Mar 20 21:19:59.809375 containerd[1485]: time="2025-03-20T21:19:59.809296137Z" level=warning msg="container event discarded" container=de3abc48bdc665541ddedcf492559b015e8e6d6c54ecab378e29eeb218a07158 type=CONTAINER_CREATED_EVENT Mar 20 21:19:59.809375 containerd[1485]: time="2025-03-20T21:19:59.809303857Z" level=warning msg="container event discarded" container=de3abc48bdc665541ddedcf492559b015e8e6d6c54ecab378e29eeb218a07158 type=CONTAINER_STARTED_EVENT Mar 20 21:19:59.820651 containerd[1485]: time="2025-03-20T21:19:59.820612808Z" level=warning msg="container event discarded" container=25aae4bb122020e866107740e1d02419fe399ef4709ba2625efe6aca7b3cc9f9 type=CONTAINER_CREATED_EVENT Mar 20 21:19:59.831845 containerd[1485]: time="2025-03-20T21:19:59.831800837Z" level=warning msg="container event discarded" container=5b814b57126748bd8c6189166eba713cdae0bff67bbc0f7e18300b807337421d type=CONTAINER_CREATED_EVENT Mar 20 21:19:59.831845 containerd[1485]: time="2025-03-20T21:19:59.831826318Z" level=warning msg="container event discarded" container=9550a2a8546bdbccb1898a840c1920a481a7835b038827b866affb4dfc23e258 type=CONTAINER_CREATED_EVENT Mar 20 21:19:59.895132 containerd[1485]: time="2025-03-20T21:19:59.895081476Z" level=warning msg="container event discarded" container=25aae4bb122020e866107740e1d02419fe399ef4709ba2625efe6aca7b3cc9f9 type=CONTAINER_STARTED_EVENT Mar 20 21:19:59.915320 containerd[1485]: time="2025-03-20T21:19:59.915278120Z" level=warning msg="container event discarded" container=9550a2a8546bdbccb1898a840c1920a481a7835b038827b866affb4dfc23e258 type=CONTAINER_STARTED_EVENT Mar 20 21:19:59.915320 containerd[1485]: time="2025-03-20T21:19:59.915317361Z" level=warning msg="container event discarded" container=5b814b57126748bd8c6189166eba713cdae0bff67bbc0f7e18300b807337421d type=CONTAINER_STARTED_EVENT Mar 20 21:20:02.386746 systemd[1]: Started sshd@64-10.0.0.54:22-10.0.0.1:36430.service - OpenSSH per-connection server daemon (10.0.0.1:36430). Mar 20 21:20:02.437718 sshd[6292]: Accepted publickey for core from 10.0.0.1 port 36430 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:20:02.438848 sshd-session[6292]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:20:02.442348 systemd-logind[1470]: New session 65 of user core. Mar 20 21:20:02.450361 systemd[1]: Started session-65.scope - Session 65 of User core. Mar 20 21:20:02.570964 sshd[6294]: Connection closed by 10.0.0.1 port 36430 Mar 20 21:20:02.571310 sshd-session[6292]: pam_unix(sshd:session): session closed for user core Mar 20 21:20:02.574773 systemd[1]: sshd@64-10.0.0.54:22-10.0.0.1:36430.service: Deactivated successfully. Mar 20 21:20:02.577242 systemd[1]: session-65.scope: Deactivated successfully. Mar 20 21:20:02.578092 systemd-logind[1470]: Session 65 logged out. Waiting for processes to exit. Mar 20 21:20:02.578942 systemd-logind[1470]: Removed session 65. Mar 20 21:20:04.732595 containerd[1485]: time="2025-03-20T21:20:04.732548767Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3298a708f055861f6f20ce36a7ee251ccfc70f10f7107e13a9514eaec3f5beb1\" id:\"4e43afd06adeb19e4d20719223d0fc13adac59c5ef32f889ed06b08bb03e0733\" pid:6321 exited_at:{seconds:1742505604 nanos:732186718}" Mar 20 21:20:07.586656 systemd[1]: Started sshd@65-10.0.0.54:22-10.0.0.1:33846.service - OpenSSH per-connection server daemon (10.0.0.1:33846). Mar 20 21:20:07.641819 sshd[6334]: Accepted publickey for core from 10.0.0.1 port 33846 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:20:07.642939 sshd-session[6334]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:20:07.646697 systemd-logind[1470]: New session 66 of user core. Mar 20 21:20:07.662340 systemd[1]: Started session-66.scope - Session 66 of User core. Mar 20 21:20:07.777557 sshd[6336]: Connection closed by 10.0.0.1 port 33846 Mar 20 21:20:07.777875 sshd-session[6334]: pam_unix(sshd:session): session closed for user core Mar 20 21:20:07.781893 systemd[1]: sshd@65-10.0.0.54:22-10.0.0.1:33846.service: Deactivated successfully. Mar 20 21:20:07.783595 systemd[1]: session-66.scope: Deactivated successfully. Mar 20 21:20:07.784186 systemd-logind[1470]: Session 66 logged out. Waiting for processes to exit. Mar 20 21:20:07.785010 systemd-logind[1470]: Removed session 66. Mar 20 21:20:11.166725 containerd[1485]: time="2025-03-20T21:20:11.166626921Z" level=warning msg="container event discarded" container=df3e91c579a1da06eee105f037b38b0efeb906f97b925fd79a928b9554540222 type=CONTAINER_CREATED_EVENT Mar 20 21:20:11.166725 containerd[1485]: time="2025-03-20T21:20:11.166692603Z" level=warning msg="container event discarded" container=df3e91c579a1da06eee105f037b38b0efeb906f97b925fd79a928b9554540222 type=CONTAINER_STARTED_EVENT Mar 20 21:20:11.188937 containerd[1485]: time="2025-03-20T21:20:11.188870461Z" level=warning msg="container event discarded" container=1a460a45179149559dee9a177cd0a67ac841188bbfaeae6ff92d5c36ac2b1c34 type=CONTAINER_CREATED_EVENT Mar 20 21:20:11.240165 containerd[1485]: time="2025-03-20T21:20:11.240101091Z" level=warning msg="container event discarded" container=1a460a45179149559dee9a177cd0a67ac841188bbfaeae6ff92d5c36ac2b1c34 type=CONTAINER_STARTED_EVENT Mar 20 21:20:12.780722 containerd[1485]: time="2025-03-20T21:20:12.780632789Z" level=warning msg="container event discarded" container=f453c31569beedde97e494dc086be6eeb7aa67a59f8477aa2d4c59219d2c5e65 type=CONTAINER_CREATED_EVENT Mar 20 21:20:12.784440 containerd[1485]: time="2025-03-20T21:20:12.780927355Z" level=warning msg="container event discarded" container=f453c31569beedde97e494dc086be6eeb7aa67a59f8477aa2d4c59219d2c5e65 type=CONTAINER_STARTED_EVENT Mar 20 21:20:12.788455 systemd[1]: Started sshd@66-10.0.0.54:22-10.0.0.1:34422.service - OpenSSH per-connection server daemon (10.0.0.1:34422). Mar 20 21:20:12.832725 sshd[6352]: Accepted publickey for core from 10.0.0.1 port 34422 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:20:12.834029 sshd-session[6352]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:20:12.838440 systemd-logind[1470]: New session 67 of user core. Mar 20 21:20:12.849356 systemd[1]: Started session-67.scope - Session 67 of User core. Mar 20 21:20:12.969123 sshd[6354]: Connection closed by 10.0.0.1 port 34422 Mar 20 21:20:12.969718 sshd-session[6352]: pam_unix(sshd:session): session closed for user core Mar 20 21:20:12.973554 systemd[1]: sshd@66-10.0.0.54:22-10.0.0.1:34422.service: Deactivated successfully. Mar 20 21:20:12.975661 systemd[1]: session-67.scope: Deactivated successfully. Mar 20 21:20:12.976425 systemd-logind[1470]: Session 67 logged out. Waiting for processes to exit. Mar 20 21:20:12.977305 systemd-logind[1470]: Removed session 67. Mar 20 21:20:15.222775 containerd[1485]: time="2025-03-20T21:20:15.222715119Z" level=warning msg="container event discarded" container=d23590733d5acc7c1adcc7593768f6eecd78985395d64a69d9264e684e357bf1 type=CONTAINER_CREATED_EVENT Mar 20 21:20:15.297095 containerd[1485]: time="2025-03-20T21:20:15.297040154Z" level=warning msg="container event discarded" container=d23590733d5acc7c1adcc7593768f6eecd78985395d64a69d9264e684e357bf1 type=CONTAINER_STARTED_EVENT Mar 20 21:20:17.982952 systemd[1]: Started sshd@67-10.0.0.54:22-10.0.0.1:34436.service - OpenSSH per-connection server daemon (10.0.0.1:34436). Mar 20 21:20:18.033910 sshd[6378]: Accepted publickey for core from 10.0.0.1 port 34436 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:20:18.035273 sshd-session[6378]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:20:18.039260 systemd-logind[1470]: New session 68 of user core. Mar 20 21:20:18.052407 systemd[1]: Started session-68.scope - Session 68 of User core. Mar 20 21:20:18.163384 sshd[6380]: Connection closed by 10.0.0.1 port 34436 Mar 20 21:20:18.163755 sshd-session[6378]: pam_unix(sshd:session): session closed for user core Mar 20 21:20:18.167310 systemd[1]: sshd@67-10.0.0.54:22-10.0.0.1:34436.service: Deactivated successfully. Mar 20 21:20:18.169164 systemd[1]: session-68.scope: Deactivated successfully. Mar 20 21:20:18.169866 systemd-logind[1470]: Session 68 logged out. Waiting for processes to exit. Mar 20 21:20:18.170986 systemd-logind[1470]: Removed session 68. Mar 20 21:20:20.034987 containerd[1485]: time="2025-03-20T21:20:20.034898888Z" level=warning msg="container event discarded" container=6689fc16b6f9db21fc1f1cf9e5fe8313930d7154ca624118ca770e06c974ad09 type=CONTAINER_CREATED_EVENT Mar 20 21:20:20.034987 containerd[1485]: time="2025-03-20T21:20:20.034959849Z" level=warning msg="container event discarded" container=6689fc16b6f9db21fc1f1cf9e5fe8313930d7154ca624118ca770e06c974ad09 type=CONTAINER_STARTED_EVENT Mar 20 21:20:20.170896 containerd[1485]: time="2025-03-20T21:20:20.170823486Z" level=warning msg="container event discarded" container=54ed274cf11ce43c3701d2bf81a57f8bf3ce02515cfbaa6de5aff9e86c50d2e6 type=CONTAINER_CREATED_EVENT Mar 20 21:20:20.170896 containerd[1485]: time="2025-03-20T21:20:20.170869607Z" level=warning msg="container event discarded" container=54ed274cf11ce43c3701d2bf81a57f8bf3ce02515cfbaa6de5aff9e86c50d2e6 type=CONTAINER_STARTED_EVENT Mar 20 21:20:21.705691 containerd[1485]: time="2025-03-20T21:20:21.705632248Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c14e33d2f5b60700e68dcfd113314dd88a94ba5b5df327532409a6162dc44321\" id:\"440b9b5e7bd11dc44354e6a3b2dd757cf02eaf4dcbd0eac10bbd7b1a283aaca9\" pid:6405 exited_at:{seconds:1742505621 nanos:705379802}" Mar 20 21:20:21.939314 containerd[1485]: time="2025-03-20T21:20:21.939214559Z" level=warning msg="container event discarded" container=cbc449a4607b3cecaadbd9ddb27b00d8ebc037a515d0affabb0b4b21de249ecd type=CONTAINER_CREATED_EVENT Mar 20 21:20:22.001800 containerd[1485]: time="2025-03-20T21:20:22.001657253Z" level=warning msg="container event discarded" container=cbc449a4607b3cecaadbd9ddb27b00d8ebc037a515d0affabb0b4b21de249ecd type=CONTAINER_STARTED_EVENT Mar 20 21:20:23.178897 systemd[1]: Started sshd@68-10.0.0.54:22-10.0.0.1:55008.service - OpenSSH per-connection server daemon (10.0.0.1:55008). Mar 20 21:20:23.238180 sshd[6421]: Accepted publickey for core from 10.0.0.1 port 55008 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:20:23.239518 sshd-session[6421]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:20:23.243729 systemd-logind[1470]: New session 69 of user core. Mar 20 21:20:23.257394 systemd[1]: Started session-69.scope - Session 69 of User core. Mar 20 21:20:23.309654 containerd[1485]: time="2025-03-20T21:20:23.309593164Z" level=warning msg="container event discarded" container=a433c0863a65b06e47e0353cf2a1ab0b9f9a9e43e65e1ae62213764aa79b5564 type=CONTAINER_CREATED_EVENT Mar 20 21:20:23.363284 containerd[1485]: time="2025-03-20T21:20:23.363181778Z" level=warning msg="container event discarded" container=a433c0863a65b06e47e0353cf2a1ab0b9f9a9e43e65e1ae62213764aa79b5564 type=CONTAINER_STARTED_EVENT Mar 20 21:20:23.373471 sshd[6423]: Connection closed by 10.0.0.1 port 55008 Mar 20 21:20:23.374073 sshd-session[6421]: pam_unix(sshd:session): session closed for user core Mar 20 21:20:23.377418 systemd[1]: sshd@68-10.0.0.54:22-10.0.0.1:55008.service: Deactivated successfully. Mar 20 21:20:23.379298 systemd[1]: session-69.scope: Deactivated successfully. Mar 20 21:20:23.380017 systemd-logind[1470]: Session 69 logged out. Waiting for processes to exit. Mar 20 21:20:23.380861 systemd-logind[1470]: Removed session 69. Mar 20 21:20:23.545665 containerd[1485]: time="2025-03-20T21:20:23.545519919Z" level=warning msg="container event discarded" container=a433c0863a65b06e47e0353cf2a1ab0b9f9a9e43e65e1ae62213764aa79b5564 type=CONTAINER_STOPPED_EVENT Mar 20 21:20:27.321174 containerd[1485]: time="2025-03-20T21:20:27.321082687Z" level=warning msg="container event discarded" container=7182e74c7e9d381b5702394a0ea0272b5db6b919096a92491e2e24db6654a75a type=CONTAINER_CREATED_EVENT Mar 20 21:20:27.385805 containerd[1485]: time="2025-03-20T21:20:27.385719311Z" level=warning msg="container event discarded" container=7182e74c7e9d381b5702394a0ea0272b5db6b919096a92491e2e24db6654a75a type=CONTAINER_STARTED_EVENT Mar 20 21:20:28.058424 containerd[1485]: time="2025-03-20T21:20:28.058346614Z" level=warning msg="container event discarded" container=7182e74c7e9d381b5702394a0ea0272b5db6b919096a92491e2e24db6654a75a type=CONTAINER_STOPPED_EVENT Mar 20 21:20:28.388405 systemd[1]: Started sshd@69-10.0.0.54:22-10.0.0.1:55022.service - OpenSSH per-connection server daemon (10.0.0.1:55022). Mar 20 21:20:28.445495 sshd[6437]: Accepted publickey for core from 10.0.0.1 port 55022 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:20:28.446699 sshd-session[6437]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:20:28.450911 systemd-logind[1470]: New session 70 of user core. Mar 20 21:20:28.461358 systemd[1]: Started session-70.scope - Session 70 of User core. Mar 20 21:20:28.572345 sshd[6439]: Connection closed by 10.0.0.1 port 55022 Mar 20 21:20:28.572694 sshd-session[6437]: pam_unix(sshd:session): session closed for user core Mar 20 21:20:28.575997 systemd[1]: sshd@69-10.0.0.54:22-10.0.0.1:55022.service: Deactivated successfully. Mar 20 21:20:28.577909 systemd[1]: session-70.scope: Deactivated successfully. Mar 20 21:20:28.579749 systemd-logind[1470]: Session 70 logged out. Waiting for processes to exit. Mar 20 21:20:28.580625 systemd-logind[1470]: Removed session 70. Mar 20 21:20:32.513674 containerd[1485]: time="2025-03-20T21:20:32.513536882Z" level=warning msg="container event discarded" container=3298a708f055861f6f20ce36a7ee251ccfc70f10f7107e13a9514eaec3f5beb1 type=CONTAINER_CREATED_EVENT Mar 20 21:20:32.608882 containerd[1485]: time="2025-03-20T21:20:32.608815141Z" level=warning msg="container event discarded" container=3298a708f055861f6f20ce36a7ee251ccfc70f10f7107e13a9514eaec3f5beb1 type=CONTAINER_STARTED_EVENT Mar 20 21:20:33.584761 systemd[1]: Started sshd@70-10.0.0.54:22-10.0.0.1:38378.service - OpenSSH per-connection server daemon (10.0.0.1:38378). Mar 20 21:20:33.632117 sshd[6453]: Accepted publickey for core from 10.0.0.1 port 38378 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:20:33.633275 sshd-session[6453]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:20:33.637681 systemd-logind[1470]: New session 71 of user core. Mar 20 21:20:33.646375 systemd[1]: Started session-71.scope - Session 71 of User core. Mar 20 21:20:33.759237 sshd[6455]: Connection closed by 10.0.0.1 port 38378 Mar 20 21:20:33.759586 sshd-session[6453]: pam_unix(sshd:session): session closed for user core Mar 20 21:20:33.763181 systemd[1]: sshd@70-10.0.0.54:22-10.0.0.1:38378.service: Deactivated successfully. Mar 20 21:20:33.764970 systemd[1]: session-71.scope: Deactivated successfully. Mar 20 21:20:33.769524 systemd-logind[1470]: Session 71 logged out. Waiting for processes to exit. Mar 20 21:20:33.771433 systemd-logind[1470]: Removed session 71. Mar 20 21:20:34.735919 containerd[1485]: time="2025-03-20T21:20:34.735869522Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3298a708f055861f6f20ce36a7ee251ccfc70f10f7107e13a9514eaec3f5beb1\" id:\"42b9fd25928314dd2a60a7fae75d58d3f7279e02d8f93180f3451b3f82f1cd3e\" pid:6479 exited_at:{seconds:1742505634 nanos:735403553}" Mar 20 21:20:38.771622 systemd[1]: Started sshd@71-10.0.0.54:22-10.0.0.1:38390.service - OpenSSH per-connection server daemon (10.0.0.1:38390). Mar 20 21:20:38.828572 sshd[6493]: Accepted publickey for core from 10.0.0.1 port 38390 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:20:38.830204 sshd-session[6493]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:20:38.834278 systemd-logind[1470]: New session 72 of user core. Mar 20 21:20:38.844361 systemd[1]: Started session-72.scope - Session 72 of User core. Mar 20 21:20:38.982263 sshd[6495]: Connection closed by 10.0.0.1 port 38390 Mar 20 21:20:38.982894 sshd-session[6493]: pam_unix(sshd:session): session closed for user core Mar 20 21:20:38.987028 systemd[1]: sshd@71-10.0.0.54:22-10.0.0.1:38390.service: Deactivated successfully. Mar 20 21:20:38.988865 systemd[1]: session-72.scope: Deactivated successfully. Mar 20 21:20:38.989585 systemd-logind[1470]: Session 72 logged out. Waiting for processes to exit. Mar 20 21:20:38.990474 systemd-logind[1470]: Removed session 72. Mar 20 21:20:39.727811 containerd[1485]: time="2025-03-20T21:20:39.727729328Z" level=warning msg="container event discarded" container=1bd1d7dec319bbb163863bfbdf9bbafd3719c64a370206bd48fddfc8374df856 type=CONTAINER_CREATED_EVENT Mar 20 21:20:39.727811 containerd[1485]: time="2025-03-20T21:20:39.727777769Z" level=warning msg="container event discarded" container=1bd1d7dec319bbb163863bfbdf9bbafd3719c64a370206bd48fddfc8374df856 type=CONTAINER_STARTED_EVENT Mar 20 21:20:39.781041 containerd[1485]: time="2025-03-20T21:20:39.780979463Z" level=warning msg="container event discarded" container=ccae1f9e695f85944ad81c141aca5b1552d328ef319db34f02107209796ff9ae type=CONTAINER_CREATED_EVENT Mar 20 21:20:39.852400 containerd[1485]: time="2025-03-20T21:20:39.852342476Z" level=warning msg="container event discarded" container=ccae1f9e695f85944ad81c141aca5b1552d328ef319db34f02107209796ff9ae type=CONTAINER_STARTED_EVENT Mar 20 21:20:40.596258 containerd[1485]: time="2025-03-20T21:20:40.595992517Z" level=warning msg="container event discarded" container=7ec20bb10b38c16b2e9a5db7fba57e5985e572fb6c605731af4bee45d12ddcd6 type=CONTAINER_CREATED_EVENT Mar 20 21:20:40.596258 containerd[1485]: time="2025-03-20T21:20:40.596039118Z" level=warning msg="container event discarded" container=7ec20bb10b38c16b2e9a5db7fba57e5985e572fb6c605731af4bee45d12ddcd6 type=CONTAINER_STARTED_EVENT Mar 20 21:20:41.634023 containerd[1485]: time="2025-03-20T21:20:41.633959028Z" level=warning msg="container event discarded" container=bb60bacaad0d905abce4377ae3f4cdab4d18db3d6b4c674d817fdcffeb4d4b17 type=CONTAINER_CREATED_EVENT Mar 20 21:20:41.634023 containerd[1485]: time="2025-03-20T21:20:41.634009509Z" level=warning msg="container event discarded" container=bb60bacaad0d905abce4377ae3f4cdab4d18db3d6b4c674d817fdcffeb4d4b17 type=CONTAINER_STARTED_EVENT Mar 20 21:20:42.005645 containerd[1485]: time="2025-03-20T21:20:42.005572011Z" level=warning msg="container event discarded" container=c14e33d2f5b60700e68dcfd113314dd88a94ba5b5df327532409a6162dc44321 type=CONTAINER_CREATED_EVENT Mar 20 21:20:42.096946 containerd[1485]: time="2025-03-20T21:20:42.096887319Z" level=warning msg="container event discarded" container=c14e33d2f5b60700e68dcfd113314dd88a94ba5b5df327532409a6162dc44321 type=CONTAINER_STARTED_EVENT Mar 20 21:20:43.785982 containerd[1485]: time="2025-03-20T21:20:43.785887177Z" level=warning msg="container event discarded" container=fdf3da8083f926e45241c057e9ea9678fb84411bc46b23959743ff765df8f1a5 type=CONTAINER_CREATED_EVENT Mar 20 21:20:43.785982 containerd[1485]: time="2025-03-20T21:20:43.785933418Z" level=warning msg="container event discarded" container=fdf3da8083f926e45241c057e9ea9678fb84411bc46b23959743ff765df8f1a5 type=CONTAINER_STARTED_EVENT Mar 20 21:20:43.785982 containerd[1485]: time="2025-03-20T21:20:43.785945138Z" level=warning msg="container event discarded" container=42bb7f7cb8701cdfea0cb1630e5d4e3ee1e28fa6aca4dab6ff1fc6e9e3affaef type=CONTAINER_CREATED_EVENT Mar 20 21:20:43.785982 containerd[1485]: time="2025-03-20T21:20:43.785952698Z" level=warning msg="container event discarded" container=42bb7f7cb8701cdfea0cb1630e5d4e3ee1e28fa6aca4dab6ff1fc6e9e3affaef type=CONTAINER_STARTED_EVENT Mar 20 21:20:43.828048 containerd[1485]: time="2025-03-20T21:20:43.820151966Z" level=warning msg="container event discarded" container=a1989d9231126907f6d0d25a2e6a98668241bee10dd196648d27bc57a2dcde5b type=CONTAINER_CREATED_EVENT Mar 20 21:20:43.877271 containerd[1485]: time="2025-03-20T21:20:43.877233799Z" level=warning msg="container event discarded" container=a1989d9231126907f6d0d25a2e6a98668241bee10dd196648d27bc57a2dcde5b type=CONTAINER_STARTED_EVENT Mar 20 21:20:43.995910 systemd[1]: Started sshd@72-10.0.0.54:22-10.0.0.1:41894.service - OpenSSH per-connection server daemon (10.0.0.1:41894). Mar 20 21:20:44.051113 sshd[6511]: Accepted publickey for core from 10.0.0.1 port 41894 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:20:44.052304 sshd-session[6511]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:20:44.058494 systemd-logind[1470]: New session 73 of user core. Mar 20 21:20:44.068353 systemd[1]: Started session-73.scope - Session 73 of User core. Mar 20 21:20:44.080676 containerd[1485]: time="2025-03-20T21:20:44.080627722Z" level=warning msg="container event discarded" container=ad4699d874faf1ba9a0794922797f614c74e415fff8508d0b1f006e22f7e35a0 type=CONTAINER_CREATED_EVENT Mar 20 21:20:44.150252 containerd[1485]: time="2025-03-20T21:20:44.150194954Z" level=warning msg="container event discarded" container=ad4699d874faf1ba9a0794922797f614c74e415fff8508d0b1f006e22f7e35a0 type=CONTAINER_STARTED_EVENT Mar 20 21:20:44.202002 sshd[6513]: Connection closed by 10.0.0.1 port 41894 Mar 20 21:20:44.202458 sshd-session[6511]: pam_unix(sshd:session): session closed for user core Mar 20 21:20:44.205170 systemd[1]: session-73.scope: Deactivated successfully. Mar 20 21:20:44.206690 systemd-logind[1470]: Session 73 logged out. Waiting for processes to exit. Mar 20 21:20:44.206849 systemd[1]: sshd@72-10.0.0.54:22-10.0.0.1:41894.service: Deactivated successfully. Mar 20 21:20:44.210416 systemd-logind[1470]: Removed session 73. Mar 20 21:20:44.340947 containerd[1485]: time="2025-03-20T21:20:44.340795220Z" level=warning msg="container event discarded" container=d6f8fec12a92c681515876f0f48394e95454d678fa466ccd8cb28302d9e49834 type=CONTAINER_CREATED_EVENT Mar 20 21:20:44.518508 containerd[1485]: time="2025-03-20T21:20:44.518448793Z" level=warning msg="container event discarded" container=d6f8fec12a92c681515876f0f48394e95454d678fa466ccd8cb28302d9e49834 type=CONTAINER_STARTED_EVENT Mar 20 21:20:45.593352 containerd[1485]: time="2025-03-20T21:20:45.593264807Z" level=warning msg="container event discarded" container=56354cbb177d440d3d5b9900021ffbe9a0248a43096667251a459d7483cb59a9 type=CONTAINER_CREATED_EVENT Mar 20 21:20:45.593352 containerd[1485]: time="2025-03-20T21:20:45.593325248Z" level=warning msg="container event discarded" container=56354cbb177d440d3d5b9900021ffbe9a0248a43096667251a459d7483cb59a9 type=CONTAINER_STARTED_EVENT Mar 20 21:20:46.470080 containerd[1485]: time="2025-03-20T21:20:46.470017199Z" level=warning msg="container event discarded" container=e0bc6deef9ced06603e2f6be06d9d8c24696117fa84278eff0ade88ffa7b8cdc type=CONTAINER_CREATED_EVENT Mar 20 21:20:46.526318 containerd[1485]: time="2025-03-20T21:20:46.526262045Z" level=warning msg="container event discarded" container=e0bc6deef9ced06603e2f6be06d9d8c24696117fa84278eff0ade88ffa7b8cdc type=CONTAINER_STARTED_EVENT Mar 20 21:20:47.449294 containerd[1485]: time="2025-03-20T21:20:47.449206591Z" level=warning msg="container event discarded" container=3cb9010637ec40861c9b311402478726839133a624aeb14954fcd935061b3e9a type=CONTAINER_CREATED_EVENT Mar 20 21:20:47.511585 containerd[1485]: time="2025-03-20T21:20:47.511505909Z" level=warning msg="container event discarded" container=3cb9010637ec40861c9b311402478726839133a624aeb14954fcd935061b3e9a type=CONTAINER_STARTED_EVENT Mar 20 21:20:47.827569 containerd[1485]: time="2025-03-20T21:20:47.827389625Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c14e33d2f5b60700e68dcfd113314dd88a94ba5b5df327532409a6162dc44321\" id:\"300429a71591a7c3cec1d36e7ff14b9c04b2dcbf5aace1f03455b3c1e39982fa\" pid:6538 exited_at:{seconds:1742505647 nanos:827115500}" Mar 20 21:20:49.214664 systemd[1]: Started sshd@73-10.0.0.54:22-10.0.0.1:41902.service - OpenSSH per-connection server daemon (10.0.0.1:41902). Mar 20 21:20:49.277049 sshd[6549]: Accepted publickey for core from 10.0.0.1 port 41902 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:20:49.278209 sshd-session[6549]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:20:49.282660 systemd-logind[1470]: New session 74 of user core. Mar 20 21:20:49.295364 systemd[1]: Started session-74.scope - Session 74 of User core. Mar 20 21:20:49.431247 sshd[6551]: Connection closed by 10.0.0.1 port 41902 Mar 20 21:20:49.431847 sshd-session[6549]: pam_unix(sshd:session): session closed for user core Mar 20 21:20:49.435171 systemd[1]: sshd@73-10.0.0.54:22-10.0.0.1:41902.service: Deactivated successfully. Mar 20 21:20:49.437071 systemd[1]: session-74.scope: Deactivated successfully. Mar 20 21:20:49.437758 systemd-logind[1470]: Session 74 logged out. Waiting for processes to exit. Mar 20 21:20:49.438631 systemd-logind[1470]: Removed session 74. Mar 20 21:20:51.716495 containerd[1485]: time="2025-03-20T21:20:51.716456542Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c14e33d2f5b60700e68dcfd113314dd88a94ba5b5df327532409a6162dc44321\" id:\"0b31ac72f9c69cccd7353704ab06369d956e3650d144d722b077cfa9e13dd72b\" pid:6576 exited_at:{seconds:1742505651 nanos:715074756}" Mar 20 21:20:54.447483 systemd[1]: Started sshd@74-10.0.0.54:22-10.0.0.1:33360.service - OpenSSH per-connection server daemon (10.0.0.1:33360). Mar 20 21:20:54.505880 sshd[6587]: Accepted publickey for core from 10.0.0.1 port 33360 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:20:54.507137 sshd-session[6587]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:20:54.511568 systemd-logind[1470]: New session 75 of user core. Mar 20 21:20:54.520404 systemd[1]: Started session-75.scope - Session 75 of User core. Mar 20 21:20:54.662925 sshd[6589]: Connection closed by 10.0.0.1 port 33360 Mar 20 21:20:54.663254 sshd-session[6587]: pam_unix(sshd:session): session closed for user core Mar 20 21:20:54.668689 systemd[1]: sshd@74-10.0.0.54:22-10.0.0.1:33360.service: Deactivated successfully. Mar 20 21:20:54.671004 systemd[1]: session-75.scope: Deactivated successfully. Mar 20 21:20:54.672454 systemd-logind[1470]: Session 75 logged out. Waiting for processes to exit. Mar 20 21:20:54.674541 systemd-logind[1470]: Removed session 75. Mar 20 21:20:59.674657 systemd[1]: Started sshd@75-10.0.0.54:22-10.0.0.1:33376.service - OpenSSH per-connection server daemon (10.0.0.1:33376). Mar 20 21:20:59.727784 sshd[6608]: Accepted publickey for core from 10.0.0.1 port 33376 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:20:59.728980 sshd-session[6608]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:20:59.733333 systemd-logind[1470]: New session 76 of user core. Mar 20 21:20:59.739384 systemd[1]: Started session-76.scope - Session 76 of User core. Mar 20 21:20:59.890667 sshd[6610]: Connection closed by 10.0.0.1 port 33376 Mar 20 21:20:59.890995 sshd-session[6608]: pam_unix(sshd:session): session closed for user core Mar 20 21:20:59.894197 systemd[1]: sshd@75-10.0.0.54:22-10.0.0.1:33376.service: Deactivated successfully. Mar 20 21:20:59.896002 systemd[1]: session-76.scope: Deactivated successfully. Mar 20 21:20:59.896974 systemd-logind[1470]: Session 76 logged out. Waiting for processes to exit. Mar 20 21:20:59.897880 systemd-logind[1470]: Removed session 76. Mar 20 21:21:04.732646 containerd[1485]: time="2025-03-20T21:21:04.732607762Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3298a708f055861f6f20ce36a7ee251ccfc70f10f7107e13a9514eaec3f5beb1\" id:\"2a166b4653b713c2eeaf426b07d561f49a71e3d75110edca9eb0e3f949c156e5\" pid:6636 exited_at:{seconds:1742505664 nanos:732340717}" Mar 20 21:21:04.905705 systemd[1]: Started sshd@76-10.0.0.54:22-10.0.0.1:55314.service - OpenSSH per-connection server daemon (10.0.0.1:55314). Mar 20 21:21:04.965129 sshd[6650]: Accepted publickey for core from 10.0.0.1 port 55314 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:21:04.966338 sshd-session[6650]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:21:04.970089 systemd-logind[1470]: New session 77 of user core. Mar 20 21:21:04.976346 systemd[1]: Started session-77.scope - Session 77 of User core. Mar 20 21:21:05.121310 sshd[6652]: Connection closed by 10.0.0.1 port 55314 Mar 20 21:21:05.121389 sshd-session[6650]: pam_unix(sshd:session): session closed for user core Mar 20 21:21:05.124990 systemd[1]: sshd@76-10.0.0.54:22-10.0.0.1:55314.service: Deactivated successfully. Mar 20 21:21:05.126678 systemd[1]: session-77.scope: Deactivated successfully. Mar 20 21:21:05.128045 systemd-logind[1470]: Session 77 logged out. Waiting for processes to exit. Mar 20 21:21:05.129008 systemd-logind[1470]: Removed session 77. Mar 20 21:21:10.134642 systemd[1]: Started sshd@77-10.0.0.54:22-10.0.0.1:55322.service - OpenSSH per-connection server daemon (10.0.0.1:55322). Mar 20 21:21:10.194837 sshd[6669]: Accepted publickey for core from 10.0.0.1 port 55322 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:21:10.196101 sshd-session[6669]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:21:10.201407 systemd-logind[1470]: New session 78 of user core. Mar 20 21:21:10.210403 systemd[1]: Started session-78.scope - Session 78 of User core. Mar 20 21:21:10.374806 sshd[6671]: Connection closed by 10.0.0.1 port 55322 Mar 20 21:21:10.375146 sshd-session[6669]: pam_unix(sshd:session): session closed for user core Mar 20 21:21:10.379789 systemd[1]: sshd@77-10.0.0.54:22-10.0.0.1:55322.service: Deactivated successfully. Mar 20 21:21:10.381727 systemd[1]: session-78.scope: Deactivated successfully. Mar 20 21:21:10.383378 systemd-logind[1470]: Session 78 logged out. Waiting for processes to exit. Mar 20 21:21:10.384839 systemd-logind[1470]: Removed session 78. Mar 20 21:21:15.387285 systemd[1]: Started sshd@78-10.0.0.54:22-10.0.0.1:58686.service - OpenSSH per-connection server daemon (10.0.0.1:58686). Mar 20 21:21:15.439120 sshd[6687]: Accepted publickey for core from 10.0.0.1 port 58686 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:21:15.440383 sshd-session[6687]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:21:15.444863 systemd-logind[1470]: New session 79 of user core. Mar 20 21:21:15.456453 systemd[1]: Started session-79.scope - Session 79 of User core. Mar 20 21:21:15.587765 sshd[6689]: Connection closed by 10.0.0.1 port 58686 Mar 20 21:21:15.588472 sshd-session[6687]: pam_unix(sshd:session): session closed for user core Mar 20 21:21:15.591577 systemd[1]: sshd@78-10.0.0.54:22-10.0.0.1:58686.service: Deactivated successfully. Mar 20 21:21:15.593458 systemd[1]: session-79.scope: Deactivated successfully. Mar 20 21:21:15.596534 systemd-logind[1470]: Session 79 logged out. Waiting for processes to exit. Mar 20 21:21:15.597508 systemd-logind[1470]: Removed session 79. Mar 20 21:21:20.601202 systemd[1]: Started sshd@79-10.0.0.54:22-10.0.0.1:58694.service - OpenSSH per-connection server daemon (10.0.0.1:58694). Mar 20 21:21:20.661210 sshd[6702]: Accepted publickey for core from 10.0.0.1 port 58694 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:21:20.661712 sshd-session[6702]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:21:20.665946 systemd-logind[1470]: New session 80 of user core. Mar 20 21:21:20.674368 systemd[1]: Started session-80.scope - Session 80 of User core. Mar 20 21:21:20.797491 sshd[6704]: Connection closed by 10.0.0.1 port 58694 Mar 20 21:21:20.797081 sshd-session[6702]: pam_unix(sshd:session): session closed for user core Mar 20 21:21:20.799964 systemd[1]: sshd@79-10.0.0.54:22-10.0.0.1:58694.service: Deactivated successfully. Mar 20 21:21:20.801845 systemd[1]: session-80.scope: Deactivated successfully. Mar 20 21:21:20.803170 systemd-logind[1470]: Session 80 logged out. Waiting for processes to exit. Mar 20 21:21:20.804164 systemd-logind[1470]: Removed session 80. Mar 20 21:21:21.720040 containerd[1485]: time="2025-03-20T21:21:21.719988548Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c14e33d2f5b60700e68dcfd113314dd88a94ba5b5df327532409a6162dc44321\" id:\"d9db77fa649121eece69006b30e74828293a96fae4f588501672ffa31b6a1239\" pid:6729 exited_at:{seconds:1742505681 nanos:719650502}" Mar 20 21:21:25.808684 systemd[1]: Started sshd@80-10.0.0.54:22-10.0.0.1:41408.service - OpenSSH per-connection server daemon (10.0.0.1:41408). Mar 20 21:21:25.864210 sshd[6739]: Accepted publickey for core from 10.0.0.1 port 41408 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:21:25.865435 sshd-session[6739]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:21:25.869284 systemd-logind[1470]: New session 81 of user core. Mar 20 21:21:25.877413 systemd[1]: Started session-81.scope - Session 81 of User core. Mar 20 21:21:25.996599 sshd[6741]: Connection closed by 10.0.0.1 port 41408 Mar 20 21:21:25.997062 sshd-session[6739]: pam_unix(sshd:session): session closed for user core Mar 20 21:21:26.000196 systemd[1]: sshd@80-10.0.0.54:22-10.0.0.1:41408.service: Deactivated successfully. Mar 20 21:21:26.002865 systemd[1]: session-81.scope: Deactivated successfully. Mar 20 21:21:26.003693 systemd-logind[1470]: Session 81 logged out. Waiting for processes to exit. Mar 20 21:21:26.004618 systemd-logind[1470]: Removed session 81. Mar 20 21:21:31.006799 systemd[1]: Started sshd@81-10.0.0.54:22-10.0.0.1:41414.service - OpenSSH per-connection server daemon (10.0.0.1:41414). Mar 20 21:21:31.056988 sshd[6754]: Accepted publickey for core from 10.0.0.1 port 41414 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:21:31.058089 sshd-session[6754]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:21:31.062088 systemd-logind[1470]: New session 82 of user core. Mar 20 21:21:31.079424 systemd[1]: Started session-82.scope - Session 82 of User core. Mar 20 21:21:31.200252 sshd[6756]: Connection closed by 10.0.0.1 port 41414 Mar 20 21:21:31.200930 sshd-session[6754]: pam_unix(sshd:session): session closed for user core Mar 20 21:21:31.203681 systemd[1]: sshd@81-10.0.0.54:22-10.0.0.1:41414.service: Deactivated successfully. Mar 20 21:21:31.205415 systemd[1]: session-82.scope: Deactivated successfully. Mar 20 21:21:31.206668 systemd-logind[1470]: Session 82 logged out. Waiting for processes to exit. Mar 20 21:21:31.207546 systemd-logind[1470]: Removed session 82. Mar 20 21:21:32.344244 kubelet[2596]: E0320 21:21:32.344156 2596 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 20 21:21:34.725896 containerd[1485]: time="2025-03-20T21:21:34.725848632Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3298a708f055861f6f20ce36a7ee251ccfc70f10f7107e13a9514eaec3f5beb1\" id:\"a2f2301f615f090ed7fb958f139eee61a69bc7e965bf49a22b3a8bf1db9df5a4\" pid:6783 exited_at:{seconds:1742505694 nanos:725478026}" Mar 20 21:21:36.212728 systemd[1]: Started sshd@82-10.0.0.54:22-10.0.0.1:42618.service - OpenSSH per-connection server daemon (10.0.0.1:42618). Mar 20 21:21:36.271275 sshd[6797]: Accepted publickey for core from 10.0.0.1 port 42618 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:21:36.271922 sshd-session[6797]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:21:36.276272 systemd-logind[1470]: New session 83 of user core. Mar 20 21:21:36.282355 systemd[1]: Started session-83.scope - Session 83 of User core. Mar 20 21:21:36.392137 sshd[6799]: Connection closed by 10.0.0.1 port 42618 Mar 20 21:21:36.392473 sshd-session[6797]: pam_unix(sshd:session): session closed for user core Mar 20 21:21:36.396344 systemd[1]: sshd@82-10.0.0.54:22-10.0.0.1:42618.service: Deactivated successfully. Mar 20 21:21:36.398675 systemd[1]: session-83.scope: Deactivated successfully. Mar 20 21:21:36.399890 systemd-logind[1470]: Session 83 logged out. Waiting for processes to exit. Mar 20 21:21:36.401208 systemd-logind[1470]: Removed session 83. Mar 20 21:21:41.344715 kubelet[2596]: E0320 21:21:41.344672 2596 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 20 21:21:41.407624 systemd[1]: Started sshd@83-10.0.0.54:22-10.0.0.1:42624.service - OpenSSH per-connection server daemon (10.0.0.1:42624). Mar 20 21:21:41.463851 sshd[6813]: Accepted publickey for core from 10.0.0.1 port 42624 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:21:41.465098 sshd-session[6813]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:21:41.469076 systemd-logind[1470]: New session 84 of user core. Mar 20 21:21:41.480413 systemd[1]: Started session-84.scope - Session 84 of User core. Mar 20 21:21:41.594263 sshd[6817]: Connection closed by 10.0.0.1 port 42624 Mar 20 21:21:41.594595 sshd-session[6813]: pam_unix(sshd:session): session closed for user core Mar 20 21:21:41.598072 systemd[1]: sshd@83-10.0.0.54:22-10.0.0.1:42624.service: Deactivated successfully. Mar 20 21:21:41.600339 systemd[1]: session-84.scope: Deactivated successfully. Mar 20 21:21:41.601590 systemd-logind[1470]: Session 84 logged out. Waiting for processes to exit. Mar 20 21:21:41.602661 systemd-logind[1470]: Removed session 84. Mar 20 21:21:46.606586 systemd[1]: Started sshd@84-10.0.0.54:22-10.0.0.1:36454.service - OpenSSH per-connection server daemon (10.0.0.1:36454). Mar 20 21:21:46.663643 sshd[6841]: Accepted publickey for core from 10.0.0.1 port 36454 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:21:46.664117 sshd-session[6841]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:21:46.667874 systemd-logind[1470]: New session 85 of user core. Mar 20 21:21:46.678418 systemd[1]: Started session-85.scope - Session 85 of User core. Mar 20 21:21:46.785994 sshd[6843]: Connection closed by 10.0.0.1 port 36454 Mar 20 21:21:46.785598 sshd-session[6841]: pam_unix(sshd:session): session closed for user core Mar 20 21:21:46.788736 systemd[1]: sshd@84-10.0.0.54:22-10.0.0.1:36454.service: Deactivated successfully. Mar 20 21:21:46.790502 systemd[1]: session-85.scope: Deactivated successfully. Mar 20 21:21:46.792688 systemd-logind[1470]: Session 85 logged out. Waiting for processes to exit. Mar 20 21:21:46.793651 systemd-logind[1470]: Removed session 85. Mar 20 21:21:47.812920 containerd[1485]: time="2025-03-20T21:21:47.812876632Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c14e33d2f5b60700e68dcfd113314dd88a94ba5b5df327532409a6162dc44321\" id:\"896e1ac11059b6ad32b1e6ad02030251e517a273b88ff6ad7047ec89c0088005\" pid:6867 exited_at:{seconds:1742505707 nanos:812452785}" Mar 20 21:21:51.720511 containerd[1485]: time="2025-03-20T21:21:51.720454067Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c14e33d2f5b60700e68dcfd113314dd88a94ba5b5df327532409a6162dc44321\" id:\"d7df8651ebabc438930600b03ca086ee45d2f9cfc20500af972c2cd5c7093df7\" pid:6889 exited_at:{seconds:1742505711 nanos:720140342}" Mar 20 21:21:51.796519 systemd[1]: Started sshd@85-10.0.0.54:22-10.0.0.1:36460.service - OpenSSH per-connection server daemon (10.0.0.1:36460). Mar 20 21:21:51.840969 sshd[6899]: Accepted publickey for core from 10.0.0.1 port 36460 ssh2: RSA SHA256:X6VVi2zGwQT4vFw/VBKa9j3CAPR/1+qaKaiwBaTCF1Y Mar 20 21:21:51.842113 sshd-session[6899]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 20 21:21:51.848117 systemd-logind[1470]: New session 86 of user core. Mar 20 21:21:51.857403 systemd[1]: Started session-86.scope - Session 86 of User core. Mar 20 21:21:51.973416 sshd[6901]: Connection closed by 10.0.0.1 port 36460 Mar 20 21:21:51.973673 sshd-session[6899]: pam_unix(sshd:session): session closed for user core Mar 20 21:21:51.977043 systemd[1]: sshd@85-10.0.0.54:22-10.0.0.1:36460.service: Deactivated successfully. Mar 20 21:21:51.978838 systemd[1]: session-86.scope: Deactivated successfully. Mar 20 21:21:51.979707 systemd-logind[1470]: Session 86 logged out. Waiting for processes to exit. Mar 20 21:21:51.980754 systemd-logind[1470]: Removed session 86.