May 14 23:51:14.938025 kernel: Linux version 6.6.89-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.2.1_p20241221 p7) 14.2.1 20241221, GNU ld (Gentoo 2.44 p1) 2.44.0) #1 SMP PREEMPT_DYNAMIC Wed May 14 22:09:34 -00 2025 May 14 23:51:14.938051 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=e0c956f61127e47bb23a2bdeb0592b0ff91bd857e2344d0bf321acb67c279f1a May 14 23:51:14.938062 kernel: BIOS-provided physical RAM map: May 14 23:51:14.938069 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable May 14 23:51:14.938076 kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000007fffff] usable May 14 23:51:14.938083 kernel: BIOS-e820: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS May 14 23:51:14.938091 kernel: BIOS-e820: [mem 0x0000000000808000-0x000000000080afff] usable May 14 23:51:14.938098 kernel: BIOS-e820: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS May 14 23:51:14.938105 kernel: BIOS-e820: [mem 0x000000000080c000-0x0000000000810fff] usable May 14 23:51:14.938112 kernel: BIOS-e820: [mem 0x0000000000811000-0x00000000008fffff] ACPI NVS May 14 23:51:14.938121 kernel: BIOS-e820: [mem 0x0000000000900000-0x000000009bd3efff] usable May 14 23:51:14.938131 kernel: BIOS-e820: [mem 0x000000009bd3f000-0x000000009bdfffff] reserved May 14 23:51:14.938148 kernel: BIOS-e820: [mem 0x000000009be00000-0x000000009c8ecfff] usable May 14 23:51:14.938164 kernel: BIOS-e820: [mem 0x000000009c8ed000-0x000000009cb6cfff] reserved May 14 23:51:14.938175 kernel: BIOS-e820: [mem 0x000000009cb6d000-0x000000009cb7efff] ACPI data May 14 23:51:14.938182 kernel: BIOS-e820: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS May 14 23:51:14.939341 kernel: BIOS-e820: [mem 0x000000009cbff000-0x000000009ce91fff] usable May 14 23:51:14.939349 kernel: BIOS-e820: [mem 0x000000009ce92000-0x000000009ce95fff] reserved May 14 23:51:14.939357 kernel: BIOS-e820: [mem 0x000000009ce96000-0x000000009ce97fff] ACPI NVS May 14 23:51:14.939372 kernel: BIOS-e820: [mem 0x000000009ce98000-0x000000009cedbfff] usable May 14 23:51:14.939381 kernel: BIOS-e820: [mem 0x000000009cedc000-0x000000009cf5ffff] reserved May 14 23:51:14.939388 kernel: BIOS-e820: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS May 14 23:51:14.939403 kernel: BIOS-e820: [mem 0x00000000e0000000-0x00000000efffffff] reserved May 14 23:51:14.939411 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved May 14 23:51:14.939418 kernel: BIOS-e820: [mem 0x00000000ffc00000-0x00000000ffffffff] reserved May 14 23:51:14.939425 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved May 14 23:51:14.939433 kernel: NX (Execute Disable) protection: active May 14 23:51:14.939444 kernel: APIC: Static calls initialized May 14 23:51:14.939451 kernel: e820: update [mem 0x9b351018-0x9b35ac57] usable ==> usable May 14 23:51:14.939459 kernel: e820: update [mem 0x9b351018-0x9b35ac57] usable ==> usable May 14 23:51:14.939476 kernel: e820: update [mem 0x9b314018-0x9b350e57] usable ==> usable May 14 23:51:14.939487 kernel: e820: update [mem 0x9b314018-0x9b350e57] usable ==> usable May 14 23:51:14.939494 kernel: extended physical RAM map: May 14 23:51:14.939501 kernel: reserve setup_data: [mem 0x0000000000000000-0x000000000009ffff] usable May 14 23:51:14.939509 kernel: reserve setup_data: [mem 0x0000000000100000-0x00000000007fffff] usable May 14 23:51:14.939516 kernel: reserve setup_data: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS May 14 23:51:14.939523 kernel: reserve setup_data: [mem 0x0000000000808000-0x000000000080afff] usable May 14 23:51:14.939530 kernel: reserve setup_data: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS May 14 23:51:14.939537 kernel: reserve setup_data: [mem 0x000000000080c000-0x0000000000810fff] usable May 14 23:51:14.939548 kernel: reserve setup_data: [mem 0x0000000000811000-0x00000000008fffff] ACPI NVS May 14 23:51:14.939559 kernel: reserve setup_data: [mem 0x0000000000900000-0x000000009b314017] usable May 14 23:51:14.939567 kernel: reserve setup_data: [mem 0x000000009b314018-0x000000009b350e57] usable May 14 23:51:14.939582 kernel: reserve setup_data: [mem 0x000000009b350e58-0x000000009b351017] usable May 14 23:51:14.939593 kernel: reserve setup_data: [mem 0x000000009b351018-0x000000009b35ac57] usable May 14 23:51:14.939600 kernel: reserve setup_data: [mem 0x000000009b35ac58-0x000000009bd3efff] usable May 14 23:51:14.939615 kernel: reserve setup_data: [mem 0x000000009bd3f000-0x000000009bdfffff] reserved May 14 23:51:14.939628 kernel: reserve setup_data: [mem 0x000000009be00000-0x000000009c8ecfff] usable May 14 23:51:14.939638 kernel: reserve setup_data: [mem 0x000000009c8ed000-0x000000009cb6cfff] reserved May 14 23:51:14.939651 kernel: reserve setup_data: [mem 0x000000009cb6d000-0x000000009cb7efff] ACPI data May 14 23:51:14.939669 kernel: reserve setup_data: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS May 14 23:51:14.939687 kernel: reserve setup_data: [mem 0x000000009cbff000-0x000000009ce91fff] usable May 14 23:51:14.939703 kernel: reserve setup_data: [mem 0x000000009ce92000-0x000000009ce95fff] reserved May 14 23:51:14.939713 kernel: reserve setup_data: [mem 0x000000009ce96000-0x000000009ce97fff] ACPI NVS May 14 23:51:14.939743 kernel: reserve setup_data: [mem 0x000000009ce98000-0x000000009cedbfff] usable May 14 23:51:14.939751 kernel: reserve setup_data: [mem 0x000000009cedc000-0x000000009cf5ffff] reserved May 14 23:51:14.939763 kernel: reserve setup_data: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS May 14 23:51:14.939770 kernel: reserve setup_data: [mem 0x00000000e0000000-0x00000000efffffff] reserved May 14 23:51:14.939778 kernel: reserve setup_data: [mem 0x00000000feffc000-0x00000000feffffff] reserved May 14 23:51:14.939788 kernel: reserve setup_data: [mem 0x00000000ffc00000-0x00000000ffffffff] reserved May 14 23:51:14.939796 kernel: reserve setup_data: [mem 0x000000fd00000000-0x000000ffffffffff] reserved May 14 23:51:14.939803 kernel: efi: EFI v2.7 by EDK II May 14 23:51:14.939811 kernel: efi: SMBIOS=0x9c988000 ACPI=0x9cb7e000 ACPI 2.0=0x9cb7e014 MEMATTR=0x9ba0d198 RNG=0x9cb73018 May 14 23:51:14.939819 kernel: random: crng init done May 14 23:51:14.939826 kernel: efi: Remove mem142: MMIO range=[0xffc00000-0xffffffff] (4MB) from e820 map May 14 23:51:14.939834 kernel: e820: remove [mem 0xffc00000-0xffffffff] reserved May 14 23:51:14.939844 kernel: secureboot: Secure boot disabled May 14 23:51:14.939854 kernel: SMBIOS 2.8 present. May 14 23:51:14.939862 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS unknown 02/02/2022 May 14 23:51:14.939869 kernel: Hypervisor detected: KVM May 14 23:51:14.939877 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 May 14 23:51:14.939884 kernel: kvm-clock: using sched offset of 3661408856 cycles May 14 23:51:14.939892 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns May 14 23:51:14.939900 kernel: tsc: Detected 2794.748 MHz processor May 14 23:51:14.939908 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved May 14 23:51:14.939916 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable May 14 23:51:14.939923 kernel: last_pfn = 0x9cedc max_arch_pfn = 0x400000000 May 14 23:51:14.939934 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs May 14 23:51:14.939942 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT May 14 23:51:14.939949 kernel: Using GB pages for direct mapping May 14 23:51:14.939957 kernel: ACPI: Early table checksum verification disabled May 14 23:51:14.939965 kernel: ACPI: RSDP 0x000000009CB7E014 000024 (v02 BOCHS ) May 14 23:51:14.939973 kernel: ACPI: XSDT 0x000000009CB7D0E8 000054 (v01 BOCHS BXPC 00000001 01000013) May 14 23:51:14.939980 kernel: ACPI: FACP 0x000000009CB79000 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) May 14 23:51:14.939988 kernel: ACPI: DSDT 0x000000009CB7A000 0021A8 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 14 23:51:14.939996 kernel: ACPI: FACS 0x000000009CBDD000 000040 May 14 23:51:14.940006 kernel: ACPI: APIC 0x000000009CB78000 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 14 23:51:14.940014 kernel: ACPI: HPET 0x000000009CB77000 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 14 23:51:14.940021 kernel: ACPI: MCFG 0x000000009CB76000 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) May 14 23:51:14.940029 kernel: ACPI: WAET 0x000000009CB75000 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 14 23:51:14.940037 kernel: ACPI: BGRT 0x000000009CB74000 000038 (v01 INTEL EDK2 00000002 01000013) May 14 23:51:14.940045 kernel: ACPI: Reserving FACP table memory at [mem 0x9cb79000-0x9cb790f3] May 14 23:51:14.940052 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cb7a000-0x9cb7c1a7] May 14 23:51:14.940060 kernel: ACPI: Reserving FACS table memory at [mem 0x9cbdd000-0x9cbdd03f] May 14 23:51:14.940067 kernel: ACPI: Reserving APIC table memory at [mem 0x9cb78000-0x9cb7808f] May 14 23:51:14.940078 kernel: ACPI: Reserving HPET table memory at [mem 0x9cb77000-0x9cb77037] May 14 23:51:14.940085 kernel: ACPI: Reserving MCFG table memory at [mem 0x9cb76000-0x9cb7603b] May 14 23:51:14.940093 kernel: ACPI: Reserving WAET table memory at [mem 0x9cb75000-0x9cb75027] May 14 23:51:14.940101 kernel: ACPI: Reserving BGRT table memory at [mem 0x9cb74000-0x9cb74037] May 14 23:51:14.940108 kernel: No NUMA configuration found May 14 23:51:14.940116 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cedbfff] May 14 23:51:14.940123 kernel: NODE_DATA(0) allocated [mem 0x9ce3a000-0x9ce3ffff] May 14 23:51:14.940131 kernel: Zone ranges: May 14 23:51:14.940139 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] May 14 23:51:14.940149 kernel: DMA32 [mem 0x0000000001000000-0x000000009cedbfff] May 14 23:51:14.940156 kernel: Normal empty May 14 23:51:14.940166 kernel: Movable zone start for each node May 14 23:51:14.940174 kernel: Early memory node ranges May 14 23:51:14.940181 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] May 14 23:51:14.940189 kernel: node 0: [mem 0x0000000000100000-0x00000000007fffff] May 14 23:51:14.940197 kernel: node 0: [mem 0x0000000000808000-0x000000000080afff] May 14 23:51:14.940204 kernel: node 0: [mem 0x000000000080c000-0x0000000000810fff] May 14 23:51:14.940212 kernel: node 0: [mem 0x0000000000900000-0x000000009bd3efff] May 14 23:51:14.940219 kernel: node 0: [mem 0x000000009be00000-0x000000009c8ecfff] May 14 23:51:14.940229 kernel: node 0: [mem 0x000000009cbff000-0x000000009ce91fff] May 14 23:51:14.940237 kernel: node 0: [mem 0x000000009ce98000-0x000000009cedbfff] May 14 23:51:14.940245 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cedbfff] May 14 23:51:14.940252 kernel: On node 0, zone DMA: 1 pages in unavailable ranges May 14 23:51:14.940260 kernel: On node 0, zone DMA: 96 pages in unavailable ranges May 14 23:51:14.940276 kernel: On node 0, zone DMA: 8 pages in unavailable ranges May 14 23:51:14.940287 kernel: On node 0, zone DMA: 1 pages in unavailable ranges May 14 23:51:14.940294 kernel: On node 0, zone DMA: 239 pages in unavailable ranges May 14 23:51:14.940302 kernel: On node 0, zone DMA32: 193 pages in unavailable ranges May 14 23:51:14.940310 kernel: On node 0, zone DMA32: 786 pages in unavailable ranges May 14 23:51:14.940320 kernel: On node 0, zone DMA32: 6 pages in unavailable ranges May 14 23:51:14.940328 kernel: On node 0, zone DMA32: 12580 pages in unavailable ranges May 14 23:51:14.940339 kernel: ACPI: PM-Timer IO Port: 0x608 May 14 23:51:14.940347 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) May 14 23:51:14.940355 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 May 14 23:51:14.940363 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) May 14 23:51:14.940373 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) May 14 23:51:14.940381 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) May 14 23:51:14.940389 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) May 14 23:51:14.940397 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) May 14 23:51:14.940405 kernel: ACPI: Using ACPI (MADT) for SMP configuration information May 14 23:51:14.940413 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 May 14 23:51:14.940420 kernel: TSC deadline timer available May 14 23:51:14.940428 kernel: smpboot: Allowing 4 CPUs, 0 hotplug CPUs May 14 23:51:14.940436 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() May 14 23:51:14.940447 kernel: kvm-guest: KVM setup pv remote TLB flush May 14 23:51:14.940455 kernel: kvm-guest: setup PV sched yield May 14 23:51:14.940463 kernel: [mem 0x9d000000-0xdfffffff] available for PCI devices May 14 23:51:14.940470 kernel: Booting paravirtualized kernel on KVM May 14 23:51:14.940481 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns May 14 23:51:14.940489 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:4 nr_cpu_ids:4 nr_node_ids:1 May 14 23:51:14.940497 kernel: percpu: Embedded 58 pages/cpu s197032 r8192 d32344 u524288 May 14 23:51:14.940505 kernel: pcpu-alloc: s197032 r8192 d32344 u524288 alloc=1*2097152 May 14 23:51:14.940513 kernel: pcpu-alloc: [0] 0 1 2 3 May 14 23:51:14.940523 kernel: kvm-guest: PV spinlocks enabled May 14 23:51:14.940531 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) May 14 23:51:14.940540 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=e0c956f61127e47bb23a2bdeb0592b0ff91bd857e2344d0bf321acb67c279f1a May 14 23:51:14.940549 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. May 14 23:51:14.940557 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) May 14 23:51:14.940567 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) May 14 23:51:14.940575 kernel: Fallback order for Node 0: 0 May 14 23:51:14.940583 kernel: Built 1 zonelists, mobility grouping on. Total pages: 629460 May 14 23:51:14.940592 kernel: Policy zone: DMA32 May 14 23:51:14.940602 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off May 14 23:51:14.940611 kernel: Memory: 2385672K/2565800K available (14336K kernel code, 2296K rwdata, 25068K rodata, 43604K init, 1468K bss, 179872K reserved, 0K cma-reserved) May 14 23:51:14.940619 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 May 14 23:51:14.940627 kernel: ftrace: allocating 37993 entries in 149 pages May 14 23:51:14.940635 kernel: ftrace: allocated 149 pages with 4 groups May 14 23:51:14.940643 kernel: Dynamic Preempt: voluntary May 14 23:51:14.940650 kernel: rcu: Preemptible hierarchical RCU implementation. May 14 23:51:14.940659 kernel: rcu: RCU event tracing is enabled. May 14 23:51:14.940667 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. May 14 23:51:14.940678 kernel: Trampoline variant of Tasks RCU enabled. May 14 23:51:14.940686 kernel: Rude variant of Tasks RCU enabled. May 14 23:51:14.940694 kernel: Tracing variant of Tasks RCU enabled. May 14 23:51:14.940702 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. May 14 23:51:14.940710 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 May 14 23:51:14.940740 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 May 14 23:51:14.940749 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. May 14 23:51:14.940757 kernel: Console: colour dummy device 80x25 May 14 23:51:14.940765 kernel: printk: console [ttyS0] enabled May 14 23:51:14.940776 kernel: ACPI: Core revision 20230628 May 14 23:51:14.940792 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns May 14 23:51:14.940814 kernel: APIC: Switch to symmetric I/O mode setup May 14 23:51:14.940823 kernel: x2apic enabled May 14 23:51:14.940844 kernel: APIC: Switched APIC routing to: physical x2apic May 14 23:51:14.940870 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() May 14 23:51:14.940889 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() May 14 23:51:14.940909 kernel: kvm-guest: setup PV IPIs May 14 23:51:14.940918 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 May 14 23:51:14.940929 kernel: tsc: Marking TSC unstable due to TSCs unsynchronized May 14 23:51:14.940938 kernel: Calibrating delay loop (skipped) preset value.. 5589.49 BogoMIPS (lpj=2794748) May 14 23:51:14.940946 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated May 14 23:51:14.940954 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 May 14 23:51:14.940962 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 May 14 23:51:14.940970 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization May 14 23:51:14.940978 kernel: Spectre V2 : Mitigation: Retpolines May 14 23:51:14.940986 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT May 14 23:51:14.940994 kernel: Spectre V2 : Enabling Speculation Barrier for firmware calls May 14 23:51:14.941005 kernel: RETBleed: Mitigation: untrained return thunk May 14 23:51:14.941013 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier May 14 23:51:14.941021 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl May 14 23:51:14.941029 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! May 14 23:51:14.941038 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. May 14 23:51:14.941046 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode May 14 23:51:14.941057 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' May 14 23:51:14.941065 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' May 14 23:51:14.941075 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' May 14 23:51:14.941083 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 May 14 23:51:14.941091 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. May 14 23:51:14.941099 kernel: Freeing SMP alternatives memory: 32K May 14 23:51:14.941107 kernel: pid_max: default: 32768 minimum: 301 May 14 23:51:14.941115 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity May 14 23:51:14.941123 kernel: landlock: Up and running. May 14 23:51:14.941131 kernel: SELinux: Initializing. May 14 23:51:14.941139 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) May 14 23:51:14.941150 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) May 14 23:51:14.941159 kernel: smpboot: CPU0: AMD EPYC 7402P 24-Core Processor (family: 0x17, model: 0x31, stepping: 0x0) May 14 23:51:14.941167 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. May 14 23:51:14.941175 kernel: RCU Tasks Rude: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. May 14 23:51:14.941183 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. May 14 23:51:14.941191 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. May 14 23:51:14.941199 kernel: ... version: 0 May 14 23:51:14.941207 kernel: ... bit width: 48 May 14 23:51:14.941225 kernel: ... generic registers: 6 May 14 23:51:14.941242 kernel: ... value mask: 0000ffffffffffff May 14 23:51:14.941250 kernel: ... max period: 00007fffffffffff May 14 23:51:14.941258 kernel: ... fixed-purpose events: 0 May 14 23:51:14.941266 kernel: ... event mask: 000000000000003f May 14 23:51:14.941274 kernel: signal: max sigframe size: 1776 May 14 23:51:14.941294 kernel: rcu: Hierarchical SRCU implementation. May 14 23:51:14.941316 kernel: rcu: Max phase no-delay instances is 400. May 14 23:51:14.941328 kernel: smp: Bringing up secondary CPUs ... May 14 23:51:14.941336 kernel: smpboot: x86: Booting SMP configuration: May 14 23:51:14.941347 kernel: .... node #0, CPUs: #1 #2 #3 May 14 23:51:14.941355 kernel: smp: Brought up 1 node, 4 CPUs May 14 23:51:14.941363 kernel: smpboot: Max logical packages: 1 May 14 23:51:14.941371 kernel: smpboot: Total of 4 processors activated (22357.98 BogoMIPS) May 14 23:51:14.941379 kernel: devtmpfs: initialized May 14 23:51:14.941387 kernel: x86/mm: Memory block size: 128MB May 14 23:51:14.941395 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00800000-0x00807fff] (32768 bytes) May 14 23:51:14.941403 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x0080b000-0x0080bfff] (4096 bytes) May 14 23:51:14.941411 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00811000-0x008fffff] (978944 bytes) May 14 23:51:14.941422 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cb7f000-0x9cbfefff] (524288 bytes) May 14 23:51:14.941442 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9ce96000-0x9ce97fff] (8192 bytes) May 14 23:51:14.941460 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cf60000-0x9cffffff] (655360 bytes) May 14 23:51:14.941482 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns May 14 23:51:14.941501 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) May 14 23:51:14.941522 kernel: pinctrl core: initialized pinctrl subsystem May 14 23:51:14.941537 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family May 14 23:51:14.941545 kernel: audit: initializing netlink subsys (disabled) May 14 23:51:14.941553 kernel: audit: type=2000 audit(1747266674.321:1): state=initialized audit_enabled=0 res=1 May 14 23:51:14.941565 kernel: thermal_sys: Registered thermal governor 'step_wise' May 14 23:51:14.941573 kernel: thermal_sys: Registered thermal governor 'user_space' May 14 23:51:14.941581 kernel: cpuidle: using governor menu May 14 23:51:14.941601 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 May 14 23:51:14.941622 kernel: dca service started, version 1.12.1 May 14 23:51:14.941643 kernel: PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xe0000000-0xefffffff] (base 0xe0000000) May 14 23:51:14.941664 kernel: PCI: Using configuration type 1 for base access May 14 23:51:14.941683 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. May 14 23:51:14.941699 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages May 14 23:51:14.941710 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page May 14 23:51:14.941734 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages May 14 23:51:14.941743 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page May 14 23:51:14.941751 kernel: ACPI: Added _OSI(Module Device) May 14 23:51:14.943014 kernel: ACPI: Added _OSI(Processor Device) May 14 23:51:14.943023 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) May 14 23:51:14.943031 kernel: ACPI: Added _OSI(Processor Aggregator Device) May 14 23:51:14.943040 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded May 14 23:51:14.943058 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC May 14 23:51:14.943083 kernel: ACPI: Interpreter enabled May 14 23:51:14.943100 kernel: ACPI: PM: (supports S0 S3 S5) May 14 23:51:14.943117 kernel: ACPI: Using IOAPIC for interrupt routing May 14 23:51:14.943140 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug May 14 23:51:14.943160 kernel: PCI: Using E820 reservations for host bridge windows May 14 23:51:14.943177 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F May 14 23:51:14.943194 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) May 14 23:51:14.943618 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] May 14 23:51:14.943982 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] May 14 23:51:14.944370 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] May 14 23:51:14.944408 kernel: PCI host bridge to bus 0000:00 May 14 23:51:14.944658 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] May 14 23:51:14.944861 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] May 14 23:51:14.945012 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] May 14 23:51:14.945137 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xdfffffff window] May 14 23:51:14.945302 kernel: pci_bus 0000:00: root bus resource [mem 0xf0000000-0xfebfffff window] May 14 23:51:14.945546 kernel: pci_bus 0000:00: root bus resource [mem 0x380000000000-0x3807ffffffff window] May 14 23:51:14.945681 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] May 14 23:51:14.945933 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 May 14 23:51:14.946087 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 May 14 23:51:14.946225 kernel: pci 0000:00:01.0: reg 0x10: [mem 0xc0000000-0xc0ffffff pref] May 14 23:51:14.946364 kernel: pci 0000:00:01.0: reg 0x18: [mem 0xc1044000-0xc1044fff] May 14 23:51:14.946514 kernel: pci 0000:00:01.0: reg 0x30: [mem 0xffff0000-0xffffffff pref] May 14 23:51:14.946661 kernel: pci 0000:00:01.0: BAR 0: assigned to efifb May 14 23:51:14.946846 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] May 14 23:51:14.947036 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 May 14 23:51:14.947238 kernel: pci 0000:00:02.0: reg 0x10: [io 0x6100-0x611f] May 14 23:51:14.947422 kernel: pci 0000:00:02.0: reg 0x14: [mem 0xc1043000-0xc1043fff] May 14 23:51:14.947618 kernel: pci 0000:00:02.0: reg 0x20: [mem 0x380000000000-0x380000003fff 64bit pref] May 14 23:51:14.947819 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 May 14 23:51:14.948038 kernel: pci 0000:00:03.0: reg 0x10: [io 0x6000-0x607f] May 14 23:51:14.948198 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xc1042000-0xc1042fff] May 14 23:51:14.948338 kernel: pci 0000:00:03.0: reg 0x20: [mem 0x380000004000-0x380000007fff 64bit pref] May 14 23:51:14.948535 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 May 14 23:51:14.948779 kernel: pci 0000:00:04.0: reg 0x10: [io 0x60e0-0x60ff] May 14 23:51:14.948942 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xc1041000-0xc1041fff] May 14 23:51:14.949107 kernel: pci 0000:00:04.0: reg 0x20: [mem 0x380000008000-0x38000000bfff 64bit pref] May 14 23:51:14.949251 kernel: pci 0000:00:04.0: reg 0x30: [mem 0xfffc0000-0xffffffff pref] May 14 23:51:14.949421 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 May 14 23:51:14.949560 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO May 14 23:51:14.949747 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 May 14 23:51:14.949925 kernel: pci 0000:00:1f.2: reg 0x20: [io 0x60c0-0x60df] May 14 23:51:14.950061 kernel: pci 0000:00:1f.2: reg 0x24: [mem 0xc1040000-0xc1040fff] May 14 23:51:14.950205 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 May 14 23:51:14.950338 kernel: pci 0000:00:1f.3: reg 0x20: [io 0x6080-0x60bf] May 14 23:51:14.950350 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 May 14 23:51:14.950359 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 May 14 23:51:14.950367 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 May 14 23:51:14.950375 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 May 14 23:51:14.950388 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 May 14 23:51:14.950396 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 May 14 23:51:14.950405 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 May 14 23:51:14.950413 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 May 14 23:51:14.950421 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 May 14 23:51:14.950429 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 May 14 23:51:14.950445 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 May 14 23:51:14.950453 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 May 14 23:51:14.950461 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 May 14 23:51:14.950472 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 May 14 23:51:14.950490 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 May 14 23:51:14.950505 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 May 14 23:51:14.950524 kernel: iommu: Default domain type: Translated May 14 23:51:14.950548 kernel: iommu: DMA domain TLB invalidation policy: lazy mode May 14 23:51:14.950568 kernel: efivars: Registered efivars operations May 14 23:51:14.950590 kernel: PCI: Using ACPI for IRQ routing May 14 23:51:14.950611 kernel: PCI: pci_cache_line_size set to 64 bytes May 14 23:51:14.950636 kernel: e820: reserve RAM buffer [mem 0x0080b000-0x008fffff] May 14 23:51:14.950663 kernel: e820: reserve RAM buffer [mem 0x00811000-0x008fffff] May 14 23:51:14.950684 kernel: e820: reserve RAM buffer [mem 0x9b314018-0x9bffffff] May 14 23:51:14.950741 kernel: e820: reserve RAM buffer [mem 0x9b351018-0x9bffffff] May 14 23:51:14.950765 kernel: e820: reserve RAM buffer [mem 0x9bd3f000-0x9bffffff] May 14 23:51:14.950778 kernel: e820: reserve RAM buffer [mem 0x9c8ed000-0x9fffffff] May 14 23:51:14.950797 kernel: e820: reserve RAM buffer [mem 0x9ce92000-0x9fffffff] May 14 23:51:14.950805 kernel: e820: reserve RAM buffer [mem 0x9cedc000-0x9fffffff] May 14 23:51:14.951827 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device May 14 23:51:14.952094 kernel: pci 0000:00:01.0: vgaarb: bridge control possible May 14 23:51:14.952503 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none May 14 23:51:14.952531 kernel: vgaarb: loaded May 14 23:51:14.952550 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 May 14 23:51:14.952571 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter May 14 23:51:14.952590 kernel: clocksource: Switched to clocksource kvm-clock May 14 23:51:14.952614 kernel: VFS: Disk quotas dquot_6.6.0 May 14 23:51:14.952633 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) May 14 23:51:14.952664 kernel: pnp: PnP ACPI init May 14 23:51:14.953293 kernel: system 00:05: [mem 0xe0000000-0xefffffff window] has been reserved May 14 23:51:14.953323 kernel: pnp: PnP ACPI: found 6 devices May 14 23:51:14.953347 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns May 14 23:51:14.953372 kernel: NET: Registered PF_INET protocol family May 14 23:51:14.953396 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) May 14 23:51:14.953470 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) May 14 23:51:14.953490 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) May 14 23:51:14.953506 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) May 14 23:51:14.953531 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) May 14 23:51:14.953550 kernel: TCP: Hash tables configured (established 32768 bind 32768) May 14 23:51:14.953562 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) May 14 23:51:14.953570 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) May 14 23:51:14.953579 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family May 14 23:51:14.953587 kernel: NET: Registered PF_XDP protocol family May 14 23:51:14.954245 kernel: pci 0000:00:04.0: can't claim BAR 6 [mem 0xfffc0000-0xffffffff pref]: no compatible bridge window May 14 23:51:14.956533 kernel: pci 0000:00:04.0: BAR 6: assigned [mem 0x9d000000-0x9d03ffff pref] May 14 23:51:14.956682 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] May 14 23:51:14.956834 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] May 14 23:51:14.956968 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] May 14 23:51:14.957264 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xdfffffff window] May 14 23:51:14.957466 kernel: pci_bus 0000:00: resource 8 [mem 0xf0000000-0xfebfffff window] May 14 23:51:14.957649 kernel: pci_bus 0000:00: resource 9 [mem 0x380000000000-0x3807ffffffff window] May 14 23:51:14.957664 kernel: PCI: CLS 0 bytes, default 64 May 14 23:51:14.957672 kernel: Initialise system trusted keyrings May 14 23:51:14.957681 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 May 14 23:51:14.957695 kernel: Key type asymmetric registered May 14 23:51:14.957703 kernel: Asymmetric key parser 'x509' registered May 14 23:51:14.957712 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) May 14 23:51:14.957740 kernel: io scheduler mq-deadline registered May 14 23:51:14.957750 kernel: io scheduler kyber registered May 14 23:51:14.957758 kernel: io scheduler bfq registered May 14 23:51:14.957767 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 May 14 23:51:14.957776 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 May 14 23:51:14.957792 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 May 14 23:51:14.957807 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 May 14 23:51:14.957818 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled May 14 23:51:14.957827 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A May 14 23:51:14.957836 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 May 14 23:51:14.957845 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 May 14 23:51:14.957854 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 May 14 23:51:14.957866 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 May 14 23:51:14.958032 kernel: rtc_cmos 00:04: RTC can wake from S4 May 14 23:51:14.958160 kernel: rtc_cmos 00:04: registered as rtc0 May 14 23:51:14.959781 kernel: rtc_cmos 00:04: setting system clock to 2025-05-14T23:51:14 UTC (1747266674) May 14 23:51:14.960238 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram May 14 23:51:14.960265 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled May 14 23:51:14.960283 kernel: efifb: probing for efifb May 14 23:51:14.960312 kernel: efifb: framebuffer at 0xc0000000, using 4000k, total 4000k May 14 23:51:14.960350 kernel: efifb: mode is 1280x800x32, linelength=5120, pages=1 May 14 23:51:14.960379 kernel: efifb: scrolling: redraw May 14 23:51:14.960411 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 May 14 23:51:14.960430 kernel: Console: switching to colour frame buffer device 160x50 May 14 23:51:14.960447 kernel: fb0: EFI VGA frame buffer device May 14 23:51:14.960469 kernel: pstore: Using crash dump compression: deflate May 14 23:51:14.960487 kernel: pstore: Registered efi_pstore as persistent store backend May 14 23:51:14.960507 kernel: NET: Registered PF_INET6 protocol family May 14 23:51:14.960525 kernel: Segment Routing with IPv6 May 14 23:51:14.960546 kernel: In-situ OAM (IOAM) with IPv6 May 14 23:51:14.960558 kernel: NET: Registered PF_PACKET protocol family May 14 23:51:14.960566 kernel: Key type dns_resolver registered May 14 23:51:14.960574 kernel: IPI shorthand broadcast: enabled May 14 23:51:14.960583 kernel: sched_clock: Marking stable (1241006138, 158086024)->(1468995966, -69903804) May 14 23:51:14.960600 kernel: registered taskstats version 1 May 14 23:51:14.960611 kernel: Loading compiled-in X.509 certificates May 14 23:51:14.960627 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.89-flatcar: 4f9bc5b8797c7efeb1fcd74892dea83a6cb9d390' May 14 23:51:14.960636 kernel: Key type .fscrypt registered May 14 23:51:14.960648 kernel: Key type fscrypt-provisioning registered May 14 23:51:14.960657 kernel: ima: No TPM chip found, activating TPM-bypass! May 14 23:51:14.960665 kernel: ima: Allocated hash algorithm: sha1 May 14 23:51:14.960674 kernel: ima: No architecture policies found May 14 23:51:14.960685 kernel: clk: Disabling unused clocks May 14 23:51:14.960696 kernel: Freeing unused kernel image (initmem) memory: 43604K May 14 23:51:14.960705 kernel: Write protecting the kernel read-only data: 40960k May 14 23:51:14.960713 kernel: Freeing unused kernel image (rodata/data gap) memory: 1556K May 14 23:51:14.960760 kernel: Run /init as init process May 14 23:51:14.960786 kernel: with arguments: May 14 23:51:14.960799 kernel: /init May 14 23:51:14.960810 kernel: with environment: May 14 23:51:14.960824 kernel: HOME=/ May 14 23:51:14.960838 kernel: TERM=linux May 14 23:51:14.960847 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a May 14 23:51:14.960872 systemd[1]: Successfully made /usr/ read-only. May 14 23:51:14.960895 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) May 14 23:51:14.960920 systemd[1]: Detected virtualization kvm. May 14 23:51:14.960930 systemd[1]: Detected architecture x86-64. May 14 23:51:14.960938 systemd[1]: Running in initrd. May 14 23:51:14.960950 systemd[1]: No hostname configured, using default hostname. May 14 23:51:14.960971 systemd[1]: Hostname set to . May 14 23:51:14.960980 systemd[1]: Initializing machine ID from VM UUID. May 14 23:51:14.960991 systemd[1]: Queued start job for default target initrd.target. May 14 23:51:14.961011 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. May 14 23:51:14.961024 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. May 14 23:51:14.961034 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... May 14 23:51:14.961043 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... May 14 23:51:14.961061 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... May 14 23:51:14.961072 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... May 14 23:51:14.961094 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... May 14 23:51:14.961115 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... May 14 23:51:14.961138 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). May 14 23:51:14.961165 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. May 14 23:51:14.961190 systemd[1]: Reached target paths.target - Path Units. May 14 23:51:14.961210 systemd[1]: Reached target slices.target - Slice Units. May 14 23:51:14.961230 systemd[1]: Reached target swap.target - Swaps. May 14 23:51:14.961253 systemd[1]: Reached target timers.target - Timer Units. May 14 23:51:14.961270 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. May 14 23:51:14.961292 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. May 14 23:51:14.963109 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). May 14 23:51:14.963119 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. May 14 23:51:14.963128 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. May 14 23:51:14.963140 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. May 14 23:51:14.963150 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. May 14 23:51:14.963159 systemd[1]: Reached target sockets.target - Socket Units. May 14 23:51:14.963168 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... May 14 23:51:14.963177 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... May 14 23:51:14.963190 systemd[1]: Finished network-cleanup.service - Network Cleanup. May 14 23:51:14.963204 systemd[1]: Starting systemd-fsck-usr.service... May 14 23:51:14.963226 systemd[1]: Starting systemd-journald.service - Journal Service... May 14 23:51:14.963249 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... May 14 23:51:14.963259 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 14 23:51:14.963272 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. May 14 23:51:14.963284 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. May 14 23:51:14.963298 systemd[1]: Finished systemd-fsck-usr.service. May 14 23:51:14.963307 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... May 14 23:51:14.963317 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. May 14 23:51:14.963359 systemd-journald[193]: Collecting audit messages is disabled. May 14 23:51:14.963385 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... May 14 23:51:14.963395 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. May 14 23:51:14.963404 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... May 14 23:51:14.963414 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. May 14 23:51:14.963423 systemd-journald[193]: Journal started May 14 23:51:14.963450 systemd-journald[193]: Runtime Journal (/run/log/journal/0661f8a4e3b44fb5833446b10e4e8ddf) is 6M, max 48.2M, 42.2M free. May 14 23:51:14.926359 systemd-modules-load[194]: Inserted module 'overlay' May 14 23:51:14.966503 systemd-modules-load[194]: Inserted module 'br_netfilter' May 14 23:51:14.967559 kernel: Bridge firewalling registered May 14 23:51:14.971305 systemd[1]: Started systemd-journald.service - Journal Service. May 14 23:51:14.972905 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. May 14 23:51:14.978938 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... May 14 23:51:14.980587 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... May 14 23:51:14.981767 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. May 14 23:51:14.985058 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. May 14 23:51:14.992412 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... May 14 23:51:15.002479 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. May 14 23:51:15.005025 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. May 14 23:51:15.010404 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... May 14 23:51:15.013048 dracut-cmdline[224]: dracut-dracut-053 May 14 23:51:15.015024 dracut-cmdline[224]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=e0c956f61127e47bb23a2bdeb0592b0ff91bd857e2344d0bf321acb67c279f1a May 14 23:51:15.074778 systemd-resolved[237]: Positive Trust Anchors: May 14 23:51:15.074863 systemd-resolved[237]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d May 14 23:51:15.074918 systemd-resolved[237]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test May 14 23:51:15.079678 systemd-resolved[237]: Defaulting to hostname 'linux'. May 14 23:51:15.081629 systemd[1]: Started systemd-resolved.service - Network Name Resolution. May 14 23:51:15.086836 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. May 14 23:51:15.121802 kernel: SCSI subsystem initialized May 14 23:51:15.130763 kernel: Loading iSCSI transport class v2.0-870. May 14 23:51:15.141753 kernel: iscsi: registered transport (tcp) May 14 23:51:15.163748 kernel: iscsi: registered transport (qla4xxx) May 14 23:51:15.163775 kernel: QLogic iSCSI HBA Driver May 14 23:51:15.222587 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. May 14 23:51:15.226938 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... May 14 23:51:15.265925 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. May 14 23:51:15.266024 kernel: device-mapper: uevent: version 1.0.3 May 14 23:51:15.266042 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com May 14 23:51:15.310763 kernel: raid6: avx2x4 gen() 29879 MB/s May 14 23:51:15.327762 kernel: raid6: avx2x2 gen() 30744 MB/s May 14 23:51:15.344889 kernel: raid6: avx2x1 gen() 25259 MB/s May 14 23:51:15.344968 kernel: raid6: using algorithm avx2x2 gen() 30744 MB/s May 14 23:51:15.362849 kernel: raid6: .... xor() 19755 MB/s, rmw enabled May 14 23:51:15.362878 kernel: raid6: using avx2x2 recovery algorithm May 14 23:51:15.385746 kernel: xor: automatically using best checksumming function avx May 14 23:51:15.538752 kernel: Btrfs loaded, zoned=no, fsverity=no May 14 23:51:15.554098 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. May 14 23:51:15.557374 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... May 14 23:51:15.585920 systemd-udevd[415]: Using default interface naming scheme 'v255'. May 14 23:51:15.592003 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. May 14 23:51:15.597231 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... May 14 23:51:15.630194 dracut-pre-trigger[423]: rd.md=0: removing MD RAID activation May 14 23:51:15.667057 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. May 14 23:51:15.670695 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... May 14 23:51:15.752104 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. May 14 23:51:15.755897 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... May 14 23:51:15.781909 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. May 14 23:51:15.785095 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. May 14 23:51:15.786880 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. May 14 23:51:15.789226 systemd[1]: Reached target remote-fs.target - Remote File Systems. May 14 23:51:15.794744 kernel: virtio_blk virtio1: 4/0/0 default/read/poll queues May 14 23:51:15.795833 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... May 14 23:51:15.799855 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) May 14 23:51:15.807395 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. May 14 23:51:15.807437 kernel: GPT:9289727 != 19775487 May 14 23:51:15.807453 kernel: GPT:Alternate GPT header not at the end of the disk. May 14 23:51:15.807468 kernel: GPT:9289727 != 19775487 May 14 23:51:15.807482 kernel: GPT: Use GNU Parted to correct GPT errors. May 14 23:51:15.807496 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 May 14 23:51:15.815762 kernel: cryptd: max_cpu_qlen set to 1000 May 14 23:51:15.822755 kernel: libata version 3.00 loaded. May 14 23:51:15.829210 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. May 14 23:51:15.843256 kernel: ahci 0000:00:1f.2: version 3.0 May 14 23:51:15.843559 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 May 14 23:51:15.843579 kernel: AVX2 version of gcm_enc/dec engaged. May 14 23:51:15.844807 kernel: AES CTR mode by8 optimization enabled May 14 23:51:15.845739 kernel: ahci 0000:00:1f.2: AHCI 0001.0000 32 slots 6 ports 1.5 Gbps 0x3f impl SATA mode May 14 23:51:15.846989 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. May 14 23:51:15.854693 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only May 14 23:51:15.847080 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. May 14 23:51:15.859070 kernel: scsi host0: ahci May 14 23:51:15.848143 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... May 14 23:51:15.863811 kernel: scsi host1: ahci May 14 23:51:15.848503 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 14 23:51:15.866187 kernel: scsi host2: ahci May 14 23:51:15.848579 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. May 14 23:51:15.869200 kernel: scsi host3: ahci May 14 23:51:15.852590 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... May 14 23:51:15.858213 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 14 23:51:15.873153 kernel: scsi host4: ahci May 14 23:51:15.875415 kernel: scsi host5: ahci May 14 23:51:15.875655 kernel: ata1: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040100 irq 34 May 14 23:51:15.875673 kernel: ata2: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040180 irq 34 May 14 23:51:15.877568 kernel: ata3: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040200 irq 34 May 14 23:51:15.877595 kernel: ata4: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040280 irq 34 May 14 23:51:15.881132 kernel: ata5: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040300 irq 34 May 14 23:51:15.881161 kernel: BTRFS: device fsid 267fa270-7a71-43aa-9209-0280512688b5 devid 1 transid 41 /dev/vda3 scanned by (udev-worker) (468) May 14 23:51:15.881183 kernel: ata6: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040380 irq 34 May 14 23:51:15.883428 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 14 23:51:15.884193 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. May 14 23:51:15.894738 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/vda6 scanned by (udev-worker) (473) May 14 23:51:15.898820 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. May 14 23:51:15.915437 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. May 14 23:51:15.922250 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. May 14 23:51:15.922569 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. May 14 23:51:15.932185 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. May 14 23:51:15.942815 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. May 14 23:51:15.945070 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... May 14 23:51:15.947132 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 14 23:51:15.972405 disk-uuid[558]: Primary Header is updated. May 14 23:51:15.972405 disk-uuid[558]: Secondary Entries is updated. May 14 23:51:15.972405 disk-uuid[558]: Secondary Header is updated. May 14 23:51:15.975347 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. May 14 23:51:15.978757 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 May 14 23:51:15.978930 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... May 14 23:51:16.013387 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. May 14 23:51:16.188756 kernel: ata1: SATA link down (SStatus 0 SControl 300) May 14 23:51:16.188845 kernel: ata2: SATA link down (SStatus 0 SControl 300) May 14 23:51:16.196744 kernel: ata4: SATA link down (SStatus 0 SControl 300) May 14 23:51:16.196785 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) May 14 23:51:16.197769 kernel: ata5: SATA link down (SStatus 0 SControl 300) May 14 23:51:16.197865 kernel: ata6: SATA link down (SStatus 0 SControl 300) May 14 23:51:16.198748 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 May 14 23:51:16.199750 kernel: ata3.00: applying bridge limits May 14 23:51:16.199767 kernel: ata3.00: configured for UDMA/100 May 14 23:51:16.200746 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 May 14 23:51:16.260757 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray May 14 23:51:16.261105 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 May 14 23:51:16.274982 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 May 14 23:51:16.989366 disk-uuid[560]: The operation has completed successfully. May 14 23:51:16.991105 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 May 14 23:51:17.023699 systemd[1]: disk-uuid.service: Deactivated successfully. May 14 23:51:17.023859 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. May 14 23:51:17.062614 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... May 14 23:51:17.078928 sh[597]: Success May 14 23:51:17.095756 kernel: device-mapper: verity: sha256 using implementation "sha256-ni" May 14 23:51:17.159275 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. May 14 23:51:17.163324 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... May 14 23:51:17.187642 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. May 14 23:51:17.194386 kernel: BTRFS info (device dm-0): first mount of filesystem 267fa270-7a71-43aa-9209-0280512688b5 May 14 23:51:17.194445 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm May 14 23:51:17.194481 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead May 14 23:51:17.195633 kernel: BTRFS info (device dm-0): disabling log replay at mount time May 14 23:51:17.196536 kernel: BTRFS info (device dm-0): using free space tree May 14 23:51:17.202602 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. May 14 23:51:17.205159 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. May 14 23:51:17.207958 systemd[1]: Starting ignition-setup.service - Ignition (setup)... May 14 23:51:17.210569 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... May 14 23:51:17.252306 kernel: BTRFS info (device vda6): first mount of filesystem 4c949817-d4f4-485b-8019-80887ee5206f May 14 23:51:17.252391 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm May 14 23:51:17.252407 kernel: BTRFS info (device vda6): using free space tree May 14 23:51:17.254756 kernel: BTRFS info (device vda6): auto enabling async discard May 14 23:51:17.260742 kernel: BTRFS info (device vda6): last unmount of filesystem 4c949817-d4f4-485b-8019-80887ee5206f May 14 23:51:17.285969 systemd[1]: Finished ignition-setup.service - Ignition (setup). May 14 23:51:17.288620 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... May 14 23:51:17.359051 ignition[712]: Ignition 2.20.0 May 14 23:51:17.359070 ignition[712]: Stage: fetch-offline May 14 23:51:17.359124 ignition[712]: no configs at "/usr/lib/ignition/base.d" May 14 23:51:17.359139 ignition[712]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" May 14 23:51:17.359264 ignition[712]: parsed url from cmdline: "" May 14 23:51:17.359269 ignition[712]: no config URL provided May 14 23:51:17.359277 ignition[712]: reading system config file "/usr/lib/ignition/user.ign" May 14 23:51:17.359289 ignition[712]: no config at "/usr/lib/ignition/user.ign" May 14 23:51:17.359330 ignition[712]: op(1): [started] loading QEMU firmware config module May 14 23:51:17.359337 ignition[712]: op(1): executing: "modprobe" "qemu_fw_cfg" May 14 23:51:17.376987 ignition[712]: op(1): [finished] loading QEMU firmware config module May 14 23:51:17.416939 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. May 14 23:51:17.421573 systemd[1]: Starting systemd-networkd.service - Network Configuration... May 14 23:51:17.426077 ignition[712]: parsing config with SHA512: aa8e222550f29c2b4d4749ab8c321d9ed1bafb748e7a67a7a80702e9638217e778036ed2104c92b39af5935b3b19ee7331c8f9e3fc420e660e65b19f25cdb92e May 14 23:51:17.433355 unknown[712]: fetched base config from "system" May 14 23:51:17.433375 unknown[712]: fetched user config from "qemu" May 14 23:51:17.435973 ignition[712]: fetch-offline: fetch-offline passed May 14 23:51:17.436107 ignition[712]: Ignition finished successfully May 14 23:51:17.439274 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). May 14 23:51:17.472023 systemd-networkd[784]: lo: Link UP May 14 23:51:17.472034 systemd-networkd[784]: lo: Gained carrier May 14 23:51:17.475411 systemd-networkd[784]: Enumeration completed May 14 23:51:17.475690 systemd[1]: Started systemd-networkd.service - Network Configuration. May 14 23:51:17.476422 systemd[1]: Reached target network.target - Network. May 14 23:51:17.478747 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). May 14 23:51:17.479944 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... May 14 23:51:17.483951 systemd-networkd[784]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 14 23:51:17.483961 systemd-networkd[784]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. May 14 23:51:17.488403 systemd-networkd[784]: eth0: Link UP May 14 23:51:17.488412 systemd-networkd[784]: eth0: Gained carrier May 14 23:51:17.488419 systemd-networkd[784]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 14 23:51:17.509786 systemd-networkd[784]: eth0: DHCPv4 address 10.0.0.142/16, gateway 10.0.0.1 acquired from 10.0.0.1 May 14 23:51:17.518914 ignition[788]: Ignition 2.20.0 May 14 23:51:17.518932 ignition[788]: Stage: kargs May 14 23:51:17.519148 ignition[788]: no configs at "/usr/lib/ignition/base.d" May 14 23:51:17.519166 ignition[788]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" May 14 23:51:17.520029 ignition[788]: kargs: kargs passed May 14 23:51:17.520079 ignition[788]: Ignition finished successfully May 14 23:51:17.527320 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). May 14 23:51:17.529585 systemd[1]: Starting ignition-disks.service - Ignition (disks)... May 14 23:51:17.590119 ignition[798]: Ignition 2.20.0 May 14 23:51:17.590135 ignition[798]: Stage: disks May 14 23:51:17.590335 ignition[798]: no configs at "/usr/lib/ignition/base.d" May 14 23:51:17.590350 ignition[798]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" May 14 23:51:17.591440 ignition[798]: disks: disks passed May 14 23:51:17.591501 ignition[798]: Ignition finished successfully May 14 23:51:17.598999 systemd[1]: Finished ignition-disks.service - Ignition (disks). May 14 23:51:17.599889 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. May 14 23:51:17.600239 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. May 14 23:51:17.600698 systemd[1]: Reached target local-fs.target - Local File Systems. May 14 23:51:17.601368 systemd[1]: Reached target sysinit.target - System Initialization. May 14 23:51:17.602012 systemd[1]: Reached target basic.target - Basic System. May 14 23:51:17.613808 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... May 14 23:51:17.645901 systemd-fsck[808]: ROOT: clean, 14/553520 files, 52654/553472 blocks May 14 23:51:17.652589 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. May 14 23:51:17.656010 systemd[1]: Mounting sysroot.mount - /sysroot... May 14 23:51:17.774748 kernel: EXT4-fs (vda9): mounted filesystem 81735587-bac5-4d9e-ae49-5642e655af7f r/w with ordered data mode. Quota mode: none. May 14 23:51:17.776020 systemd[1]: Mounted sysroot.mount - /sysroot. May 14 23:51:17.777713 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. May 14 23:51:17.780609 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... May 14 23:51:17.782827 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... May 14 23:51:17.784107 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. May 14 23:51:17.784160 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). May 14 23:51:17.784196 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. May 14 23:51:17.800601 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. May 14 23:51:17.805377 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/vda6 scanned by mount (816) May 14 23:51:17.802864 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... May 14 23:51:17.810051 kernel: BTRFS info (device vda6): first mount of filesystem 4c949817-d4f4-485b-8019-80887ee5206f May 14 23:51:17.810073 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm May 14 23:51:17.810089 kernel: BTRFS info (device vda6): using free space tree May 14 23:51:17.811759 kernel: BTRFS info (device vda6): auto enabling async discard May 14 23:51:17.814779 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. May 14 23:51:17.850346 initrd-setup-root[840]: cut: /sysroot/etc/passwd: No such file or directory May 14 23:51:17.856010 initrd-setup-root[847]: cut: /sysroot/etc/group: No such file or directory May 14 23:51:17.861325 initrd-setup-root[854]: cut: /sysroot/etc/shadow: No such file or directory May 14 23:51:17.867276 initrd-setup-root[861]: cut: /sysroot/etc/gshadow: No such file or directory May 14 23:51:18.032909 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. May 14 23:51:18.034890 systemd[1]: Starting ignition-mount.service - Ignition (mount)... May 14 23:51:18.038463 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... May 14 23:51:18.060757 kernel: BTRFS info (device vda6): last unmount of filesystem 4c949817-d4f4-485b-8019-80887ee5206f May 14 23:51:18.078974 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. May 14 23:51:18.101432 ignition[930]: INFO : Ignition 2.20.0 May 14 23:51:18.101432 ignition[930]: INFO : Stage: mount May 14 23:51:18.103595 ignition[930]: INFO : no configs at "/usr/lib/ignition/base.d" May 14 23:51:18.103595 ignition[930]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" May 14 23:51:18.103595 ignition[930]: INFO : mount: mount passed May 14 23:51:18.103595 ignition[930]: INFO : Ignition finished successfully May 14 23:51:18.110099 systemd[1]: Finished ignition-mount.service - Ignition (mount). May 14 23:51:18.111868 systemd[1]: Starting ignition-files.service - Ignition (files)... May 14 23:51:18.194237 systemd[1]: sysroot-oem.mount: Deactivated successfully. May 14 23:51:18.196786 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... May 14 23:51:18.222768 kernel: BTRFS: device label OEM devid 1 transid 16 /dev/vda6 scanned by mount (943) May 14 23:51:18.225140 kernel: BTRFS info (device vda6): first mount of filesystem 4c949817-d4f4-485b-8019-80887ee5206f May 14 23:51:18.225239 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm May 14 23:51:18.225254 kernel: BTRFS info (device vda6): using free space tree May 14 23:51:18.229759 kernel: BTRFS info (device vda6): auto enabling async discard May 14 23:51:18.231777 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. May 14 23:51:18.269468 ignition[960]: INFO : Ignition 2.20.0 May 14 23:51:18.269468 ignition[960]: INFO : Stage: files May 14 23:51:18.271804 ignition[960]: INFO : no configs at "/usr/lib/ignition/base.d" May 14 23:51:18.271804 ignition[960]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" May 14 23:51:18.271804 ignition[960]: DEBUG : files: compiled without relabeling support, skipping May 14 23:51:18.275810 ignition[960]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" May 14 23:51:18.275810 ignition[960]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" May 14 23:51:18.275810 ignition[960]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" May 14 23:51:18.275810 ignition[960]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" May 14 23:51:18.281778 ignition[960]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" May 14 23:51:18.281778 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" May 14 23:51:18.281778 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-amd64.tar.gz: attempt #1 May 14 23:51:18.276167 unknown[960]: wrote ssh authorized keys file for user: core May 14 23:51:18.437148 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK May 14 23:51:18.925532 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" May 14 23:51:18.927994 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" May 14 23:51:18.930093 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" May 14 23:51:18.931907 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" May 14 23:51:18.933657 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" May 14 23:51:18.935531 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" May 14 23:51:18.937393 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" May 14 23:51:18.939283 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" May 14 23:51:18.941130 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" May 14 23:51:18.943066 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" May 14 23:51:18.945138 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" May 14 23:51:18.947089 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" May 14 23:51:18.949783 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" May 14 23:51:18.952343 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" May 14 23:51:18.954655 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.30.1-x86-64.raw: attempt #1 May 14 23:51:19.086110 systemd-networkd[784]: eth0: Gained IPv6LL May 14 23:51:19.430126 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK May 14 23:51:19.761894 ignition[960]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" May 14 23:51:19.761894 ignition[960]: INFO : files: op(b): [started] processing unit "prepare-helm.service" May 14 23:51:19.765705 ignition[960]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" May 14 23:51:19.767969 ignition[960]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" May 14 23:51:19.767969 ignition[960]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" May 14 23:51:19.767969 ignition[960]: INFO : files: op(d): [started] processing unit "coreos-metadata.service" May 14 23:51:19.772367 ignition[960]: INFO : files: op(d): op(e): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" May 14 23:51:19.774347 ignition[960]: INFO : files: op(d): op(e): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" May 14 23:51:19.774347 ignition[960]: INFO : files: op(d): [finished] processing unit "coreos-metadata.service" May 14 23:51:19.777557 ignition[960]: INFO : files: op(f): [started] setting preset to disabled for "coreos-metadata.service" May 14 23:51:19.795586 ignition[960]: INFO : files: op(f): op(10): [started] removing enablement symlink(s) for "coreos-metadata.service" May 14 23:51:19.802084 ignition[960]: INFO : files: op(f): op(10): [finished] removing enablement symlink(s) for "coreos-metadata.service" May 14 23:51:19.803948 ignition[960]: INFO : files: op(f): [finished] setting preset to disabled for "coreos-metadata.service" May 14 23:51:19.803948 ignition[960]: INFO : files: op(11): [started] setting preset to enabled for "prepare-helm.service" May 14 23:51:19.807138 ignition[960]: INFO : files: op(11): [finished] setting preset to enabled for "prepare-helm.service" May 14 23:51:19.808861 ignition[960]: INFO : files: createResultFile: createFiles: op(12): [started] writing file "/sysroot/etc/.ignition-result.json" May 14 23:51:19.810936 ignition[960]: INFO : files: createResultFile: createFiles: op(12): [finished] writing file "/sysroot/etc/.ignition-result.json" May 14 23:51:19.812863 ignition[960]: INFO : files: files passed May 14 23:51:19.813710 ignition[960]: INFO : Ignition finished successfully May 14 23:51:19.817353 systemd[1]: Finished ignition-files.service - Ignition (files). May 14 23:51:19.819348 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... May 14 23:51:19.822466 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... May 14 23:51:19.850088 systemd[1]: ignition-quench.service: Deactivated successfully. May 14 23:51:19.850262 systemd[1]: Finished ignition-quench.service - Ignition (record completion). May 14 23:51:19.854695 initrd-setup-root-after-ignition[989]: grep: /sysroot/oem/oem-release: No such file or directory May 14 23:51:19.856215 initrd-setup-root-after-ignition[991]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory May 14 23:51:19.856215 initrd-setup-root-after-ignition[991]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory May 14 23:51:19.861036 initrd-setup-root-after-ignition[995]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory May 14 23:51:19.858107 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. May 14 23:51:19.861320 systemd[1]: Reached target ignition-complete.target - Ignition Complete. May 14 23:51:19.864679 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... May 14 23:51:19.935456 systemd[1]: initrd-parse-etc.service: Deactivated successfully. May 14 23:51:19.935608 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. May 14 23:51:19.938769 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. May 14 23:51:19.941536 systemd[1]: Reached target initrd.target - Initrd Default Target. May 14 23:51:19.942159 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. May 14 23:51:19.943312 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... May 14 23:51:19.974294 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. May 14 23:51:19.978671 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... May 14 23:51:20.010271 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. May 14 23:51:20.010667 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. May 14 23:51:20.013270 systemd[1]: Stopped target timers.target - Timer Units. May 14 23:51:20.013603 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. May 14 23:51:20.013771 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. May 14 23:51:20.017664 systemd[1]: Stopped target initrd.target - Initrd Default Target. May 14 23:51:20.018215 systemd[1]: Stopped target basic.target - Basic System. May 14 23:51:20.018562 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. May 14 23:51:20.019117 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. May 14 23:51:20.019456 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. May 14 23:51:20.020028 systemd[1]: Stopped target remote-fs.target - Remote File Systems. May 14 23:51:20.020371 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. May 14 23:51:20.020778 systemd[1]: Stopped target sysinit.target - System Initialization. May 14 23:51:20.021284 systemd[1]: Stopped target local-fs.target - Local File Systems. May 14 23:51:20.021642 systemd[1]: Stopped target swap.target - Swaps. May 14 23:51:20.040031 systemd[1]: dracut-pre-mount.service: Deactivated successfully. May 14 23:51:20.040978 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. May 14 23:51:20.043981 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. May 14 23:51:20.046182 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). May 14 23:51:20.046772 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. May 14 23:51:20.047088 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. May 14 23:51:20.049123 systemd[1]: dracut-initqueue.service: Deactivated successfully. May 14 23:51:20.049233 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. May 14 23:51:20.054204 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. May 14 23:51:20.054323 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). May 14 23:51:20.054710 systemd[1]: Stopped target paths.target - Path Units. May 14 23:51:20.058238 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. May 14 23:51:20.061839 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. May 14 23:51:20.064708 systemd[1]: Stopped target slices.target - Slice Units. May 14 23:51:20.066857 systemd[1]: Stopped target sockets.target - Socket Units. May 14 23:51:20.067369 systemd[1]: iscsid.socket: Deactivated successfully. May 14 23:51:20.067476 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. May 14 23:51:20.069164 systemd[1]: iscsiuio.socket: Deactivated successfully. May 14 23:51:20.069247 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. May 14 23:51:20.071099 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. May 14 23:51:20.071219 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. May 14 23:51:20.072812 systemd[1]: ignition-files.service: Deactivated successfully. May 14 23:51:20.072917 systemd[1]: Stopped ignition-files.service - Ignition (files). May 14 23:51:20.075818 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... May 14 23:51:20.076659 systemd[1]: kmod-static-nodes.service: Deactivated successfully. May 14 23:51:20.076803 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. May 14 23:51:20.079437 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... May 14 23:51:20.087019 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. May 14 23:51:20.088249 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. May 14 23:51:20.090891 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. May 14 23:51:20.092163 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. May 14 23:51:20.099480 systemd[1]: initrd-cleanup.service: Deactivated successfully. May 14 23:51:20.099661 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. May 14 23:51:20.111932 ignition[1016]: INFO : Ignition 2.20.0 May 14 23:51:20.111932 ignition[1016]: INFO : Stage: umount May 14 23:51:20.114066 ignition[1016]: INFO : no configs at "/usr/lib/ignition/base.d" May 14 23:51:20.114066 ignition[1016]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" May 14 23:51:20.114066 ignition[1016]: INFO : umount: umount passed May 14 23:51:20.114066 ignition[1016]: INFO : Ignition finished successfully May 14 23:51:20.120322 systemd[1]: ignition-mount.service: Deactivated successfully. May 14 23:51:20.120472 systemd[1]: Stopped ignition-mount.service - Ignition (mount). May 14 23:51:20.123499 systemd[1]: sysroot-boot.mount: Deactivated successfully. May 14 23:51:20.124082 systemd[1]: Stopped target network.target - Network. May 14 23:51:20.124484 systemd[1]: ignition-disks.service: Deactivated successfully. May 14 23:51:20.124543 systemd[1]: Stopped ignition-disks.service - Ignition (disks). May 14 23:51:20.125144 systemd[1]: ignition-kargs.service: Deactivated successfully. May 14 23:51:20.125202 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). May 14 23:51:20.125474 systemd[1]: ignition-setup.service: Deactivated successfully. May 14 23:51:20.125522 systemd[1]: Stopped ignition-setup.service - Ignition (setup). May 14 23:51:20.125989 systemd[1]: ignition-setup-pre.service: Deactivated successfully. May 14 23:51:20.126035 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. May 14 23:51:20.132799 systemd[1]: Stopping systemd-networkd.service - Network Configuration... May 14 23:51:20.134566 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... May 14 23:51:20.146519 systemd[1]: systemd-resolved.service: Deactivated successfully. May 14 23:51:20.146709 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. May 14 23:51:20.151665 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. May 14 23:51:20.152107 systemd[1]: systemd-networkd.service: Deactivated successfully. May 14 23:51:20.152278 systemd[1]: Stopped systemd-networkd.service - Network Configuration. May 14 23:51:20.156125 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. May 14 23:51:20.156948 systemd[1]: systemd-networkd.socket: Deactivated successfully. May 14 23:51:20.157017 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. May 14 23:51:20.160396 systemd[1]: Stopping network-cleanup.service - Network Cleanup... May 14 23:51:20.161053 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. May 14 23:51:20.161125 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. May 14 23:51:20.161460 systemd[1]: systemd-sysctl.service: Deactivated successfully. May 14 23:51:20.161522 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. May 14 23:51:20.167002 systemd[1]: systemd-modules-load.service: Deactivated successfully. May 14 23:51:20.167060 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. May 14 23:51:20.167394 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. May 14 23:51:20.167440 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. May 14 23:51:20.172124 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... May 14 23:51:20.173518 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. May 14 23:51:20.173587 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. May 14 23:51:20.186158 systemd[1]: network-cleanup.service: Deactivated successfully. May 14 23:51:20.186280 systemd[1]: Stopped network-cleanup.service - Network Cleanup. May 14 23:51:20.195473 systemd[1]: systemd-udevd.service: Deactivated successfully. May 14 23:51:20.195664 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. May 14 23:51:20.196359 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. May 14 23:51:20.196408 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. May 14 23:51:20.199242 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. May 14 23:51:20.199282 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. May 14 23:51:20.199543 systemd[1]: dracut-pre-udev.service: Deactivated successfully. May 14 23:51:20.199593 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. May 14 23:51:20.221627 systemd[1]: dracut-cmdline.service: Deactivated successfully. May 14 23:51:20.221702 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. May 14 23:51:20.224546 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. May 14 23:51:20.224604 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. May 14 23:51:20.228492 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... May 14 23:51:20.228996 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. May 14 23:51:20.229052 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. May 14 23:51:20.233045 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 14 23:51:20.233097 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. May 14 23:51:20.237194 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. May 14 23:51:20.237260 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. May 14 23:51:20.254003 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. May 14 23:51:20.254132 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. May 14 23:51:20.274786 systemd[1]: sysroot-boot.service: Deactivated successfully. May 14 23:51:20.274919 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. May 14 23:51:20.276012 systemd[1]: Reached target initrd-switch-root.target - Switch Root. May 14 23:51:20.277960 systemd[1]: initrd-setup-root.service: Deactivated successfully. May 14 23:51:20.278026 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. May 14 23:51:20.279154 systemd[1]: Starting initrd-switch-root.service - Switch Root... May 14 23:51:20.303434 systemd[1]: Switching root. May 14 23:51:20.341112 systemd-journald[193]: Journal stopped May 14 23:51:21.856383 systemd-journald[193]: Received SIGTERM from PID 1 (systemd). May 14 23:51:21.856484 kernel: SELinux: policy capability network_peer_controls=1 May 14 23:51:21.856508 kernel: SELinux: policy capability open_perms=1 May 14 23:51:21.856521 kernel: SELinux: policy capability extended_socket_class=1 May 14 23:51:21.856535 kernel: SELinux: policy capability always_check_network=0 May 14 23:51:21.856554 kernel: SELinux: policy capability cgroup_seclabel=1 May 14 23:51:21.856566 kernel: SELinux: policy capability nnp_nosuid_transition=1 May 14 23:51:21.856577 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 May 14 23:51:21.856608 kernel: SELinux: policy capability ioctl_skip_cloexec=0 May 14 23:51:21.856620 kernel: audit: type=1403 audit(1747266680.874:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 May 14 23:51:21.856634 systemd[1]: Successfully loaded SELinux policy in 42.626ms. May 14 23:51:21.856661 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 14.177ms. May 14 23:51:21.856675 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) May 14 23:51:21.856688 systemd[1]: Detected virtualization kvm. May 14 23:51:21.856700 systemd[1]: Detected architecture x86-64. May 14 23:51:21.856713 systemd[1]: Detected first boot. May 14 23:51:21.856744 systemd[1]: Initializing machine ID from VM UUID. May 14 23:51:21.856757 zram_generator::config[1063]: No configuration found. May 14 23:51:21.856782 kernel: Guest personality initialized and is inactive May 14 23:51:21.856794 kernel: VMCI host device registered (name=vmci, major=10, minor=125) May 14 23:51:21.856805 kernel: Initialized host personality May 14 23:51:21.856818 kernel: NET: Registered PF_VSOCK protocol family May 14 23:51:21.856830 systemd[1]: Populated /etc with preset unit settings. May 14 23:51:21.856843 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. May 14 23:51:21.856857 systemd[1]: initrd-switch-root.service: Deactivated successfully. May 14 23:51:21.856875 systemd[1]: Stopped initrd-switch-root.service - Switch Root. May 14 23:51:21.856887 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. May 14 23:51:21.856900 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. May 14 23:51:21.856913 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. May 14 23:51:21.856925 systemd[1]: Created slice system-getty.slice - Slice /system/getty. May 14 23:51:21.856938 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. May 14 23:51:21.856950 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. May 14 23:51:21.856966 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. May 14 23:51:21.856981 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. May 14 23:51:21.856994 systemd[1]: Created slice user.slice - User and Session Slice. May 14 23:51:21.857007 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. May 14 23:51:21.857020 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. May 14 23:51:21.857033 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. May 14 23:51:21.857045 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. May 14 23:51:21.857058 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. May 14 23:51:21.857071 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... May 14 23:51:21.857086 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... May 14 23:51:21.857099 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). May 14 23:51:21.857112 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. May 14 23:51:21.857124 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. May 14 23:51:21.857137 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. May 14 23:51:21.857152 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. May 14 23:51:21.857164 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. May 14 23:51:21.857177 systemd[1]: Reached target remote-fs.target - Remote File Systems. May 14 23:51:21.857189 systemd[1]: Reached target slices.target - Slice Units. May 14 23:51:21.857204 systemd[1]: Reached target swap.target - Swaps. May 14 23:51:21.857217 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. May 14 23:51:21.857230 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. May 14 23:51:21.857242 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. May 14 23:51:21.857255 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. May 14 23:51:21.857272 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. May 14 23:51:21.857284 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. May 14 23:51:21.857297 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. May 14 23:51:21.857310 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... May 14 23:51:21.857326 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... May 14 23:51:21.857338 systemd[1]: Mounting media.mount - External Media Directory... May 14 23:51:21.857351 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). May 14 23:51:21.857364 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... May 14 23:51:21.857376 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... May 14 23:51:21.857389 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... May 14 23:51:21.857402 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). May 14 23:51:21.857415 systemd[1]: Reached target machines.target - Containers. May 14 23:51:21.857429 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... May 14 23:51:21.857448 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 14 23:51:21.857461 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... May 14 23:51:21.857473 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... May 14 23:51:21.857485 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... May 14 23:51:21.857498 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... May 14 23:51:21.857510 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... May 14 23:51:21.857523 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... May 14 23:51:21.857535 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... May 14 23:51:21.857550 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). May 14 23:51:21.857563 systemd[1]: systemd-fsck-root.service: Deactivated successfully. May 14 23:51:21.857575 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. May 14 23:51:21.857588 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. May 14 23:51:21.857608 systemd[1]: Stopped systemd-fsck-usr.service. May 14 23:51:21.857622 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 14 23:51:21.857634 systemd[1]: Starting systemd-journald.service - Journal Service... May 14 23:51:21.857647 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... May 14 23:51:21.857659 kernel: fuse: init (API version 7.39) May 14 23:51:21.857674 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... May 14 23:51:21.857687 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... May 14 23:51:21.857700 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... May 14 23:51:21.857726 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... May 14 23:51:21.857742 kernel: loop: module loaded May 14 23:51:21.857755 systemd[1]: verity-setup.service: Deactivated successfully. May 14 23:51:21.857767 systemd[1]: Stopped verity-setup.service. May 14 23:51:21.857779 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). May 14 23:51:21.857813 systemd-journald[1138]: Collecting audit messages is disabled. May 14 23:51:21.857843 systemd-journald[1138]: Journal started May 14 23:51:21.857871 systemd-journald[1138]: Runtime Journal (/run/log/journal/0661f8a4e3b44fb5833446b10e4e8ddf) is 6M, max 48.2M, 42.2M free. May 14 23:51:21.866956 kernel: ACPI: bus type drm_connector registered May 14 23:51:21.593578 systemd[1]: Queued start job for default target multi-user.target. May 14 23:51:21.608711 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. May 14 23:51:21.609349 systemd[1]: systemd-journald.service: Deactivated successfully. May 14 23:51:21.870316 systemd[1]: Started systemd-journald.service - Journal Service. May 14 23:51:21.871228 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. May 14 23:51:21.872548 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. May 14 23:51:21.873917 systemd[1]: Mounted media.mount - External Media Directory. May 14 23:51:21.875105 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. May 14 23:51:21.876377 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. May 14 23:51:21.877699 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. May 14 23:51:21.879111 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. May 14 23:51:21.880743 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. May 14 23:51:21.882376 systemd[1]: modprobe@configfs.service: Deactivated successfully. May 14 23:51:21.882627 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. May 14 23:51:21.884223 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 14 23:51:21.884462 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. May 14 23:51:21.886150 systemd[1]: modprobe@drm.service: Deactivated successfully. May 14 23:51:21.886379 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. May 14 23:51:21.887829 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 14 23:51:21.888054 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. May 14 23:51:21.889916 systemd[1]: modprobe@fuse.service: Deactivated successfully. May 14 23:51:21.890144 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. May 14 23:51:21.891587 systemd[1]: modprobe@loop.service: Deactivated successfully. May 14 23:51:21.891841 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. May 14 23:51:21.893329 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. May 14 23:51:21.894861 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. May 14 23:51:21.896483 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. May 14 23:51:21.898333 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. May 14 23:51:21.915326 systemd[1]: Reached target network-pre.target - Preparation for Network. May 14 23:51:21.918374 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... May 14 23:51:21.921353 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... May 14 23:51:21.922788 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). May 14 23:51:21.922825 systemd[1]: Reached target local-fs.target - Local File Systems. May 14 23:51:21.925237 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. May 14 23:51:21.928254 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... May 14 23:51:21.933221 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... May 14 23:51:21.935085 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 14 23:51:21.939868 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... May 14 23:51:21.942631 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... May 14 23:51:21.944020 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 14 23:51:21.946875 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... May 14 23:51:21.948927 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. May 14 23:51:21.951019 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... May 14 23:51:21.955096 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... May 14 23:51:21.961187 systemd[1]: Starting systemd-sysusers.service - Create System Users... May 14 23:51:21.964614 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. May 14 23:51:21.969616 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. May 14 23:51:21.974898 systemd-journald[1138]: Time spent on flushing to /var/log/journal/0661f8a4e3b44fb5833446b10e4e8ddf is 14.440ms for 1056 entries. May 14 23:51:21.974898 systemd-journald[1138]: System Journal (/var/log/journal/0661f8a4e3b44fb5833446b10e4e8ddf) is 8M, max 195.6M, 187.6M free. May 14 23:51:22.008188 systemd-journald[1138]: Received client request to flush runtime journal. May 14 23:51:22.008285 kernel: loop0: detected capacity change from 0 to 210664 May 14 23:51:21.971632 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. May 14 23:51:21.978036 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. May 14 23:51:21.988835 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. May 14 23:51:21.992671 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... May 14 23:51:21.997324 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. May 14 23:51:22.004178 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. May 14 23:51:22.009597 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... May 14 23:51:22.011465 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. May 14 23:51:22.030216 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher May 14 23:51:22.036826 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. May 14 23:51:22.042159 systemd[1]: Finished systemd-sysusers.service - Create System Users. May 14 23:51:22.048045 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... May 14 23:51:22.049883 udevadm[1196]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. May 14 23:51:22.070746 kernel: loop1: detected capacity change from 0 to 109808 May 14 23:51:22.081424 systemd-tmpfiles[1202]: ACLs are not supported, ignoring. May 14 23:51:22.081446 systemd-tmpfiles[1202]: ACLs are not supported, ignoring. May 14 23:51:22.090170 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. May 14 23:51:22.112758 kernel: loop2: detected capacity change from 0 to 151640 May 14 23:51:22.152750 kernel: loop3: detected capacity change from 0 to 210664 May 14 23:51:22.164963 kernel: loop4: detected capacity change from 0 to 109808 May 14 23:51:22.179777 kernel: loop5: detected capacity change from 0 to 151640 May 14 23:51:22.196473 (sd-merge)[1207]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. May 14 23:51:22.197174 (sd-merge)[1207]: Merged extensions into '/usr'. May 14 23:51:22.204571 systemd[1]: Reload requested from client PID 1183 ('systemd-sysext') (unit systemd-sysext.service)... May 14 23:51:22.204598 systemd[1]: Reloading... May 14 23:51:22.274394 zram_generator::config[1238]: No configuration found. May 14 23:51:22.330660 ldconfig[1178]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. May 14 23:51:22.398863 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 14 23:51:22.469600 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. May 14 23:51:22.470254 systemd[1]: Reloading finished in 265 ms. May 14 23:51:22.490826 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. May 14 23:51:22.492456 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. May 14 23:51:22.508489 systemd[1]: Starting ensure-sysext.service... May 14 23:51:22.510752 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... May 14 23:51:22.522881 systemd[1]: Reload requested from client PID 1272 ('systemctl') (unit ensure-sysext.service)... May 14 23:51:22.522898 systemd[1]: Reloading... May 14 23:51:22.541845 systemd-tmpfiles[1274]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. May 14 23:51:22.542286 systemd-tmpfiles[1274]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. May 14 23:51:22.545064 systemd-tmpfiles[1274]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. May 14 23:51:22.545572 systemd-tmpfiles[1274]: ACLs are not supported, ignoring. May 14 23:51:22.545890 systemd-tmpfiles[1274]: ACLs are not supported, ignoring. May 14 23:51:22.551541 systemd-tmpfiles[1274]: Detected autofs mount point /boot during canonicalization of boot. May 14 23:51:22.554760 systemd-tmpfiles[1274]: Skipping /boot May 14 23:51:22.577126 systemd-tmpfiles[1274]: Detected autofs mount point /boot during canonicalization of boot. May 14 23:51:22.577150 systemd-tmpfiles[1274]: Skipping /boot May 14 23:51:22.584949 zram_generator::config[1309]: No configuration found. May 14 23:51:22.764909 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 14 23:51:22.846517 systemd[1]: Reloading finished in 323 ms. May 14 23:51:22.859981 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. May 14 23:51:22.878917 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. May 14 23:51:22.889766 systemd[1]: Starting audit-rules.service - Load Audit Rules... May 14 23:51:22.892762 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... May 14 23:51:22.900821 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... May 14 23:51:22.905488 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... May 14 23:51:22.910410 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... May 14 23:51:22.916094 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... May 14 23:51:22.929083 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). May 14 23:51:22.929295 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 14 23:51:22.931996 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... May 14 23:51:22.936008 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... May 14 23:51:22.939383 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... May 14 23:51:22.941911 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 14 23:51:22.942036 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 14 23:51:22.946856 systemd[1]: Starting systemd-userdbd.service - User Database Manager... May 14 23:51:22.948476 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). May 14 23:51:22.950384 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. May 14 23:51:22.955698 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 14 23:51:22.955971 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. May 14 23:51:22.959382 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 14 23:51:22.960214 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. May 14 23:51:22.962104 systemd[1]: modprobe@loop.service: Deactivated successfully. May 14 23:51:22.962331 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. May 14 23:51:22.975015 augenrules[1375]: No rules May 14 23:51:22.979315 systemd[1]: audit-rules.service: Deactivated successfully. May 14 23:51:22.980526 systemd[1]: Finished audit-rules.service - Load Audit Rules. May 14 23:51:22.980637 systemd-udevd[1349]: Using default interface naming scheme 'v255'. May 14 23:51:22.982430 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. May 14 23:51:22.988109 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. May 14 23:51:22.994634 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). May 14 23:51:22.998995 systemd[1]: Starting audit-rules.service - Load Audit Rules... May 14 23:51:23.000123 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 14 23:51:23.002011 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... May 14 23:51:23.015654 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... May 14 23:51:23.021177 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... May 14 23:51:23.025327 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... May 14 23:51:23.026518 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 14 23:51:23.026644 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 14 23:51:23.028854 systemd[1]: Starting systemd-update-done.service - Update is Completed... May 14 23:51:23.030701 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). May 14 23:51:23.030823 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). May 14 23:51:23.042744 systemd[1]: Started systemd-userdbd.service - User Database Manager. May 14 23:51:23.044436 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. May 14 23:51:23.054190 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 14 23:51:23.054516 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. May 14 23:51:23.056654 systemd[1]: modprobe@drm.service: Deactivated successfully. May 14 23:51:23.057984 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. May 14 23:51:23.059840 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 14 23:51:23.060102 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. May 14 23:51:23.061203 augenrules[1386]: /sbin/augenrules: No change May 14 23:51:23.061851 systemd[1]: modprobe@loop.service: Deactivated successfully. May 14 23:51:23.062098 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. May 14 23:51:23.063842 systemd[1]: Finished systemd-update-done.service - Update is Completed. May 14 23:51:23.069652 systemd[1]: Finished ensure-sysext.service. May 14 23:51:23.086543 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. May 14 23:51:23.090865 systemd[1]: Starting systemd-networkd.service - Network Configuration... May 14 23:51:23.092239 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 14 23:51:23.092323 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. May 14 23:51:23.094677 augenrules[1435]: No rules May 14 23:51:23.095993 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... May 14 23:51:23.098287 systemd[1]: audit-rules.service: Deactivated successfully. May 14 23:51:23.098798 systemd[1]: Finished audit-rules.service - Load Audit Rules. May 14 23:51:23.123766 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 41 scanned by (udev-worker) (1409) May 14 23:51:23.139890 systemd-resolved[1346]: Positive Trust Anchors: May 14 23:51:23.139909 systemd-resolved[1346]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d May 14 23:51:23.139941 systemd-resolved[1346]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test May 14 23:51:23.143547 systemd-resolved[1346]: Defaulting to hostname 'linux'. May 14 23:51:23.162211 systemd[1]: Started systemd-resolved.service - Network Name Resolution. May 14 23:51:23.177662 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. May 14 23:51:23.191749 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input2 May 14 23:51:23.194790 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. May 14 23:51:23.196341 systemd[1]: Reached target time-set.target - System Time Set. May 14 23:51:23.198789 kernel: ACPI: button: Power Button [PWRF] May 14 23:51:23.201747 systemd-networkd[1437]: lo: Link UP May 14 23:51:23.201760 systemd-networkd[1437]: lo: Gained carrier May 14 23:51:23.205794 systemd-networkd[1437]: Enumeration completed May 14 23:51:23.206261 systemd-networkd[1437]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 14 23:51:23.206269 systemd-networkd[1437]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. May 14 23:51:23.206969 systemd-networkd[1437]: eth0: Link UP May 14 23:51:23.206977 systemd-networkd[1437]: eth0: Gained carrier May 14 23:51:23.206991 systemd-networkd[1437]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 14 23:51:23.208021 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. May 14 23:51:23.209506 systemd[1]: Started systemd-networkd.service - Network Configuration. May 14 23:51:23.215145 kernel: i801_smbus 0000:00:1f.3: Enabling SMBus device May 14 23:51:23.215459 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt May 14 23:51:23.216697 kernel: i2c i2c-0: 1/1 memory slots populated (from DMI) May 14 23:51:23.216965 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD May 14 23:51:23.215066 systemd[1]: Reached target network.target - Network. May 14 23:51:23.220033 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... May 14 23:51:23.224337 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input3 May 14 23:51:23.222814 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... May 14 23:51:23.223255 systemd-networkd[1437]: eth0: DHCPv4 address 10.0.0.142/16, gateway 10.0.0.1 acquired from 10.0.0.1 May 14 23:51:23.224395 systemd-timesyncd[1439]: Network configuration changed, trying to establish connection. May 14 23:51:24.255083 systemd-resolved[1346]: Clock change detected. Flushing caches. May 14 23:51:24.255130 systemd-timesyncd[1439]: Contacted time server 10.0.0.1:123 (10.0.0.1). May 14 23:51:24.255183 systemd-timesyncd[1439]: Initial clock synchronization to Wed 2025-05-14 23:51:24.253611 UTC. May 14 23:51:24.255463 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... May 14 23:51:24.282078 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. May 14 23:51:24.289000 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. May 14 23:51:24.311540 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 14 23:51:24.357326 kernel: mousedev: PS/2 mouse device common for all mice May 14 23:51:24.362099 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 14 23:51:24.362477 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. May 14 23:51:24.365720 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 14 23:51:24.371391 kernel: kvm_amd: TSC scaling supported May 14 23:51:24.371426 kernel: kvm_amd: Nested Virtualization enabled May 14 23:51:24.371440 kernel: kvm_amd: Nested Paging enabled May 14 23:51:24.371452 kernel: kvm_amd: LBR virtualization supported May 14 23:51:24.372470 kernel: kvm_amd: Virtual VMLOAD VMSAVE supported May 14 23:51:24.372488 kernel: kvm_amd: Virtual GIF supported May 14 23:51:24.392366 kernel: EDAC MC: Ver: 3.0.0 May 14 23:51:24.420629 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. May 14 23:51:24.422643 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... May 14 23:51:24.440742 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. May 14 23:51:24.449659 lvm[1471]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. May 14 23:51:24.486916 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. May 14 23:51:24.488608 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. May 14 23:51:24.489760 systemd[1]: Reached target sysinit.target - System Initialization. May 14 23:51:24.490969 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. May 14 23:51:24.492283 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. May 14 23:51:24.493799 systemd[1]: Started logrotate.timer - Daily rotation of log files. May 14 23:51:24.495086 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. May 14 23:51:24.496498 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. May 14 23:51:24.497818 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). May 14 23:51:24.497842 systemd[1]: Reached target paths.target - Path Units. May 14 23:51:24.498783 systemd[1]: Reached target timers.target - Timer Units. May 14 23:51:24.500652 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. May 14 23:51:24.503650 systemd[1]: Starting docker.socket - Docker Socket for the API... May 14 23:51:24.507377 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). May 14 23:51:24.508850 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). May 14 23:51:24.510156 systemd[1]: Reached target ssh-access.target - SSH Access Available. May 14 23:51:24.514259 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. May 14 23:51:24.516086 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. May 14 23:51:24.519188 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... May 14 23:51:24.521119 systemd[1]: Listening on docker.socket - Docker Socket for the API. May 14 23:51:24.522437 systemd[1]: Reached target sockets.target - Socket Units. May 14 23:51:24.523489 systemd[1]: Reached target basic.target - Basic System. May 14 23:51:24.524567 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. May 14 23:51:24.524613 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. May 14 23:51:24.525980 systemd[1]: Starting containerd.service - containerd container runtime... May 14 23:51:24.528514 systemd[1]: Starting dbus.service - D-Bus System Message Bus... May 14 23:51:24.531370 lvm[1477]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. May 14 23:51:24.533398 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... May 14 23:51:24.535983 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... May 14 23:51:24.537108 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). May 14 23:51:24.540670 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... May 14 23:51:24.542600 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... May 14 23:51:24.545256 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... May 14 23:51:24.549830 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... May 14 23:51:24.551565 jq[1480]: false May 14 23:51:24.557516 systemd[1]: Starting systemd-logind.service - User Login Management... May 14 23:51:24.559274 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). May 14 23:51:24.564687 extend-filesystems[1481]: Found loop3 May 14 23:51:24.565022 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. May 14 23:51:24.580955 extend-filesystems[1481]: Found loop4 May 14 23:51:24.580955 extend-filesystems[1481]: Found loop5 May 14 23:51:24.580955 extend-filesystems[1481]: Found sr0 May 14 23:51:24.580955 extend-filesystems[1481]: Found vda May 14 23:51:24.580955 extend-filesystems[1481]: Found vda1 May 14 23:51:24.580955 extend-filesystems[1481]: Found vda2 May 14 23:51:24.580955 extend-filesystems[1481]: Found vda3 May 14 23:51:24.580955 extend-filesystems[1481]: Found usr May 14 23:51:24.580955 extend-filesystems[1481]: Found vda4 May 14 23:51:24.580955 extend-filesystems[1481]: Found vda6 May 14 23:51:24.580955 extend-filesystems[1481]: Found vda7 May 14 23:51:24.580955 extend-filesystems[1481]: Found vda9 May 14 23:51:24.580955 extend-filesystems[1481]: Checking size of /dev/vda9 May 14 23:51:24.579652 systemd[1]: Starting update-engine.service - Update Engine... May 14 23:51:24.589811 dbus-daemon[1479]: [system] SELinux support is enabled May 14 23:51:24.618674 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 41 scanned by (udev-worker) (1418) May 14 23:51:24.582218 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... May 14 23:51:24.584278 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. May 14 23:51:24.619005 jq[1497]: true May 14 23:51:24.586433 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. May 14 23:51:24.586754 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. May 14 23:51:24.587248 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. May 14 23:51:24.587757 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. May 14 23:51:24.590193 systemd[1]: Started dbus.service - D-Bus System Message Bus. May 14 23:51:24.603867 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). May 14 23:51:24.603953 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. May 14 23:51:24.609790 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). May 14 23:51:24.609817 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. May 14 23:51:24.618898 systemd[1]: motdgen.service: Deactivated successfully. May 14 23:51:24.619485 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. May 14 23:51:24.632931 extend-filesystems[1481]: Resized partition /dev/vda9 May 14 23:51:24.637434 extend-filesystems[1511]: resize2fs 1.47.2 (1-Jan-2025) May 14 23:51:24.642958 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks May 14 23:51:24.642986 update_engine[1492]: I20250514 23:51:24.637125 1492 main.cc:92] Flatcar Update Engine starting May 14 23:51:24.643211 jq[1503]: true May 14 23:51:24.649793 update_engine[1492]: I20250514 23:51:24.648779 1492 update_check_scheduler.cc:74] Next update check in 10m48s May 14 23:51:24.648654 (ntainerd)[1510]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR May 14 23:51:24.652998 tar[1502]: linux-amd64/helm May 14 23:51:24.666570 systemd[1]: Started update-engine.service - Update Engine. May 14 23:51:24.671051 systemd[1]: Started locksmithd.service - Cluster reboot manager. May 14 23:51:24.681470 kernel: EXT4-fs (vda9): resized filesystem to 1864699 May 14 23:51:24.711690 extend-filesystems[1511]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required May 14 23:51:24.711690 extend-filesystems[1511]: old_desc_blocks = 1, new_desc_blocks = 1 May 14 23:51:24.711690 extend-filesystems[1511]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. May 14 23:51:24.722161 extend-filesystems[1481]: Resized filesystem in /dev/vda9 May 14 23:51:24.714826 systemd-logind[1487]: Watching system buttons on /dev/input/event1 (Power Button) May 14 23:51:24.714855 systemd-logind[1487]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) May 14 23:51:24.715146 systemd[1]: extend-filesystems.service: Deactivated successfully. May 14 23:51:24.715474 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. May 14 23:51:24.715886 systemd-logind[1487]: New seat seat0. May 14 23:51:24.720506 systemd[1]: Started systemd-logind.service - User Login Management. May 14 23:51:24.730692 locksmithd[1531]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" May 14 23:51:24.847390 bash[1533]: Updated "/home/core/.ssh/authorized_keys" May 14 23:51:24.849309 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. May 14 23:51:24.853688 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. May 14 23:51:24.903674 containerd[1510]: time="2025-05-14T23:51:24Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 May 14 23:51:24.904619 containerd[1510]: time="2025-05-14T23:51:24.904561514Z" level=info msg="starting containerd" revision=88aa2f531d6c2922003cc7929e51daf1c14caa0a version=v2.0.1 May 14 23:51:24.914558 containerd[1510]: time="2025-05-14T23:51:24.914514793Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="6.522µs" May 14 23:51:24.914558 containerd[1510]: time="2025-05-14T23:51:24.914543697Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 May 14 23:51:24.914558 containerd[1510]: time="2025-05-14T23:51:24.914561811Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 May 14 23:51:24.914765 containerd[1510]: time="2025-05-14T23:51:24.914737641Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 May 14 23:51:24.914765 containerd[1510]: time="2025-05-14T23:51:24.914758230Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 May 14 23:51:24.914817 containerd[1510]: time="2025-05-14T23:51:24.914781894Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 May 14 23:51:24.914870 containerd[1510]: time="2025-05-14T23:51:24.914850353Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 May 14 23:51:24.914870 containerd[1510]: time="2025-05-14T23:51:24.914865291Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 May 14 23:51:24.915147 containerd[1510]: time="2025-05-14T23:51:24.915112534Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 May 14 23:51:24.915147 containerd[1510]: time="2025-05-14T23:51:24.915138944Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 May 14 23:51:24.915190 containerd[1510]: time="2025-05-14T23:51:24.915149664Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 May 14 23:51:24.915190 containerd[1510]: time="2025-05-14T23:51:24.915159242Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 May 14 23:51:24.915285 containerd[1510]: time="2025-05-14T23:51:24.915257817Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 May 14 23:51:24.915611 containerd[1510]: time="2025-05-14T23:51:24.915579330Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 May 14 23:51:24.915640 containerd[1510]: time="2025-05-14T23:51:24.915621018Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 May 14 23:51:24.915640 containerd[1510]: time="2025-05-14T23:51:24.915631217Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 May 14 23:51:24.915679 containerd[1510]: time="2025-05-14T23:51:24.915662947Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 May 14 23:51:24.916109 containerd[1510]: time="2025-05-14T23:51:24.915905101Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 May 14 23:51:24.916109 containerd[1510]: time="2025-05-14T23:51:24.916043701Z" level=info msg="metadata content store policy set" policy=shared May 14 23:51:24.922210 sshd_keygen[1501]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 May 14 23:51:24.924811 containerd[1510]: time="2025-05-14T23:51:24.923140372Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 May 14 23:51:24.924811 containerd[1510]: time="2025-05-14T23:51:24.923199894Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 May 14 23:51:24.924811 containerd[1510]: time="2025-05-14T23:51:24.923216856Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 May 14 23:51:24.924811 containerd[1510]: time="2025-05-14T23:51:24.923230431Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 May 14 23:51:24.924811 containerd[1510]: time="2025-05-14T23:51:24.923242354Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 May 14 23:51:24.924811 containerd[1510]: time="2025-05-14T23:51:24.923252953Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 May 14 23:51:24.924811 containerd[1510]: time="2025-05-14T23:51:24.923264485Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 May 14 23:51:24.924811 containerd[1510]: time="2025-05-14T23:51:24.923277059Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 May 14 23:51:24.924811 containerd[1510]: time="2025-05-14T23:51:24.923288961Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 May 14 23:51:24.924811 containerd[1510]: time="2025-05-14T23:51:24.923313487Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 May 14 23:51:24.924811 containerd[1510]: time="2025-05-14T23:51:24.923323115Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 May 14 23:51:24.924811 containerd[1510]: time="2025-05-14T23:51:24.923335298Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 May 14 23:51:24.924811 containerd[1510]: time="2025-05-14T23:51:24.923489477Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 May 14 23:51:24.924811 containerd[1510]: time="2025-05-14T23:51:24.923510897Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 May 14 23:51:24.925087 containerd[1510]: time="2025-05-14T23:51:24.923532217Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 May 14 23:51:24.925087 containerd[1510]: time="2025-05-14T23:51:24.923543769Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 May 14 23:51:24.925087 containerd[1510]: time="2025-05-14T23:51:24.923555230Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 May 14 23:51:24.925087 containerd[1510]: time="2025-05-14T23:51:24.923564999Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 May 14 23:51:24.925087 containerd[1510]: time="2025-05-14T23:51:24.923577162Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 May 14 23:51:24.925087 containerd[1510]: time="2025-05-14T23:51:24.923587852Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 May 14 23:51:24.925087 containerd[1510]: time="2025-05-14T23:51:24.923599584Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 May 14 23:51:24.925087 containerd[1510]: time="2025-05-14T23:51:24.923610735Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 May 14 23:51:24.925087 containerd[1510]: time="2025-05-14T23:51:24.923620653Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 May 14 23:51:24.925087 containerd[1510]: time="2025-05-14T23:51:24.923681217Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" May 14 23:51:24.925087 containerd[1510]: time="2025-05-14T23:51:24.923713457Z" level=info msg="Start snapshots syncer" May 14 23:51:24.925087 containerd[1510]: time="2025-05-14T23:51:24.923734737Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 May 14 23:51:24.925437 containerd[1510]: time="2025-05-14T23:51:24.923953147Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" May 14 23:51:24.925437 containerd[1510]: time="2025-05-14T23:51:24.923997781Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 May 14 23:51:24.925557 containerd[1510]: time="2025-05-14T23:51:24.924065658Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 May 14 23:51:24.925557 containerd[1510]: time="2025-05-14T23:51:24.924176145Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 May 14 23:51:24.925557 containerd[1510]: time="2025-05-14T23:51:24.924194940Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 May 14 23:51:24.925557 containerd[1510]: time="2025-05-14T23:51:24.924207334Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 May 14 23:51:24.925557 containerd[1510]: time="2025-05-14T23:51:24.924218234Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 May 14 23:51:24.925557 containerd[1510]: time="2025-05-14T23:51:24.924229165Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 May 14 23:51:24.925557 containerd[1510]: time="2025-05-14T23:51:24.924239013Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 May 14 23:51:24.925557 containerd[1510]: time="2025-05-14T23:51:24.924256346Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 May 14 23:51:24.925557 containerd[1510]: time="2025-05-14T23:51:24.924278597Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 May 14 23:51:24.925557 containerd[1510]: time="2025-05-14T23:51:24.924289718Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 May 14 23:51:24.925557 containerd[1510]: time="2025-05-14T23:51:24.924317270Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 May 14 23:51:24.925557 containerd[1510]: time="2025-05-14T23:51:24.924342838Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 May 14 23:51:24.925557 containerd[1510]: time="2025-05-14T23:51:24.924355171Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 May 14 23:51:24.925557 containerd[1510]: time="2025-05-14T23:51:24.924367965Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 May 14 23:51:24.925808 containerd[1510]: time="2025-05-14T23:51:24.924380569Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 May 14 23:51:24.925808 containerd[1510]: time="2025-05-14T23:51:24.924390567Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 May 14 23:51:24.925808 containerd[1510]: time="2025-05-14T23:51:24.924415945Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 May 14 23:51:24.925808 containerd[1510]: time="2025-05-14T23:51:24.924429550Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 May 14 23:51:24.925808 containerd[1510]: time="2025-05-14T23:51:24.924441433Z" level=info msg="runtime interface created" May 14 23:51:24.925808 containerd[1510]: time="2025-05-14T23:51:24.924446602Z" level=info msg="created NRI interface" May 14 23:51:24.925808 containerd[1510]: time="2025-05-14T23:51:24.924454607Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 May 14 23:51:24.925808 containerd[1510]: time="2025-05-14T23:51:24.924466199Z" level=info msg="Connect containerd service" May 14 23:51:24.925808 containerd[1510]: time="2025-05-14T23:51:24.924496626Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" May 14 23:51:24.927475 containerd[1510]: time="2025-05-14T23:51:24.927397558Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" May 14 23:51:24.949330 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. May 14 23:51:24.953289 systemd[1]: Starting issuegen.service - Generate /run/issue... May 14 23:51:24.979481 systemd[1]: issuegen.service: Deactivated successfully. May 14 23:51:24.979772 systemd[1]: Finished issuegen.service - Generate /run/issue. May 14 23:51:24.998590 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... May 14 23:51:25.021681 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. May 14 23:51:25.026366 systemd[1]: Started getty@tty1.service - Getty on tty1. May 14 23:51:25.031236 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. May 14 23:51:25.032676 systemd[1]: Reached target getty.target - Login Prompts. May 14 23:51:25.033735 containerd[1510]: time="2025-05-14T23:51:25.033687297Z" level=info msg="Start subscribing containerd event" May 14 23:51:25.033874 containerd[1510]: time="2025-05-14T23:51:25.033838911Z" level=info msg="Start recovering state" May 14 23:51:25.034025 containerd[1510]: time="2025-05-14T23:51:25.034010122Z" level=info msg="Start event monitor" May 14 23:51:25.034176 containerd[1510]: time="2025-05-14T23:51:25.034157559Z" level=info msg="Start cni network conf syncer for default" May 14 23:51:25.035456 containerd[1510]: time="2025-05-14T23:51:25.035384681Z" level=info msg="Start streaming server" May 14 23:51:25.035493 containerd[1510]: time="2025-05-14T23:51:25.033756126Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc May 14 23:51:25.035555 containerd[1510]: time="2025-05-14T23:51:25.035531827Z" level=info msg=serving... address=/run/containerd/containerd.sock May 14 23:51:25.035555 containerd[1510]: time="2025-05-14T23:51:25.035465933Z" level=info msg="Registered namespace \"k8s.io\" with NRI" May 14 23:51:25.035599 containerd[1510]: time="2025-05-14T23:51:25.035563516Z" level=info msg="runtime interface starting up..." May 14 23:51:25.035599 containerd[1510]: time="2025-05-14T23:51:25.035570399Z" level=info msg="starting plugins..." May 14 23:51:25.035599 containerd[1510]: time="2025-05-14T23:51:25.035592821Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" May 14 23:51:25.035763 systemd[1]: Started containerd.service - containerd container runtime. May 14 23:51:25.045944 containerd[1510]: time="2025-05-14T23:51:25.045914892Z" level=info msg="containerd successfully booted in 0.143217s" May 14 23:51:25.115883 tar[1502]: linux-amd64/LICENSE May 14 23:51:25.115883 tar[1502]: linux-amd64/README.md May 14 23:51:25.137571 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. May 14 23:51:26.002723 systemd-networkd[1437]: eth0: Gained IPv6LL May 14 23:51:26.008863 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. May 14 23:51:26.011907 systemd[1]: Reached target network-online.target - Network is Online. May 14 23:51:26.016349 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... May 14 23:51:26.044562 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 14 23:51:26.047968 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... May 14 23:51:26.104348 systemd[1]: coreos-metadata.service: Deactivated successfully. May 14 23:51:26.104731 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. May 14 23:51:26.107784 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. May 14 23:51:26.110098 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. May 14 23:51:27.849562 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 14 23:51:27.877899 (kubelet)[1605]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 14 23:51:27.878660 systemd[1]: Reached target multi-user.target - Multi-User System. May 14 23:51:27.889238 systemd[1]: Startup finished in 1.383s (kernel) + 6.151s (initrd) + 6.027s (userspace) = 13.561s. May 14 23:51:28.597128 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. May 14 23:51:28.601535 systemd[1]: Started sshd@0-10.0.0.142:22-10.0.0.1:56948.service - OpenSSH per-connection server daemon (10.0.0.1:56948). May 14 23:51:28.676944 sshd[1617]: Accepted publickey for core from 10.0.0.1 port 56948 ssh2: RSA SHA256:lk2TkYBEL43KPVbrGyh3Ro8IB8NGN6uTNXzFyrYR01I May 14 23:51:28.679347 sshd-session[1617]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 23:51:28.690903 systemd-logind[1487]: New session 1 of user core. May 14 23:51:28.692349 systemd[1]: Created slice user-500.slice - User Slice of UID 500. May 14 23:51:28.693680 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... May 14 23:51:28.734078 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. May 14 23:51:28.739134 systemd[1]: Starting user@500.service - User Manager for UID 500... May 14 23:51:28.757447 (systemd)[1622]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) May 14 23:51:28.759976 systemd-logind[1487]: New session c1 of user core. May 14 23:51:28.844911 kubelet[1605]: E0514 23:51:28.844831 1605 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 14 23:51:28.850002 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 14 23:51:28.850346 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 14 23:51:28.850878 systemd[1]: kubelet.service: Consumed 1.559s CPU time, 247.5M memory peak. May 14 23:51:28.955571 systemd[1622]: Queued start job for default target default.target. May 14 23:51:28.972586 systemd[1622]: Created slice app.slice - User Application Slice. May 14 23:51:28.972629 systemd[1622]: Reached target paths.target - Paths. May 14 23:51:28.972685 systemd[1622]: Reached target timers.target - Timers. May 14 23:51:28.975485 systemd[1622]: Starting dbus.socket - D-Bus User Message Bus Socket... May 14 23:51:28.988573 systemd[1622]: Listening on dbus.socket - D-Bus User Message Bus Socket. May 14 23:51:28.988777 systemd[1622]: Reached target sockets.target - Sockets. May 14 23:51:28.988848 systemd[1622]: Reached target basic.target - Basic System. May 14 23:51:28.988911 systemd[1622]: Reached target default.target - Main User Target. May 14 23:51:28.988958 systemd[1622]: Startup finished in 220ms. May 14 23:51:28.989129 systemd[1]: Started user@500.service - User Manager for UID 500. May 14 23:51:28.991219 systemd[1]: Started session-1.scope - Session 1 of User core. May 14 23:51:29.061761 systemd[1]: Started sshd@1-10.0.0.142:22-10.0.0.1:56954.service - OpenSSH per-connection server daemon (10.0.0.1:56954). May 14 23:51:29.110479 sshd[1635]: Accepted publickey for core from 10.0.0.1 port 56954 ssh2: RSA SHA256:lk2TkYBEL43KPVbrGyh3Ro8IB8NGN6uTNXzFyrYR01I May 14 23:51:29.112071 sshd-session[1635]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 23:51:29.117281 systemd-logind[1487]: New session 2 of user core. May 14 23:51:29.142762 systemd[1]: Started session-2.scope - Session 2 of User core. May 14 23:51:29.199769 sshd[1637]: Connection closed by 10.0.0.1 port 56954 May 14 23:51:29.200236 sshd-session[1635]: pam_unix(sshd:session): session closed for user core May 14 23:51:29.220658 systemd[1]: sshd@1-10.0.0.142:22-10.0.0.1:56954.service: Deactivated successfully. May 14 23:51:29.223038 systemd[1]: session-2.scope: Deactivated successfully. May 14 23:51:29.225462 systemd-logind[1487]: Session 2 logged out. Waiting for processes to exit. May 14 23:51:29.227363 systemd[1]: Started sshd@2-10.0.0.142:22-10.0.0.1:56964.service - OpenSSH per-connection server daemon (10.0.0.1:56964). May 14 23:51:29.228828 systemd-logind[1487]: Removed session 2. May 14 23:51:29.273871 sshd[1642]: Accepted publickey for core from 10.0.0.1 port 56964 ssh2: RSA SHA256:lk2TkYBEL43KPVbrGyh3Ro8IB8NGN6uTNXzFyrYR01I May 14 23:51:29.275610 sshd-session[1642]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 23:51:29.281031 systemd-logind[1487]: New session 3 of user core. May 14 23:51:29.290639 systemd[1]: Started session-3.scope - Session 3 of User core. May 14 23:51:29.344598 sshd[1645]: Connection closed by 10.0.0.1 port 56964 May 14 23:51:29.345212 sshd-session[1642]: pam_unix(sshd:session): session closed for user core May 14 23:51:29.355221 systemd[1]: sshd@2-10.0.0.142:22-10.0.0.1:56964.service: Deactivated successfully. May 14 23:51:29.357721 systemd[1]: session-3.scope: Deactivated successfully. May 14 23:51:29.359367 systemd-logind[1487]: Session 3 logged out. Waiting for processes to exit. May 14 23:51:29.361178 systemd[1]: Started sshd@3-10.0.0.142:22-10.0.0.1:56968.service - OpenSSH per-connection server daemon (10.0.0.1:56968). May 14 23:51:29.362353 systemd-logind[1487]: Removed session 3. May 14 23:51:29.420512 sshd[1650]: Accepted publickey for core from 10.0.0.1 port 56968 ssh2: RSA SHA256:lk2TkYBEL43KPVbrGyh3Ro8IB8NGN6uTNXzFyrYR01I May 14 23:51:29.422559 sshd-session[1650]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 23:51:29.427941 systemd-logind[1487]: New session 4 of user core. May 14 23:51:29.437529 systemd[1]: Started session-4.scope - Session 4 of User core. May 14 23:51:29.497646 sshd[1653]: Connection closed by 10.0.0.1 port 56968 May 14 23:51:29.498026 sshd-session[1650]: pam_unix(sshd:session): session closed for user core May 14 23:51:29.512397 systemd[1]: sshd@3-10.0.0.142:22-10.0.0.1:56968.service: Deactivated successfully. May 14 23:51:29.514401 systemd[1]: session-4.scope: Deactivated successfully. May 14 23:51:29.516152 systemd-logind[1487]: Session 4 logged out. Waiting for processes to exit. May 14 23:51:29.517602 systemd[1]: Started sshd@4-10.0.0.142:22-10.0.0.1:56978.service - OpenSSH per-connection server daemon (10.0.0.1:56978). May 14 23:51:29.518355 systemd-logind[1487]: Removed session 4. May 14 23:51:29.570525 sshd[1658]: Accepted publickey for core from 10.0.0.1 port 56978 ssh2: RSA SHA256:lk2TkYBEL43KPVbrGyh3Ro8IB8NGN6uTNXzFyrYR01I May 14 23:51:29.572279 sshd-session[1658]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 23:51:29.579515 systemd-logind[1487]: New session 5 of user core. May 14 23:51:29.589594 systemd[1]: Started session-5.scope - Session 5 of User core. May 14 23:51:29.655987 sudo[1662]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 May 14 23:51:29.656455 sudo[1662]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 14 23:51:29.674912 sudo[1662]: pam_unix(sudo:session): session closed for user root May 14 23:51:29.676469 sshd[1661]: Connection closed by 10.0.0.1 port 56978 May 14 23:51:29.676937 sshd-session[1658]: pam_unix(sshd:session): session closed for user core May 14 23:51:29.697841 systemd[1]: sshd@4-10.0.0.142:22-10.0.0.1:56978.service: Deactivated successfully. May 14 23:51:29.699825 systemd[1]: session-5.scope: Deactivated successfully. May 14 23:51:29.701891 systemd-logind[1487]: Session 5 logged out. Waiting for processes to exit. May 14 23:51:29.703205 systemd[1]: Started sshd@5-10.0.0.142:22-10.0.0.1:56988.service - OpenSSH per-connection server daemon (10.0.0.1:56988). May 14 23:51:29.704113 systemd-logind[1487]: Removed session 5. May 14 23:51:29.759516 sshd[1667]: Accepted publickey for core from 10.0.0.1 port 56988 ssh2: RSA SHA256:lk2TkYBEL43KPVbrGyh3Ro8IB8NGN6uTNXzFyrYR01I May 14 23:51:29.761576 sshd-session[1667]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 23:51:29.767678 systemd-logind[1487]: New session 6 of user core. May 14 23:51:29.777579 systemd[1]: Started session-6.scope - Session 6 of User core. May 14 23:51:29.835558 sudo[1672]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules May 14 23:51:29.835892 sudo[1672]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 14 23:51:29.840825 sudo[1672]: pam_unix(sudo:session): session closed for user root May 14 23:51:29.847986 sudo[1671]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules May 14 23:51:29.848355 sudo[1671]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 14 23:51:29.860753 systemd[1]: Starting audit-rules.service - Load Audit Rules... May 14 23:51:29.913964 augenrules[1694]: No rules May 14 23:51:29.916091 systemd[1]: audit-rules.service: Deactivated successfully. May 14 23:51:29.916458 systemd[1]: Finished audit-rules.service - Load Audit Rules. May 14 23:51:29.917860 sudo[1671]: pam_unix(sudo:session): session closed for user root May 14 23:51:29.919426 sshd[1670]: Connection closed by 10.0.0.1 port 56988 May 14 23:51:29.919750 sshd-session[1667]: pam_unix(sshd:session): session closed for user core May 14 23:51:29.928424 systemd[1]: sshd@5-10.0.0.142:22-10.0.0.1:56988.service: Deactivated successfully. May 14 23:51:29.930429 systemd[1]: session-6.scope: Deactivated successfully. May 14 23:51:29.932073 systemd-logind[1487]: Session 6 logged out. Waiting for processes to exit. May 14 23:51:29.933425 systemd[1]: Started sshd@6-10.0.0.142:22-10.0.0.1:57002.service - OpenSSH per-connection server daemon (10.0.0.1:57002). May 14 23:51:29.934243 systemd-logind[1487]: Removed session 6. May 14 23:51:29.984604 sshd[1702]: Accepted publickey for core from 10.0.0.1 port 57002 ssh2: RSA SHA256:lk2TkYBEL43KPVbrGyh3Ro8IB8NGN6uTNXzFyrYR01I May 14 23:51:29.986528 sshd-session[1702]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 23:51:29.991449 systemd-logind[1487]: New session 7 of user core. May 14 23:51:30.001604 systemd[1]: Started session-7.scope - Session 7 of User core. May 14 23:51:30.059983 sudo[1706]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh May 14 23:51:30.060464 sudo[1706]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 14 23:51:31.438788 systemd[1]: Starting docker.service - Docker Application Container Engine... May 14 23:51:31.452728 (dockerd)[1727]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU May 14 23:51:32.101486 dockerd[1727]: time="2025-05-14T23:51:32.099817798Z" level=info msg="Starting up" May 14 23:51:32.103554 dockerd[1727]: time="2025-05-14T23:51:32.103525332Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" May 14 23:51:32.762057 dockerd[1727]: time="2025-05-14T23:51:32.761947684Z" level=info msg="Loading containers: start." May 14 23:51:33.288329 kernel: Initializing XFRM netlink socket May 14 23:51:33.430602 systemd-networkd[1437]: docker0: Link UP May 14 23:51:33.519831 dockerd[1727]: time="2025-05-14T23:51:33.519754493Z" level=info msg="Loading containers: done." May 14 23:51:33.551789 dockerd[1727]: time="2025-05-14T23:51:33.551616137Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 May 14 23:51:33.551789 dockerd[1727]: time="2025-05-14T23:51:33.551744337Z" level=info msg="Docker daemon" commit=c710b88579fcb5e0d53f96dcae976d79323b9166 containerd-snapshotter=false storage-driver=overlay2 version=27.4.1 May 14 23:51:33.552065 dockerd[1727]: time="2025-05-14T23:51:33.551916169Z" level=info msg="Daemon has completed initialization" May 14 23:51:33.601883 dockerd[1727]: time="2025-05-14T23:51:33.601776403Z" level=info msg="API listen on /run/docker.sock" May 14 23:51:33.602047 systemd[1]: Started docker.service - Docker Application Container Engine. May 14 23:51:34.751711 containerd[1510]: time="2025-05-14T23:51:34.751640129Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.30.12\"" May 14 23:51:36.021916 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1416347595.mount: Deactivated successfully. May 14 23:51:38.490257 containerd[1510]: time="2025-05-14T23:51:38.490176895Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.30.12\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:51:38.547650 containerd[1510]: time="2025-05-14T23:51:38.547522068Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.30.12: active requests=0, bytes read=32674873" May 14 23:51:38.584718 containerd[1510]: time="2025-05-14T23:51:38.584649779Z" level=info msg="ImageCreate event name:\"sha256:e113c59aa22f0650435e2a3ed64aadb01e87f3d2835aa3825fe078cd39699bfb\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:51:38.607168 containerd[1510]: time="2025-05-14T23:51:38.607103917Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:4878682f7a044274d42399a6316ef452c5411aafd4ad99cc57de7235ca490e4e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:51:38.608372 containerd[1510]: time="2025-05-14T23:51:38.608326019Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.30.12\" with image id \"sha256:e113c59aa22f0650435e2a3ed64aadb01e87f3d2835aa3825fe078cd39699bfb\", repo tag \"registry.k8s.io/kube-apiserver:v1.30.12\", repo digest \"registry.k8s.io/kube-apiserver@sha256:4878682f7a044274d42399a6316ef452c5411aafd4ad99cc57de7235ca490e4e\", size \"32671673\" in 3.856642299s" May 14 23:51:38.608419 containerd[1510]: time="2025-05-14T23:51:38.608381613Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.30.12\" returns image reference \"sha256:e113c59aa22f0650435e2a3ed64aadb01e87f3d2835aa3825fe078cd39699bfb\"" May 14 23:51:38.628219 containerd[1510]: time="2025-05-14T23:51:38.628177505Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.30.12\"" May 14 23:51:39.100577 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. May 14 23:51:39.102269 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 14 23:51:39.291071 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 14 23:51:39.296581 (kubelet)[2014]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 14 23:51:39.418596 kubelet[2014]: E0514 23:51:39.418403 2014 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 14 23:51:39.426325 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 14 23:51:39.426554 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 14 23:51:39.427126 systemd[1]: kubelet.service: Consumed 231ms CPU time, 96.9M memory peak. May 14 23:51:40.587823 containerd[1510]: time="2025-05-14T23:51:40.587737740Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.30.12\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:51:40.588804 containerd[1510]: time="2025-05-14T23:51:40.588724961Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.30.12: active requests=0, bytes read=29617534" May 14 23:51:40.590235 containerd[1510]: time="2025-05-14T23:51:40.590191071Z" level=info msg="ImageCreate event name:\"sha256:70742b7b7d90a618a1fa06d89248dbe2c291c19d7f75f4ad60a69d0454dbbac8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:51:40.592852 containerd[1510]: time="2025-05-14T23:51:40.592792872Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:3a36711d0409d565b370a18d0c19339e93d4f1b1f2b3fd382eb31c714c463b74\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:51:40.593988 containerd[1510]: time="2025-05-14T23:51:40.593936286Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.30.12\" with image id \"sha256:70742b7b7d90a618a1fa06d89248dbe2c291c19d7f75f4ad60a69d0454dbbac8\", repo tag \"registry.k8s.io/kube-controller-manager:v1.30.12\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:3a36711d0409d565b370a18d0c19339e93d4f1b1f2b3fd382eb31c714c463b74\", size \"31105907\" in 1.965714358s" May 14 23:51:40.593988 containerd[1510]: time="2025-05-14T23:51:40.593975880Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.30.12\" returns image reference \"sha256:70742b7b7d90a618a1fa06d89248dbe2c291c19d7f75f4ad60a69d0454dbbac8\"" May 14 23:51:40.615513 containerd[1510]: time="2025-05-14T23:51:40.615456522Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.30.12\"" May 14 23:51:41.645319 containerd[1510]: time="2025-05-14T23:51:41.645224052Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.30.12\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:51:41.645940 containerd[1510]: time="2025-05-14T23:51:41.645873750Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.30.12: active requests=0, bytes read=17903682" May 14 23:51:41.647422 containerd[1510]: time="2025-05-14T23:51:41.647366130Z" level=info msg="ImageCreate event name:\"sha256:c0b91cfea9f9a1c09fc5d056f3a015e52604fd0d63671ff5bf31e642402ef05d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:51:41.649720 containerd[1510]: time="2025-05-14T23:51:41.649689427Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:521c843d01025be7d4e246ddee8cde74556eb9813c606d6db9f0f03236f6d029\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:51:41.650674 containerd[1510]: time="2025-05-14T23:51:41.650626956Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.30.12\" with image id \"sha256:c0b91cfea9f9a1c09fc5d056f3a015e52604fd0d63671ff5bf31e642402ef05d\", repo tag \"registry.k8s.io/kube-scheduler:v1.30.12\", repo digest \"registry.k8s.io/kube-scheduler@sha256:521c843d01025be7d4e246ddee8cde74556eb9813c606d6db9f0f03236f6d029\", size \"19392073\" in 1.03511483s" May 14 23:51:41.650674 containerd[1510]: time="2025-05-14T23:51:41.650665238Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.30.12\" returns image reference \"sha256:c0b91cfea9f9a1c09fc5d056f3a015e52604fd0d63671ff5bf31e642402ef05d\"" May 14 23:51:41.671180 containerd[1510]: time="2025-05-14T23:51:41.671139682Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.12\"" May 14 23:51:43.057542 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1767425106.mount: Deactivated successfully. May 14 23:51:43.413824 containerd[1510]: time="2025-05-14T23:51:43.413656806Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.30.12\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:51:43.414886 containerd[1510]: time="2025-05-14T23:51:43.414836599Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.30.12: active requests=0, bytes read=29185817" May 14 23:51:43.416111 containerd[1510]: time="2025-05-14T23:51:43.416059202Z" level=info msg="ImageCreate event name:\"sha256:c9356fea5d151501039907c3ba870272461396117eabc74063632616f4e31b2b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:51:43.417837 containerd[1510]: time="2025-05-14T23:51:43.417805979Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:ea8c7d5392acf6b0c11ebba78301e1a6c2dc6abcd7544102ed578e49d1c82f15\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:51:43.418261 containerd[1510]: time="2025-05-14T23:51:43.418229553Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.30.12\" with image id \"sha256:c9356fea5d151501039907c3ba870272461396117eabc74063632616f4e31b2b\", repo tag \"registry.k8s.io/kube-proxy:v1.30.12\", repo digest \"registry.k8s.io/kube-proxy@sha256:ea8c7d5392acf6b0c11ebba78301e1a6c2dc6abcd7544102ed578e49d1c82f15\", size \"29184836\" in 1.746864218s" May 14 23:51:43.418289 containerd[1510]: time="2025-05-14T23:51:43.418259329Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.12\" returns image reference \"sha256:c9356fea5d151501039907c3ba870272461396117eabc74063632616f4e31b2b\"" May 14 23:51:43.440562 containerd[1510]: time="2025-05-14T23:51:43.440513421Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\"" May 14 23:51:43.928052 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3391376734.mount: Deactivated successfully. May 14 23:51:45.439332 containerd[1510]: time="2025-05-14T23:51:45.439207791Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:51:45.440273 containerd[1510]: time="2025-05-14T23:51:45.440153915Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.1: active requests=0, bytes read=18185761" May 14 23:51:45.442334 containerd[1510]: time="2025-05-14T23:51:45.442269123Z" level=info msg="ImageCreate event name:\"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:51:45.450108 containerd[1510]: time="2025-05-14T23:51:45.450065857Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:51:45.451094 containerd[1510]: time="2025-05-14T23:51:45.451025147Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.1\" with image id \"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\", size \"18182961\" in 2.010463295s" May 14 23:51:45.451094 containerd[1510]: time="2025-05-14T23:51:45.451067536Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\" returns image reference \"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\"" May 14 23:51:45.478109 containerd[1510]: time="2025-05-14T23:51:45.478060137Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\"" May 14 23:51:46.020159 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3300870581.mount: Deactivated successfully. May 14 23:51:46.027453 containerd[1510]: time="2025-05-14T23:51:46.027397456Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:51:46.028271 containerd[1510]: time="2025-05-14T23:51:46.028207245Z" level=info msg="stop pulling image registry.k8s.io/pause:3.9: active requests=0, bytes read=322290" May 14 23:51:46.029531 containerd[1510]: time="2025-05-14T23:51:46.029479021Z" level=info msg="ImageCreate event name:\"sha256:e6f1816883972d4be47bd48879a08919b96afcd344132622e4d444987919323c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:51:46.031544 containerd[1510]: time="2025-05-14T23:51:46.031511583Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:51:46.033949 containerd[1510]: time="2025-05-14T23:51:46.032990698Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.9\" with image id \"sha256:e6f1816883972d4be47bd48879a08919b96afcd344132622e4d444987919323c\", repo tag \"registry.k8s.io/pause:3.9\", repo digest \"registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097\", size \"321520\" in 554.891057ms" May 14 23:51:46.033949 containerd[1510]: time="2025-05-14T23:51:46.033024531Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\" returns image reference \"sha256:e6f1816883972d4be47bd48879a08919b96afcd344132622e4d444987919323c\"" May 14 23:51:46.054477 containerd[1510]: time="2025-05-14T23:51:46.054428369Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.12-0\"" May 14 23:51:46.676760 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4266230090.mount: Deactivated successfully. May 14 23:51:48.557943 containerd[1510]: time="2025-05-14T23:51:48.557879057Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.12-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:51:48.558760 containerd[1510]: time="2025-05-14T23:51:48.558700016Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.12-0: active requests=0, bytes read=57238571" May 14 23:51:48.560089 containerd[1510]: time="2025-05-14T23:51:48.560048526Z" level=info msg="ImageCreate event name:\"sha256:3861cfcd7c04ccac1f062788eca39487248527ef0c0cfd477a83d7691a75a899\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:51:48.562729 containerd[1510]: time="2025-05-14T23:51:48.562697875Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:44a8e24dcbba3470ee1fee21d5e88d128c936e9b55d4bc51fbef8086f8ed123b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:51:48.563698 containerd[1510]: time="2025-05-14T23:51:48.563651494Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.12-0\" with image id \"sha256:3861cfcd7c04ccac1f062788eca39487248527ef0c0cfd477a83d7691a75a899\", repo tag \"registry.k8s.io/etcd:3.5.12-0\", repo digest \"registry.k8s.io/etcd@sha256:44a8e24dcbba3470ee1fee21d5e88d128c936e9b55d4bc51fbef8086f8ed123b\", size \"57236178\" in 2.509185976s" May 14 23:51:48.563755 containerd[1510]: time="2025-05-14T23:51:48.563699624Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.12-0\" returns image reference \"sha256:3861cfcd7c04ccac1f062788eca39487248527ef0c0cfd477a83d7691a75a899\"" May 14 23:51:49.482217 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. May 14 23:51:49.484443 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 14 23:51:49.696033 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 14 23:51:49.716880 (kubelet)[2280]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 14 23:51:49.762382 kubelet[2280]: E0514 23:51:49.762193 2280 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 14 23:51:49.767257 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 14 23:51:49.767532 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 14 23:51:49.767973 systemd[1]: kubelet.service: Consumed 229ms CPU time, 98.1M memory peak. May 14 23:51:51.015558 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. May 14 23:51:51.015752 systemd[1]: kubelet.service: Consumed 229ms CPU time, 98.1M memory peak. May 14 23:51:51.018307 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 14 23:51:51.054443 systemd[1]: Reload requested from client PID 2296 ('systemctl') (unit session-7.scope)... May 14 23:51:51.054477 systemd[1]: Reloading... May 14 23:51:51.146344 zram_generator::config[2339]: No configuration found. May 14 23:51:51.862209 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 14 23:51:51.974910 systemd[1]: Reloading finished in 919 ms. May 14 23:51:52.046566 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... May 14 23:51:52.048588 systemd[1]: kubelet.service: Deactivated successfully. May 14 23:51:52.048898 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. May 14 23:51:52.048954 systemd[1]: kubelet.service: Consumed 162ms CPU time, 83.7M memory peak. May 14 23:51:52.050768 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 14 23:51:52.235738 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 14 23:51:52.247697 (kubelet)[2389]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS May 14 23:51:52.287331 kubelet[2389]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 14 23:51:52.287331 kubelet[2389]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. May 14 23:51:52.287331 kubelet[2389]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 14 23:51:52.287331 kubelet[2389]: I0514 23:51:52.285988 2389 server.go:205] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" May 14 23:51:52.539659 kubelet[2389]: I0514 23:51:52.539510 2389 server.go:484] "Kubelet version" kubeletVersion="v1.30.1" May 14 23:51:52.539659 kubelet[2389]: I0514 23:51:52.539544 2389 server.go:486] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" May 14 23:51:52.539822 kubelet[2389]: I0514 23:51:52.539796 2389 server.go:927] "Client rotation is on, will bootstrap in background" May 14 23:51:52.558117 kubelet[2389]: I0514 23:51:52.558061 2389 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" May 14 23:51:52.572631 kubelet[2389]: E0514 23:51:52.572559 2389 certificate_manager.go:562] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://10.0.0.142:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 10.0.0.142:6443: connect: connection refused May 14 23:51:52.595977 kubelet[2389]: I0514 23:51:52.595926 2389 server.go:742] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" May 14 23:51:52.608662 kubelet[2389]: I0514 23:51:52.608570 2389 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] May 14 23:51:52.608862 kubelet[2389]: I0514 23:51:52.608648 2389 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} May 14 23:51:52.611110 kubelet[2389]: I0514 23:51:52.611077 2389 topology_manager.go:138] "Creating topology manager with none policy" May 14 23:51:52.611110 kubelet[2389]: I0514 23:51:52.611098 2389 container_manager_linux.go:301] "Creating device plugin manager" May 14 23:51:52.611287 kubelet[2389]: I0514 23:51:52.611265 2389 state_mem.go:36] "Initialized new in-memory state store" May 14 23:51:52.612414 kubelet[2389]: I0514 23:51:52.612379 2389 kubelet.go:400] "Attempting to sync node with API server" May 14 23:51:52.612414 kubelet[2389]: I0514 23:51:52.612399 2389 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" May 14 23:51:52.612487 kubelet[2389]: I0514 23:51:52.612422 2389 kubelet.go:312] "Adding apiserver pod source" May 14 23:51:52.612487 kubelet[2389]: I0514 23:51:52.612441 2389 apiserver.go:42] "Waiting for node sync before watching apiserver pods" May 14 23:51:52.613206 kubelet[2389]: W0514 23:51:52.613135 2389 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.142:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.142:6443: connect: connection refused May 14 23:51:52.613267 kubelet[2389]: E0514 23:51:52.613208 2389 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.0.0.142:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.142:6443: connect: connection refused May 14 23:51:52.613678 kubelet[2389]: W0514 23:51:52.613612 2389 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.142:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.142:6443: connect: connection refused May 14 23:51:52.613678 kubelet[2389]: E0514 23:51:52.613677 2389 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.0.0.142:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.142:6443: connect: connection refused May 14 23:51:52.619078 kubelet[2389]: I0514 23:51:52.619051 2389 kuberuntime_manager.go:261] "Container runtime initialized" containerRuntime="containerd" version="v2.0.1" apiVersion="v1" May 14 23:51:52.620844 kubelet[2389]: I0514 23:51:52.620809 2389 kubelet.go:815] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" May 14 23:51:52.620930 kubelet[2389]: W0514 23:51:52.620870 2389 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. May 14 23:51:52.621539 kubelet[2389]: I0514 23:51:52.621520 2389 server.go:1264] "Started kubelet" May 14 23:51:52.621778 kubelet[2389]: I0514 23:51:52.621733 2389 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 May 14 23:51:52.621845 kubelet[2389]: I0514 23:51:52.621772 2389 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 May 14 23:51:52.622143 kubelet[2389]: I0514 23:51:52.622109 2389 server.go:227] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" May 14 23:51:52.623458 kubelet[2389]: I0514 23:51:52.623418 2389 server.go:455] "Adding debug handlers to kubelet server" May 14 23:51:52.623970 kubelet[2389]: I0514 23:51:52.623930 2389 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" May 14 23:51:52.628834 kubelet[2389]: E0514 23:51:52.628806 2389 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 14 23:51:52.628902 kubelet[2389]: I0514 23:51:52.628850 2389 volume_manager.go:291] "Starting Kubelet Volume Manager" May 14 23:51:52.629676 kubelet[2389]: I0514 23:51:52.628950 2389 desired_state_of_world_populator.go:149] "Desired state populator starts to run" May 14 23:51:52.629676 kubelet[2389]: I0514 23:51:52.629010 2389 reconciler.go:26] "Reconciler: start to sync state" May 14 23:51:52.629676 kubelet[2389]: E0514 23:51:52.629411 2389 kubelet.go:1467] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" May 14 23:51:52.629676 kubelet[2389]: W0514 23:51:52.629397 2389 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.142:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.142:6443: connect: connection refused May 14 23:51:52.629676 kubelet[2389]: E0514 23:51:52.629451 2389 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.0.0.142:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.142:6443: connect: connection refused May 14 23:51:52.629980 kubelet[2389]: E0514 23:51:52.629945 2389 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.142:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.142:6443: connect: connection refused" interval="200ms" May 14 23:51:52.630052 kubelet[2389]: I0514 23:51:52.630024 2389 factory.go:221] Registration of the systemd container factory successfully May 14 23:51:52.630129 kubelet[2389]: I0514 23:51:52.630109 2389 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory May 14 23:51:52.630912 kubelet[2389]: I0514 23:51:52.630877 2389 factory.go:221] Registration of the containerd container factory successfully May 14 23:51:52.642058 kubelet[2389]: E0514 23:51:52.641943 2389 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.142:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.142:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.183f89d155ab5656 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-05-14 23:51:52.621504086 +0000 UTC m=+0.369512481,LastTimestamp:2025-05-14 23:51:52.621504086 +0000 UTC m=+0.369512481,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" May 14 23:51:52.649227 kubelet[2389]: I0514 23:51:52.649167 2389 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" May 14 23:51:52.650898 kubelet[2389]: I0514 23:51:52.650872 2389 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" May 14 23:51:52.651687 kubelet[2389]: I0514 23:51:52.651019 2389 status_manager.go:217] "Starting to sync pod status with apiserver" May 14 23:51:52.651687 kubelet[2389]: I0514 23:51:52.651048 2389 kubelet.go:2337] "Starting kubelet main sync loop" May 14 23:51:52.651687 kubelet[2389]: E0514 23:51:52.651099 2389 kubelet.go:2361] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" May 14 23:51:52.651687 kubelet[2389]: W0514 23:51:52.651637 2389 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.142:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.142:6443: connect: connection refused May 14 23:51:52.651687 kubelet[2389]: E0514 23:51:52.651666 2389 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.0.0.142:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.142:6443: connect: connection refused May 14 23:51:52.675012 kubelet[2389]: I0514 23:51:52.674970 2389 cpu_manager.go:214] "Starting CPU manager" policy="none" May 14 23:51:52.675012 kubelet[2389]: I0514 23:51:52.674997 2389 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" May 14 23:51:52.675012 kubelet[2389]: I0514 23:51:52.675018 2389 state_mem.go:36] "Initialized new in-memory state store" May 14 23:51:52.730718 kubelet[2389]: I0514 23:51:52.730665 2389 kubelet_node_status.go:73] "Attempting to register node" node="localhost" May 14 23:51:52.731120 kubelet[2389]: E0514 23:51:52.731073 2389 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.142:6443/api/v1/nodes\": dial tcp 10.0.0.142:6443: connect: connection refused" node="localhost" May 14 23:51:52.751610 kubelet[2389]: E0514 23:51:52.751555 2389 kubelet.go:2361] "Skipping pod synchronization" err="container runtime status check may not have completed yet" May 14 23:51:52.831529 kubelet[2389]: E0514 23:51:52.831378 2389 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.142:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.142:6443: connect: connection refused" interval="400ms" May 14 23:51:52.866778 kubelet[2389]: I0514 23:51:52.866715 2389 policy_none.go:49] "None policy: Start" May 14 23:51:52.867497 kubelet[2389]: I0514 23:51:52.867468 2389 memory_manager.go:170] "Starting memorymanager" policy="None" May 14 23:51:52.867563 kubelet[2389]: I0514 23:51:52.867512 2389 state_mem.go:35] "Initializing new in-memory state store" May 14 23:51:52.876077 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. May 14 23:51:52.892672 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. May 14 23:51:52.896329 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. May 14 23:51:52.910276 kubelet[2389]: I0514 23:51:52.910244 2389 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" May 14 23:51:52.910571 kubelet[2389]: I0514 23:51:52.910525 2389 container_log_manager.go:186] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" May 14 23:51:52.910714 kubelet[2389]: I0514 23:51:52.910693 2389 plugin_manager.go:118] "Starting Kubelet Plugin Manager" May 14 23:51:52.911657 kubelet[2389]: E0514 23:51:52.911639 2389 eviction_manager.go:282] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" May 14 23:51:52.933097 kubelet[2389]: I0514 23:51:52.933054 2389 kubelet_node_status.go:73] "Attempting to register node" node="localhost" May 14 23:51:52.933464 kubelet[2389]: E0514 23:51:52.933429 2389 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.142:6443/api/v1/nodes\": dial tcp 10.0.0.142:6443: connect: connection refused" node="localhost" May 14 23:51:52.952649 kubelet[2389]: I0514 23:51:52.952588 2389 topology_manager.go:215] "Topology Admit Handler" podUID="5c0805abca5da2a4f81d6d9b39961eac" podNamespace="kube-system" podName="kube-apiserver-localhost" May 14 23:51:52.953554 kubelet[2389]: I0514 23:51:52.953522 2389 topology_manager.go:215] "Topology Admit Handler" podUID="b20b39a8540dba87b5883a6f0f602dba" podNamespace="kube-system" podName="kube-controller-manager-localhost" May 14 23:51:52.954278 kubelet[2389]: I0514 23:51:52.954246 2389 topology_manager.go:215] "Topology Admit Handler" podUID="6ece95f10dbffa04b25ec3439a115512" podNamespace="kube-system" podName="kube-scheduler-localhost" May 14 23:51:52.961024 systemd[1]: Created slice kubepods-burstable-podb20b39a8540dba87b5883a6f0f602dba.slice - libcontainer container kubepods-burstable-podb20b39a8540dba87b5883a6f0f602dba.slice. May 14 23:51:52.974081 systemd[1]: Created slice kubepods-burstable-pod5c0805abca5da2a4f81d6d9b39961eac.slice - libcontainer container kubepods-burstable-pod5c0805abca5da2a4f81d6d9b39961eac.slice. May 14 23:51:52.994982 systemd[1]: Created slice kubepods-burstable-pod6ece95f10dbffa04b25ec3439a115512.slice - libcontainer container kubepods-burstable-pod6ece95f10dbffa04b25ec3439a115512.slice. May 14 23:51:53.031015 kubelet[2389]: I0514 23:51:53.030976 2389 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 14 23:51:53.031170 kubelet[2389]: I0514 23:51:53.031018 2389 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/5c0805abca5da2a4f81d6d9b39961eac-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"5c0805abca5da2a4f81d6d9b39961eac\") " pod="kube-system/kube-apiserver-localhost" May 14 23:51:53.031170 kubelet[2389]: I0514 23:51:53.031044 2389 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/5c0805abca5da2a4f81d6d9b39961eac-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"5c0805abca5da2a4f81d6d9b39961eac\") " pod="kube-system/kube-apiserver-localhost" May 14 23:51:53.031170 kubelet[2389]: I0514 23:51:53.031065 2389 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/5c0805abca5da2a4f81d6d9b39961eac-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"5c0805abca5da2a4f81d6d9b39961eac\") " pod="kube-system/kube-apiserver-localhost" May 14 23:51:53.031170 kubelet[2389]: I0514 23:51:53.031088 2389 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 14 23:51:53.031170 kubelet[2389]: I0514 23:51:53.031107 2389 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 14 23:51:53.031284 kubelet[2389]: I0514 23:51:53.031127 2389 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 14 23:51:53.031284 kubelet[2389]: I0514 23:51:53.031163 2389 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 14 23:51:53.031284 kubelet[2389]: I0514 23:51:53.031182 2389 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/6ece95f10dbffa04b25ec3439a115512-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"6ece95f10dbffa04b25ec3439a115512\") " pod="kube-system/kube-scheduler-localhost" May 14 23:51:53.232700 kubelet[2389]: E0514 23:51:53.232555 2389 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.142:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.142:6443: connect: connection refused" interval="800ms" May 14 23:51:53.273443 containerd[1510]: time="2025-05-14T23:51:53.273397171Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:b20b39a8540dba87b5883a6f0f602dba,Namespace:kube-system,Attempt:0,}" May 14 23:51:53.293153 containerd[1510]: time="2025-05-14T23:51:53.293098705Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:5c0805abca5da2a4f81d6d9b39961eac,Namespace:kube-system,Attempt:0,}" May 14 23:51:53.297776 containerd[1510]: time="2025-05-14T23:51:53.297740442Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:6ece95f10dbffa04b25ec3439a115512,Namespace:kube-system,Attempt:0,}" May 14 23:51:53.335465 kubelet[2389]: I0514 23:51:53.335433 2389 kubelet_node_status.go:73] "Attempting to register node" node="localhost" May 14 23:51:53.335943 kubelet[2389]: E0514 23:51:53.335766 2389 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.142:6443/api/v1/nodes\": dial tcp 10.0.0.142:6443: connect: connection refused" node="localhost" May 14 23:51:53.659702 kubelet[2389]: W0514 23:51:53.659497 2389 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.142:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.142:6443: connect: connection refused May 14 23:51:53.659702 kubelet[2389]: E0514 23:51:53.659557 2389 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.0.0.142:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.142:6443: connect: connection refused May 14 23:51:53.840078 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2335994157.mount: Deactivated successfully. May 14 23:51:53.849726 kubelet[2389]: W0514 23:51:53.849636 2389 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.142:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.142:6443: connect: connection refused May 14 23:51:53.849726 kubelet[2389]: E0514 23:51:53.849723 2389 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.0.0.142:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.142:6443: connect: connection refused May 14 23:51:53.853211 containerd[1510]: time="2025-05-14T23:51:53.853140023Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 14 23:51:53.870042 containerd[1510]: time="2025-05-14T23:51:53.869917152Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" May 14 23:51:54.034159 kubelet[2389]: E0514 23:51:54.034014 2389 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.142:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.142:6443: connect: connection refused" interval="1.6s" May 14 23:51:54.137219 kubelet[2389]: I0514 23:51:54.137185 2389 kubelet_node_status.go:73] "Attempting to register node" node="localhost" May 14 23:51:54.137444 kubelet[2389]: W0514 23:51:54.137399 2389 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.142:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.142:6443: connect: connection refused May 14 23:51:54.137488 kubelet[2389]: E0514 23:51:54.137449 2389 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.0.0.142:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.142:6443: connect: connection refused May 14 23:51:54.137488 kubelet[2389]: E0514 23:51:54.137441 2389 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.142:6443/api/v1/nodes\": dial tcp 10.0.0.142:6443: connect: connection refused" node="localhost" May 14 23:51:54.215216 kubelet[2389]: W0514 23:51:54.215123 2389 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.142:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.142:6443: connect: connection refused May 14 23:51:54.215216 kubelet[2389]: E0514 23:51:54.215196 2389 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.0.0.142:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.142:6443: connect: connection refused May 14 23:51:54.724477 containerd[1510]: time="2025-05-14T23:51:54.724395589Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 14 23:51:54.765220 kubelet[2389]: E0514 23:51:54.765169 2389 certificate_manager.go:562] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://10.0.0.142:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 10.0.0.142:6443: connect: connection refused May 14 23:51:54.812487 containerd[1510]: time="2025-05-14T23:51:54.812420829Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 14 23:51:54.815929 containerd[1510]: time="2025-05-14T23:51:54.815878004Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=0" May 14 23:51:54.816879 containerd[1510]: time="2025-05-14T23:51:54.816840078Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 14 23:51:54.817940 containerd[1510]: time="2025-05-14T23:51:54.817897642Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=0" May 14 23:51:54.818935 containerd[1510]: time="2025-05-14T23:51:54.818904771Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 14 23:51:54.819716 containerd[1510]: time="2025-05-14T23:51:54.819661901Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 1.542483889s" May 14 23:51:54.823320 containerd[1510]: time="2025-05-14T23:51:54.823277052Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 1.527862523s" May 14 23:51:54.823885 containerd[1510]: time="2025-05-14T23:51:54.823850889Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 1.521569079s" May 14 23:51:54.863370 containerd[1510]: time="2025-05-14T23:51:54.863283804Z" level=info msg="connecting to shim 6e3234b88530f242ca8bd3ec186819469aa1351964c577a972c54a5eedbc047b" address="unix:///run/containerd/s/3d3b4189f68489ffcc9f76cfe54cdb014034ace4c0ecc1c14111089a95ec11b6" namespace=k8s.io protocol=ttrpc version=3 May 14 23:51:54.880674 containerd[1510]: time="2025-05-14T23:51:54.880615853Z" level=info msg="connecting to shim df525df5d63349ba39c7908c69f79fd5fca801fba01cd51ab270bb2a6d94353f" address="unix:///run/containerd/s/c08dd49897151bf1f9579267ccb911c6e9ceafe23d6016f5a77bf91f393a670d" namespace=k8s.io protocol=ttrpc version=3 May 14 23:51:54.890978 containerd[1510]: time="2025-05-14T23:51:54.890915503Z" level=info msg="connecting to shim 507802de7f2d4d2d9a3553a482865eaf513bba7b6dcae459860649c629dc093d" address="unix:///run/containerd/s/e650e371d2f654d16cea73c1aecfcf8757d52539c01ec348399af774da323c99" namespace=k8s.io protocol=ttrpc version=3 May 14 23:51:54.928520 systemd[1]: Started cri-containerd-6e3234b88530f242ca8bd3ec186819469aa1351964c577a972c54a5eedbc047b.scope - libcontainer container 6e3234b88530f242ca8bd3ec186819469aa1351964c577a972c54a5eedbc047b. May 14 23:51:54.936443 systemd[1]: Started cri-containerd-df525df5d63349ba39c7908c69f79fd5fca801fba01cd51ab270bb2a6d94353f.scope - libcontainer container df525df5d63349ba39c7908c69f79fd5fca801fba01cd51ab270bb2a6d94353f. May 14 23:51:54.948096 systemd[1]: Started cri-containerd-507802de7f2d4d2d9a3553a482865eaf513bba7b6dcae459860649c629dc093d.scope - libcontainer container 507802de7f2d4d2d9a3553a482865eaf513bba7b6dcae459860649c629dc093d. May 14 23:51:55.143842 containerd[1510]: time="2025-05-14T23:51:55.143582353Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:b20b39a8540dba87b5883a6f0f602dba,Namespace:kube-system,Attempt:0,} returns sandbox id \"6e3234b88530f242ca8bd3ec186819469aa1351964c577a972c54a5eedbc047b\"" May 14 23:51:55.147933 containerd[1510]: time="2025-05-14T23:51:55.147880726Z" level=info msg="CreateContainer within sandbox \"6e3234b88530f242ca8bd3ec186819469aa1351964c577a972c54a5eedbc047b\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" May 14 23:51:55.206421 containerd[1510]: time="2025-05-14T23:51:55.206353644Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:5c0805abca5da2a4f81d6d9b39961eac,Namespace:kube-system,Attempt:0,} returns sandbox id \"507802de7f2d4d2d9a3553a482865eaf513bba7b6dcae459860649c629dc093d\"" May 14 23:51:55.209084 containerd[1510]: time="2025-05-14T23:51:55.209027700Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:6ece95f10dbffa04b25ec3439a115512,Namespace:kube-system,Attempt:0,} returns sandbox id \"df525df5d63349ba39c7908c69f79fd5fca801fba01cd51ab270bb2a6d94353f\"" May 14 23:51:55.209241 containerd[1510]: time="2025-05-14T23:51:55.209135142Z" level=info msg="CreateContainer within sandbox \"507802de7f2d4d2d9a3553a482865eaf513bba7b6dcae459860649c629dc093d\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" May 14 23:51:55.211434 containerd[1510]: time="2025-05-14T23:51:55.211405430Z" level=info msg="CreateContainer within sandbox \"df525df5d63349ba39c7908c69f79fd5fca801fba01cd51ab270bb2a6d94353f\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" May 14 23:51:55.271830 containerd[1510]: time="2025-05-14T23:51:55.271755760Z" level=info msg="Container 52705e901648e94c55786949cb2d936f9ad42774be9a7aeb6601258fc7411429: CDI devices from CRI Config.CDIDevices: []" May 14 23:51:55.276345 containerd[1510]: time="2025-05-14T23:51:55.276286208Z" level=info msg="Container 89f74a2a01dec82598e5be020b73688c509cba458cba517bec0f70393142e4a6: CDI devices from CRI Config.CDIDevices: []" May 14 23:51:55.288397 containerd[1510]: time="2025-05-14T23:51:55.288336190Z" level=info msg="Container a576471c90294ef76c470f7dfed6933ea451196706a43a6da3a727748dc9ba4d: CDI devices from CRI Config.CDIDevices: []" May 14 23:51:55.289331 containerd[1510]: time="2025-05-14T23:51:55.289264091Z" level=info msg="CreateContainer within sandbox \"507802de7f2d4d2d9a3553a482865eaf513bba7b6dcae459860649c629dc093d\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"89f74a2a01dec82598e5be020b73688c509cba458cba517bec0f70393142e4a6\"" May 14 23:51:55.290114 containerd[1510]: time="2025-05-14T23:51:55.290088928Z" level=info msg="StartContainer for \"89f74a2a01dec82598e5be020b73688c509cba458cba517bec0f70393142e4a6\"" May 14 23:51:55.291310 containerd[1510]: time="2025-05-14T23:51:55.291273059Z" level=info msg="connecting to shim 89f74a2a01dec82598e5be020b73688c509cba458cba517bec0f70393142e4a6" address="unix:///run/containerd/s/e650e371d2f654d16cea73c1aecfcf8757d52539c01ec348399af774da323c99" protocol=ttrpc version=3 May 14 23:51:55.293484 containerd[1510]: time="2025-05-14T23:51:55.293456304Z" level=info msg="CreateContainer within sandbox \"6e3234b88530f242ca8bd3ec186819469aa1351964c577a972c54a5eedbc047b\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"52705e901648e94c55786949cb2d936f9ad42774be9a7aeb6601258fc7411429\"" May 14 23:51:55.294020 containerd[1510]: time="2025-05-14T23:51:55.293976500Z" level=info msg="StartContainer for \"52705e901648e94c55786949cb2d936f9ad42774be9a7aeb6601258fc7411429\"" May 14 23:51:55.295458 containerd[1510]: time="2025-05-14T23:51:55.295417984Z" level=info msg="connecting to shim 52705e901648e94c55786949cb2d936f9ad42774be9a7aeb6601258fc7411429" address="unix:///run/containerd/s/3d3b4189f68489ffcc9f76cfe54cdb014034ace4c0ecc1c14111089a95ec11b6" protocol=ttrpc version=3 May 14 23:51:55.297978 containerd[1510]: time="2025-05-14T23:51:55.297941347Z" level=info msg="CreateContainer within sandbox \"df525df5d63349ba39c7908c69f79fd5fca801fba01cd51ab270bb2a6d94353f\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"a576471c90294ef76c470f7dfed6933ea451196706a43a6da3a727748dc9ba4d\"" May 14 23:51:55.298394 containerd[1510]: time="2025-05-14T23:51:55.298369490Z" level=info msg="StartContainer for \"a576471c90294ef76c470f7dfed6933ea451196706a43a6da3a727748dc9ba4d\"" May 14 23:51:55.299374 containerd[1510]: time="2025-05-14T23:51:55.299347013Z" level=info msg="connecting to shim a576471c90294ef76c470f7dfed6933ea451196706a43a6da3a727748dc9ba4d" address="unix:///run/containerd/s/c08dd49897151bf1f9579267ccb911c6e9ceafe23d6016f5a77bf91f393a670d" protocol=ttrpc version=3 May 14 23:51:55.317487 systemd[1]: Started cri-containerd-89f74a2a01dec82598e5be020b73688c509cba458cba517bec0f70393142e4a6.scope - libcontainer container 89f74a2a01dec82598e5be020b73688c509cba458cba517bec0f70393142e4a6. May 14 23:51:55.329655 systemd[1]: Started cri-containerd-52705e901648e94c55786949cb2d936f9ad42774be9a7aeb6601258fc7411429.scope - libcontainer container 52705e901648e94c55786949cb2d936f9ad42774be9a7aeb6601258fc7411429. May 14 23:51:55.332919 systemd[1]: Started cri-containerd-a576471c90294ef76c470f7dfed6933ea451196706a43a6da3a727748dc9ba4d.scope - libcontainer container a576471c90294ef76c470f7dfed6933ea451196706a43a6da3a727748dc9ba4d. May 14 23:51:55.404818 containerd[1510]: time="2025-05-14T23:51:55.404659880Z" level=info msg="StartContainer for \"a576471c90294ef76c470f7dfed6933ea451196706a43a6da3a727748dc9ba4d\" returns successfully" May 14 23:51:55.408442 containerd[1510]: time="2025-05-14T23:51:55.408274410Z" level=info msg="StartContainer for \"89f74a2a01dec82598e5be020b73688c509cba458cba517bec0f70393142e4a6\" returns successfully" May 14 23:51:55.415358 containerd[1510]: time="2025-05-14T23:51:55.414389230Z" level=info msg="StartContainer for \"52705e901648e94c55786949cb2d936f9ad42774be9a7aeb6601258fc7411429\" returns successfully" May 14 23:51:55.741883 kubelet[2389]: I0514 23:51:55.741696 2389 kubelet_node_status.go:73] "Attempting to register node" node="localhost" May 14 23:51:57.107769 kubelet[2389]: E0514 23:51:57.107702 2389 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" May 14 23:51:57.225605 kubelet[2389]: I0514 23:51:57.225543 2389 kubelet_node_status.go:76] "Successfully registered node" node="localhost" May 14 23:51:57.617039 kubelet[2389]: I0514 23:51:57.616999 2389 apiserver.go:52] "Watching apiserver" May 14 23:51:57.629998 kubelet[2389]: I0514 23:51:57.629945 2389 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" May 14 23:51:57.678198 kubelet[2389]: E0514 23:51:57.678143 2389 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-localhost" May 14 23:51:59.677693 systemd[1]: Reload requested from client PID 2660 ('systemctl') (unit session-7.scope)... May 14 23:51:59.677713 systemd[1]: Reloading... May 14 23:51:59.758373 zram_generator::config[2707]: No configuration found. May 14 23:51:59.868079 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 14 23:52:00.011075 systemd[1]: Reloading finished in 332 ms. May 14 23:52:00.035785 kubelet[2389]: I0514 23:52:00.035742 2389 dynamic_cafile_content.go:171] "Shutting down controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" May 14 23:52:00.035859 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... May 14 23:52:00.052978 systemd[1]: kubelet.service: Deactivated successfully. May 14 23:52:00.053372 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. May 14 23:52:00.053442 systemd[1]: kubelet.service: Consumed 919ms CPU time, 119M memory peak. May 14 23:52:00.055849 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 14 23:52:00.256236 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 14 23:52:00.276657 (kubelet)[2749]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS May 14 23:52:00.321893 kubelet[2749]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 14 23:52:00.321893 kubelet[2749]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. May 14 23:52:00.321893 kubelet[2749]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 14 23:52:00.322392 kubelet[2749]: I0514 23:52:00.321939 2749 server.go:205] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" May 14 23:52:00.328185 kubelet[2749]: I0514 23:52:00.328126 2749 server.go:484] "Kubelet version" kubeletVersion="v1.30.1" May 14 23:52:00.328185 kubelet[2749]: I0514 23:52:00.328164 2749 server.go:486] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" May 14 23:52:00.328423 kubelet[2749]: I0514 23:52:00.328388 2749 server.go:927] "Client rotation is on, will bootstrap in background" May 14 23:52:00.330063 kubelet[2749]: I0514 23:52:00.330017 2749 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". May 14 23:52:00.331284 kubelet[2749]: I0514 23:52:00.331256 2749 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" May 14 23:52:00.345335 kubelet[2749]: I0514 23:52:00.343932 2749 server.go:742] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" May 14 23:52:00.345335 kubelet[2749]: I0514 23:52:00.344255 2749 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] May 14 23:52:00.345335 kubelet[2749]: I0514 23:52:00.344355 2749 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} May 14 23:52:00.345335 kubelet[2749]: I0514 23:52:00.344616 2749 topology_manager.go:138] "Creating topology manager with none policy" May 14 23:52:00.345680 kubelet[2749]: I0514 23:52:00.344629 2749 container_manager_linux.go:301] "Creating device plugin manager" May 14 23:52:00.345680 kubelet[2749]: I0514 23:52:00.344694 2749 state_mem.go:36] "Initialized new in-memory state store" May 14 23:52:00.345680 kubelet[2749]: I0514 23:52:00.344834 2749 kubelet.go:400] "Attempting to sync node with API server" May 14 23:52:00.345680 kubelet[2749]: I0514 23:52:00.344853 2749 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" May 14 23:52:00.345680 kubelet[2749]: I0514 23:52:00.344886 2749 kubelet.go:312] "Adding apiserver pod source" May 14 23:52:00.345680 kubelet[2749]: I0514 23:52:00.344911 2749 apiserver.go:42] "Waiting for node sync before watching apiserver pods" May 14 23:52:00.346599 kubelet[2749]: I0514 23:52:00.346558 2749 kuberuntime_manager.go:261] "Container runtime initialized" containerRuntime="containerd" version="v2.0.1" apiVersion="v1" May 14 23:52:00.346846 kubelet[2749]: I0514 23:52:00.346821 2749 kubelet.go:815] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" May 14 23:52:00.349372 kubelet[2749]: I0514 23:52:00.347501 2749 server.go:1264] "Started kubelet" May 14 23:52:00.349372 kubelet[2749]: I0514 23:52:00.347940 2749 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 May 14 23:52:00.349372 kubelet[2749]: I0514 23:52:00.349139 2749 server.go:455] "Adding debug handlers to kubelet server" May 14 23:52:00.350919 kubelet[2749]: I0514 23:52:00.350857 2749 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 May 14 23:52:00.351260 kubelet[2749]: I0514 23:52:00.351232 2749 server.go:227] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" May 14 23:52:00.351815 kubelet[2749]: I0514 23:52:00.351739 2749 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" May 14 23:52:00.353608 kubelet[2749]: I0514 23:52:00.353576 2749 volume_manager.go:291] "Starting Kubelet Volume Manager" May 14 23:52:00.353756 kubelet[2749]: I0514 23:52:00.353721 2749 desired_state_of_world_populator.go:149] "Desired state populator starts to run" May 14 23:52:00.353939 kubelet[2749]: I0514 23:52:00.353912 2749 reconciler.go:26] "Reconciler: start to sync state" May 14 23:52:00.355985 kubelet[2749]: E0514 23:52:00.355947 2749 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 14 23:52:00.361815 kubelet[2749]: I0514 23:52:00.361771 2749 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory May 14 23:52:00.362825 kubelet[2749]: E0514 23:52:00.362792 2749 kubelet.go:1467] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" May 14 23:52:00.364936 kubelet[2749]: I0514 23:52:00.364903 2749 factory.go:221] Registration of the containerd container factory successfully May 14 23:52:00.364936 kubelet[2749]: I0514 23:52:00.364938 2749 factory.go:221] Registration of the systemd container factory successfully May 14 23:52:00.371548 kubelet[2749]: I0514 23:52:00.371480 2749 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" May 14 23:52:00.376978 kubelet[2749]: I0514 23:52:00.376944 2749 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" May 14 23:52:00.377900 kubelet[2749]: I0514 23:52:00.377200 2749 status_manager.go:217] "Starting to sync pod status with apiserver" May 14 23:52:00.377900 kubelet[2749]: I0514 23:52:00.377238 2749 kubelet.go:2337] "Starting kubelet main sync loop" May 14 23:52:00.377900 kubelet[2749]: E0514 23:52:00.377292 2749 kubelet.go:2361] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" May 14 23:52:00.412552 kubelet[2749]: I0514 23:52:00.412507 2749 cpu_manager.go:214] "Starting CPU manager" policy="none" May 14 23:52:00.412552 kubelet[2749]: I0514 23:52:00.412529 2749 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" May 14 23:52:00.412552 kubelet[2749]: I0514 23:52:00.412562 2749 state_mem.go:36] "Initialized new in-memory state store" May 14 23:52:00.413236 kubelet[2749]: I0514 23:52:00.412767 2749 state_mem.go:88] "Updated default CPUSet" cpuSet="" May 14 23:52:00.413236 kubelet[2749]: I0514 23:52:00.412778 2749 state_mem.go:96] "Updated CPUSet assignments" assignments={} May 14 23:52:00.413236 kubelet[2749]: I0514 23:52:00.412797 2749 policy_none.go:49] "None policy: Start" May 14 23:52:00.413631 kubelet[2749]: I0514 23:52:00.413601 2749 memory_manager.go:170] "Starting memorymanager" policy="None" May 14 23:52:00.413631 kubelet[2749]: I0514 23:52:00.413627 2749 state_mem.go:35] "Initializing new in-memory state store" May 14 23:52:00.413781 kubelet[2749]: I0514 23:52:00.413760 2749 state_mem.go:75] "Updated machine memory state" May 14 23:52:00.418991 kubelet[2749]: I0514 23:52:00.418963 2749 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" May 14 23:52:00.419458 kubelet[2749]: I0514 23:52:00.419198 2749 container_log_manager.go:186] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" May 14 23:52:00.419458 kubelet[2749]: I0514 23:52:00.419323 2749 plugin_manager.go:118] "Starting Kubelet Plugin Manager" May 14 23:52:00.458010 kubelet[2749]: I0514 23:52:00.457970 2749 kubelet_node_status.go:73] "Attempting to register node" node="localhost" May 14 23:52:00.478354 kubelet[2749]: I0514 23:52:00.478228 2749 topology_manager.go:215] "Topology Admit Handler" podUID="5c0805abca5da2a4f81d6d9b39961eac" podNamespace="kube-system" podName="kube-apiserver-localhost" May 14 23:52:00.478449 kubelet[2749]: I0514 23:52:00.478425 2749 topology_manager.go:215] "Topology Admit Handler" podUID="b20b39a8540dba87b5883a6f0f602dba" podNamespace="kube-system" podName="kube-controller-manager-localhost" May 14 23:52:00.478581 kubelet[2749]: I0514 23:52:00.478548 2749 topology_manager.go:215] "Topology Admit Handler" podUID="6ece95f10dbffa04b25ec3439a115512" podNamespace="kube-system" podName="kube-scheduler-localhost" May 14 23:52:00.555141 kubelet[2749]: I0514 23:52:00.555002 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/6ece95f10dbffa04b25ec3439a115512-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"6ece95f10dbffa04b25ec3439a115512\") " pod="kube-system/kube-scheduler-localhost" May 14 23:52:00.555141 kubelet[2749]: I0514 23:52:00.555058 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/5c0805abca5da2a4f81d6d9b39961eac-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"5c0805abca5da2a4f81d6d9b39961eac\") " pod="kube-system/kube-apiserver-localhost" May 14 23:52:00.555141 kubelet[2749]: I0514 23:52:00.555086 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 14 23:52:00.555141 kubelet[2749]: I0514 23:52:00.555111 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 14 23:52:00.555141 kubelet[2749]: I0514 23:52:00.555132 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 14 23:52:00.555385 kubelet[2749]: I0514 23:52:00.555152 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 14 23:52:00.555385 kubelet[2749]: I0514 23:52:00.555171 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/5c0805abca5da2a4f81d6d9b39961eac-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"5c0805abca5da2a4f81d6d9b39961eac\") " pod="kube-system/kube-apiserver-localhost" May 14 23:52:00.555385 kubelet[2749]: I0514 23:52:00.555249 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/5c0805abca5da2a4f81d6d9b39961eac-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"5c0805abca5da2a4f81d6d9b39961eac\") " pod="kube-system/kube-apiserver-localhost" May 14 23:52:00.555385 kubelet[2749]: I0514 23:52:00.555291 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 14 23:52:00.560393 kubelet[2749]: I0514 23:52:00.560348 2749 kubelet_node_status.go:112] "Node was previously registered" node="localhost" May 14 23:52:00.560556 kubelet[2749]: I0514 23:52:00.560433 2749 kubelet_node_status.go:76] "Successfully registered node" node="localhost" May 14 23:52:01.347340 kubelet[2749]: I0514 23:52:01.347064 2749 apiserver.go:52] "Watching apiserver" May 14 23:52:01.353942 kubelet[2749]: I0514 23:52:01.353871 2749 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" May 14 23:52:01.414343 kubelet[2749]: E0514 23:52:01.413148 2749 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" May 14 23:52:01.441027 kubelet[2749]: I0514 23:52:01.440931 2749 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=1.440907175 podStartE2EDuration="1.440907175s" podCreationTimestamp="2025-05-14 23:52:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-14 23:52:01.433166877 +0000 UTC m=+1.152397657" watchObservedRunningTime="2025-05-14 23:52:01.440907175 +0000 UTC m=+1.160137946" May 14 23:52:01.451334 kubelet[2749]: I0514 23:52:01.450081 2749 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.450063651 podStartE2EDuration="1.450063651s" podCreationTimestamp="2025-05-14 23:52:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-14 23:52:01.441222967 +0000 UTC m=+1.160453737" watchObservedRunningTime="2025-05-14 23:52:01.450063651 +0000 UTC m=+1.169294421" May 14 23:52:01.451334 kubelet[2749]: I0514 23:52:01.450178 2749 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.450175445 podStartE2EDuration="1.450175445s" podCreationTimestamp="2025-05-14 23:52:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-14 23:52:01.449817232 +0000 UTC m=+1.169048002" watchObservedRunningTime="2025-05-14 23:52:01.450175445 +0000 UTC m=+1.169406215" May 14 23:52:05.101438 sudo[1706]: pam_unix(sudo:session): session closed for user root May 14 23:52:05.103599 sshd[1705]: Connection closed by 10.0.0.1 port 57002 May 14 23:52:05.104530 sshd-session[1702]: pam_unix(sshd:session): session closed for user core May 14 23:52:05.110582 systemd[1]: sshd@6-10.0.0.142:22-10.0.0.1:57002.service: Deactivated successfully. May 14 23:52:05.113460 systemd[1]: session-7.scope: Deactivated successfully. May 14 23:52:05.114387 systemd[1]: session-7.scope: Consumed 5.425s CPU time, 231.3M memory peak. May 14 23:52:05.116829 systemd-logind[1487]: Session 7 logged out. Waiting for processes to exit. May 14 23:52:05.117982 systemd-logind[1487]: Removed session 7. May 14 23:52:10.329254 update_engine[1492]: I20250514 23:52:10.329139 1492 update_attempter.cc:509] Updating boot flags... May 14 23:52:10.431698 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 41 scanned by (udev-worker) (2844) May 14 23:52:10.484339 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 41 scanned by (udev-worker) (2841) May 14 23:52:13.126322 kubelet[2749]: I0514 23:52:13.126132 2749 kuberuntime_manager.go:1523] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" May 14 23:52:13.129340 containerd[1510]: time="2025-05-14T23:52:13.127631419Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." May 14 23:52:13.129875 kubelet[2749]: I0514 23:52:13.128117 2749 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" May 14 23:52:13.930220 kubelet[2749]: I0514 23:52:13.930132 2749 topology_manager.go:215] "Topology Admit Handler" podUID="274277db-d244-4ab0-90d6-186fe1aaecaf" podNamespace="kube-system" podName="kube-proxy-gj9j5" May 14 23:52:13.937903 systemd[1]: Created slice kubepods-besteffort-pod274277db_d244_4ab0_90d6_186fe1aaecaf.slice - libcontainer container kubepods-besteffort-pod274277db_d244_4ab0_90d6_186fe1aaecaf.slice. May 14 23:52:13.979523 kubelet[2749]: I0514 23:52:13.979449 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/274277db-d244-4ab0-90d6-186fe1aaecaf-lib-modules\") pod \"kube-proxy-gj9j5\" (UID: \"274277db-d244-4ab0-90d6-186fe1aaecaf\") " pod="kube-system/kube-proxy-gj9j5" May 14 23:52:13.979523 kubelet[2749]: I0514 23:52:13.979510 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/274277db-d244-4ab0-90d6-186fe1aaecaf-kube-proxy\") pod \"kube-proxy-gj9j5\" (UID: \"274277db-d244-4ab0-90d6-186fe1aaecaf\") " pod="kube-system/kube-proxy-gj9j5" May 14 23:52:13.979523 kubelet[2749]: I0514 23:52:13.979540 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/274277db-d244-4ab0-90d6-186fe1aaecaf-xtables-lock\") pod \"kube-proxy-gj9j5\" (UID: \"274277db-d244-4ab0-90d6-186fe1aaecaf\") " pod="kube-system/kube-proxy-gj9j5" May 14 23:52:13.979790 kubelet[2749]: I0514 23:52:13.979589 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggztm\" (UniqueName: \"kubernetes.io/projected/274277db-d244-4ab0-90d6-186fe1aaecaf-kube-api-access-ggztm\") pod \"kube-proxy-gj9j5\" (UID: \"274277db-d244-4ab0-90d6-186fe1aaecaf\") " pod="kube-system/kube-proxy-gj9j5" May 14 23:52:14.104870 kubelet[2749]: E0514 23:52:14.104814 2749 projected.go:294] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found May 14 23:52:14.104870 kubelet[2749]: E0514 23:52:14.104862 2749 projected.go:200] Error preparing data for projected volume kube-api-access-ggztm for pod kube-system/kube-proxy-gj9j5: configmap "kube-root-ca.crt" not found May 14 23:52:14.105076 kubelet[2749]: E0514 23:52:14.104942 2749 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/274277db-d244-4ab0-90d6-186fe1aaecaf-kube-api-access-ggztm podName:274277db-d244-4ab0-90d6-186fe1aaecaf nodeName:}" failed. No retries permitted until 2025-05-14 23:52:14.604917868 +0000 UTC m=+14.324148638 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-ggztm" (UniqueName: "kubernetes.io/projected/274277db-d244-4ab0-90d6-186fe1aaecaf-kube-api-access-ggztm") pod "kube-proxy-gj9j5" (UID: "274277db-d244-4ab0-90d6-186fe1aaecaf") : configmap "kube-root-ca.crt" not found May 14 23:52:14.294190 kubelet[2749]: I0514 23:52:14.293467 2749 topology_manager.go:215] "Topology Admit Handler" podUID="227bb07c-b4a1-48fd-8276-76d883e0594c" podNamespace="tigera-operator" podName="tigera-operator-797db67f8-rgzlp" May 14 23:52:14.296707 kubelet[2749]: W0514 23:52:14.296624 2749 reflector.go:547] object-"tigera-operator"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:localhost" cannot list resource "configmaps" in API group "" in the namespace "tigera-operator": no relationship found between node 'localhost' and this object May 14 23:52:14.296707 kubelet[2749]: E0514 23:52:14.296693 2749 reflector.go:150] object-"tigera-operator"/"kube-root-ca.crt": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:localhost" cannot list resource "configmaps" in API group "" in the namespace "tigera-operator": no relationship found between node 'localhost' and this object May 14 23:52:14.296920 kubelet[2749]: W0514 23:52:14.296736 2749 reflector.go:547] object-"tigera-operator"/"kubernetes-services-endpoint": failed to list *v1.ConfigMap: configmaps "kubernetes-services-endpoint" is forbidden: User "system:node:localhost" cannot list resource "configmaps" in API group "" in the namespace "tigera-operator": no relationship found between node 'localhost' and this object May 14 23:52:14.296920 kubelet[2749]: E0514 23:52:14.296757 2749 reflector.go:150] object-"tigera-operator"/"kubernetes-services-endpoint": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps "kubernetes-services-endpoint" is forbidden: User "system:node:localhost" cannot list resource "configmaps" in API group "" in the namespace "tigera-operator": no relationship found between node 'localhost' and this object May 14 23:52:14.303902 systemd[1]: Created slice kubepods-besteffort-pod227bb07c_b4a1_48fd_8276_76d883e0594c.slice - libcontainer container kubepods-besteffort-pod227bb07c_b4a1_48fd_8276_76d883e0594c.slice. May 14 23:52:14.382992 kubelet[2749]: I0514 23:52:14.382925 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckgtt\" (UniqueName: \"kubernetes.io/projected/227bb07c-b4a1-48fd-8276-76d883e0594c-kube-api-access-ckgtt\") pod \"tigera-operator-797db67f8-rgzlp\" (UID: \"227bb07c-b4a1-48fd-8276-76d883e0594c\") " pod="tigera-operator/tigera-operator-797db67f8-rgzlp" May 14 23:52:14.382992 kubelet[2749]: I0514 23:52:14.382990 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/227bb07c-b4a1-48fd-8276-76d883e0594c-var-lib-calico\") pod \"tigera-operator-797db67f8-rgzlp\" (UID: \"227bb07c-b4a1-48fd-8276-76d883e0594c\") " pod="tigera-operator/tigera-operator-797db67f8-rgzlp" May 14 23:52:14.850366 containerd[1510]: time="2025-05-14T23:52:14.850285759Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-gj9j5,Uid:274277db-d244-4ab0-90d6-186fe1aaecaf,Namespace:kube-system,Attempt:0,}" May 14 23:52:14.899070 containerd[1510]: time="2025-05-14T23:52:14.899008418Z" level=info msg="connecting to shim 226325b6f2d804b348567d249f578c1051afe4836e47693d9bfde1fd1c378e07" address="unix:///run/containerd/s/111983cc8311998bb65f18da68763937643fc44d86d1593cbc047af94dd31cde" namespace=k8s.io protocol=ttrpc version=3 May 14 23:52:14.946477 systemd[1]: Started cri-containerd-226325b6f2d804b348567d249f578c1051afe4836e47693d9bfde1fd1c378e07.scope - libcontainer container 226325b6f2d804b348567d249f578c1051afe4836e47693d9bfde1fd1c378e07. May 14 23:52:14.973866 containerd[1510]: time="2025-05-14T23:52:14.973818289Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-gj9j5,Uid:274277db-d244-4ab0-90d6-186fe1aaecaf,Namespace:kube-system,Attempt:0,} returns sandbox id \"226325b6f2d804b348567d249f578c1051afe4836e47693d9bfde1fd1c378e07\"" May 14 23:52:14.976282 containerd[1510]: time="2025-05-14T23:52:14.976259127Z" level=info msg="CreateContainer within sandbox \"226325b6f2d804b348567d249f578c1051afe4836e47693d9bfde1fd1c378e07\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" May 14 23:52:14.989102 containerd[1510]: time="2025-05-14T23:52:14.989050027Z" level=info msg="Container cf637163783f840f4bf7ceb0c9fab20b2666925337acb161d99e5344b08d4357: CDI devices from CRI Config.CDIDevices: []" May 14 23:52:14.998737 containerd[1510]: time="2025-05-14T23:52:14.998681501Z" level=info msg="CreateContainer within sandbox \"226325b6f2d804b348567d249f578c1051afe4836e47693d9bfde1fd1c378e07\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"cf637163783f840f4bf7ceb0c9fab20b2666925337acb161d99e5344b08d4357\"" May 14 23:52:14.999335 containerd[1510]: time="2025-05-14T23:52:14.999307453Z" level=info msg="StartContainer for \"cf637163783f840f4bf7ceb0c9fab20b2666925337acb161d99e5344b08d4357\"" May 14 23:52:15.000883 containerd[1510]: time="2025-05-14T23:52:15.000853461Z" level=info msg="connecting to shim cf637163783f840f4bf7ceb0c9fab20b2666925337acb161d99e5344b08d4357" address="unix:///run/containerd/s/111983cc8311998bb65f18da68763937643fc44d86d1593cbc047af94dd31cde" protocol=ttrpc version=3 May 14 23:52:15.023451 systemd[1]: Started cri-containerd-cf637163783f840f4bf7ceb0c9fab20b2666925337acb161d99e5344b08d4357.scope - libcontainer container cf637163783f840f4bf7ceb0c9fab20b2666925337acb161d99e5344b08d4357. May 14 23:52:15.075343 containerd[1510]: time="2025-05-14T23:52:15.075273960Z" level=info msg="StartContainer for \"cf637163783f840f4bf7ceb0c9fab20b2666925337acb161d99e5344b08d4357\" returns successfully" May 14 23:52:15.489476 kubelet[2749]: E0514 23:52:15.489411 2749 projected.go:294] Couldn't get configMap tigera-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition May 14 23:52:15.489476 kubelet[2749]: E0514 23:52:15.489457 2749 projected.go:200] Error preparing data for projected volume kube-api-access-ckgtt for pod tigera-operator/tigera-operator-797db67f8-rgzlp: failed to sync configmap cache: timed out waiting for the condition May 14 23:52:15.490043 kubelet[2749]: E0514 23:52:15.489530 2749 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/227bb07c-b4a1-48fd-8276-76d883e0594c-kube-api-access-ckgtt podName:227bb07c-b4a1-48fd-8276-76d883e0594c nodeName:}" failed. No retries permitted until 2025-05-14 23:52:15.989504933 +0000 UTC m=+15.708735723 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-ckgtt" (UniqueName: "kubernetes.io/projected/227bb07c-b4a1-48fd-8276-76d883e0594c-kube-api-access-ckgtt") pod "tigera-operator-797db67f8-rgzlp" (UID: "227bb07c-b4a1-48fd-8276-76d883e0594c") : failed to sync configmap cache: timed out waiting for the condition May 14 23:52:16.108847 containerd[1510]: time="2025-05-14T23:52:16.108759927Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-797db67f8-rgzlp,Uid:227bb07c-b4a1-48fd-8276-76d883e0594c,Namespace:tigera-operator,Attempt:0,}" May 14 23:52:16.245944 containerd[1510]: time="2025-05-14T23:52:16.245869035Z" level=info msg="connecting to shim 46420c588cdb29b941aaa0ef276aa0927d8bf1baa4178cccb897552c8cc3b93c" address="unix:///run/containerd/s/0bff6fdd6281fe856771a1b8a8e77574762ff0560e0c093e13ce926e1b927506" namespace=k8s.io protocol=ttrpc version=3 May 14 23:52:16.279510 systemd[1]: Started cri-containerd-46420c588cdb29b941aaa0ef276aa0927d8bf1baa4178cccb897552c8cc3b93c.scope - libcontainer container 46420c588cdb29b941aaa0ef276aa0927d8bf1baa4178cccb897552c8cc3b93c. May 14 23:52:16.337324 containerd[1510]: time="2025-05-14T23:52:16.337254146Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-797db67f8-rgzlp,Uid:227bb07c-b4a1-48fd-8276-76d883e0594c,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"46420c588cdb29b941aaa0ef276aa0927d8bf1baa4178cccb897552c8cc3b93c\"" May 14 23:52:16.342914 containerd[1510]: time="2025-05-14T23:52:16.342840123Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.36.7\"" May 14 23:52:18.838168 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount83945674.mount: Deactivated successfully. May 14 23:52:19.442698 containerd[1510]: time="2025-05-14T23:52:19.442622606Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.36.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:52:19.445275 containerd[1510]: time="2025-05-14T23:52:19.445185446Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.36.7: active requests=0, bytes read=22002662" May 14 23:52:19.447330 containerd[1510]: time="2025-05-14T23:52:19.447254506Z" level=info msg="ImageCreate event name:\"sha256:e9b19fa62f476f04e5840eb65a0f71b49c7b9f4ceede31675409ddc218bb5578\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:52:19.451672 containerd[1510]: time="2025-05-14T23:52:19.451598682Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:a4a44422d8f2a14e0aaea2031ccb5580f2bf68218c9db444450c1888743305e9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:52:19.452149 containerd[1510]: time="2025-05-14T23:52:19.452087103Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.36.7\" with image id \"sha256:e9b19fa62f476f04e5840eb65a0f71b49c7b9f4ceede31675409ddc218bb5578\", repo tag \"quay.io/tigera/operator:v1.36.7\", repo digest \"quay.io/tigera/operator@sha256:a4a44422d8f2a14e0aaea2031ccb5580f2bf68218c9db444450c1888743305e9\", size \"21998657\" in 3.109184592s" May 14 23:52:19.452149 containerd[1510]: time="2025-05-14T23:52:19.452123912Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.36.7\" returns image reference \"sha256:e9b19fa62f476f04e5840eb65a0f71b49c7b9f4ceede31675409ddc218bb5578\"" May 14 23:52:19.454965 containerd[1510]: time="2025-05-14T23:52:19.454924140Z" level=info msg="CreateContainer within sandbox \"46420c588cdb29b941aaa0ef276aa0927d8bf1baa4178cccb897552c8cc3b93c\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" May 14 23:52:19.488397 containerd[1510]: time="2025-05-14T23:52:19.488339945Z" level=info msg="Container 811f242e79ac47dcf2bf5a5bfa48637b2805c93969242cfa9da1e2f13d218fbb: CDI devices from CRI Config.CDIDevices: []" May 14 23:52:19.510445 containerd[1510]: time="2025-05-14T23:52:19.510394721Z" level=info msg="CreateContainer within sandbox \"46420c588cdb29b941aaa0ef276aa0927d8bf1baa4178cccb897552c8cc3b93c\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"811f242e79ac47dcf2bf5a5bfa48637b2805c93969242cfa9da1e2f13d218fbb\"" May 14 23:52:19.510799 containerd[1510]: time="2025-05-14T23:52:19.510770478Z" level=info msg="StartContainer for \"811f242e79ac47dcf2bf5a5bfa48637b2805c93969242cfa9da1e2f13d218fbb\"" May 14 23:52:19.511893 containerd[1510]: time="2025-05-14T23:52:19.511824716Z" level=info msg="connecting to shim 811f242e79ac47dcf2bf5a5bfa48637b2805c93969242cfa9da1e2f13d218fbb" address="unix:///run/containerd/s/0bff6fdd6281fe856771a1b8a8e77574762ff0560e0c093e13ce926e1b927506" protocol=ttrpc version=3 May 14 23:52:19.541662 systemd[1]: Started cri-containerd-811f242e79ac47dcf2bf5a5bfa48637b2805c93969242cfa9da1e2f13d218fbb.scope - libcontainer container 811f242e79ac47dcf2bf5a5bfa48637b2805c93969242cfa9da1e2f13d218fbb. May 14 23:52:19.579292 containerd[1510]: time="2025-05-14T23:52:19.578967694Z" level=info msg="StartContainer for \"811f242e79ac47dcf2bf5a5bfa48637b2805c93969242cfa9da1e2f13d218fbb\" returns successfully" May 14 23:52:20.442932 kubelet[2749]: I0514 23:52:20.442848 2749 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-gj9j5" podStartSLOduration=7.442824177 podStartE2EDuration="7.442824177s" podCreationTimestamp="2025-05-14 23:52:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-14 23:52:15.433515317 +0000 UTC m=+15.152746087" watchObservedRunningTime="2025-05-14 23:52:20.442824177 +0000 UTC m=+20.162054937" May 14 23:52:22.425289 kubelet[2749]: I0514 23:52:22.425133 2749 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-797db67f8-rgzlp" podStartSLOduration=5.310542886 podStartE2EDuration="8.425060087s" podCreationTimestamp="2025-05-14 23:52:14 +0000 UTC" firstStartedPulling="2025-05-14 23:52:16.338719589 +0000 UTC m=+16.057950359" lastFinishedPulling="2025-05-14 23:52:19.45323678 +0000 UTC m=+19.172467560" observedRunningTime="2025-05-14 23:52:20.443058097 +0000 UTC m=+20.162288867" watchObservedRunningTime="2025-05-14 23:52:22.425060087 +0000 UTC m=+22.144290857" May 14 23:52:22.428194 kubelet[2749]: I0514 23:52:22.427226 2749 topology_manager.go:215] "Topology Admit Handler" podUID="9c23eb35-5980-4b22-b637-416dfdc9a74c" podNamespace="calico-system" podName="calico-typha-7c6786c468-hlrct" May 14 23:52:22.449852 systemd[1]: Created slice kubepods-besteffort-pod9c23eb35_5980_4b22_b637_416dfdc9a74c.slice - libcontainer container kubepods-besteffort-pod9c23eb35_5980_4b22_b637_416dfdc9a74c.slice. May 14 23:52:22.532630 kubelet[2749]: I0514 23:52:22.532567 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/9c23eb35-5980-4b22-b637-416dfdc9a74c-typha-certs\") pod \"calico-typha-7c6786c468-hlrct\" (UID: \"9c23eb35-5980-4b22-b637-416dfdc9a74c\") " pod="calico-system/calico-typha-7c6786c468-hlrct" May 14 23:52:22.532630 kubelet[2749]: I0514 23:52:22.532631 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9c23eb35-5980-4b22-b637-416dfdc9a74c-tigera-ca-bundle\") pod \"calico-typha-7c6786c468-hlrct\" (UID: \"9c23eb35-5980-4b22-b637-416dfdc9a74c\") " pod="calico-system/calico-typha-7c6786c468-hlrct" May 14 23:52:22.532881 kubelet[2749]: I0514 23:52:22.532659 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlcfn\" (UniqueName: \"kubernetes.io/projected/9c23eb35-5980-4b22-b637-416dfdc9a74c-kube-api-access-hlcfn\") pod \"calico-typha-7c6786c468-hlrct\" (UID: \"9c23eb35-5980-4b22-b637-416dfdc9a74c\") " pod="calico-system/calico-typha-7c6786c468-hlrct" May 14 23:52:22.560009 kubelet[2749]: I0514 23:52:22.559934 2749 topology_manager.go:215] "Topology Admit Handler" podUID="5a918280-6248-4ddf-a32f-be6fcc2b4883" podNamespace="calico-system" podName="calico-node-m5llq" May 14 23:52:22.572077 systemd[1]: Created slice kubepods-besteffort-pod5a918280_6248_4ddf_a32f_be6fcc2b4883.slice - libcontainer container kubepods-besteffort-pod5a918280_6248_4ddf_a32f_be6fcc2b4883.slice. May 14 23:52:22.633694 kubelet[2749]: I0514 23:52:22.633598 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/5a918280-6248-4ddf-a32f-be6fcc2b4883-var-run-calico\") pod \"calico-node-m5llq\" (UID: \"5a918280-6248-4ddf-a32f-be6fcc2b4883\") " pod="calico-system/calico-node-m5llq" May 14 23:52:22.633694 kubelet[2749]: I0514 23:52:22.633657 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/5a918280-6248-4ddf-a32f-be6fcc2b4883-var-lib-calico\") pod \"calico-node-m5llq\" (UID: \"5a918280-6248-4ddf-a32f-be6fcc2b4883\") " pod="calico-system/calico-node-m5llq" May 14 23:52:22.633694 kubelet[2749]: I0514 23:52:22.633679 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/5a918280-6248-4ddf-a32f-be6fcc2b4883-cni-bin-dir\") pod \"calico-node-m5llq\" (UID: \"5a918280-6248-4ddf-a32f-be6fcc2b4883\") " pod="calico-system/calico-node-m5llq" May 14 23:52:22.633694 kubelet[2749]: I0514 23:52:22.633702 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5a918280-6248-4ddf-a32f-be6fcc2b4883-lib-modules\") pod \"calico-node-m5llq\" (UID: \"5a918280-6248-4ddf-a32f-be6fcc2b4883\") " pod="calico-system/calico-node-m5llq" May 14 23:52:22.634010 kubelet[2749]: I0514 23:52:22.633743 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qjpb\" (UniqueName: \"kubernetes.io/projected/5a918280-6248-4ddf-a32f-be6fcc2b4883-kube-api-access-6qjpb\") pod \"calico-node-m5llq\" (UID: \"5a918280-6248-4ddf-a32f-be6fcc2b4883\") " pod="calico-system/calico-node-m5llq" May 14 23:52:22.634010 kubelet[2749]: I0514 23:52:22.633766 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/5a918280-6248-4ddf-a32f-be6fcc2b4883-xtables-lock\") pod \"calico-node-m5llq\" (UID: \"5a918280-6248-4ddf-a32f-be6fcc2b4883\") " pod="calico-system/calico-node-m5llq" May 14 23:52:22.634010 kubelet[2749]: I0514 23:52:22.633785 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5a918280-6248-4ddf-a32f-be6fcc2b4883-tigera-ca-bundle\") pod \"calico-node-m5llq\" (UID: \"5a918280-6248-4ddf-a32f-be6fcc2b4883\") " pod="calico-system/calico-node-m5llq" May 14 23:52:22.634010 kubelet[2749]: I0514 23:52:22.633804 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/5a918280-6248-4ddf-a32f-be6fcc2b4883-cni-net-dir\") pod \"calico-node-m5llq\" (UID: \"5a918280-6248-4ddf-a32f-be6fcc2b4883\") " pod="calico-system/calico-node-m5llq" May 14 23:52:22.634010 kubelet[2749]: I0514 23:52:22.633823 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/5a918280-6248-4ddf-a32f-be6fcc2b4883-policysync\") pod \"calico-node-m5llq\" (UID: \"5a918280-6248-4ddf-a32f-be6fcc2b4883\") " pod="calico-system/calico-node-m5llq" May 14 23:52:22.634189 kubelet[2749]: I0514 23:52:22.633856 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/5a918280-6248-4ddf-a32f-be6fcc2b4883-node-certs\") pod \"calico-node-m5llq\" (UID: \"5a918280-6248-4ddf-a32f-be6fcc2b4883\") " pod="calico-system/calico-node-m5llq" May 14 23:52:22.634189 kubelet[2749]: I0514 23:52:22.633876 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/5a918280-6248-4ddf-a32f-be6fcc2b4883-cni-log-dir\") pod \"calico-node-m5llq\" (UID: \"5a918280-6248-4ddf-a32f-be6fcc2b4883\") " pod="calico-system/calico-node-m5llq" May 14 23:52:22.634189 kubelet[2749]: I0514 23:52:22.633895 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/5a918280-6248-4ddf-a32f-be6fcc2b4883-flexvol-driver-host\") pod \"calico-node-m5llq\" (UID: \"5a918280-6248-4ddf-a32f-be6fcc2b4883\") " pod="calico-system/calico-node-m5llq" May 14 23:52:22.669428 kubelet[2749]: I0514 23:52:22.665883 2749 topology_manager.go:215] "Topology Admit Handler" podUID="8f0b6425-d6bf-493f-ada4-58a2a3096286" podNamespace="calico-system" podName="csi-node-driver-96wg8" May 14 23:52:22.669428 kubelet[2749]: E0514 23:52:22.667847 2749 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-96wg8" podUID="8f0b6425-d6bf-493f-ada4-58a2a3096286" May 14 23:52:22.735692 kubelet[2749]: I0514 23:52:22.734804 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/8f0b6425-d6bf-493f-ada4-58a2a3096286-varrun\") pod \"csi-node-driver-96wg8\" (UID: \"8f0b6425-d6bf-493f-ada4-58a2a3096286\") " pod="calico-system/csi-node-driver-96wg8" May 14 23:52:22.735692 kubelet[2749]: I0514 23:52:22.734869 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/8f0b6425-d6bf-493f-ada4-58a2a3096286-registration-dir\") pod \"csi-node-driver-96wg8\" (UID: \"8f0b6425-d6bf-493f-ada4-58a2a3096286\") " pod="calico-system/csi-node-driver-96wg8" May 14 23:52:22.735692 kubelet[2749]: I0514 23:52:22.734895 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5vk5\" (UniqueName: \"kubernetes.io/projected/8f0b6425-d6bf-493f-ada4-58a2a3096286-kube-api-access-q5vk5\") pod \"csi-node-driver-96wg8\" (UID: \"8f0b6425-d6bf-493f-ada4-58a2a3096286\") " pod="calico-system/csi-node-driver-96wg8" May 14 23:52:22.735692 kubelet[2749]: I0514 23:52:22.735001 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8f0b6425-d6bf-493f-ada4-58a2a3096286-kubelet-dir\") pod \"csi-node-driver-96wg8\" (UID: \"8f0b6425-d6bf-493f-ada4-58a2a3096286\") " pod="calico-system/csi-node-driver-96wg8" May 14 23:52:22.735692 kubelet[2749]: I0514 23:52:22.735073 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/8f0b6425-d6bf-493f-ada4-58a2a3096286-socket-dir\") pod \"csi-node-driver-96wg8\" (UID: \"8f0b6425-d6bf-493f-ada4-58a2a3096286\") " pod="calico-system/csi-node-driver-96wg8" May 14 23:52:22.741558 kubelet[2749]: E0514 23:52:22.741472 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:22.741727 kubelet[2749]: W0514 23:52:22.741705 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:22.741810 kubelet[2749]: E0514 23:52:22.741793 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:22.745473 kubelet[2749]: E0514 23:52:22.745420 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:22.745473 kubelet[2749]: W0514 23:52:22.745456 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:22.745640 kubelet[2749]: E0514 23:52:22.745484 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:22.752992 kubelet[2749]: E0514 23:52:22.752952 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:22.752992 kubelet[2749]: W0514 23:52:22.752979 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:22.752992 kubelet[2749]: E0514 23:52:22.753005 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:22.757106 containerd[1510]: time="2025-05-14T23:52:22.756837991Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-7c6786c468-hlrct,Uid:9c23eb35-5980-4b22-b637-416dfdc9a74c,Namespace:calico-system,Attempt:0,}" May 14 23:52:22.796015 containerd[1510]: time="2025-05-14T23:52:22.795898193Z" level=info msg="connecting to shim 208fc061d43c5c87c2ebe9c1672ef033d91d73ae1f549fd3b37c69bcc887036f" address="unix:///run/containerd/s/e4d70620dafce04d7c136327f3b4446b69ab71de516cdb48b726d1afaa272ace" namespace=k8s.io protocol=ttrpc version=3 May 14 23:52:22.829531 systemd[1]: Started cri-containerd-208fc061d43c5c87c2ebe9c1672ef033d91d73ae1f549fd3b37c69bcc887036f.scope - libcontainer container 208fc061d43c5c87c2ebe9c1672ef033d91d73ae1f549fd3b37c69bcc887036f. May 14 23:52:22.836209 kubelet[2749]: E0514 23:52:22.836161 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:22.836209 kubelet[2749]: W0514 23:52:22.836187 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:22.836209 kubelet[2749]: E0514 23:52:22.836212 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:22.836611 kubelet[2749]: E0514 23:52:22.836594 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:22.836611 kubelet[2749]: W0514 23:52:22.836607 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:22.836723 kubelet[2749]: E0514 23:52:22.836622 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:22.837002 kubelet[2749]: E0514 23:52:22.836926 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:22.837002 kubelet[2749]: W0514 23:52:22.836938 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:22.837002 kubelet[2749]: E0514 23:52:22.836950 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:22.837437 kubelet[2749]: E0514 23:52:22.837406 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:22.837437 kubelet[2749]: W0514 23:52:22.837418 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:22.837437 kubelet[2749]: E0514 23:52:22.837431 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:22.838012 kubelet[2749]: E0514 23:52:22.837983 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:22.838012 kubelet[2749]: W0514 23:52:22.837996 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:22.838127 kubelet[2749]: E0514 23:52:22.838076 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:22.838236 kubelet[2749]: E0514 23:52:22.838212 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:22.838236 kubelet[2749]: W0514 23:52:22.838223 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:22.838400 kubelet[2749]: E0514 23:52:22.838369 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:22.839107 kubelet[2749]: E0514 23:52:22.838565 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:22.839107 kubelet[2749]: W0514 23:52:22.838576 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:22.839107 kubelet[2749]: E0514 23:52:22.838671 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:22.839107 kubelet[2749]: E0514 23:52:22.838827 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:22.839107 kubelet[2749]: W0514 23:52:22.838835 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:22.839107 kubelet[2749]: E0514 23:52:22.838918 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:22.839107 kubelet[2749]: E0514 23:52:22.839050 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:22.839107 kubelet[2749]: W0514 23:52:22.839056 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:22.839440 kubelet[2749]: E0514 23:52:22.839135 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:22.839440 kubelet[2749]: E0514 23:52:22.839247 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:22.839440 kubelet[2749]: W0514 23:52:22.839253 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:22.839440 kubelet[2749]: E0514 23:52:22.839266 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:22.839581 kubelet[2749]: E0514 23:52:22.839572 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:22.839620 kubelet[2749]: W0514 23:52:22.839581 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:22.839620 kubelet[2749]: E0514 23:52:22.839604 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:22.840190 kubelet[2749]: E0514 23:52:22.839862 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:22.840190 kubelet[2749]: W0514 23:52:22.839874 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:22.840190 kubelet[2749]: E0514 23:52:22.840010 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:22.843506 kubelet[2749]: E0514 23:52:22.843468 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:22.843506 kubelet[2749]: W0514 23:52:22.843485 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:22.843715 kubelet[2749]: E0514 23:52:22.843613 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:22.843762 kubelet[2749]: E0514 23:52:22.843734 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:22.843762 kubelet[2749]: W0514 23:52:22.843742 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:22.843855 kubelet[2749]: E0514 23:52:22.843833 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:22.843965 kubelet[2749]: E0514 23:52:22.843943 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:22.843965 kubelet[2749]: W0514 23:52:22.843955 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:22.844072 kubelet[2749]: E0514 23:52:22.844050 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:22.844205 kubelet[2749]: E0514 23:52:22.844177 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:22.844205 kubelet[2749]: W0514 23:52:22.844188 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:22.844382 kubelet[2749]: E0514 23:52:22.844356 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:22.844486 kubelet[2749]: E0514 23:52:22.844466 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:22.844486 kubelet[2749]: W0514 23:52:22.844478 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:22.844605 kubelet[2749]: E0514 23:52:22.844585 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:22.844775 kubelet[2749]: E0514 23:52:22.844745 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:22.844775 kubelet[2749]: W0514 23:52:22.844765 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:22.844848 kubelet[2749]: E0514 23:52:22.844782 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:22.845126 kubelet[2749]: E0514 23:52:22.845098 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:22.845126 kubelet[2749]: W0514 23:52:22.845116 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:22.845247 kubelet[2749]: E0514 23:52:22.845218 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:22.845586 kubelet[2749]: E0514 23:52:22.845548 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:22.845586 kubelet[2749]: W0514 23:52:22.845580 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:22.845731 kubelet[2749]: E0514 23:52:22.845706 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:22.845889 kubelet[2749]: E0514 23:52:22.845853 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:22.845889 kubelet[2749]: W0514 23:52:22.845885 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:22.845998 kubelet[2749]: E0514 23:52:22.845977 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:22.846382 kubelet[2749]: E0514 23:52:22.846357 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:22.846382 kubelet[2749]: W0514 23:52:22.846376 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:22.846487 kubelet[2749]: E0514 23:52:22.846467 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:22.846835 kubelet[2749]: E0514 23:52:22.846788 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:22.846835 kubelet[2749]: W0514 23:52:22.846809 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:22.846911 kubelet[2749]: E0514 23:52:22.846839 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:22.847225 kubelet[2749]: E0514 23:52:22.847187 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:22.847225 kubelet[2749]: W0514 23:52:22.847220 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:22.847317 kubelet[2749]: E0514 23:52:22.847237 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:22.847807 kubelet[2749]: E0514 23:52:22.847784 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:22.847807 kubelet[2749]: W0514 23:52:22.847804 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:22.847874 kubelet[2749]: E0514 23:52:22.847817 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:22.858065 kubelet[2749]: E0514 23:52:22.858011 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:22.858065 kubelet[2749]: W0514 23:52:22.858039 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:22.858065 kubelet[2749]: E0514 23:52:22.858067 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:22.878178 containerd[1510]: time="2025-05-14T23:52:22.878107548Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-m5llq,Uid:5a918280-6248-4ddf-a32f-be6fcc2b4883,Namespace:calico-system,Attempt:0,}" May 14 23:52:23.925466 containerd[1510]: time="2025-05-14T23:52:23.925156569Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-7c6786c468-hlrct,Uid:9c23eb35-5980-4b22-b637-416dfdc9a74c,Namespace:calico-system,Attempt:0,} returns sandbox id \"208fc061d43c5c87c2ebe9c1672ef033d91d73ae1f549fd3b37c69bcc887036f\"" May 14 23:52:23.931381 containerd[1510]: time="2025-05-14T23:52:23.931167074Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.29.3\"" May 14 23:52:24.377835 kubelet[2749]: E0514 23:52:24.377744 2749 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-96wg8" podUID="8f0b6425-d6bf-493f-ada4-58a2a3096286" May 14 23:52:25.082372 containerd[1510]: time="2025-05-14T23:52:25.082286754Z" level=info msg="connecting to shim 8f9fa26a1fa44c203c6923f82ed62cd8f77fec5a5b9463691529e42246e6aee7" address="unix:///run/containerd/s/336e40f2e2183dfd0f34d097e5c6cab0f947f7c46bb4579ae03ff572ea0fecda" namespace=k8s.io protocol=ttrpc version=3 May 14 23:52:25.110521 systemd[1]: Started cri-containerd-8f9fa26a1fa44c203c6923f82ed62cd8f77fec5a5b9463691529e42246e6aee7.scope - libcontainer container 8f9fa26a1fa44c203c6923f82ed62cd8f77fec5a5b9463691529e42246e6aee7. May 14 23:52:25.240461 containerd[1510]: time="2025-05-14T23:52:25.240388444Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-m5llq,Uid:5a918280-6248-4ddf-a32f-be6fcc2b4883,Namespace:calico-system,Attempt:0,} returns sandbox id \"8f9fa26a1fa44c203c6923f82ed62cd8f77fec5a5b9463691529e42246e6aee7\"" May 14 23:52:26.377965 kubelet[2749]: E0514 23:52:26.377903 2749 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-96wg8" podUID="8f0b6425-d6bf-493f-ada4-58a2a3096286" May 14 23:52:28.377884 kubelet[2749]: E0514 23:52:28.377691 2749 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-96wg8" podUID="8f0b6425-d6bf-493f-ada4-58a2a3096286" May 14 23:52:29.267418 containerd[1510]: time="2025-05-14T23:52:29.267345493Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:52:29.277276 containerd[1510]: time="2025-05-14T23:52:29.277213950Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.29.3: active requests=0, bytes read=30426870" May 14 23:52:29.279177 containerd[1510]: time="2025-05-14T23:52:29.279126536Z" level=info msg="ImageCreate event name:\"sha256:bde24a3cb8851b59372b76b3ad78f8028d1a915ffed82c6cc6256f34e500bd3d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:52:29.282391 containerd[1510]: time="2025-05-14T23:52:29.282335761Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:f5516aa6a78f00931d2625f3012dcf2c69d141ce41483b8d59c6ec6330a18620\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:52:29.282949 containerd[1510]: time="2025-05-14T23:52:29.282909971Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.29.3\" with image id \"sha256:bde24a3cb8851b59372b76b3ad78f8028d1a915ffed82c6cc6256f34e500bd3d\", repo tag \"ghcr.io/flatcar/calico/typha:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:f5516aa6a78f00931d2625f3012dcf2c69d141ce41483b8d59c6ec6330a18620\", size \"31919484\" in 5.351704093s" May 14 23:52:29.282949 containerd[1510]: time="2025-05-14T23:52:29.282941941Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.29.3\" returns image reference \"sha256:bde24a3cb8851b59372b76b3ad78f8028d1a915ffed82c6cc6256f34e500bd3d\"" May 14 23:52:29.283854 containerd[1510]: time="2025-05-14T23:52:29.283822757Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\"" May 14 23:52:29.292568 containerd[1510]: time="2025-05-14T23:52:29.292235405Z" level=info msg="CreateContainer within sandbox \"208fc061d43c5c87c2ebe9c1672ef033d91d73ae1f549fd3b37c69bcc887036f\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" May 14 23:52:29.303914 containerd[1510]: time="2025-05-14T23:52:29.303862039Z" level=info msg="Container a40e5ed3b7650962e38fa56950b05e12b01bd2bd13deaa804a24ff778dd23e25: CDI devices from CRI Config.CDIDevices: []" May 14 23:52:29.313519 containerd[1510]: time="2025-05-14T23:52:29.313458744Z" level=info msg="CreateContainer within sandbox \"208fc061d43c5c87c2ebe9c1672ef033d91d73ae1f549fd3b37c69bcc887036f\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"a40e5ed3b7650962e38fa56950b05e12b01bd2bd13deaa804a24ff778dd23e25\"" May 14 23:52:29.314228 containerd[1510]: time="2025-05-14T23:52:29.314178197Z" level=info msg="StartContainer for \"a40e5ed3b7650962e38fa56950b05e12b01bd2bd13deaa804a24ff778dd23e25\"" May 14 23:52:29.318325 containerd[1510]: time="2025-05-14T23:52:29.317426935Z" level=info msg="connecting to shim a40e5ed3b7650962e38fa56950b05e12b01bd2bd13deaa804a24ff778dd23e25" address="unix:///run/containerd/s/e4d70620dafce04d7c136327f3b4446b69ab71de516cdb48b726d1afaa272ace" protocol=ttrpc version=3 May 14 23:52:29.352525 systemd[1]: Started cri-containerd-a40e5ed3b7650962e38fa56950b05e12b01bd2bd13deaa804a24ff778dd23e25.scope - libcontainer container a40e5ed3b7650962e38fa56950b05e12b01bd2bd13deaa804a24ff778dd23e25. May 14 23:52:29.411941 containerd[1510]: time="2025-05-14T23:52:29.411873060Z" level=info msg="StartContainer for \"a40e5ed3b7650962e38fa56950b05e12b01bd2bd13deaa804a24ff778dd23e25\" returns successfully" May 14 23:52:29.536125 kubelet[2749]: E0514 23:52:29.535964 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:29.536125 kubelet[2749]: W0514 23:52:29.535995 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:29.536125 kubelet[2749]: E0514 23:52:29.536016 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:29.536908 kubelet[2749]: E0514 23:52:29.536506 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:29.536908 kubelet[2749]: W0514 23:52:29.536592 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:29.536908 kubelet[2749]: E0514 23:52:29.536602 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:29.536908 kubelet[2749]: E0514 23:52:29.536866 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:29.536908 kubelet[2749]: W0514 23:52:29.536874 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:29.536908 kubelet[2749]: E0514 23:52:29.536882 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:29.537148 kubelet[2749]: E0514 23:52:29.537111 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:29.537148 kubelet[2749]: W0514 23:52:29.537132 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:29.537148 kubelet[2749]: E0514 23:52:29.537139 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:29.537522 kubelet[2749]: E0514 23:52:29.537495 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:29.537522 kubelet[2749]: W0514 23:52:29.537508 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:29.537522 kubelet[2749]: E0514 23:52:29.537516 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:29.537732 kubelet[2749]: E0514 23:52:29.537707 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:29.537732 kubelet[2749]: W0514 23:52:29.537719 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:29.537732 kubelet[2749]: E0514 23:52:29.537728 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:29.537928 kubelet[2749]: E0514 23:52:29.537911 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:29.537928 kubelet[2749]: W0514 23:52:29.537922 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:29.537928 kubelet[2749]: E0514 23:52:29.537930 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:29.538140 kubelet[2749]: E0514 23:52:29.538115 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:29.538140 kubelet[2749]: W0514 23:52:29.538128 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:29.538140 kubelet[2749]: E0514 23:52:29.538135 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:29.538491 kubelet[2749]: E0514 23:52:29.538448 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:29.538491 kubelet[2749]: W0514 23:52:29.538481 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:29.538569 kubelet[2749]: E0514 23:52:29.538515 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:29.538876 kubelet[2749]: E0514 23:52:29.538856 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:29.538876 kubelet[2749]: W0514 23:52:29.538869 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:29.538930 kubelet[2749]: E0514 23:52:29.538880 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:29.539153 kubelet[2749]: E0514 23:52:29.539120 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:29.539153 kubelet[2749]: W0514 23:52:29.539131 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:29.539153 kubelet[2749]: E0514 23:52:29.539141 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:29.539427 kubelet[2749]: E0514 23:52:29.539399 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:29.539427 kubelet[2749]: W0514 23:52:29.539421 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:29.539503 kubelet[2749]: E0514 23:52:29.539432 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:29.539665 kubelet[2749]: E0514 23:52:29.539647 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:29.539665 kubelet[2749]: W0514 23:52:29.539657 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:29.539724 kubelet[2749]: E0514 23:52:29.539665 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:29.539896 kubelet[2749]: E0514 23:52:29.539878 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:29.539896 kubelet[2749]: W0514 23:52:29.539889 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:29.539954 kubelet[2749]: E0514 23:52:29.539897 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:29.540115 kubelet[2749]: E0514 23:52:29.540103 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:29.540115 kubelet[2749]: W0514 23:52:29.540114 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:29.540167 kubelet[2749]: E0514 23:52:29.540122 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:29.585675 kubelet[2749]: E0514 23:52:29.585644 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:29.585675 kubelet[2749]: W0514 23:52:29.585665 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:29.585675 kubelet[2749]: E0514 23:52:29.585684 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:29.586030 kubelet[2749]: E0514 23:52:29.586002 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:29.586066 kubelet[2749]: W0514 23:52:29.586045 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:29.586093 kubelet[2749]: E0514 23:52:29.586079 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:29.586437 kubelet[2749]: E0514 23:52:29.586418 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:29.586437 kubelet[2749]: W0514 23:52:29.586434 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:29.586503 kubelet[2749]: E0514 23:52:29.586456 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:29.587031 kubelet[2749]: E0514 23:52:29.586963 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:29.587031 kubelet[2749]: W0514 23:52:29.587010 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:29.587110 kubelet[2749]: E0514 23:52:29.587049 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:29.587389 kubelet[2749]: E0514 23:52:29.587365 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:29.587389 kubelet[2749]: W0514 23:52:29.587382 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:29.587470 kubelet[2749]: E0514 23:52:29.587399 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:29.587657 kubelet[2749]: E0514 23:52:29.587634 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:29.587657 kubelet[2749]: W0514 23:52:29.587649 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:29.587705 kubelet[2749]: E0514 23:52:29.587680 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:29.587885 kubelet[2749]: E0514 23:52:29.587864 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:29.587885 kubelet[2749]: W0514 23:52:29.587879 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:29.587941 kubelet[2749]: E0514 23:52:29.587904 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:29.588099 kubelet[2749]: E0514 23:52:29.588084 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:29.588099 kubelet[2749]: W0514 23:52:29.588097 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:29.588150 kubelet[2749]: E0514 23:52:29.588122 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:29.588377 kubelet[2749]: E0514 23:52:29.588362 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:29.588420 kubelet[2749]: W0514 23:52:29.588376 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:29.588420 kubelet[2749]: E0514 23:52:29.588392 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:29.588669 kubelet[2749]: E0514 23:52:29.588649 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:29.588669 kubelet[2749]: W0514 23:52:29.588662 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:29.588716 kubelet[2749]: E0514 23:52:29.588677 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:29.588885 kubelet[2749]: E0514 23:52:29.588873 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:29.588917 kubelet[2749]: W0514 23:52:29.588884 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:29.588917 kubelet[2749]: E0514 23:52:29.588898 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:29.589144 kubelet[2749]: E0514 23:52:29.589131 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:29.589144 kubelet[2749]: W0514 23:52:29.589142 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:29.589199 kubelet[2749]: E0514 23:52:29.589157 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:29.589591 kubelet[2749]: E0514 23:52:29.589565 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:29.589591 kubelet[2749]: W0514 23:52:29.589582 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:29.589651 kubelet[2749]: E0514 23:52:29.589598 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:29.589829 kubelet[2749]: E0514 23:52:29.589808 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:29.589829 kubelet[2749]: W0514 23:52:29.589820 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:29.589877 kubelet[2749]: E0514 23:52:29.589833 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:29.590058 kubelet[2749]: E0514 23:52:29.590038 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:29.590058 kubelet[2749]: W0514 23:52:29.590054 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:29.590116 kubelet[2749]: E0514 23:52:29.590069 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:29.590283 kubelet[2749]: E0514 23:52:29.590269 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:29.590283 kubelet[2749]: W0514 23:52:29.590281 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:29.590383 kubelet[2749]: E0514 23:52:29.590310 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:29.590589 kubelet[2749]: E0514 23:52:29.590553 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:29.590589 kubelet[2749]: W0514 23:52:29.590577 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:29.590589 kubelet[2749]: E0514 23:52:29.590587 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:29.591067 kubelet[2749]: E0514 23:52:29.591047 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:29.591067 kubelet[2749]: W0514 23:52:29.591058 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:29.591067 kubelet[2749]: E0514 23:52:29.591067 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:29.729340 kubelet[2749]: I0514 23:52:29.728623 2749 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-7c6786c468-hlrct" podStartSLOduration=2.375407395 podStartE2EDuration="7.72860607s" podCreationTimestamp="2025-05-14 23:52:22 +0000 UTC" firstStartedPulling="2025-05-14 23:52:23.930491021 +0000 UTC m=+23.649721801" lastFinishedPulling="2025-05-14 23:52:29.283689706 +0000 UTC m=+29.002920476" observedRunningTime="2025-05-14 23:52:29.724943062 +0000 UTC m=+29.444173833" watchObservedRunningTime="2025-05-14 23:52:29.72860607 +0000 UTC m=+29.447836840" May 14 23:52:30.050142 systemd[1]: Started sshd@7-10.0.0.142:22-10.0.0.1:57844.service - OpenSSH per-connection server daemon (10.0.0.1:57844). May 14 23:52:30.114563 sshd[3371]: Accepted publickey for core from 10.0.0.1 port 57844 ssh2: RSA SHA256:lk2TkYBEL43KPVbrGyh3Ro8IB8NGN6uTNXzFyrYR01I May 14 23:52:30.116507 sshd-session[3371]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 23:52:30.126102 systemd-logind[1487]: New session 8 of user core. May 14 23:52:30.137523 systemd[1]: Started session-8.scope - Session 8 of User core. May 14 23:52:30.267337 sshd[3374]: Connection closed by 10.0.0.1 port 57844 May 14 23:52:30.267692 sshd-session[3371]: pam_unix(sshd:session): session closed for user core May 14 23:52:30.272787 systemd[1]: sshd@7-10.0.0.142:22-10.0.0.1:57844.service: Deactivated successfully. May 14 23:52:30.275599 systemd[1]: session-8.scope: Deactivated successfully. May 14 23:52:30.276673 systemd-logind[1487]: Session 8 logged out. Waiting for processes to exit. May 14 23:52:30.277927 systemd-logind[1487]: Removed session 8. May 14 23:52:30.377735 kubelet[2749]: E0514 23:52:30.377666 2749 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-96wg8" podUID="8f0b6425-d6bf-493f-ada4-58a2a3096286" May 14 23:52:30.546377 kubelet[2749]: E0514 23:52:30.546327 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:30.547178 kubelet[2749]: W0514 23:52:30.546964 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:30.547178 kubelet[2749]: E0514 23:52:30.547006 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:30.547410 kubelet[2749]: E0514 23:52:30.547372 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:30.547410 kubelet[2749]: W0514 23:52:30.547396 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:30.547410 kubelet[2749]: E0514 23:52:30.547412 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:30.547694 kubelet[2749]: E0514 23:52:30.547675 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:30.547694 kubelet[2749]: W0514 23:52:30.547690 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:30.547760 kubelet[2749]: E0514 23:52:30.547700 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:30.547978 kubelet[2749]: E0514 23:52:30.547953 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:30.547978 kubelet[2749]: W0514 23:52:30.547968 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:30.548027 kubelet[2749]: E0514 23:52:30.547977 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:30.548241 kubelet[2749]: E0514 23:52:30.548225 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:30.548241 kubelet[2749]: W0514 23:52:30.548238 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:30.548290 kubelet[2749]: E0514 23:52:30.548248 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:30.548581 kubelet[2749]: E0514 23:52:30.548555 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:30.548581 kubelet[2749]: W0514 23:52:30.548571 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:30.548629 kubelet[2749]: E0514 23:52:30.548582 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:30.548845 kubelet[2749]: E0514 23:52:30.548822 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:30.548845 kubelet[2749]: W0514 23:52:30.548837 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:30.548910 kubelet[2749]: E0514 23:52:30.548846 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:30.549117 kubelet[2749]: E0514 23:52:30.549092 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:30.549117 kubelet[2749]: W0514 23:52:30.549108 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:30.549172 kubelet[2749]: E0514 23:52:30.549118 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:30.549384 kubelet[2749]: E0514 23:52:30.549369 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:30.549432 kubelet[2749]: W0514 23:52:30.549381 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:30.549432 kubelet[2749]: E0514 23:52:30.549404 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:30.549648 kubelet[2749]: E0514 23:52:30.549625 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:30.549648 kubelet[2749]: W0514 23:52:30.549639 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:30.549697 kubelet[2749]: E0514 23:52:30.549649 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:30.549918 kubelet[2749]: E0514 23:52:30.549895 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:30.549918 kubelet[2749]: W0514 23:52:30.549910 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:30.549976 kubelet[2749]: E0514 23:52:30.549920 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:30.550150 kubelet[2749]: E0514 23:52:30.550133 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:30.550150 kubelet[2749]: W0514 23:52:30.550146 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:30.550202 kubelet[2749]: E0514 23:52:30.550156 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:30.550545 kubelet[2749]: E0514 23:52:30.550529 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:30.550545 kubelet[2749]: W0514 23:52:30.550541 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:30.550608 kubelet[2749]: E0514 23:52:30.550554 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:30.550812 kubelet[2749]: E0514 23:52:30.550795 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:30.550812 kubelet[2749]: W0514 23:52:30.550808 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:30.550874 kubelet[2749]: E0514 23:52:30.550817 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:30.551076 kubelet[2749]: E0514 23:52:30.551061 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:30.551076 kubelet[2749]: W0514 23:52:30.551073 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:30.551129 kubelet[2749]: E0514 23:52:30.551082 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:30.591770 kubelet[2749]: E0514 23:52:30.591710 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:30.591770 kubelet[2749]: W0514 23:52:30.591744 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:30.591770 kubelet[2749]: E0514 23:52:30.591771 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:30.592095 kubelet[2749]: E0514 23:52:30.592067 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:30.592095 kubelet[2749]: W0514 23:52:30.592081 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:30.592147 kubelet[2749]: E0514 23:52:30.592097 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:30.592459 kubelet[2749]: E0514 23:52:30.592431 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:30.592459 kubelet[2749]: W0514 23:52:30.592448 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:30.592533 kubelet[2749]: E0514 23:52:30.592465 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:30.592784 kubelet[2749]: E0514 23:52:30.592748 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:30.592784 kubelet[2749]: W0514 23:52:30.592765 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:30.592876 kubelet[2749]: E0514 23:52:30.592788 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:30.593084 kubelet[2749]: E0514 23:52:30.593052 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:30.593084 kubelet[2749]: W0514 23:52:30.593067 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:30.593084 kubelet[2749]: E0514 23:52:30.593083 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:30.593350 kubelet[2749]: E0514 23:52:30.593330 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:30.593350 kubelet[2749]: W0514 23:52:30.593343 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:30.593450 kubelet[2749]: E0514 23:52:30.593360 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:30.593829 kubelet[2749]: E0514 23:52:30.593788 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:30.593829 kubelet[2749]: W0514 23:52:30.593825 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:30.593926 kubelet[2749]: E0514 23:52:30.593862 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:30.594717 kubelet[2749]: E0514 23:52:30.594694 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:30.594717 kubelet[2749]: W0514 23:52:30.594710 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:30.594820 kubelet[2749]: E0514 23:52:30.594776 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:30.595004 kubelet[2749]: E0514 23:52:30.594980 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:30.595004 kubelet[2749]: W0514 23:52:30.594995 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:30.595084 kubelet[2749]: E0514 23:52:30.595045 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:30.595321 kubelet[2749]: E0514 23:52:30.595281 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:30.595321 kubelet[2749]: W0514 23:52:30.595318 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:30.595402 kubelet[2749]: E0514 23:52:30.595369 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:30.595633 kubelet[2749]: E0514 23:52:30.595600 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:30.595633 kubelet[2749]: W0514 23:52:30.595619 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:30.595633 kubelet[2749]: E0514 23:52:30.595643 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:30.596054 kubelet[2749]: E0514 23:52:30.595978 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:30.596054 kubelet[2749]: W0514 23:52:30.595990 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:30.596054 kubelet[2749]: E0514 23:52:30.596005 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:30.599326 kubelet[2749]: E0514 23:52:30.596503 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:30.599326 kubelet[2749]: W0514 23:52:30.596522 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:30.599326 kubelet[2749]: E0514 23:52:30.596541 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:30.599326 kubelet[2749]: E0514 23:52:30.596877 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:30.599326 kubelet[2749]: W0514 23:52:30.596890 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:30.599326 kubelet[2749]: E0514 23:52:30.596937 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:30.599326 kubelet[2749]: E0514 23:52:30.597280 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:30.599326 kubelet[2749]: W0514 23:52:30.597289 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:30.599326 kubelet[2749]: E0514 23:52:30.597349 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:30.599326 kubelet[2749]: E0514 23:52:30.597705 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:30.599691 kubelet[2749]: W0514 23:52:30.597716 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:30.599691 kubelet[2749]: E0514 23:52:30.597836 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:30.599691 kubelet[2749]: E0514 23:52:30.598579 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:30.599691 kubelet[2749]: W0514 23:52:30.598620 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:30.599691 kubelet[2749]: E0514 23:52:30.598639 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:30.599691 kubelet[2749]: E0514 23:52:30.598926 2749 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 23:52:30.599691 kubelet[2749]: W0514 23:52:30.598937 2749 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 23:52:30.599691 kubelet[2749]: E0514 23:52:30.598960 2749 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 23:52:31.303492 containerd[1510]: time="2025-05-14T23:52:31.303418086Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:52:31.304252 containerd[1510]: time="2025-05-14T23:52:31.304189947Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3: active requests=0, bytes read=5366937" May 14 23:52:31.305647 containerd[1510]: time="2025-05-14T23:52:31.305613083Z" level=info msg="ImageCreate event name:\"sha256:0ceddb3add2e9955cbb604f666245e259f30b1d6683c428f8748359e83d238a5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:52:31.308485 containerd[1510]: time="2025-05-14T23:52:31.308454434Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:eeaa2bb4f9b1aa61adde43ce6dea95eee89291f96963548e108d9a2dfbc5edd1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:52:31.309395 containerd[1510]: time="2025-05-14T23:52:31.309330761Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\" with image id \"sha256:0ceddb3add2e9955cbb604f666245e259f30b1d6683c428f8748359e83d238a5\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:eeaa2bb4f9b1aa61adde43ce6dea95eee89291f96963548e108d9a2dfbc5edd1\", size \"6859519\" in 2.025474452s" May 14 23:52:31.309456 containerd[1510]: time="2025-05-14T23:52:31.309401124Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\" returns image reference \"sha256:0ceddb3add2e9955cbb604f666245e259f30b1d6683c428f8748359e83d238a5\"" May 14 23:52:31.311466 containerd[1510]: time="2025-05-14T23:52:31.311423455Z" level=info msg="CreateContainer within sandbox \"8f9fa26a1fa44c203c6923f82ed62cd8f77fec5a5b9463691529e42246e6aee7\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" May 14 23:52:31.323288 containerd[1510]: time="2025-05-14T23:52:31.323225300Z" level=info msg="Container 478707a089bd655e341f92773dd10b0b3fb8ca8ceb508eecffb41eb4cf4a0a06: CDI devices from CRI Config.CDIDevices: []" May 14 23:52:31.332216 containerd[1510]: time="2025-05-14T23:52:31.332170105Z" level=info msg="CreateContainer within sandbox \"8f9fa26a1fa44c203c6923f82ed62cd8f77fec5a5b9463691529e42246e6aee7\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"478707a089bd655e341f92773dd10b0b3fb8ca8ceb508eecffb41eb4cf4a0a06\"" May 14 23:52:31.332694 containerd[1510]: time="2025-05-14T23:52:31.332654736Z" level=info msg="StartContainer for \"478707a089bd655e341f92773dd10b0b3fb8ca8ceb508eecffb41eb4cf4a0a06\"" May 14 23:52:31.334136 containerd[1510]: time="2025-05-14T23:52:31.334106886Z" level=info msg="connecting to shim 478707a089bd655e341f92773dd10b0b3fb8ca8ceb508eecffb41eb4cf4a0a06" address="unix:///run/containerd/s/336e40f2e2183dfd0f34d097e5c6cab0f947f7c46bb4579ae03ff572ea0fecda" protocol=ttrpc version=3 May 14 23:52:31.361667 systemd[1]: Started cri-containerd-478707a089bd655e341f92773dd10b0b3fb8ca8ceb508eecffb41eb4cf4a0a06.scope - libcontainer container 478707a089bd655e341f92773dd10b0b3fb8ca8ceb508eecffb41eb4cf4a0a06. May 14 23:52:31.411565 containerd[1510]: time="2025-05-14T23:52:31.411506182Z" level=info msg="StartContainer for \"478707a089bd655e341f92773dd10b0b3fb8ca8ceb508eecffb41eb4cf4a0a06\" returns successfully" May 14 23:52:31.426083 systemd[1]: cri-containerd-478707a089bd655e341f92773dd10b0b3fb8ca8ceb508eecffb41eb4cf4a0a06.scope: Deactivated successfully. May 14 23:52:31.429045 containerd[1510]: time="2025-05-14T23:52:31.429000518Z" level=info msg="received exit event container_id:\"478707a089bd655e341f92773dd10b0b3fb8ca8ceb508eecffb41eb4cf4a0a06\" id:\"478707a089bd655e341f92773dd10b0b3fb8ca8ceb508eecffb41eb4cf4a0a06\" pid:3466 exited_at:{seconds:1747266751 nanos:428524253}" May 14 23:52:31.429184 containerd[1510]: time="2025-05-14T23:52:31.429133108Z" level=info msg="TaskExit event in podsandbox handler container_id:\"478707a089bd655e341f92773dd10b0b3fb8ca8ceb508eecffb41eb4cf4a0a06\" id:\"478707a089bd655e341f92773dd10b0b3fb8ca8ceb508eecffb41eb4cf4a0a06\" pid:3466 exited_at:{seconds:1747266751 nanos:428524253}" May 14 23:52:31.459864 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-478707a089bd655e341f92773dd10b0b3fb8ca8ceb508eecffb41eb4cf4a0a06-rootfs.mount: Deactivated successfully. May 14 23:52:32.378879 kubelet[2749]: E0514 23:52:32.378765 2749 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-96wg8" podUID="8f0b6425-d6bf-493f-ada4-58a2a3096286" May 14 23:52:32.473644 containerd[1510]: time="2025-05-14T23:52:32.473482945Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.3\"" May 14 23:52:34.378453 kubelet[2749]: E0514 23:52:34.378354 2749 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-96wg8" podUID="8f0b6425-d6bf-493f-ada4-58a2a3096286" May 14 23:52:35.282848 systemd[1]: Started sshd@8-10.0.0.142:22-10.0.0.1:57852.service - OpenSSH per-connection server daemon (10.0.0.1:57852). May 14 23:52:35.332409 sshd[3509]: Accepted publickey for core from 10.0.0.1 port 57852 ssh2: RSA SHA256:lk2TkYBEL43KPVbrGyh3Ro8IB8NGN6uTNXzFyrYR01I May 14 23:52:35.333968 sshd-session[3509]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 23:52:35.340368 systemd-logind[1487]: New session 9 of user core. May 14 23:52:35.349446 systemd[1]: Started session-9.scope - Session 9 of User core. May 14 23:52:35.496327 sshd[3511]: Connection closed by 10.0.0.1 port 57852 May 14 23:52:35.494670 sshd-session[3509]: pam_unix(sshd:session): session closed for user core May 14 23:52:35.498724 systemd[1]: sshd@8-10.0.0.142:22-10.0.0.1:57852.service: Deactivated successfully. May 14 23:52:35.501688 systemd[1]: session-9.scope: Deactivated successfully. May 14 23:52:35.503206 systemd-logind[1487]: Session 9 logged out. Waiting for processes to exit. May 14 23:52:35.505835 systemd-logind[1487]: Removed session 9. May 14 23:52:36.266231 containerd[1510]: time="2025-05-14T23:52:36.266151892Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:52:36.267325 containerd[1510]: time="2025-05-14T23:52:36.267260455Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.29.3: active requests=0, bytes read=97793683" May 14 23:52:36.268625 containerd[1510]: time="2025-05-14T23:52:36.268590163Z" level=info msg="ImageCreate event name:\"sha256:a140d04be1bc987bae0a1b9159e1dcb85751c448830efbdb3494207cf602b2d9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:52:36.270932 containerd[1510]: time="2025-05-14T23:52:36.270884233Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:4505ec8f976470994b6a94295a4dabac0cb98375db050e959a22603e00ada90b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:52:36.271448 containerd[1510]: time="2025-05-14T23:52:36.271418487Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.29.3\" with image id \"sha256:a140d04be1bc987bae0a1b9159e1dcb85751c448830efbdb3494207cf602b2d9\", repo tag \"ghcr.io/flatcar/calico/cni:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:4505ec8f976470994b6a94295a4dabac0cb98375db050e959a22603e00ada90b\", size \"99286305\" in 3.797879476s" May 14 23:52:36.271448 containerd[1510]: time="2025-05-14T23:52:36.271448654Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.3\" returns image reference \"sha256:a140d04be1bc987bae0a1b9159e1dcb85751c448830efbdb3494207cf602b2d9\"" May 14 23:52:36.287439 containerd[1510]: time="2025-05-14T23:52:36.287365095Z" level=info msg="CreateContainer within sandbox \"8f9fa26a1fa44c203c6923f82ed62cd8f77fec5a5b9463691529e42246e6aee7\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" May 14 23:52:36.298672 containerd[1510]: time="2025-05-14T23:52:36.298607265Z" level=info msg="Container 37f35fbbdf86058a05529ff9d1d38d391a90ca5283bf71b709e49670c000b486: CDI devices from CRI Config.CDIDevices: []" May 14 23:52:36.309683 containerd[1510]: time="2025-05-14T23:52:36.309621989Z" level=info msg="CreateContainer within sandbox \"8f9fa26a1fa44c203c6923f82ed62cd8f77fec5a5b9463691529e42246e6aee7\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"37f35fbbdf86058a05529ff9d1d38d391a90ca5283bf71b709e49670c000b486\"" May 14 23:52:36.313204 containerd[1510]: time="2025-05-14T23:52:36.313159295Z" level=info msg="StartContainer for \"37f35fbbdf86058a05529ff9d1d38d391a90ca5283bf71b709e49670c000b486\"" May 14 23:52:36.315216 containerd[1510]: time="2025-05-14T23:52:36.315178949Z" level=info msg="connecting to shim 37f35fbbdf86058a05529ff9d1d38d391a90ca5283bf71b709e49670c000b486" address="unix:///run/containerd/s/336e40f2e2183dfd0f34d097e5c6cab0f947f7c46bb4579ae03ff572ea0fecda" protocol=ttrpc version=3 May 14 23:52:36.345819 systemd[1]: Started cri-containerd-37f35fbbdf86058a05529ff9d1d38d391a90ca5283bf71b709e49670c000b486.scope - libcontainer container 37f35fbbdf86058a05529ff9d1d38d391a90ca5283bf71b709e49670c000b486. May 14 23:52:36.381267 kubelet[2749]: E0514 23:52:36.381198 2749 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-96wg8" podUID="8f0b6425-d6bf-493f-ada4-58a2a3096286" May 14 23:52:36.438636 containerd[1510]: time="2025-05-14T23:52:36.438579579Z" level=info msg="StartContainer for \"37f35fbbdf86058a05529ff9d1d38d391a90ca5283bf71b709e49670c000b486\" returns successfully" May 14 23:52:38.378707 kubelet[2749]: E0514 23:52:38.378623 2749 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-96wg8" podUID="8f0b6425-d6bf-493f-ada4-58a2a3096286" May 14 23:52:38.649668 systemd[1]: cri-containerd-37f35fbbdf86058a05529ff9d1d38d391a90ca5283bf71b709e49670c000b486.scope: Deactivated successfully. May 14 23:52:38.650156 systemd[1]: cri-containerd-37f35fbbdf86058a05529ff9d1d38d391a90ca5283bf71b709e49670c000b486.scope: Consumed 652ms CPU time, 160.8M memory peak, 24K read from disk, 154M written to disk. May 14 23:52:38.651137 containerd[1510]: time="2025-05-14T23:52:38.650938928Z" level=info msg="received exit event container_id:\"37f35fbbdf86058a05529ff9d1d38d391a90ca5283bf71b709e49670c000b486\" id:\"37f35fbbdf86058a05529ff9d1d38d391a90ca5283bf71b709e49670c000b486\" pid:3541 exited_at:{seconds:1747266758 nanos:650571858}" May 14 23:52:38.651137 containerd[1510]: time="2025-05-14T23:52:38.651084280Z" level=info msg="TaskExit event in podsandbox handler container_id:\"37f35fbbdf86058a05529ff9d1d38d391a90ca5283bf71b709e49670c000b486\" id:\"37f35fbbdf86058a05529ff9d1d38d391a90ca5283bf71b709e49670c000b486\" pid:3541 exited_at:{seconds:1747266758 nanos:650571858}" May 14 23:52:38.681223 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-37f35fbbdf86058a05529ff9d1d38d391a90ca5283bf71b709e49670c000b486-rootfs.mount: Deactivated successfully. May 14 23:52:38.736159 kubelet[2749]: I0514 23:52:38.736114 2749 kubelet_node_status.go:497] "Fast updating node status as it just became ready" May 14 23:52:38.839939 kubelet[2749]: I0514 23:52:38.837701 2749 topology_manager.go:215] "Topology Admit Handler" podUID="d0da08d5-ebd0-447d-b3dc-5ae18d60d27e" podNamespace="kube-system" podName="coredns-7db6d8ff4d-5bhtm" May 14 23:52:38.850167 kubelet[2749]: I0514 23:52:38.850120 2749 topology_manager.go:215] "Topology Admit Handler" podUID="d5000901-d9c1-41f3-9444-837fe018e119" podNamespace="kube-system" podName="coredns-7db6d8ff4d-xq4zr" May 14 23:52:38.850838 kubelet[2749]: I0514 23:52:38.850815 2749 topology_manager.go:215] "Topology Admit Handler" podUID="aaa27ebc-a52f-4ebe-9a1d-8508c4e3e499" podNamespace="calico-system" podName="calico-kube-controllers-57dfd6df95-fhrn2" May 14 23:52:38.852425 kubelet[2749]: I0514 23:52:38.852248 2749 topology_manager.go:215] "Topology Admit Handler" podUID="3b94c16d-c98d-4f96-8337-ac0791c35c61" podNamespace="calico-apiserver" podName="calico-apiserver-78f88fbddb-wtw8n" May 14 23:52:38.855387 kubelet[2749]: I0514 23:52:38.855344 2749 topology_manager.go:215] "Topology Admit Handler" podUID="e558cb7f-428e-4fae-a45c-cc274d3aa581" podNamespace="calico-apiserver" podName="calico-apiserver-78f88fbddb-c68xl" May 14 23:52:38.856896 systemd[1]: Created slice kubepods-burstable-podd0da08d5_ebd0_447d_b3dc_5ae18d60d27e.slice - libcontainer container kubepods-burstable-podd0da08d5_ebd0_447d_b3dc_5ae18d60d27e.slice. May 14 23:52:38.867344 systemd[1]: Created slice kubepods-besteffort-pod3b94c16d_c98d_4f96_8337_ac0791c35c61.slice - libcontainer container kubepods-besteffort-pod3b94c16d_c98d_4f96_8337_ac0791c35c61.slice. May 14 23:52:38.873131 systemd[1]: Created slice kubepods-besteffort-podaaa27ebc_a52f_4ebe_9a1d_8508c4e3e499.slice - libcontainer container kubepods-besteffort-podaaa27ebc_a52f_4ebe_9a1d_8508c4e3e499.slice. May 14 23:52:38.879570 systemd[1]: Created slice kubepods-burstable-podd5000901_d9c1_41f3_9444_837fe018e119.slice - libcontainer container kubepods-burstable-podd5000901_d9c1_41f3_9444_837fe018e119.slice. May 14 23:52:38.885674 systemd[1]: Created slice kubepods-besteffort-pode558cb7f_428e_4fae_a45c_cc274d3aa581.slice - libcontainer container kubepods-besteffort-pode558cb7f_428e_4fae_a45c_cc274d3aa581.slice. May 14 23:52:38.971783 kubelet[2749]: I0514 23:52:38.971525 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/3b94c16d-c98d-4f96-8337-ac0791c35c61-calico-apiserver-certs\") pod \"calico-apiserver-78f88fbddb-wtw8n\" (UID: \"3b94c16d-c98d-4f96-8337-ac0791c35c61\") " pod="calico-apiserver/calico-apiserver-78f88fbddb-wtw8n" May 14 23:52:38.971783 kubelet[2749]: I0514 23:52:38.971613 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wzs7\" (UniqueName: \"kubernetes.io/projected/aaa27ebc-a52f-4ebe-9a1d-8508c4e3e499-kube-api-access-6wzs7\") pod \"calico-kube-controllers-57dfd6df95-fhrn2\" (UID: \"aaa27ebc-a52f-4ebe-9a1d-8508c4e3e499\") " pod="calico-system/calico-kube-controllers-57dfd6df95-fhrn2" May 14 23:52:38.971783 kubelet[2749]: I0514 23:52:38.971651 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/e558cb7f-428e-4fae-a45c-cc274d3aa581-calico-apiserver-certs\") pod \"calico-apiserver-78f88fbddb-c68xl\" (UID: \"e558cb7f-428e-4fae-a45c-cc274d3aa581\") " pod="calico-apiserver/calico-apiserver-78f88fbddb-c68xl" May 14 23:52:38.971783 kubelet[2749]: I0514 23:52:38.971676 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbqm8\" (UniqueName: \"kubernetes.io/projected/d0da08d5-ebd0-447d-b3dc-5ae18d60d27e-kube-api-access-rbqm8\") pod \"coredns-7db6d8ff4d-5bhtm\" (UID: \"d0da08d5-ebd0-447d-b3dc-5ae18d60d27e\") " pod="kube-system/coredns-7db6d8ff4d-5bhtm" May 14 23:52:38.971783 kubelet[2749]: I0514 23:52:38.971703 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d5000901-d9c1-41f3-9444-837fe018e119-config-volume\") pod \"coredns-7db6d8ff4d-xq4zr\" (UID: \"d5000901-d9c1-41f3-9444-837fe018e119\") " pod="kube-system/coredns-7db6d8ff4d-xq4zr" May 14 23:52:38.972140 kubelet[2749]: I0514 23:52:38.971727 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aaa27ebc-a52f-4ebe-9a1d-8508c4e3e499-tigera-ca-bundle\") pod \"calico-kube-controllers-57dfd6df95-fhrn2\" (UID: \"aaa27ebc-a52f-4ebe-9a1d-8508c4e3e499\") " pod="calico-system/calico-kube-controllers-57dfd6df95-fhrn2" May 14 23:52:38.972140 kubelet[2749]: I0514 23:52:38.971751 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnkf2\" (UniqueName: \"kubernetes.io/projected/d5000901-d9c1-41f3-9444-837fe018e119-kube-api-access-jnkf2\") pod \"coredns-7db6d8ff4d-xq4zr\" (UID: \"d5000901-d9c1-41f3-9444-837fe018e119\") " pod="kube-system/coredns-7db6d8ff4d-xq4zr" May 14 23:52:38.972140 kubelet[2749]: I0514 23:52:38.971780 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0da08d5-ebd0-447d-b3dc-5ae18d60d27e-config-volume\") pod \"coredns-7db6d8ff4d-5bhtm\" (UID: \"d0da08d5-ebd0-447d-b3dc-5ae18d60d27e\") " pod="kube-system/coredns-7db6d8ff4d-5bhtm" May 14 23:52:38.972140 kubelet[2749]: I0514 23:52:38.971805 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4hjq\" (UniqueName: \"kubernetes.io/projected/e558cb7f-428e-4fae-a45c-cc274d3aa581-kube-api-access-l4hjq\") pod \"calico-apiserver-78f88fbddb-c68xl\" (UID: \"e558cb7f-428e-4fae-a45c-cc274d3aa581\") " pod="calico-apiserver/calico-apiserver-78f88fbddb-c68xl" May 14 23:52:38.972140 kubelet[2749]: I0514 23:52:38.971830 2749 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vcvt\" (UniqueName: \"kubernetes.io/projected/3b94c16d-c98d-4f96-8337-ac0791c35c61-kube-api-access-6vcvt\") pod \"calico-apiserver-78f88fbddb-wtw8n\" (UID: \"3b94c16d-c98d-4f96-8337-ac0791c35c61\") " pod="calico-apiserver/calico-apiserver-78f88fbddb-wtw8n" May 14 23:52:39.465261 containerd[1510]: time="2025-05-14T23:52:39.465166697Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5bhtm,Uid:d0da08d5-ebd0-447d-b3dc-5ae18d60d27e,Namespace:kube-system,Attempt:0,}" May 14 23:52:39.471728 containerd[1510]: time="2025-05-14T23:52:39.471680061Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-78f88fbddb-wtw8n,Uid:3b94c16d-c98d-4f96-8337-ac0791c35c61,Namespace:calico-apiserver,Attempt:0,}" May 14 23:52:39.477800 containerd[1510]: time="2025-05-14T23:52:39.477733200Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-57dfd6df95-fhrn2,Uid:aaa27ebc-a52f-4ebe-9a1d-8508c4e3e499,Namespace:calico-system,Attempt:0,}" May 14 23:52:39.484913 containerd[1510]: time="2025-05-14T23:52:39.484846459Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-xq4zr,Uid:d5000901-d9c1-41f3-9444-837fe018e119,Namespace:kube-system,Attempt:0,}" May 14 23:52:39.490383 containerd[1510]: time="2025-05-14T23:52:39.490316714Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-78f88fbddb-c68xl,Uid:e558cb7f-428e-4fae-a45c-cc274d3aa581,Namespace:calico-apiserver,Attempt:0,}" May 14 23:52:39.513523 containerd[1510]: time="2025-05-14T23:52:39.513255516Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.3\"" May 14 23:52:39.606040 containerd[1510]: time="2025-05-14T23:52:39.605955522Z" level=error msg="Failed to destroy network for sandbox \"dfcd84869b7df3cbed24bcc3646a1aceeae3f47773b592b7e6d868cecb24d293\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 14 23:52:39.611753 containerd[1510]: time="2025-05-14T23:52:39.611567582Z" level=error msg="Failed to destroy network for sandbox \"edf85d83bf055f9662fd3e202d36bbd8c91ecc4de7a54c8b8b9acf8319b2dcbb\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 14 23:52:39.611895 containerd[1510]: time="2025-05-14T23:52:39.611642493Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-78f88fbddb-wtw8n,Uid:3b94c16d-c98d-4f96-8337-ac0791c35c61,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"dfcd84869b7df3cbed24bcc3646a1aceeae3f47773b592b7e6d868cecb24d293\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 14 23:52:39.614376 kubelet[2749]: E0514 23:52:39.612136 2749 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"dfcd84869b7df3cbed24bcc3646a1aceeae3f47773b592b7e6d868cecb24d293\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 14 23:52:39.614376 kubelet[2749]: E0514 23:52:39.612260 2749 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"dfcd84869b7df3cbed24bcc3646a1aceeae3f47773b592b7e6d868cecb24d293\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-78f88fbddb-wtw8n" May 14 23:52:39.614376 kubelet[2749]: E0514 23:52:39.612291 2749 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"dfcd84869b7df3cbed24bcc3646a1aceeae3f47773b592b7e6d868cecb24d293\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-78f88fbddb-wtw8n" May 14 23:52:39.615009 kubelet[2749]: E0514 23:52:39.612470 2749 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-78f88fbddb-wtw8n_calico-apiserver(3b94c16d-c98d-4f96-8337-ac0791c35c61)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-78f88fbddb-wtw8n_calico-apiserver(3b94c16d-c98d-4f96-8337-ac0791c35c61)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"dfcd84869b7df3cbed24bcc3646a1aceeae3f47773b592b7e6d868cecb24d293\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-78f88fbddb-wtw8n" podUID="3b94c16d-c98d-4f96-8337-ac0791c35c61" May 14 23:52:39.615417 containerd[1510]: time="2025-05-14T23:52:39.615254947Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-57dfd6df95-fhrn2,Uid:aaa27ebc-a52f-4ebe-9a1d-8508c4e3e499,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"edf85d83bf055f9662fd3e202d36bbd8c91ecc4de7a54c8b8b9acf8319b2dcbb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 14 23:52:39.615849 kubelet[2749]: E0514 23:52:39.615624 2749 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"edf85d83bf055f9662fd3e202d36bbd8c91ecc4de7a54c8b8b9acf8319b2dcbb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 14 23:52:39.615849 kubelet[2749]: E0514 23:52:39.615711 2749 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"edf85d83bf055f9662fd3e202d36bbd8c91ecc4de7a54c8b8b9acf8319b2dcbb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-57dfd6df95-fhrn2" May 14 23:52:39.615849 kubelet[2749]: E0514 23:52:39.615739 2749 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"edf85d83bf055f9662fd3e202d36bbd8c91ecc4de7a54c8b8b9acf8319b2dcbb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-57dfd6df95-fhrn2" May 14 23:52:39.616016 kubelet[2749]: E0514 23:52:39.615794 2749 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-57dfd6df95-fhrn2_calico-system(aaa27ebc-a52f-4ebe-9a1d-8508c4e3e499)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-57dfd6df95-fhrn2_calico-system(aaa27ebc-a52f-4ebe-9a1d-8508c4e3e499)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"edf85d83bf055f9662fd3e202d36bbd8c91ecc4de7a54c8b8b9acf8319b2dcbb\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-57dfd6df95-fhrn2" podUID="aaa27ebc-a52f-4ebe-9a1d-8508c4e3e499" May 14 23:52:39.623362 containerd[1510]: time="2025-05-14T23:52:39.623211041Z" level=error msg="Failed to destroy network for sandbox \"805b5d35a4f350ae627798a4f62ba48a19bc372f1917f62c07c12579814bd22b\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 14 23:52:39.625970 containerd[1510]: time="2025-05-14T23:52:39.625904350Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5bhtm,Uid:d0da08d5-ebd0-447d-b3dc-5ae18d60d27e,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"805b5d35a4f350ae627798a4f62ba48a19bc372f1917f62c07c12579814bd22b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 14 23:52:39.626579 kubelet[2749]: E0514 23:52:39.626504 2749 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"805b5d35a4f350ae627798a4f62ba48a19bc372f1917f62c07c12579814bd22b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 14 23:52:39.626756 kubelet[2749]: E0514 23:52:39.626588 2749 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"805b5d35a4f350ae627798a4f62ba48a19bc372f1917f62c07c12579814bd22b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-5bhtm" May 14 23:52:39.626756 kubelet[2749]: E0514 23:52:39.626621 2749 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"805b5d35a4f350ae627798a4f62ba48a19bc372f1917f62c07c12579814bd22b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-5bhtm" May 14 23:52:39.626756 kubelet[2749]: E0514 23:52:39.626678 2749 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-5bhtm_kube-system(d0da08d5-ebd0-447d-b3dc-5ae18d60d27e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-5bhtm_kube-system(d0da08d5-ebd0-447d-b3dc-5ae18d60d27e)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"805b5d35a4f350ae627798a4f62ba48a19bc372f1917f62c07c12579814bd22b\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-5bhtm" podUID="d0da08d5-ebd0-447d-b3dc-5ae18d60d27e" May 14 23:52:39.630444 containerd[1510]: time="2025-05-14T23:52:39.630383803Z" level=error msg="Failed to destroy network for sandbox \"a98199034f929c52da30b3c8b8a9f93d36acf9651188804e39a52cdeafc3d250\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 14 23:52:39.640935 containerd[1510]: time="2025-05-14T23:52:39.640854488Z" level=error msg="Failed to destroy network for sandbox \"d38622cf91ccfdf99c080247bdcfea3cef786efbb6e15336f89720111b815661\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 14 23:52:39.651763 containerd[1510]: time="2025-05-14T23:52:39.651675532Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-xq4zr,Uid:d5000901-d9c1-41f3-9444-837fe018e119,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"a98199034f929c52da30b3c8b8a9f93d36acf9651188804e39a52cdeafc3d250\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 14 23:52:39.652379 kubelet[2749]: E0514 23:52:39.652032 2749 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a98199034f929c52da30b3c8b8a9f93d36acf9651188804e39a52cdeafc3d250\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 14 23:52:39.652379 kubelet[2749]: E0514 23:52:39.652118 2749 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a98199034f929c52da30b3c8b8a9f93d36acf9651188804e39a52cdeafc3d250\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-xq4zr" May 14 23:52:39.652379 kubelet[2749]: E0514 23:52:39.652149 2749 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a98199034f929c52da30b3c8b8a9f93d36acf9651188804e39a52cdeafc3d250\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-xq4zr" May 14 23:52:39.652551 kubelet[2749]: E0514 23:52:39.652213 2749 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-xq4zr_kube-system(d5000901-d9c1-41f3-9444-837fe018e119)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-xq4zr_kube-system(d5000901-d9c1-41f3-9444-837fe018e119)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"a98199034f929c52da30b3c8b8a9f93d36acf9651188804e39a52cdeafc3d250\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-xq4zr" podUID="d5000901-d9c1-41f3-9444-837fe018e119" May 14 23:52:39.688395 containerd[1510]: time="2025-05-14T23:52:39.688258190Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-78f88fbddb-c68xl,Uid:e558cb7f-428e-4fae-a45c-cc274d3aa581,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"d38622cf91ccfdf99c080247bdcfea3cef786efbb6e15336f89720111b815661\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 14 23:52:39.688729 kubelet[2749]: E0514 23:52:39.688664 2749 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d38622cf91ccfdf99c080247bdcfea3cef786efbb6e15336f89720111b815661\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 14 23:52:39.688797 kubelet[2749]: E0514 23:52:39.688751 2749 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d38622cf91ccfdf99c080247bdcfea3cef786efbb6e15336f89720111b815661\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-78f88fbddb-c68xl" May 14 23:52:39.688797 kubelet[2749]: E0514 23:52:39.688777 2749 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d38622cf91ccfdf99c080247bdcfea3cef786efbb6e15336f89720111b815661\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-78f88fbddb-c68xl" May 14 23:52:39.688879 kubelet[2749]: E0514 23:52:39.688836 2749 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-78f88fbddb-c68xl_calico-apiserver(e558cb7f-428e-4fae-a45c-cc274d3aa581)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-78f88fbddb-c68xl_calico-apiserver(e558cb7f-428e-4fae-a45c-cc274d3aa581)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"d38622cf91ccfdf99c080247bdcfea3cef786efbb6e15336f89720111b815661\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-78f88fbddb-c68xl" podUID="e558cb7f-428e-4fae-a45c-cc274d3aa581" May 14 23:52:40.385744 systemd[1]: Created slice kubepods-besteffort-pod8f0b6425_d6bf_493f_ada4_58a2a3096286.slice - libcontainer container kubepods-besteffort-pod8f0b6425_d6bf_493f_ada4_58a2a3096286.slice. May 14 23:52:40.388733 containerd[1510]: time="2025-05-14T23:52:40.388687701Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-96wg8,Uid:8f0b6425-d6bf-493f-ada4-58a2a3096286,Namespace:calico-system,Attempt:0,}" May 14 23:52:40.513156 systemd[1]: Started sshd@9-10.0.0.142:22-10.0.0.1:54024.service - OpenSSH per-connection server daemon (10.0.0.1:54024). May 14 23:52:40.521379 containerd[1510]: time="2025-05-14T23:52:40.521327800Z" level=error msg="Failed to destroy network for sandbox \"57be01ba8e191f3bfcabb247d4f16a6fbee0cd79ec94d1974313477cd6829c4b\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 14 23:52:40.524585 systemd[1]: run-netns-cni\x2dc7c421a9\x2d7c0b\x2dbac5\x2dccf3\x2de108510542e6.mount: Deactivated successfully. May 14 23:52:40.549101 containerd[1510]: time="2025-05-14T23:52:40.549002373Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-96wg8,Uid:8f0b6425-d6bf-493f-ada4-58a2a3096286,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"57be01ba8e191f3bfcabb247d4f16a6fbee0cd79ec94d1974313477cd6829c4b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 14 23:52:40.549455 kubelet[2749]: E0514 23:52:40.549382 2749 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"57be01ba8e191f3bfcabb247d4f16a6fbee0cd79ec94d1974313477cd6829c4b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 14 23:52:40.549547 kubelet[2749]: E0514 23:52:40.549464 2749 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"57be01ba8e191f3bfcabb247d4f16a6fbee0cd79ec94d1974313477cd6829c4b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-96wg8" May 14 23:52:40.549547 kubelet[2749]: E0514 23:52:40.549495 2749 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"57be01ba8e191f3bfcabb247d4f16a6fbee0cd79ec94d1974313477cd6829c4b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-96wg8" May 14 23:52:40.549632 kubelet[2749]: E0514 23:52:40.549561 2749 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-96wg8_calico-system(8f0b6425-d6bf-493f-ada4-58a2a3096286)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-96wg8_calico-system(8f0b6425-d6bf-493f-ada4-58a2a3096286)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"57be01ba8e191f3bfcabb247d4f16a6fbee0cd79ec94d1974313477cd6829c4b\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-96wg8" podUID="8f0b6425-d6bf-493f-ada4-58a2a3096286" May 14 23:52:40.586109 sshd[3799]: Accepted publickey for core from 10.0.0.1 port 54024 ssh2: RSA SHA256:lk2TkYBEL43KPVbrGyh3Ro8IB8NGN6uTNXzFyrYR01I May 14 23:52:40.588270 sshd-session[3799]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 23:52:40.594268 systemd-logind[1487]: New session 10 of user core. May 14 23:52:40.605652 systemd[1]: Started session-10.scope - Session 10 of User core. May 14 23:52:40.765010 sshd[3802]: Connection closed by 10.0.0.1 port 54024 May 14 23:52:40.765392 sshd-session[3799]: pam_unix(sshd:session): session closed for user core May 14 23:52:40.776393 systemd[1]: sshd@9-10.0.0.142:22-10.0.0.1:54024.service: Deactivated successfully. May 14 23:52:40.779024 systemd[1]: session-10.scope: Deactivated successfully. May 14 23:52:40.781429 systemd-logind[1487]: Session 10 logged out. Waiting for processes to exit. May 14 23:52:40.783334 systemd[1]: Started sshd@10-10.0.0.142:22-10.0.0.1:54026.service - OpenSSH per-connection server daemon (10.0.0.1:54026). May 14 23:52:40.784256 systemd-logind[1487]: Removed session 10. May 14 23:52:40.861193 sshd[3816]: Accepted publickey for core from 10.0.0.1 port 54026 ssh2: RSA SHA256:lk2TkYBEL43KPVbrGyh3Ro8IB8NGN6uTNXzFyrYR01I May 14 23:52:40.863357 sshd-session[3816]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 23:52:40.869545 systemd-logind[1487]: New session 11 of user core. May 14 23:52:40.878448 systemd[1]: Started session-11.scope - Session 11 of User core. May 14 23:52:41.298725 sshd[3819]: Connection closed by 10.0.0.1 port 54026 May 14 23:52:41.299426 sshd-session[3816]: pam_unix(sshd:session): session closed for user core May 14 23:52:41.312532 systemd[1]: sshd@10-10.0.0.142:22-10.0.0.1:54026.service: Deactivated successfully. May 14 23:52:41.315517 systemd[1]: session-11.scope: Deactivated successfully. May 14 23:52:41.317953 systemd-logind[1487]: Session 11 logged out. Waiting for processes to exit. May 14 23:52:41.320446 systemd[1]: Started sshd@11-10.0.0.142:22-10.0.0.1:54032.service - OpenSSH per-connection server daemon (10.0.0.1:54032). May 14 23:52:41.322278 systemd-logind[1487]: Removed session 11. May 14 23:52:41.379160 sshd[3830]: Accepted publickey for core from 10.0.0.1 port 54032 ssh2: RSA SHA256:lk2TkYBEL43KPVbrGyh3Ro8IB8NGN6uTNXzFyrYR01I May 14 23:52:41.381060 sshd-session[3830]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 23:52:41.386728 systemd-logind[1487]: New session 12 of user core. May 14 23:52:41.396494 systemd[1]: Started session-12.scope - Session 12 of User core. May 14 23:52:41.695219 sshd[3833]: Connection closed by 10.0.0.1 port 54032 May 14 23:52:41.695611 sshd-session[3830]: pam_unix(sshd:session): session closed for user core May 14 23:52:41.700719 systemd[1]: sshd@11-10.0.0.142:22-10.0.0.1:54032.service: Deactivated successfully. May 14 23:52:41.703016 systemd[1]: session-12.scope: Deactivated successfully. May 14 23:52:41.703899 systemd-logind[1487]: Session 12 logged out. Waiting for processes to exit. May 14 23:52:41.704856 systemd-logind[1487]: Removed session 12. May 14 23:52:44.922572 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2250283480.mount: Deactivated successfully. May 14 23:52:46.148467 containerd[1510]: time="2025-05-14T23:52:46.148385771Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:52:46.153802 containerd[1510]: time="2025-05-14T23:52:46.153710737Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.29.3: active requests=0, bytes read=144068748" May 14 23:52:46.156900 containerd[1510]: time="2025-05-14T23:52:46.156821928Z" level=info msg="ImageCreate event name:\"sha256:042163432abcec06b8077b24973b223a5f4cfdb35d85c3816f5d07a13d51afae\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:52:46.159902 containerd[1510]: time="2025-05-14T23:52:46.159827079Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:750e267b4f8217e0ca9e4107228370190d1a2499b72112ad04370ab9b4553916\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:52:46.160602 containerd[1510]: time="2025-05-14T23:52:46.160549314Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.29.3\" with image id \"sha256:042163432abcec06b8077b24973b223a5f4cfdb35d85c3816f5d07a13d51afae\", repo tag \"ghcr.io/flatcar/calico/node:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/node@sha256:750e267b4f8217e0ca9e4107228370190d1a2499b72112ad04370ab9b4553916\", size \"144068610\" in 6.647239266s" May 14 23:52:46.160669 containerd[1510]: time="2025-05-14T23:52:46.160605962Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.3\" returns image reference \"sha256:042163432abcec06b8077b24973b223a5f4cfdb35d85c3816f5d07a13d51afae\"" May 14 23:52:46.177529 containerd[1510]: time="2025-05-14T23:52:46.177447117Z" level=info msg="CreateContainer within sandbox \"8f9fa26a1fa44c203c6923f82ed62cd8f77fec5a5b9463691529e42246e6aee7\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" May 14 23:52:46.219619 containerd[1510]: time="2025-05-14T23:52:46.219369105Z" level=info msg="Container 6bebd036ec15cc329550bdaaaec133f7641f67426d0a97dbcf4742f65440836d: CDI devices from CRI Config.CDIDevices: []" May 14 23:52:46.254213 containerd[1510]: time="2025-05-14T23:52:46.254145982Z" level=info msg="CreateContainer within sandbox \"8f9fa26a1fa44c203c6923f82ed62cd8f77fec5a5b9463691529e42246e6aee7\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"6bebd036ec15cc329550bdaaaec133f7641f67426d0a97dbcf4742f65440836d\"" May 14 23:52:46.254944 containerd[1510]: time="2025-05-14T23:52:46.254775273Z" level=info msg="StartContainer for \"6bebd036ec15cc329550bdaaaec133f7641f67426d0a97dbcf4742f65440836d\"" May 14 23:52:46.256854 containerd[1510]: time="2025-05-14T23:52:46.256788342Z" level=info msg="connecting to shim 6bebd036ec15cc329550bdaaaec133f7641f67426d0a97dbcf4742f65440836d" address="unix:///run/containerd/s/336e40f2e2183dfd0f34d097e5c6cab0f947f7c46bb4579ae03ff572ea0fecda" protocol=ttrpc version=3 May 14 23:52:46.281508 systemd[1]: Started cri-containerd-6bebd036ec15cc329550bdaaaec133f7641f67426d0a97dbcf4742f65440836d.scope - libcontainer container 6bebd036ec15cc329550bdaaaec133f7641f67426d0a97dbcf4742f65440836d. May 14 23:52:46.341756 containerd[1510]: time="2025-05-14T23:52:46.341687440Z" level=info msg="StartContainer for \"6bebd036ec15cc329550bdaaaec133f7641f67426d0a97dbcf4742f65440836d\" returns successfully" May 14 23:52:46.416960 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. May 14 23:52:46.417172 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. May 14 23:52:46.544560 kubelet[2749]: I0514 23:52:46.544479 2749 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-m5llq" podStartSLOduration=3.62488761 podStartE2EDuration="24.54446003s" podCreationTimestamp="2025-05-14 23:52:22 +0000 UTC" firstStartedPulling="2025-05-14 23:52:25.241927029 +0000 UTC m=+24.961157799" lastFinishedPulling="2025-05-14 23:52:46.161499438 +0000 UTC m=+45.880730219" observedRunningTime="2025-05-14 23:52:46.544456804 +0000 UTC m=+46.263687584" watchObservedRunningTime="2025-05-14 23:52:46.54446003 +0000 UTC m=+46.263690800" May 14 23:52:46.709638 systemd[1]: Started sshd@12-10.0.0.142:22-10.0.0.1:54034.service - OpenSSH per-connection server daemon (10.0.0.1:54034). May 14 23:52:46.782528 sshd[3906]: Accepted publickey for core from 10.0.0.1 port 54034 ssh2: RSA SHA256:lk2TkYBEL43KPVbrGyh3Ro8IB8NGN6uTNXzFyrYR01I May 14 23:52:46.785021 sshd-session[3906]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 23:52:46.793668 systemd-logind[1487]: New session 13 of user core. May 14 23:52:46.806443 systemd[1]: Started session-13.scope - Session 13 of User core. May 14 23:52:46.979475 sshd[3910]: Connection closed by 10.0.0.1 port 54034 May 14 23:52:46.979899 sshd-session[3906]: pam_unix(sshd:session): session closed for user core May 14 23:52:46.985092 systemd[1]: sshd@12-10.0.0.142:22-10.0.0.1:54034.service: Deactivated successfully. May 14 23:52:46.987956 systemd[1]: session-13.scope: Deactivated successfully. May 14 23:52:46.988957 systemd-logind[1487]: Session 13 logged out. Waiting for processes to exit. May 14 23:52:46.990264 systemd-logind[1487]: Removed session 13. May 14 23:52:48.296334 kernel: bpftool[4067]: memfd_create() called without MFD_EXEC or MFD_NOEXEC_SEAL set May 14 23:52:48.556474 systemd-networkd[1437]: vxlan.calico: Link UP May 14 23:52:48.556484 systemd-networkd[1437]: vxlan.calico: Gained carrier May 14 23:52:48.808752 kubelet[2749]: I0514 23:52:48.808563 2749 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 14 23:52:49.005865 containerd[1510]: time="2025-05-14T23:52:49.005809707Z" level=info msg="TaskExit event in podsandbox handler container_id:\"6bebd036ec15cc329550bdaaaec133f7641f67426d0a97dbcf4742f65440836d\" id:\"7891ba5a4d1b2da626aa13144ce9d386eb42707637cd075c8a7a7609a33be622\" pid:4151 exit_status:1 exited_at:{seconds:1747266769 nanos:5135442}" May 14 23:52:49.090960 containerd[1510]: time="2025-05-14T23:52:49.090923486Z" level=info msg="TaskExit event in podsandbox handler container_id:\"6bebd036ec15cc329550bdaaaec133f7641f67426d0a97dbcf4742f65440836d\" id:\"64e1e8b1cd3081cd9afd9e4c5b39a3870370060990dde71523dcf2851721687f\" pid:4174 exit_status:1 exited_at:{seconds:1747266769 nanos:90616750}" May 14 23:52:50.379658 containerd[1510]: time="2025-05-14T23:52:50.379589750Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-78f88fbddb-wtw8n,Uid:3b94c16d-c98d-4f96-8337-ac0791c35c61,Namespace:calico-apiserver,Attempt:0,}" May 14 23:52:50.546480 systemd-networkd[1437]: vxlan.calico: Gained IPv6LL May 14 23:52:51.248841 systemd-networkd[1437]: cali622705137f5: Link UP May 14 23:52:51.249464 systemd-networkd[1437]: cali622705137f5: Gained carrier May 14 23:52:51.265076 containerd[1510]: 2025-05-14 23:52:51.115 [INFO][4193] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--78f88fbddb--wtw8n-eth0 calico-apiserver-78f88fbddb- calico-apiserver 3b94c16d-c98d-4f96-8337-ac0791c35c61 768 0 2025-05-14 23:52:22 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:78f88fbddb projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-78f88fbddb-wtw8n eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali622705137f5 [] []}} ContainerID="155a92f9eec8ef61b6b2f8d758ded3690f3b190b0ef6f444ef42f3dff801c046" Namespace="calico-apiserver" Pod="calico-apiserver-78f88fbddb-wtw8n" WorkloadEndpoint="localhost-k8s-calico--apiserver--78f88fbddb--wtw8n-" May 14 23:52:51.265076 containerd[1510]: 2025-05-14 23:52:51.115 [INFO][4193] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="155a92f9eec8ef61b6b2f8d758ded3690f3b190b0ef6f444ef42f3dff801c046" Namespace="calico-apiserver" Pod="calico-apiserver-78f88fbddb-wtw8n" WorkloadEndpoint="localhost-k8s-calico--apiserver--78f88fbddb--wtw8n-eth0" May 14 23:52:51.265076 containerd[1510]: 2025-05-14 23:52:51.198 [INFO][4208] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="155a92f9eec8ef61b6b2f8d758ded3690f3b190b0ef6f444ef42f3dff801c046" HandleID="k8s-pod-network.155a92f9eec8ef61b6b2f8d758ded3690f3b190b0ef6f444ef42f3dff801c046" Workload="localhost-k8s-calico--apiserver--78f88fbddb--wtw8n-eth0" May 14 23:52:51.265433 containerd[1510]: 2025-05-14 23:52:51.209 [INFO][4208] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="155a92f9eec8ef61b6b2f8d758ded3690f3b190b0ef6f444ef42f3dff801c046" HandleID="k8s-pod-network.155a92f9eec8ef61b6b2f8d758ded3690f3b190b0ef6f444ef42f3dff801c046" Workload="localhost-k8s-calico--apiserver--78f88fbddb--wtw8n-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000264870), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-78f88fbddb-wtw8n", "timestamp":"2025-05-14 23:52:51.198840972 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 14 23:52:51.265433 containerd[1510]: 2025-05-14 23:52:51.209 [INFO][4208] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 14 23:52:51.265433 containerd[1510]: 2025-05-14 23:52:51.209 [INFO][4208] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 14 23:52:51.265433 containerd[1510]: 2025-05-14 23:52:51.209 [INFO][4208] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' May 14 23:52:51.265433 containerd[1510]: 2025-05-14 23:52:51.211 [INFO][4208] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.155a92f9eec8ef61b6b2f8d758ded3690f3b190b0ef6f444ef42f3dff801c046" host="localhost" May 14 23:52:51.265433 containerd[1510]: 2025-05-14 23:52:51.216 [INFO][4208] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" May 14 23:52:51.265433 containerd[1510]: 2025-05-14 23:52:51.221 [INFO][4208] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" May 14 23:52:51.265433 containerd[1510]: 2025-05-14 23:52:51.224 [INFO][4208] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" May 14 23:52:51.265433 containerd[1510]: 2025-05-14 23:52:51.228 [INFO][4208] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" May 14 23:52:51.265433 containerd[1510]: 2025-05-14 23:52:51.228 [INFO][4208] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.155a92f9eec8ef61b6b2f8d758ded3690f3b190b0ef6f444ef42f3dff801c046" host="localhost" May 14 23:52:51.265805 containerd[1510]: 2025-05-14 23:52:51.230 [INFO][4208] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.155a92f9eec8ef61b6b2f8d758ded3690f3b190b0ef6f444ef42f3dff801c046 May 14 23:52:51.265805 containerd[1510]: 2025-05-14 23:52:51.236 [INFO][4208] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.155a92f9eec8ef61b6b2f8d758ded3690f3b190b0ef6f444ef42f3dff801c046" host="localhost" May 14 23:52:51.265805 containerd[1510]: 2025-05-14 23:52:51.241 [INFO][4208] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.129/26] block=192.168.88.128/26 handle="k8s-pod-network.155a92f9eec8ef61b6b2f8d758ded3690f3b190b0ef6f444ef42f3dff801c046" host="localhost" May 14 23:52:51.265805 containerd[1510]: 2025-05-14 23:52:51.241 [INFO][4208] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.129/26] handle="k8s-pod-network.155a92f9eec8ef61b6b2f8d758ded3690f3b190b0ef6f444ef42f3dff801c046" host="localhost" May 14 23:52:51.265805 containerd[1510]: 2025-05-14 23:52:51.241 [INFO][4208] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 14 23:52:51.265805 containerd[1510]: 2025-05-14 23:52:51.241 [INFO][4208] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.129/26] IPv6=[] ContainerID="155a92f9eec8ef61b6b2f8d758ded3690f3b190b0ef6f444ef42f3dff801c046" HandleID="k8s-pod-network.155a92f9eec8ef61b6b2f8d758ded3690f3b190b0ef6f444ef42f3dff801c046" Workload="localhost-k8s-calico--apiserver--78f88fbddb--wtw8n-eth0" May 14 23:52:51.266009 containerd[1510]: 2025-05-14 23:52:51.244 [INFO][4193] cni-plugin/k8s.go 386: Populated endpoint ContainerID="155a92f9eec8ef61b6b2f8d758ded3690f3b190b0ef6f444ef42f3dff801c046" Namespace="calico-apiserver" Pod="calico-apiserver-78f88fbddb-wtw8n" WorkloadEndpoint="localhost-k8s-calico--apiserver--78f88fbddb--wtw8n-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--78f88fbddb--wtw8n-eth0", GenerateName:"calico-apiserver-78f88fbddb-", Namespace:"calico-apiserver", SelfLink:"", UID:"3b94c16d-c98d-4f96-8337-ac0791c35c61", ResourceVersion:"768", Generation:0, CreationTimestamp:time.Date(2025, time.May, 14, 23, 52, 22, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"78f88fbddb", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-78f88fbddb-wtw8n", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali622705137f5", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 14 23:52:51.266109 containerd[1510]: 2025-05-14 23:52:51.245 [INFO][4193] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.129/32] ContainerID="155a92f9eec8ef61b6b2f8d758ded3690f3b190b0ef6f444ef42f3dff801c046" Namespace="calico-apiserver" Pod="calico-apiserver-78f88fbddb-wtw8n" WorkloadEndpoint="localhost-k8s-calico--apiserver--78f88fbddb--wtw8n-eth0" May 14 23:52:51.266109 containerd[1510]: 2025-05-14 23:52:51.245 [INFO][4193] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali622705137f5 ContainerID="155a92f9eec8ef61b6b2f8d758ded3690f3b190b0ef6f444ef42f3dff801c046" Namespace="calico-apiserver" Pod="calico-apiserver-78f88fbddb-wtw8n" WorkloadEndpoint="localhost-k8s-calico--apiserver--78f88fbddb--wtw8n-eth0" May 14 23:52:51.266109 containerd[1510]: 2025-05-14 23:52:51.249 [INFO][4193] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="155a92f9eec8ef61b6b2f8d758ded3690f3b190b0ef6f444ef42f3dff801c046" Namespace="calico-apiserver" Pod="calico-apiserver-78f88fbddb-wtw8n" WorkloadEndpoint="localhost-k8s-calico--apiserver--78f88fbddb--wtw8n-eth0" May 14 23:52:51.266222 containerd[1510]: 2025-05-14 23:52:51.249 [INFO][4193] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="155a92f9eec8ef61b6b2f8d758ded3690f3b190b0ef6f444ef42f3dff801c046" Namespace="calico-apiserver" Pod="calico-apiserver-78f88fbddb-wtw8n" WorkloadEndpoint="localhost-k8s-calico--apiserver--78f88fbddb--wtw8n-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--78f88fbddb--wtw8n-eth0", GenerateName:"calico-apiserver-78f88fbddb-", Namespace:"calico-apiserver", SelfLink:"", UID:"3b94c16d-c98d-4f96-8337-ac0791c35c61", ResourceVersion:"768", Generation:0, CreationTimestamp:time.Date(2025, time.May, 14, 23, 52, 22, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"78f88fbddb", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"155a92f9eec8ef61b6b2f8d758ded3690f3b190b0ef6f444ef42f3dff801c046", Pod:"calico-apiserver-78f88fbddb-wtw8n", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali622705137f5", MAC:"de:64:cc:70:cf:4e", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 14 23:52:51.266325 containerd[1510]: 2025-05-14 23:52:51.260 [INFO][4193] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="155a92f9eec8ef61b6b2f8d758ded3690f3b190b0ef6f444ef42f3dff801c046" Namespace="calico-apiserver" Pod="calico-apiserver-78f88fbddb-wtw8n" WorkloadEndpoint="localhost-k8s-calico--apiserver--78f88fbddb--wtw8n-eth0" May 14 23:52:51.367418 containerd[1510]: time="2025-05-14T23:52:51.367332524Z" level=info msg="connecting to shim 155a92f9eec8ef61b6b2f8d758ded3690f3b190b0ef6f444ef42f3dff801c046" address="unix:///run/containerd/s/decac4b3035eb208e7c5e9f3523c1d0237be9489cdeeefda1e2e5f987bbdd6e6" namespace=k8s.io protocol=ttrpc version=3 May 14 23:52:51.403498 systemd[1]: Started cri-containerd-155a92f9eec8ef61b6b2f8d758ded3690f3b190b0ef6f444ef42f3dff801c046.scope - libcontainer container 155a92f9eec8ef61b6b2f8d758ded3690f3b190b0ef6f444ef42f3dff801c046. May 14 23:52:51.418630 systemd-resolved[1346]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address May 14 23:52:51.458148 containerd[1510]: time="2025-05-14T23:52:51.458092076Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-78f88fbddb-wtw8n,Uid:3b94c16d-c98d-4f96-8337-ac0791c35c61,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"155a92f9eec8ef61b6b2f8d758ded3690f3b190b0ef6f444ef42f3dff801c046\"" May 14 23:52:51.460088 containerd[1510]: time="2025-05-14T23:52:51.460001710Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.3\"" May 14 23:52:51.994278 systemd[1]: Started sshd@13-10.0.0.142:22-10.0.0.1:40968.service - OpenSSH per-connection server daemon (10.0.0.1:40968). May 14 23:52:52.045697 sshd[4276]: Accepted publickey for core from 10.0.0.1 port 40968 ssh2: RSA SHA256:lk2TkYBEL43KPVbrGyh3Ro8IB8NGN6uTNXzFyrYR01I May 14 23:52:52.048158 sshd-session[4276]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 23:52:52.053869 systemd-logind[1487]: New session 14 of user core. May 14 23:52:52.060479 systemd[1]: Started session-14.scope - Session 14 of User core. May 14 23:52:52.180391 sshd[4278]: Connection closed by 10.0.0.1 port 40968 May 14 23:52:52.180743 sshd-session[4276]: pam_unix(sshd:session): session closed for user core May 14 23:52:52.185384 systemd[1]: sshd@13-10.0.0.142:22-10.0.0.1:40968.service: Deactivated successfully. May 14 23:52:52.188031 systemd[1]: session-14.scope: Deactivated successfully. May 14 23:52:52.188976 systemd-logind[1487]: Session 14 logged out. Waiting for processes to exit. May 14 23:52:52.190113 systemd-logind[1487]: Removed session 14. May 14 23:52:52.378868 containerd[1510]: time="2025-05-14T23:52:52.378813439Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-57dfd6df95-fhrn2,Uid:aaa27ebc-a52f-4ebe-9a1d-8508c4e3e499,Namespace:calico-system,Attempt:0,}" May 14 23:52:52.379079 containerd[1510]: time="2025-05-14T23:52:52.378824800Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5bhtm,Uid:d0da08d5-ebd0-447d-b3dc-5ae18d60d27e,Namespace:kube-system,Attempt:0,}" May 14 23:52:52.530501 systemd-networkd[1437]: cali622705137f5: Gained IPv6LL May 14 23:52:52.896275 systemd-networkd[1437]: calief6e0cbb5b5: Link UP May 14 23:52:52.896546 systemd-networkd[1437]: calief6e0cbb5b5: Gained carrier May 14 23:52:52.918500 containerd[1510]: 2025-05-14 23:52:52.648 [INFO][4291] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--7db6d8ff4d--5bhtm-eth0 coredns-7db6d8ff4d- kube-system d0da08d5-ebd0-447d-b3dc-5ae18d60d27e 764 0 2025-05-14 23:52:14 +0000 UTC map[k8s-app:kube-dns pod-template-hash:7db6d8ff4d projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-7db6d8ff4d-5bhtm eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] calief6e0cbb5b5 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="1e53d328ef6e4d4882bd71353d6e3ac065d280406d845623c500383ae136c728" Namespace="kube-system" Pod="coredns-7db6d8ff4d-5bhtm" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--5bhtm-" May 14 23:52:52.918500 containerd[1510]: 2025-05-14 23:52:52.649 [INFO][4291] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="1e53d328ef6e4d4882bd71353d6e3ac065d280406d845623c500383ae136c728" Namespace="kube-system" Pod="coredns-7db6d8ff4d-5bhtm" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--5bhtm-eth0" May 14 23:52:52.918500 containerd[1510]: 2025-05-14 23:52:52.720 [INFO][4304] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="1e53d328ef6e4d4882bd71353d6e3ac065d280406d845623c500383ae136c728" HandleID="k8s-pod-network.1e53d328ef6e4d4882bd71353d6e3ac065d280406d845623c500383ae136c728" Workload="localhost-k8s-coredns--7db6d8ff4d--5bhtm-eth0" May 14 23:52:52.919029 containerd[1510]: 2025-05-14 23:52:52.728 [INFO][4304] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="1e53d328ef6e4d4882bd71353d6e3ac065d280406d845623c500383ae136c728" HandleID="k8s-pod-network.1e53d328ef6e4d4882bd71353d6e3ac065d280406d845623c500383ae136c728" Workload="localhost-k8s-coredns--7db6d8ff4d--5bhtm-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000297170), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-7db6d8ff4d-5bhtm", "timestamp":"2025-05-14 23:52:52.7209176 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 14 23:52:52.919029 containerd[1510]: 2025-05-14 23:52:52.729 [INFO][4304] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 14 23:52:52.919029 containerd[1510]: 2025-05-14 23:52:52.729 [INFO][4304] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 14 23:52:52.919029 containerd[1510]: 2025-05-14 23:52:52.729 [INFO][4304] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' May 14 23:52:52.919029 containerd[1510]: 2025-05-14 23:52:52.730 [INFO][4304] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.1e53d328ef6e4d4882bd71353d6e3ac065d280406d845623c500383ae136c728" host="localhost" May 14 23:52:52.919029 containerd[1510]: 2025-05-14 23:52:52.734 [INFO][4304] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" May 14 23:52:52.919029 containerd[1510]: 2025-05-14 23:52:52.739 [INFO][4304] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" May 14 23:52:52.919029 containerd[1510]: 2025-05-14 23:52:52.740 [INFO][4304] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" May 14 23:52:52.919029 containerd[1510]: 2025-05-14 23:52:52.742 [INFO][4304] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" May 14 23:52:52.919029 containerd[1510]: 2025-05-14 23:52:52.742 [INFO][4304] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.1e53d328ef6e4d4882bd71353d6e3ac065d280406d845623c500383ae136c728" host="localhost" May 14 23:52:52.919478 containerd[1510]: 2025-05-14 23:52:52.743 [INFO][4304] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.1e53d328ef6e4d4882bd71353d6e3ac065d280406d845623c500383ae136c728 May 14 23:52:52.919478 containerd[1510]: 2025-05-14 23:52:52.760 [INFO][4304] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.1e53d328ef6e4d4882bd71353d6e3ac065d280406d845623c500383ae136c728" host="localhost" May 14 23:52:52.919478 containerd[1510]: 2025-05-14 23:52:52.888 [INFO][4304] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.130/26] block=192.168.88.128/26 handle="k8s-pod-network.1e53d328ef6e4d4882bd71353d6e3ac065d280406d845623c500383ae136c728" host="localhost" May 14 23:52:52.919478 containerd[1510]: 2025-05-14 23:52:52.888 [INFO][4304] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.130/26] handle="k8s-pod-network.1e53d328ef6e4d4882bd71353d6e3ac065d280406d845623c500383ae136c728" host="localhost" May 14 23:52:52.919478 containerd[1510]: 2025-05-14 23:52:52.888 [INFO][4304] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 14 23:52:52.919478 containerd[1510]: 2025-05-14 23:52:52.888 [INFO][4304] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.130/26] IPv6=[] ContainerID="1e53d328ef6e4d4882bd71353d6e3ac065d280406d845623c500383ae136c728" HandleID="k8s-pod-network.1e53d328ef6e4d4882bd71353d6e3ac065d280406d845623c500383ae136c728" Workload="localhost-k8s-coredns--7db6d8ff4d--5bhtm-eth0" May 14 23:52:52.919597 containerd[1510]: 2025-05-14 23:52:52.892 [INFO][4291] cni-plugin/k8s.go 386: Populated endpoint ContainerID="1e53d328ef6e4d4882bd71353d6e3ac065d280406d845623c500383ae136c728" Namespace="kube-system" Pod="coredns-7db6d8ff4d-5bhtm" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--5bhtm-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7db6d8ff4d--5bhtm-eth0", GenerateName:"coredns-7db6d8ff4d-", Namespace:"kube-system", SelfLink:"", UID:"d0da08d5-ebd0-447d-b3dc-5ae18d60d27e", ResourceVersion:"764", Generation:0, CreationTimestamp:time.Date(2025, time.May, 14, 23, 52, 14, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7db6d8ff4d", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-7db6d8ff4d-5bhtm", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calief6e0cbb5b5", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} May 14 23:52:52.919918 containerd[1510]: 2025-05-14 23:52:52.892 [INFO][4291] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.130/32] ContainerID="1e53d328ef6e4d4882bd71353d6e3ac065d280406d845623c500383ae136c728" Namespace="kube-system" Pod="coredns-7db6d8ff4d-5bhtm" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--5bhtm-eth0" May 14 23:52:52.919918 containerd[1510]: 2025-05-14 23:52:52.892 [INFO][4291] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calief6e0cbb5b5 ContainerID="1e53d328ef6e4d4882bd71353d6e3ac065d280406d845623c500383ae136c728" Namespace="kube-system" Pod="coredns-7db6d8ff4d-5bhtm" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--5bhtm-eth0" May 14 23:52:52.919918 containerd[1510]: 2025-05-14 23:52:52.895 [INFO][4291] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="1e53d328ef6e4d4882bd71353d6e3ac065d280406d845623c500383ae136c728" Namespace="kube-system" Pod="coredns-7db6d8ff4d-5bhtm" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--5bhtm-eth0" May 14 23:52:52.920020 containerd[1510]: 2025-05-14 23:52:52.895 [INFO][4291] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="1e53d328ef6e4d4882bd71353d6e3ac065d280406d845623c500383ae136c728" Namespace="kube-system" Pod="coredns-7db6d8ff4d-5bhtm" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--5bhtm-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7db6d8ff4d--5bhtm-eth0", GenerateName:"coredns-7db6d8ff4d-", Namespace:"kube-system", SelfLink:"", UID:"d0da08d5-ebd0-447d-b3dc-5ae18d60d27e", ResourceVersion:"764", Generation:0, CreationTimestamp:time.Date(2025, time.May, 14, 23, 52, 14, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7db6d8ff4d", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"1e53d328ef6e4d4882bd71353d6e3ac065d280406d845623c500383ae136c728", Pod:"coredns-7db6d8ff4d-5bhtm", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calief6e0cbb5b5", MAC:"46:ad:9d:d6:ff:aa", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} May 14 23:52:52.920020 containerd[1510]: 2025-05-14 23:52:52.914 [INFO][4291] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="1e53d328ef6e4d4882bd71353d6e3ac065d280406d845623c500383ae136c728" Namespace="kube-system" Pod="coredns-7db6d8ff4d-5bhtm" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--5bhtm-eth0" May 14 23:52:52.964956 containerd[1510]: time="2025-05-14T23:52:52.964906532Z" level=info msg="connecting to shim 1e53d328ef6e4d4882bd71353d6e3ac065d280406d845623c500383ae136c728" address="unix:///run/containerd/s/2241723ec0f61bc4a6915ddfb893708d3f029df74b9d2d0f0595748b29b20553" namespace=k8s.io protocol=ttrpc version=3 May 14 23:52:53.003511 systemd[1]: Started cri-containerd-1e53d328ef6e4d4882bd71353d6e3ac065d280406d845623c500383ae136c728.scope - libcontainer container 1e53d328ef6e4d4882bd71353d6e3ac065d280406d845623c500383ae136c728. May 14 23:52:53.019760 systemd-resolved[1346]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address May 14 23:52:53.040275 systemd-networkd[1437]: calib61933789e5: Link UP May 14 23:52:53.040949 systemd-networkd[1437]: calib61933789e5: Gained carrier May 14 23:52:53.109558 containerd[1510]: 2025-05-14 23:52:52.891 [INFO][4312] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--kube--controllers--57dfd6df95--fhrn2-eth0 calico-kube-controllers-57dfd6df95- calico-system aaa27ebc-a52f-4ebe-9a1d-8508c4e3e499 771 0 2025-05-14 23:52:22 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:57dfd6df95 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s localhost calico-kube-controllers-57dfd6df95-fhrn2 eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] calib61933789e5 [] []}} ContainerID="9dcc7e7dcb03816f78c3144c5e0d390722433131d7b41383d2c4187f062647f0" Namespace="calico-system" Pod="calico-kube-controllers-57dfd6df95-fhrn2" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--57dfd6df95--fhrn2-" May 14 23:52:53.109558 containerd[1510]: 2025-05-14 23:52:52.891 [INFO][4312] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="9dcc7e7dcb03816f78c3144c5e0d390722433131d7b41383d2c4187f062647f0" Namespace="calico-system" Pod="calico-kube-controllers-57dfd6df95-fhrn2" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--57dfd6df95--fhrn2-eth0" May 14 23:52:53.109558 containerd[1510]: 2025-05-14 23:52:52.946 [INFO][4331] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="9dcc7e7dcb03816f78c3144c5e0d390722433131d7b41383d2c4187f062647f0" HandleID="k8s-pod-network.9dcc7e7dcb03816f78c3144c5e0d390722433131d7b41383d2c4187f062647f0" Workload="localhost-k8s-calico--kube--controllers--57dfd6df95--fhrn2-eth0" May 14 23:52:53.109558 containerd[1510]: 2025-05-14 23:52:52.955 [INFO][4331] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="9dcc7e7dcb03816f78c3144c5e0d390722433131d7b41383d2c4187f062647f0" HandleID="k8s-pod-network.9dcc7e7dcb03816f78c3144c5e0d390722433131d7b41383d2c4187f062647f0" Workload="localhost-k8s-calico--kube--controllers--57dfd6df95--fhrn2-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00027fb10), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"calico-kube-controllers-57dfd6df95-fhrn2", "timestamp":"2025-05-14 23:52:52.946212728 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 14 23:52:53.109558 containerd[1510]: 2025-05-14 23:52:52.955 [INFO][4331] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 14 23:52:53.109558 containerd[1510]: 2025-05-14 23:52:52.955 [INFO][4331] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 14 23:52:53.109558 containerd[1510]: 2025-05-14 23:52:52.956 [INFO][4331] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' May 14 23:52:53.109558 containerd[1510]: 2025-05-14 23:52:52.958 [INFO][4331] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.9dcc7e7dcb03816f78c3144c5e0d390722433131d7b41383d2c4187f062647f0" host="localhost" May 14 23:52:53.109558 containerd[1510]: 2025-05-14 23:52:52.965 [INFO][4331] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" May 14 23:52:53.109558 containerd[1510]: 2025-05-14 23:52:52.973 [INFO][4331] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" May 14 23:52:53.109558 containerd[1510]: 2025-05-14 23:52:52.975 [INFO][4331] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" May 14 23:52:53.109558 containerd[1510]: 2025-05-14 23:52:52.983 [INFO][4331] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" May 14 23:52:53.109558 containerd[1510]: 2025-05-14 23:52:52.983 [INFO][4331] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.9dcc7e7dcb03816f78c3144c5e0d390722433131d7b41383d2c4187f062647f0" host="localhost" May 14 23:52:53.109558 containerd[1510]: 2025-05-14 23:52:52.985 [INFO][4331] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.9dcc7e7dcb03816f78c3144c5e0d390722433131d7b41383d2c4187f062647f0 May 14 23:52:53.109558 containerd[1510]: 2025-05-14 23:52:53.019 [INFO][4331] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.9dcc7e7dcb03816f78c3144c5e0d390722433131d7b41383d2c4187f062647f0" host="localhost" May 14 23:52:53.109558 containerd[1510]: 2025-05-14 23:52:53.034 [INFO][4331] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.131/26] block=192.168.88.128/26 handle="k8s-pod-network.9dcc7e7dcb03816f78c3144c5e0d390722433131d7b41383d2c4187f062647f0" host="localhost" May 14 23:52:53.109558 containerd[1510]: 2025-05-14 23:52:53.034 [INFO][4331] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.131/26] handle="k8s-pod-network.9dcc7e7dcb03816f78c3144c5e0d390722433131d7b41383d2c4187f062647f0" host="localhost" May 14 23:52:53.109558 containerd[1510]: 2025-05-14 23:52:53.034 [INFO][4331] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 14 23:52:53.109558 containerd[1510]: 2025-05-14 23:52:53.034 [INFO][4331] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.131/26] IPv6=[] ContainerID="9dcc7e7dcb03816f78c3144c5e0d390722433131d7b41383d2c4187f062647f0" HandleID="k8s-pod-network.9dcc7e7dcb03816f78c3144c5e0d390722433131d7b41383d2c4187f062647f0" Workload="localhost-k8s-calico--kube--controllers--57dfd6df95--fhrn2-eth0" May 14 23:52:53.110386 containerd[1510]: 2025-05-14 23:52:53.037 [INFO][4312] cni-plugin/k8s.go 386: Populated endpoint ContainerID="9dcc7e7dcb03816f78c3144c5e0d390722433131d7b41383d2c4187f062647f0" Namespace="calico-system" Pod="calico-kube-controllers-57dfd6df95-fhrn2" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--57dfd6df95--fhrn2-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--57dfd6df95--fhrn2-eth0", GenerateName:"calico-kube-controllers-57dfd6df95-", Namespace:"calico-system", SelfLink:"", UID:"aaa27ebc-a52f-4ebe-9a1d-8508c4e3e499", ResourceVersion:"771", Generation:0, CreationTimestamp:time.Date(2025, time.May, 14, 23, 52, 22, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"57dfd6df95", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-kube-controllers-57dfd6df95-fhrn2", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calib61933789e5", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 14 23:52:53.110386 containerd[1510]: 2025-05-14 23:52:53.037 [INFO][4312] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.131/32] ContainerID="9dcc7e7dcb03816f78c3144c5e0d390722433131d7b41383d2c4187f062647f0" Namespace="calico-system" Pod="calico-kube-controllers-57dfd6df95-fhrn2" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--57dfd6df95--fhrn2-eth0" May 14 23:52:53.110386 containerd[1510]: 2025-05-14 23:52:53.037 [INFO][4312] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calib61933789e5 ContainerID="9dcc7e7dcb03816f78c3144c5e0d390722433131d7b41383d2c4187f062647f0" Namespace="calico-system" Pod="calico-kube-controllers-57dfd6df95-fhrn2" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--57dfd6df95--fhrn2-eth0" May 14 23:52:53.110386 containerd[1510]: 2025-05-14 23:52:53.041 [INFO][4312] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="9dcc7e7dcb03816f78c3144c5e0d390722433131d7b41383d2c4187f062647f0" Namespace="calico-system" Pod="calico-kube-controllers-57dfd6df95-fhrn2" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--57dfd6df95--fhrn2-eth0" May 14 23:52:53.110386 containerd[1510]: 2025-05-14 23:52:53.041 [INFO][4312] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="9dcc7e7dcb03816f78c3144c5e0d390722433131d7b41383d2c4187f062647f0" Namespace="calico-system" Pod="calico-kube-controllers-57dfd6df95-fhrn2" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--57dfd6df95--fhrn2-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--57dfd6df95--fhrn2-eth0", GenerateName:"calico-kube-controllers-57dfd6df95-", Namespace:"calico-system", SelfLink:"", UID:"aaa27ebc-a52f-4ebe-9a1d-8508c4e3e499", ResourceVersion:"771", Generation:0, CreationTimestamp:time.Date(2025, time.May, 14, 23, 52, 22, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"57dfd6df95", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"9dcc7e7dcb03816f78c3144c5e0d390722433131d7b41383d2c4187f062647f0", Pod:"calico-kube-controllers-57dfd6df95-fhrn2", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calib61933789e5", MAC:"e2:6a:ba:19:da:ee", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 14 23:52:53.110386 containerd[1510]: 2025-05-14 23:52:53.105 [INFO][4312] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="9dcc7e7dcb03816f78c3144c5e0d390722433131d7b41383d2c4187f062647f0" Namespace="calico-system" Pod="calico-kube-controllers-57dfd6df95-fhrn2" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--57dfd6df95--fhrn2-eth0" May 14 23:52:53.115474 containerd[1510]: time="2025-05-14T23:52:53.115428229Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5bhtm,Uid:d0da08d5-ebd0-447d-b3dc-5ae18d60d27e,Namespace:kube-system,Attempt:0,} returns sandbox id \"1e53d328ef6e4d4882bd71353d6e3ac065d280406d845623c500383ae136c728\"" May 14 23:52:53.118887 containerd[1510]: time="2025-05-14T23:52:53.118855402Z" level=info msg="CreateContainer within sandbox \"1e53d328ef6e4d4882bd71353d6e3ac065d280406d845623c500383ae136c728\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" May 14 23:52:53.509072 containerd[1510]: time="2025-05-14T23:52:53.509017130Z" level=info msg="Container dae74e536b2552c95e4d08708dc4e8cd0b9283db99c012a3352f022248b4c15d: CDI devices from CRI Config.CDIDevices: []" May 14 23:52:53.523719 containerd[1510]: time="2025-05-14T23:52:53.523603986Z" level=info msg="CreateContainer within sandbox \"1e53d328ef6e4d4882bd71353d6e3ac065d280406d845623c500383ae136c728\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"dae74e536b2552c95e4d08708dc4e8cd0b9283db99c012a3352f022248b4c15d\"" May 14 23:52:53.524802 containerd[1510]: time="2025-05-14T23:52:53.524775815Z" level=info msg="StartContainer for \"dae74e536b2552c95e4d08708dc4e8cd0b9283db99c012a3352f022248b4c15d\"" May 14 23:52:53.526328 containerd[1510]: time="2025-05-14T23:52:53.525891058Z" level=info msg="connecting to shim dae74e536b2552c95e4d08708dc4e8cd0b9283db99c012a3352f022248b4c15d" address="unix:///run/containerd/s/2241723ec0f61bc4a6915ddfb893708d3f029df74b9d2d0f0595748b29b20553" protocol=ttrpc version=3 May 14 23:52:53.555498 systemd[1]: Started cri-containerd-dae74e536b2552c95e4d08708dc4e8cd0b9283db99c012a3352f022248b4c15d.scope - libcontainer container dae74e536b2552c95e4d08708dc4e8cd0b9283db99c012a3352f022248b4c15d. May 14 23:52:53.560366 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount876853774.mount: Deactivated successfully. May 14 23:52:53.638046 containerd[1510]: time="2025-05-14T23:52:53.637997719Z" level=info msg="connecting to shim 9dcc7e7dcb03816f78c3144c5e0d390722433131d7b41383d2c4187f062647f0" address="unix:///run/containerd/s/961f062b58e2de23d9ad00676f8cd38ca1edbb528df4302ec990d83b42bf6b2e" namespace=k8s.io protocol=ttrpc version=3 May 14 23:52:53.667507 systemd[1]: Started cri-containerd-9dcc7e7dcb03816f78c3144c5e0d390722433131d7b41383d2c4187f062647f0.scope - libcontainer container 9dcc7e7dcb03816f78c3144c5e0d390722433131d7b41383d2c4187f062647f0. May 14 23:52:53.685864 systemd-resolved[1346]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address May 14 23:52:53.708144 containerd[1510]: time="2025-05-14T23:52:53.708084401Z" level=info msg="StartContainer for \"dae74e536b2552c95e4d08708dc4e8cd0b9283db99c012a3352f022248b4c15d\" returns successfully" May 14 23:52:53.759069 containerd[1510]: time="2025-05-14T23:52:53.758881282Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-57dfd6df95-fhrn2,Uid:aaa27ebc-a52f-4ebe-9a1d-8508c4e3e499,Namespace:calico-system,Attempt:0,} returns sandbox id \"9dcc7e7dcb03816f78c3144c5e0d390722433131d7b41383d2c4187f062647f0\"" May 14 23:52:54.194483 systemd-networkd[1437]: calib61933789e5: Gained IPv6LL May 14 23:52:54.380138 containerd[1510]: time="2025-05-14T23:52:54.380070378Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-xq4zr,Uid:d5000901-d9c1-41f3-9444-837fe018e119,Namespace:kube-system,Attempt:0,}" May 14 23:52:54.502836 systemd-networkd[1437]: calic1fd6edbd56: Link UP May 14 23:52:54.503687 systemd-networkd[1437]: calic1fd6edbd56: Gained carrier May 14 23:52:54.519523 containerd[1510]: 2025-05-14 23:52:54.425 [INFO][4502] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--7db6d8ff4d--xq4zr-eth0 coredns-7db6d8ff4d- kube-system d5000901-d9c1-41f3-9444-837fe018e119 770 0 2025-05-14 23:52:14 +0000 UTC map[k8s-app:kube-dns pod-template-hash:7db6d8ff4d projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-7db6d8ff4d-xq4zr eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] calic1fd6edbd56 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="73f38fdd8dbd653352694c3fc224a33e604360f3e21087f56d309937271229a2" Namespace="kube-system" Pod="coredns-7db6d8ff4d-xq4zr" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--xq4zr-" May 14 23:52:54.519523 containerd[1510]: 2025-05-14 23:52:54.426 [INFO][4502] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="73f38fdd8dbd653352694c3fc224a33e604360f3e21087f56d309937271229a2" Namespace="kube-system" Pod="coredns-7db6d8ff4d-xq4zr" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--xq4zr-eth0" May 14 23:52:54.519523 containerd[1510]: 2025-05-14 23:52:54.459 [INFO][4517] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="73f38fdd8dbd653352694c3fc224a33e604360f3e21087f56d309937271229a2" HandleID="k8s-pod-network.73f38fdd8dbd653352694c3fc224a33e604360f3e21087f56d309937271229a2" Workload="localhost-k8s-coredns--7db6d8ff4d--xq4zr-eth0" May 14 23:52:54.519523 containerd[1510]: 2025-05-14 23:52:54.468 [INFO][4517] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="73f38fdd8dbd653352694c3fc224a33e604360f3e21087f56d309937271229a2" HandleID="k8s-pod-network.73f38fdd8dbd653352694c3fc224a33e604360f3e21087f56d309937271229a2" Workload="localhost-k8s-coredns--7db6d8ff4d--xq4zr-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0000c6500), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-7db6d8ff4d-xq4zr", "timestamp":"2025-05-14 23:52:54.459789893 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 14 23:52:54.519523 containerd[1510]: 2025-05-14 23:52:54.468 [INFO][4517] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 14 23:52:54.519523 containerd[1510]: 2025-05-14 23:52:54.468 [INFO][4517] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 14 23:52:54.519523 containerd[1510]: 2025-05-14 23:52:54.468 [INFO][4517] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' May 14 23:52:54.519523 containerd[1510]: 2025-05-14 23:52:54.470 [INFO][4517] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.73f38fdd8dbd653352694c3fc224a33e604360f3e21087f56d309937271229a2" host="localhost" May 14 23:52:54.519523 containerd[1510]: 2025-05-14 23:52:54.473 [INFO][4517] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" May 14 23:52:54.519523 containerd[1510]: 2025-05-14 23:52:54.477 [INFO][4517] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" May 14 23:52:54.519523 containerd[1510]: 2025-05-14 23:52:54.479 [INFO][4517] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" May 14 23:52:54.519523 containerd[1510]: 2025-05-14 23:52:54.481 [INFO][4517] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" May 14 23:52:54.519523 containerd[1510]: 2025-05-14 23:52:54.481 [INFO][4517] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.73f38fdd8dbd653352694c3fc224a33e604360f3e21087f56d309937271229a2" host="localhost" May 14 23:52:54.519523 containerd[1510]: 2025-05-14 23:52:54.483 [INFO][4517] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.73f38fdd8dbd653352694c3fc224a33e604360f3e21087f56d309937271229a2 May 14 23:52:54.519523 containerd[1510]: 2025-05-14 23:52:54.487 [INFO][4517] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.73f38fdd8dbd653352694c3fc224a33e604360f3e21087f56d309937271229a2" host="localhost" May 14 23:52:54.519523 containerd[1510]: 2025-05-14 23:52:54.492 [INFO][4517] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.132/26] block=192.168.88.128/26 handle="k8s-pod-network.73f38fdd8dbd653352694c3fc224a33e604360f3e21087f56d309937271229a2" host="localhost" May 14 23:52:54.519523 containerd[1510]: 2025-05-14 23:52:54.492 [INFO][4517] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.132/26] handle="k8s-pod-network.73f38fdd8dbd653352694c3fc224a33e604360f3e21087f56d309937271229a2" host="localhost" May 14 23:52:54.519523 containerd[1510]: 2025-05-14 23:52:54.492 [INFO][4517] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 14 23:52:54.519523 containerd[1510]: 2025-05-14 23:52:54.492 [INFO][4517] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.132/26] IPv6=[] ContainerID="73f38fdd8dbd653352694c3fc224a33e604360f3e21087f56d309937271229a2" HandleID="k8s-pod-network.73f38fdd8dbd653352694c3fc224a33e604360f3e21087f56d309937271229a2" Workload="localhost-k8s-coredns--7db6d8ff4d--xq4zr-eth0" May 14 23:52:54.520116 containerd[1510]: 2025-05-14 23:52:54.497 [INFO][4502] cni-plugin/k8s.go 386: Populated endpoint ContainerID="73f38fdd8dbd653352694c3fc224a33e604360f3e21087f56d309937271229a2" Namespace="kube-system" Pod="coredns-7db6d8ff4d-xq4zr" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--xq4zr-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7db6d8ff4d--xq4zr-eth0", GenerateName:"coredns-7db6d8ff4d-", Namespace:"kube-system", SelfLink:"", UID:"d5000901-d9c1-41f3-9444-837fe018e119", ResourceVersion:"770", Generation:0, CreationTimestamp:time.Date(2025, time.May, 14, 23, 52, 14, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7db6d8ff4d", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-7db6d8ff4d-xq4zr", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calic1fd6edbd56", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} May 14 23:52:54.520116 containerd[1510]: 2025-05-14 23:52:54.497 [INFO][4502] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.132/32] ContainerID="73f38fdd8dbd653352694c3fc224a33e604360f3e21087f56d309937271229a2" Namespace="kube-system" Pod="coredns-7db6d8ff4d-xq4zr" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--xq4zr-eth0" May 14 23:52:54.520116 containerd[1510]: 2025-05-14 23:52:54.497 [INFO][4502] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calic1fd6edbd56 ContainerID="73f38fdd8dbd653352694c3fc224a33e604360f3e21087f56d309937271229a2" Namespace="kube-system" Pod="coredns-7db6d8ff4d-xq4zr" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--xq4zr-eth0" May 14 23:52:54.520116 containerd[1510]: 2025-05-14 23:52:54.505 [INFO][4502] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="73f38fdd8dbd653352694c3fc224a33e604360f3e21087f56d309937271229a2" Namespace="kube-system" Pod="coredns-7db6d8ff4d-xq4zr" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--xq4zr-eth0" May 14 23:52:54.520116 containerd[1510]: 2025-05-14 23:52:54.505 [INFO][4502] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="73f38fdd8dbd653352694c3fc224a33e604360f3e21087f56d309937271229a2" Namespace="kube-system" Pod="coredns-7db6d8ff4d-xq4zr" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--xq4zr-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7db6d8ff4d--xq4zr-eth0", GenerateName:"coredns-7db6d8ff4d-", Namespace:"kube-system", SelfLink:"", UID:"d5000901-d9c1-41f3-9444-837fe018e119", ResourceVersion:"770", Generation:0, CreationTimestamp:time.Date(2025, time.May, 14, 23, 52, 14, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7db6d8ff4d", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"73f38fdd8dbd653352694c3fc224a33e604360f3e21087f56d309937271229a2", Pod:"coredns-7db6d8ff4d-xq4zr", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calic1fd6edbd56", MAC:"0a:3c:ec:c8:63:ff", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} May 14 23:52:54.520116 containerd[1510]: 2025-05-14 23:52:54.516 [INFO][4502] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="73f38fdd8dbd653352694c3fc224a33e604360f3e21087f56d309937271229a2" Namespace="kube-system" Pod="coredns-7db6d8ff4d-xq4zr" WorkloadEndpoint="localhost-k8s-coredns--7db6d8ff4d--xq4zr-eth0" May 14 23:52:54.581339 kubelet[2749]: I0514 23:52:54.580854 2749 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7db6d8ff4d-5bhtm" podStartSLOduration=40.58083414 podStartE2EDuration="40.58083414s" podCreationTimestamp="2025-05-14 23:52:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-14 23:52:54.566845018 +0000 UTC m=+54.286075788" watchObservedRunningTime="2025-05-14 23:52:54.58083414 +0000 UTC m=+54.300064920" May 14 23:52:54.621542 containerd[1510]: time="2025-05-14T23:52:54.620810902Z" level=info msg="connecting to shim 73f38fdd8dbd653352694c3fc224a33e604360f3e21087f56d309937271229a2" address="unix:///run/containerd/s/1bb7170e015d5d206d898641493b62e1f820d67564e0743a4c91583eabaa1a11" namespace=k8s.io protocol=ttrpc version=3 May 14 23:52:54.676520 systemd[1]: Started cri-containerd-73f38fdd8dbd653352694c3fc224a33e604360f3e21087f56d309937271229a2.scope - libcontainer container 73f38fdd8dbd653352694c3fc224a33e604360f3e21087f56d309937271229a2. May 14 23:52:54.693009 systemd-resolved[1346]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address May 14 23:52:54.898592 systemd-networkd[1437]: calief6e0cbb5b5: Gained IPv6LL May 14 23:52:55.063977 containerd[1510]: time="2025-05-14T23:52:55.063908966Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-xq4zr,Uid:d5000901-d9c1-41f3-9444-837fe018e119,Namespace:kube-system,Attempt:0,} returns sandbox id \"73f38fdd8dbd653352694c3fc224a33e604360f3e21087f56d309937271229a2\"" May 14 23:52:55.066425 containerd[1510]: time="2025-05-14T23:52:55.066374432Z" level=info msg="CreateContainer within sandbox \"73f38fdd8dbd653352694c3fc224a33e604360f3e21087f56d309937271229a2\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" May 14 23:52:55.100189 containerd[1510]: time="2025-05-14T23:52:55.100129848Z" level=info msg="Container 6f01b87172514be0aca3cbb39a1d54bdaffda7583751fa779293c54d3063be72: CDI devices from CRI Config.CDIDevices: []" May 14 23:52:55.108363 containerd[1510]: time="2025-05-14T23:52:55.108258111Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:52:55.109362 containerd[1510]: time="2025-05-14T23:52:55.109161507Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.29.3: active requests=0, bytes read=43021437" May 14 23:52:55.109680 containerd[1510]: time="2025-05-14T23:52:55.109629474Z" level=info msg="CreateContainer within sandbox \"73f38fdd8dbd653352694c3fc224a33e604360f3e21087f56d309937271229a2\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"6f01b87172514be0aca3cbb39a1d54bdaffda7583751fa779293c54d3063be72\"" May 14 23:52:55.112012 containerd[1510]: time="2025-05-14T23:52:55.110465263Z" level=info msg="StartContainer for \"6f01b87172514be0aca3cbb39a1d54bdaffda7583751fa779293c54d3063be72\"" May 14 23:52:55.112012 containerd[1510]: time="2025-05-14T23:52:55.111190403Z" level=info msg="ImageCreate event name:\"sha256:b1960e792987d99ee8f3583d7354dcd25a683cf854e8f10322ca7eeb83128532\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:52:55.112427 containerd[1510]: time="2025-05-14T23:52:55.112395656Z" level=info msg="connecting to shim 6f01b87172514be0aca3cbb39a1d54bdaffda7583751fa779293c54d3063be72" address="unix:///run/containerd/s/1bb7170e015d5d206d898641493b62e1f820d67564e0743a4c91583eabaa1a11" protocol=ttrpc version=3 May 14 23:52:55.116308 containerd[1510]: time="2025-05-14T23:52:55.116235491Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:bcb659f25f9aebaa389ed1dbb65edb39478ddf82c57d07d8da474e8cab38d77b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:52:55.119055 containerd[1510]: time="2025-05-14T23:52:55.117017950Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.29.3\" with image id \"sha256:b1960e792987d99ee8f3583d7354dcd25a683cf854e8f10322ca7eeb83128532\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:bcb659f25f9aebaa389ed1dbb65edb39478ddf82c57d07d8da474e8cab38d77b\", size \"44514075\" in 3.656971175s" May 14 23:52:55.119055 containerd[1510]: time="2025-05-14T23:52:55.117050451Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.3\" returns image reference \"sha256:b1960e792987d99ee8f3583d7354dcd25a683cf854e8f10322ca7eeb83128532\"" May 14 23:52:55.122607 containerd[1510]: time="2025-05-14T23:52:55.121048914Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\"" May 14 23:52:55.123054 containerd[1510]: time="2025-05-14T23:52:55.123016236Z" level=info msg="CreateContainer within sandbox \"155a92f9eec8ef61b6b2f8d758ded3690f3b190b0ef6f444ef42f3dff801c046\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" May 14 23:52:55.141008 systemd[1]: Started cri-containerd-6f01b87172514be0aca3cbb39a1d54bdaffda7583751fa779293c54d3063be72.scope - libcontainer container 6f01b87172514be0aca3cbb39a1d54bdaffda7583751fa779293c54d3063be72. May 14 23:52:55.143238 containerd[1510]: time="2025-05-14T23:52:55.143168583Z" level=info msg="Container 0a478f6607ee2389f6a2a87b54e61d705094d758b382ea35dfb24f02c3a9778a: CDI devices from CRI Config.CDIDevices: []" May 14 23:52:55.155253 containerd[1510]: time="2025-05-14T23:52:55.155060350Z" level=info msg="CreateContainer within sandbox \"155a92f9eec8ef61b6b2f8d758ded3690f3b190b0ef6f444ef42f3dff801c046\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"0a478f6607ee2389f6a2a87b54e61d705094d758b382ea35dfb24f02c3a9778a\"" May 14 23:52:55.159407 containerd[1510]: time="2025-05-14T23:52:55.157231534Z" level=info msg="StartContainer for \"0a478f6607ee2389f6a2a87b54e61d705094d758b382ea35dfb24f02c3a9778a\"" May 14 23:52:55.159407 containerd[1510]: time="2025-05-14T23:52:55.158821968Z" level=info msg="connecting to shim 0a478f6607ee2389f6a2a87b54e61d705094d758b382ea35dfb24f02c3a9778a" address="unix:///run/containerd/s/decac4b3035eb208e7c5e9f3523c1d0237be9489cdeeefda1e2e5f987bbdd6e6" protocol=ttrpc version=3 May 14 23:52:55.186628 systemd[1]: Started cri-containerd-0a478f6607ee2389f6a2a87b54e61d705094d758b382ea35dfb24f02c3a9778a.scope - libcontainer container 0a478f6607ee2389f6a2a87b54e61d705094d758b382ea35dfb24f02c3a9778a. May 14 23:52:55.206367 containerd[1510]: time="2025-05-14T23:52:55.206280509Z" level=info msg="StartContainer for \"6f01b87172514be0aca3cbb39a1d54bdaffda7583751fa779293c54d3063be72\" returns successfully" May 14 23:52:55.287257 containerd[1510]: time="2025-05-14T23:52:55.287207745Z" level=info msg="StartContainer for \"0a478f6607ee2389f6a2a87b54e61d705094d758b382ea35dfb24f02c3a9778a\" returns successfully" May 14 23:52:55.379065 containerd[1510]: time="2025-05-14T23:52:55.379006964Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-78f88fbddb-c68xl,Uid:e558cb7f-428e-4fae-a45c-cc274d3aa581,Namespace:calico-apiserver,Attempt:0,}" May 14 23:52:55.519878 systemd-networkd[1437]: cali04b0ed547f4: Link UP May 14 23:52:55.520134 systemd-networkd[1437]: cali04b0ed547f4: Gained carrier May 14 23:52:55.543746 containerd[1510]: 2025-05-14 23:52:55.438 [INFO][4664] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--78f88fbddb--c68xl-eth0 calico-apiserver-78f88fbddb- calico-apiserver e558cb7f-428e-4fae-a45c-cc274d3aa581 769 0 2025-05-14 23:52:22 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:78f88fbddb projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-78f88fbddb-c68xl eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali04b0ed547f4 [] []}} ContainerID="4f4bc667d65a0fb43d566a520efd1f1d90ec74798a55fc1a55054a3ea5d2cde2" Namespace="calico-apiserver" Pod="calico-apiserver-78f88fbddb-c68xl" WorkloadEndpoint="localhost-k8s-calico--apiserver--78f88fbddb--c68xl-" May 14 23:52:55.543746 containerd[1510]: 2025-05-14 23:52:55.438 [INFO][4664] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="4f4bc667d65a0fb43d566a520efd1f1d90ec74798a55fc1a55054a3ea5d2cde2" Namespace="calico-apiserver" Pod="calico-apiserver-78f88fbddb-c68xl" WorkloadEndpoint="localhost-k8s-calico--apiserver--78f88fbddb--c68xl-eth0" May 14 23:52:55.543746 containerd[1510]: 2025-05-14 23:52:55.474 [INFO][4678] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="4f4bc667d65a0fb43d566a520efd1f1d90ec74798a55fc1a55054a3ea5d2cde2" HandleID="k8s-pod-network.4f4bc667d65a0fb43d566a520efd1f1d90ec74798a55fc1a55054a3ea5d2cde2" Workload="localhost-k8s-calico--apiserver--78f88fbddb--c68xl-eth0" May 14 23:52:55.543746 containerd[1510]: 2025-05-14 23:52:55.483 [INFO][4678] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="4f4bc667d65a0fb43d566a520efd1f1d90ec74798a55fc1a55054a3ea5d2cde2" HandleID="k8s-pod-network.4f4bc667d65a0fb43d566a520efd1f1d90ec74798a55fc1a55054a3ea5d2cde2" Workload="localhost-k8s-calico--apiserver--78f88fbddb--c68xl-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002e97a0), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-78f88fbddb-c68xl", "timestamp":"2025-05-14 23:52:55.474093923 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 14 23:52:55.543746 containerd[1510]: 2025-05-14 23:52:55.483 [INFO][4678] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 14 23:52:55.543746 containerd[1510]: 2025-05-14 23:52:55.483 [INFO][4678] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 14 23:52:55.543746 containerd[1510]: 2025-05-14 23:52:55.483 [INFO][4678] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' May 14 23:52:55.543746 containerd[1510]: 2025-05-14 23:52:55.485 [INFO][4678] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.4f4bc667d65a0fb43d566a520efd1f1d90ec74798a55fc1a55054a3ea5d2cde2" host="localhost" May 14 23:52:55.543746 containerd[1510]: 2025-05-14 23:52:55.489 [INFO][4678] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" May 14 23:52:55.543746 containerd[1510]: 2025-05-14 23:52:55.493 [INFO][4678] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" May 14 23:52:55.543746 containerd[1510]: 2025-05-14 23:52:55.494 [INFO][4678] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" May 14 23:52:55.543746 containerd[1510]: 2025-05-14 23:52:55.496 [INFO][4678] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" May 14 23:52:55.543746 containerd[1510]: 2025-05-14 23:52:55.496 [INFO][4678] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.4f4bc667d65a0fb43d566a520efd1f1d90ec74798a55fc1a55054a3ea5d2cde2" host="localhost" May 14 23:52:55.543746 containerd[1510]: 2025-05-14 23:52:55.498 [INFO][4678] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.4f4bc667d65a0fb43d566a520efd1f1d90ec74798a55fc1a55054a3ea5d2cde2 May 14 23:52:55.543746 containerd[1510]: 2025-05-14 23:52:55.501 [INFO][4678] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.4f4bc667d65a0fb43d566a520efd1f1d90ec74798a55fc1a55054a3ea5d2cde2" host="localhost" May 14 23:52:55.543746 containerd[1510]: 2025-05-14 23:52:55.510 [INFO][4678] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.133/26] block=192.168.88.128/26 handle="k8s-pod-network.4f4bc667d65a0fb43d566a520efd1f1d90ec74798a55fc1a55054a3ea5d2cde2" host="localhost" May 14 23:52:55.543746 containerd[1510]: 2025-05-14 23:52:55.510 [INFO][4678] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.133/26] handle="k8s-pod-network.4f4bc667d65a0fb43d566a520efd1f1d90ec74798a55fc1a55054a3ea5d2cde2" host="localhost" May 14 23:52:55.543746 containerd[1510]: 2025-05-14 23:52:55.510 [INFO][4678] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 14 23:52:55.543746 containerd[1510]: 2025-05-14 23:52:55.510 [INFO][4678] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.133/26] IPv6=[] ContainerID="4f4bc667d65a0fb43d566a520efd1f1d90ec74798a55fc1a55054a3ea5d2cde2" HandleID="k8s-pod-network.4f4bc667d65a0fb43d566a520efd1f1d90ec74798a55fc1a55054a3ea5d2cde2" Workload="localhost-k8s-calico--apiserver--78f88fbddb--c68xl-eth0" May 14 23:52:55.544992 containerd[1510]: 2025-05-14 23:52:55.514 [INFO][4664] cni-plugin/k8s.go 386: Populated endpoint ContainerID="4f4bc667d65a0fb43d566a520efd1f1d90ec74798a55fc1a55054a3ea5d2cde2" Namespace="calico-apiserver" Pod="calico-apiserver-78f88fbddb-c68xl" WorkloadEndpoint="localhost-k8s-calico--apiserver--78f88fbddb--c68xl-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--78f88fbddb--c68xl-eth0", GenerateName:"calico-apiserver-78f88fbddb-", Namespace:"calico-apiserver", SelfLink:"", UID:"e558cb7f-428e-4fae-a45c-cc274d3aa581", ResourceVersion:"769", Generation:0, CreationTimestamp:time.Date(2025, time.May, 14, 23, 52, 22, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"78f88fbddb", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-78f88fbddb-c68xl", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali04b0ed547f4", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 14 23:52:55.544992 containerd[1510]: 2025-05-14 23:52:55.515 [INFO][4664] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.133/32] ContainerID="4f4bc667d65a0fb43d566a520efd1f1d90ec74798a55fc1a55054a3ea5d2cde2" Namespace="calico-apiserver" Pod="calico-apiserver-78f88fbddb-c68xl" WorkloadEndpoint="localhost-k8s-calico--apiserver--78f88fbddb--c68xl-eth0" May 14 23:52:55.544992 containerd[1510]: 2025-05-14 23:52:55.515 [INFO][4664] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali04b0ed547f4 ContainerID="4f4bc667d65a0fb43d566a520efd1f1d90ec74798a55fc1a55054a3ea5d2cde2" Namespace="calico-apiserver" Pod="calico-apiserver-78f88fbddb-c68xl" WorkloadEndpoint="localhost-k8s-calico--apiserver--78f88fbddb--c68xl-eth0" May 14 23:52:55.544992 containerd[1510]: 2025-05-14 23:52:55.517 [INFO][4664] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="4f4bc667d65a0fb43d566a520efd1f1d90ec74798a55fc1a55054a3ea5d2cde2" Namespace="calico-apiserver" Pod="calico-apiserver-78f88fbddb-c68xl" WorkloadEndpoint="localhost-k8s-calico--apiserver--78f88fbddb--c68xl-eth0" May 14 23:52:55.544992 containerd[1510]: 2025-05-14 23:52:55.518 [INFO][4664] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="4f4bc667d65a0fb43d566a520efd1f1d90ec74798a55fc1a55054a3ea5d2cde2" Namespace="calico-apiserver" Pod="calico-apiserver-78f88fbddb-c68xl" WorkloadEndpoint="localhost-k8s-calico--apiserver--78f88fbddb--c68xl-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--78f88fbddb--c68xl-eth0", GenerateName:"calico-apiserver-78f88fbddb-", Namespace:"calico-apiserver", SelfLink:"", UID:"e558cb7f-428e-4fae-a45c-cc274d3aa581", ResourceVersion:"769", Generation:0, CreationTimestamp:time.Date(2025, time.May, 14, 23, 52, 22, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"78f88fbddb", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"4f4bc667d65a0fb43d566a520efd1f1d90ec74798a55fc1a55054a3ea5d2cde2", Pod:"calico-apiserver-78f88fbddb-c68xl", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali04b0ed547f4", MAC:"92:96:c1:45:cd:c0", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 14 23:52:55.544992 containerd[1510]: 2025-05-14 23:52:55.537 [INFO][4664] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="4f4bc667d65a0fb43d566a520efd1f1d90ec74798a55fc1a55054a3ea5d2cde2" Namespace="calico-apiserver" Pod="calico-apiserver-78f88fbddb-c68xl" WorkloadEndpoint="localhost-k8s-calico--apiserver--78f88fbddb--c68xl-eth0" May 14 23:52:55.585025 kubelet[2749]: I0514 23:52:55.584919 2749 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7db6d8ff4d-xq4zr" podStartSLOduration=41.584903607 podStartE2EDuration="41.584903607s" podCreationTimestamp="2025-05-14 23:52:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-14 23:52:55.58404177 +0000 UTC m=+55.303272541" watchObservedRunningTime="2025-05-14 23:52:55.584903607 +0000 UTC m=+55.304134377" May 14 23:52:55.601870 containerd[1510]: time="2025-05-14T23:52:55.601806156Z" level=info msg="connecting to shim 4f4bc667d65a0fb43d566a520efd1f1d90ec74798a55fc1a55054a3ea5d2cde2" address="unix:///run/containerd/s/2e00f3d19d05739c8408c201c0e14a81c949516030e2c89117215a51d34ad1a2" namespace=k8s.io protocol=ttrpc version=3 May 14 23:52:55.647518 systemd[1]: Started cri-containerd-4f4bc667d65a0fb43d566a520efd1f1d90ec74798a55fc1a55054a3ea5d2cde2.scope - libcontainer container 4f4bc667d65a0fb43d566a520efd1f1d90ec74798a55fc1a55054a3ea5d2cde2. May 14 23:52:55.663568 systemd-resolved[1346]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address May 14 23:52:55.666468 systemd-networkd[1437]: calic1fd6edbd56: Gained IPv6LL May 14 23:52:55.702324 containerd[1510]: time="2025-05-14T23:52:55.702260668Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-78f88fbddb-c68xl,Uid:e558cb7f-428e-4fae-a45c-cc274d3aa581,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"4f4bc667d65a0fb43d566a520efd1f1d90ec74798a55fc1a55054a3ea5d2cde2\"" May 14 23:52:55.706674 containerd[1510]: time="2025-05-14T23:52:55.706646248Z" level=info msg="CreateContainer within sandbox \"4f4bc667d65a0fb43d566a520efd1f1d90ec74798a55fc1a55054a3ea5d2cde2\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" May 14 23:52:55.717846 containerd[1510]: time="2025-05-14T23:52:55.717797243Z" level=info msg="Container 3f1f4481ab449f397010810df8ed6afaf4f5061b6bbe2a85acf0d230d89fc3c3: CDI devices from CRI Config.CDIDevices: []" May 14 23:52:55.726659 containerd[1510]: time="2025-05-14T23:52:55.726614270Z" level=info msg="CreateContainer within sandbox \"4f4bc667d65a0fb43d566a520efd1f1d90ec74798a55fc1a55054a3ea5d2cde2\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"3f1f4481ab449f397010810df8ed6afaf4f5061b6bbe2a85acf0d230d89fc3c3\"" May 14 23:52:55.728572 containerd[1510]: time="2025-05-14T23:52:55.728494909Z" level=info msg="StartContainer for \"3f1f4481ab449f397010810df8ed6afaf4f5061b6bbe2a85acf0d230d89fc3c3\"" May 14 23:52:55.732589 containerd[1510]: time="2025-05-14T23:52:55.732535681Z" level=info msg="connecting to shim 3f1f4481ab449f397010810df8ed6afaf4f5061b6bbe2a85acf0d230d89fc3c3" address="unix:///run/containerd/s/2e00f3d19d05739c8408c201c0e14a81c949516030e2c89117215a51d34ad1a2" protocol=ttrpc version=3 May 14 23:52:55.758460 systemd[1]: Started cri-containerd-3f1f4481ab449f397010810df8ed6afaf4f5061b6bbe2a85acf0d230d89fc3c3.scope - libcontainer container 3f1f4481ab449f397010810df8ed6afaf4f5061b6bbe2a85acf0d230d89fc3c3. May 14 23:52:55.889240 containerd[1510]: time="2025-05-14T23:52:55.889174919Z" level=info msg="StartContainer for \"3f1f4481ab449f397010810df8ed6afaf4f5061b6bbe2a85acf0d230d89fc3c3\" returns successfully" May 14 23:52:56.383564 containerd[1510]: time="2025-05-14T23:52:56.383503929Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-96wg8,Uid:8f0b6425-d6bf-493f-ada4-58a2a3096286,Namespace:calico-system,Attempt:0,}" May 14 23:52:56.572358 kubelet[2749]: I0514 23:52:56.572322 2749 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 14 23:52:56.606290 kubelet[2749]: I0514 23:52:56.606216 2749 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-78f88fbddb-wtw8n" podStartSLOduration=30.945776262 podStartE2EDuration="34.606193593s" podCreationTimestamp="2025-05-14 23:52:22 +0000 UTC" firstStartedPulling="2025-05-14 23:52:51.459762712 +0000 UTC m=+51.178993492" lastFinishedPulling="2025-05-14 23:52:55.120180043 +0000 UTC m=+54.839410823" observedRunningTime="2025-05-14 23:52:55.634833726 +0000 UTC m=+55.354064496" watchObservedRunningTime="2025-05-14 23:52:56.606193593 +0000 UTC m=+56.325424363" May 14 23:52:56.607097 kubelet[2749]: I0514 23:52:56.606347 2749 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-78f88fbddb-c68xl" podStartSLOduration=34.60634149 podStartE2EDuration="34.60634149s" podCreationTimestamp="2025-05-14 23:52:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-14 23:52:56.60578716 +0000 UTC m=+56.325017930" watchObservedRunningTime="2025-05-14 23:52:56.60634149 +0000 UTC m=+56.325572260" May 14 23:52:56.622749 systemd-networkd[1437]: cali8730933c9ff: Link UP May 14 23:52:56.624280 systemd-networkd[1437]: cali8730933c9ff: Gained carrier May 14 23:52:56.645656 containerd[1510]: 2025-05-14 23:52:56.447 [INFO][4783] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-csi--node--driver--96wg8-eth0 csi-node-driver- calico-system 8f0b6425-d6bf-493f-ada4-58a2a3096286 602 0 2025-05-14 23:52:22 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:55b7b4b9d k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s localhost csi-node-driver-96wg8 eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali8730933c9ff [] []}} ContainerID="ead6a2b5ece0504762e0be90f168c1cc06970ec01464f436e26a49c3daa21955" Namespace="calico-system" Pod="csi-node-driver-96wg8" WorkloadEndpoint="localhost-k8s-csi--node--driver--96wg8-" May 14 23:52:56.645656 containerd[1510]: 2025-05-14 23:52:56.447 [INFO][4783] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="ead6a2b5ece0504762e0be90f168c1cc06970ec01464f436e26a49c3daa21955" Namespace="calico-system" Pod="csi-node-driver-96wg8" WorkloadEndpoint="localhost-k8s-csi--node--driver--96wg8-eth0" May 14 23:52:56.645656 containerd[1510]: 2025-05-14 23:52:56.488 [INFO][4797] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="ead6a2b5ece0504762e0be90f168c1cc06970ec01464f436e26a49c3daa21955" HandleID="k8s-pod-network.ead6a2b5ece0504762e0be90f168c1cc06970ec01464f436e26a49c3daa21955" Workload="localhost-k8s-csi--node--driver--96wg8-eth0" May 14 23:52:56.645656 containerd[1510]: 2025-05-14 23:52:56.498 [INFO][4797] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="ead6a2b5ece0504762e0be90f168c1cc06970ec01464f436e26a49c3daa21955" HandleID="k8s-pod-network.ead6a2b5ece0504762e0be90f168c1cc06970ec01464f436e26a49c3daa21955" Workload="localhost-k8s-csi--node--driver--96wg8-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000365370), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"csi-node-driver-96wg8", "timestamp":"2025-05-14 23:52:56.488694221 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 14 23:52:56.645656 containerd[1510]: 2025-05-14 23:52:56.498 [INFO][4797] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 14 23:52:56.645656 containerd[1510]: 2025-05-14 23:52:56.498 [INFO][4797] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 14 23:52:56.645656 containerd[1510]: 2025-05-14 23:52:56.498 [INFO][4797] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' May 14 23:52:56.645656 containerd[1510]: 2025-05-14 23:52:56.499 [INFO][4797] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.ead6a2b5ece0504762e0be90f168c1cc06970ec01464f436e26a49c3daa21955" host="localhost" May 14 23:52:56.645656 containerd[1510]: 2025-05-14 23:52:56.516 [INFO][4797] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" May 14 23:52:56.645656 containerd[1510]: 2025-05-14 23:52:56.537 [INFO][4797] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" May 14 23:52:56.645656 containerd[1510]: 2025-05-14 23:52:56.538 [INFO][4797] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" May 14 23:52:56.645656 containerd[1510]: 2025-05-14 23:52:56.592 [INFO][4797] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" May 14 23:52:56.645656 containerd[1510]: 2025-05-14 23:52:56.593 [INFO][4797] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.ead6a2b5ece0504762e0be90f168c1cc06970ec01464f436e26a49c3daa21955" host="localhost" May 14 23:52:56.645656 containerd[1510]: 2025-05-14 23:52:56.600 [INFO][4797] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.ead6a2b5ece0504762e0be90f168c1cc06970ec01464f436e26a49c3daa21955 May 14 23:52:56.645656 containerd[1510]: 2025-05-14 23:52:56.608 [INFO][4797] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.ead6a2b5ece0504762e0be90f168c1cc06970ec01464f436e26a49c3daa21955" host="localhost" May 14 23:52:56.645656 containerd[1510]: 2025-05-14 23:52:56.615 [INFO][4797] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.134/26] block=192.168.88.128/26 handle="k8s-pod-network.ead6a2b5ece0504762e0be90f168c1cc06970ec01464f436e26a49c3daa21955" host="localhost" May 14 23:52:56.645656 containerd[1510]: 2025-05-14 23:52:56.615 [INFO][4797] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.134/26] handle="k8s-pod-network.ead6a2b5ece0504762e0be90f168c1cc06970ec01464f436e26a49c3daa21955" host="localhost" May 14 23:52:56.645656 containerd[1510]: 2025-05-14 23:52:56.615 [INFO][4797] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 14 23:52:56.645656 containerd[1510]: 2025-05-14 23:52:56.615 [INFO][4797] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.134/26] IPv6=[] ContainerID="ead6a2b5ece0504762e0be90f168c1cc06970ec01464f436e26a49c3daa21955" HandleID="k8s-pod-network.ead6a2b5ece0504762e0be90f168c1cc06970ec01464f436e26a49c3daa21955" Workload="localhost-k8s-csi--node--driver--96wg8-eth0" May 14 23:52:56.646742 containerd[1510]: 2025-05-14 23:52:56.619 [INFO][4783] cni-plugin/k8s.go 386: Populated endpoint ContainerID="ead6a2b5ece0504762e0be90f168c1cc06970ec01464f436e26a49c3daa21955" Namespace="calico-system" Pod="csi-node-driver-96wg8" WorkloadEndpoint="localhost-k8s-csi--node--driver--96wg8-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--96wg8-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"8f0b6425-d6bf-493f-ada4-58a2a3096286", ResourceVersion:"602", Generation:0, CreationTimestamp:time.Date(2025, time.May, 14, 23, 52, 22, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"55b7b4b9d", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"csi-node-driver-96wg8", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali8730933c9ff", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 14 23:52:56.646742 containerd[1510]: 2025-05-14 23:52:56.619 [INFO][4783] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.134/32] ContainerID="ead6a2b5ece0504762e0be90f168c1cc06970ec01464f436e26a49c3daa21955" Namespace="calico-system" Pod="csi-node-driver-96wg8" WorkloadEndpoint="localhost-k8s-csi--node--driver--96wg8-eth0" May 14 23:52:56.646742 containerd[1510]: 2025-05-14 23:52:56.619 [INFO][4783] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali8730933c9ff ContainerID="ead6a2b5ece0504762e0be90f168c1cc06970ec01464f436e26a49c3daa21955" Namespace="calico-system" Pod="csi-node-driver-96wg8" WorkloadEndpoint="localhost-k8s-csi--node--driver--96wg8-eth0" May 14 23:52:56.646742 containerd[1510]: 2025-05-14 23:52:56.623 [INFO][4783] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="ead6a2b5ece0504762e0be90f168c1cc06970ec01464f436e26a49c3daa21955" Namespace="calico-system" Pod="csi-node-driver-96wg8" WorkloadEndpoint="localhost-k8s-csi--node--driver--96wg8-eth0" May 14 23:52:56.646742 containerd[1510]: 2025-05-14 23:52:56.623 [INFO][4783] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="ead6a2b5ece0504762e0be90f168c1cc06970ec01464f436e26a49c3daa21955" Namespace="calico-system" Pod="csi-node-driver-96wg8" WorkloadEndpoint="localhost-k8s-csi--node--driver--96wg8-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--96wg8-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"8f0b6425-d6bf-493f-ada4-58a2a3096286", ResourceVersion:"602", Generation:0, CreationTimestamp:time.Date(2025, time.May, 14, 23, 52, 22, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"55b7b4b9d", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"ead6a2b5ece0504762e0be90f168c1cc06970ec01464f436e26a49c3daa21955", Pod:"csi-node-driver-96wg8", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali8730933c9ff", MAC:"d2:44:31:89:dd:29", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 14 23:52:56.646742 containerd[1510]: 2025-05-14 23:52:56.638 [INFO][4783] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="ead6a2b5ece0504762e0be90f168c1cc06970ec01464f436e26a49c3daa21955" Namespace="calico-system" Pod="csi-node-driver-96wg8" WorkloadEndpoint="localhost-k8s-csi--node--driver--96wg8-eth0" May 14 23:52:56.687357 containerd[1510]: time="2025-05-14T23:52:56.687289302Z" level=info msg="connecting to shim ead6a2b5ece0504762e0be90f168c1cc06970ec01464f436e26a49c3daa21955" address="unix:///run/containerd/s/646a44ee9a17cc62a3c967d820bb12a38663d4a6e9f71aeaba72b926a3725168" namespace=k8s.io protocol=ttrpc version=3 May 14 23:52:56.729481 systemd[1]: Started cri-containerd-ead6a2b5ece0504762e0be90f168c1cc06970ec01464f436e26a49c3daa21955.scope - libcontainer container ead6a2b5ece0504762e0be90f168c1cc06970ec01464f436e26a49c3daa21955. May 14 23:52:56.746084 systemd-resolved[1346]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address May 14 23:52:56.765041 containerd[1510]: time="2025-05-14T23:52:56.765004186Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-96wg8,Uid:8f0b6425-d6bf-493f-ada4-58a2a3096286,Namespace:calico-system,Attempt:0,} returns sandbox id \"ead6a2b5ece0504762e0be90f168c1cc06970ec01464f436e26a49c3daa21955\"" May 14 23:52:57.074556 systemd-networkd[1437]: cali04b0ed547f4: Gained IPv6LL May 14 23:52:57.174129 containerd[1510]: time="2025-05-14T23:52:57.174051206Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:52:57.175268 containerd[1510]: time="2025-05-14T23:52:57.175145700Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.29.3: active requests=0, bytes read=34789138" May 14 23:52:57.176801 containerd[1510]: time="2025-05-14T23:52:57.176752174Z" level=info msg="ImageCreate event name:\"sha256:4e982138231b3653a012db4f21ed5e7be69afd5f553dba38cf7e88f0ed740b94\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:52:57.179244 containerd[1510]: time="2025-05-14T23:52:57.179180992Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:feaab0197035d474845e0f8137a99a78cab274f0a3cac4d5485cf9b1bdf9ffa9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:52:57.180421 containerd[1510]: time="2025-05-14T23:52:57.180340528Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\" with image id \"sha256:4e982138231b3653a012db4f21ed5e7be69afd5f553dba38cf7e88f0ed740b94\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:feaab0197035d474845e0f8137a99a78cab274f0a3cac4d5485cf9b1bdf9ffa9\", size \"36281728\" in 2.059244777s" May 14 23:52:57.180421 containerd[1510]: time="2025-05-14T23:52:57.180382828Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\" returns image reference \"sha256:4e982138231b3653a012db4f21ed5e7be69afd5f553dba38cf7e88f0ed740b94\"" May 14 23:52:57.181934 containerd[1510]: time="2025-05-14T23:52:57.181793244Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.3\"" May 14 23:52:57.200373 containerd[1510]: time="2025-05-14T23:52:57.199096133Z" level=info msg="CreateContainer within sandbox \"9dcc7e7dcb03816f78c3144c5e0d390722433131d7b41383d2c4187f062647f0\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" May 14 23:52:57.204884 systemd[1]: Started sshd@14-10.0.0.142:22-10.0.0.1:40972.service - OpenSSH per-connection server daemon (10.0.0.1:40972). May 14 23:52:57.212539 containerd[1510]: time="2025-05-14T23:52:57.212484847Z" level=info msg="Container b07927927be2c36d35386b3094e3735dc4eb30a740e3157b0bf20c820bde0512: CDI devices from CRI Config.CDIDevices: []" May 14 23:52:57.223579 containerd[1510]: time="2025-05-14T23:52:57.223175557Z" level=info msg="CreateContainer within sandbox \"9dcc7e7dcb03816f78c3144c5e0d390722433131d7b41383d2c4187f062647f0\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"b07927927be2c36d35386b3094e3735dc4eb30a740e3157b0bf20c820bde0512\"" May 14 23:52:57.224626 containerd[1510]: time="2025-05-14T23:52:57.224576446Z" level=info msg="StartContainer for \"b07927927be2c36d35386b3094e3735dc4eb30a740e3157b0bf20c820bde0512\"" May 14 23:52:57.225891 containerd[1510]: time="2025-05-14T23:52:57.225857510Z" level=info msg="connecting to shim b07927927be2c36d35386b3094e3735dc4eb30a740e3157b0bf20c820bde0512" address="unix:///run/containerd/s/961f062b58e2de23d9ad00676f8cd38ca1edbb528df4302ec990d83b42bf6b2e" protocol=ttrpc version=3 May 14 23:52:57.257456 systemd[1]: Started cri-containerd-b07927927be2c36d35386b3094e3735dc4eb30a740e3157b0bf20c820bde0512.scope - libcontainer container b07927927be2c36d35386b3094e3735dc4eb30a740e3157b0bf20c820bde0512. May 14 23:52:57.270271 sshd[4873]: Accepted publickey for core from 10.0.0.1 port 40972 ssh2: RSA SHA256:lk2TkYBEL43KPVbrGyh3Ro8IB8NGN6uTNXzFyrYR01I May 14 23:52:57.272336 sshd-session[4873]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 23:52:57.279135 systemd-logind[1487]: New session 15 of user core. May 14 23:52:57.286437 systemd[1]: Started session-15.scope - Session 15 of User core. May 14 23:52:57.408362 containerd[1510]: time="2025-05-14T23:52:57.406849718Z" level=info msg="StartContainer for \"b07927927be2c36d35386b3094e3735dc4eb30a740e3157b0bf20c820bde0512\" returns successfully" May 14 23:52:57.513948 sshd[4894]: Connection closed by 10.0.0.1 port 40972 May 14 23:52:57.514399 sshd-session[4873]: pam_unix(sshd:session): session closed for user core May 14 23:52:57.519991 systemd[1]: sshd@14-10.0.0.142:22-10.0.0.1:40972.service: Deactivated successfully. May 14 23:52:57.523805 systemd[1]: session-15.scope: Deactivated successfully. May 14 23:52:57.526085 systemd-logind[1487]: Session 15 logged out. Waiting for processes to exit. May 14 23:52:57.527602 systemd-logind[1487]: Removed session 15. May 14 23:52:57.577718 kubelet[2749]: I0514 23:52:57.577674 2749 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 14 23:52:57.612568 kubelet[2749]: I0514 23:52:57.612493 2749 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-57dfd6df95-fhrn2" podStartSLOduration=32.192671881 podStartE2EDuration="35.61246961s" podCreationTimestamp="2025-05-14 23:52:22 +0000 UTC" firstStartedPulling="2025-05-14 23:52:53.76147551 +0000 UTC m=+53.480706280" lastFinishedPulling="2025-05-14 23:52:57.181273239 +0000 UTC m=+56.900504009" observedRunningTime="2025-05-14 23:52:57.612450194 +0000 UTC m=+57.331680964" watchObservedRunningTime="2025-05-14 23:52:57.61246961 +0000 UTC m=+57.331700380" May 14 23:52:58.290637 systemd-networkd[1437]: cali8730933c9ff: Gained IPv6LL May 14 23:52:58.664167 containerd[1510]: time="2025-05-14T23:52:58.660446407Z" level=info msg="TaskExit event in podsandbox handler container_id:\"b07927927be2c36d35386b3094e3735dc4eb30a740e3157b0bf20c820bde0512\" id:\"7989bb0bd13b710905f83b4c62c11c06beb83b12231a794a189e5401316b4b0e\" pid:4935 exited_at:{seconds:1747266778 nanos:659382260}" May 14 23:52:59.341091 containerd[1510]: time="2025-05-14T23:52:59.341009422Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:52:59.360566 containerd[1510]: time="2025-05-14T23:52:59.360471770Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.29.3: active requests=0, bytes read=7912898" May 14 23:52:59.364516 containerd[1510]: time="2025-05-14T23:52:59.364441810Z" level=info msg="ImageCreate event name:\"sha256:4c37db5645f4075f8b8170eea8f14e340cb13550e0a392962f1f211ded741505\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:52:59.367522 containerd[1510]: time="2025-05-14T23:52:59.367467818Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:72455a36febc7c56ec8881007f4805caed5764026a0694e4f86a2503209b2d31\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:52:59.368008 containerd[1510]: time="2025-05-14T23:52:59.367972344Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.29.3\" with image id \"sha256:4c37db5645f4075f8b8170eea8f14e340cb13550e0a392962f1f211ded741505\", repo tag \"ghcr.io/flatcar/calico/csi:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:72455a36febc7c56ec8881007f4805caed5764026a0694e4f86a2503209b2d31\", size \"9405520\" in 2.18614151s" May 14 23:52:59.368008 containerd[1510]: time="2025-05-14T23:52:59.368005967Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.3\" returns image reference \"sha256:4c37db5645f4075f8b8170eea8f14e340cb13550e0a392962f1f211ded741505\"" May 14 23:52:59.370093 containerd[1510]: time="2025-05-14T23:52:59.369974160Z" level=info msg="CreateContainer within sandbox \"ead6a2b5ece0504762e0be90f168c1cc06970ec01464f436e26a49c3daa21955\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" May 14 23:52:59.395265 containerd[1510]: time="2025-05-14T23:52:59.395222896Z" level=info msg="Container fa303ef20cd421eb18214ae9bd1b6684633081a7f0d971c3dc4e91c5219188b7: CDI devices from CRI Config.CDIDevices: []" May 14 23:52:59.404623 containerd[1510]: time="2025-05-14T23:52:59.404582549Z" level=info msg="CreateContainer within sandbox \"ead6a2b5ece0504762e0be90f168c1cc06970ec01464f436e26a49c3daa21955\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"fa303ef20cd421eb18214ae9bd1b6684633081a7f0d971c3dc4e91c5219188b7\"" May 14 23:52:59.405177 containerd[1510]: time="2025-05-14T23:52:59.405148481Z" level=info msg="StartContainer for \"fa303ef20cd421eb18214ae9bd1b6684633081a7f0d971c3dc4e91c5219188b7\"" May 14 23:52:59.406436 containerd[1510]: time="2025-05-14T23:52:59.406398597Z" level=info msg="connecting to shim fa303ef20cd421eb18214ae9bd1b6684633081a7f0d971c3dc4e91c5219188b7" address="unix:///run/containerd/s/646a44ee9a17cc62a3c967d820bb12a38663d4a6e9f71aeaba72b926a3725168" protocol=ttrpc version=3 May 14 23:52:59.427531 systemd[1]: Started cri-containerd-fa303ef20cd421eb18214ae9bd1b6684633081a7f0d971c3dc4e91c5219188b7.scope - libcontainer container fa303ef20cd421eb18214ae9bd1b6684633081a7f0d971c3dc4e91c5219188b7. May 14 23:52:59.569108 containerd[1510]: time="2025-05-14T23:52:59.569060823Z" level=info msg="StartContainer for \"fa303ef20cd421eb18214ae9bd1b6684633081a7f0d971c3dc4e91c5219188b7\" returns successfully" May 14 23:52:59.570585 containerd[1510]: time="2025-05-14T23:52:59.570562331Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\"" May 14 23:53:00.813609 kubelet[2749]: I0514 23:53:00.813557 2749 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 14 23:53:00.933390 containerd[1510]: time="2025-05-14T23:53:00.933257215Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:53:00.934434 containerd[1510]: time="2025-05-14T23:53:00.934369068Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3: active requests=0, bytes read=13991773" May 14 23:53:00.935585 containerd[1510]: time="2025-05-14T23:53:00.935549388Z" level=info msg="ImageCreate event name:\"sha256:e909e2ccf54404290b577fbddd190d036984deed184001767f820b0dddf77fd9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:53:00.937771 containerd[1510]: time="2025-05-14T23:53:00.937739398Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:3f15090a9bb45773d1fd019455ec3d3f3746f3287c35d8013e497b38d8237324\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 23:53:00.938330 containerd[1510]: time="2025-05-14T23:53:00.938283903Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\" with image id \"sha256:e909e2ccf54404290b577fbddd190d036984deed184001767f820b0dddf77fd9\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:3f15090a9bb45773d1fd019455ec3d3f3746f3287c35d8013e497b38d8237324\", size \"15484347\" in 1.367693719s" May 14 23:53:00.938406 containerd[1510]: time="2025-05-14T23:53:00.938333426Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\" returns image reference \"sha256:e909e2ccf54404290b577fbddd190d036984deed184001767f820b0dddf77fd9\"" May 14 23:53:00.940960 containerd[1510]: time="2025-05-14T23:53:00.940894093Z" level=info msg="CreateContainer within sandbox \"ead6a2b5ece0504762e0be90f168c1cc06970ec01464f436e26a49c3daa21955\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" May 14 23:53:00.950234 containerd[1510]: time="2025-05-14T23:53:00.949268187Z" level=info msg="Container c0509e685cd517154ef8d0e2555334178b271a104571c5625762a67ae6324880: CDI devices from CRI Config.CDIDevices: []" May 14 23:53:00.961613 containerd[1510]: time="2025-05-14T23:53:00.961553439Z" level=info msg="CreateContainer within sandbox \"ead6a2b5ece0504762e0be90f168c1cc06970ec01464f436e26a49c3daa21955\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"c0509e685cd517154ef8d0e2555334178b271a104571c5625762a67ae6324880\"" May 14 23:53:00.962153 containerd[1510]: time="2025-05-14T23:53:00.962128100Z" level=info msg="StartContainer for \"c0509e685cd517154ef8d0e2555334178b271a104571c5625762a67ae6324880\"" May 14 23:53:00.963918 containerd[1510]: time="2025-05-14T23:53:00.963893211Z" level=info msg="connecting to shim c0509e685cd517154ef8d0e2555334178b271a104571c5625762a67ae6324880" address="unix:///run/containerd/s/646a44ee9a17cc62a3c967d820bb12a38663d4a6e9f71aeaba72b926a3725168" protocol=ttrpc version=3 May 14 23:53:00.999683 systemd[1]: Started cri-containerd-c0509e685cd517154ef8d0e2555334178b271a104571c5625762a67ae6324880.scope - libcontainer container c0509e685cd517154ef8d0e2555334178b271a104571c5625762a67ae6324880. May 14 23:53:01.053186 containerd[1510]: time="2025-05-14T23:53:01.053011688Z" level=info msg="StartContainer for \"c0509e685cd517154ef8d0e2555334178b271a104571c5625762a67ae6324880\" returns successfully" May 14 23:53:01.485014 kubelet[2749]: I0514 23:53:01.484952 2749 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 May 14 23:53:01.485014 kubelet[2749]: I0514 23:53:01.485001 2749 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock May 14 23:53:01.630491 kubelet[2749]: I0514 23:53:01.630414 2749 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-96wg8" podStartSLOduration=35.457566513 podStartE2EDuration="39.630393706s" podCreationTimestamp="2025-05-14 23:52:22 +0000 UTC" firstStartedPulling="2025-05-14 23:52:56.766433357 +0000 UTC m=+56.485664127" lastFinishedPulling="2025-05-14 23:53:00.93926055 +0000 UTC m=+60.658491320" observedRunningTime="2025-05-14 23:53:01.630221766 +0000 UTC m=+61.349452536" watchObservedRunningTime="2025-05-14 23:53:01.630393706 +0000 UTC m=+61.349624486" May 14 23:53:02.528086 systemd[1]: Started sshd@15-10.0.0.142:22-10.0.0.1:59222.service - OpenSSH per-connection server daemon (10.0.0.1:59222). May 14 23:53:02.600247 sshd[5027]: Accepted publickey for core from 10.0.0.1 port 59222 ssh2: RSA SHA256:lk2TkYBEL43KPVbrGyh3Ro8IB8NGN6uTNXzFyrYR01I May 14 23:53:02.602164 sshd-session[5027]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 23:53:02.607088 systemd-logind[1487]: New session 16 of user core. May 14 23:53:02.621477 systemd[1]: Started session-16.scope - Session 16 of User core. May 14 23:53:02.750262 sshd[5029]: Connection closed by 10.0.0.1 port 59222 May 14 23:53:02.750727 sshd-session[5027]: pam_unix(sshd:session): session closed for user core May 14 23:53:02.759895 systemd[1]: sshd@15-10.0.0.142:22-10.0.0.1:59222.service: Deactivated successfully. May 14 23:53:02.762110 systemd[1]: session-16.scope: Deactivated successfully. May 14 23:53:02.764124 systemd-logind[1487]: Session 16 logged out. Waiting for processes to exit. May 14 23:53:02.766151 systemd[1]: Started sshd@16-10.0.0.142:22-10.0.0.1:59234.service - OpenSSH per-connection server daemon (10.0.0.1:59234). May 14 23:53:02.767421 systemd-logind[1487]: Removed session 16. May 14 23:53:02.812275 sshd[5043]: Accepted publickey for core from 10.0.0.1 port 59234 ssh2: RSA SHA256:lk2TkYBEL43KPVbrGyh3Ro8IB8NGN6uTNXzFyrYR01I May 14 23:53:02.814089 sshd-session[5043]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 23:53:02.819534 systemd-logind[1487]: New session 17 of user core. May 14 23:53:02.827537 systemd[1]: Started session-17.scope - Session 17 of User core. May 14 23:53:03.790568 sshd[5046]: Connection closed by 10.0.0.1 port 59234 May 14 23:53:03.790958 sshd-session[5043]: pam_unix(sshd:session): session closed for user core May 14 23:53:03.801565 systemd[1]: sshd@16-10.0.0.142:22-10.0.0.1:59234.service: Deactivated successfully. May 14 23:53:03.803894 systemd[1]: session-17.scope: Deactivated successfully. May 14 23:53:03.805716 systemd-logind[1487]: Session 17 logged out. Waiting for processes to exit. May 14 23:53:03.807686 systemd[1]: Started sshd@17-10.0.0.142:22-10.0.0.1:59244.service - OpenSSH per-connection server daemon (10.0.0.1:59244). May 14 23:53:03.808693 systemd-logind[1487]: Removed session 17. May 14 23:53:03.866324 sshd[5056]: Accepted publickey for core from 10.0.0.1 port 59244 ssh2: RSA SHA256:lk2TkYBEL43KPVbrGyh3Ro8IB8NGN6uTNXzFyrYR01I May 14 23:53:03.868169 sshd-session[5056]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 23:53:03.873852 systemd-logind[1487]: New session 18 of user core. May 14 23:53:03.884492 systemd[1]: Started session-18.scope - Session 18 of User core. May 14 23:53:05.463076 sshd[5059]: Connection closed by 10.0.0.1 port 59244 May 14 23:53:05.463337 sshd-session[5056]: pam_unix(sshd:session): session closed for user core May 14 23:53:05.476502 systemd[1]: sshd@17-10.0.0.142:22-10.0.0.1:59244.service: Deactivated successfully. May 14 23:53:05.478839 systemd[1]: session-18.scope: Deactivated successfully. May 14 23:53:05.479827 systemd-logind[1487]: Session 18 logged out. Waiting for processes to exit. May 14 23:53:05.482100 systemd[1]: Started sshd@18-10.0.0.142:22-10.0.0.1:59260.service - OpenSSH per-connection server daemon (10.0.0.1:59260). May 14 23:53:05.483666 systemd-logind[1487]: Removed session 18. May 14 23:53:05.537994 sshd[5077]: Accepted publickey for core from 10.0.0.1 port 59260 ssh2: RSA SHA256:lk2TkYBEL43KPVbrGyh3Ro8IB8NGN6uTNXzFyrYR01I May 14 23:53:05.539508 sshd-session[5077]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 23:53:05.544493 systemd-logind[1487]: New session 19 of user core. May 14 23:53:05.554479 systemd[1]: Started session-19.scope - Session 19 of User core. May 14 23:53:05.763388 sshd[5080]: Connection closed by 10.0.0.1 port 59260 May 14 23:53:05.765142 sshd-session[5077]: pam_unix(sshd:session): session closed for user core May 14 23:53:05.775322 systemd[1]: sshd@18-10.0.0.142:22-10.0.0.1:59260.service: Deactivated successfully. May 14 23:53:05.778186 systemd[1]: session-19.scope: Deactivated successfully. May 14 23:53:05.779269 systemd-logind[1487]: Session 19 logged out. Waiting for processes to exit. May 14 23:53:05.782825 systemd[1]: Started sshd@19-10.0.0.142:22-10.0.0.1:59268.service - OpenSSH per-connection server daemon (10.0.0.1:59268). May 14 23:53:05.783832 systemd-logind[1487]: Removed session 19. May 14 23:53:05.830525 sshd[5091]: Accepted publickey for core from 10.0.0.1 port 59268 ssh2: RSA SHA256:lk2TkYBEL43KPVbrGyh3Ro8IB8NGN6uTNXzFyrYR01I May 14 23:53:05.832474 sshd-session[5091]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 23:53:05.837436 systemd-logind[1487]: New session 20 of user core. May 14 23:53:05.851618 systemd[1]: Started session-20.scope - Session 20 of User core. May 14 23:53:05.981975 sshd[5094]: Connection closed by 10.0.0.1 port 59268 May 14 23:53:05.982481 sshd-session[5091]: pam_unix(sshd:session): session closed for user core May 14 23:53:05.986540 systemd[1]: sshd@19-10.0.0.142:22-10.0.0.1:59268.service: Deactivated successfully. May 14 23:53:05.988914 systemd[1]: session-20.scope: Deactivated successfully. May 14 23:53:05.989853 systemd-logind[1487]: Session 20 logged out. Waiting for processes to exit. May 14 23:53:05.991204 systemd-logind[1487]: Removed session 20. May 14 23:53:09.530355 containerd[1510]: time="2025-05-14T23:53:09.530252034Z" level=info msg="TaskExit event in podsandbox handler container_id:\"b07927927be2c36d35386b3094e3735dc4eb30a740e3157b0bf20c820bde0512\" id:\"3380ae6de80e4b5d6d2f74e1c9fb6bc051528874da0403c60ce8da5f96d46429\" pid:5127 exited_at:{seconds:1747266789 nanos:529980124}" May 14 23:53:11.000263 systemd[1]: Started sshd@20-10.0.0.142:22-10.0.0.1:45588.service - OpenSSH per-connection server daemon (10.0.0.1:45588). May 14 23:53:11.053769 sshd[5143]: Accepted publickey for core from 10.0.0.1 port 45588 ssh2: RSA SHA256:lk2TkYBEL43KPVbrGyh3Ro8IB8NGN6uTNXzFyrYR01I May 14 23:53:11.055829 sshd-session[5143]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 23:53:11.060888 systemd-logind[1487]: New session 21 of user core. May 14 23:53:11.072550 systemd[1]: Started session-21.scope - Session 21 of User core. May 14 23:53:11.186718 sshd[5145]: Connection closed by 10.0.0.1 port 45588 May 14 23:53:11.187329 sshd-session[5143]: pam_unix(sshd:session): session closed for user core May 14 23:53:11.192435 systemd[1]: sshd@20-10.0.0.142:22-10.0.0.1:45588.service: Deactivated successfully. May 14 23:53:11.194859 systemd[1]: session-21.scope: Deactivated successfully. May 14 23:53:11.195566 systemd-logind[1487]: Session 21 logged out. Waiting for processes to exit. May 14 23:53:11.196660 systemd-logind[1487]: Removed session 21. May 14 23:53:16.201047 systemd[1]: Started sshd@21-10.0.0.142:22-10.0.0.1:45592.service - OpenSSH per-connection server daemon (10.0.0.1:45592). May 14 23:53:16.264920 sshd[5160]: Accepted publickey for core from 10.0.0.1 port 45592 ssh2: RSA SHA256:lk2TkYBEL43KPVbrGyh3Ro8IB8NGN6uTNXzFyrYR01I May 14 23:53:16.266844 sshd-session[5160]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 23:53:16.271258 systemd-logind[1487]: New session 22 of user core. May 14 23:53:16.279433 systemd[1]: Started session-22.scope - Session 22 of User core. May 14 23:53:16.328760 kubelet[2749]: I0514 23:53:16.328711 2749 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 14 23:53:16.407452 sshd[5162]: Connection closed by 10.0.0.1 port 45592 May 14 23:53:16.407845 sshd-session[5160]: pam_unix(sshd:session): session closed for user core May 14 23:53:16.412283 systemd[1]: sshd@21-10.0.0.142:22-10.0.0.1:45592.service: Deactivated successfully. May 14 23:53:16.414905 systemd[1]: session-22.scope: Deactivated successfully. May 14 23:53:16.415758 systemd-logind[1487]: Session 22 logged out. Waiting for processes to exit. May 14 23:53:16.416938 systemd-logind[1487]: Removed session 22. May 14 23:53:18.880644 containerd[1510]: time="2025-05-14T23:53:18.880576701Z" level=info msg="TaskExit event in podsandbox handler container_id:\"6bebd036ec15cc329550bdaaaec133f7641f67426d0a97dbcf4742f65440836d\" id:\"554ebf51c2da6d9cc21e8d9ec1b356425b230eee878e68e99e3c82085190e967\" pid:5189 exited_at:{seconds:1747266798 nanos:880204722}" May 14 23:53:21.450598 systemd[1]: Started sshd@22-10.0.0.142:22-10.0.0.1:35996.service - OpenSSH per-connection server daemon (10.0.0.1:35996). May 14 23:53:21.528197 sshd[5205]: Accepted publickey for core from 10.0.0.1 port 35996 ssh2: RSA SHA256:lk2TkYBEL43KPVbrGyh3Ro8IB8NGN6uTNXzFyrYR01I May 14 23:53:21.530584 sshd-session[5205]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 23:53:21.541365 systemd-logind[1487]: New session 23 of user core. May 14 23:53:21.558759 systemd[1]: Started session-23.scope - Session 23 of User core. May 14 23:53:21.753466 sshd[5207]: Connection closed by 10.0.0.1 port 35996 May 14 23:53:21.754058 sshd-session[5205]: pam_unix(sshd:session): session closed for user core May 14 23:53:21.759973 systemd[1]: sshd@22-10.0.0.142:22-10.0.0.1:35996.service: Deactivated successfully. May 14 23:53:21.762813 systemd[1]: session-23.scope: Deactivated successfully. May 14 23:53:21.764369 systemd-logind[1487]: Session 23 logged out. Waiting for processes to exit. May 14 23:53:21.765568 systemd-logind[1487]: Removed session 23. May 14 23:53:23.187432 containerd[1510]: time="2025-05-14T23:53:23.187373070Z" level=info msg="TaskExit event in podsandbox handler container_id:\"b07927927be2c36d35386b3094e3735dc4eb30a740e3157b0bf20c820bde0512\" id:\"0dbeebb7ade4e2d8d7ec31a28dcf0562e4c923c0c254b9205eb1ec8de3b515b9\" pid:5232 exited_at:{seconds:1747266803 nanos:187083660}" May 14 23:53:26.785752 systemd[1]: Started sshd@23-10.0.0.142:22-10.0.0.1:36008.service - OpenSSH per-connection server daemon (10.0.0.1:36008). May 14 23:53:26.941989 sshd[5244]: Accepted publickey for core from 10.0.0.1 port 36008 ssh2: RSA SHA256:lk2TkYBEL43KPVbrGyh3Ro8IB8NGN6uTNXzFyrYR01I May 14 23:53:26.942728 sshd-session[5244]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 23:53:26.956008 systemd-logind[1487]: New session 24 of user core. May 14 23:53:26.971575 systemd[1]: Started session-24.scope - Session 24 of User core. May 14 23:53:27.152501 sshd[5246]: Connection closed by 10.0.0.1 port 36008 May 14 23:53:27.155966 sshd-session[5244]: pam_unix(sshd:session): session closed for user core May 14 23:53:27.166526 systemd[1]: sshd@23-10.0.0.142:22-10.0.0.1:36008.service: Deactivated successfully. May 14 23:53:27.175861 systemd[1]: session-24.scope: Deactivated successfully. May 14 23:53:27.178439 systemd-logind[1487]: Session 24 logged out. Waiting for processes to exit. May 14 23:53:27.180096 systemd-logind[1487]: Removed session 24.