May 14 05:02:23.843604 kernel: Linux version 6.12.20-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.2.1_p20241221 p7) 14.2.1 20241221, GNU ld (Gentoo 2.44 p1) 2.44.0) #1 SMP PREEMPT_DYNAMIC Wed May 14 03:42:56 -00 2025 May 14 05:02:23.843628 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=bd5d20a479abde3485dc2e7b97a54e804895b9926289ae86f84794bef32a40f3 May 14 05:02:23.843639 kernel: BIOS-provided physical RAM map: May 14 05:02:23.843646 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable May 14 05:02:23.843652 kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000007fffff] usable May 14 05:02:23.843659 kernel: BIOS-e820: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS May 14 05:02:23.843667 kernel: BIOS-e820: [mem 0x0000000000808000-0x000000000080afff] usable May 14 05:02:23.843673 kernel: BIOS-e820: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS May 14 05:02:23.843682 kernel: BIOS-e820: [mem 0x000000000080c000-0x0000000000810fff] usable May 14 05:02:23.843688 kernel: BIOS-e820: [mem 0x0000000000811000-0x00000000008fffff] ACPI NVS May 14 05:02:23.843694 kernel: BIOS-e820: [mem 0x0000000000900000-0x000000009bd3efff] usable May 14 05:02:23.843714 kernel: BIOS-e820: [mem 0x000000009bd3f000-0x000000009bdfffff] reserved May 14 05:02:23.843720 kernel: BIOS-e820: [mem 0x000000009be00000-0x000000009c8ecfff] usable May 14 05:02:23.843727 kernel: BIOS-e820: [mem 0x000000009c8ed000-0x000000009cb6cfff] reserved May 14 05:02:23.843737 kernel: BIOS-e820: [mem 0x000000009cb6d000-0x000000009cb7efff] ACPI data May 14 05:02:23.843745 kernel: BIOS-e820: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS May 14 05:02:23.843755 kernel: BIOS-e820: [mem 0x000000009cbff000-0x000000009ce90fff] usable May 14 05:02:23.843762 kernel: BIOS-e820: [mem 0x000000009ce91000-0x000000009ce94fff] reserved May 14 05:02:23.843775 kernel: BIOS-e820: [mem 0x000000009ce95000-0x000000009ce96fff] ACPI NVS May 14 05:02:23.843782 kernel: BIOS-e820: [mem 0x000000009ce97000-0x000000009cedbfff] usable May 14 05:02:23.843789 kernel: BIOS-e820: [mem 0x000000009cedc000-0x000000009cf5ffff] reserved May 14 05:02:23.843796 kernel: BIOS-e820: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS May 14 05:02:23.843803 kernel: BIOS-e820: [mem 0x00000000e0000000-0x00000000efffffff] reserved May 14 05:02:23.843810 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved May 14 05:02:23.843817 kernel: BIOS-e820: [mem 0x00000000ffc00000-0x00000000ffffffff] reserved May 14 05:02:23.843826 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved May 14 05:02:23.843833 kernel: NX (Execute Disable) protection: active May 14 05:02:23.843840 kernel: APIC: Static calls initialized May 14 05:02:23.843847 kernel: e820: update [mem 0x9b320018-0x9b329c57] usable ==> usable May 14 05:02:23.843854 kernel: e820: update [mem 0x9b2e3018-0x9b31fe57] usable ==> usable May 14 05:02:23.843861 kernel: extended physical RAM map: May 14 05:02:23.843868 kernel: reserve setup_data: [mem 0x0000000000000000-0x000000000009ffff] usable May 14 05:02:23.843875 kernel: reserve setup_data: [mem 0x0000000000100000-0x00000000007fffff] usable May 14 05:02:23.843882 kernel: reserve setup_data: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS May 14 05:02:23.843889 kernel: reserve setup_data: [mem 0x0000000000808000-0x000000000080afff] usable May 14 05:02:23.843896 kernel: reserve setup_data: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS May 14 05:02:23.843905 kernel: reserve setup_data: [mem 0x000000000080c000-0x0000000000810fff] usable May 14 05:02:23.843912 kernel: reserve setup_data: [mem 0x0000000000811000-0x00000000008fffff] ACPI NVS May 14 05:02:23.843919 kernel: reserve setup_data: [mem 0x0000000000900000-0x000000009b2e3017] usable May 14 05:02:23.843926 kernel: reserve setup_data: [mem 0x000000009b2e3018-0x000000009b31fe57] usable May 14 05:02:23.843936 kernel: reserve setup_data: [mem 0x000000009b31fe58-0x000000009b320017] usable May 14 05:02:23.843944 kernel: reserve setup_data: [mem 0x000000009b320018-0x000000009b329c57] usable May 14 05:02:23.843953 kernel: reserve setup_data: [mem 0x000000009b329c58-0x000000009bd3efff] usable May 14 05:02:23.843960 kernel: reserve setup_data: [mem 0x000000009bd3f000-0x000000009bdfffff] reserved May 14 05:02:23.843968 kernel: reserve setup_data: [mem 0x000000009be00000-0x000000009c8ecfff] usable May 14 05:02:23.843975 kernel: reserve setup_data: [mem 0x000000009c8ed000-0x000000009cb6cfff] reserved May 14 05:02:23.843982 kernel: reserve setup_data: [mem 0x000000009cb6d000-0x000000009cb7efff] ACPI data May 14 05:02:23.843989 kernel: reserve setup_data: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS May 14 05:02:23.843997 kernel: reserve setup_data: [mem 0x000000009cbff000-0x000000009ce90fff] usable May 14 05:02:23.844006 kernel: reserve setup_data: [mem 0x000000009ce91000-0x000000009ce94fff] reserved May 14 05:02:23.844014 kernel: reserve setup_data: [mem 0x000000009ce95000-0x000000009ce96fff] ACPI NVS May 14 05:02:23.844023 kernel: reserve setup_data: [mem 0x000000009ce97000-0x000000009cedbfff] usable May 14 05:02:23.844030 kernel: reserve setup_data: [mem 0x000000009cedc000-0x000000009cf5ffff] reserved May 14 05:02:23.844037 kernel: reserve setup_data: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS May 14 05:02:23.844044 kernel: reserve setup_data: [mem 0x00000000e0000000-0x00000000efffffff] reserved May 14 05:02:23.844051 kernel: reserve setup_data: [mem 0x00000000feffc000-0x00000000feffffff] reserved May 14 05:02:23.844059 kernel: reserve setup_data: [mem 0x00000000ffc00000-0x00000000ffffffff] reserved May 14 05:02:23.844066 kernel: reserve setup_data: [mem 0x000000fd00000000-0x000000ffffffffff] reserved May 14 05:02:23.844073 kernel: efi: EFI v2.7 by EDK II May 14 05:02:23.844080 kernel: efi: SMBIOS=0x9c988000 ACPI=0x9cb7e000 ACPI 2.0=0x9cb7e014 MEMATTR=0x9b9e4198 RNG=0x9cb73018 May 14 05:02:23.844088 kernel: random: crng init done May 14 05:02:23.844098 kernel: efi: Remove mem149: MMIO range=[0xffc00000-0xffffffff] (4MB) from e820 map May 14 05:02:23.844105 kernel: e820: remove [mem 0xffc00000-0xffffffff] reserved May 14 05:02:23.844116 kernel: secureboot: Secure boot disabled May 14 05:02:23.844123 kernel: SMBIOS 2.8 present. May 14 05:02:23.844130 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS unknown 02/02/2022 May 14 05:02:23.844138 kernel: DMI: Memory slots populated: 1/1 May 14 05:02:23.844145 kernel: Hypervisor detected: KVM May 14 05:02:23.844154 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 May 14 05:02:23.844162 kernel: kvm-clock: using sched offset of 4217715159 cycles May 14 05:02:23.844171 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns May 14 05:02:23.844180 kernel: tsc: Detected 2794.748 MHz processor May 14 05:02:23.844188 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved May 14 05:02:23.844197 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable May 14 05:02:23.844205 kernel: last_pfn = 0x9cedc max_arch_pfn = 0x400000000 May 14 05:02:23.844212 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs May 14 05:02:23.844220 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT May 14 05:02:23.844227 kernel: Using GB pages for direct mapping May 14 05:02:23.844234 kernel: ACPI: Early table checksum verification disabled May 14 05:02:23.844242 kernel: ACPI: RSDP 0x000000009CB7E014 000024 (v02 BOCHS ) May 14 05:02:23.844249 kernel: ACPI: XSDT 0x000000009CB7D0E8 000054 (v01 BOCHS BXPC 00000001 01000013) May 14 05:02:23.844257 kernel: ACPI: FACP 0x000000009CB79000 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) May 14 05:02:23.844266 kernel: ACPI: DSDT 0x000000009CB7A000 0021A8 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 14 05:02:23.844274 kernel: ACPI: FACS 0x000000009CBDD000 000040 May 14 05:02:23.844281 kernel: ACPI: APIC 0x000000009CB78000 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 14 05:02:23.844289 kernel: ACPI: HPET 0x000000009CB77000 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 14 05:02:23.844296 kernel: ACPI: MCFG 0x000000009CB76000 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) May 14 05:02:23.844313 kernel: ACPI: WAET 0x000000009CB75000 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) May 14 05:02:23.844321 kernel: ACPI: BGRT 0x000000009CB74000 000038 (v01 INTEL EDK2 00000002 01000013) May 14 05:02:23.844328 kernel: ACPI: Reserving FACP table memory at [mem 0x9cb79000-0x9cb790f3] May 14 05:02:23.844347 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cb7a000-0x9cb7c1a7] May 14 05:02:23.844356 kernel: ACPI: Reserving FACS table memory at [mem 0x9cbdd000-0x9cbdd03f] May 14 05:02:23.844364 kernel: ACPI: Reserving APIC table memory at [mem 0x9cb78000-0x9cb7808f] May 14 05:02:23.844371 kernel: ACPI: Reserving HPET table memory at [mem 0x9cb77000-0x9cb77037] May 14 05:02:23.844379 kernel: ACPI: Reserving MCFG table memory at [mem 0x9cb76000-0x9cb7603b] May 14 05:02:23.844386 kernel: ACPI: Reserving WAET table memory at [mem 0x9cb75000-0x9cb75027] May 14 05:02:23.844393 kernel: ACPI: Reserving BGRT table memory at [mem 0x9cb74000-0x9cb74037] May 14 05:02:23.844405 kernel: No NUMA configuration found May 14 05:02:23.844412 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cedbfff] May 14 05:02:23.844419 kernel: NODE_DATA(0) allocated [mem 0x9ce36dc0-0x9ce3dfff] May 14 05:02:23.844429 kernel: Zone ranges: May 14 05:02:23.844437 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] May 14 05:02:23.844444 kernel: DMA32 [mem 0x0000000001000000-0x000000009cedbfff] May 14 05:02:23.844451 kernel: Normal empty May 14 05:02:23.844459 kernel: Device empty May 14 05:02:23.844466 kernel: Movable zone start for each node May 14 05:02:23.844473 kernel: Early memory node ranges May 14 05:02:23.844480 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] May 14 05:02:23.844488 kernel: node 0: [mem 0x0000000000100000-0x00000000007fffff] May 14 05:02:23.844498 kernel: node 0: [mem 0x0000000000808000-0x000000000080afff] May 14 05:02:23.844507 kernel: node 0: [mem 0x000000000080c000-0x0000000000810fff] May 14 05:02:23.844515 kernel: node 0: [mem 0x0000000000900000-0x000000009bd3efff] May 14 05:02:23.844522 kernel: node 0: [mem 0x000000009be00000-0x000000009c8ecfff] May 14 05:02:23.844529 kernel: node 0: [mem 0x000000009cbff000-0x000000009ce90fff] May 14 05:02:23.844537 kernel: node 0: [mem 0x000000009ce97000-0x000000009cedbfff] May 14 05:02:23.844544 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cedbfff] May 14 05:02:23.844551 kernel: On node 0, zone DMA: 1 pages in unavailable ranges May 14 05:02:23.844559 kernel: On node 0, zone DMA: 96 pages in unavailable ranges May 14 05:02:23.844575 kernel: On node 0, zone DMA: 8 pages in unavailable ranges May 14 05:02:23.844583 kernel: On node 0, zone DMA: 1 pages in unavailable ranges May 14 05:02:23.844590 kernel: On node 0, zone DMA: 239 pages in unavailable ranges May 14 05:02:23.844598 kernel: On node 0, zone DMA32: 193 pages in unavailable ranges May 14 05:02:23.844607 kernel: On node 0, zone DMA32: 786 pages in unavailable ranges May 14 05:02:23.844615 kernel: On node 0, zone DMA32: 6 pages in unavailable ranges May 14 05:02:23.844623 kernel: On node 0, zone DMA32: 12580 pages in unavailable ranges May 14 05:02:23.844631 kernel: ACPI: PM-Timer IO Port: 0x608 May 14 05:02:23.844638 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) May 14 05:02:23.844648 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 May 14 05:02:23.844656 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) May 14 05:02:23.844663 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) May 14 05:02:23.844671 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) May 14 05:02:23.844679 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) May 14 05:02:23.844687 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) May 14 05:02:23.844694 kernel: ACPI: Using ACPI (MADT) for SMP configuration information May 14 05:02:23.844746 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 May 14 05:02:23.844754 kernel: TSC deadline timer available May 14 05:02:23.844764 kernel: CPU topo: Max. logical packages: 1 May 14 05:02:23.844778 kernel: CPU topo: Max. logical dies: 1 May 14 05:02:23.844786 kernel: CPU topo: Max. dies per package: 1 May 14 05:02:23.844794 kernel: CPU topo: Max. threads per core: 1 May 14 05:02:23.844801 kernel: CPU topo: Num. cores per package: 4 May 14 05:02:23.844809 kernel: CPU topo: Num. threads per package: 4 May 14 05:02:23.844816 kernel: CPU topo: Allowing 4 present CPUs plus 0 hotplug CPUs May 14 05:02:23.844824 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() May 14 05:02:23.844831 kernel: kvm-guest: KVM setup pv remote TLB flush May 14 05:02:23.844841 kernel: kvm-guest: setup PV sched yield May 14 05:02:23.844849 kernel: [mem 0x9d000000-0xdfffffff] available for PCI devices May 14 05:02:23.844857 kernel: Booting paravirtualized kernel on KVM May 14 05:02:23.844865 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns May 14 05:02:23.844873 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:4 nr_cpu_ids:4 nr_node_ids:1 May 14 05:02:23.844880 kernel: percpu: Embedded 60 pages/cpu s207832 r8192 d29736 u524288 May 14 05:02:23.844888 kernel: pcpu-alloc: s207832 r8192 d29736 u524288 alloc=1*2097152 May 14 05:02:23.844896 kernel: pcpu-alloc: [0] 0 1 2 3 May 14 05:02:23.844903 kernel: kvm-guest: PV spinlocks enabled May 14 05:02:23.844913 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) May 14 05:02:23.844922 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=bd5d20a479abde3485dc2e7b97a54e804895b9926289ae86f84794bef32a40f3 May 14 05:02:23.844933 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. May 14 05:02:23.844941 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) May 14 05:02:23.844949 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) May 14 05:02:23.844957 kernel: Fallback order for Node 0: 0 May 14 05:02:23.844964 kernel: Built 1 zonelists, mobility grouping on. Total pages: 641450 May 14 05:02:23.844972 kernel: Policy zone: DMA32 May 14 05:02:23.844982 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off May 14 05:02:23.844989 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 May 14 05:02:23.844997 kernel: ftrace: allocating 40065 entries in 157 pages May 14 05:02:23.845005 kernel: ftrace: allocated 157 pages with 5 groups May 14 05:02:23.845012 kernel: Dynamic Preempt: voluntary May 14 05:02:23.845020 kernel: rcu: Preemptible hierarchical RCU implementation. May 14 05:02:23.845028 kernel: rcu: RCU event tracing is enabled. May 14 05:02:23.845037 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. May 14 05:02:23.845045 kernel: Trampoline variant of Tasks RCU enabled. May 14 05:02:23.845053 kernel: Rude variant of Tasks RCU enabled. May 14 05:02:23.845062 kernel: Tracing variant of Tasks RCU enabled. May 14 05:02:23.845070 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. May 14 05:02:23.845078 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 May 14 05:02:23.845086 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. May 14 05:02:23.845094 kernel: RCU Tasks Rude: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. May 14 05:02:23.845102 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4. May 14 05:02:23.845109 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 May 14 05:02:23.845117 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. May 14 05:02:23.845127 kernel: Console: colour dummy device 80x25 May 14 05:02:23.845135 kernel: printk: legacy console [ttyS0] enabled May 14 05:02:23.845143 kernel: ACPI: Core revision 20240827 May 14 05:02:23.845150 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns May 14 05:02:23.845158 kernel: APIC: Switch to symmetric I/O mode setup May 14 05:02:23.845166 kernel: x2apic enabled May 14 05:02:23.845174 kernel: APIC: Switched APIC routing to: physical x2apic May 14 05:02:23.845181 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() May 14 05:02:23.845189 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() May 14 05:02:23.845199 kernel: kvm-guest: setup PV IPIs May 14 05:02:23.845207 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 May 14 05:02:23.845215 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x2848df6a9de, max_idle_ns: 440795280912 ns May 14 05:02:23.845223 kernel: Calibrating delay loop (skipped) preset value.. 5589.49 BogoMIPS (lpj=2794748) May 14 05:02:23.845230 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated May 14 05:02:23.845238 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 May 14 05:02:23.845246 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 May 14 05:02:23.845253 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization May 14 05:02:23.845261 kernel: Spectre V2 : Mitigation: Retpolines May 14 05:02:23.845271 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch May 14 05:02:23.845279 kernel: Spectre V2 : Spectre v2 / SpectreRSB : Filling RSB on VMEXIT May 14 05:02:23.845286 kernel: Spectre V2 : Enabling Speculation Barrier for firmware calls May 14 05:02:23.845294 kernel: RETBleed: Mitigation: untrained return thunk May 14 05:02:23.845304 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier May 14 05:02:23.845312 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl May 14 05:02:23.845320 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! May 14 05:02:23.845328 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. May 14 05:02:23.845336 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode May 14 05:02:23.845346 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' May 14 05:02:23.845353 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' May 14 05:02:23.845361 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' May 14 05:02:23.845369 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 May 14 05:02:23.845377 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. May 14 05:02:23.845384 kernel: Freeing SMP alternatives memory: 32K May 14 05:02:23.845392 kernel: pid_max: default: 32768 minimum: 301 May 14 05:02:23.845400 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima May 14 05:02:23.845409 kernel: landlock: Up and running. May 14 05:02:23.845417 kernel: SELinux: Initializing. May 14 05:02:23.845425 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) May 14 05:02:23.845433 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) May 14 05:02:23.845440 kernel: smpboot: CPU0: AMD EPYC 7402P 24-Core Processor (family: 0x17, model: 0x31, stepping: 0x0) May 14 05:02:23.845448 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. May 14 05:02:23.845456 kernel: ... version: 0 May 14 05:02:23.845463 kernel: ... bit width: 48 May 14 05:02:23.845471 kernel: ... generic registers: 6 May 14 05:02:23.845479 kernel: ... value mask: 0000ffffffffffff May 14 05:02:23.845488 kernel: ... max period: 00007fffffffffff May 14 05:02:23.845496 kernel: ... fixed-purpose events: 0 May 14 05:02:23.845504 kernel: ... event mask: 000000000000003f May 14 05:02:23.845511 kernel: signal: max sigframe size: 1776 May 14 05:02:23.845519 kernel: rcu: Hierarchical SRCU implementation. May 14 05:02:23.845527 kernel: rcu: Max phase no-delay instances is 400. May 14 05:02:23.845534 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level May 14 05:02:23.845542 kernel: smp: Bringing up secondary CPUs ... May 14 05:02:23.845550 kernel: smpboot: x86: Booting SMP configuration: May 14 05:02:23.845559 kernel: .... node #0, CPUs: #1 #2 #3 May 14 05:02:23.845567 kernel: smp: Brought up 1 node, 4 CPUs May 14 05:02:23.845575 kernel: smpboot: Total of 4 processors activated (22357.98 BogoMIPS) May 14 05:02:23.845583 kernel: Memory: 2422664K/2565800K available (14336K kernel code, 2438K rwdata, 9944K rodata, 54416K init, 2544K bss, 137196K reserved, 0K cma-reserved) May 14 05:02:23.845591 kernel: devtmpfs: initialized May 14 05:02:23.845598 kernel: x86/mm: Memory block size: 128MB May 14 05:02:23.845606 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00800000-0x00807fff] (32768 bytes) May 14 05:02:23.845614 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x0080b000-0x0080bfff] (4096 bytes) May 14 05:02:23.845622 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00811000-0x008fffff] (978944 bytes) May 14 05:02:23.845631 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cb7f000-0x9cbfefff] (524288 bytes) May 14 05:02:23.845639 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9ce95000-0x9ce96fff] (8192 bytes) May 14 05:02:23.845647 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cf60000-0x9cffffff] (655360 bytes) May 14 05:02:23.845655 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns May 14 05:02:23.845663 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) May 14 05:02:23.845670 kernel: pinctrl core: initialized pinctrl subsystem May 14 05:02:23.845678 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family May 14 05:02:23.845686 kernel: audit: initializing netlink subsys (disabled) May 14 05:02:23.845696 kernel: audit: type=2000 audit(1747198942.341:1): state=initialized audit_enabled=0 res=1 May 14 05:02:23.845719 kernel: thermal_sys: Registered thermal governor 'step_wise' May 14 05:02:23.845737 kernel: thermal_sys: Registered thermal governor 'user_space' May 14 05:02:23.845752 kernel: cpuidle: using governor menu May 14 05:02:23.845760 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 May 14 05:02:23.845773 kernel: dca service started, version 1.12.1 May 14 05:02:23.845782 kernel: PCI: ECAM [mem 0xe0000000-0xefffffff] (base 0xe0000000) for domain 0000 [bus 00-ff] May 14 05:02:23.845789 kernel: PCI: Using configuration type 1 for base access May 14 05:02:23.845797 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. May 14 05:02:23.845808 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages May 14 05:02:23.845816 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page May 14 05:02:23.845823 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages May 14 05:02:23.845831 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page May 14 05:02:23.845839 kernel: ACPI: Added _OSI(Module Device) May 14 05:02:23.845847 kernel: ACPI: Added _OSI(Processor Device) May 14 05:02:23.845854 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) May 14 05:02:23.845862 kernel: ACPI: Added _OSI(Processor Aggregator Device) May 14 05:02:23.845870 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded May 14 05:02:23.845879 kernel: ACPI: Interpreter enabled May 14 05:02:23.845887 kernel: ACPI: PM: (supports S0 S3 S5) May 14 05:02:23.845895 kernel: ACPI: Using IOAPIC for interrupt routing May 14 05:02:23.845902 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug May 14 05:02:23.845910 kernel: PCI: Using E820 reservations for host bridge windows May 14 05:02:23.845918 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F May 14 05:02:23.845926 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) May 14 05:02:23.846099 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] May 14 05:02:23.846222 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] May 14 05:02:23.846335 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] May 14 05:02:23.846346 kernel: PCI host bridge to bus 0000:00 May 14 05:02:23.846474 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] May 14 05:02:23.846581 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] May 14 05:02:23.846695 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] May 14 05:02:23.846824 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xdfffffff window] May 14 05:02:23.846939 kernel: pci_bus 0000:00: root bus resource [mem 0xf0000000-0xfebfffff window] May 14 05:02:23.847044 kernel: pci_bus 0000:00: root bus resource [mem 0x380000000000-0x3807ffffffff window] May 14 05:02:23.847147 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] May 14 05:02:23.848433 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 conventional PCI endpoint May 14 05:02:23.848630 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 conventional PCI endpoint May 14 05:02:23.848821 kernel: pci 0000:00:01.0: BAR 0 [mem 0xc0000000-0xc0ffffff pref] May 14 05:02:23.848965 kernel: pci 0000:00:01.0: BAR 2 [mem 0xc1044000-0xc1044fff] May 14 05:02:23.849082 kernel: pci 0000:00:01.0: ROM [mem 0xffff0000-0xffffffff pref] May 14 05:02:23.849202 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] May 14 05:02:23.849326 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 conventional PCI endpoint May 14 05:02:23.849445 kernel: pci 0000:00:02.0: BAR 0 [io 0x6100-0x611f] May 14 05:02:23.849560 kernel: pci 0000:00:02.0: BAR 1 [mem 0xc1043000-0xc1043fff] May 14 05:02:23.849679 kernel: pci 0000:00:02.0: BAR 4 [mem 0x380000000000-0x380000003fff 64bit pref] May 14 05:02:23.849839 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 conventional PCI endpoint May 14 05:02:23.849963 kernel: pci 0000:00:03.0: BAR 0 [io 0x6000-0x607f] May 14 05:02:23.850080 kernel: pci 0000:00:03.0: BAR 1 [mem 0xc1042000-0xc1042fff] May 14 05:02:23.850195 kernel: pci 0000:00:03.0: BAR 4 [mem 0x380000004000-0x380000007fff 64bit pref] May 14 05:02:23.850320 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 conventional PCI endpoint May 14 05:02:23.850436 kernel: pci 0000:00:04.0: BAR 0 [io 0x60e0-0x60ff] May 14 05:02:23.850551 kernel: pci 0000:00:04.0: BAR 1 [mem 0xc1041000-0xc1041fff] May 14 05:02:23.850673 kernel: pci 0000:00:04.0: BAR 4 [mem 0x380000008000-0x38000000bfff 64bit pref] May 14 05:02:23.850865 kernel: pci 0000:00:04.0: ROM [mem 0xfffc0000-0xffffffff pref] May 14 05:02:23.850992 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 conventional PCI endpoint May 14 05:02:23.851109 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO May 14 05:02:23.851232 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 conventional PCI endpoint May 14 05:02:23.851347 kernel: pci 0000:00:1f.2: BAR 4 [io 0x60c0-0x60df] May 14 05:02:23.851462 kernel: pci 0000:00:1f.2: BAR 5 [mem 0xc1040000-0xc1040fff] May 14 05:02:23.851589 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 conventional PCI endpoint May 14 05:02:23.851776 kernel: pci 0000:00:1f.3: BAR 4 [io 0x6080-0x60bf] May 14 05:02:23.851790 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 May 14 05:02:23.851798 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 May 14 05:02:23.851806 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 May 14 05:02:23.851814 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 May 14 05:02:23.851822 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 May 14 05:02:23.851830 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 May 14 05:02:23.851842 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 May 14 05:02:23.851850 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 May 14 05:02:23.851858 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 May 14 05:02:23.851866 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 May 14 05:02:23.851874 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 May 14 05:02:23.851882 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 May 14 05:02:23.851890 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 May 14 05:02:23.851898 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 May 14 05:02:23.851906 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 May 14 05:02:23.851916 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 May 14 05:02:23.851925 kernel: iommu: Default domain type: Translated May 14 05:02:23.851933 kernel: iommu: DMA domain TLB invalidation policy: lazy mode May 14 05:02:23.851941 kernel: efivars: Registered efivars operations May 14 05:02:23.851949 kernel: PCI: Using ACPI for IRQ routing May 14 05:02:23.851957 kernel: PCI: pci_cache_line_size set to 64 bytes May 14 05:02:23.851966 kernel: e820: reserve RAM buffer [mem 0x0080b000-0x008fffff] May 14 05:02:23.851974 kernel: e820: reserve RAM buffer [mem 0x00811000-0x008fffff] May 14 05:02:23.851982 kernel: e820: reserve RAM buffer [mem 0x9b2e3018-0x9bffffff] May 14 05:02:23.851992 kernel: e820: reserve RAM buffer [mem 0x9b320018-0x9bffffff] May 14 05:02:23.852000 kernel: e820: reserve RAM buffer [mem 0x9bd3f000-0x9bffffff] May 14 05:02:23.852008 kernel: e820: reserve RAM buffer [mem 0x9c8ed000-0x9fffffff] May 14 05:02:23.852016 kernel: e820: reserve RAM buffer [mem 0x9ce91000-0x9fffffff] May 14 05:02:23.852024 kernel: e820: reserve RAM buffer [mem 0x9cedc000-0x9fffffff] May 14 05:02:23.852142 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device May 14 05:02:23.852255 kernel: pci 0000:00:01.0: vgaarb: bridge control possible May 14 05:02:23.852367 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none May 14 05:02:23.852382 kernel: vgaarb: loaded May 14 05:02:23.852390 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 May 14 05:02:23.852398 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter May 14 05:02:23.852406 kernel: clocksource: Switched to clocksource kvm-clock May 14 05:02:23.852414 kernel: VFS: Disk quotas dquot_6.6.0 May 14 05:02:23.852423 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) May 14 05:02:23.852431 kernel: pnp: PnP ACPI init May 14 05:02:23.852591 kernel: system 00:05: [mem 0xe0000000-0xefffffff window] has been reserved May 14 05:02:23.852610 kernel: pnp: PnP ACPI: found 6 devices May 14 05:02:23.852618 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns May 14 05:02:23.852627 kernel: NET: Registered PF_INET protocol family May 14 05:02:23.852635 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) May 14 05:02:23.852643 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) May 14 05:02:23.852652 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) May 14 05:02:23.852660 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) May 14 05:02:23.852672 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) May 14 05:02:23.852686 kernel: TCP: Hash tables configured (established 32768 bind 32768) May 14 05:02:23.852698 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) May 14 05:02:23.852721 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) May 14 05:02:23.852729 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family May 14 05:02:23.852737 kernel: NET: Registered PF_XDP protocol family May 14 05:02:23.852867 kernel: pci 0000:00:04.0: ROM [mem 0xfffc0000-0xffffffff pref]: can't claim; no compatible bridge window May 14 05:02:23.852984 kernel: pci 0000:00:04.0: ROM [mem 0x9d000000-0x9d03ffff pref]: assigned May 14 05:02:23.853096 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] May 14 05:02:23.853202 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] May 14 05:02:23.853310 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] May 14 05:02:23.854691 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xdfffffff window] May 14 05:02:23.854827 kernel: pci_bus 0000:00: resource 8 [mem 0xf0000000-0xfebfffff window] May 14 05:02:23.854933 kernel: pci_bus 0000:00: resource 9 [mem 0x380000000000-0x3807ffffffff window] May 14 05:02:23.854944 kernel: PCI: CLS 0 bytes, default 64 May 14 05:02:23.854953 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x2848df6a9de, max_idle_ns: 440795280912 ns May 14 05:02:23.854962 kernel: Initialise system trusted keyrings May 14 05:02:23.854974 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 May 14 05:02:23.854982 kernel: Key type asymmetric registered May 14 05:02:23.854991 kernel: Asymmetric key parser 'x509' registered May 14 05:02:23.854999 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) May 14 05:02:23.855007 kernel: io scheduler mq-deadline registered May 14 05:02:23.855015 kernel: io scheduler kyber registered May 14 05:02:23.855023 kernel: io scheduler bfq registered May 14 05:02:23.855034 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 May 14 05:02:23.855043 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 May 14 05:02:23.855052 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 May 14 05:02:23.855060 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 May 14 05:02:23.855068 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled May 14 05:02:23.855076 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A May 14 05:02:23.855084 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 May 14 05:02:23.855093 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 May 14 05:02:23.855102 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 May 14 05:02:23.855113 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 May 14 05:02:23.855237 kernel: rtc_cmos 00:04: RTC can wake from S4 May 14 05:02:23.855350 kernel: rtc_cmos 00:04: registered as rtc0 May 14 05:02:23.855459 kernel: rtc_cmos 00:04: setting system clock to 2025-05-14T05:02:23 UTC (1747198943) May 14 05:02:23.855566 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram May 14 05:02:23.855577 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled May 14 05:02:23.855585 kernel: efifb: probing for efifb May 14 05:02:23.855597 kernel: efifb: framebuffer at 0xc0000000, using 4000k, total 4000k May 14 05:02:23.855606 kernel: efifb: mode is 1280x800x32, linelength=5120, pages=1 May 14 05:02:23.855614 kernel: efifb: scrolling: redraw May 14 05:02:23.855622 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 May 14 05:02:23.855631 kernel: Console: switching to colour frame buffer device 160x50 May 14 05:02:23.855639 kernel: fb0: EFI VGA frame buffer device May 14 05:02:23.855647 kernel: pstore: Using crash dump compression: deflate May 14 05:02:23.855655 kernel: pstore: Registered efi_pstore as persistent store backend May 14 05:02:23.855663 kernel: NET: Registered PF_INET6 protocol family May 14 05:02:23.855672 kernel: Segment Routing with IPv6 May 14 05:02:23.855682 kernel: In-situ OAM (IOAM) with IPv6 May 14 05:02:23.855690 kernel: NET: Registered PF_PACKET protocol family May 14 05:02:23.855698 kernel: Key type dns_resolver registered May 14 05:02:23.855719 kernel: IPI shorthand broadcast: enabled May 14 05:02:23.855727 kernel: sched_clock: Marking stable (3067002332, 158908581)->(3251878214, -25967301) May 14 05:02:23.855736 kernel: registered taskstats version 1 May 14 05:02:23.855744 kernel: Loading compiled-in X.509 certificates May 14 05:02:23.855752 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.20-flatcar: de56839f264dfa1264ece2be0efda2f53967cc2a' May 14 05:02:23.855760 kernel: Demotion targets for Node 0: null May 14 05:02:23.855777 kernel: Key type .fscrypt registered May 14 05:02:23.855786 kernel: Key type fscrypt-provisioning registered May 14 05:02:23.855794 kernel: ima: No TPM chip found, activating TPM-bypass! May 14 05:02:23.855803 kernel: ima: Allocated hash algorithm: sha1 May 14 05:02:23.855811 kernel: ima: No architecture policies found May 14 05:02:23.855819 kernel: clk: Disabling unused clocks May 14 05:02:23.855827 kernel: Warning: unable to open an initial console. May 14 05:02:23.855836 kernel: Freeing unused kernel image (initmem) memory: 54416K May 14 05:02:23.855847 kernel: Write protecting the kernel read-only data: 24576k May 14 05:02:23.855855 kernel: Freeing unused kernel image (rodata/data gap) memory: 296K May 14 05:02:23.855863 kernel: Run /init as init process May 14 05:02:23.855871 kernel: with arguments: May 14 05:02:23.855879 kernel: /init May 14 05:02:23.855888 kernel: with environment: May 14 05:02:23.855896 kernel: HOME=/ May 14 05:02:23.855904 kernel: TERM=linux May 14 05:02:23.855912 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a May 14 05:02:23.855921 systemd[1]: Successfully made /usr/ read-only. May 14 05:02:23.855936 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) May 14 05:02:23.855946 systemd[1]: Detected virtualization kvm. May 14 05:02:23.855954 systemd[1]: Detected architecture x86-64. May 14 05:02:23.855962 systemd[1]: Running in initrd. May 14 05:02:23.855971 systemd[1]: No hostname configured, using default hostname. May 14 05:02:23.855980 systemd[1]: Hostname set to . May 14 05:02:23.855991 systemd[1]: Initializing machine ID from VM UUID. May 14 05:02:23.856000 systemd[1]: Queued start job for default target initrd.target. May 14 05:02:23.856008 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. May 14 05:02:23.856017 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. May 14 05:02:23.856026 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... May 14 05:02:23.856035 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... May 14 05:02:23.856044 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... May 14 05:02:23.856053 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... May 14 05:02:23.856066 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... May 14 05:02:23.856074 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... May 14 05:02:23.856083 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). May 14 05:02:23.856092 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. May 14 05:02:23.856101 systemd[1]: Reached target paths.target - Path Units. May 14 05:02:23.856109 systemd[1]: Reached target slices.target - Slice Units. May 14 05:02:23.856120 systemd[1]: Reached target swap.target - Swaps. May 14 05:02:23.856129 systemd[1]: Reached target timers.target - Timer Units. May 14 05:02:23.856139 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. May 14 05:02:23.856148 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. May 14 05:02:23.856158 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). May 14 05:02:23.856169 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. May 14 05:02:23.856179 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. May 14 05:02:23.856191 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. May 14 05:02:23.856200 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. May 14 05:02:23.856211 systemd[1]: Reached target sockets.target - Socket Units. May 14 05:02:23.856222 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... May 14 05:02:23.856231 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... May 14 05:02:23.856240 systemd[1]: Finished network-cleanup.service - Network Cleanup. May 14 05:02:23.856249 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). May 14 05:02:23.856258 systemd[1]: Starting systemd-fsck-usr.service... May 14 05:02:23.856267 systemd[1]: Starting systemd-journald.service - Journal Service... May 14 05:02:23.856275 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... May 14 05:02:23.856284 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 14 05:02:23.856292 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. May 14 05:02:23.856304 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. May 14 05:02:23.856313 systemd[1]: Finished systemd-fsck-usr.service. May 14 05:02:23.856322 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... May 14 05:02:23.856353 systemd-journald[220]: Collecting audit messages is disabled. May 14 05:02:23.856379 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. May 14 05:02:23.856388 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... May 14 05:02:23.856398 systemd-journald[220]: Journal started May 14 05:02:23.856420 systemd-journald[220]: Runtime Journal (/run/log/journal/c4bf09972df041de81d977387eb4d3f9) is 6M, max 48.5M, 42.4M free. May 14 05:02:23.842333 systemd-modules-load[221]: Inserted module 'overlay' May 14 05:02:23.859745 systemd[1]: Started systemd-journald.service - Journal Service. May 14 05:02:23.868719 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. May 14 05:02:23.870844 systemd-modules-load[221]: Inserted module 'br_netfilter' May 14 05:02:23.871841 kernel: Bridge firewalling registered May 14 05:02:23.872385 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... May 14 05:02:23.873824 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. May 14 05:02:23.874289 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. May 14 05:02:23.880120 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... May 14 05:02:23.883152 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... May 14 05:02:23.887180 systemd-tmpfiles[239]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. May 14 05:02:23.890838 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. May 14 05:02:23.892336 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. May 14 05:02:23.901622 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. May 14 05:02:23.904122 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... May 14 05:02:23.917859 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. May 14 05:02:23.919033 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... May 14 05:02:23.937687 dracut-cmdline[263]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=bd5d20a479abde3485dc2e7b97a54e804895b9926289ae86f84794bef32a40f3 May 14 05:02:23.958755 systemd-resolved[256]: Positive Trust Anchors: May 14 05:02:23.958779 systemd-resolved[256]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d May 14 05:02:23.958811 systemd-resolved[256]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test May 14 05:02:23.961966 systemd-resolved[256]: Defaulting to hostname 'linux'. May 14 05:02:23.963314 systemd[1]: Started systemd-resolved.service - Network Name Resolution. May 14 05:02:23.968046 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. May 14 05:02:24.053739 kernel: SCSI subsystem initialized May 14 05:02:24.062730 kernel: Loading iSCSI transport class v2.0-870. May 14 05:02:24.072740 kernel: iscsi: registered transport (tcp) May 14 05:02:24.096733 kernel: iscsi: registered transport (qla4xxx) May 14 05:02:24.096811 kernel: QLogic iSCSI HBA Driver May 14 05:02:24.116403 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... May 14 05:02:24.133952 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. May 14 05:02:24.134717 systemd[1]: Reached target network-pre.target - Preparation for Network. May 14 05:02:24.191274 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. May 14 05:02:24.193380 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... May 14 05:02:24.252728 kernel: raid6: avx2x4 gen() 30489 MB/s May 14 05:02:24.269722 kernel: raid6: avx2x2 gen() 30522 MB/s May 14 05:02:24.286815 kernel: raid6: avx2x1 gen() 25971 MB/s May 14 05:02:24.286830 kernel: raid6: using algorithm avx2x2 gen() 30522 MB/s May 14 05:02:24.304818 kernel: raid6: .... xor() 19941 MB/s, rmw enabled May 14 05:02:24.304842 kernel: raid6: using avx2x2 recovery algorithm May 14 05:02:24.325731 kernel: xor: automatically using best checksumming function avx May 14 05:02:24.490760 kernel: Btrfs loaded, zoned=no, fsverity=no May 14 05:02:24.499943 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. May 14 05:02:24.502136 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... May 14 05:02:24.538594 systemd-udevd[472]: Using default interface naming scheme 'v255'. May 14 05:02:24.544671 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. May 14 05:02:24.546617 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... May 14 05:02:24.575367 dracut-pre-trigger[476]: rd.md=0: removing MD RAID activation May 14 05:02:24.604212 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. May 14 05:02:24.608014 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... May 14 05:02:24.675642 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. May 14 05:02:24.680806 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... May 14 05:02:24.711754 kernel: virtio_blk virtio1: 4/0/0 default/read/poll queues May 14 05:02:24.735278 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) May 14 05:02:24.735563 kernel: cryptd: max_cpu_qlen set to 1000 May 14 05:02:24.735584 kernel: AES CTR mode by8 optimization enabled May 14 05:02:24.735604 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input2 May 14 05:02:24.735627 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. May 14 05:02:24.735646 kernel: GPT:9289727 != 19775487 May 14 05:02:24.735665 kernel: GPT:Alternate GPT header not at the end of the disk. May 14 05:02:24.735685 kernel: GPT:9289727 != 19775487 May 14 05:02:24.735701 kernel: GPT: Use GNU Parted to correct GPT errors. May 14 05:02:24.735748 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 May 14 05:02:24.739507 kernel: libata version 3.00 loaded. May 14 05:02:24.753773 kernel: ahci 0000:00:1f.2: version 3.0 May 14 05:02:24.787014 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 May 14 05:02:24.787033 kernel: ahci 0000:00:1f.2: AHCI vers 0001.0000, 32 command slots, 1.5 Gbps, SATA mode May 14 05:02:24.787186 kernel: ahci 0000:00:1f.2: 6/6 ports implemented (port mask 0x3f) May 14 05:02:24.787322 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only May 14 05:02:24.787452 kernel: scsi host0: ahci May 14 05:02:24.787606 kernel: scsi host1: ahci May 14 05:02:24.787772 kernel: scsi host2: ahci May 14 05:02:24.787914 kernel: scsi host3: ahci May 14 05:02:24.788052 kernel: scsi host4: ahci May 14 05:02:24.788185 kernel: scsi host5: ahci May 14 05:02:24.788317 kernel: ata1: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040100 irq 34 lpm-pol 0 May 14 05:02:24.788333 kernel: ata2: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040180 irq 34 lpm-pol 0 May 14 05:02:24.788343 kernel: ata3: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040200 irq 34 lpm-pol 0 May 14 05:02:24.788353 kernel: ata4: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040280 irq 34 lpm-pol 0 May 14 05:02:24.788363 kernel: ata5: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040300 irq 34 lpm-pol 0 May 14 05:02:24.788373 kernel: ata6: SATA max UDMA/133 abar m4096@0xc1040000 port 0xc1040380 irq 34 lpm-pol 0 May 14 05:02:24.788875 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. May 14 05:02:24.800060 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. May 14 05:02:24.816277 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. May 14 05:02:24.825176 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. May 14 05:02:24.828424 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. May 14 05:02:24.837973 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... May 14 05:02:24.839186 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 14 05:02:24.839272 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. May 14 05:02:24.841826 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... May 14 05:02:24.847046 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 14 05:02:24.849338 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. May 14 05:02:24.857368 disk-uuid[630]: Primary Header is updated. May 14 05:02:24.857368 disk-uuid[630]: Secondary Entries is updated. May 14 05:02:24.857368 disk-uuid[630]: Secondary Header is updated. May 14 05:02:24.860767 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 May 14 05:02:24.865744 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 May 14 05:02:24.874884 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. May 14 05:02:25.094418 kernel: ata2: SATA link down (SStatus 0 SControl 300) May 14 05:02:25.094472 kernel: ata5: SATA link down (SStatus 0 SControl 300) May 14 05:02:25.094484 kernel: ata4: SATA link down (SStatus 0 SControl 300) May 14 05:02:25.094497 kernel: ata1: SATA link down (SStatus 0 SControl 300) May 14 05:02:25.095739 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) May 14 05:02:25.096726 kernel: ata6: SATA link down (SStatus 0 SControl 300) May 14 05:02:25.097857 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 May 14 05:02:25.097923 kernel: ata3.00: applying bridge limits May 14 05:02:25.097935 kernel: ata3.00: configured for UDMA/100 May 14 05:02:25.098734 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 May 14 05:02:25.155735 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray May 14 05:02:25.181363 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 May 14 05:02:25.181377 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 May 14 05:02:25.652545 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. May 14 05:02:25.655470 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. May 14 05:02:25.658090 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. May 14 05:02:25.660550 systemd[1]: Reached target remote-fs.target - Remote File Systems. May 14 05:02:25.663995 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... May 14 05:02:25.689101 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. May 14 05:02:25.865743 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 May 14 05:02:25.866531 disk-uuid[634]: The operation has completed successfully. May 14 05:02:25.897865 systemd[1]: disk-uuid.service: Deactivated successfully. May 14 05:02:25.897988 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. May 14 05:02:25.928412 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... May 14 05:02:25.955940 sh[665]: Success May 14 05:02:25.974529 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. May 14 05:02:25.974570 kernel: device-mapper: uevent: version 1.0.3 May 14 05:02:25.974583 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev May 14 05:02:25.983725 kernel: device-mapper: verity: sha256 using shash "sha256-ni" May 14 05:02:26.012844 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. May 14 05:02:26.016665 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... May 14 05:02:26.035000 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. May 14 05:02:26.040853 kernel: BTRFS info: 'norecovery' is for compatibility only, recommended to use 'rescue=nologreplay' May 14 05:02:26.040875 kernel: BTRFS: device fsid 522ba959-9153-4a92-926e-3277bc1060e7 devid 1 transid 40 /dev/mapper/usr (253:0) scanned by mount (677) May 14 05:02:26.042106 kernel: BTRFS info (device dm-0): first mount of filesystem 522ba959-9153-4a92-926e-3277bc1060e7 May 14 05:02:26.042135 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm May 14 05:02:26.042982 kernel: BTRFS info (device dm-0): using free-space-tree May 14 05:02:26.047398 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. May 14 05:02:26.049484 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. May 14 05:02:26.051718 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. May 14 05:02:26.054255 systemd[1]: Starting ignition-setup.service - Ignition (setup)... May 14 05:02:26.057034 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... May 14 05:02:26.090070 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/vda6 (254:6) scanned by mount (710) May 14 05:02:26.090118 kernel: BTRFS info (device vda6): first mount of filesystem 27ac52bc-c86c-4e09-9b91-c3f9e8d3f2a0 May 14 05:02:26.090129 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm May 14 05:02:26.090965 kernel: BTRFS info (device vda6): using free-space-tree May 14 05:02:26.098731 kernel: BTRFS info (device vda6): last unmount of filesystem 27ac52bc-c86c-4e09-9b91-c3f9e8d3f2a0 May 14 05:02:26.099128 systemd[1]: Finished ignition-setup.service - Ignition (setup). May 14 05:02:26.102010 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... May 14 05:02:26.217692 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. May 14 05:02:26.220146 systemd[1]: Starting systemd-networkd.service - Network Configuration... May 14 05:02:26.236218 ignition[755]: Ignition 2.21.0 May 14 05:02:26.236229 ignition[755]: Stage: fetch-offline May 14 05:02:26.236257 ignition[755]: no configs at "/usr/lib/ignition/base.d" May 14 05:02:26.236270 ignition[755]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" May 14 05:02:26.236354 ignition[755]: parsed url from cmdline: "" May 14 05:02:26.236358 ignition[755]: no config URL provided May 14 05:02:26.236363 ignition[755]: reading system config file "/usr/lib/ignition/user.ign" May 14 05:02:26.236371 ignition[755]: no config at "/usr/lib/ignition/user.ign" May 14 05:02:26.236394 ignition[755]: op(1): [started] loading QEMU firmware config module May 14 05:02:26.236400 ignition[755]: op(1): executing: "modprobe" "qemu_fw_cfg" May 14 05:02:26.244081 ignition[755]: op(1): [finished] loading QEMU firmware config module May 14 05:02:26.271496 systemd-networkd[851]: lo: Link UP May 14 05:02:26.271505 systemd-networkd[851]: lo: Gained carrier May 14 05:02:26.272998 systemd-networkd[851]: Enumeration completed May 14 05:02:26.273322 systemd-networkd[851]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 14 05:02:26.273326 systemd-networkd[851]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. May 14 05:02:26.284413 systemd[1]: Started systemd-networkd.service - Network Configuration. May 14 05:02:26.284875 systemd-networkd[851]: eth0: Link UP May 14 05:02:26.284879 systemd-networkd[851]: eth0: Gained carrier May 14 05:02:26.284887 systemd-networkd[851]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 14 05:02:26.296918 systemd[1]: Reached target network.target - Network. May 14 05:02:26.300753 systemd-networkd[851]: eth0: DHCPv4 address 10.0.0.15/16, gateway 10.0.0.1 acquired from 10.0.0.1 May 14 05:02:26.306902 ignition[755]: parsing config with SHA512: 253e4449ac55c0375975ad4f652e00bc068a3636e6aff3b1f97ad922b1132d003dc0a6dffd362f117e15ad7ab785ec4f5172276e87614fe04a9ec64346016a55 May 14 05:02:26.310533 unknown[755]: fetched base config from "system" May 14 05:02:26.310544 unknown[755]: fetched user config from "qemu" May 14 05:02:26.310865 ignition[755]: fetch-offline: fetch-offline passed May 14 05:02:26.310914 ignition[755]: Ignition finished successfully May 14 05:02:26.315379 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). May 14 05:02:26.317914 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). May 14 05:02:26.318764 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... May 14 05:02:26.389354 ignition[860]: Ignition 2.21.0 May 14 05:02:26.389367 ignition[860]: Stage: kargs May 14 05:02:26.389496 ignition[860]: no configs at "/usr/lib/ignition/base.d" May 14 05:02:26.389506 ignition[860]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" May 14 05:02:26.391452 ignition[860]: kargs: kargs passed May 14 05:02:26.391509 ignition[860]: Ignition finished successfully May 14 05:02:26.395979 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). May 14 05:02:26.399191 systemd[1]: Starting ignition-disks.service - Ignition (disks)... May 14 05:02:26.429078 ignition[868]: Ignition 2.21.0 May 14 05:02:26.429091 ignition[868]: Stage: disks May 14 05:02:26.430465 ignition[868]: no configs at "/usr/lib/ignition/base.d" May 14 05:02:26.430485 ignition[868]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" May 14 05:02:26.434538 ignition[868]: disks: disks passed May 14 05:02:26.434596 ignition[868]: Ignition finished successfully May 14 05:02:26.438877 systemd[1]: Finished ignition-disks.service - Ignition (disks). May 14 05:02:26.440221 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. May 14 05:02:26.441524 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. May 14 05:02:26.442079 systemd[1]: Reached target local-fs.target - Local File Systems. May 14 05:02:26.442448 systemd[1]: Reached target sysinit.target - System Initialization. May 14 05:02:26.443000 systemd[1]: Reached target basic.target - Basic System. May 14 05:02:26.444260 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... May 14 05:02:26.473464 systemd-resolved[256]: Detected conflict on linux IN A 10.0.0.15 May 14 05:02:26.473477 systemd-resolved[256]: Hostname conflict, changing published hostname from 'linux' to 'linux10'. May 14 05:02:26.476753 systemd-fsck[879]: ROOT: clean, 15/553520 files, 52789/553472 blocks May 14 05:02:26.485847 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. May 14 05:02:26.490440 systemd[1]: Mounting sysroot.mount - /sysroot... May 14 05:02:26.598754 kernel: EXT4-fs (vda9): mounted filesystem 7fda6268-ffdc-406a-8662-dffb0e9a24fa r/w with ordered data mode. Quota mode: none. May 14 05:02:26.599260 systemd[1]: Mounted sysroot.mount - /sysroot. May 14 05:02:26.601347 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. May 14 05:02:26.606323 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... May 14 05:02:26.608752 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... May 14 05:02:26.610655 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. May 14 05:02:26.610720 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). May 14 05:02:26.612441 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. May 14 05:02:26.618634 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. May 14 05:02:26.621766 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... May 14 05:02:26.625021 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/vda6 (254:6) scanned by mount (888) May 14 05:02:26.627161 kernel: BTRFS info (device vda6): first mount of filesystem 27ac52bc-c86c-4e09-9b91-c3f9e8d3f2a0 May 14 05:02:26.627183 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm May 14 05:02:26.627194 kernel: BTRFS info (device vda6): using free-space-tree May 14 05:02:26.631509 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. May 14 05:02:26.659077 initrd-setup-root[912]: cut: /sysroot/etc/passwd: No such file or directory May 14 05:02:26.663139 initrd-setup-root[919]: cut: /sysroot/etc/group: No such file or directory May 14 05:02:26.667754 initrd-setup-root[926]: cut: /sysroot/etc/shadow: No such file or directory May 14 05:02:26.672208 initrd-setup-root[933]: cut: /sysroot/etc/gshadow: No such file or directory May 14 05:02:26.751653 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. May 14 05:02:26.755312 systemd[1]: Starting ignition-mount.service - Ignition (mount)... May 14 05:02:26.756829 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... May 14 05:02:26.789742 kernel: BTRFS info (device vda6): last unmount of filesystem 27ac52bc-c86c-4e09-9b91-c3f9e8d3f2a0 May 14 05:02:26.800840 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. May 14 05:02:26.818268 ignition[1003]: INFO : Ignition 2.21.0 May 14 05:02:26.818268 ignition[1003]: INFO : Stage: mount May 14 05:02:26.820266 ignition[1003]: INFO : no configs at "/usr/lib/ignition/base.d" May 14 05:02:26.820266 ignition[1003]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" May 14 05:02:26.824653 ignition[1003]: INFO : mount: mount passed May 14 05:02:26.825517 ignition[1003]: INFO : Ignition finished successfully May 14 05:02:26.829110 systemd[1]: Finished ignition-mount.service - Ignition (mount). May 14 05:02:26.831446 systemd[1]: Starting ignition-files.service - Ignition (files)... May 14 05:02:27.040152 systemd[1]: sysroot-oem.mount: Deactivated successfully. May 14 05:02:27.041873 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... May 14 05:02:27.072741 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/vda6 (254:6) scanned by mount (1015) May 14 05:02:27.072827 kernel: BTRFS info (device vda6): first mount of filesystem 27ac52bc-c86c-4e09-9b91-c3f9e8d3f2a0 May 14 05:02:27.074952 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm May 14 05:02:27.074966 kernel: BTRFS info (device vda6): using free-space-tree May 14 05:02:27.079467 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. May 14 05:02:27.118208 ignition[1032]: INFO : Ignition 2.21.0 May 14 05:02:27.118208 ignition[1032]: INFO : Stage: files May 14 05:02:27.118208 ignition[1032]: INFO : no configs at "/usr/lib/ignition/base.d" May 14 05:02:27.122110 ignition[1032]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" May 14 05:02:27.122110 ignition[1032]: DEBUG : files: compiled without relabeling support, skipping May 14 05:02:27.122110 ignition[1032]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" May 14 05:02:27.122110 ignition[1032]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" May 14 05:02:27.128406 ignition[1032]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" May 14 05:02:27.128406 ignition[1032]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" May 14 05:02:27.128406 ignition[1032]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" May 14 05:02:27.128406 ignition[1032]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" May 14 05:02:27.128406 ignition[1032]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.0-linux-amd64.tar.gz: attempt #1 May 14 05:02:27.123781 unknown[1032]: wrote ssh authorized keys file for user: core May 14 05:02:27.172005 ignition[1032]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK May 14 05:02:27.457818 systemd-networkd[851]: eth0: Gained IPv6LL May 14 05:02:27.698308 ignition[1032]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.0-linux-amd64.tar.gz" May 14 05:02:27.700846 ignition[1032]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" May 14 05:02:27.700846 ignition[1032]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" May 14 05:02:27.700846 ignition[1032]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" May 14 05:02:27.700846 ignition[1032]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" May 14 05:02:27.700846 ignition[1032]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" May 14 05:02:27.700846 ignition[1032]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" May 14 05:02:27.700846 ignition[1032]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" May 14 05:02:27.700846 ignition[1032]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" May 14 05:02:27.715919 ignition[1032]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" May 14 05:02:27.715919 ignition[1032]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" May 14 05:02:27.715919 ignition[1032]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.0-x86-64.raw" May 14 05:02:27.715919 ignition[1032]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.0-x86-64.raw" May 14 05:02:27.715919 ignition[1032]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.0-x86-64.raw" May 14 05:02:27.715919 ignition[1032]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.32.0-x86-64.raw: attempt #1 May 14 05:02:28.295295 ignition[1032]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK May 14 05:02:30.182286 ignition[1032]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.0-x86-64.raw" May 14 05:02:30.182286 ignition[1032]: INFO : files: op(b): [started] processing unit "prepare-helm.service" May 14 05:02:30.186629 ignition[1032]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" May 14 05:02:30.188992 ignition[1032]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" May 14 05:02:30.188992 ignition[1032]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" May 14 05:02:30.188992 ignition[1032]: INFO : files: op(d): [started] processing unit "coreos-metadata.service" May 14 05:02:30.194305 ignition[1032]: INFO : files: op(d): op(e): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" May 14 05:02:30.194305 ignition[1032]: INFO : files: op(d): op(e): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" May 14 05:02:30.194305 ignition[1032]: INFO : files: op(d): [finished] processing unit "coreos-metadata.service" May 14 05:02:30.194305 ignition[1032]: INFO : files: op(f): [started] setting preset to disabled for "coreos-metadata.service" May 14 05:02:30.209296 ignition[1032]: INFO : files: op(f): op(10): [started] removing enablement symlink(s) for "coreos-metadata.service" May 14 05:02:30.214978 ignition[1032]: INFO : files: op(f): op(10): [finished] removing enablement symlink(s) for "coreos-metadata.service" May 14 05:02:30.216897 ignition[1032]: INFO : files: op(f): [finished] setting preset to disabled for "coreos-metadata.service" May 14 05:02:30.216897 ignition[1032]: INFO : files: op(11): [started] setting preset to enabled for "prepare-helm.service" May 14 05:02:30.216897 ignition[1032]: INFO : files: op(11): [finished] setting preset to enabled for "prepare-helm.service" May 14 05:02:30.216897 ignition[1032]: INFO : files: createResultFile: createFiles: op(12): [started] writing file "/sysroot/etc/.ignition-result.json" May 14 05:02:30.216897 ignition[1032]: INFO : files: createResultFile: createFiles: op(12): [finished] writing file "/sysroot/etc/.ignition-result.json" May 14 05:02:30.216897 ignition[1032]: INFO : files: files passed May 14 05:02:30.216897 ignition[1032]: INFO : Ignition finished successfully May 14 05:02:30.227328 systemd[1]: Finished ignition-files.service - Ignition (files). May 14 05:02:30.231051 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... May 14 05:02:30.234041 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... May 14 05:02:30.245128 systemd[1]: ignition-quench.service: Deactivated successfully. May 14 05:02:30.245271 systemd[1]: Finished ignition-quench.service - Ignition (record completion). May 14 05:02:30.249795 initrd-setup-root-after-ignition[1061]: grep: /sysroot/oem/oem-release: No such file or directory May 14 05:02:30.253661 initrd-setup-root-after-ignition[1063]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory May 14 05:02:30.255681 initrd-setup-root-after-ignition[1063]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory May 14 05:02:30.257327 initrd-setup-root-after-ignition[1067]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory May 14 05:02:30.256162 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. May 14 05:02:30.258573 systemd[1]: Reached target ignition-complete.target - Ignition Complete. May 14 05:02:30.261698 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... May 14 05:02:30.309861 systemd[1]: initrd-parse-etc.service: Deactivated successfully. May 14 05:02:30.310018 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. May 14 05:02:30.311830 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. May 14 05:02:30.312094 systemd[1]: Reached target initrd.target - Initrd Default Target. May 14 05:02:30.316514 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. May 14 05:02:30.318977 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... May 14 05:02:30.348917 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. May 14 05:02:30.351055 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... May 14 05:02:30.372615 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. May 14 05:02:30.373212 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. May 14 05:02:30.373625 systemd[1]: Stopped target timers.target - Timer Units. May 14 05:02:30.374179 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. May 14 05:02:30.374339 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. May 14 05:02:30.380496 systemd[1]: Stopped target initrd.target - Initrd Default Target. May 14 05:02:30.381073 systemd[1]: Stopped target basic.target - Basic System. May 14 05:02:30.381447 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. May 14 05:02:30.382023 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. May 14 05:02:30.382405 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. May 14 05:02:30.382814 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. May 14 05:02:30.383364 systemd[1]: Stopped target remote-fs.target - Remote File Systems. May 14 05:02:30.383778 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. May 14 05:02:30.384342 systemd[1]: Stopped target sysinit.target - System Initialization. May 14 05:02:30.384747 systemd[1]: Stopped target local-fs.target - Local File Systems. May 14 05:02:30.385291 systemd[1]: Stopped target swap.target - Swaps. May 14 05:02:30.385642 systemd[1]: dracut-pre-mount.service: Deactivated successfully. May 14 05:02:30.385807 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. May 14 05:02:30.407347 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. May 14 05:02:30.408116 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). May 14 05:02:30.408461 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. May 14 05:02:30.408621 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. May 14 05:02:30.412464 systemd[1]: dracut-initqueue.service: Deactivated successfully. May 14 05:02:30.412633 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. May 14 05:02:30.415184 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. May 14 05:02:30.415329 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). May 14 05:02:30.418080 systemd[1]: Stopped target paths.target - Path Units. May 14 05:02:30.420113 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. May 14 05:02:30.425240 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. May 14 05:02:30.425849 systemd[1]: Stopped target slices.target - Slice Units. May 14 05:02:30.426381 systemd[1]: Stopped target sockets.target - Socket Units. May 14 05:02:30.426789 systemd[1]: iscsid.socket: Deactivated successfully. May 14 05:02:30.426914 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. May 14 05:02:30.427361 systemd[1]: iscsiuio.socket: Deactivated successfully. May 14 05:02:30.427470 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. May 14 05:02:30.433566 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. May 14 05:02:30.433766 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. May 14 05:02:30.435483 systemd[1]: ignition-files.service: Deactivated successfully. May 14 05:02:30.435636 systemd[1]: Stopped ignition-files.service - Ignition (files). May 14 05:02:30.439256 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... May 14 05:02:30.440638 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... May 14 05:02:30.443035 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. May 14 05:02:30.443186 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. May 14 05:02:30.445193 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. May 14 05:02:30.445338 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. May 14 05:02:30.454807 systemd[1]: initrd-cleanup.service: Deactivated successfully. May 14 05:02:30.454939 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. May 14 05:02:30.469450 ignition[1088]: INFO : Ignition 2.21.0 May 14 05:02:30.469450 ignition[1088]: INFO : Stage: umount May 14 05:02:30.472628 ignition[1088]: INFO : no configs at "/usr/lib/ignition/base.d" May 14 05:02:30.472628 ignition[1088]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" May 14 05:02:30.474966 ignition[1088]: INFO : umount: umount passed May 14 05:02:30.474966 ignition[1088]: INFO : Ignition finished successfully May 14 05:02:30.475224 systemd[1]: ignition-mount.service: Deactivated successfully. May 14 05:02:30.475368 systemd[1]: Stopped ignition-mount.service - Ignition (mount). May 14 05:02:30.476851 systemd[1]: Stopped target network.target - Network. May 14 05:02:30.478177 systemd[1]: ignition-disks.service: Deactivated successfully. May 14 05:02:30.478263 systemd[1]: Stopped ignition-disks.service - Ignition (disks). May 14 05:02:30.480139 systemd[1]: ignition-kargs.service: Deactivated successfully. May 14 05:02:30.480200 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). May 14 05:02:30.480443 systemd[1]: ignition-setup.service: Deactivated successfully. May 14 05:02:30.480489 systemd[1]: Stopped ignition-setup.service - Ignition (setup). May 14 05:02:30.480954 systemd[1]: ignition-setup-pre.service: Deactivated successfully. May 14 05:02:30.480996 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. May 14 05:02:30.481422 systemd[1]: Stopping systemd-networkd.service - Network Configuration... May 14 05:02:30.487172 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... May 14 05:02:30.488608 systemd[1]: sysroot-boot.mount: Deactivated successfully. May 14 05:02:30.494373 systemd[1]: sysroot-boot.service: Deactivated successfully. May 14 05:02:30.494525 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. May 14 05:02:30.495831 systemd[1]: initrd-setup-root.service: Deactivated successfully. May 14 05:02:30.495926 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. May 14 05:02:30.498356 systemd[1]: systemd-resolved.service: Deactivated successfully. May 14 05:02:30.498490 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. May 14 05:02:30.503143 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. May 14 05:02:30.503901 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. May 14 05:02:30.503991 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. May 14 05:02:30.507848 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. May 14 05:02:30.514012 systemd[1]: systemd-networkd.service: Deactivated successfully. May 14 05:02:30.514169 systemd[1]: Stopped systemd-networkd.service - Network Configuration. May 14 05:02:30.518180 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. May 14 05:02:30.518344 systemd[1]: Stopped target network-pre.target - Preparation for Network. May 14 05:02:30.519238 systemd[1]: systemd-networkd.socket: Deactivated successfully. May 14 05:02:30.519276 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. May 14 05:02:30.526539 systemd[1]: Stopping network-cleanup.service - Network Cleanup... May 14 05:02:30.527647 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. May 14 05:02:30.527798 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. May 14 05:02:30.529436 systemd[1]: systemd-sysctl.service: Deactivated successfully. May 14 05:02:30.529480 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. May 14 05:02:30.533517 systemd[1]: systemd-modules-load.service: Deactivated successfully. May 14 05:02:30.533563 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. May 14 05:02:30.534146 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... May 14 05:02:30.536064 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. May 14 05:02:30.556522 systemd[1]: systemd-udevd.service: Deactivated successfully. May 14 05:02:30.561875 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. May 14 05:02:30.562527 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. May 14 05:02:30.562583 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. May 14 05:02:30.564775 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. May 14 05:02:30.564811 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. May 14 05:02:30.565061 systemd[1]: dracut-pre-udev.service: Deactivated successfully. May 14 05:02:30.565108 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. May 14 05:02:30.569974 systemd[1]: dracut-cmdline.service: Deactivated successfully. May 14 05:02:30.570021 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. May 14 05:02:30.572848 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. May 14 05:02:30.572944 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. May 14 05:02:30.576942 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... May 14 05:02:30.577321 systemd[1]: systemd-network-generator.service: Deactivated successfully. May 14 05:02:30.577372 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. May 14 05:02:30.582279 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. May 14 05:02:30.582371 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. May 14 05:02:30.586686 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. May 14 05:02:30.586751 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. May 14 05:02:30.590111 systemd[1]: kmod-static-nodes.service: Deactivated successfully. May 14 05:02:30.590153 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. May 14 05:02:30.590664 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 14 05:02:30.590721 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. May 14 05:02:30.596248 systemd[1]: network-cleanup.service: Deactivated successfully. May 14 05:02:30.601953 systemd[1]: Stopped network-cleanup.service - Network Cleanup. May 14 05:02:30.611728 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. May 14 05:02:30.611860 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. May 14 05:02:30.612506 systemd[1]: Reached target initrd-switch-root.target - Switch Root. May 14 05:02:30.617322 systemd[1]: Starting initrd-switch-root.service - Switch Root... May 14 05:02:30.642893 systemd[1]: Switching root. May 14 05:02:30.682440 systemd-journald[220]: Journal stopped May 14 05:02:31.785779 systemd-journald[220]: Received SIGTERM from PID 1 (systemd). May 14 05:02:31.785847 kernel: SELinux: policy capability network_peer_controls=1 May 14 05:02:31.785861 kernel: SELinux: policy capability open_perms=1 May 14 05:02:31.785872 kernel: SELinux: policy capability extended_socket_class=1 May 14 05:02:31.785884 kernel: SELinux: policy capability always_check_network=0 May 14 05:02:31.785895 kernel: SELinux: policy capability cgroup_seclabel=1 May 14 05:02:31.785906 kernel: SELinux: policy capability nnp_nosuid_transition=1 May 14 05:02:31.785917 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 May 14 05:02:31.785936 kernel: SELinux: policy capability ioctl_skip_cloexec=0 May 14 05:02:31.785953 kernel: SELinux: policy capability userspace_initial_context=0 May 14 05:02:31.785969 kernel: audit: type=1403 audit(1747198951.030:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 May 14 05:02:31.785982 systemd[1]: Successfully loaded SELinux policy in 47.235ms. May 14 05:02:31.785997 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 12.186ms. May 14 05:02:31.786010 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) May 14 05:02:31.786023 systemd[1]: Detected virtualization kvm. May 14 05:02:31.786040 systemd[1]: Detected architecture x86-64. May 14 05:02:31.786052 systemd[1]: Detected first boot. May 14 05:02:31.786066 systemd[1]: Initializing machine ID from VM UUID. May 14 05:02:31.786078 zram_generator::config[1134]: No configuration found. May 14 05:02:31.786096 kernel: Guest personality initialized and is inactive May 14 05:02:31.786109 kernel: VMCI host device registered (name=vmci, major=10, minor=125) May 14 05:02:31.786120 kernel: Initialized host personality May 14 05:02:31.786132 kernel: NET: Registered PF_VSOCK protocol family May 14 05:02:31.786143 systemd[1]: Populated /etc with preset unit settings. May 14 05:02:31.786156 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. May 14 05:02:31.786170 systemd[1]: initrd-switch-root.service: Deactivated successfully. May 14 05:02:31.786182 systemd[1]: Stopped initrd-switch-root.service - Switch Root. May 14 05:02:31.786196 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. May 14 05:02:31.786208 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. May 14 05:02:31.786220 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. May 14 05:02:31.786232 systemd[1]: Created slice system-getty.slice - Slice /system/getty. May 14 05:02:31.786244 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. May 14 05:02:31.786256 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. May 14 05:02:31.786268 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. May 14 05:02:31.786283 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. May 14 05:02:31.786295 systemd[1]: Created slice user.slice - User and Session Slice. May 14 05:02:31.786312 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. May 14 05:02:31.786324 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. May 14 05:02:31.786336 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. May 14 05:02:31.786348 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. May 14 05:02:31.786360 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. May 14 05:02:31.786375 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... May 14 05:02:31.786387 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... May 14 05:02:31.786399 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). May 14 05:02:31.786411 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. May 14 05:02:31.786423 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. May 14 05:02:31.786435 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. May 14 05:02:31.786447 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. May 14 05:02:31.786461 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. May 14 05:02:31.786473 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. May 14 05:02:31.786487 systemd[1]: Reached target remote-fs.target - Remote File Systems. May 14 05:02:31.786499 systemd[1]: Reached target slices.target - Slice Units. May 14 05:02:31.786511 systemd[1]: Reached target swap.target - Swaps. May 14 05:02:31.786523 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. May 14 05:02:31.786542 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. May 14 05:02:31.786555 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. May 14 05:02:31.786567 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. May 14 05:02:31.786579 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. May 14 05:02:31.786591 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. May 14 05:02:31.786604 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. May 14 05:02:31.786618 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... May 14 05:02:31.786630 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... May 14 05:02:31.786642 systemd[1]: Mounting media.mount - External Media Directory... May 14 05:02:31.786654 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). May 14 05:02:31.786667 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... May 14 05:02:31.786678 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... May 14 05:02:31.786690 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... May 14 05:02:31.786715 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). May 14 05:02:31.786729 systemd[1]: Reached target machines.target - Containers. May 14 05:02:31.786742 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... May 14 05:02:31.786755 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 14 05:02:31.786767 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... May 14 05:02:31.786779 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... May 14 05:02:31.786791 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... May 14 05:02:31.786803 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... May 14 05:02:31.786814 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... May 14 05:02:31.786826 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... May 14 05:02:31.786840 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... May 14 05:02:31.786852 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). May 14 05:02:31.786864 systemd[1]: systemd-fsck-root.service: Deactivated successfully. May 14 05:02:31.786876 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. May 14 05:02:31.786888 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. May 14 05:02:31.786900 systemd[1]: Stopped systemd-fsck-usr.service. May 14 05:02:31.786913 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 14 05:02:31.786925 systemd[1]: Starting systemd-journald.service - Journal Service... May 14 05:02:31.786939 kernel: fuse: init (API version 7.41) May 14 05:02:31.786951 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... May 14 05:02:31.786963 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... May 14 05:02:31.786976 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... May 14 05:02:31.786988 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... May 14 05:02:31.787002 kernel: loop: module loaded May 14 05:02:31.787014 kernel: ACPI: bus type drm_connector registered May 14 05:02:31.787026 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... May 14 05:02:31.787038 systemd[1]: verity-setup.service: Deactivated successfully. May 14 05:02:31.787049 systemd[1]: Stopped verity-setup.service. May 14 05:02:31.787062 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). May 14 05:02:31.787074 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. May 14 05:02:31.787086 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. May 14 05:02:31.787099 systemd[1]: Mounted media.mount - External Media Directory. May 14 05:02:31.787113 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. May 14 05:02:31.787146 systemd-journald[1207]: Collecting audit messages is disabled. May 14 05:02:31.787169 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. May 14 05:02:31.787180 systemd-journald[1207]: Journal started May 14 05:02:31.787205 systemd-journald[1207]: Runtime Journal (/run/log/journal/c4bf09972df041de81d977387eb4d3f9) is 6M, max 48.5M, 42.4M free. May 14 05:02:31.544769 systemd[1]: Queued start job for default target multi-user.target. May 14 05:02:31.566669 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. May 14 05:02:31.567131 systemd[1]: systemd-journald.service: Deactivated successfully. May 14 05:02:31.788761 systemd[1]: Started systemd-journald.service - Journal Service. May 14 05:02:31.790002 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. May 14 05:02:31.791317 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. May 14 05:02:31.792861 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. May 14 05:02:31.794392 systemd[1]: modprobe@configfs.service: Deactivated successfully. May 14 05:02:31.794626 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. May 14 05:02:31.796101 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 14 05:02:31.796315 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. May 14 05:02:31.797908 systemd[1]: modprobe@drm.service: Deactivated successfully. May 14 05:02:31.798124 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. May 14 05:02:31.799468 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 14 05:02:31.799695 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. May 14 05:02:31.801213 systemd[1]: modprobe@fuse.service: Deactivated successfully. May 14 05:02:31.801425 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. May 14 05:02:31.802818 systemd[1]: modprobe@loop.service: Deactivated successfully. May 14 05:02:31.803029 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. May 14 05:02:31.804596 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. May 14 05:02:31.806057 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. May 14 05:02:31.807629 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. May 14 05:02:31.809250 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. May 14 05:02:31.823466 systemd[1]: Reached target network-pre.target - Preparation for Network. May 14 05:02:31.826201 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... May 14 05:02:31.828516 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... May 14 05:02:31.829802 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). May 14 05:02:31.829886 systemd[1]: Reached target local-fs.target - Local File Systems. May 14 05:02:31.832018 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. May 14 05:02:31.840857 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... May 14 05:02:31.842753 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 14 05:02:31.844228 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... May 14 05:02:31.846900 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... May 14 05:02:31.847364 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 14 05:02:31.849321 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... May 14 05:02:31.851011 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. May 14 05:02:31.867837 systemd-journald[1207]: Time spent on flushing to /var/log/journal/c4bf09972df041de81d977387eb4d3f9 is 13.710ms for 1061 entries. May 14 05:02:31.867837 systemd-journald[1207]: System Journal (/var/log/journal/c4bf09972df041de81d977387eb4d3f9) is 8M, max 195.6M, 187.6M free. May 14 05:02:31.901466 systemd-journald[1207]: Received client request to flush runtime journal. May 14 05:02:31.901545 kernel: loop0: detected capacity change from 0 to 146240 May 14 05:02:31.864070 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... May 14 05:02:31.869809 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... May 14 05:02:31.873567 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... May 14 05:02:31.882093 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. May 14 05:02:31.883693 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. May 14 05:02:31.885960 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. May 14 05:02:31.888578 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. May 14 05:02:31.897188 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. May 14 05:02:31.899603 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. May 14 05:02:31.904877 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... May 14 05:02:31.906782 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. May 14 05:02:31.914694 systemd-tmpfiles[1255]: ACLs are not supported, ignoring. May 14 05:02:31.914733 systemd-tmpfiles[1255]: ACLs are not supported, ignoring. May 14 05:02:31.925871 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. May 14 05:02:31.927144 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher May 14 05:02:31.930460 systemd[1]: Starting systemd-sysusers.service - Create System Users... May 14 05:02:31.943815 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. May 14 05:02:31.951231 kernel: loop1: detected capacity change from 0 to 218376 May 14 05:02:31.972999 kernel: loop2: detected capacity change from 0 to 113872 May 14 05:02:31.973734 systemd[1]: Finished systemd-sysusers.service - Create System Users. May 14 05:02:31.977825 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... May 14 05:02:32.006190 systemd-tmpfiles[1275]: ACLs are not supported, ignoring. May 14 05:02:32.006212 systemd-tmpfiles[1275]: ACLs are not supported, ignoring. May 14 05:02:32.009759 kernel: loop3: detected capacity change from 0 to 146240 May 14 05:02:32.012318 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. May 14 05:02:32.024744 kernel: loop4: detected capacity change from 0 to 218376 May 14 05:02:32.033733 kernel: loop5: detected capacity change from 0 to 113872 May 14 05:02:32.044381 (sd-merge)[1278]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. May 14 05:02:32.046009 (sd-merge)[1278]: Merged extensions into '/usr'. May 14 05:02:32.052349 systemd[1]: Reload requested from client PID 1253 ('systemd-sysext') (unit systemd-sysext.service)... May 14 05:02:32.052506 systemd[1]: Reloading... May 14 05:02:32.118733 zram_generator::config[1304]: No configuration found. May 14 05:02:32.191218 ldconfig[1248]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. May 14 05:02:32.244013 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 14 05:02:32.339091 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. May 14 05:02:32.339763 systemd[1]: Reloading finished in 286 ms. May 14 05:02:32.366802 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. May 14 05:02:32.368545 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. May 14 05:02:32.386477 systemd[1]: Starting ensure-sysext.service... May 14 05:02:32.388676 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... May 14 05:02:32.399456 systemd[1]: Reload requested from client PID 1342 ('systemctl') (unit ensure-sysext.service)... May 14 05:02:32.399473 systemd[1]: Reloading... May 14 05:02:32.412837 systemd-tmpfiles[1344]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. May 14 05:02:32.413313 systemd-tmpfiles[1344]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. May 14 05:02:32.413687 systemd-tmpfiles[1344]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. May 14 05:02:32.414108 systemd-tmpfiles[1344]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. May 14 05:02:32.415189 systemd-tmpfiles[1344]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. May 14 05:02:32.415629 systemd-tmpfiles[1344]: ACLs are not supported, ignoring. May 14 05:02:32.415742 systemd-tmpfiles[1344]: ACLs are not supported, ignoring. May 14 05:02:32.421033 systemd-tmpfiles[1344]: Detected autofs mount point /boot during canonicalization of boot. May 14 05:02:32.421049 systemd-tmpfiles[1344]: Skipping /boot May 14 05:02:32.435567 systemd-tmpfiles[1344]: Detected autofs mount point /boot during canonicalization of boot. May 14 05:02:32.435755 systemd-tmpfiles[1344]: Skipping /boot May 14 05:02:32.460827 zram_generator::config[1371]: No configuration found. May 14 05:02:32.558845 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 14 05:02:32.645796 systemd[1]: Reloading finished in 245 ms. May 14 05:02:32.666276 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. May 14 05:02:32.695117 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. May 14 05:02:32.705247 systemd[1]: Starting audit-rules.service - Load Audit Rules... May 14 05:02:32.707919 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... May 14 05:02:32.717634 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... May 14 05:02:32.721318 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... May 14 05:02:32.724570 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... May 14 05:02:32.727774 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... May 14 05:02:32.732531 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). May 14 05:02:32.732714 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 14 05:02:32.735012 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... May 14 05:02:32.738052 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... May 14 05:02:32.741033 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... May 14 05:02:32.742229 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 14 05:02:32.742397 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 14 05:02:32.750988 systemd[1]: Starting systemd-userdbd.service - User Database Manager... May 14 05:02:32.752126 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). May 14 05:02:32.754742 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. May 14 05:02:32.756621 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 14 05:02:32.757035 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. May 14 05:02:32.759670 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 14 05:02:32.759942 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. May 14 05:02:32.762382 systemd[1]: modprobe@loop.service: Deactivated successfully. May 14 05:02:32.762967 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. May 14 05:02:32.770567 systemd-udevd[1414]: Using default interface naming scheme 'v255'. May 14 05:02:32.773146 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). May 14 05:02:32.773563 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 14 05:02:32.776934 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... May 14 05:02:32.780620 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... May 14 05:02:32.786107 augenrules[1445]: No rules May 14 05:02:32.793015 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... May 14 05:02:32.794289 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 14 05:02:32.794449 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 14 05:02:32.796927 systemd[1]: Starting systemd-update-done.service - Update is Completed... May 14 05:02:32.798122 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). May 14 05:02:32.800572 systemd[1]: audit-rules.service: Deactivated successfully. May 14 05:02:32.801740 systemd[1]: Finished audit-rules.service - Load Audit Rules. May 14 05:02:32.803717 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. May 14 05:02:32.806051 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 14 05:02:32.806330 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. May 14 05:02:32.808112 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. May 14 05:02:32.810580 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 14 05:02:32.810909 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. May 14 05:02:32.813257 systemd[1]: modprobe@loop.service: Deactivated successfully. May 14 05:02:32.813553 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. May 14 05:02:32.815742 systemd[1]: Finished systemd-update-done.service - Update is Completed. May 14 05:02:32.826294 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. May 14 05:02:32.828285 systemd[1]: Started systemd-userdbd.service - User Database Manager. May 14 05:02:32.843203 systemd[1]: Finished ensure-sysext.service. May 14 05:02:32.846832 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). May 14 05:02:32.848776 systemd[1]: Starting audit-rules.service - Load Audit Rules... May 14 05:02:32.851012 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 14 05:02:32.852153 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... May 14 05:02:32.859881 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... May 14 05:02:32.865308 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... May 14 05:02:32.868165 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... May 14 05:02:32.869314 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 14 05:02:32.869359 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 14 05:02:32.872056 systemd[1]: Starting systemd-networkd.service - Network Configuration... May 14 05:02:32.875886 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... May 14 05:02:32.877073 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). May 14 05:02:32.877110 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). May 14 05:02:32.878002 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 14 05:02:32.880464 augenrules[1491]: /sbin/augenrules: No change May 14 05:02:32.883245 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. May 14 05:02:32.889014 systemd[1]: modprobe@drm.service: Deactivated successfully. May 14 05:02:32.889254 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. May 14 05:02:32.890966 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 14 05:02:32.891182 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. May 14 05:02:32.893179 systemd[1]: modprobe@loop.service: Deactivated successfully. May 14 05:02:32.893391 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. May 14 05:02:32.904001 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 14 05:02:32.904074 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. May 14 05:02:32.910331 augenrules[1519]: No rules May 14 05:02:32.913333 systemd[1]: audit-rules.service: Deactivated successfully. May 14 05:02:32.913699 systemd[1]: Finished audit-rules.service - Load Audit Rules. May 14 05:02:32.919872 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. May 14 05:02:32.971475 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. May 14 05:02:32.971742 kernel: mousedev: PS/2 mouse device common for all mice May 14 05:02:32.975251 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... May 14 05:02:32.986728 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 May 14 05:02:32.998721 kernel: ACPI: button: Power Button [PWRF] May 14 05:02:33.001285 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. May 14 05:02:33.017944 systemd-resolved[1413]: Positive Trust Anchors: May 14 05:02:33.017965 systemd-resolved[1413]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d May 14 05:02:33.017996 systemd-resolved[1413]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test May 14 05:02:33.023870 systemd-resolved[1413]: Defaulting to hostname 'linux'. May 14 05:02:33.026246 systemd[1]: Started systemd-resolved.service - Network Name Resolution. May 14 05:02:33.027567 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. May 14 05:02:33.030907 kernel: i801_smbus 0000:00:1f.3: Enabling SMBus device May 14 05:02:33.032109 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt May 14 05:02:33.032271 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD May 14 05:02:33.081393 systemd-networkd[1502]: lo: Link UP May 14 05:02:33.081403 systemd-networkd[1502]: lo: Gained carrier May 14 05:02:33.084294 systemd-networkd[1502]: Enumeration completed May 14 05:02:33.084674 systemd-networkd[1502]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 14 05:02:33.084678 systemd-networkd[1502]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. May 14 05:02:33.084794 systemd[1]: Started systemd-networkd.service - Network Configuration. May 14 05:02:33.086054 systemd[1]: Reached target network.target - Network. May 14 05:02:33.088884 systemd-networkd[1502]: eth0: Link UP May 14 05:02:33.089054 systemd-networkd[1502]: eth0: Gained carrier May 14 05:02:33.089072 systemd-networkd[1502]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 14 05:02:33.089364 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... May 14 05:02:33.091859 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... May 14 05:02:33.105768 systemd-networkd[1502]: eth0: DHCPv4 address 10.0.0.15/16, gateway 10.0.0.1 acquired from 10.0.0.1 May 14 05:02:33.111623 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. May 14 05:02:33.113657 systemd[1]: Reached target time-set.target - System Time Set. May 14 05:02:34.971496 systemd-resolved[1413]: Clock change detected. Flushing caches. May 14 05:02:34.971573 systemd-timesyncd[1504]: Contacted time server 10.0.0.1:123 (10.0.0.1). May 14 05:02:34.971619 systemd-timesyncd[1504]: Initial clock synchronization to Wed 2025-05-14 05:02:34.971462 UTC. May 14 05:02:34.994037 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 14 05:02:35.025760 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. May 14 05:02:35.037321 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 14 05:02:35.037651 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. May 14 05:02:35.041448 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. May 14 05:02:35.048676 kernel: kvm_amd: TSC scaling supported May 14 05:02:35.048726 kernel: kvm_amd: Nested Virtualization enabled May 14 05:02:35.048740 kernel: kvm_amd: Nested Paging enabled May 14 05:02:35.048752 kernel: kvm_amd: LBR virtualization supported May 14 05:02:35.049755 kernel: kvm_amd: Virtual VMLOAD VMSAVE supported May 14 05:02:35.049777 kernel: kvm_amd: Virtual GIF supported May 14 05:02:35.050643 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 14 05:02:35.072423 kernel: EDAC MC: Ver: 3.0.0 May 14 05:02:35.110046 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. May 14 05:02:35.111492 systemd[1]: Reached target sysinit.target - System Initialization. May 14 05:02:35.112673 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. May 14 05:02:35.113911 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. May 14 05:02:35.115148 systemd[1]: Started google-oslogin-cache.timer - NSS cache refresh timer. May 14 05:02:35.116451 systemd[1]: Started logrotate.timer - Daily rotation of log files. May 14 05:02:35.117799 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. May 14 05:02:35.119142 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. May 14 05:02:35.120437 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). May 14 05:02:35.120469 systemd[1]: Reached target paths.target - Path Units. May 14 05:02:35.121435 systemd[1]: Reached target timers.target - Timer Units. May 14 05:02:35.123442 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. May 14 05:02:35.126326 systemd[1]: Starting docker.socket - Docker Socket for the API... May 14 05:02:35.129869 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). May 14 05:02:35.131282 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). May 14 05:02:35.132541 systemd[1]: Reached target ssh-access.target - SSH Access Available. May 14 05:02:35.140909 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. May 14 05:02:35.142383 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. May 14 05:02:35.144257 systemd[1]: Listening on docker.socket - Docker Socket for the API. May 14 05:02:35.146052 systemd[1]: Reached target sockets.target - Socket Units. May 14 05:02:35.147003 systemd[1]: Reached target basic.target - Basic System. May 14 05:02:35.147963 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. May 14 05:02:35.147993 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. May 14 05:02:35.148963 systemd[1]: Starting containerd.service - containerd container runtime... May 14 05:02:35.150903 systemd[1]: Starting dbus.service - D-Bus System Message Bus... May 14 05:02:35.152743 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... May 14 05:02:35.155031 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... May 14 05:02:35.158226 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... May 14 05:02:35.159258 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). May 14 05:02:35.160466 systemd[1]: Starting google-oslogin-cache.service - NSS cache refresh... May 14 05:02:35.163303 jq[1574]: false May 14 05:02:35.163726 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... May 14 05:02:35.166467 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... May 14 05:02:35.168919 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... May 14 05:02:35.173593 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... May 14 05:02:35.175813 google_oslogin_nss_cache[1576]: oslogin_cache_refresh[1576]: Refreshing passwd entry cache May 14 05:02:35.175817 oslogin_cache_refresh[1576]: Refreshing passwd entry cache May 14 05:02:35.180478 systemd[1]: Starting systemd-logind.service - User Login Management... May 14 05:02:35.182268 extend-filesystems[1575]: Found loop3 May 14 05:02:35.183666 extend-filesystems[1575]: Found loop4 May 14 05:02:35.183666 extend-filesystems[1575]: Found loop5 May 14 05:02:35.183666 extend-filesystems[1575]: Found sr0 May 14 05:02:35.183666 extend-filesystems[1575]: Found vda May 14 05:02:35.183666 extend-filesystems[1575]: Found vda1 May 14 05:02:35.183666 extend-filesystems[1575]: Found vda2 May 14 05:02:35.183666 extend-filesystems[1575]: Found vda3 May 14 05:02:35.183666 extend-filesystems[1575]: Found usr May 14 05:02:35.183666 extend-filesystems[1575]: Found vda4 May 14 05:02:35.183666 extend-filesystems[1575]: Found vda6 May 14 05:02:35.183666 extend-filesystems[1575]: Found vda7 May 14 05:02:35.183666 extend-filesystems[1575]: Found vda9 May 14 05:02:35.183666 extend-filesystems[1575]: Checking size of /dev/vda9 May 14 05:02:35.186353 oslogin_cache_refresh[1576]: Failure getting users, quitting May 14 05:02:35.193262 google_oslogin_nss_cache[1576]: oslogin_cache_refresh[1576]: Failure getting users, quitting May 14 05:02:35.193262 google_oslogin_nss_cache[1576]: oslogin_cache_refresh[1576]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. May 14 05:02:35.193262 google_oslogin_nss_cache[1576]: oslogin_cache_refresh[1576]: Refreshing group entry cache May 14 05:02:35.183823 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). May 14 05:02:35.186368 oslogin_cache_refresh[1576]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. May 14 05:02:35.184414 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. May 14 05:02:35.186423 oslogin_cache_refresh[1576]: Refreshing group entry cache May 14 05:02:35.194622 systemd[1]: Starting update-engine.service - Update Engine... May 14 05:02:35.195253 google_oslogin_nss_cache[1576]: oslogin_cache_refresh[1576]: Failure getting groups, quitting May 14 05:02:35.195253 google_oslogin_nss_cache[1576]: oslogin_cache_refresh[1576]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. May 14 05:02:35.195297 extend-filesystems[1575]: Resized partition /dev/vda9 May 14 05:02:35.193583 oslogin_cache_refresh[1576]: Failure getting groups, quitting May 14 05:02:35.202929 extend-filesystems[1593]: resize2fs 1.47.2 (1-Jan-2025) May 14 05:02:35.206323 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks May 14 05:02:35.196543 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... May 14 05:02:35.193598 oslogin_cache_refresh[1576]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. May 14 05:02:35.199866 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. May 14 05:02:35.203012 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. May 14 05:02:35.203288 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. May 14 05:02:35.203663 systemd[1]: google-oslogin-cache.service: Deactivated successfully. May 14 05:02:35.204619 systemd[1]: Finished google-oslogin-cache.service - NSS cache refresh. May 14 05:02:35.207781 systemd[1]: motdgen.service: Deactivated successfully. May 14 05:02:35.208187 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. May 14 05:02:35.213458 jq[1594]: true May 14 05:02:35.213875 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. May 14 05:02:35.214116 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. May 14 05:02:35.228033 kernel: EXT4-fs (vda9): resized filesystem to 1864699 May 14 05:02:35.230760 (ntainerd)[1601]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR May 14 05:02:35.235373 jq[1600]: true May 14 05:02:35.242497 update_engine[1589]: I20250514 05:02:35.242431 1589 main.cc:92] Flatcar Update Engine starting May 14 05:02:35.266848 extend-filesystems[1593]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required May 14 05:02:35.266848 extend-filesystems[1593]: old_desc_blocks = 1, new_desc_blocks = 1 May 14 05:02:35.266848 extend-filesystems[1593]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. May 14 05:02:35.270894 extend-filesystems[1575]: Resized filesystem in /dev/vda9 May 14 05:02:35.269755 systemd[1]: extend-filesystems.service: Deactivated successfully. May 14 05:02:35.270165 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. May 14 05:02:35.281091 tar[1598]: linux-amd64/LICENSE May 14 05:02:35.281091 tar[1598]: linux-amd64/helm May 14 05:02:35.286052 systemd-logind[1587]: Watching system buttons on /dev/input/event2 (Power Button) May 14 05:02:35.286084 systemd-logind[1587]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) May 14 05:02:35.288675 systemd-logind[1587]: New seat seat0. May 14 05:02:35.290866 systemd[1]: Started systemd-logind.service - User Login Management. May 14 05:02:35.292502 dbus-daemon[1572]: [system] SELinux support is enabled May 14 05:02:35.292877 systemd[1]: Started dbus.service - D-Bus System Message Bus. May 14 05:02:35.296183 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). May 14 05:02:35.296215 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. May 14 05:02:35.297035 dbus-daemon[1572]: [system] Successfully activated service 'org.freedesktop.systemd1' May 14 05:02:35.297560 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). May 14 05:02:35.297580 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. May 14 05:02:35.297886 update_engine[1589]: I20250514 05:02:35.297832 1589 update_check_scheduler.cc:74] Next update check in 6m43s May 14 05:02:35.300606 systemd[1]: Started update-engine.service - Update Engine. May 14 05:02:35.303391 systemd[1]: Started locksmithd.service - Cluster reboot manager. May 14 05:02:35.318801 bash[1631]: Updated "/home/core/.ssh/authorized_keys" May 14 05:02:35.322799 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. May 14 05:02:35.325013 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. May 14 05:02:35.345201 locksmithd[1632]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" May 14 05:02:35.388934 sshd_keygen[1599]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 May 14 05:02:35.411867 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. May 14 05:02:35.415045 systemd[1]: Starting issuegen.service - Generate /run/issue... May 14 05:02:35.427351 containerd[1601]: time="2025-05-14T05:02:35Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 May 14 05:02:35.428055 containerd[1601]: time="2025-05-14T05:02:35.428020886Z" level=info msg="starting containerd" revision=06b99ca80cdbfbc6cc8bd567021738c9af2b36ce version=v2.0.4 May 14 05:02:35.434011 systemd[1]: issuegen.service: Deactivated successfully. May 14 05:02:35.434293 systemd[1]: Finished issuegen.service - Generate /run/issue. May 14 05:02:35.436335 containerd[1601]: time="2025-05-14T05:02:35.436273125Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="11.372µs" May 14 05:02:35.436335 containerd[1601]: time="2025-05-14T05:02:35.436325092Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 May 14 05:02:35.436426 containerd[1601]: time="2025-05-14T05:02:35.436346743Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 May 14 05:02:35.436672 containerd[1601]: time="2025-05-14T05:02:35.436592514Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 May 14 05:02:35.436672 containerd[1601]: time="2025-05-14T05:02:35.436615116Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 May 14 05:02:35.436672 containerd[1601]: time="2025-05-14T05:02:35.436642688Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 May 14 05:02:35.436746 containerd[1601]: time="2025-05-14T05:02:35.436706658Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 May 14 05:02:35.436746 containerd[1601]: time="2025-05-14T05:02:35.436718159Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 May 14 05:02:35.437978 containerd[1601]: time="2025-05-14T05:02:35.437068096Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 May 14 05:02:35.437978 containerd[1601]: time="2025-05-14T05:02:35.437088544Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 May 14 05:02:35.437978 containerd[1601]: time="2025-05-14T05:02:35.437101158Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 May 14 05:02:35.437978 containerd[1601]: time="2025-05-14T05:02:35.437110445Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 May 14 05:02:35.437978 containerd[1601]: time="2025-05-14T05:02:35.437224469Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 May 14 05:02:35.437978 containerd[1601]: time="2025-05-14T05:02:35.437490588Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 May 14 05:02:35.437978 containerd[1601]: time="2025-05-14T05:02:35.437520404Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 May 14 05:02:35.437978 containerd[1601]: time="2025-05-14T05:02:35.437530864Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 May 14 05:02:35.437978 containerd[1601]: time="2025-05-14T05:02:35.437585797Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 May 14 05:02:35.437978 containerd[1601]: time="2025-05-14T05:02:35.437942135Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 May 14 05:02:35.437521 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... May 14 05:02:35.438495 containerd[1601]: time="2025-05-14T05:02:35.438008520Z" level=info msg="metadata content store policy set" policy=shared May 14 05:02:35.446483 containerd[1601]: time="2025-05-14T05:02:35.446377127Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 May 14 05:02:35.446534 containerd[1601]: time="2025-05-14T05:02:35.446518933Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 May 14 05:02:35.446557 containerd[1601]: time="2025-05-14T05:02:35.446542527Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 May 14 05:02:35.446577 containerd[1601]: time="2025-05-14T05:02:35.446556704Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 May 14 05:02:35.446577 containerd[1601]: time="2025-05-14T05:02:35.446570339Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 May 14 05:02:35.446612 containerd[1601]: time="2025-05-14T05:02:35.446581530Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 May 14 05:02:35.446612 containerd[1601]: time="2025-05-14T05:02:35.446594384Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 May 14 05:02:35.446612 containerd[1601]: time="2025-05-14T05:02:35.446606277Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 May 14 05:02:35.446674 containerd[1601]: time="2025-05-14T05:02:35.446617638Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 May 14 05:02:35.446674 containerd[1601]: time="2025-05-14T05:02:35.446628328Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 May 14 05:02:35.446674 containerd[1601]: time="2025-05-14T05:02:35.446637926Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 May 14 05:02:35.446674 containerd[1601]: time="2025-05-14T05:02:35.446650850Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 May 14 05:02:35.446832 containerd[1601]: time="2025-05-14T05:02:35.446805931Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 May 14 05:02:35.446856 containerd[1601]: time="2025-05-14T05:02:35.446831008Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 May 14 05:02:35.446856 containerd[1601]: time="2025-05-14T05:02:35.446845686Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 May 14 05:02:35.446897 containerd[1601]: time="2025-05-14T05:02:35.446857678Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 May 14 05:02:35.446897 containerd[1601]: time="2025-05-14T05:02:35.446868068Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 May 14 05:02:35.446897 containerd[1601]: time="2025-05-14T05:02:35.446877976Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 May 14 05:02:35.446897 containerd[1601]: time="2025-05-14T05:02:35.446890359Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 May 14 05:02:35.446979 containerd[1601]: time="2025-05-14T05:02:35.446901320Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 May 14 05:02:35.446979 containerd[1601]: time="2025-05-14T05:02:35.446913563Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 May 14 05:02:35.446979 containerd[1601]: time="2025-05-14T05:02:35.446929333Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 May 14 05:02:35.446979 containerd[1601]: time="2025-05-14T05:02:35.446940483Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 May 14 05:02:35.447052 containerd[1601]: time="2025-05-14T05:02:35.447013761Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" May 14 05:02:35.447052 containerd[1601]: time="2025-05-14T05:02:35.447027196Z" level=info msg="Start snapshots syncer" May 14 05:02:35.447090 containerd[1601]: time="2025-05-14T05:02:35.447055860Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 May 14 05:02:35.447356 containerd[1601]: time="2025-05-14T05:02:35.447309436Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" May 14 05:02:35.447485 containerd[1601]: time="2025-05-14T05:02:35.447367695Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 May 14 05:02:35.447485 containerd[1601]: time="2025-05-14T05:02:35.447468063Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 May 14 05:02:35.447586 containerd[1601]: time="2025-05-14T05:02:35.447567389Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 May 14 05:02:35.447609 containerd[1601]: time="2025-05-14T05:02:35.447591004Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 May 14 05:02:35.447609 containerd[1601]: time="2025-05-14T05:02:35.447602996Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 May 14 05:02:35.447646 containerd[1601]: time="2025-05-14T05:02:35.447613766Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 May 14 05:02:35.447646 containerd[1601]: time="2025-05-14T05:02:35.447628614Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 May 14 05:02:35.447646 containerd[1601]: time="2025-05-14T05:02:35.447640286Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 May 14 05:02:35.447704 containerd[1601]: time="2025-05-14T05:02:35.447651527Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 May 14 05:02:35.447704 containerd[1601]: time="2025-05-14T05:02:35.447676163Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 May 14 05:02:35.447704 containerd[1601]: time="2025-05-14T05:02:35.447688256Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 May 14 05:02:35.447704 containerd[1601]: time="2025-05-14T05:02:35.447699026Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 May 14 05:02:35.447773 containerd[1601]: time="2025-05-14T05:02:35.447752036Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 May 14 05:02:35.447773 containerd[1601]: time="2025-05-14T05:02:35.447768316Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 May 14 05:02:35.447867 containerd[1601]: time="2025-05-14T05:02:35.447779167Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 May 14 05:02:35.447867 containerd[1601]: time="2025-05-14T05:02:35.447790338Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 May 14 05:02:35.447867 containerd[1601]: time="2025-05-14T05:02:35.447799014Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 May 14 05:02:35.447867 containerd[1601]: time="2025-05-14T05:02:35.447809213Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 May 14 05:02:35.447867 containerd[1601]: time="2025-05-14T05:02:35.447823309Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 May 14 05:02:35.447867 containerd[1601]: time="2025-05-14T05:02:35.447842616Z" level=info msg="runtime interface created" May 14 05:02:35.447867 containerd[1601]: time="2025-05-14T05:02:35.447847865Z" level=info msg="created NRI interface" May 14 05:02:35.447867 containerd[1601]: time="2025-05-14T05:02:35.447856672Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 May 14 05:02:35.447867 containerd[1601]: time="2025-05-14T05:02:35.447867973Z" level=info msg="Connect containerd service" May 14 05:02:35.448026 containerd[1601]: time="2025-05-14T05:02:35.447891487Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" May 14 05:02:35.448701 containerd[1601]: time="2025-05-14T05:02:35.448669887Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" May 14 05:02:35.461616 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. May 14 05:02:35.465808 systemd[1]: Started getty@tty1.service - Getty on tty1. May 14 05:02:35.470598 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. May 14 05:02:35.471969 systemd[1]: Reached target getty.target - Login Prompts. May 14 05:02:35.537971 containerd[1601]: time="2025-05-14T05:02:35.537895409Z" level=info msg="Start subscribing containerd event" May 14 05:02:35.538108 containerd[1601]: time="2025-05-14T05:02:35.537985217Z" level=info msg="Start recovering state" May 14 05:02:35.538108 containerd[1601]: time="2025-05-14T05:02:35.538014733Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc May 14 05:02:35.538108 containerd[1601]: time="2025-05-14T05:02:35.538080636Z" level=info msg=serving... address=/run/containerd/containerd.sock May 14 05:02:35.538174 containerd[1601]: time="2025-05-14T05:02:35.538142653Z" level=info msg="Start event monitor" May 14 05:02:35.538174 containerd[1601]: time="2025-05-14T05:02:35.538161378Z" level=info msg="Start cni network conf syncer for default" May 14 05:02:35.538211 containerd[1601]: time="2025-05-14T05:02:35.538182247Z" level=info msg="Start streaming server" May 14 05:02:35.538211 containerd[1601]: time="2025-05-14T05:02:35.538200221Z" level=info msg="Registered namespace \"k8s.io\" with NRI" May 14 05:02:35.538211 containerd[1601]: time="2025-05-14T05:02:35.538208576Z" level=info msg="runtime interface starting up..." May 14 05:02:35.538268 containerd[1601]: time="2025-05-14T05:02:35.538217012Z" level=info msg="starting plugins..." May 14 05:02:35.538268 containerd[1601]: time="2025-05-14T05:02:35.538243271Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" May 14 05:02:35.538622 systemd[1]: Started containerd.service - containerd container runtime. May 14 05:02:35.539062 containerd[1601]: time="2025-05-14T05:02:35.539037631Z" level=info msg="containerd successfully booted in 0.112165s" May 14 05:02:35.707238 tar[1598]: linux-amd64/README.md May 14 05:02:35.730485 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. May 14 05:02:36.479627 systemd-networkd[1502]: eth0: Gained IPv6LL May 14 05:02:36.482639 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. May 14 05:02:36.486092 systemd[1]: Reached target network-online.target - Network is Online. May 14 05:02:36.488736 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... May 14 05:02:36.491548 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 14 05:02:36.498972 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... May 14 05:02:36.522926 systemd[1]: coreos-metadata.service: Deactivated successfully. May 14 05:02:36.523237 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. May 14 05:02:36.525250 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. May 14 05:02:36.527710 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. May 14 05:02:37.180054 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 14 05:02:37.181941 systemd[1]: Reached target multi-user.target - Multi-User System. May 14 05:02:37.183493 systemd[1]: Startup finished in 3.151s (kernel) + 7.381s (initrd) + 4.341s (userspace) = 14.874s. May 14 05:02:37.183817 (kubelet)[1703]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 14 05:02:37.576973 kubelet[1703]: E0514 05:02:37.576857 1703 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 14 05:02:37.580491 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 14 05:02:37.580725 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 14 05:02:37.581080 systemd[1]: kubelet.service: Consumed 918ms CPU time, 252.2M memory peak. May 14 05:02:39.133461 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. May 14 05:02:39.134609 systemd[1]: Started sshd@0-10.0.0.15:22-10.0.0.1:52992.service - OpenSSH per-connection server daemon (10.0.0.1:52992). May 14 05:02:39.208811 sshd[1716]: Accepted publickey for core from 10.0.0.1 port 52992 ssh2: RSA SHA256:9779yoEmBEYtokxLadky4y4rhX8tiwkjz8vWSk0iWXU May 14 05:02:39.210753 sshd-session[1716]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 05:02:39.223515 systemd-logind[1587]: New session 1 of user core. May 14 05:02:39.225036 systemd[1]: Created slice user-500.slice - User Slice of UID 500. May 14 05:02:39.226268 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... May 14 05:02:39.263210 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. May 14 05:02:39.265922 systemd[1]: Starting user@500.service - User Manager for UID 500... May 14 05:02:39.286651 (systemd)[1720]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) May 14 05:02:39.288784 systemd-logind[1587]: New session c1 of user core. May 14 05:02:39.428914 systemd[1720]: Queued start job for default target default.target. May 14 05:02:39.445630 systemd[1720]: Created slice app.slice - User Application Slice. May 14 05:02:39.445653 systemd[1720]: Reached target paths.target - Paths. May 14 05:02:39.445694 systemd[1720]: Reached target timers.target - Timers. May 14 05:02:39.447184 systemd[1720]: Starting dbus.socket - D-Bus User Message Bus Socket... May 14 05:02:39.457704 systemd[1720]: Listening on dbus.socket - D-Bus User Message Bus Socket. May 14 05:02:39.457841 systemd[1720]: Reached target sockets.target - Sockets. May 14 05:02:39.457891 systemd[1720]: Reached target basic.target - Basic System. May 14 05:02:39.457942 systemd[1720]: Reached target default.target - Main User Target. May 14 05:02:39.457984 systemd[1720]: Startup finished in 162ms. May 14 05:02:39.458320 systemd[1]: Started user@500.service - User Manager for UID 500. May 14 05:02:39.459921 systemd[1]: Started session-1.scope - Session 1 of User core. May 14 05:02:39.531556 systemd[1]: Started sshd@1-10.0.0.15:22-10.0.0.1:52998.service - OpenSSH per-connection server daemon (10.0.0.1:52998). May 14 05:02:39.590369 sshd[1731]: Accepted publickey for core from 10.0.0.1 port 52998 ssh2: RSA SHA256:9779yoEmBEYtokxLadky4y4rhX8tiwkjz8vWSk0iWXU May 14 05:02:39.591553 sshd-session[1731]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 05:02:39.595874 systemd-logind[1587]: New session 2 of user core. May 14 05:02:39.609517 systemd[1]: Started session-2.scope - Session 2 of User core. May 14 05:02:39.663841 sshd[1733]: Connection closed by 10.0.0.1 port 52998 May 14 05:02:39.664222 sshd-session[1731]: pam_unix(sshd:session): session closed for user core May 14 05:02:39.676463 systemd[1]: sshd@1-10.0.0.15:22-10.0.0.1:52998.service: Deactivated successfully. May 14 05:02:39.678115 systemd[1]: session-2.scope: Deactivated successfully. May 14 05:02:39.678878 systemd-logind[1587]: Session 2 logged out. Waiting for processes to exit. May 14 05:02:39.681409 systemd[1]: Started sshd@2-10.0.0.15:22-10.0.0.1:53010.service - OpenSSH per-connection server daemon (10.0.0.1:53010). May 14 05:02:39.682150 systemd-logind[1587]: Removed session 2. May 14 05:02:39.730459 sshd[1739]: Accepted publickey for core from 10.0.0.1 port 53010 ssh2: RSA SHA256:9779yoEmBEYtokxLadky4y4rhX8tiwkjz8vWSk0iWXU May 14 05:02:39.731753 sshd-session[1739]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 05:02:39.736115 systemd-logind[1587]: New session 3 of user core. May 14 05:02:39.745524 systemd[1]: Started session-3.scope - Session 3 of User core. May 14 05:02:39.795487 sshd[1741]: Connection closed by 10.0.0.1 port 53010 May 14 05:02:39.795862 sshd-session[1739]: pam_unix(sshd:session): session closed for user core May 14 05:02:39.804841 systemd[1]: sshd@2-10.0.0.15:22-10.0.0.1:53010.service: Deactivated successfully. May 14 05:02:39.806344 systemd[1]: session-3.scope: Deactivated successfully. May 14 05:02:39.807114 systemd-logind[1587]: Session 3 logged out. Waiting for processes to exit. May 14 05:02:39.809828 systemd[1]: Started sshd@3-10.0.0.15:22-10.0.0.1:53018.service - OpenSSH per-connection server daemon (10.0.0.1:53018). May 14 05:02:39.810537 systemd-logind[1587]: Removed session 3. May 14 05:02:39.856579 sshd[1747]: Accepted publickey for core from 10.0.0.1 port 53018 ssh2: RSA SHA256:9779yoEmBEYtokxLadky4y4rhX8tiwkjz8vWSk0iWXU May 14 05:02:39.857852 sshd-session[1747]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 05:02:39.862780 systemd-logind[1587]: New session 4 of user core. May 14 05:02:39.872518 systemd[1]: Started session-4.scope - Session 4 of User core. May 14 05:02:39.927698 sshd[1749]: Connection closed by 10.0.0.1 port 53018 May 14 05:02:39.928189 sshd-session[1747]: pam_unix(sshd:session): session closed for user core May 14 05:02:39.945089 systemd[1]: sshd@3-10.0.0.15:22-10.0.0.1:53018.service: Deactivated successfully. May 14 05:02:39.946786 systemd[1]: session-4.scope: Deactivated successfully. May 14 05:02:39.947561 systemd-logind[1587]: Session 4 logged out. Waiting for processes to exit. May 14 05:02:39.950603 systemd[1]: Started sshd@4-10.0.0.15:22-10.0.0.1:53020.service - OpenSSH per-connection server daemon (10.0.0.1:53020). May 14 05:02:39.951331 systemd-logind[1587]: Removed session 4. May 14 05:02:40.002018 sshd[1755]: Accepted publickey for core from 10.0.0.1 port 53020 ssh2: RSA SHA256:9779yoEmBEYtokxLadky4y4rhX8tiwkjz8vWSk0iWXU May 14 05:02:40.003708 sshd-session[1755]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 05:02:40.008355 systemd-logind[1587]: New session 5 of user core. May 14 05:02:40.017548 systemd[1]: Started session-5.scope - Session 5 of User core. May 14 05:02:40.076784 sudo[1758]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 May 14 05:02:40.077097 sudo[1758]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 14 05:02:40.094725 sudo[1758]: pam_unix(sudo:session): session closed for user root May 14 05:02:40.096544 sshd[1757]: Connection closed by 10.0.0.1 port 53020 May 14 05:02:40.096910 sshd-session[1755]: pam_unix(sshd:session): session closed for user core May 14 05:02:40.117728 systemd[1]: sshd@4-10.0.0.15:22-10.0.0.1:53020.service: Deactivated successfully. May 14 05:02:40.119944 systemd[1]: session-5.scope: Deactivated successfully. May 14 05:02:40.120809 systemd-logind[1587]: Session 5 logged out. Waiting for processes to exit. May 14 05:02:40.124665 systemd[1]: Started sshd@5-10.0.0.15:22-10.0.0.1:53028.service - OpenSSH per-connection server daemon (10.0.0.1:53028). May 14 05:02:40.125340 systemd-logind[1587]: Removed session 5. May 14 05:02:40.174093 sshd[1764]: Accepted publickey for core from 10.0.0.1 port 53028 ssh2: RSA SHA256:9779yoEmBEYtokxLadky4y4rhX8tiwkjz8vWSk0iWXU May 14 05:02:40.175846 sshd-session[1764]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 05:02:40.180286 systemd-logind[1587]: New session 6 of user core. May 14 05:02:40.189559 systemd[1]: Started session-6.scope - Session 6 of User core. May 14 05:02:40.244012 sudo[1768]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules May 14 05:02:40.244314 sudo[1768]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 14 05:02:40.252615 sudo[1768]: pam_unix(sudo:session): session closed for user root May 14 05:02:40.259202 sudo[1767]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules May 14 05:02:40.259521 sudo[1767]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 14 05:02:40.270106 systemd[1]: Starting audit-rules.service - Load Audit Rules... May 14 05:02:40.326600 augenrules[1790]: No rules May 14 05:02:40.328407 systemd[1]: audit-rules.service: Deactivated successfully. May 14 05:02:40.328685 systemd[1]: Finished audit-rules.service - Load Audit Rules. May 14 05:02:40.329814 sudo[1767]: pam_unix(sudo:session): session closed for user root May 14 05:02:40.331338 sshd[1766]: Connection closed by 10.0.0.1 port 53028 May 14 05:02:40.331706 sshd-session[1764]: pam_unix(sshd:session): session closed for user core May 14 05:02:40.345144 systemd[1]: sshd@5-10.0.0.15:22-10.0.0.1:53028.service: Deactivated successfully. May 14 05:02:40.347720 systemd[1]: session-6.scope: Deactivated successfully. May 14 05:02:40.348541 systemd-logind[1587]: Session 6 logged out. Waiting for processes to exit. May 14 05:02:40.353031 systemd[1]: Started sshd@6-10.0.0.15:22-10.0.0.1:53044.service - OpenSSH per-connection server daemon (10.0.0.1:53044). May 14 05:02:40.353685 systemd-logind[1587]: Removed session 6. May 14 05:02:40.406738 sshd[1799]: Accepted publickey for core from 10.0.0.1 port 53044 ssh2: RSA SHA256:9779yoEmBEYtokxLadky4y4rhX8tiwkjz8vWSk0iWXU May 14 05:02:40.408180 sshd-session[1799]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 05:02:40.412491 systemd-logind[1587]: New session 7 of user core. May 14 05:02:40.426584 systemd[1]: Started session-7.scope - Session 7 of User core. May 14 05:02:40.479714 sudo[1802]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh May 14 05:02:40.480008 sudo[1802]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 14 05:02:40.901902 systemd[1]: Starting docker.service - Docker Application Container Engine... May 14 05:02:40.919041 (dockerd)[1822]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU May 14 05:02:41.314984 dockerd[1822]: time="2025-05-14T05:02:41.314825248Z" level=info msg="Starting up" May 14 05:02:41.316676 dockerd[1822]: time="2025-05-14T05:02:41.316644641Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" May 14 05:02:41.616295 dockerd[1822]: time="2025-05-14T05:02:41.616173972Z" level=info msg="Loading containers: start." May 14 05:02:41.628423 kernel: Initializing XFRM netlink socket May 14 05:02:41.873962 systemd-networkd[1502]: docker0: Link UP May 14 05:02:41.879680 dockerd[1822]: time="2025-05-14T05:02:41.879639608Z" level=info msg="Loading containers: done." May 14 05:02:41.894496 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck955501236-merged.mount: Deactivated successfully. May 14 05:02:41.896814 dockerd[1822]: time="2025-05-14T05:02:41.896774098Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 May 14 05:02:41.896884 dockerd[1822]: time="2025-05-14T05:02:41.896861892Z" level=info msg="Docker daemon" commit=bbd0a17ccc67e48d4a69393287b7fcc4f0578683 containerd-snapshotter=false storage-driver=overlay2 version=28.0.1 May 14 05:02:41.896986 dockerd[1822]: time="2025-05-14T05:02:41.896960678Z" level=info msg="Initializing buildkit" May 14 05:02:41.926636 dockerd[1822]: time="2025-05-14T05:02:41.926571459Z" level=info msg="Completed buildkit initialization" May 14 05:02:41.933372 dockerd[1822]: time="2025-05-14T05:02:41.933319266Z" level=info msg="Daemon has completed initialization" May 14 05:02:41.933513 dockerd[1822]: time="2025-05-14T05:02:41.933427098Z" level=info msg="API listen on /run/docker.sock" May 14 05:02:41.933648 systemd[1]: Started docker.service - Docker Application Container Engine. May 14 05:02:43.178419 containerd[1601]: time="2025-05-14T05:02:43.178311946Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.4\"" May 14 05:02:43.843434 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3011813926.mount: Deactivated successfully. May 14 05:02:44.957536 containerd[1601]: time="2025-05-14T05:02:44.957471732Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.32.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:02:44.958229 containerd[1601]: time="2025-05-14T05:02:44.958203775Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.32.4: active requests=0, bytes read=28682879" May 14 05:02:44.959473 containerd[1601]: time="2025-05-14T05:02:44.959420818Z" level=info msg="ImageCreate event name:\"sha256:1c20c8797e48698afa3380793df2f1fb260e3209df72d8e864e1bc73af8336e5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:02:44.961823 containerd[1601]: time="2025-05-14T05:02:44.961780904Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:631c6cc78b2862be4fed7df3384a643ef7297eebadae22e8ef9cbe2e19b6386f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:02:44.962857 containerd[1601]: time="2025-05-14T05:02:44.962759891Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.32.4\" with image id \"sha256:1c20c8797e48698afa3380793df2f1fb260e3209df72d8e864e1bc73af8336e5\", repo tag \"registry.k8s.io/kube-apiserver:v1.32.4\", repo digest \"registry.k8s.io/kube-apiserver@sha256:631c6cc78b2862be4fed7df3384a643ef7297eebadae22e8ef9cbe2e19b6386f\", size \"28679679\" in 1.784391919s" May 14 05:02:44.962902 containerd[1601]: time="2025-05-14T05:02:44.962862393Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.4\" returns image reference \"sha256:1c20c8797e48698afa3380793df2f1fb260e3209df72d8e864e1bc73af8336e5\"" May 14 05:02:44.963479 containerd[1601]: time="2025-05-14T05:02:44.963413176Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.4\"" May 14 05:02:46.333502 containerd[1601]: time="2025-05-14T05:02:46.333446938Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.32.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:02:46.334233 containerd[1601]: time="2025-05-14T05:02:46.334203988Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.32.4: active requests=0, bytes read=24779589" May 14 05:02:46.335412 containerd[1601]: time="2025-05-14T05:02:46.335335280Z" level=info msg="ImageCreate event name:\"sha256:4db5364cd5509e0fc8e9f821fbc4b31ed79d4c9ae21809d22030ad67d530a61a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:02:46.339595 containerd[1601]: time="2025-05-14T05:02:46.339549825Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.32.4\" with image id \"sha256:4db5364cd5509e0fc8e9f821fbc4b31ed79d4c9ae21809d22030ad67d530a61a\", repo tag \"registry.k8s.io/kube-controller-manager:v1.32.4\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:25e29187ea66f0ff9b9a00114849c3a30b649005c900a8b2a69e3f3fa56448fb\", size \"26267962\" in 1.376064944s" May 14 05:02:46.339595 containerd[1601]: time="2025-05-14T05:02:46.339589600Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.4\" returns image reference \"sha256:4db5364cd5509e0fc8e9f821fbc4b31ed79d4c9ae21809d22030ad67d530a61a\"" May 14 05:02:46.340085 containerd[1601]: time="2025-05-14T05:02:46.340012212Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.4\"" May 14 05:02:46.340149 containerd[1601]: time="2025-05-14T05:02:46.340105177Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:25e29187ea66f0ff9b9a00114849c3a30b649005c900a8b2a69e3f3fa56448fb\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:02:47.778823 containerd[1601]: time="2025-05-14T05:02:47.778743251Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.32.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:02:47.779581 containerd[1601]: time="2025-05-14T05:02:47.779516731Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.32.4: active requests=0, bytes read=19169938" May 14 05:02:47.780701 containerd[1601]: time="2025-05-14T05:02:47.780650819Z" level=info msg="ImageCreate event name:\"sha256:70a252485ed1f2e8332b6f0a5f8f57443bfbc3c480228f8dcd82ad5ab5cc4000\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:02:47.782998 containerd[1601]: time="2025-05-14T05:02:47.782951905Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:09c55f8dac59a4b8e5e354140f5a4bdd6fa9bd95c42d6bcba6782ed37c31b5a2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:02:47.783877 containerd[1601]: time="2025-05-14T05:02:47.783824131Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.32.4\" with image id \"sha256:70a252485ed1f2e8332b6f0a5f8f57443bfbc3c480228f8dcd82ad5ab5cc4000\", repo tag \"registry.k8s.io/kube-scheduler:v1.32.4\", repo digest \"registry.k8s.io/kube-scheduler@sha256:09c55f8dac59a4b8e5e354140f5a4bdd6fa9bd95c42d6bcba6782ed37c31b5a2\", size \"20658329\" in 1.443783756s" May 14 05:02:47.783877 containerd[1601]: time="2025-05-14T05:02:47.783873133Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.4\" returns image reference \"sha256:70a252485ed1f2e8332b6f0a5f8f57443bfbc3c480228f8dcd82ad5ab5cc4000\"" May 14 05:02:47.784425 containerd[1601]: time="2025-05-14T05:02:47.784380374Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.4\"" May 14 05:02:47.831181 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. May 14 05:02:47.833187 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 14 05:02:48.106229 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 14 05:02:48.124850 (kubelet)[2103]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 14 05:02:48.271221 kubelet[2103]: E0514 05:02:48.271154 2103 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 14 05:02:48.283741 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 14 05:02:48.283945 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 14 05:02:48.284320 systemd[1]: kubelet.service: Consumed 291ms CPU time, 104.2M memory peak. May 14 05:02:49.311160 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2104283292.mount: Deactivated successfully. May 14 05:02:50.075470 containerd[1601]: time="2025-05-14T05:02:50.075370868Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.32.4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:02:50.076158 containerd[1601]: time="2025-05-14T05:02:50.076094976Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.32.4: active requests=0, bytes read=30917856" May 14 05:02:50.077343 containerd[1601]: time="2025-05-14T05:02:50.077303122Z" level=info msg="ImageCreate event name:\"sha256:608f0c8bf7f9651ca79f170235ea5eefb978a0c1da132e7477a88ad37d171ad3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:02:50.079923 containerd[1601]: time="2025-05-14T05:02:50.079875657Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:152638222ecf265eb8e5352e3c50e8fc520994e8ffcff1ee1490c975f7fc2b36\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:02:50.080426 containerd[1601]: time="2025-05-14T05:02:50.080350157Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.32.4\" with image id \"sha256:608f0c8bf7f9651ca79f170235ea5eefb978a0c1da132e7477a88ad37d171ad3\", repo tag \"registry.k8s.io/kube-proxy:v1.32.4\", repo digest \"registry.k8s.io/kube-proxy@sha256:152638222ecf265eb8e5352e3c50e8fc520994e8ffcff1ee1490c975f7fc2b36\", size \"30916875\" in 2.29591484s" May 14 05:02:50.080426 containerd[1601]: time="2025-05-14T05:02:50.080380274Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.4\" returns image reference \"sha256:608f0c8bf7f9651ca79f170235ea5eefb978a0c1da132e7477a88ad37d171ad3\"" May 14 05:02:50.080981 containerd[1601]: time="2025-05-14T05:02:50.080885451Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" May 14 05:02:50.631138 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4119042268.mount: Deactivated successfully. May 14 05:02:51.362477 containerd[1601]: time="2025-05-14T05:02:51.362415956Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:02:51.363164 containerd[1601]: time="2025-05-14T05:02:51.363139574Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=18565241" May 14 05:02:51.364451 containerd[1601]: time="2025-05-14T05:02:51.364391432Z" level=info msg="ImageCreate event name:\"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:02:51.366797 containerd[1601]: time="2025-05-14T05:02:51.366765335Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:02:51.367721 containerd[1601]: time="2025-05-14T05:02:51.367666835Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"18562039\" in 1.286755225s" May 14 05:02:51.367768 containerd[1601]: time="2025-05-14T05:02:51.367722760Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\"" May 14 05:02:51.368156 containerd[1601]: time="2025-05-14T05:02:51.368131307Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" May 14 05:02:51.813915 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3918726202.mount: Deactivated successfully. May 14 05:02:51.820060 containerd[1601]: time="2025-05-14T05:02:51.820009711Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 14 05:02:51.820706 containerd[1601]: time="2025-05-14T05:02:51.820660202Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" May 14 05:02:51.821954 containerd[1601]: time="2025-05-14T05:02:51.821923631Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 14 05:02:51.823898 containerd[1601]: time="2025-05-14T05:02:51.823872066Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 14 05:02:51.824440 containerd[1601]: time="2025-05-14T05:02:51.824385008Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 456.228354ms" May 14 05:02:51.824440 containerd[1601]: time="2025-05-14T05:02:51.824438428Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" May 14 05:02:51.824912 containerd[1601]: time="2025-05-14T05:02:51.824892009Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\"" May 14 05:02:52.353498 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1730018400.mount: Deactivated successfully. May 14 05:02:54.419276 containerd[1601]: time="2025-05-14T05:02:54.419214687Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.16-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:02:54.419898 containerd[1601]: time="2025-05-14T05:02:54.419847714Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.16-0: active requests=0, bytes read=57551360" May 14 05:02:54.420926 containerd[1601]: time="2025-05-14T05:02:54.420897584Z" level=info msg="ImageCreate event name:\"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:02:54.423331 containerd[1601]: time="2025-05-14T05:02:54.423296604Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:02:54.424270 containerd[1601]: time="2025-05-14T05:02:54.424239572Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.16-0\" with image id \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\", repo tag \"registry.k8s.io/etcd:3.5.16-0\", repo digest \"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\", size \"57680541\" in 2.599324159s" May 14 05:02:54.424307 containerd[1601]: time="2025-05-14T05:02:54.424269218Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\" returns image reference \"sha256:a9e7e6b294baf1695fccb862d956c5d3ad8510e1e4ca1535f35dc09f247abbfc\"" May 14 05:02:57.731589 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. May 14 05:02:57.731788 systemd[1]: kubelet.service: Consumed 291ms CPU time, 104.2M memory peak. May 14 05:02:57.734031 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 14 05:02:57.759150 systemd[1]: Reload requested from client PID 2260 ('systemctl') (unit session-7.scope)... May 14 05:02:57.759172 systemd[1]: Reloading... May 14 05:02:57.838425 zram_generator::config[2300]: No configuration found. May 14 05:02:58.063575 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 14 05:02:58.177427 systemd[1]: Reloading finished in 417 ms. May 14 05:02:58.263087 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM May 14 05:02:58.263186 systemd[1]: kubelet.service: Failed with result 'signal'. May 14 05:02:58.263486 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. May 14 05:02:58.263544 systemd[1]: kubelet.service: Consumed 141ms CPU time, 91.8M memory peak. May 14 05:02:58.265343 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 14 05:02:58.416318 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 14 05:02:58.420521 (kubelet)[2351]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS May 14 05:02:58.457534 kubelet[2351]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 14 05:02:58.457534 kubelet[2351]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. May 14 05:02:58.457534 kubelet[2351]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 14 05:02:58.457928 kubelet[2351]: I0514 05:02:58.457619 2351 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" May 14 05:02:58.716067 kubelet[2351]: I0514 05:02:58.715937 2351 server.go:520] "Kubelet version" kubeletVersion="v1.32.0" May 14 05:02:58.716067 kubelet[2351]: I0514 05:02:58.715979 2351 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" May 14 05:02:58.716273 kubelet[2351]: I0514 05:02:58.716251 2351 server.go:954] "Client rotation is on, will bootstrap in background" May 14 05:02:58.738841 kubelet[2351]: E0514 05:02:58.738799 2351 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.0.0.15:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.0.0.15:6443: connect: connection refused" logger="UnhandledError" May 14 05:02:58.739939 kubelet[2351]: I0514 05:02:58.739913 2351 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" May 14 05:02:58.748933 kubelet[2351]: I0514 05:02:58.748522 2351 server.go:1444] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" May 14 05:02:58.754405 kubelet[2351]: I0514 05:02:58.754357 2351 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" May 14 05:02:58.754650 kubelet[2351]: I0514 05:02:58.754614 2351 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] May 14 05:02:58.754810 kubelet[2351]: I0514 05:02:58.754645 2351 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} May 14 05:02:58.754903 kubelet[2351]: I0514 05:02:58.754813 2351 topology_manager.go:138] "Creating topology manager with none policy" May 14 05:02:58.754903 kubelet[2351]: I0514 05:02:58.754822 2351 container_manager_linux.go:304] "Creating device plugin manager" May 14 05:02:58.755381 kubelet[2351]: I0514 05:02:58.755364 2351 state_mem.go:36] "Initialized new in-memory state store" May 14 05:02:58.757835 kubelet[2351]: I0514 05:02:58.757812 2351 kubelet.go:446] "Attempting to sync node with API server" May 14 05:02:58.757835 kubelet[2351]: I0514 05:02:58.757829 2351 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" May 14 05:02:58.757906 kubelet[2351]: I0514 05:02:58.757849 2351 kubelet.go:352] "Adding apiserver pod source" May 14 05:02:58.757906 kubelet[2351]: I0514 05:02:58.757859 2351 apiserver.go:42] "Waiting for node sync before watching apiserver pods" May 14 05:02:58.760923 kubelet[2351]: I0514 05:02:58.760886 2351 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v2.0.4" apiVersion="v1" May 14 05:02:58.761238 kubelet[2351]: I0514 05:02:58.761210 2351 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" May 14 05:02:58.761276 kubelet[2351]: W0514 05:02:58.761269 2351 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. May 14 05:02:58.762580 kubelet[2351]: W0514 05:02:58.762255 2351 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.15:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.15:6443: connect: connection refused May 14 05:02:58.762580 kubelet[2351]: E0514 05:02:58.762306 2351 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.0.0.15:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.15:6443: connect: connection refused" logger="UnhandledError" May 14 05:02:58.763171 kubelet[2351]: W0514 05:02:58.763123 2351 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.15:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.0.0.15:6443: connect: connection refused May 14 05:02:58.763198 kubelet[2351]: E0514 05:02:58.763183 2351 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.0.0.15:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.0.0.15:6443: connect: connection refused" logger="UnhandledError" May 14 05:02:58.763681 kubelet[2351]: I0514 05:02:58.763652 2351 watchdog_linux.go:99] "Systemd watchdog is not enabled" May 14 05:02:58.763741 kubelet[2351]: I0514 05:02:58.763689 2351 server.go:1287] "Started kubelet" May 14 05:02:58.763853 kubelet[2351]: I0514 05:02:58.763815 2351 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 May 14 05:02:58.764820 kubelet[2351]: I0514 05:02:58.764757 2351 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 May 14 05:02:58.765106 kubelet[2351]: I0514 05:02:58.765075 2351 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" May 14 05:02:58.766177 kubelet[2351]: I0514 05:02:58.766147 2351 server.go:490] "Adding debug handlers to kubelet server" May 14 05:02:58.767009 kubelet[2351]: I0514 05:02:58.766335 2351 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" May 14 05:02:58.767154 kubelet[2351]: E0514 05:02:58.767094 2351 kubelet.go:1561] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" May 14 05:02:58.768195 kubelet[2351]: E0514 05:02:58.767258 2351 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" May 14 05:02:58.768195 kubelet[2351]: I0514 05:02:58.767292 2351 volume_manager.go:297] "Starting Kubelet Volume Manager" May 14 05:02:58.768195 kubelet[2351]: I0514 05:02:58.767305 2351 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" May 14 05:02:58.768195 kubelet[2351]: I0514 05:02:58.767712 2351 desired_state_of_world_populator.go:149] "Desired state populator starts to run" May 14 05:02:58.768195 kubelet[2351]: I0514 05:02:58.767755 2351 reconciler.go:26] "Reconciler: start to sync state" May 14 05:02:58.768195 kubelet[2351]: W0514 05:02:58.768114 2351 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.15:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.15:6443: connect: connection refused May 14 05:02:58.768195 kubelet[2351]: E0514 05:02:58.768152 2351 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.0.0.15:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.15:6443: connect: connection refused" logger="UnhandledError" May 14 05:02:58.769303 kubelet[2351]: E0514 05:02:58.768811 2351 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.15:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.15:6443: connect: connection refused" interval="200ms" May 14 05:02:58.771680 kubelet[2351]: I0514 05:02:58.771653 2351 factory.go:221] Registration of the containerd container factory successfully May 14 05:02:58.771680 kubelet[2351]: I0514 05:02:58.771683 2351 factory.go:221] Registration of the systemd container factory successfully May 14 05:02:58.771794 kubelet[2351]: I0514 05:02:58.771765 2351 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory May 14 05:02:58.772314 kubelet[2351]: E0514 05:02:58.771341 2351 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.15:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.15:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.183f4c36d122396b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-05-14 05:02:58.763667819 +0000 UTC m=+0.339563324,LastTimestamp:2025-05-14 05:02:58.763667819 +0000 UTC m=+0.339563324,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" May 14 05:02:58.784585 kubelet[2351]: I0514 05:02:58.784547 2351 cpu_manager.go:221] "Starting CPU manager" policy="none" May 14 05:02:58.784585 kubelet[2351]: I0514 05:02:58.784566 2351 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" May 14 05:02:58.784585 kubelet[2351]: I0514 05:02:58.784583 2351 state_mem.go:36] "Initialized new in-memory state store" May 14 05:02:58.787417 kubelet[2351]: I0514 05:02:58.787356 2351 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" May 14 05:02:58.788615 kubelet[2351]: I0514 05:02:58.788593 2351 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" May 14 05:02:58.788685 kubelet[2351]: I0514 05:02:58.788620 2351 status_manager.go:227] "Starting to sync pod status with apiserver" May 14 05:02:58.788685 kubelet[2351]: I0514 05:02:58.788643 2351 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." May 14 05:02:58.788685 kubelet[2351]: I0514 05:02:58.788650 2351 kubelet.go:2388] "Starting kubelet main sync loop" May 14 05:02:58.788747 kubelet[2351]: E0514 05:02:58.788696 2351 kubelet.go:2412] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" May 14 05:02:58.790114 kubelet[2351]: W0514 05:02:58.789249 2351 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.15:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.15:6443: connect: connection refused May 14 05:02:58.790114 kubelet[2351]: E0514 05:02:58.789284 2351 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.0.0.15:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.15:6443: connect: connection refused" logger="UnhandledError" May 14 05:02:58.867944 kubelet[2351]: E0514 05:02:58.867897 2351 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" May 14 05:02:58.889268 kubelet[2351]: E0514 05:02:58.889214 2351 kubelet.go:2412] "Skipping pod synchronization" err="container runtime status check may not have completed yet" May 14 05:02:58.968689 kubelet[2351]: E0514 05:02:58.968542 2351 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" May 14 05:02:58.970144 kubelet[2351]: E0514 05:02:58.970092 2351 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.15:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.15:6443: connect: connection refused" interval="400ms" May 14 05:02:59.069518 kubelet[2351]: E0514 05:02:59.069457 2351 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" May 14 05:02:59.089566 kubelet[2351]: E0514 05:02:59.089525 2351 kubelet.go:2412] "Skipping pod synchronization" err="container runtime status check may not have completed yet" May 14 05:02:59.116803 kubelet[2351]: I0514 05:02:59.116765 2351 policy_none.go:49] "None policy: Start" May 14 05:02:59.116849 kubelet[2351]: I0514 05:02:59.116823 2351 memory_manager.go:186] "Starting memorymanager" policy="None" May 14 05:02:59.116849 kubelet[2351]: I0514 05:02:59.116838 2351 state_mem.go:35] "Initializing new in-memory state store" May 14 05:02:59.126082 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. May 14 05:02:59.137022 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. May 14 05:02:59.140738 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. May 14 05:02:59.159632 kubelet[2351]: I0514 05:02:59.159576 2351 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" May 14 05:02:59.159846 kubelet[2351]: I0514 05:02:59.159826 2351 eviction_manager.go:189] "Eviction manager: starting control loop" May 14 05:02:59.159895 kubelet[2351]: I0514 05:02:59.159844 2351 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" May 14 05:02:59.160133 kubelet[2351]: I0514 05:02:59.160112 2351 plugin_manager.go:118] "Starting Kubelet Plugin Manager" May 14 05:02:59.160935 kubelet[2351]: E0514 05:02:59.160880 2351 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" May 14 05:02:59.160935 kubelet[2351]: E0514 05:02:59.160929 2351 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" May 14 05:02:59.261585 kubelet[2351]: I0514 05:02:59.261473 2351 kubelet_node_status.go:76] "Attempting to register node" node="localhost" May 14 05:02:59.261872 kubelet[2351]: E0514 05:02:59.261842 2351 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://10.0.0.15:6443/api/v1/nodes\": dial tcp 10.0.0.15:6443: connect: connection refused" node="localhost" May 14 05:02:59.371457 kubelet[2351]: E0514 05:02:59.371375 2351 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.15:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.15:6443: connect: connection refused" interval="800ms" May 14 05:02:59.463638 kubelet[2351]: I0514 05:02:59.463597 2351 kubelet_node_status.go:76] "Attempting to register node" node="localhost" May 14 05:02:59.463999 kubelet[2351]: E0514 05:02:59.463943 2351 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://10.0.0.15:6443/api/v1/nodes\": dial tcp 10.0.0.15:6443: connect: connection refused" node="localhost" May 14 05:02:59.498524 systemd[1]: Created slice kubepods-burstable-pod4ef2178d5f83051e6eb1068e93ce70df.slice - libcontainer container kubepods-burstable-pod4ef2178d5f83051e6eb1068e93ce70df.slice. May 14 05:02:59.518335 kubelet[2351]: E0514 05:02:59.518249 2351 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" May 14 05:02:59.520811 systemd[1]: Created slice kubepods-burstable-pod5386fe11ed933ab82453de11903c7f47.slice - libcontainer container kubepods-burstable-pod5386fe11ed933ab82453de11903c7f47.slice. May 14 05:02:59.522593 kubelet[2351]: E0514 05:02:59.522557 2351 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" May 14 05:02:59.542253 systemd[1]: Created slice kubepods-burstable-pod2980a8ab51edc665be10a02e33130e15.slice - libcontainer container kubepods-burstable-pod2980a8ab51edc665be10a02e33130e15.slice. May 14 05:02:59.543976 kubelet[2351]: E0514 05:02:59.543945 2351 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" May 14 05:02:59.572548 kubelet[2351]: I0514 05:02:59.572484 2351 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/4ef2178d5f83051e6eb1068e93ce70df-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"4ef2178d5f83051e6eb1068e93ce70df\") " pod="kube-system/kube-apiserver-localhost" May 14 05:02:59.572548 kubelet[2351]: I0514 05:02:59.572539 2351 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/5386fe11ed933ab82453de11903c7f47-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"5386fe11ed933ab82453de11903c7f47\") " pod="kube-system/kube-controller-manager-localhost" May 14 05:02:59.572609 kubelet[2351]: I0514 05:02:59.572565 2351 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/5386fe11ed933ab82453de11903c7f47-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"5386fe11ed933ab82453de11903c7f47\") " pod="kube-system/kube-controller-manager-localhost" May 14 05:02:59.572642 kubelet[2351]: I0514 05:02:59.572602 2351 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/4ef2178d5f83051e6eb1068e93ce70df-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"4ef2178d5f83051e6eb1068e93ce70df\") " pod="kube-system/kube-apiserver-localhost" May 14 05:02:59.572695 kubelet[2351]: I0514 05:02:59.572660 2351 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/4ef2178d5f83051e6eb1068e93ce70df-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"4ef2178d5f83051e6eb1068e93ce70df\") " pod="kube-system/kube-apiserver-localhost" May 14 05:02:59.572746 kubelet[2351]: I0514 05:02:59.572728 2351 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/5386fe11ed933ab82453de11903c7f47-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"5386fe11ed933ab82453de11903c7f47\") " pod="kube-system/kube-controller-manager-localhost" May 14 05:02:59.572770 kubelet[2351]: I0514 05:02:59.572746 2351 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/5386fe11ed933ab82453de11903c7f47-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"5386fe11ed933ab82453de11903c7f47\") " pod="kube-system/kube-controller-manager-localhost" May 14 05:02:59.572770 kubelet[2351]: I0514 05:02:59.572762 2351 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/5386fe11ed933ab82453de11903c7f47-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"5386fe11ed933ab82453de11903c7f47\") " pod="kube-system/kube-controller-manager-localhost" May 14 05:02:59.572809 kubelet[2351]: I0514 05:02:59.572786 2351 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/2980a8ab51edc665be10a02e33130e15-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"2980a8ab51edc665be10a02e33130e15\") " pod="kube-system/kube-scheduler-localhost" May 14 05:02:59.576938 kubelet[2351]: W0514 05:02:59.576885 2351 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.15:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.15:6443: connect: connection refused May 14 05:02:59.576938 kubelet[2351]: E0514 05:02:59.576935 2351 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.0.0.15:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.0.0.15:6443: connect: connection refused" logger="UnhandledError" May 14 05:02:59.819632 kubelet[2351]: E0514 05:02:59.819499 2351 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:02:59.820165 containerd[1601]: time="2025-05-14T05:02:59.820106055Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:4ef2178d5f83051e6eb1068e93ce70df,Namespace:kube-system,Attempt:0,}" May 14 05:02:59.823562 kubelet[2351]: E0514 05:02:59.823517 2351 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:02:59.823940 containerd[1601]: time="2025-05-14T05:02:59.823912125Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:5386fe11ed933ab82453de11903c7f47,Namespace:kube-system,Attempt:0,}" May 14 05:02:59.845250 kubelet[2351]: E0514 05:02:59.845213 2351 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:02:59.845531 containerd[1601]: time="2025-05-14T05:02:59.845479469Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:2980a8ab51edc665be10a02e33130e15,Namespace:kube-system,Attempt:0,}" May 14 05:02:59.865563 kubelet[2351]: I0514 05:02:59.865542 2351 kubelet_node_status.go:76] "Attempting to register node" node="localhost" May 14 05:02:59.865938 kubelet[2351]: E0514 05:02:59.865909 2351 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://10.0.0.15:6443/api/v1/nodes\": dial tcp 10.0.0.15:6443: connect: connection refused" node="localhost" May 14 05:02:59.918718 kubelet[2351]: W0514 05:02:59.918670 2351 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.15:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.15:6443: connect: connection refused May 14 05:02:59.918718 kubelet[2351]: E0514 05:02:59.918702 2351 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.0.0.15:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.0.0.15:6443: connect: connection refused" logger="UnhandledError" May 14 05:02:59.952949 containerd[1601]: time="2025-05-14T05:02:59.952869361Z" level=info msg="connecting to shim 1defce71c5080de071151d34ccb39afe4223a836f9d2487e8d7d806b34570b42" address="unix:///run/containerd/s/003bedb507a677a3e3695ad348bc9ac8ecb17fc3fda90e262951031565508438" namespace=k8s.io protocol=ttrpc version=3 May 14 05:02:59.961796 containerd[1601]: time="2025-05-14T05:02:59.961731524Z" level=info msg="connecting to shim 336f7da1e4d23b5bacd1626e7726a0a82298044ef4b3e1a1aa0c5e87c19b746c" address="unix:///run/containerd/s/0cd830b36f34f6eff93a2b3242fbaa40fd70c212b56906f1959cd7fec7f03ad1" namespace=k8s.io protocol=ttrpc version=3 May 14 05:02:59.973493 containerd[1601]: time="2025-05-14T05:02:59.973419157Z" level=info msg="connecting to shim 54c449977658c080a797fd8661ce53ecf2f98f79f2c74676409161c5a858fac7" address="unix:///run/containerd/s/817e1c4c03a055044163dae1fc4053a876210f255d95d659010b22b4200723e2" namespace=k8s.io protocol=ttrpc version=3 May 14 05:02:59.986548 systemd[1]: Started cri-containerd-1defce71c5080de071151d34ccb39afe4223a836f9d2487e8d7d806b34570b42.scope - libcontainer container 1defce71c5080de071151d34ccb39afe4223a836f9d2487e8d7d806b34570b42. May 14 05:02:59.988025 systemd[1]: Started cri-containerd-336f7da1e4d23b5bacd1626e7726a0a82298044ef4b3e1a1aa0c5e87c19b746c.scope - libcontainer container 336f7da1e4d23b5bacd1626e7726a0a82298044ef4b3e1a1aa0c5e87c19b746c. May 14 05:02:59.999534 systemd[1]: Started cri-containerd-54c449977658c080a797fd8661ce53ecf2f98f79f2c74676409161c5a858fac7.scope - libcontainer container 54c449977658c080a797fd8661ce53ecf2f98f79f2c74676409161c5a858fac7. May 14 05:03:00.009071 kubelet[2351]: W0514 05:03:00.009016 2351 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.15:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.15:6443: connect: connection refused May 14 05:03:00.009149 kubelet[2351]: E0514 05:03:00.009080 2351 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.0.0.15:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 10.0.0.15:6443: connect: connection refused" logger="UnhandledError" May 14 05:03:00.035099 containerd[1601]: time="2025-05-14T05:03:00.035044158Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:4ef2178d5f83051e6eb1068e93ce70df,Namespace:kube-system,Attempt:0,} returns sandbox id \"1defce71c5080de071151d34ccb39afe4223a836f9d2487e8d7d806b34570b42\"" May 14 05:03:00.036858 kubelet[2351]: E0514 05:03:00.036826 2351 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:00.039356 containerd[1601]: time="2025-05-14T05:03:00.039306232Z" level=info msg="CreateContainer within sandbox \"1defce71c5080de071151d34ccb39afe4223a836f9d2487e8d7d806b34570b42\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" May 14 05:03:00.040702 containerd[1601]: time="2025-05-14T05:03:00.040664480Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:5386fe11ed933ab82453de11903c7f47,Namespace:kube-system,Attempt:0,} returns sandbox id \"336f7da1e4d23b5bacd1626e7726a0a82298044ef4b3e1a1aa0c5e87c19b746c\"" May 14 05:03:00.041294 kubelet[2351]: E0514 05:03:00.041198 2351 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:00.044565 containerd[1601]: time="2025-05-14T05:03:00.044531203Z" level=info msg="CreateContainer within sandbox \"336f7da1e4d23b5bacd1626e7726a0a82298044ef4b3e1a1aa0c5e87c19b746c\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" May 14 05:03:00.048802 containerd[1601]: time="2025-05-14T05:03:00.048771126Z" level=info msg="Container 0125cd70a16c1cd7dc3e0d5935a202413016a1865c4ff592c66d8aa7f6bd85f2: CDI devices from CRI Config.CDIDevices: []" May 14 05:03:00.058236 containerd[1601]: time="2025-05-14T05:03:00.058198198Z" level=info msg="CreateContainer within sandbox \"1defce71c5080de071151d34ccb39afe4223a836f9d2487e8d7d806b34570b42\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"0125cd70a16c1cd7dc3e0d5935a202413016a1865c4ff592c66d8aa7f6bd85f2\"" May 14 05:03:00.058850 containerd[1601]: time="2025-05-14T05:03:00.058829432Z" level=info msg="StartContainer for \"0125cd70a16c1cd7dc3e0d5935a202413016a1865c4ff592c66d8aa7f6bd85f2\"" May 14 05:03:00.058900 containerd[1601]: time="2025-05-14T05:03:00.058878845Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:2980a8ab51edc665be10a02e33130e15,Namespace:kube-system,Attempt:0,} returns sandbox id \"54c449977658c080a797fd8661ce53ecf2f98f79f2c74676409161c5a858fac7\"" May 14 05:03:00.059256 containerd[1601]: time="2025-05-14T05:03:00.059233470Z" level=info msg="Container f92ab7ab5c1de2d08b18cc6a4c2c5d8f78b8a26de80db58f58e97f33cd654ec1: CDI devices from CRI Config.CDIDevices: []" May 14 05:03:00.059477 kubelet[2351]: E0514 05:03:00.059452 2351 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:00.059827 containerd[1601]: time="2025-05-14T05:03:00.059801666Z" level=info msg="connecting to shim 0125cd70a16c1cd7dc3e0d5935a202413016a1865c4ff592c66d8aa7f6bd85f2" address="unix:///run/containerd/s/003bedb507a677a3e3695ad348bc9ac8ecb17fc3fda90e262951031565508438" protocol=ttrpc version=3 May 14 05:03:00.061023 containerd[1601]: time="2025-05-14T05:03:00.061001757Z" level=info msg="CreateContainer within sandbox \"54c449977658c080a797fd8661ce53ecf2f98f79f2c74676409161c5a858fac7\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" May 14 05:03:00.073698 containerd[1601]: time="2025-05-14T05:03:00.073619404Z" level=info msg="CreateContainer within sandbox \"336f7da1e4d23b5bacd1626e7726a0a82298044ef4b3e1a1aa0c5e87c19b746c\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"f92ab7ab5c1de2d08b18cc6a4c2c5d8f78b8a26de80db58f58e97f33cd654ec1\"" May 14 05:03:00.074938 containerd[1601]: time="2025-05-14T05:03:00.073965784Z" level=info msg="StartContainer for \"f92ab7ab5c1de2d08b18cc6a4c2c5d8f78b8a26de80db58f58e97f33cd654ec1\"" May 14 05:03:00.075047 containerd[1601]: time="2025-05-14T05:03:00.075028978Z" level=info msg="connecting to shim f92ab7ab5c1de2d08b18cc6a4c2c5d8f78b8a26de80db58f58e97f33cd654ec1" address="unix:///run/containerd/s/0cd830b36f34f6eff93a2b3242fbaa40fd70c212b56906f1959cd7fec7f03ad1" protocol=ttrpc version=3 May 14 05:03:00.078376 containerd[1601]: time="2025-05-14T05:03:00.078357782Z" level=info msg="Container ec1bdf083adcc171fb189ad2605db1da39fca99b227a0cf4512e025a31da1a42: CDI devices from CRI Config.CDIDevices: []" May 14 05:03:00.080541 systemd[1]: Started cri-containerd-0125cd70a16c1cd7dc3e0d5935a202413016a1865c4ff592c66d8aa7f6bd85f2.scope - libcontainer container 0125cd70a16c1cd7dc3e0d5935a202413016a1865c4ff592c66d8aa7f6bd85f2. May 14 05:03:00.086012 containerd[1601]: time="2025-05-14T05:03:00.085978306Z" level=info msg="CreateContainer within sandbox \"54c449977658c080a797fd8661ce53ecf2f98f79f2c74676409161c5a858fac7\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"ec1bdf083adcc171fb189ad2605db1da39fca99b227a0cf4512e025a31da1a42\"" May 14 05:03:00.086600 containerd[1601]: time="2025-05-14T05:03:00.086572280Z" level=info msg="StartContainer for \"ec1bdf083adcc171fb189ad2605db1da39fca99b227a0cf4512e025a31da1a42\"" May 14 05:03:00.087610 containerd[1601]: time="2025-05-14T05:03:00.087581102Z" level=info msg="connecting to shim ec1bdf083adcc171fb189ad2605db1da39fca99b227a0cf4512e025a31da1a42" address="unix:///run/containerd/s/817e1c4c03a055044163dae1fc4053a876210f255d95d659010b22b4200723e2" protocol=ttrpc version=3 May 14 05:03:00.097681 systemd[1]: Started cri-containerd-f92ab7ab5c1de2d08b18cc6a4c2c5d8f78b8a26de80db58f58e97f33cd654ec1.scope - libcontainer container f92ab7ab5c1de2d08b18cc6a4c2c5d8f78b8a26de80db58f58e97f33cd654ec1. May 14 05:03:00.107667 systemd[1]: Started cri-containerd-ec1bdf083adcc171fb189ad2605db1da39fca99b227a0cf4512e025a31da1a42.scope - libcontainer container ec1bdf083adcc171fb189ad2605db1da39fca99b227a0cf4512e025a31da1a42. May 14 05:03:00.142549 containerd[1601]: time="2025-05-14T05:03:00.142422389Z" level=info msg="StartContainer for \"0125cd70a16c1cd7dc3e0d5935a202413016a1865c4ff592c66d8aa7f6bd85f2\" returns successfully" May 14 05:03:00.158083 containerd[1601]: time="2025-05-14T05:03:00.158038470Z" level=info msg="StartContainer for \"f92ab7ab5c1de2d08b18cc6a4c2c5d8f78b8a26de80db58f58e97f33cd654ec1\" returns successfully" May 14 05:03:00.161916 containerd[1601]: time="2025-05-14T05:03:00.161703034Z" level=info msg="StartContainer for \"ec1bdf083adcc171fb189ad2605db1da39fca99b227a0cf4512e025a31da1a42\" returns successfully" May 14 05:03:00.171883 kubelet[2351]: E0514 05:03:00.171838 2351 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.15:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.15:6443: connect: connection refused" interval="1.6s" May 14 05:03:00.667805 kubelet[2351]: I0514 05:03:00.667728 2351 kubelet_node_status.go:76] "Attempting to register node" node="localhost" May 14 05:03:00.796984 kubelet[2351]: E0514 05:03:00.796794 2351 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" May 14 05:03:00.797313 kubelet[2351]: E0514 05:03:00.797237 2351 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:00.811415 kubelet[2351]: E0514 05:03:00.810242 2351 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" May 14 05:03:00.811709 kubelet[2351]: E0514 05:03:00.811694 2351 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:00.815630 kubelet[2351]: E0514 05:03:00.815473 2351 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" May 14 05:03:00.815630 kubelet[2351]: E0514 05:03:00.815599 2351 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:01.104250 kubelet[2351]: I0514 05:03:01.104042 2351 kubelet_node_status.go:79] "Successfully registered node" node="localhost" May 14 05:03:01.104250 kubelet[2351]: E0514 05:03:01.104088 2351 kubelet_node_status.go:549] "Error updating node status, will retry" err="error getting node \"localhost\": node \"localhost\" not found" May 14 05:03:01.120727 kubelet[2351]: E0514 05:03:01.120684 2351 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" May 14 05:03:01.221655 kubelet[2351]: E0514 05:03:01.221611 2351 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" May 14 05:03:01.322738 kubelet[2351]: E0514 05:03:01.322687 2351 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" May 14 05:03:01.423514 kubelet[2351]: E0514 05:03:01.423384 2351 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" May 14 05:03:01.524017 kubelet[2351]: E0514 05:03:01.523973 2351 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" May 14 05:03:01.624797 kubelet[2351]: E0514 05:03:01.624719 2351 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" May 14 05:03:01.725541 kubelet[2351]: E0514 05:03:01.725374 2351 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" May 14 05:03:01.816584 kubelet[2351]: E0514 05:03:01.816544 2351 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" May 14 05:03:01.816686 kubelet[2351]: E0514 05:03:01.816642 2351 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:01.816686 kubelet[2351]: E0514 05:03:01.816656 2351 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" May 14 05:03:01.816764 kubelet[2351]: E0514 05:03:01.816746 2351 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:01.826080 kubelet[2351]: E0514 05:03:01.826058 2351 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" May 14 05:03:01.926722 kubelet[2351]: E0514 05:03:01.926679 2351 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" May 14 05:03:02.027558 kubelet[2351]: E0514 05:03:02.027419 2351 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" May 14 05:03:02.128094 kubelet[2351]: E0514 05:03:02.128046 2351 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" May 14 05:03:02.228855 kubelet[2351]: E0514 05:03:02.228788 2351 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" May 14 05:03:02.329722 kubelet[2351]: E0514 05:03:02.329568 2351 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" May 14 05:03:02.430173 kubelet[2351]: E0514 05:03:02.430111 2351 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" May 14 05:03:02.530847 kubelet[2351]: E0514 05:03:02.530790 2351 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" May 14 05:03:02.631485 kubelet[2351]: E0514 05:03:02.631360 2351 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" May 14 05:03:02.731891 kubelet[2351]: E0514 05:03:02.731859 2351 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" May 14 05:03:02.817899 kubelet[2351]: E0514 05:03:02.817857 2351 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" May 14 05:03:02.818008 kubelet[2351]: E0514 05:03:02.817986 2351 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:02.832197 kubelet[2351]: E0514 05:03:02.832164 2351 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" May 14 05:03:02.858455 systemd[1]: Reload requested from client PID 2625 ('systemctl') (unit session-7.scope)... May 14 05:03:02.858470 systemd[1]: Reloading... May 14 05:03:02.933041 kubelet[2351]: E0514 05:03:02.932272 2351 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" May 14 05:03:02.934451 zram_generator::config[2668]: No configuration found. May 14 05:03:03.022287 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 14 05:03:03.068829 kubelet[2351]: I0514 05:03:03.068797 2351 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" May 14 05:03:03.075015 kubelet[2351]: I0514 05:03:03.074970 2351 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" May 14 05:03:03.078256 kubelet[2351]: I0514 05:03:03.078236 2351 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" May 14 05:03:03.148231 systemd[1]: Reloading finished in 289 ms. May 14 05:03:03.183003 kubelet[2351]: I0514 05:03:03.182948 2351 dynamic_cafile_content.go:175] "Shutting down controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" May 14 05:03:03.183226 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... May 14 05:03:03.206931 systemd[1]: kubelet.service: Deactivated successfully. May 14 05:03:03.207313 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. May 14 05:03:03.207388 systemd[1]: kubelet.service: Consumed 724ms CPU time, 124.6M memory peak. May 14 05:03:03.209733 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 14 05:03:03.399705 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 14 05:03:03.412832 (kubelet)[2713]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS May 14 05:03:03.448238 kubelet[2713]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 14 05:03:03.448238 kubelet[2713]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. May 14 05:03:03.448238 kubelet[2713]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 14 05:03:03.448238 kubelet[2713]: I0514 05:03:03.448039 2713 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" May 14 05:03:03.454678 kubelet[2713]: I0514 05:03:03.454647 2713 server.go:520] "Kubelet version" kubeletVersion="v1.32.0" May 14 05:03:03.454678 kubelet[2713]: I0514 05:03:03.454669 2713 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" May 14 05:03:03.454933 kubelet[2713]: I0514 05:03:03.454916 2713 server.go:954] "Client rotation is on, will bootstrap in background" May 14 05:03:03.455953 kubelet[2713]: I0514 05:03:03.455934 2713 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". May 14 05:03:03.457912 kubelet[2713]: I0514 05:03:03.457887 2713 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" May 14 05:03:03.461274 kubelet[2713]: I0514 05:03:03.461245 2713 server.go:1444] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" May 14 05:03:03.466787 kubelet[2713]: I0514 05:03:03.466754 2713 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" May 14 05:03:03.467018 kubelet[2713]: I0514 05:03:03.466983 2713 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] May 14 05:03:03.467162 kubelet[2713]: I0514 05:03:03.467014 2713 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} May 14 05:03:03.467235 kubelet[2713]: I0514 05:03:03.467169 2713 topology_manager.go:138] "Creating topology manager with none policy" May 14 05:03:03.467235 kubelet[2713]: I0514 05:03:03.467177 2713 container_manager_linux.go:304] "Creating device plugin manager" May 14 05:03:03.467235 kubelet[2713]: I0514 05:03:03.467214 2713 state_mem.go:36] "Initialized new in-memory state store" May 14 05:03:03.467381 kubelet[2713]: I0514 05:03:03.467367 2713 kubelet.go:446] "Attempting to sync node with API server" May 14 05:03:03.467443 kubelet[2713]: I0514 05:03:03.467382 2713 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" May 14 05:03:03.467443 kubelet[2713]: I0514 05:03:03.467414 2713 kubelet.go:352] "Adding apiserver pod source" May 14 05:03:03.467443 kubelet[2713]: I0514 05:03:03.467424 2713 apiserver.go:42] "Waiting for node sync before watching apiserver pods" May 14 05:03:03.468129 kubelet[2713]: I0514 05:03:03.468047 2713 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v2.0.4" apiVersion="v1" May 14 05:03:03.468575 kubelet[2713]: I0514 05:03:03.468440 2713 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" May 14 05:03:03.468879 kubelet[2713]: I0514 05:03:03.468829 2713 watchdog_linux.go:99] "Systemd watchdog is not enabled" May 14 05:03:03.468879 kubelet[2713]: I0514 05:03:03.468855 2713 server.go:1287] "Started kubelet" May 14 05:03:03.469843 kubelet[2713]: I0514 05:03:03.469780 2713 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 May 14 05:03:03.470242 kubelet[2713]: I0514 05:03:03.469773 2713 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 May 14 05:03:03.470726 kubelet[2713]: I0514 05:03:03.470714 2713 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" May 14 05:03:03.472204 kubelet[2713]: I0514 05:03:03.472170 2713 server.go:490] "Adding debug handlers to kubelet server" May 14 05:03:03.473257 kubelet[2713]: I0514 05:03:03.473239 2713 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" May 14 05:03:03.473551 kubelet[2713]: I0514 05:03:03.473508 2713 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" May 14 05:03:03.475582 kubelet[2713]: E0514 05:03:03.475370 2713 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"localhost\" not found" May 14 05:03:03.476050 kubelet[2713]: I0514 05:03:03.476030 2713 volume_manager.go:297] "Starting Kubelet Volume Manager" May 14 05:03:03.477039 kubelet[2713]: I0514 05:03:03.477026 2713 desired_state_of_world_populator.go:149] "Desired state populator starts to run" May 14 05:03:03.477269 kubelet[2713]: I0514 05:03:03.477259 2713 reconciler.go:26] "Reconciler: start to sync state" May 14 05:03:03.478750 kubelet[2713]: I0514 05:03:03.478462 2713 factory.go:221] Registration of the systemd container factory successfully May 14 05:03:03.478750 kubelet[2713]: I0514 05:03:03.478557 2713 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory May 14 05:03:03.480035 kubelet[2713]: I0514 05:03:03.479654 2713 factory.go:221] Registration of the containerd container factory successfully May 14 05:03:03.480315 kubelet[2713]: E0514 05:03:03.480291 2713 kubelet.go:1561] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" May 14 05:03:03.487569 kubelet[2713]: I0514 05:03:03.487524 2713 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" May 14 05:03:03.488764 kubelet[2713]: I0514 05:03:03.488743 2713 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" May 14 05:03:03.489197 kubelet[2713]: I0514 05:03:03.488913 2713 status_manager.go:227] "Starting to sync pod status with apiserver" May 14 05:03:03.489197 kubelet[2713]: I0514 05:03:03.488940 2713 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." May 14 05:03:03.489197 kubelet[2713]: I0514 05:03:03.488949 2713 kubelet.go:2388] "Starting kubelet main sync loop" May 14 05:03:03.489197 kubelet[2713]: E0514 05:03:03.489003 2713 kubelet.go:2412] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" May 14 05:03:03.511606 kubelet[2713]: I0514 05:03:03.511575 2713 cpu_manager.go:221] "Starting CPU manager" policy="none" May 14 05:03:03.511606 kubelet[2713]: I0514 05:03:03.511592 2713 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" May 14 05:03:03.511606 kubelet[2713]: I0514 05:03:03.511609 2713 state_mem.go:36] "Initialized new in-memory state store" May 14 05:03:03.511788 kubelet[2713]: I0514 05:03:03.511769 2713 state_mem.go:88] "Updated default CPUSet" cpuSet="" May 14 05:03:03.511809 kubelet[2713]: I0514 05:03:03.511786 2713 state_mem.go:96] "Updated CPUSet assignments" assignments={} May 14 05:03:03.511809 kubelet[2713]: I0514 05:03:03.511803 2713 policy_none.go:49] "None policy: Start" May 14 05:03:03.511852 kubelet[2713]: I0514 05:03:03.511812 2713 memory_manager.go:186] "Starting memorymanager" policy="None" May 14 05:03:03.511852 kubelet[2713]: I0514 05:03:03.511821 2713 state_mem.go:35] "Initializing new in-memory state store" May 14 05:03:03.511953 kubelet[2713]: I0514 05:03:03.511912 2713 state_mem.go:75] "Updated machine memory state" May 14 05:03:03.515866 kubelet[2713]: I0514 05:03:03.515839 2713 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" May 14 05:03:03.516024 kubelet[2713]: I0514 05:03:03.516007 2713 eviction_manager.go:189] "Eviction manager: starting control loop" May 14 05:03:03.516069 kubelet[2713]: I0514 05:03:03.516023 2713 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" May 14 05:03:03.516364 kubelet[2713]: I0514 05:03:03.516190 2713 plugin_manager.go:118] "Starting Kubelet Plugin Manager" May 14 05:03:03.516888 kubelet[2713]: E0514 05:03:03.516832 2713 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" May 14 05:03:03.590058 kubelet[2713]: I0514 05:03:03.590018 2713 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" May 14 05:03:03.590193 kubelet[2713]: I0514 05:03:03.590166 2713 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" May 14 05:03:03.590312 kubelet[2713]: I0514 05:03:03.590284 2713 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" May 14 05:03:03.595622 kubelet[2713]: E0514 05:03:03.595573 2713 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-controller-manager-localhost\" already exists" pod="kube-system/kube-controller-manager-localhost" May 14 05:03:03.596021 kubelet[2713]: E0514 05:03:03.596003 2713 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost" May 14 05:03:03.596217 kubelet[2713]: E0514 05:03:03.596186 2713 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" May 14 05:03:03.620132 kubelet[2713]: I0514 05:03:03.620117 2713 kubelet_node_status.go:76] "Attempting to register node" node="localhost" May 14 05:03:03.625032 kubelet[2713]: I0514 05:03:03.625012 2713 kubelet_node_status.go:125] "Node was previously registered" node="localhost" May 14 05:03:03.625088 kubelet[2713]: I0514 05:03:03.625070 2713 kubelet_node_status.go:79] "Successfully registered node" node="localhost" May 14 05:03:03.677522 kubelet[2713]: I0514 05:03:03.677485 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/4ef2178d5f83051e6eb1068e93ce70df-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"4ef2178d5f83051e6eb1068e93ce70df\") " pod="kube-system/kube-apiserver-localhost" May 14 05:03:03.677522 kubelet[2713]: I0514 05:03:03.677519 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/5386fe11ed933ab82453de11903c7f47-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"5386fe11ed933ab82453de11903c7f47\") " pod="kube-system/kube-controller-manager-localhost" May 14 05:03:03.677619 kubelet[2713]: I0514 05:03:03.677539 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/5386fe11ed933ab82453de11903c7f47-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"5386fe11ed933ab82453de11903c7f47\") " pod="kube-system/kube-controller-manager-localhost" May 14 05:03:03.677619 kubelet[2713]: I0514 05:03:03.677552 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/5386fe11ed933ab82453de11903c7f47-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"5386fe11ed933ab82453de11903c7f47\") " pod="kube-system/kube-controller-manager-localhost" May 14 05:03:03.677619 kubelet[2713]: I0514 05:03:03.677570 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/5386fe11ed933ab82453de11903c7f47-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"5386fe11ed933ab82453de11903c7f47\") " pod="kube-system/kube-controller-manager-localhost" May 14 05:03:03.677619 kubelet[2713]: I0514 05:03:03.677585 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/5386fe11ed933ab82453de11903c7f47-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"5386fe11ed933ab82453de11903c7f47\") " pod="kube-system/kube-controller-manager-localhost" May 14 05:03:03.677619 kubelet[2713]: I0514 05:03:03.677602 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/4ef2178d5f83051e6eb1068e93ce70df-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"4ef2178d5f83051e6eb1068e93ce70df\") " pod="kube-system/kube-apiserver-localhost" May 14 05:03:03.677736 kubelet[2713]: I0514 05:03:03.677619 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/4ef2178d5f83051e6eb1068e93ce70df-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"4ef2178d5f83051e6eb1068e93ce70df\") " pod="kube-system/kube-apiserver-localhost" May 14 05:03:03.677736 kubelet[2713]: I0514 05:03:03.677660 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/2980a8ab51edc665be10a02e33130e15-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"2980a8ab51edc665be10a02e33130e15\") " pod="kube-system/kube-scheduler-localhost" May 14 05:03:03.896798 kubelet[2713]: E0514 05:03:03.896698 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:03.896798 kubelet[2713]: E0514 05:03:03.896750 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:03.896949 kubelet[2713]: E0514 05:03:03.896796 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:04.467986 kubelet[2713]: I0514 05:03:04.467941 2713 apiserver.go:52] "Watching apiserver" May 14 05:03:04.477681 kubelet[2713]: I0514 05:03:04.477624 2713 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" May 14 05:03:04.500798 kubelet[2713]: I0514 05:03:04.500346 2713 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" May 14 05:03:04.500798 kubelet[2713]: E0514 05:03:04.500471 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:04.500798 kubelet[2713]: E0514 05:03:04.500732 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:04.506692 kubelet[2713]: E0514 05:03:04.506644 2713 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" May 14 05:03:04.506842 kubelet[2713]: E0514 05:03:04.506819 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:04.521378 kubelet[2713]: I0514 05:03:04.521250 2713 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.5211881219999999 podStartE2EDuration="1.521188122s" podCreationTimestamp="2025-05-14 05:03:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-14 05:03:04.521151393 +0000 UTC m=+1.103638311" watchObservedRunningTime="2025-05-14 05:03:04.521188122 +0000 UTC m=+1.103675040" May 14 05:03:04.540453 kubelet[2713]: I0514 05:03:04.540365 2713 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=1.540347159 podStartE2EDuration="1.540347159s" podCreationTimestamp="2025-05-14 05:03:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-14 05:03:04.529024971 +0000 UTC m=+1.111511889" watchObservedRunningTime="2025-05-14 05:03:04.540347159 +0000 UTC m=+1.122834077" May 14 05:03:04.540785 kubelet[2713]: I0514 05:03:04.540490 2713 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.540485388 podStartE2EDuration="1.540485388s" podCreationTimestamp="2025-05-14 05:03:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-14 05:03:04.539929946 +0000 UTC m=+1.122416864" watchObservedRunningTime="2025-05-14 05:03:04.540485388 +0000 UTC m=+1.122972336" May 14 05:03:05.501858 kubelet[2713]: E0514 05:03:05.501816 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:05.502294 kubelet[2713]: E0514 05:03:05.502277 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:06.750291 kubelet[2713]: E0514 05:03:06.750238 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:07.829862 kubelet[2713]: I0514 05:03:07.829824 2713 kuberuntime_manager.go:1702] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" May 14 05:03:07.830346 containerd[1601]: time="2025-05-14T05:03:07.830124591Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." May 14 05:03:07.830681 kubelet[2713]: I0514 05:03:07.830425 2713 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" May 14 05:03:08.154869 sudo[1802]: pam_unix(sudo:session): session closed for user root May 14 05:03:08.156270 sshd[1801]: Connection closed by 10.0.0.1 port 53044 May 14 05:03:08.156664 sshd-session[1799]: pam_unix(sshd:session): session closed for user core May 14 05:03:08.160748 systemd[1]: sshd@6-10.0.0.15:22-10.0.0.1:53044.service: Deactivated successfully. May 14 05:03:08.163262 systemd[1]: session-7.scope: Deactivated successfully. May 14 05:03:08.163500 systemd[1]: session-7.scope: Consumed 5.161s CPU time, 219M memory peak. May 14 05:03:08.164689 systemd-logind[1587]: Session 7 logged out. Waiting for processes to exit. May 14 05:03:08.166100 systemd-logind[1587]: Removed session 7. May 14 05:03:08.678273 kubelet[2713]: I0514 05:03:08.678229 2713 status_manager.go:890] "Failed to get status for pod" podUID="67875d60-26fe-4673-8f34-bf5eabbbc358" pod="kube-system/kube-proxy-b9xjt" err="pods \"kube-proxy-b9xjt\" is forbidden: User \"system:node:localhost\" cannot get resource \"pods\" in API group \"\" in the namespace \"kube-system\": no relationship found between node 'localhost' and this object" May 14 05:03:08.683895 systemd[1]: Created slice kubepods-besteffort-pod67875d60_26fe_4673_8f34_bf5eabbbc358.slice - libcontainer container kubepods-besteffort-pod67875d60_26fe_4673_8f34_bf5eabbbc358.slice. May 14 05:03:08.710527 kubelet[2713]: I0514 05:03:08.710485 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/67875d60-26fe-4673-8f34-bf5eabbbc358-lib-modules\") pod \"kube-proxy-b9xjt\" (UID: \"67875d60-26fe-4673-8f34-bf5eabbbc358\") " pod="kube-system/kube-proxy-b9xjt" May 14 05:03:08.710527 kubelet[2713]: I0514 05:03:08.710523 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/67875d60-26fe-4673-8f34-bf5eabbbc358-kube-proxy\") pod \"kube-proxy-b9xjt\" (UID: \"67875d60-26fe-4673-8f34-bf5eabbbc358\") " pod="kube-system/kube-proxy-b9xjt" May 14 05:03:08.710527 kubelet[2713]: I0514 05:03:08.710537 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/67875d60-26fe-4673-8f34-bf5eabbbc358-xtables-lock\") pod \"kube-proxy-b9xjt\" (UID: \"67875d60-26fe-4673-8f34-bf5eabbbc358\") " pod="kube-system/kube-proxy-b9xjt" May 14 05:03:08.710527 kubelet[2713]: I0514 05:03:08.710552 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5zxw\" (UniqueName: \"kubernetes.io/projected/67875d60-26fe-4673-8f34-bf5eabbbc358-kube-api-access-z5zxw\") pod \"kube-proxy-b9xjt\" (UID: \"67875d60-26fe-4673-8f34-bf5eabbbc358\") " pod="kube-system/kube-proxy-b9xjt" May 14 05:03:08.815335 kubelet[2713]: E0514 05:03:08.815286 2713 projected.go:288] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found May 14 05:03:08.815335 kubelet[2713]: E0514 05:03:08.815319 2713 projected.go:194] Error preparing data for projected volume kube-api-access-z5zxw for pod kube-system/kube-proxy-b9xjt: configmap "kube-root-ca.crt" not found May 14 05:03:08.815524 kubelet[2713]: E0514 05:03:08.815370 2713 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/67875d60-26fe-4673-8f34-bf5eabbbc358-kube-api-access-z5zxw podName:67875d60-26fe-4673-8f34-bf5eabbbc358 nodeName:}" failed. No retries permitted until 2025-05-14 05:03:09.315349805 +0000 UTC m=+5.897836723 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-z5zxw" (UniqueName: "kubernetes.io/projected/67875d60-26fe-4673-8f34-bf5eabbbc358-kube-api-access-z5zxw") pod "kube-proxy-b9xjt" (UID: "67875d60-26fe-4673-8f34-bf5eabbbc358") : configmap "kube-root-ca.crt" not found May 14 05:03:08.950736 systemd[1]: Created slice kubepods-besteffort-podb001fb32_3030_40a8_a87b_46011221d288.slice - libcontainer container kubepods-besteffort-podb001fb32_3030_40a8_a87b_46011221d288.slice. May 14 05:03:09.013102 kubelet[2713]: I0514 05:03:09.013054 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/b001fb32-3030-40a8-a87b-46011221d288-var-lib-calico\") pod \"tigera-operator-789496d6f5-xvbbv\" (UID: \"b001fb32-3030-40a8-a87b-46011221d288\") " pod="tigera-operator/tigera-operator-789496d6f5-xvbbv" May 14 05:03:09.013505 kubelet[2713]: I0514 05:03:09.013117 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qdxq\" (UniqueName: \"kubernetes.io/projected/b001fb32-3030-40a8-a87b-46011221d288-kube-api-access-2qdxq\") pod \"tigera-operator-789496d6f5-xvbbv\" (UID: \"b001fb32-3030-40a8-a87b-46011221d288\") " pod="tigera-operator/tigera-operator-789496d6f5-xvbbv" May 14 05:03:09.254771 containerd[1601]: time="2025-05-14T05:03:09.254671061Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-789496d6f5-xvbbv,Uid:b001fb32-3030-40a8-a87b-46011221d288,Namespace:tigera-operator,Attempt:0,}" May 14 05:03:09.293932 containerd[1601]: time="2025-05-14T05:03:09.293886031Z" level=info msg="connecting to shim 29b7d9f242ee4893b95bba8114278ce03f96f76d3971a9f42d2c9cc7c9a64e68" address="unix:///run/containerd/s/5c79aa105a8fccbcf10f4c142c79298fc9550910780635547af66692fbe69b75" namespace=k8s.io protocol=ttrpc version=3 May 14 05:03:09.350543 systemd[1]: Started cri-containerd-29b7d9f242ee4893b95bba8114278ce03f96f76d3971a9f42d2c9cc7c9a64e68.scope - libcontainer container 29b7d9f242ee4893b95bba8114278ce03f96f76d3971a9f42d2c9cc7c9a64e68. May 14 05:03:09.388228 containerd[1601]: time="2025-05-14T05:03:09.388178646Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-789496d6f5-xvbbv,Uid:b001fb32-3030-40a8-a87b-46011221d288,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"29b7d9f242ee4893b95bba8114278ce03f96f76d3971a9f42d2c9cc7c9a64e68\"" May 14 05:03:09.389770 containerd[1601]: time="2025-05-14T05:03:09.389744921Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.36.7\"" May 14 05:03:09.595767 kubelet[2713]: E0514 05:03:09.595645 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:09.596411 containerd[1601]: time="2025-05-14T05:03:09.596359458Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-b9xjt,Uid:67875d60-26fe-4673-8f34-bf5eabbbc358,Namespace:kube-system,Attempt:0,}" May 14 05:03:09.621696 containerd[1601]: time="2025-05-14T05:03:09.621634386Z" level=info msg="connecting to shim 4a53562f47b3192adf3d92241b120445d464e122596a97ad5bef9f93b70dfe7a" address="unix:///run/containerd/s/1f0432c6c75bc9e2eae2fa8da143ba2f9c22da1978f7140ddb9c51c704b28c29" namespace=k8s.io protocol=ttrpc version=3 May 14 05:03:09.644578 systemd[1]: Started cri-containerd-4a53562f47b3192adf3d92241b120445d464e122596a97ad5bef9f93b70dfe7a.scope - libcontainer container 4a53562f47b3192adf3d92241b120445d464e122596a97ad5bef9f93b70dfe7a. May 14 05:03:09.668602 containerd[1601]: time="2025-05-14T05:03:09.668557058Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-b9xjt,Uid:67875d60-26fe-4673-8f34-bf5eabbbc358,Namespace:kube-system,Attempt:0,} returns sandbox id \"4a53562f47b3192adf3d92241b120445d464e122596a97ad5bef9f93b70dfe7a\"" May 14 05:03:09.669320 kubelet[2713]: E0514 05:03:09.669283 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:09.673158 containerd[1601]: time="2025-05-14T05:03:09.672645631Z" level=info msg="CreateContainer within sandbox \"4a53562f47b3192adf3d92241b120445d464e122596a97ad5bef9f93b70dfe7a\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" May 14 05:03:09.689855 containerd[1601]: time="2025-05-14T05:03:09.689805493Z" level=info msg="Container 1d8cffc7478ac1ef34e842146d92d3bc1fbd1c5cb36952c0f49be15529b2a047: CDI devices from CRI Config.CDIDevices: []" May 14 05:03:09.698067 containerd[1601]: time="2025-05-14T05:03:09.698014689Z" level=info msg="CreateContainer within sandbox \"4a53562f47b3192adf3d92241b120445d464e122596a97ad5bef9f93b70dfe7a\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"1d8cffc7478ac1ef34e842146d92d3bc1fbd1c5cb36952c0f49be15529b2a047\"" May 14 05:03:09.698744 containerd[1601]: time="2025-05-14T05:03:09.698653620Z" level=info msg="StartContainer for \"1d8cffc7478ac1ef34e842146d92d3bc1fbd1c5cb36952c0f49be15529b2a047\"" May 14 05:03:09.699970 containerd[1601]: time="2025-05-14T05:03:09.699941801Z" level=info msg="connecting to shim 1d8cffc7478ac1ef34e842146d92d3bc1fbd1c5cb36952c0f49be15529b2a047" address="unix:///run/containerd/s/1f0432c6c75bc9e2eae2fa8da143ba2f9c22da1978f7140ddb9c51c704b28c29" protocol=ttrpc version=3 May 14 05:03:09.730542 systemd[1]: Started cri-containerd-1d8cffc7478ac1ef34e842146d92d3bc1fbd1c5cb36952c0f49be15529b2a047.scope - libcontainer container 1d8cffc7478ac1ef34e842146d92d3bc1fbd1c5cb36952c0f49be15529b2a047. May 14 05:03:09.772178 containerd[1601]: time="2025-05-14T05:03:09.772138839Z" level=info msg="StartContainer for \"1d8cffc7478ac1ef34e842146d92d3bc1fbd1c5cb36952c0f49be15529b2a047\" returns successfully" May 14 05:03:10.513633 kubelet[2713]: E0514 05:03:10.513589 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:10.522994 kubelet[2713]: I0514 05:03:10.522930 2713 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-b9xjt" podStartSLOduration=2.522908778 podStartE2EDuration="2.522908778s" podCreationTimestamp="2025-05-14 05:03:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-14 05:03:10.522843683 +0000 UTC m=+7.105330601" watchObservedRunningTime="2025-05-14 05:03:10.522908778 +0000 UTC m=+7.105395696" May 14 05:03:11.116771 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4217651699.mount: Deactivated successfully. May 14 05:03:11.328822 kubelet[2713]: E0514 05:03:11.328776 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:11.515495 kubelet[2713]: E0514 05:03:11.514988 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:13.070623 kubelet[2713]: E0514 05:03:13.070591 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:13.518171 kubelet[2713]: E0514 05:03:13.518024 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:14.082126 containerd[1601]: time="2025-05-14T05:03:14.082067786Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.36.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:03:14.082829 containerd[1601]: time="2025-05-14T05:03:14.082799986Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.36.7: active requests=0, bytes read=22002662" May 14 05:03:14.084125 containerd[1601]: time="2025-05-14T05:03:14.084072949Z" level=info msg="ImageCreate event name:\"sha256:e9b19fa62f476f04e5840eb65a0f71b49c7b9f4ceede31675409ddc218bb5578\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:03:14.086084 containerd[1601]: time="2025-05-14T05:03:14.086044348Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:a4a44422d8f2a14e0aaea2031ccb5580f2bf68218c9db444450c1888743305e9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:03:14.086651 containerd[1601]: time="2025-05-14T05:03:14.086612124Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.36.7\" with image id \"sha256:e9b19fa62f476f04e5840eb65a0f71b49c7b9f4ceede31675409ddc218bb5578\", repo tag \"quay.io/tigera/operator:v1.36.7\", repo digest \"quay.io/tigera/operator@sha256:a4a44422d8f2a14e0aaea2031ccb5580f2bf68218c9db444450c1888743305e9\", size \"21998657\" in 4.696835211s" May 14 05:03:14.086695 containerd[1601]: time="2025-05-14T05:03:14.086651268Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.36.7\" returns image reference \"sha256:e9b19fa62f476f04e5840eb65a0f71b49c7b9f4ceede31675409ddc218bb5578\"" May 14 05:03:14.088410 containerd[1601]: time="2025-05-14T05:03:14.088366187Z" level=info msg="CreateContainer within sandbox \"29b7d9f242ee4893b95bba8114278ce03f96f76d3971a9f42d2c9cc7c9a64e68\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" May 14 05:03:14.096465 containerd[1601]: time="2025-05-14T05:03:14.096429563Z" level=info msg="Container 7eed832c0dd7ba864710e20cc2457715ba738f1dce79c674bb2e55af1b8967b3: CDI devices from CRI Config.CDIDevices: []" May 14 05:03:14.102423 containerd[1601]: time="2025-05-14T05:03:14.102381031Z" level=info msg="CreateContainer within sandbox \"29b7d9f242ee4893b95bba8114278ce03f96f76d3971a9f42d2c9cc7c9a64e68\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"7eed832c0dd7ba864710e20cc2457715ba738f1dce79c674bb2e55af1b8967b3\"" May 14 05:03:14.102793 containerd[1601]: time="2025-05-14T05:03:14.102769273Z" level=info msg="StartContainer for \"7eed832c0dd7ba864710e20cc2457715ba738f1dce79c674bb2e55af1b8967b3\"" May 14 05:03:14.103570 containerd[1601]: time="2025-05-14T05:03:14.103547862Z" level=info msg="connecting to shim 7eed832c0dd7ba864710e20cc2457715ba738f1dce79c674bb2e55af1b8967b3" address="unix:///run/containerd/s/5c79aa105a8fccbcf10f4c142c79298fc9550910780635547af66692fbe69b75" protocol=ttrpc version=3 May 14 05:03:14.125520 systemd[1]: Started cri-containerd-7eed832c0dd7ba864710e20cc2457715ba738f1dce79c674bb2e55af1b8967b3.scope - libcontainer container 7eed832c0dd7ba864710e20cc2457715ba738f1dce79c674bb2e55af1b8967b3. May 14 05:03:14.153371 containerd[1601]: time="2025-05-14T05:03:14.153316079Z" level=info msg="StartContainer for \"7eed832c0dd7ba864710e20cc2457715ba738f1dce79c674bb2e55af1b8967b3\" returns successfully" May 14 05:03:14.530643 kubelet[2713]: I0514 05:03:14.530570 2713 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-789496d6f5-xvbbv" podStartSLOduration=1.832635755 podStartE2EDuration="6.530552831s" podCreationTimestamp="2025-05-14 05:03:08 +0000 UTC" firstStartedPulling="2025-05-14 05:03:09.389256832 +0000 UTC m=+5.971743760" lastFinishedPulling="2025-05-14 05:03:14.087173918 +0000 UTC m=+10.669660836" observedRunningTime="2025-05-14 05:03:14.530505821 +0000 UTC m=+11.112992739" watchObservedRunningTime="2025-05-14 05:03:14.530552831 +0000 UTC m=+11.113039749" May 14 05:03:16.755480 kubelet[2713]: E0514 05:03:16.755371 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:17.086643 systemd[1]: Created slice kubepods-besteffort-poda9fdcd33_966f_4ee3_b1a0_c76dec250762.slice - libcontainer container kubepods-besteffort-poda9fdcd33_966f_4ee3_b1a0_c76dec250762.slice. May 14 05:03:17.143612 systemd[1]: Created slice kubepods-besteffort-pod03036838_da63_4d30_9dfe_9ac2d68f1daa.slice - libcontainer container kubepods-besteffort-pod03036838_da63_4d30_9dfe_9ac2d68f1daa.slice. May 14 05:03:17.162001 kubelet[2713]: I0514 05:03:17.161942 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/03036838-da63-4d30-9dfe-9ac2d68f1daa-xtables-lock\") pod \"calico-node-cchm7\" (UID: \"03036838-da63-4d30-9dfe-9ac2d68f1daa\") " pod="calico-system/calico-node-cchm7" May 14 05:03:17.162001 kubelet[2713]: I0514 05:03:17.161987 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/03036838-da63-4d30-9dfe-9ac2d68f1daa-node-certs\") pod \"calico-node-cchm7\" (UID: \"03036838-da63-4d30-9dfe-9ac2d68f1daa\") " pod="calico-system/calico-node-cchm7" May 14 05:03:17.162001 kubelet[2713]: I0514 05:03:17.162003 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/03036838-da63-4d30-9dfe-9ac2d68f1daa-cni-bin-dir\") pod \"calico-node-cchm7\" (UID: \"03036838-da63-4d30-9dfe-9ac2d68f1daa\") " pod="calico-system/calico-node-cchm7" May 14 05:03:17.162186 kubelet[2713]: I0514 05:03:17.162019 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/03036838-da63-4d30-9dfe-9ac2d68f1daa-cni-log-dir\") pod \"calico-node-cchm7\" (UID: \"03036838-da63-4d30-9dfe-9ac2d68f1daa\") " pod="calico-system/calico-node-cchm7" May 14 05:03:17.162186 kubelet[2713]: I0514 05:03:17.162035 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/a9fdcd33-966f-4ee3-b1a0-c76dec250762-typha-certs\") pod \"calico-typha-6bd746d57b-vsjwv\" (UID: \"a9fdcd33-966f-4ee3-b1a0-c76dec250762\") " pod="calico-system/calico-typha-6bd746d57b-vsjwv" May 14 05:03:17.162186 kubelet[2713]: I0514 05:03:17.162050 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a9fdcd33-966f-4ee3-b1a0-c76dec250762-tigera-ca-bundle\") pod \"calico-typha-6bd746d57b-vsjwv\" (UID: \"a9fdcd33-966f-4ee3-b1a0-c76dec250762\") " pod="calico-system/calico-typha-6bd746d57b-vsjwv" May 14 05:03:17.162186 kubelet[2713]: I0514 05:03:17.162067 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/03036838-da63-4d30-9dfe-9ac2d68f1daa-policysync\") pod \"calico-node-cchm7\" (UID: \"03036838-da63-4d30-9dfe-9ac2d68f1daa\") " pod="calico-system/calico-node-cchm7" May 14 05:03:17.162186 kubelet[2713]: I0514 05:03:17.162079 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/03036838-da63-4d30-9dfe-9ac2d68f1daa-cni-net-dir\") pod \"calico-node-cchm7\" (UID: \"03036838-da63-4d30-9dfe-9ac2d68f1daa\") " pod="calico-system/calico-node-cchm7" May 14 05:03:17.162304 kubelet[2713]: I0514 05:03:17.162107 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/03036838-da63-4d30-9dfe-9ac2d68f1daa-flexvol-driver-host\") pod \"calico-node-cchm7\" (UID: \"03036838-da63-4d30-9dfe-9ac2d68f1daa\") " pod="calico-system/calico-node-cchm7" May 14 05:03:17.162304 kubelet[2713]: I0514 05:03:17.162123 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7s8fp\" (UniqueName: \"kubernetes.io/projected/03036838-da63-4d30-9dfe-9ac2d68f1daa-kube-api-access-7s8fp\") pod \"calico-node-cchm7\" (UID: \"03036838-da63-4d30-9dfe-9ac2d68f1daa\") " pod="calico-system/calico-node-cchm7" May 14 05:03:17.162304 kubelet[2713]: I0514 05:03:17.162170 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/03036838-da63-4d30-9dfe-9ac2d68f1daa-tigera-ca-bundle\") pod \"calico-node-cchm7\" (UID: \"03036838-da63-4d30-9dfe-9ac2d68f1daa\") " pod="calico-system/calico-node-cchm7" May 14 05:03:17.162304 kubelet[2713]: I0514 05:03:17.162205 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qgch\" (UniqueName: \"kubernetes.io/projected/a9fdcd33-966f-4ee3-b1a0-c76dec250762-kube-api-access-8qgch\") pod \"calico-typha-6bd746d57b-vsjwv\" (UID: \"a9fdcd33-966f-4ee3-b1a0-c76dec250762\") " pod="calico-system/calico-typha-6bd746d57b-vsjwv" May 14 05:03:17.162304 kubelet[2713]: I0514 05:03:17.162226 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/03036838-da63-4d30-9dfe-9ac2d68f1daa-lib-modules\") pod \"calico-node-cchm7\" (UID: \"03036838-da63-4d30-9dfe-9ac2d68f1daa\") " pod="calico-system/calico-node-cchm7" May 14 05:03:17.162451 kubelet[2713]: I0514 05:03:17.162262 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/03036838-da63-4d30-9dfe-9ac2d68f1daa-var-run-calico\") pod \"calico-node-cchm7\" (UID: \"03036838-da63-4d30-9dfe-9ac2d68f1daa\") " pod="calico-system/calico-node-cchm7" May 14 05:03:17.162451 kubelet[2713]: I0514 05:03:17.162279 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/03036838-da63-4d30-9dfe-9ac2d68f1daa-var-lib-calico\") pod \"calico-node-cchm7\" (UID: \"03036838-da63-4d30-9dfe-9ac2d68f1daa\") " pod="calico-system/calico-node-cchm7" May 14 05:03:17.248163 kubelet[2713]: E0514 05:03:17.248108 2713 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-vwdck" podUID="f37146d1-125d-4167-a0d9-8ac3149fed78" May 14 05:03:17.263168 kubelet[2713]: I0514 05:03:17.263121 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/f37146d1-125d-4167-a0d9-8ac3149fed78-varrun\") pod \"csi-node-driver-vwdck\" (UID: \"f37146d1-125d-4167-a0d9-8ac3149fed78\") " pod="calico-system/csi-node-driver-vwdck" May 14 05:03:17.263168 kubelet[2713]: I0514 05:03:17.263165 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/f37146d1-125d-4167-a0d9-8ac3149fed78-registration-dir\") pod \"csi-node-driver-vwdck\" (UID: \"f37146d1-125d-4167-a0d9-8ac3149fed78\") " pod="calico-system/csi-node-driver-vwdck" May 14 05:03:17.263385 kubelet[2713]: I0514 05:03:17.263245 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/f37146d1-125d-4167-a0d9-8ac3149fed78-socket-dir\") pod \"csi-node-driver-vwdck\" (UID: \"f37146d1-125d-4167-a0d9-8ac3149fed78\") " pod="calico-system/csi-node-driver-vwdck" May 14 05:03:17.263385 kubelet[2713]: I0514 05:03:17.263352 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f37146d1-125d-4167-a0d9-8ac3149fed78-kubelet-dir\") pod \"csi-node-driver-vwdck\" (UID: \"f37146d1-125d-4167-a0d9-8ac3149fed78\") " pod="calico-system/csi-node-driver-vwdck" May 14 05:03:17.263451 kubelet[2713]: I0514 05:03:17.263424 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8tx4\" (UniqueName: \"kubernetes.io/projected/f37146d1-125d-4167-a0d9-8ac3149fed78-kube-api-access-f8tx4\") pod \"csi-node-driver-vwdck\" (UID: \"f37146d1-125d-4167-a0d9-8ac3149fed78\") " pod="calico-system/csi-node-driver-vwdck" May 14 05:03:17.268309 kubelet[2713]: E0514 05:03:17.268276 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:17.268309 kubelet[2713]: W0514 05:03:17.268300 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:17.268540 kubelet[2713]: E0514 05:03:17.268324 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:17.278421 kubelet[2713]: E0514 05:03:17.277332 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:17.278421 kubelet[2713]: W0514 05:03:17.278324 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:17.278421 kubelet[2713]: E0514 05:03:17.278341 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:17.281419 kubelet[2713]: E0514 05:03:17.280659 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:17.281419 kubelet[2713]: W0514 05:03:17.280685 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:17.281419 kubelet[2713]: E0514 05:03:17.280705 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:17.283432 kubelet[2713]: E0514 05:03:17.282679 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:17.283577 kubelet[2713]: W0514 05:03:17.283562 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:17.283672 kubelet[2713]: E0514 05:03:17.283639 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:17.284522 kubelet[2713]: E0514 05:03:17.284473 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:17.284522 kubelet[2713]: W0514 05:03:17.284490 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:17.284522 kubelet[2713]: E0514 05:03:17.284502 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:17.364165 kubelet[2713]: E0514 05:03:17.364050 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:17.364165 kubelet[2713]: W0514 05:03:17.364070 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:17.364165 kubelet[2713]: E0514 05:03:17.364098 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:17.364870 kubelet[2713]: E0514 05:03:17.364373 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:17.364870 kubelet[2713]: W0514 05:03:17.364385 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:17.364870 kubelet[2713]: E0514 05:03:17.364421 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:17.364870 kubelet[2713]: E0514 05:03:17.364768 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:17.364870 kubelet[2713]: W0514 05:03:17.364794 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:17.364870 kubelet[2713]: E0514 05:03:17.364824 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:17.365772 kubelet[2713]: E0514 05:03:17.365135 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:17.365772 kubelet[2713]: W0514 05:03:17.365148 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:17.365772 kubelet[2713]: E0514 05:03:17.365163 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:17.365772 kubelet[2713]: E0514 05:03:17.365484 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:17.365772 kubelet[2713]: W0514 05:03:17.365492 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:17.365772 kubelet[2713]: E0514 05:03:17.365511 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:17.365970 kubelet[2713]: E0514 05:03:17.365911 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:17.365970 kubelet[2713]: W0514 05:03:17.365938 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:17.366018 kubelet[2713]: E0514 05:03:17.365969 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:17.366195 kubelet[2713]: E0514 05:03:17.366166 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:17.366195 kubelet[2713]: W0514 05:03:17.366181 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:17.366262 kubelet[2713]: E0514 05:03:17.366247 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:17.366414 kubelet[2713]: E0514 05:03:17.366376 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:17.366414 kubelet[2713]: W0514 05:03:17.366390 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:17.366509 kubelet[2713]: E0514 05:03:17.366491 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:17.366619 kubelet[2713]: E0514 05:03:17.366603 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:17.366619 kubelet[2713]: W0514 05:03:17.366616 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:17.366679 kubelet[2713]: E0514 05:03:17.366657 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:17.366842 kubelet[2713]: E0514 05:03:17.366826 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:17.366842 kubelet[2713]: W0514 05:03:17.366840 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:17.366918 kubelet[2713]: E0514 05:03:17.366900 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:17.367182 kubelet[2713]: E0514 05:03:17.367058 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:17.367182 kubelet[2713]: W0514 05:03:17.367075 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:17.367182 kubelet[2713]: E0514 05:03:17.367104 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:17.367695 kubelet[2713]: E0514 05:03:17.367473 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:17.367695 kubelet[2713]: W0514 05:03:17.367488 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:17.367695 kubelet[2713]: E0514 05:03:17.367506 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:17.367820 kubelet[2713]: E0514 05:03:17.367712 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:17.367820 kubelet[2713]: W0514 05:03:17.367722 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:17.367820 kubelet[2713]: E0514 05:03:17.367738 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:17.367905 kubelet[2713]: E0514 05:03:17.367900 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:17.367948 kubelet[2713]: W0514 05:03:17.367907 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:17.367986 kubelet[2713]: E0514 05:03:17.367933 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:17.368083 kubelet[2713]: E0514 05:03:17.368064 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:17.368083 kubelet[2713]: W0514 05:03:17.368075 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:17.368159 kubelet[2713]: E0514 05:03:17.368114 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:17.368276 kubelet[2713]: E0514 05:03:17.368261 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:17.368276 kubelet[2713]: W0514 05:03:17.368270 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:17.368346 kubelet[2713]: E0514 05:03:17.368294 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:17.368496 kubelet[2713]: E0514 05:03:17.368474 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:17.368496 kubelet[2713]: W0514 05:03:17.368484 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:17.368555 kubelet[2713]: E0514 05:03:17.368518 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:17.368699 kubelet[2713]: E0514 05:03:17.368677 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:17.368699 kubelet[2713]: W0514 05:03:17.368689 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:17.368699 kubelet[2713]: E0514 05:03:17.368700 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:17.368872 kubelet[2713]: E0514 05:03:17.368858 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:17.368872 kubelet[2713]: W0514 05:03:17.368868 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:17.368970 kubelet[2713]: E0514 05:03:17.368881 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:17.369057 kubelet[2713]: E0514 05:03:17.369024 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:17.369057 kubelet[2713]: W0514 05:03:17.369039 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:17.369057 kubelet[2713]: E0514 05:03:17.369054 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:17.369255 kubelet[2713]: E0514 05:03:17.369235 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:17.369255 kubelet[2713]: W0514 05:03:17.369248 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:17.369336 kubelet[2713]: E0514 05:03:17.369265 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:17.369759 kubelet[2713]: E0514 05:03:17.369741 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:17.369759 kubelet[2713]: W0514 05:03:17.369757 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:17.369831 kubelet[2713]: E0514 05:03:17.369775 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:17.370127 kubelet[2713]: E0514 05:03:17.370109 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:17.370127 kubelet[2713]: W0514 05:03:17.370126 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:17.370248 kubelet[2713]: E0514 05:03:17.370224 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:17.370517 kubelet[2713]: E0514 05:03:17.370472 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:17.370517 kubelet[2713]: W0514 05:03:17.370486 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:17.370517 kubelet[2713]: E0514 05:03:17.370499 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:17.370757 kubelet[2713]: E0514 05:03:17.370739 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:17.370757 kubelet[2713]: W0514 05:03:17.370754 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:17.370832 kubelet[2713]: E0514 05:03:17.370767 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:17.376903 kubelet[2713]: E0514 05:03:17.376851 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:17.376903 kubelet[2713]: W0514 05:03:17.376867 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:17.376903 kubelet[2713]: E0514 05:03:17.376879 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:17.390112 kubelet[2713]: E0514 05:03:17.390088 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:17.390602 containerd[1601]: time="2025-05-14T05:03:17.390565276Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-6bd746d57b-vsjwv,Uid:a9fdcd33-966f-4ee3-b1a0-c76dec250762,Namespace:calico-system,Attempt:0,}" May 14 05:03:17.432809 containerd[1601]: time="2025-05-14T05:03:17.432753402Z" level=info msg="connecting to shim 0aa5745c7ff598b71d6b61b29399b9021705eef79277a7da9d70cd86b0bd0c0f" address="unix:///run/containerd/s/a7e214628a3c8cedf4cf8f34c1cede4e7709bca18c5b86b28634bc47a66815ce" namespace=k8s.io protocol=ttrpc version=3 May 14 05:03:17.446480 kubelet[2713]: E0514 05:03:17.446444 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:17.447086 containerd[1601]: time="2025-05-14T05:03:17.447026660Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-cchm7,Uid:03036838-da63-4d30-9dfe-9ac2d68f1daa,Namespace:calico-system,Attempt:0,}" May 14 05:03:17.458594 systemd[1]: Started cri-containerd-0aa5745c7ff598b71d6b61b29399b9021705eef79277a7da9d70cd86b0bd0c0f.scope - libcontainer container 0aa5745c7ff598b71d6b61b29399b9021705eef79277a7da9d70cd86b0bd0c0f. May 14 05:03:17.472500 containerd[1601]: time="2025-05-14T05:03:17.472366294Z" level=info msg="connecting to shim 3601474911d55aef52f0a90124d287de4236d2e328eede33e262631c9388e41a" address="unix:///run/containerd/s/77e1234b6ea2bee922c8f1119f0fbf5107d4fd1008dd0105c7c7ebe949c23c89" namespace=k8s.io protocol=ttrpc version=3 May 14 05:03:17.500661 systemd[1]: Started cri-containerd-3601474911d55aef52f0a90124d287de4236d2e328eede33e262631c9388e41a.scope - libcontainer container 3601474911d55aef52f0a90124d287de4236d2e328eede33e262631c9388e41a. May 14 05:03:17.514263 containerd[1601]: time="2025-05-14T05:03:17.514154319Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-6bd746d57b-vsjwv,Uid:a9fdcd33-966f-4ee3-b1a0-c76dec250762,Namespace:calico-system,Attempt:0,} returns sandbox id \"0aa5745c7ff598b71d6b61b29399b9021705eef79277a7da9d70cd86b0bd0c0f\"" May 14 05:03:17.516663 kubelet[2713]: E0514 05:03:17.516620 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:17.518502 containerd[1601]: time="2025-05-14T05:03:17.518463940Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.29.3\"" May 14 05:03:17.528734 kubelet[2713]: E0514 05:03:17.528488 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:17.536762 containerd[1601]: time="2025-05-14T05:03:17.536704907Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-cchm7,Uid:03036838-da63-4d30-9dfe-9ac2d68f1daa,Namespace:calico-system,Attempt:0,} returns sandbox id \"3601474911d55aef52f0a90124d287de4236d2e328eede33e262631c9388e41a\"" May 14 05:03:17.537700 kubelet[2713]: E0514 05:03:17.537673 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:17.559238 kubelet[2713]: E0514 05:03:17.559212 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:17.559238 kubelet[2713]: W0514 05:03:17.559232 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:17.559372 kubelet[2713]: E0514 05:03:17.559253 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:17.559513 kubelet[2713]: E0514 05:03:17.559485 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:17.559513 kubelet[2713]: W0514 05:03:17.559507 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:17.559573 kubelet[2713]: E0514 05:03:17.559516 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:17.559728 kubelet[2713]: E0514 05:03:17.559702 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:17.559728 kubelet[2713]: W0514 05:03:17.559720 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:17.559774 kubelet[2713]: E0514 05:03:17.559733 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:17.560000 kubelet[2713]: E0514 05:03:17.559981 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:17.560000 kubelet[2713]: W0514 05:03:17.559994 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:17.560000 kubelet[2713]: E0514 05:03:17.560002 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:17.560194 kubelet[2713]: E0514 05:03:17.560176 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:17.560194 kubelet[2713]: W0514 05:03:17.560188 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:17.560240 kubelet[2713]: E0514 05:03:17.560197 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:19.493809 kubelet[2713]: E0514 05:03:19.493755 2713 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-vwdck" podUID="f37146d1-125d-4167-a0d9-8ac3149fed78" May 14 05:03:20.311520 update_engine[1589]: I20250514 05:03:20.311450 1589 update_attempter.cc:509] Updating boot flags... May 14 05:03:20.320064 containerd[1601]: time="2025-05-14T05:03:20.320015993Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:03:20.321097 containerd[1601]: time="2025-05-14T05:03:20.320874023Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.29.3: active requests=0, bytes read=30426870" May 14 05:03:20.322045 containerd[1601]: time="2025-05-14T05:03:20.322008629Z" level=info msg="ImageCreate event name:\"sha256:bde24a3cb8851b59372b76b3ad78f8028d1a915ffed82c6cc6256f34e500bd3d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:03:20.324116 containerd[1601]: time="2025-05-14T05:03:20.324067651Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:f5516aa6a78f00931d2625f3012dcf2c69d141ce41483b8d59c6ec6330a18620\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:03:20.324364 containerd[1601]: time="2025-05-14T05:03:20.324343195Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.29.3\" with image id \"sha256:bde24a3cb8851b59372b76b3ad78f8028d1a915ffed82c6cc6256f34e500bd3d\", repo tag \"ghcr.io/flatcar/calico/typha:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:f5516aa6a78f00931d2625f3012dcf2c69d141ce41483b8d59c6ec6330a18620\", size \"31919484\" in 2.805783975s" May 14 05:03:20.324364 containerd[1601]: time="2025-05-14T05:03:20.324370137Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.29.3\" returns image reference \"sha256:bde24a3cb8851b59372b76b3ad78f8028d1a915ffed82c6cc6256f34e500bd3d\"" May 14 05:03:20.326010 containerd[1601]: time="2025-05-14T05:03:20.325650880Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\"" May 14 05:03:20.337534 containerd[1601]: time="2025-05-14T05:03:20.337473907Z" level=info msg="CreateContainer within sandbox \"0aa5745c7ff598b71d6b61b29399b9021705eef79277a7da9d70cd86b0bd0c0f\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" May 14 05:03:20.349650 containerd[1601]: time="2025-05-14T05:03:20.348843011Z" level=info msg="Container 33a3378037b5234006dd64e5abc940cacd8478f36904965747e69667be9442d3: CDI devices from CRI Config.CDIDevices: []" May 14 05:03:20.359918 containerd[1601]: time="2025-05-14T05:03:20.359872588Z" level=info msg="CreateContainer within sandbox \"0aa5745c7ff598b71d6b61b29399b9021705eef79277a7da9d70cd86b0bd0c0f\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"33a3378037b5234006dd64e5abc940cacd8478f36904965747e69667be9442d3\"" May 14 05:03:20.362763 containerd[1601]: time="2025-05-14T05:03:20.362737934Z" level=info msg="StartContainer for \"33a3378037b5234006dd64e5abc940cacd8478f36904965747e69667be9442d3\"" May 14 05:03:20.365666 containerd[1601]: time="2025-05-14T05:03:20.365617585Z" level=info msg="connecting to shim 33a3378037b5234006dd64e5abc940cacd8478f36904965747e69667be9442d3" address="unix:///run/containerd/s/a7e214628a3c8cedf4cf8f34c1cede4e7709bca18c5b86b28634bc47a66815ce" protocol=ttrpc version=3 May 14 05:03:20.442671 systemd[1]: Started cri-containerd-33a3378037b5234006dd64e5abc940cacd8478f36904965747e69667be9442d3.scope - libcontainer container 33a3378037b5234006dd64e5abc940cacd8478f36904965747e69667be9442d3. May 14 05:03:20.569512 containerd[1601]: time="2025-05-14T05:03:20.567605547Z" level=info msg="StartContainer for \"33a3378037b5234006dd64e5abc940cacd8478f36904965747e69667be9442d3\" returns successfully" May 14 05:03:21.490132 kubelet[2713]: E0514 05:03:21.490072 2713 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-vwdck" podUID="f37146d1-125d-4167-a0d9-8ac3149fed78" May 14 05:03:21.538832 kubelet[2713]: E0514 05:03:21.538801 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:21.547361 kubelet[2713]: I0514 05:03:21.547290 2713 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-6bd746d57b-vsjwv" podStartSLOduration=1.739833395 podStartE2EDuration="4.547272527s" podCreationTimestamp="2025-05-14 05:03:17 +0000 UTC" firstStartedPulling="2025-05-14 05:03:17.518064669 +0000 UTC m=+14.100551587" lastFinishedPulling="2025-05-14 05:03:20.325503801 +0000 UTC m=+16.907990719" observedRunningTime="2025-05-14 05:03:21.546566606 +0000 UTC m=+18.129053524" watchObservedRunningTime="2025-05-14 05:03:21.547272527 +0000 UTC m=+18.129759435" May 14 05:03:21.582496 kubelet[2713]: E0514 05:03:21.582456 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:21.582496 kubelet[2713]: W0514 05:03:21.582478 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:21.582496 kubelet[2713]: E0514 05:03:21.582498 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:21.582689 kubelet[2713]: E0514 05:03:21.582683 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:21.582714 kubelet[2713]: W0514 05:03:21.582690 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:21.582714 kubelet[2713]: E0514 05:03:21.582698 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:21.582865 kubelet[2713]: E0514 05:03:21.582844 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:21.582865 kubelet[2713]: W0514 05:03:21.582854 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:21.582865 kubelet[2713]: E0514 05:03:21.582862 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:21.583056 kubelet[2713]: E0514 05:03:21.583038 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:21.583056 kubelet[2713]: W0514 05:03:21.583048 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:21.583056 kubelet[2713]: E0514 05:03:21.583055 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:21.583240 kubelet[2713]: E0514 05:03:21.583219 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:21.583240 kubelet[2713]: W0514 05:03:21.583234 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:21.583292 kubelet[2713]: E0514 05:03:21.583242 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:21.583415 kubelet[2713]: E0514 05:03:21.583390 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:21.583415 kubelet[2713]: W0514 05:03:21.583412 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:21.583470 kubelet[2713]: E0514 05:03:21.583419 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:21.583567 kubelet[2713]: E0514 05:03:21.583554 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:21.583567 kubelet[2713]: W0514 05:03:21.583563 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:21.583610 kubelet[2713]: E0514 05:03:21.583570 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:21.583720 kubelet[2713]: E0514 05:03:21.583706 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:21.583720 kubelet[2713]: W0514 05:03:21.583715 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:21.583770 kubelet[2713]: E0514 05:03:21.583722 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:21.583878 kubelet[2713]: E0514 05:03:21.583865 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:21.583878 kubelet[2713]: W0514 05:03:21.583874 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:21.583921 kubelet[2713]: E0514 05:03:21.583882 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:21.584026 kubelet[2713]: E0514 05:03:21.584013 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:21.584026 kubelet[2713]: W0514 05:03:21.584022 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:21.584076 kubelet[2713]: E0514 05:03:21.584029 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:21.584178 kubelet[2713]: E0514 05:03:21.584165 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:21.584178 kubelet[2713]: W0514 05:03:21.584174 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:21.584222 kubelet[2713]: E0514 05:03:21.584181 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:21.584339 kubelet[2713]: E0514 05:03:21.584326 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:21.584339 kubelet[2713]: W0514 05:03:21.584335 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:21.584410 kubelet[2713]: E0514 05:03:21.584341 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:21.584547 kubelet[2713]: E0514 05:03:21.584525 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:21.584547 kubelet[2713]: W0514 05:03:21.584536 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:21.584592 kubelet[2713]: E0514 05:03:21.584555 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:21.584734 kubelet[2713]: E0514 05:03:21.584720 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:21.584734 kubelet[2713]: W0514 05:03:21.584730 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:21.584784 kubelet[2713]: E0514 05:03:21.584739 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:21.584921 kubelet[2713]: E0514 05:03:21.584908 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:21.584921 kubelet[2713]: W0514 05:03:21.584917 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:21.584967 kubelet[2713]: E0514 05:03:21.584934 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:21.597354 kubelet[2713]: E0514 05:03:21.597319 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:21.597354 kubelet[2713]: W0514 05:03:21.597341 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:21.597354 kubelet[2713]: E0514 05:03:21.597360 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:21.597593 kubelet[2713]: E0514 05:03:21.597569 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:21.597593 kubelet[2713]: W0514 05:03:21.597580 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:21.597593 kubelet[2713]: E0514 05:03:21.597595 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:21.597787 kubelet[2713]: E0514 05:03:21.597764 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:21.597787 kubelet[2713]: W0514 05:03:21.597779 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:21.597839 kubelet[2713]: E0514 05:03:21.597794 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:21.598005 kubelet[2713]: E0514 05:03:21.597985 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:21.598005 kubelet[2713]: W0514 05:03:21.598002 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:21.598068 kubelet[2713]: E0514 05:03:21.598017 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:21.598196 kubelet[2713]: E0514 05:03:21.598181 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:21.598196 kubelet[2713]: W0514 05:03:21.598192 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:21.598254 kubelet[2713]: E0514 05:03:21.598204 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:21.598382 kubelet[2713]: E0514 05:03:21.598368 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:21.598382 kubelet[2713]: W0514 05:03:21.598379 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:21.598452 kubelet[2713]: E0514 05:03:21.598392 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:21.598566 kubelet[2713]: E0514 05:03:21.598551 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:21.598566 kubelet[2713]: W0514 05:03:21.598562 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:21.598616 kubelet[2713]: E0514 05:03:21.598574 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:21.598762 kubelet[2713]: E0514 05:03:21.598747 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:21.598762 kubelet[2713]: W0514 05:03:21.598758 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:21.598817 kubelet[2713]: E0514 05:03:21.598771 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:21.598930 kubelet[2713]: E0514 05:03:21.598918 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:21.598930 kubelet[2713]: W0514 05:03:21.598927 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:21.598982 kubelet[2713]: E0514 05:03:21.598937 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:21.599100 kubelet[2713]: E0514 05:03:21.599085 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:21.599100 kubelet[2713]: W0514 05:03:21.599097 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:21.599149 kubelet[2713]: E0514 05:03:21.599109 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:21.599281 kubelet[2713]: E0514 05:03:21.599267 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:21.599281 kubelet[2713]: W0514 05:03:21.599276 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:21.599346 kubelet[2713]: E0514 05:03:21.599288 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:21.599515 kubelet[2713]: E0514 05:03:21.599490 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:21.599515 kubelet[2713]: W0514 05:03:21.599500 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:21.599515 kubelet[2713]: E0514 05:03:21.599513 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:21.599748 kubelet[2713]: E0514 05:03:21.599732 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:21.599748 kubelet[2713]: W0514 05:03:21.599745 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:21.599802 kubelet[2713]: E0514 05:03:21.599758 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:21.599944 kubelet[2713]: E0514 05:03:21.599930 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:21.599944 kubelet[2713]: W0514 05:03:21.599940 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:21.600000 kubelet[2713]: E0514 05:03:21.599952 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:21.600225 kubelet[2713]: E0514 05:03:21.600205 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:21.600225 kubelet[2713]: W0514 05:03:21.600217 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:21.600279 kubelet[2713]: E0514 05:03:21.600230 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:21.600504 kubelet[2713]: E0514 05:03:21.600480 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:21.600504 kubelet[2713]: W0514 05:03:21.600491 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:21.600560 kubelet[2713]: E0514 05:03:21.600506 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:21.600680 kubelet[2713]: E0514 05:03:21.600666 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:21.600680 kubelet[2713]: W0514 05:03:21.600676 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:21.600733 kubelet[2713]: E0514 05:03:21.600684 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:21.600950 kubelet[2713]: E0514 05:03:21.600935 2713 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input May 14 05:03:21.600950 kubelet[2713]: W0514 05:03:21.600947 2713 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" May 14 05:03:21.601002 kubelet[2713]: E0514 05:03:21.600955 2713 plugins.go:695] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" May 14 05:03:21.993191 containerd[1601]: time="2025-05-14T05:03:21.993140871Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:03:21.994275 containerd[1601]: time="2025-05-14T05:03:21.994252260Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3: active requests=0, bytes read=5366937" May 14 05:03:21.995348 containerd[1601]: time="2025-05-14T05:03:21.995321801Z" level=info msg="ImageCreate event name:\"sha256:0ceddb3add2e9955cbb604f666245e259f30b1d6683c428f8748359e83d238a5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:03:21.997257 containerd[1601]: time="2025-05-14T05:03:21.997211820Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:eeaa2bb4f9b1aa61adde43ce6dea95eee89291f96963548e108d9a2dfbc5edd1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:03:21.997645 containerd[1601]: time="2025-05-14T05:03:21.997605307Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\" with image id \"sha256:0ceddb3add2e9955cbb604f666245e259f30b1d6683c428f8748359e83d238a5\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:eeaa2bb4f9b1aa61adde43ce6dea95eee89291f96963548e108d9a2dfbc5edd1\", size \"6859519\" in 1.671927105s" May 14 05:03:21.997645 containerd[1601]: time="2025-05-14T05:03:21.997634091Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\" returns image reference \"sha256:0ceddb3add2e9955cbb604f666245e259f30b1d6683c428f8748359e83d238a5\"" May 14 05:03:21.999294 containerd[1601]: time="2025-05-14T05:03:21.999267854Z" level=info msg="CreateContainer within sandbox \"3601474911d55aef52f0a90124d287de4236d2e328eede33e262631c9388e41a\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" May 14 05:03:22.009654 containerd[1601]: time="2025-05-14T05:03:22.009620678Z" level=info msg="Container 4f69e6cb5691bc2cac3c350e267cfc08ddea5b7a1a9e4449a1689e738681e9b3: CDI devices from CRI Config.CDIDevices: []" May 14 05:03:22.017537 containerd[1601]: time="2025-05-14T05:03:22.017497723Z" level=info msg="CreateContainer within sandbox \"3601474911d55aef52f0a90124d287de4236d2e328eede33e262631c9388e41a\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"4f69e6cb5691bc2cac3c350e267cfc08ddea5b7a1a9e4449a1689e738681e9b3\"" May 14 05:03:22.017953 containerd[1601]: time="2025-05-14T05:03:22.017928942Z" level=info msg="StartContainer for \"4f69e6cb5691bc2cac3c350e267cfc08ddea5b7a1a9e4449a1689e738681e9b3\"" May 14 05:03:22.019157 containerd[1601]: time="2025-05-14T05:03:22.019120121Z" level=info msg="connecting to shim 4f69e6cb5691bc2cac3c350e267cfc08ddea5b7a1a9e4449a1689e738681e9b3" address="unix:///run/containerd/s/77e1234b6ea2bee922c8f1119f0fbf5107d4fd1008dd0105c7c7ebe949c23c89" protocol=ttrpc version=3 May 14 05:03:22.044536 systemd[1]: Started cri-containerd-4f69e6cb5691bc2cac3c350e267cfc08ddea5b7a1a9e4449a1689e738681e9b3.scope - libcontainer container 4f69e6cb5691bc2cac3c350e267cfc08ddea5b7a1a9e4449a1689e738681e9b3. May 14 05:03:22.086130 containerd[1601]: time="2025-05-14T05:03:22.086061987Z" level=info msg="StartContainer for \"4f69e6cb5691bc2cac3c350e267cfc08ddea5b7a1a9e4449a1689e738681e9b3\" returns successfully" May 14 05:03:22.097913 systemd[1]: cri-containerd-4f69e6cb5691bc2cac3c350e267cfc08ddea5b7a1a9e4449a1689e738681e9b3.scope: Deactivated successfully. May 14 05:03:22.100634 containerd[1601]: time="2025-05-14T05:03:22.100585171Z" level=info msg="received exit event container_id:\"4f69e6cb5691bc2cac3c350e267cfc08ddea5b7a1a9e4449a1689e738681e9b3\" id:\"4f69e6cb5691bc2cac3c350e267cfc08ddea5b7a1a9e4449a1689e738681e9b3\" pid:3356 exited_at:{seconds:1747199002 nanos:100096012}" May 14 05:03:22.100748 containerd[1601]: time="2025-05-14T05:03:22.100678528Z" level=info msg="TaskExit event in podsandbox handler container_id:\"4f69e6cb5691bc2cac3c350e267cfc08ddea5b7a1a9e4449a1689e738681e9b3\" id:\"4f69e6cb5691bc2cac3c350e267cfc08ddea5b7a1a9e4449a1689e738681e9b3\" pid:3356 exited_at:{seconds:1747199002 nanos:100096012}" May 14 05:03:22.124405 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-4f69e6cb5691bc2cac3c350e267cfc08ddea5b7a1a9e4449a1689e738681e9b3-rootfs.mount: Deactivated successfully. May 14 05:03:22.541672 kubelet[2713]: I0514 05:03:22.541642 2713 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 14 05:03:22.542147 kubelet[2713]: E0514 05:03:22.542001 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:22.542590 kubelet[2713]: E0514 05:03:22.542442 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:22.542800 containerd[1601]: time="2025-05-14T05:03:22.542776488Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.3\"" May 14 05:03:23.490164 kubelet[2713]: E0514 05:03:23.490119 2713 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-vwdck" podUID="f37146d1-125d-4167-a0d9-8ac3149fed78" May 14 05:03:25.489979 kubelet[2713]: E0514 05:03:25.489898 2713 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-vwdck" podUID="f37146d1-125d-4167-a0d9-8ac3149fed78" May 14 05:03:27.250580 containerd[1601]: time="2025-05-14T05:03:27.250513207Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:03:27.251250 containerd[1601]: time="2025-05-14T05:03:27.251186321Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.29.3: active requests=0, bytes read=97793683" May 14 05:03:27.252276 containerd[1601]: time="2025-05-14T05:03:27.252233361Z" level=info msg="ImageCreate event name:\"sha256:a140d04be1bc987bae0a1b9159e1dcb85751c448830efbdb3494207cf602b2d9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:03:27.254185 containerd[1601]: time="2025-05-14T05:03:27.254122955Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:4505ec8f976470994b6a94295a4dabac0cb98375db050e959a22603e00ada90b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:03:27.254663 containerd[1601]: time="2025-05-14T05:03:27.254616859Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.29.3\" with image id \"sha256:a140d04be1bc987bae0a1b9159e1dcb85751c448830efbdb3494207cf602b2d9\", repo tag \"ghcr.io/flatcar/calico/cni:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:4505ec8f976470994b6a94295a4dabac0cb98375db050e959a22603e00ada90b\", size \"99286305\" in 4.711734581s" May 14 05:03:27.254663 containerd[1601]: time="2025-05-14T05:03:27.254651975Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.3\" returns image reference \"sha256:a140d04be1bc987bae0a1b9159e1dcb85751c448830efbdb3494207cf602b2d9\"" May 14 05:03:27.256827 containerd[1601]: time="2025-05-14T05:03:27.256790170Z" level=info msg="CreateContainer within sandbox \"3601474911d55aef52f0a90124d287de4236d2e328eede33e262631c9388e41a\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" May 14 05:03:27.263922 containerd[1601]: time="2025-05-14T05:03:27.263880278Z" level=info msg="Container 1a0eb37de9590178f6bf4f192a3bebaf5070f261f896e7c7ab7aa441c4defbf2: CDI devices from CRI Config.CDIDevices: []" May 14 05:03:27.273278 containerd[1601]: time="2025-05-14T05:03:27.273213681Z" level=info msg="CreateContainer within sandbox \"3601474911d55aef52f0a90124d287de4236d2e328eede33e262631c9388e41a\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"1a0eb37de9590178f6bf4f192a3bebaf5070f261f896e7c7ab7aa441c4defbf2\"" May 14 05:03:27.273813 containerd[1601]: time="2025-05-14T05:03:27.273786524Z" level=info msg="StartContainer for \"1a0eb37de9590178f6bf4f192a3bebaf5070f261f896e7c7ab7aa441c4defbf2\"" May 14 05:03:27.275107 containerd[1601]: time="2025-05-14T05:03:27.274989079Z" level=info msg="connecting to shim 1a0eb37de9590178f6bf4f192a3bebaf5070f261f896e7c7ab7aa441c4defbf2" address="unix:///run/containerd/s/77e1234b6ea2bee922c8f1119f0fbf5107d4fd1008dd0105c7c7ebe949c23c89" protocol=ttrpc version=3 May 14 05:03:27.298607 systemd[1]: Started cri-containerd-1a0eb37de9590178f6bf4f192a3bebaf5070f261f896e7c7ab7aa441c4defbf2.scope - libcontainer container 1a0eb37de9590178f6bf4f192a3bebaf5070f261f896e7c7ab7aa441c4defbf2. May 14 05:03:27.346110 containerd[1601]: time="2025-05-14T05:03:27.346060527Z" level=info msg="StartContainer for \"1a0eb37de9590178f6bf4f192a3bebaf5070f261f896e7c7ab7aa441c4defbf2\" returns successfully" May 14 05:03:27.490381 kubelet[2713]: E0514 05:03:27.490321 2713 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-vwdck" podUID="f37146d1-125d-4167-a0d9-8ac3149fed78" May 14 05:03:27.553475 kubelet[2713]: E0514 05:03:27.553311 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:28.555233 kubelet[2713]: E0514 05:03:28.555179 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:28.594452 containerd[1601]: time="2025-05-14T05:03:28.594388554Z" level=error msg="failed to reload cni configuration after receiving fs change event(WRITE \"/etc/cni/net.d/calico-kubeconfig\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" May 14 05:03:28.597114 systemd[1]: cri-containerd-1a0eb37de9590178f6bf4f192a3bebaf5070f261f896e7c7ab7aa441c4defbf2.scope: Deactivated successfully. May 14 05:03:28.597877 systemd[1]: cri-containerd-1a0eb37de9590178f6bf4f192a3bebaf5070f261f896e7c7ab7aa441c4defbf2.scope: Consumed 566ms CPU time, 160.3M memory peak, 8K read from disk, 154M written to disk. May 14 05:03:28.598710 containerd[1601]: time="2025-05-14T05:03:28.598661041Z" level=info msg="received exit event container_id:\"1a0eb37de9590178f6bf4f192a3bebaf5070f261f896e7c7ab7aa441c4defbf2\" id:\"1a0eb37de9590178f6bf4f192a3bebaf5070f261f896e7c7ab7aa441c4defbf2\" pid:3414 exited_at:{seconds:1747199008 nanos:598416378}" May 14 05:03:28.598786 containerd[1601]: time="2025-05-14T05:03:28.598726194Z" level=info msg="TaskExit event in podsandbox handler container_id:\"1a0eb37de9590178f6bf4f192a3bebaf5070f261f896e7c7ab7aa441c4defbf2\" id:\"1a0eb37de9590178f6bf4f192a3bebaf5070f261f896e7c7ab7aa441c4defbf2\" pid:3414 exited_at:{seconds:1747199008 nanos:598416378}" May 14 05:03:28.621090 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-1a0eb37de9590178f6bf4f192a3bebaf5070f261f896e7c7ab7aa441c4defbf2-rootfs.mount: Deactivated successfully. May 14 05:03:28.630664 kubelet[2713]: I0514 05:03:28.630617 2713 kubelet_node_status.go:502] "Fast updating node status as it just became ready" May 14 05:03:28.753549 systemd[1]: Created slice kubepods-burstable-podd9aa914a_a52c_4b5f_ad56_47d2689175df.slice - libcontainer container kubepods-burstable-podd9aa914a_a52c_4b5f_ad56_47d2689175df.slice. May 14 05:03:28.761304 systemd[1]: Created slice kubepods-besteffort-pod7bb22493_3b00_4ec5_bf18_61c580662509.slice - libcontainer container kubepods-besteffort-pod7bb22493_3b00_4ec5_bf18_61c580662509.slice. May 14 05:03:28.769161 systemd[1]: Created slice kubepods-burstable-podc3637dd0_8b45_4255_a7d5_8005497de667.slice - libcontainer container kubepods-burstable-podc3637dd0_8b45_4255_a7d5_8005497de667.slice. May 14 05:03:28.776793 systemd[1]: Created slice kubepods-besteffort-pod82ca362d_c267_44f4_b9b7_f4f80c01a070.slice - libcontainer container kubepods-besteffort-pod82ca362d_c267_44f4_b9b7_f4f80c01a070.slice. May 14 05:03:28.781763 systemd[1]: Created slice kubepods-besteffort-pod19fede1c_9ba1_49bc_a0f6_8714dc44887e.slice - libcontainer container kubepods-besteffort-pod19fede1c_9ba1_49bc_a0f6_8714dc44887e.slice. May 14 05:03:28.923972 kubelet[2713]: I0514 05:03:28.923907 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkfhz\" (UniqueName: \"kubernetes.io/projected/d9aa914a-a52c-4b5f-ad56-47d2689175df-kube-api-access-vkfhz\") pod \"coredns-668d6bf9bc-gkkg2\" (UID: \"d9aa914a-a52c-4b5f-ad56-47d2689175df\") " pod="kube-system/coredns-668d6bf9bc-gkkg2" May 14 05:03:28.923972 kubelet[2713]: I0514 05:03:28.923963 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbbzk\" (UniqueName: \"kubernetes.io/projected/82ca362d-c267-44f4-b9b7-f4f80c01a070-kube-api-access-nbbzk\") pod \"calico-apiserver-bbffb6fd-5t7tc\" (UID: \"82ca362d-c267-44f4-b9b7-f4f80c01a070\") " pod="calico-apiserver/calico-apiserver-bbffb6fd-5t7tc" May 14 05:03:28.923972 kubelet[2713]: I0514 05:03:28.923983 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68wjd\" (UniqueName: \"kubernetes.io/projected/7bb22493-3b00-4ec5-bf18-61c580662509-kube-api-access-68wjd\") pod \"calico-apiserver-bbffb6fd-kw4jm\" (UID: \"7bb22493-3b00-4ec5-bf18-61c580662509\") " pod="calico-apiserver/calico-apiserver-bbffb6fd-kw4jm" May 14 05:03:28.924200 kubelet[2713]: I0514 05:03:28.924001 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d9aa914a-a52c-4b5f-ad56-47d2689175df-config-volume\") pod \"coredns-668d6bf9bc-gkkg2\" (UID: \"d9aa914a-a52c-4b5f-ad56-47d2689175df\") " pod="kube-system/coredns-668d6bf9bc-gkkg2" May 14 05:03:28.924200 kubelet[2713]: I0514 05:03:28.924019 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtt9h\" (UniqueName: \"kubernetes.io/projected/c3637dd0-8b45-4255-a7d5-8005497de667-kube-api-access-vtt9h\") pod \"coredns-668d6bf9bc-dmf4p\" (UID: \"c3637dd0-8b45-4255-a7d5-8005497de667\") " pod="kube-system/coredns-668d6bf9bc-dmf4p" May 14 05:03:28.924200 kubelet[2713]: I0514 05:03:28.924032 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/82ca362d-c267-44f4-b9b7-f4f80c01a070-calico-apiserver-certs\") pod \"calico-apiserver-bbffb6fd-5t7tc\" (UID: \"82ca362d-c267-44f4-b9b7-f4f80c01a070\") " pod="calico-apiserver/calico-apiserver-bbffb6fd-5t7tc" May 14 05:03:28.924200 kubelet[2713]: I0514 05:03:28.924046 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8qh7\" (UniqueName: \"kubernetes.io/projected/19fede1c-9ba1-49bc-a0f6-8714dc44887e-kube-api-access-t8qh7\") pod \"calico-kube-controllers-6479477dcb-f59jw\" (UID: \"19fede1c-9ba1-49bc-a0f6-8714dc44887e\") " pod="calico-system/calico-kube-controllers-6479477dcb-f59jw" May 14 05:03:28.924200 kubelet[2713]: I0514 05:03:28.924064 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c3637dd0-8b45-4255-a7d5-8005497de667-config-volume\") pod \"coredns-668d6bf9bc-dmf4p\" (UID: \"c3637dd0-8b45-4255-a7d5-8005497de667\") " pod="kube-system/coredns-668d6bf9bc-dmf4p" May 14 05:03:28.924341 kubelet[2713]: I0514 05:03:28.924078 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/19fede1c-9ba1-49bc-a0f6-8714dc44887e-tigera-ca-bundle\") pod \"calico-kube-controllers-6479477dcb-f59jw\" (UID: \"19fede1c-9ba1-49bc-a0f6-8714dc44887e\") " pod="calico-system/calico-kube-controllers-6479477dcb-f59jw" May 14 05:03:28.924341 kubelet[2713]: I0514 05:03:28.924141 2713 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/7bb22493-3b00-4ec5-bf18-61c580662509-calico-apiserver-certs\") pod \"calico-apiserver-bbffb6fd-kw4jm\" (UID: \"7bb22493-3b00-4ec5-bf18-61c580662509\") " pod="calico-apiserver/calico-apiserver-bbffb6fd-kw4jm" May 14 05:03:29.058936 kubelet[2713]: E0514 05:03:29.058884 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:29.059565 containerd[1601]: time="2025-05-14T05:03:29.059519935Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-gkkg2,Uid:d9aa914a-a52c-4b5f-ad56-47d2689175df,Namespace:kube-system,Attempt:0,}" May 14 05:03:29.066444 containerd[1601]: time="2025-05-14T05:03:29.066384265Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-bbffb6fd-kw4jm,Uid:7bb22493-3b00-4ec5-bf18-61c580662509,Namespace:calico-apiserver,Attempt:0,}" May 14 05:03:29.073123 kubelet[2713]: E0514 05:03:29.073056 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:29.073925 containerd[1601]: time="2025-05-14T05:03:29.073886800Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-dmf4p,Uid:c3637dd0-8b45-4255-a7d5-8005497de667,Namespace:kube-system,Attempt:0,}" May 14 05:03:29.082700 containerd[1601]: time="2025-05-14T05:03:29.082645018Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-bbffb6fd-5t7tc,Uid:82ca362d-c267-44f4-b9b7-f4f80c01a070,Namespace:calico-apiserver,Attempt:0,}" May 14 05:03:29.085721 containerd[1601]: time="2025-05-14T05:03:29.084856847Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6479477dcb-f59jw,Uid:19fede1c-9ba1-49bc-a0f6-8714dc44887e,Namespace:calico-system,Attempt:0,}" May 14 05:03:29.160286 containerd[1601]: time="2025-05-14T05:03:29.160187138Z" level=error msg="Failed to destroy network for sandbox \"c754e367e4344af0b73accd0e68abb8470c3beb53788ea521351766a9b7d4e61\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 14 05:03:29.161755 containerd[1601]: time="2025-05-14T05:03:29.161558730Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-bbffb6fd-kw4jm,Uid:7bb22493-3b00-4ec5-bf18-61c580662509,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"c754e367e4344af0b73accd0e68abb8470c3beb53788ea521351766a9b7d4e61\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 14 05:03:29.161886 kubelet[2713]: E0514 05:03:29.161796 2713 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c754e367e4344af0b73accd0e68abb8470c3beb53788ea521351766a9b7d4e61\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 14 05:03:29.161886 kubelet[2713]: E0514 05:03:29.161874 2713 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c754e367e4344af0b73accd0e68abb8470c3beb53788ea521351766a9b7d4e61\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-bbffb6fd-kw4jm" May 14 05:03:29.161974 kubelet[2713]: E0514 05:03:29.161894 2713 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c754e367e4344af0b73accd0e68abb8470c3beb53788ea521351766a9b7d4e61\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-bbffb6fd-kw4jm" May 14 05:03:29.161974 kubelet[2713]: E0514 05:03:29.161934 2713 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-bbffb6fd-kw4jm_calico-apiserver(7bb22493-3b00-4ec5-bf18-61c580662509)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-bbffb6fd-kw4jm_calico-apiserver(7bb22493-3b00-4ec5-bf18-61c580662509)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"c754e367e4344af0b73accd0e68abb8470c3beb53788ea521351766a9b7d4e61\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-bbffb6fd-kw4jm" podUID="7bb22493-3b00-4ec5-bf18-61c580662509" May 14 05:03:29.163307 containerd[1601]: time="2025-05-14T05:03:29.163162870Z" level=error msg="Failed to destroy network for sandbox \"1c30af2eeefc19b0e8e7a8cf5355e6f4e468c5383c5388528211e59d6bc39722\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 14 05:03:29.164583 containerd[1601]: time="2025-05-14T05:03:29.164547426Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-gkkg2,Uid:d9aa914a-a52c-4b5f-ad56-47d2689175df,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"1c30af2eeefc19b0e8e7a8cf5355e6f4e468c5383c5388528211e59d6bc39722\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 14 05:03:29.164957 kubelet[2713]: E0514 05:03:29.164926 2713 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1c30af2eeefc19b0e8e7a8cf5355e6f4e468c5383c5388528211e59d6bc39722\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 14 05:03:29.165062 kubelet[2713]: E0514 05:03:29.164968 2713 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1c30af2eeefc19b0e8e7a8cf5355e6f4e468c5383c5388528211e59d6bc39722\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-gkkg2" May 14 05:03:29.165062 kubelet[2713]: E0514 05:03:29.164982 2713 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1c30af2eeefc19b0e8e7a8cf5355e6f4e468c5383c5388528211e59d6bc39722\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-gkkg2" May 14 05:03:29.165857 kubelet[2713]: E0514 05:03:29.165495 2713 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-gkkg2_kube-system(d9aa914a-a52c-4b5f-ad56-47d2689175df)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-gkkg2_kube-system(d9aa914a-a52c-4b5f-ad56-47d2689175df)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"1c30af2eeefc19b0e8e7a8cf5355e6f4e468c5383c5388528211e59d6bc39722\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-gkkg2" podUID="d9aa914a-a52c-4b5f-ad56-47d2689175df" May 14 05:03:29.175817 containerd[1601]: time="2025-05-14T05:03:29.175273653Z" level=error msg="Failed to destroy network for sandbox \"24c53fee4a455d56affb09d81d49c687dc8f747f42d1a7601e500ff6de0c017f\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 14 05:03:29.175817 containerd[1601]: time="2025-05-14T05:03:29.175711430Z" level=error msg="Failed to destroy network for sandbox \"4b8da234c500cfd6821e7563881165c474a731e5f5f228668d30a42b8eea6f53\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 14 05:03:29.176831 containerd[1601]: time="2025-05-14T05:03:29.176781141Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-dmf4p,Uid:c3637dd0-8b45-4255-a7d5-8005497de667,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"24c53fee4a455d56affb09d81d49c687dc8f747f42d1a7601e500ff6de0c017f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 14 05:03:29.177117 kubelet[2713]: E0514 05:03:29.177057 2713 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"24c53fee4a455d56affb09d81d49c687dc8f747f42d1a7601e500ff6de0c017f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 14 05:03:29.177176 kubelet[2713]: E0514 05:03:29.177140 2713 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"24c53fee4a455d56affb09d81d49c687dc8f747f42d1a7601e500ff6de0c017f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-dmf4p" May 14 05:03:29.177203 kubelet[2713]: E0514 05:03:29.177169 2713 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"24c53fee4a455d56affb09d81d49c687dc8f747f42d1a7601e500ff6de0c017f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-668d6bf9bc-dmf4p" May 14 05:03:29.177283 kubelet[2713]: E0514 05:03:29.177244 2713 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-668d6bf9bc-dmf4p_kube-system(c3637dd0-8b45-4255-a7d5-8005497de667)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-668d6bf9bc-dmf4p_kube-system(c3637dd0-8b45-4255-a7d5-8005497de667)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"24c53fee4a455d56affb09d81d49c687dc8f747f42d1a7601e500ff6de0c017f\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-668d6bf9bc-dmf4p" podUID="c3637dd0-8b45-4255-a7d5-8005497de667" May 14 05:03:29.178574 containerd[1601]: time="2025-05-14T05:03:29.178265325Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-bbffb6fd-5t7tc,Uid:82ca362d-c267-44f4-b9b7-f4f80c01a070,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"4b8da234c500cfd6821e7563881165c474a731e5f5f228668d30a42b8eea6f53\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 14 05:03:29.178637 kubelet[2713]: E0514 05:03:29.178478 2713 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4b8da234c500cfd6821e7563881165c474a731e5f5f228668d30a42b8eea6f53\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 14 05:03:29.178637 kubelet[2713]: E0514 05:03:29.178613 2713 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4b8da234c500cfd6821e7563881165c474a731e5f5f228668d30a42b8eea6f53\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-bbffb6fd-5t7tc" May 14 05:03:29.178697 kubelet[2713]: E0514 05:03:29.178634 2713 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4b8da234c500cfd6821e7563881165c474a731e5f5f228668d30a42b8eea6f53\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-bbffb6fd-5t7tc" May 14 05:03:29.178755 kubelet[2713]: E0514 05:03:29.178708 2713 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-bbffb6fd-5t7tc_calico-apiserver(82ca362d-c267-44f4-b9b7-f4f80c01a070)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-bbffb6fd-5t7tc_calico-apiserver(82ca362d-c267-44f4-b9b7-f4f80c01a070)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"4b8da234c500cfd6821e7563881165c474a731e5f5f228668d30a42b8eea6f53\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-bbffb6fd-5t7tc" podUID="82ca362d-c267-44f4-b9b7-f4f80c01a070" May 14 05:03:29.180956 containerd[1601]: time="2025-05-14T05:03:29.180921815Z" level=error msg="Failed to destroy network for sandbox \"8a28ecd1cdf136aa724e11efe14d72fcfe8148584abc7b18bdae6d2e89a64aeb\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 14 05:03:29.182332 containerd[1601]: time="2025-05-14T05:03:29.182297012Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6479477dcb-f59jw,Uid:19fede1c-9ba1-49bc-a0f6-8714dc44887e,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"8a28ecd1cdf136aa724e11efe14d72fcfe8148584abc7b18bdae6d2e89a64aeb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 14 05:03:29.182631 kubelet[2713]: E0514 05:03:29.182488 2713 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8a28ecd1cdf136aa724e11efe14d72fcfe8148584abc7b18bdae6d2e89a64aeb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 14 05:03:29.182631 kubelet[2713]: E0514 05:03:29.182555 2713 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8a28ecd1cdf136aa724e11efe14d72fcfe8148584abc7b18bdae6d2e89a64aeb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-6479477dcb-f59jw" May 14 05:03:29.182631 kubelet[2713]: E0514 05:03:29.182573 2713 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8a28ecd1cdf136aa724e11efe14d72fcfe8148584abc7b18bdae6d2e89a64aeb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-6479477dcb-f59jw" May 14 05:03:29.182774 kubelet[2713]: E0514 05:03:29.182616 2713 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-6479477dcb-f59jw_calico-system(19fede1c-9ba1-49bc-a0f6-8714dc44887e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-6479477dcb-f59jw_calico-system(19fede1c-9ba1-49bc-a0f6-8714dc44887e)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"8a28ecd1cdf136aa724e11efe14d72fcfe8148584abc7b18bdae6d2e89a64aeb\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-6479477dcb-f59jw" podUID="19fede1c-9ba1-49bc-a0f6-8714dc44887e" May 14 05:03:29.495262 systemd[1]: Created slice kubepods-besteffort-podf37146d1_125d_4167_a0d9_8ac3149fed78.slice - libcontainer container kubepods-besteffort-podf37146d1_125d_4167_a0d9_8ac3149fed78.slice. May 14 05:03:29.497488 containerd[1601]: time="2025-05-14T05:03:29.497456120Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-vwdck,Uid:f37146d1-125d-4167-a0d9-8ac3149fed78,Namespace:calico-system,Attempt:0,}" May 14 05:03:29.544549 containerd[1601]: time="2025-05-14T05:03:29.544502180Z" level=error msg="Failed to destroy network for sandbox \"bb6895ad074fc4be7a472eef774a9764e0d9b0e93f5bb278c3fccfdaaff265ab\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 14 05:03:29.545683 containerd[1601]: time="2025-05-14T05:03:29.545653315Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-vwdck,Uid:f37146d1-125d-4167-a0d9-8ac3149fed78,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"bb6895ad074fc4be7a472eef774a9764e0d9b0e93f5bb278c3fccfdaaff265ab\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 14 05:03:29.545895 kubelet[2713]: E0514 05:03:29.545854 2713 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bb6895ad074fc4be7a472eef774a9764e0d9b0e93f5bb278c3fccfdaaff265ab\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" May 14 05:03:29.545948 kubelet[2713]: E0514 05:03:29.545915 2713 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bb6895ad074fc4be7a472eef774a9764e0d9b0e93f5bb278c3fccfdaaff265ab\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-vwdck" May 14 05:03:29.545948 kubelet[2713]: E0514 05:03:29.545934 2713 kuberuntime_manager.go:1237] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"bb6895ad074fc4be7a472eef774a9764e0d9b0e93f5bb278c3fccfdaaff265ab\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-vwdck" May 14 05:03:29.546021 kubelet[2713]: E0514 05:03:29.545977 2713 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-vwdck_calico-system(f37146d1-125d-4167-a0d9-8ac3149fed78)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-vwdck_calico-system(f37146d1-125d-4167-a0d9-8ac3149fed78)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"bb6895ad074fc4be7a472eef774a9764e0d9b0e93f5bb278c3fccfdaaff265ab\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-vwdck" podUID="f37146d1-125d-4167-a0d9-8ac3149fed78" May 14 05:03:29.559723 kubelet[2713]: E0514 05:03:29.559697 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:29.560825 containerd[1601]: time="2025-05-14T05:03:29.560572895Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.3\"" May 14 05:03:32.769262 systemd[1]: Started sshd@7-10.0.0.15:22-10.0.0.1:54682.service - OpenSSH per-connection server daemon (10.0.0.1:54682). May 14 05:03:32.830900 sshd[3677]: Accepted publickey for core from 10.0.0.1 port 54682 ssh2: RSA SHA256:9779yoEmBEYtokxLadky4y4rhX8tiwkjz8vWSk0iWXU May 14 05:03:32.832676 sshd-session[3677]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 05:03:32.837843 systemd-logind[1587]: New session 8 of user core. May 14 05:03:32.848623 systemd[1]: Started session-8.scope - Session 8 of User core. May 14 05:03:32.970455 sshd[3679]: Connection closed by 10.0.0.1 port 54682 May 14 05:03:32.970711 sshd-session[3677]: pam_unix(sshd:session): session closed for user core May 14 05:03:32.975528 systemd[1]: sshd@7-10.0.0.15:22-10.0.0.1:54682.service: Deactivated successfully. May 14 05:03:32.978173 systemd[1]: session-8.scope: Deactivated successfully. May 14 05:03:32.979209 systemd-logind[1587]: Session 8 logged out. Waiting for processes to exit. May 14 05:03:32.980808 systemd-logind[1587]: Removed session 8. May 14 05:03:35.827430 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount429179793.mount: Deactivated successfully. May 14 05:03:36.589096 containerd[1601]: time="2025-05-14T05:03:36.589020373Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:03:36.589828 containerd[1601]: time="2025-05-14T05:03:36.589783943Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.29.3: active requests=0, bytes read=144068748" May 14 05:03:36.590956 containerd[1601]: time="2025-05-14T05:03:36.590921346Z" level=info msg="ImageCreate event name:\"sha256:042163432abcec06b8077b24973b223a5f4cfdb35d85c3816f5d07a13d51afae\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:03:36.592794 containerd[1601]: time="2025-05-14T05:03:36.592749391Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:750e267b4f8217e0ca9e4107228370190d1a2499b72112ad04370ab9b4553916\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:03:36.593211 containerd[1601]: time="2025-05-14T05:03:36.593177206Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.29.3\" with image id \"sha256:042163432abcec06b8077b24973b223a5f4cfdb35d85c3816f5d07a13d51afae\", repo tag \"ghcr.io/flatcar/calico/node:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/node@sha256:750e267b4f8217e0ca9e4107228370190d1a2499b72112ad04370ab9b4553916\", size \"144068610\" in 7.032569907s" May 14 05:03:36.593211 containerd[1601]: time="2025-05-14T05:03:36.593205460Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.3\" returns image reference \"sha256:042163432abcec06b8077b24973b223a5f4cfdb35d85c3816f5d07a13d51afae\"" May 14 05:03:36.600600 containerd[1601]: time="2025-05-14T05:03:36.600564087Z" level=info msg="CreateContainer within sandbox \"3601474911d55aef52f0a90124d287de4236d2e328eede33e262631c9388e41a\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" May 14 05:03:36.610767 containerd[1601]: time="2025-05-14T05:03:36.610732328Z" level=info msg="Container 02ddbce4998bf831dc35ac719c47bc7fdbc2458ddad1416a7fb98143145b3a0e: CDI devices from CRI Config.CDIDevices: []" May 14 05:03:36.638700 containerd[1601]: time="2025-05-14T05:03:36.638660527Z" level=info msg="CreateContainer within sandbox \"3601474911d55aef52f0a90124d287de4236d2e328eede33e262631c9388e41a\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"02ddbce4998bf831dc35ac719c47bc7fdbc2458ddad1416a7fb98143145b3a0e\"" May 14 05:03:36.639155 containerd[1601]: time="2025-05-14T05:03:36.639123930Z" level=info msg="StartContainer for \"02ddbce4998bf831dc35ac719c47bc7fdbc2458ddad1416a7fb98143145b3a0e\"" May 14 05:03:36.640518 containerd[1601]: time="2025-05-14T05:03:36.640461592Z" level=info msg="connecting to shim 02ddbce4998bf831dc35ac719c47bc7fdbc2458ddad1416a7fb98143145b3a0e" address="unix:///run/containerd/s/77e1234b6ea2bee922c8f1119f0fbf5107d4fd1008dd0105c7c7ebe949c23c89" protocol=ttrpc version=3 May 14 05:03:36.661568 systemd[1]: Started cri-containerd-02ddbce4998bf831dc35ac719c47bc7fdbc2458ddad1416a7fb98143145b3a0e.scope - libcontainer container 02ddbce4998bf831dc35ac719c47bc7fdbc2458ddad1416a7fb98143145b3a0e. May 14 05:03:36.816156 containerd[1601]: time="2025-05-14T05:03:36.816117294Z" level=info msg="StartContainer for \"02ddbce4998bf831dc35ac719c47bc7fdbc2458ddad1416a7fb98143145b3a0e\" returns successfully" May 14 05:03:36.835072 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. May 14 05:03:36.835193 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. May 14 05:03:36.980426 kubelet[2713]: E0514 05:03:36.980113 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:36.996516 kubelet[2713]: I0514 05:03:36.996449 2713 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-cchm7" podStartSLOduration=0.941288491 podStartE2EDuration="19.996424682s" podCreationTimestamp="2025-05-14 05:03:17 +0000 UTC" firstStartedPulling="2025-05-14 05:03:17.538605008 +0000 UTC m=+14.121091926" lastFinishedPulling="2025-05-14 05:03:36.593741199 +0000 UTC m=+33.176228117" observedRunningTime="2025-05-14 05:03:36.995735673 +0000 UTC m=+33.578222581" watchObservedRunningTime="2025-05-14 05:03:36.996424682 +0000 UTC m=+33.578911600" May 14 05:03:37.986633 systemd[1]: Started sshd@8-10.0.0.15:22-10.0.0.1:54694.service - OpenSSH per-connection server daemon (10.0.0.1:54694). May 14 05:03:38.044159 sshd[3763]: Accepted publickey for core from 10.0.0.1 port 54694 ssh2: RSA SHA256:9779yoEmBEYtokxLadky4y4rhX8tiwkjz8vWSk0iWXU May 14 05:03:38.045381 sshd-session[3763]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 05:03:38.049728 systemd-logind[1587]: New session 9 of user core. May 14 05:03:38.059529 systemd[1]: Started session-9.scope - Session 9 of User core. May 14 05:03:38.232787 sshd[3765]: Connection closed by 10.0.0.1 port 54694 May 14 05:03:38.233124 sshd-session[3763]: pam_unix(sshd:session): session closed for user core May 14 05:03:38.237616 systemd[1]: sshd@8-10.0.0.15:22-10.0.0.1:54694.service: Deactivated successfully. May 14 05:03:38.239992 systemd[1]: session-9.scope: Deactivated successfully. May 14 05:03:38.240872 systemd-logind[1587]: Session 9 logged out. Waiting for processes to exit. May 14 05:03:38.242207 systemd-logind[1587]: Removed session 9. May 14 05:03:40.454966 kubelet[2713]: I0514 05:03:40.454888 2713 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 14 05:03:40.455516 kubelet[2713]: E0514 05:03:40.455460 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:40.490415 kubelet[2713]: E0514 05:03:40.490312 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:40.490802 containerd[1601]: time="2025-05-14T05:03:40.490763449Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-dmf4p,Uid:c3637dd0-8b45-4255-a7d5-8005497de667,Namespace:kube-system,Attempt:0,}" May 14 05:03:40.594287 containerd[1601]: time="2025-05-14T05:03:40.594239258Z" level=info msg="TaskExit event in podsandbox handler container_id:\"02ddbce4998bf831dc35ac719c47bc7fdbc2458ddad1416a7fb98143145b3a0e\" id:\"c596be1db43027adc702c729d2ad33320a119f6949574c02f46a63992792e4a1\" pid:3940 exit_status:1 exited_at:{seconds:1747199020 nanos:531169191}" May 14 05:03:40.618635 containerd[1601]: time="2025-05-14T05:03:40.618577213Z" level=info msg="TaskExit event in podsandbox handler container_id:\"02ddbce4998bf831dc35ac719c47bc7fdbc2458ddad1416a7fb98143145b3a0e\" id:\"eae78547243d6e7bb720115c38095339aaca1d8b9c7ec53a98f4445c499e6fa1\" pid:3965 exit_status:1 exited_at:{seconds:1747199020 nanos:618246310}" May 14 05:03:40.763962 systemd-networkd[1502]: cali47860cd406b: Link UP May 14 05:03:40.764443 systemd-networkd[1502]: cali47860cd406b: Gained carrier May 14 05:03:40.775604 containerd[1601]: 2025-05-14 05:03:40.639 [INFO][3977] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist May 14 05:03:40.775604 containerd[1601]: 2025-05-14 05:03:40.654 [INFO][3977] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--668d6bf9bc--dmf4p-eth0 coredns-668d6bf9bc- kube-system c3637dd0-8b45-4255-a7d5-8005497de667 696 0 2025-05-14 05:03:08 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-668d6bf9bc-dmf4p eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali47860cd406b [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="e11490967aaf29e6987c692510d6d55005602a15a41e7be7946ab07b63ae67d3" Namespace="kube-system" Pod="coredns-668d6bf9bc-dmf4p" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--dmf4p-" May 14 05:03:40.775604 containerd[1601]: 2025-05-14 05:03:40.654 [INFO][3977] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="e11490967aaf29e6987c692510d6d55005602a15a41e7be7946ab07b63ae67d3" Namespace="kube-system" Pod="coredns-668d6bf9bc-dmf4p" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--dmf4p-eth0" May 14 05:03:40.775604 containerd[1601]: 2025-05-14 05:03:40.709 [INFO][3992] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="e11490967aaf29e6987c692510d6d55005602a15a41e7be7946ab07b63ae67d3" HandleID="k8s-pod-network.e11490967aaf29e6987c692510d6d55005602a15a41e7be7946ab07b63ae67d3" Workload="localhost-k8s-coredns--668d6bf9bc--dmf4p-eth0" May 14 05:03:40.776009 containerd[1601]: 2025-05-14 05:03:40.721 [INFO][3992] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="e11490967aaf29e6987c692510d6d55005602a15a41e7be7946ab07b63ae67d3" HandleID="k8s-pod-network.e11490967aaf29e6987c692510d6d55005602a15a41e7be7946ab07b63ae67d3" Workload="localhost-k8s-coredns--668d6bf9bc--dmf4p-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000051e70), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-668d6bf9bc-dmf4p", "timestamp":"2025-05-14 05:03:40.70973873 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 14 05:03:40.776009 containerd[1601]: 2025-05-14 05:03:40.721 [INFO][3992] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 14 05:03:40.776009 containerd[1601]: 2025-05-14 05:03:40.721 [INFO][3992] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 14 05:03:40.776009 containerd[1601]: 2025-05-14 05:03:40.721 [INFO][3992] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' May 14 05:03:40.776009 containerd[1601]: 2025-05-14 05:03:40.728 [INFO][3992] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.e11490967aaf29e6987c692510d6d55005602a15a41e7be7946ab07b63ae67d3" host="localhost" May 14 05:03:40.776009 containerd[1601]: 2025-05-14 05:03:40.733 [INFO][3992] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" May 14 05:03:40.776009 containerd[1601]: 2025-05-14 05:03:40.738 [INFO][3992] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" May 14 05:03:40.776009 containerd[1601]: 2025-05-14 05:03:40.739 [INFO][3992] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" May 14 05:03:40.776009 containerd[1601]: 2025-05-14 05:03:40.741 [INFO][3992] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" May 14 05:03:40.776009 containerd[1601]: 2025-05-14 05:03:40.741 [INFO][3992] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.e11490967aaf29e6987c692510d6d55005602a15a41e7be7946ab07b63ae67d3" host="localhost" May 14 05:03:40.776294 containerd[1601]: 2025-05-14 05:03:40.742 [INFO][3992] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.e11490967aaf29e6987c692510d6d55005602a15a41e7be7946ab07b63ae67d3 May 14 05:03:40.776294 containerd[1601]: 2025-05-14 05:03:40.748 [INFO][3992] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.e11490967aaf29e6987c692510d6d55005602a15a41e7be7946ab07b63ae67d3" host="localhost" May 14 05:03:40.776294 containerd[1601]: 2025-05-14 05:03:40.752 [INFO][3992] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.129/26] block=192.168.88.128/26 handle="k8s-pod-network.e11490967aaf29e6987c692510d6d55005602a15a41e7be7946ab07b63ae67d3" host="localhost" May 14 05:03:40.776294 containerd[1601]: 2025-05-14 05:03:40.752 [INFO][3992] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.129/26] handle="k8s-pod-network.e11490967aaf29e6987c692510d6d55005602a15a41e7be7946ab07b63ae67d3" host="localhost" May 14 05:03:40.776294 containerd[1601]: 2025-05-14 05:03:40.752 [INFO][3992] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 14 05:03:40.776294 containerd[1601]: 2025-05-14 05:03:40.752 [INFO][3992] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.129/26] IPv6=[] ContainerID="e11490967aaf29e6987c692510d6d55005602a15a41e7be7946ab07b63ae67d3" HandleID="k8s-pod-network.e11490967aaf29e6987c692510d6d55005602a15a41e7be7946ab07b63ae67d3" Workload="localhost-k8s-coredns--668d6bf9bc--dmf4p-eth0" May 14 05:03:40.776525 containerd[1601]: 2025-05-14 05:03:40.756 [INFO][3977] cni-plugin/k8s.go 386: Populated endpoint ContainerID="e11490967aaf29e6987c692510d6d55005602a15a41e7be7946ab07b63ae67d3" Namespace="kube-system" Pod="coredns-668d6bf9bc-dmf4p" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--dmf4p-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--668d6bf9bc--dmf4p-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"c3637dd0-8b45-4255-a7d5-8005497de667", ResourceVersion:"696", Generation:0, CreationTimestamp:time.Date(2025, time.May, 14, 5, 3, 8, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-668d6bf9bc-dmf4p", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali47860cd406b", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} May 14 05:03:40.776616 containerd[1601]: 2025-05-14 05:03:40.756 [INFO][3977] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.129/32] ContainerID="e11490967aaf29e6987c692510d6d55005602a15a41e7be7946ab07b63ae67d3" Namespace="kube-system" Pod="coredns-668d6bf9bc-dmf4p" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--dmf4p-eth0" May 14 05:03:40.776616 containerd[1601]: 2025-05-14 05:03:40.756 [INFO][3977] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali47860cd406b ContainerID="e11490967aaf29e6987c692510d6d55005602a15a41e7be7946ab07b63ae67d3" Namespace="kube-system" Pod="coredns-668d6bf9bc-dmf4p" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--dmf4p-eth0" May 14 05:03:40.776616 containerd[1601]: 2025-05-14 05:03:40.763 [INFO][3977] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="e11490967aaf29e6987c692510d6d55005602a15a41e7be7946ab07b63ae67d3" Namespace="kube-system" Pod="coredns-668d6bf9bc-dmf4p" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--dmf4p-eth0" May 14 05:03:40.776710 containerd[1601]: 2025-05-14 05:03:40.764 [INFO][3977] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="e11490967aaf29e6987c692510d6d55005602a15a41e7be7946ab07b63ae67d3" Namespace="kube-system" Pod="coredns-668d6bf9bc-dmf4p" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--dmf4p-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--668d6bf9bc--dmf4p-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"c3637dd0-8b45-4255-a7d5-8005497de667", ResourceVersion:"696", Generation:0, CreationTimestamp:time.Date(2025, time.May, 14, 5, 3, 8, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"e11490967aaf29e6987c692510d6d55005602a15a41e7be7946ab07b63ae67d3", Pod:"coredns-668d6bf9bc-dmf4p", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali47860cd406b", MAC:"06:c9:ac:4d:94:a8", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} May 14 05:03:40.776710 containerd[1601]: 2025-05-14 05:03:40.772 [INFO][3977] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="e11490967aaf29e6987c692510d6d55005602a15a41e7be7946ab07b63ae67d3" Namespace="kube-system" Pod="coredns-668d6bf9bc-dmf4p" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--dmf4p-eth0" May 14 05:03:40.985344 kubelet[2713]: I0514 05:03:40.985266 2713 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 14 05:03:40.986197 kubelet[2713]: E0514 05:03:40.986158 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:41.036313 containerd[1601]: time="2025-05-14T05:03:41.036167933Z" level=info msg="connecting to shim e11490967aaf29e6987c692510d6d55005602a15a41e7be7946ab07b63ae67d3" address="unix:///run/containerd/s/8bb3fd27360cc8aef498768d9914913c81a3962f529693eff6c689a6cb371d08" namespace=k8s.io protocol=ttrpc version=3 May 14 05:03:41.066587 systemd[1]: Started cri-containerd-e11490967aaf29e6987c692510d6d55005602a15a41e7be7946ab07b63ae67d3.scope - libcontainer container e11490967aaf29e6987c692510d6d55005602a15a41e7be7946ab07b63ae67d3. May 14 05:03:41.079707 systemd-resolved[1413]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address May 14 05:03:41.110330 containerd[1601]: time="2025-05-14T05:03:41.110280224Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-dmf4p,Uid:c3637dd0-8b45-4255-a7d5-8005497de667,Namespace:kube-system,Attempt:0,} returns sandbox id \"e11490967aaf29e6987c692510d6d55005602a15a41e7be7946ab07b63ae67d3\"" May 14 05:03:41.111119 kubelet[2713]: E0514 05:03:41.111090 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:41.112912 containerd[1601]: time="2025-05-14T05:03:41.112855501Z" level=info msg="CreateContainer within sandbox \"e11490967aaf29e6987c692510d6d55005602a15a41e7be7946ab07b63ae67d3\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" May 14 05:03:41.123590 containerd[1601]: time="2025-05-14T05:03:41.123541272Z" level=info msg="Container d9c96db3fd158ed9a2468c23a5ad38bc39f9d039a2d14a5787c6e3c3bfb980f9: CDI devices from CRI Config.CDIDevices: []" May 14 05:03:41.127422 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2037286365.mount: Deactivated successfully. May 14 05:03:41.130798 containerd[1601]: time="2025-05-14T05:03:41.130761511Z" level=info msg="CreateContainer within sandbox \"e11490967aaf29e6987c692510d6d55005602a15a41e7be7946ab07b63ae67d3\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"d9c96db3fd158ed9a2468c23a5ad38bc39f9d039a2d14a5787c6e3c3bfb980f9\"" May 14 05:03:41.131246 containerd[1601]: time="2025-05-14T05:03:41.131189367Z" level=info msg="StartContainer for \"d9c96db3fd158ed9a2468c23a5ad38bc39f9d039a2d14a5787c6e3c3bfb980f9\"" May 14 05:03:41.132137 containerd[1601]: time="2025-05-14T05:03:41.132109538Z" level=info msg="connecting to shim d9c96db3fd158ed9a2468c23a5ad38bc39f9d039a2d14a5787c6e3c3bfb980f9" address="unix:///run/containerd/s/8bb3fd27360cc8aef498768d9914913c81a3962f529693eff6c689a6cb371d08" protocol=ttrpc version=3 May 14 05:03:41.157564 systemd[1]: Started cri-containerd-d9c96db3fd158ed9a2468c23a5ad38bc39f9d039a2d14a5787c6e3c3bfb980f9.scope - libcontainer container d9c96db3fd158ed9a2468c23a5ad38bc39f9d039a2d14a5787c6e3c3bfb980f9. May 14 05:03:41.189952 containerd[1601]: time="2025-05-14T05:03:41.189910000Z" level=info msg="StartContainer for \"d9c96db3fd158ed9a2468c23a5ad38bc39f9d039a2d14a5787c6e3c3bfb980f9\" returns successfully" May 14 05:03:41.649258 systemd-networkd[1502]: vxlan.calico: Link UP May 14 05:03:41.649758 systemd-networkd[1502]: vxlan.calico: Gained carrier May 14 05:03:41.990183 kubelet[2713]: E0514 05:03:41.990134 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:41.990702 kubelet[2713]: E0514 05:03:41.990321 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:41.999943 kubelet[2713]: I0514 05:03:41.999897 2713 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-dmf4p" podStartSLOduration=33.999881874 podStartE2EDuration="33.999881874s" podCreationTimestamp="2025-05-14 05:03:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-14 05:03:41.999022827 +0000 UTC m=+38.581509745" watchObservedRunningTime="2025-05-14 05:03:41.999881874 +0000 UTC m=+38.582368792" May 14 05:03:42.463625 systemd-networkd[1502]: cali47860cd406b: Gained IPv6LL May 14 05:03:42.490320 containerd[1601]: time="2025-05-14T05:03:42.490255991Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6479477dcb-f59jw,Uid:19fede1c-9ba1-49bc-a0f6-8714dc44887e,Namespace:calico-system,Attempt:0,}" May 14 05:03:42.490716 containerd[1601]: time="2025-05-14T05:03:42.490336532Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-bbffb6fd-5t7tc,Uid:82ca362d-c267-44f4-b9b7-f4f80c01a070,Namespace:calico-apiserver,Attempt:0,}" May 14 05:03:42.490716 containerd[1601]: time="2025-05-14T05:03:42.490429918Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-bbffb6fd-kw4jm,Uid:7bb22493-3b00-4ec5-bf18-61c580662509,Namespace:calico-apiserver,Attempt:0,}" May 14 05:03:42.618081 systemd-networkd[1502]: calic7d81e15d8d: Link UP May 14 05:03:42.618539 systemd-networkd[1502]: calic7d81e15d8d: Gained carrier May 14 05:03:42.633106 containerd[1601]: 2025-05-14 05:03:42.540 [INFO][4233] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--kube--controllers--6479477dcb--f59jw-eth0 calico-kube-controllers-6479477dcb- calico-system 19fede1c-9ba1-49bc-a0f6-8714dc44887e 697 0 2025-05-14 05:03:17 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:6479477dcb projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s localhost calico-kube-controllers-6479477dcb-f59jw eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] calic7d81e15d8d [] []}} ContainerID="19c60bc325dca1b7b906d84b6bd0ee9fab1743101dd38e240652f54645f5142a" Namespace="calico-system" Pod="calico-kube-controllers-6479477dcb-f59jw" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6479477dcb--f59jw-" May 14 05:03:42.633106 containerd[1601]: 2025-05-14 05:03:42.541 [INFO][4233] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="19c60bc325dca1b7b906d84b6bd0ee9fab1743101dd38e240652f54645f5142a" Namespace="calico-system" Pod="calico-kube-controllers-6479477dcb-f59jw" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6479477dcb--f59jw-eth0" May 14 05:03:42.633106 containerd[1601]: 2025-05-14 05:03:42.573 [INFO][4278] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="19c60bc325dca1b7b906d84b6bd0ee9fab1743101dd38e240652f54645f5142a" HandleID="k8s-pod-network.19c60bc325dca1b7b906d84b6bd0ee9fab1743101dd38e240652f54645f5142a" Workload="localhost-k8s-calico--kube--controllers--6479477dcb--f59jw-eth0" May 14 05:03:42.633106 containerd[1601]: 2025-05-14 05:03:42.584 [INFO][4278] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="19c60bc325dca1b7b906d84b6bd0ee9fab1743101dd38e240652f54645f5142a" HandleID="k8s-pod-network.19c60bc325dca1b7b906d84b6bd0ee9fab1743101dd38e240652f54645f5142a" Workload="localhost-k8s-calico--kube--controllers--6479477dcb--f59jw-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0003176a0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"calico-kube-controllers-6479477dcb-f59jw", "timestamp":"2025-05-14 05:03:42.573674334 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 14 05:03:42.633106 containerd[1601]: 2025-05-14 05:03:42.584 [INFO][4278] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 14 05:03:42.633106 containerd[1601]: 2025-05-14 05:03:42.584 [INFO][4278] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 14 05:03:42.633106 containerd[1601]: 2025-05-14 05:03:42.584 [INFO][4278] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' May 14 05:03:42.633106 containerd[1601]: 2025-05-14 05:03:42.586 [INFO][4278] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.19c60bc325dca1b7b906d84b6bd0ee9fab1743101dd38e240652f54645f5142a" host="localhost" May 14 05:03:42.633106 containerd[1601]: 2025-05-14 05:03:42.593 [INFO][4278] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" May 14 05:03:42.633106 containerd[1601]: 2025-05-14 05:03:42.597 [INFO][4278] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" May 14 05:03:42.633106 containerd[1601]: 2025-05-14 05:03:42.599 [INFO][4278] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" May 14 05:03:42.633106 containerd[1601]: 2025-05-14 05:03:42.601 [INFO][4278] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" May 14 05:03:42.633106 containerd[1601]: 2025-05-14 05:03:42.601 [INFO][4278] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.19c60bc325dca1b7b906d84b6bd0ee9fab1743101dd38e240652f54645f5142a" host="localhost" May 14 05:03:42.633106 containerd[1601]: 2025-05-14 05:03:42.603 [INFO][4278] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.19c60bc325dca1b7b906d84b6bd0ee9fab1743101dd38e240652f54645f5142a May 14 05:03:42.633106 containerd[1601]: 2025-05-14 05:03:42.608 [INFO][4278] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.19c60bc325dca1b7b906d84b6bd0ee9fab1743101dd38e240652f54645f5142a" host="localhost" May 14 05:03:42.633106 containerd[1601]: 2025-05-14 05:03:42.612 [INFO][4278] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.130/26] block=192.168.88.128/26 handle="k8s-pod-network.19c60bc325dca1b7b906d84b6bd0ee9fab1743101dd38e240652f54645f5142a" host="localhost" May 14 05:03:42.633106 containerd[1601]: 2025-05-14 05:03:42.612 [INFO][4278] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.130/26] handle="k8s-pod-network.19c60bc325dca1b7b906d84b6bd0ee9fab1743101dd38e240652f54645f5142a" host="localhost" May 14 05:03:42.633106 containerd[1601]: 2025-05-14 05:03:42.612 [INFO][4278] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 14 05:03:42.633106 containerd[1601]: 2025-05-14 05:03:42.612 [INFO][4278] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.130/26] IPv6=[] ContainerID="19c60bc325dca1b7b906d84b6bd0ee9fab1743101dd38e240652f54645f5142a" HandleID="k8s-pod-network.19c60bc325dca1b7b906d84b6bd0ee9fab1743101dd38e240652f54645f5142a" Workload="localhost-k8s-calico--kube--controllers--6479477dcb--f59jw-eth0" May 14 05:03:42.634232 containerd[1601]: 2025-05-14 05:03:42.615 [INFO][4233] cni-plugin/k8s.go 386: Populated endpoint ContainerID="19c60bc325dca1b7b906d84b6bd0ee9fab1743101dd38e240652f54645f5142a" Namespace="calico-system" Pod="calico-kube-controllers-6479477dcb-f59jw" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6479477dcb--f59jw-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--6479477dcb--f59jw-eth0", GenerateName:"calico-kube-controllers-6479477dcb-", Namespace:"calico-system", SelfLink:"", UID:"19fede1c-9ba1-49bc-a0f6-8714dc44887e", ResourceVersion:"697", Generation:0, CreationTimestamp:time.Date(2025, time.May, 14, 5, 3, 17, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"6479477dcb", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-kube-controllers-6479477dcb-f59jw", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calic7d81e15d8d", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 14 05:03:42.634232 containerd[1601]: 2025-05-14 05:03:42.616 [INFO][4233] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.130/32] ContainerID="19c60bc325dca1b7b906d84b6bd0ee9fab1743101dd38e240652f54645f5142a" Namespace="calico-system" Pod="calico-kube-controllers-6479477dcb-f59jw" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6479477dcb--f59jw-eth0" May 14 05:03:42.634232 containerd[1601]: 2025-05-14 05:03:42.616 [INFO][4233] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calic7d81e15d8d ContainerID="19c60bc325dca1b7b906d84b6bd0ee9fab1743101dd38e240652f54645f5142a" Namespace="calico-system" Pod="calico-kube-controllers-6479477dcb-f59jw" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6479477dcb--f59jw-eth0" May 14 05:03:42.634232 containerd[1601]: 2025-05-14 05:03:42.618 [INFO][4233] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="19c60bc325dca1b7b906d84b6bd0ee9fab1743101dd38e240652f54645f5142a" Namespace="calico-system" Pod="calico-kube-controllers-6479477dcb-f59jw" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6479477dcb--f59jw-eth0" May 14 05:03:42.634232 containerd[1601]: 2025-05-14 05:03:42.620 [INFO][4233] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="19c60bc325dca1b7b906d84b6bd0ee9fab1743101dd38e240652f54645f5142a" Namespace="calico-system" Pod="calico-kube-controllers-6479477dcb-f59jw" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6479477dcb--f59jw-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--6479477dcb--f59jw-eth0", GenerateName:"calico-kube-controllers-6479477dcb-", Namespace:"calico-system", SelfLink:"", UID:"19fede1c-9ba1-49bc-a0f6-8714dc44887e", ResourceVersion:"697", Generation:0, CreationTimestamp:time.Date(2025, time.May, 14, 5, 3, 17, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"6479477dcb", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"19c60bc325dca1b7b906d84b6bd0ee9fab1743101dd38e240652f54645f5142a", Pod:"calico-kube-controllers-6479477dcb-f59jw", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calic7d81e15d8d", MAC:"2a:08:79:10:40:7a", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 14 05:03:42.634232 containerd[1601]: 2025-05-14 05:03:42.628 [INFO][4233] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="19c60bc325dca1b7b906d84b6bd0ee9fab1743101dd38e240652f54645f5142a" Namespace="calico-system" Pod="calico-kube-controllers-6479477dcb-f59jw" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--6479477dcb--f59jw-eth0" May 14 05:03:42.673258 containerd[1601]: time="2025-05-14T05:03:42.673199052Z" level=info msg="connecting to shim 19c60bc325dca1b7b906d84b6bd0ee9fab1743101dd38e240652f54645f5142a" address="unix:///run/containerd/s/639b5588637fc05f03392a71700e0ed957a7e4b62120821818fecde669678930" namespace=k8s.io protocol=ttrpc version=3 May 14 05:03:42.703545 systemd[1]: Started cri-containerd-19c60bc325dca1b7b906d84b6bd0ee9fab1743101dd38e240652f54645f5142a.scope - libcontainer container 19c60bc325dca1b7b906d84b6bd0ee9fab1743101dd38e240652f54645f5142a. May 14 05:03:42.720974 systemd-resolved[1413]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address May 14 05:03:42.724280 systemd-networkd[1502]: cali0f6fd710712: Link UP May 14 05:03:42.725435 systemd-networkd[1502]: cali0f6fd710712: Gained carrier May 14 05:03:42.739130 containerd[1601]: 2025-05-14 05:03:42.550 [INFO][4244] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--bbffb6fd--kw4jm-eth0 calico-apiserver-bbffb6fd- calico-apiserver 7bb22493-3b00-4ec5-bf18-61c580662509 698 0 2025-05-14 05:03:17 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:bbffb6fd projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-bbffb6fd-kw4jm eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali0f6fd710712 [] []}} ContainerID="579dde6dd65bb084cc426da55e91b87142056a1acd12ff786be6053f3abc0b02" Namespace="calico-apiserver" Pod="calico-apiserver-bbffb6fd-kw4jm" WorkloadEndpoint="localhost-k8s-calico--apiserver--bbffb6fd--kw4jm-" May 14 05:03:42.739130 containerd[1601]: 2025-05-14 05:03:42.550 [INFO][4244] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="579dde6dd65bb084cc426da55e91b87142056a1acd12ff786be6053f3abc0b02" Namespace="calico-apiserver" Pod="calico-apiserver-bbffb6fd-kw4jm" WorkloadEndpoint="localhost-k8s-calico--apiserver--bbffb6fd--kw4jm-eth0" May 14 05:03:42.739130 containerd[1601]: 2025-05-14 05:03:42.587 [INFO][4284] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="579dde6dd65bb084cc426da55e91b87142056a1acd12ff786be6053f3abc0b02" HandleID="k8s-pod-network.579dde6dd65bb084cc426da55e91b87142056a1acd12ff786be6053f3abc0b02" Workload="localhost-k8s-calico--apiserver--bbffb6fd--kw4jm-eth0" May 14 05:03:42.739130 containerd[1601]: 2025-05-14 05:03:42.594 [INFO][4284] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="579dde6dd65bb084cc426da55e91b87142056a1acd12ff786be6053f3abc0b02" HandleID="k8s-pod-network.579dde6dd65bb084cc426da55e91b87142056a1acd12ff786be6053f3abc0b02" Workload="localhost-k8s-calico--apiserver--bbffb6fd--kw4jm-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000308330), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-bbffb6fd-kw4jm", "timestamp":"2025-05-14 05:03:42.587965975 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 14 05:03:42.739130 containerd[1601]: 2025-05-14 05:03:42.595 [INFO][4284] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 14 05:03:42.739130 containerd[1601]: 2025-05-14 05:03:42.612 [INFO][4284] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 14 05:03:42.739130 containerd[1601]: 2025-05-14 05:03:42.612 [INFO][4284] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' May 14 05:03:42.739130 containerd[1601]: 2025-05-14 05:03:42.688 [INFO][4284] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.579dde6dd65bb084cc426da55e91b87142056a1acd12ff786be6053f3abc0b02" host="localhost" May 14 05:03:42.739130 containerd[1601]: 2025-05-14 05:03:42.693 [INFO][4284] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" May 14 05:03:42.739130 containerd[1601]: 2025-05-14 05:03:42.697 [INFO][4284] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" May 14 05:03:42.739130 containerd[1601]: 2025-05-14 05:03:42.700 [INFO][4284] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" May 14 05:03:42.739130 containerd[1601]: 2025-05-14 05:03:42.702 [INFO][4284] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" May 14 05:03:42.739130 containerd[1601]: 2025-05-14 05:03:42.702 [INFO][4284] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.579dde6dd65bb084cc426da55e91b87142056a1acd12ff786be6053f3abc0b02" host="localhost" May 14 05:03:42.739130 containerd[1601]: 2025-05-14 05:03:42.704 [INFO][4284] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.579dde6dd65bb084cc426da55e91b87142056a1acd12ff786be6053f3abc0b02 May 14 05:03:42.739130 containerd[1601]: 2025-05-14 05:03:42.710 [INFO][4284] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.579dde6dd65bb084cc426da55e91b87142056a1acd12ff786be6053f3abc0b02" host="localhost" May 14 05:03:42.739130 containerd[1601]: 2025-05-14 05:03:42.717 [INFO][4284] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.131/26] block=192.168.88.128/26 handle="k8s-pod-network.579dde6dd65bb084cc426da55e91b87142056a1acd12ff786be6053f3abc0b02" host="localhost" May 14 05:03:42.739130 containerd[1601]: 2025-05-14 05:03:42.717 [INFO][4284] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.131/26] handle="k8s-pod-network.579dde6dd65bb084cc426da55e91b87142056a1acd12ff786be6053f3abc0b02" host="localhost" May 14 05:03:42.739130 containerd[1601]: 2025-05-14 05:03:42.717 [INFO][4284] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 14 05:03:42.739130 containerd[1601]: 2025-05-14 05:03:42.717 [INFO][4284] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.131/26] IPv6=[] ContainerID="579dde6dd65bb084cc426da55e91b87142056a1acd12ff786be6053f3abc0b02" HandleID="k8s-pod-network.579dde6dd65bb084cc426da55e91b87142056a1acd12ff786be6053f3abc0b02" Workload="localhost-k8s-calico--apiserver--bbffb6fd--kw4jm-eth0" May 14 05:03:42.739713 containerd[1601]: 2025-05-14 05:03:42.721 [INFO][4244] cni-plugin/k8s.go 386: Populated endpoint ContainerID="579dde6dd65bb084cc426da55e91b87142056a1acd12ff786be6053f3abc0b02" Namespace="calico-apiserver" Pod="calico-apiserver-bbffb6fd-kw4jm" WorkloadEndpoint="localhost-k8s-calico--apiserver--bbffb6fd--kw4jm-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--bbffb6fd--kw4jm-eth0", GenerateName:"calico-apiserver-bbffb6fd-", Namespace:"calico-apiserver", SelfLink:"", UID:"7bb22493-3b00-4ec5-bf18-61c580662509", ResourceVersion:"698", Generation:0, CreationTimestamp:time.Date(2025, time.May, 14, 5, 3, 17, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"bbffb6fd", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-bbffb6fd-kw4jm", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali0f6fd710712", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 14 05:03:42.739713 containerd[1601]: 2025-05-14 05:03:42.721 [INFO][4244] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.131/32] ContainerID="579dde6dd65bb084cc426da55e91b87142056a1acd12ff786be6053f3abc0b02" Namespace="calico-apiserver" Pod="calico-apiserver-bbffb6fd-kw4jm" WorkloadEndpoint="localhost-k8s-calico--apiserver--bbffb6fd--kw4jm-eth0" May 14 05:03:42.739713 containerd[1601]: 2025-05-14 05:03:42.721 [INFO][4244] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali0f6fd710712 ContainerID="579dde6dd65bb084cc426da55e91b87142056a1acd12ff786be6053f3abc0b02" Namespace="calico-apiserver" Pod="calico-apiserver-bbffb6fd-kw4jm" WorkloadEndpoint="localhost-k8s-calico--apiserver--bbffb6fd--kw4jm-eth0" May 14 05:03:42.739713 containerd[1601]: 2025-05-14 05:03:42.725 [INFO][4244] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="579dde6dd65bb084cc426da55e91b87142056a1acd12ff786be6053f3abc0b02" Namespace="calico-apiserver" Pod="calico-apiserver-bbffb6fd-kw4jm" WorkloadEndpoint="localhost-k8s-calico--apiserver--bbffb6fd--kw4jm-eth0" May 14 05:03:42.739713 containerd[1601]: 2025-05-14 05:03:42.726 [INFO][4244] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="579dde6dd65bb084cc426da55e91b87142056a1acd12ff786be6053f3abc0b02" Namespace="calico-apiserver" Pod="calico-apiserver-bbffb6fd-kw4jm" WorkloadEndpoint="localhost-k8s-calico--apiserver--bbffb6fd--kw4jm-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--bbffb6fd--kw4jm-eth0", GenerateName:"calico-apiserver-bbffb6fd-", Namespace:"calico-apiserver", SelfLink:"", UID:"7bb22493-3b00-4ec5-bf18-61c580662509", ResourceVersion:"698", Generation:0, CreationTimestamp:time.Date(2025, time.May, 14, 5, 3, 17, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"bbffb6fd", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"579dde6dd65bb084cc426da55e91b87142056a1acd12ff786be6053f3abc0b02", Pod:"calico-apiserver-bbffb6fd-kw4jm", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali0f6fd710712", MAC:"62:e1:f1:28:26:87", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 14 05:03:42.739713 containerd[1601]: 2025-05-14 05:03:42.735 [INFO][4244] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="579dde6dd65bb084cc426da55e91b87142056a1acd12ff786be6053f3abc0b02" Namespace="calico-apiserver" Pod="calico-apiserver-bbffb6fd-kw4jm" WorkloadEndpoint="localhost-k8s-calico--apiserver--bbffb6fd--kw4jm-eth0" May 14 05:03:42.764544 containerd[1601]: time="2025-05-14T05:03:42.764485861Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6479477dcb-f59jw,Uid:19fede1c-9ba1-49bc-a0f6-8714dc44887e,Namespace:calico-system,Attempt:0,} returns sandbox id \"19c60bc325dca1b7b906d84b6bd0ee9fab1743101dd38e240652f54645f5142a\"" May 14 05:03:42.768859 containerd[1601]: time="2025-05-14T05:03:42.768824105Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\"" May 14 05:03:42.769595 containerd[1601]: time="2025-05-14T05:03:42.769557254Z" level=info msg="connecting to shim 579dde6dd65bb084cc426da55e91b87142056a1acd12ff786be6053f3abc0b02" address="unix:///run/containerd/s/ad2107fdabc14e9eb59a6aa6e660346b4360665c154b002aef60521db0069244" namespace=k8s.io protocol=ttrpc version=3 May 14 05:03:42.795591 systemd[1]: Started cri-containerd-579dde6dd65bb084cc426da55e91b87142056a1acd12ff786be6053f3abc0b02.scope - libcontainer container 579dde6dd65bb084cc426da55e91b87142056a1acd12ff786be6053f3abc0b02. May 14 05:03:42.810087 systemd-resolved[1413]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address May 14 05:03:42.819338 systemd-networkd[1502]: cali09d777db576: Link UP May 14 05:03:42.819980 systemd-networkd[1502]: cali09d777db576: Gained carrier May 14 05:03:42.974604 containerd[1601]: time="2025-05-14T05:03:42.974472851Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-bbffb6fd-kw4jm,Uid:7bb22493-3b00-4ec5-bf18-61c580662509,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"579dde6dd65bb084cc426da55e91b87142056a1acd12ff786be6053f3abc0b02\"" May 14 05:03:42.991291 containerd[1601]: 2025-05-14 05:03:42.559 [INFO][4255] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--bbffb6fd--5t7tc-eth0 calico-apiserver-bbffb6fd- calico-apiserver 82ca362d-c267-44f4-b9b7-f4f80c01a070 699 0 2025-05-14 05:03:17 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:bbffb6fd projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-bbffb6fd-5t7tc eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali09d777db576 [] []}} ContainerID="a2a9caea92810841b3fc4c0f6a1be62821bd645aa7f46186755f0176c401ea39" Namespace="calico-apiserver" Pod="calico-apiserver-bbffb6fd-5t7tc" WorkloadEndpoint="localhost-k8s-calico--apiserver--bbffb6fd--5t7tc-" May 14 05:03:42.991291 containerd[1601]: 2025-05-14 05:03:42.559 [INFO][4255] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="a2a9caea92810841b3fc4c0f6a1be62821bd645aa7f46186755f0176c401ea39" Namespace="calico-apiserver" Pod="calico-apiserver-bbffb6fd-5t7tc" WorkloadEndpoint="localhost-k8s-calico--apiserver--bbffb6fd--5t7tc-eth0" May 14 05:03:42.991291 containerd[1601]: 2025-05-14 05:03:42.599 [INFO][4291] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="a2a9caea92810841b3fc4c0f6a1be62821bd645aa7f46186755f0176c401ea39" HandleID="k8s-pod-network.a2a9caea92810841b3fc4c0f6a1be62821bd645aa7f46186755f0176c401ea39" Workload="localhost-k8s-calico--apiserver--bbffb6fd--5t7tc-eth0" May 14 05:03:42.991291 containerd[1601]: 2025-05-14 05:03:42.683 [INFO][4291] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="a2a9caea92810841b3fc4c0f6a1be62821bd645aa7f46186755f0176c401ea39" HandleID="k8s-pod-network.a2a9caea92810841b3fc4c0f6a1be62821bd645aa7f46186755f0176c401ea39" Workload="localhost-k8s-calico--apiserver--bbffb6fd--5t7tc-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000332720), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-bbffb6fd-5t7tc", "timestamp":"2025-05-14 05:03:42.599730992 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 14 05:03:42.991291 containerd[1601]: 2025-05-14 05:03:42.683 [INFO][4291] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 14 05:03:42.991291 containerd[1601]: 2025-05-14 05:03:42.717 [INFO][4291] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 14 05:03:42.991291 containerd[1601]: 2025-05-14 05:03:42.717 [INFO][4291] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' May 14 05:03:42.991291 containerd[1601]: 2025-05-14 05:03:42.788 [INFO][4291] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.a2a9caea92810841b3fc4c0f6a1be62821bd645aa7f46186755f0176c401ea39" host="localhost" May 14 05:03:42.991291 containerd[1601]: 2025-05-14 05:03:42.794 [INFO][4291] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" May 14 05:03:42.991291 containerd[1601]: 2025-05-14 05:03:42.798 [INFO][4291] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" May 14 05:03:42.991291 containerd[1601]: 2025-05-14 05:03:42.800 [INFO][4291] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" May 14 05:03:42.991291 containerd[1601]: 2025-05-14 05:03:42.802 [INFO][4291] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" May 14 05:03:42.991291 containerd[1601]: 2025-05-14 05:03:42.802 [INFO][4291] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.a2a9caea92810841b3fc4c0f6a1be62821bd645aa7f46186755f0176c401ea39" host="localhost" May 14 05:03:42.991291 containerd[1601]: 2025-05-14 05:03:42.804 [INFO][4291] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.a2a9caea92810841b3fc4c0f6a1be62821bd645aa7f46186755f0176c401ea39 May 14 05:03:42.991291 containerd[1601]: 2025-05-14 05:03:42.808 [INFO][4291] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.a2a9caea92810841b3fc4c0f6a1be62821bd645aa7f46186755f0176c401ea39" host="localhost" May 14 05:03:42.991291 containerd[1601]: 2025-05-14 05:03:42.813 [INFO][4291] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.132/26] block=192.168.88.128/26 handle="k8s-pod-network.a2a9caea92810841b3fc4c0f6a1be62821bd645aa7f46186755f0176c401ea39" host="localhost" May 14 05:03:42.991291 containerd[1601]: 2025-05-14 05:03:42.813 [INFO][4291] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.132/26] handle="k8s-pod-network.a2a9caea92810841b3fc4c0f6a1be62821bd645aa7f46186755f0176c401ea39" host="localhost" May 14 05:03:42.991291 containerd[1601]: 2025-05-14 05:03:42.813 [INFO][4291] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 14 05:03:42.991291 containerd[1601]: 2025-05-14 05:03:42.813 [INFO][4291] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.132/26] IPv6=[] ContainerID="a2a9caea92810841b3fc4c0f6a1be62821bd645aa7f46186755f0176c401ea39" HandleID="k8s-pod-network.a2a9caea92810841b3fc4c0f6a1be62821bd645aa7f46186755f0176c401ea39" Workload="localhost-k8s-calico--apiserver--bbffb6fd--5t7tc-eth0" May 14 05:03:42.992341 containerd[1601]: 2025-05-14 05:03:42.817 [INFO][4255] cni-plugin/k8s.go 386: Populated endpoint ContainerID="a2a9caea92810841b3fc4c0f6a1be62821bd645aa7f46186755f0176c401ea39" Namespace="calico-apiserver" Pod="calico-apiserver-bbffb6fd-5t7tc" WorkloadEndpoint="localhost-k8s-calico--apiserver--bbffb6fd--5t7tc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--bbffb6fd--5t7tc-eth0", GenerateName:"calico-apiserver-bbffb6fd-", Namespace:"calico-apiserver", SelfLink:"", UID:"82ca362d-c267-44f4-b9b7-f4f80c01a070", ResourceVersion:"699", Generation:0, CreationTimestamp:time.Date(2025, time.May, 14, 5, 3, 17, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"bbffb6fd", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-bbffb6fd-5t7tc", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali09d777db576", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 14 05:03:42.992341 containerd[1601]: 2025-05-14 05:03:42.817 [INFO][4255] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.132/32] ContainerID="a2a9caea92810841b3fc4c0f6a1be62821bd645aa7f46186755f0176c401ea39" Namespace="calico-apiserver" Pod="calico-apiserver-bbffb6fd-5t7tc" WorkloadEndpoint="localhost-k8s-calico--apiserver--bbffb6fd--5t7tc-eth0" May 14 05:03:42.992341 containerd[1601]: 2025-05-14 05:03:42.817 [INFO][4255] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali09d777db576 ContainerID="a2a9caea92810841b3fc4c0f6a1be62821bd645aa7f46186755f0176c401ea39" Namespace="calico-apiserver" Pod="calico-apiserver-bbffb6fd-5t7tc" WorkloadEndpoint="localhost-k8s-calico--apiserver--bbffb6fd--5t7tc-eth0" May 14 05:03:42.992341 containerd[1601]: 2025-05-14 05:03:42.820 [INFO][4255] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="a2a9caea92810841b3fc4c0f6a1be62821bd645aa7f46186755f0176c401ea39" Namespace="calico-apiserver" Pod="calico-apiserver-bbffb6fd-5t7tc" WorkloadEndpoint="localhost-k8s-calico--apiserver--bbffb6fd--5t7tc-eth0" May 14 05:03:42.992341 containerd[1601]: 2025-05-14 05:03:42.820 [INFO][4255] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="a2a9caea92810841b3fc4c0f6a1be62821bd645aa7f46186755f0176c401ea39" Namespace="calico-apiserver" Pod="calico-apiserver-bbffb6fd-5t7tc" WorkloadEndpoint="localhost-k8s-calico--apiserver--bbffb6fd--5t7tc-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--bbffb6fd--5t7tc-eth0", GenerateName:"calico-apiserver-bbffb6fd-", Namespace:"calico-apiserver", SelfLink:"", UID:"82ca362d-c267-44f4-b9b7-f4f80c01a070", ResourceVersion:"699", Generation:0, CreationTimestamp:time.Date(2025, time.May, 14, 5, 3, 17, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"bbffb6fd", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"a2a9caea92810841b3fc4c0f6a1be62821bd645aa7f46186755f0176c401ea39", Pod:"calico-apiserver-bbffb6fd-5t7tc", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali09d777db576", MAC:"1a:40:13:54:76:a8", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 14 05:03:42.992341 containerd[1601]: 2025-05-14 05:03:42.986 [INFO][4255] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="a2a9caea92810841b3fc4c0f6a1be62821bd645aa7f46186755f0176c401ea39" Namespace="calico-apiserver" Pod="calico-apiserver-bbffb6fd-5t7tc" WorkloadEndpoint="localhost-k8s-calico--apiserver--bbffb6fd--5t7tc-eth0" May 14 05:03:42.996872 kubelet[2713]: E0514 05:03:42.996837 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:43.064610 containerd[1601]: time="2025-05-14T05:03:43.064556502Z" level=info msg="connecting to shim a2a9caea92810841b3fc4c0f6a1be62821bd645aa7f46186755f0176c401ea39" address="unix:///run/containerd/s/d974314019a4dae73031b984ebb955bb10145480262a9cc997af1ae274ec7a62" namespace=k8s.io protocol=ttrpc version=3 May 14 05:03:43.094634 systemd[1]: Started cri-containerd-a2a9caea92810841b3fc4c0f6a1be62821bd645aa7f46186755f0176c401ea39.scope - libcontainer container a2a9caea92810841b3fc4c0f6a1be62821bd645aa7f46186755f0176c401ea39. May 14 05:03:43.109959 systemd-resolved[1413]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address May 14 05:03:43.144812 containerd[1601]: time="2025-05-14T05:03:43.144755557Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-bbffb6fd-5t7tc,Uid:82ca362d-c267-44f4-b9b7-f4f80c01a070,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"a2a9caea92810841b3fc4c0f6a1be62821bd645aa7f46186755f0176c401ea39\"" May 14 05:03:43.255468 systemd[1]: Started sshd@9-10.0.0.15:22-10.0.0.1:38032.service - OpenSSH per-connection server daemon (10.0.0.1:38032). May 14 05:03:43.310066 sshd[4488]: Accepted publickey for core from 10.0.0.1 port 38032 ssh2: RSA SHA256:9779yoEmBEYtokxLadky4y4rhX8tiwkjz8vWSk0iWXU May 14 05:03:43.311518 sshd-session[4488]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 05:03:43.315990 systemd-logind[1587]: New session 10 of user core. May 14 05:03:43.327523 systemd[1]: Started session-10.scope - Session 10 of User core. May 14 05:03:43.424573 systemd-networkd[1502]: vxlan.calico: Gained IPv6LL May 14 05:03:43.452226 sshd[4491]: Connection closed by 10.0.0.1 port 38032 May 14 05:03:43.452598 sshd-session[4488]: pam_unix(sshd:session): session closed for user core May 14 05:03:43.457253 systemd[1]: sshd@9-10.0.0.15:22-10.0.0.1:38032.service: Deactivated successfully. May 14 05:03:43.459791 systemd[1]: session-10.scope: Deactivated successfully. May 14 05:03:43.460802 systemd-logind[1587]: Session 10 logged out. Waiting for processes to exit. May 14 05:03:43.462814 systemd-logind[1587]: Removed session 10. May 14 05:03:43.999210 kubelet[2713]: E0514 05:03:43.999166 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:43.999849 systemd-networkd[1502]: calic7d81e15d8d: Gained IPv6LL May 14 05:03:44.191582 systemd-networkd[1502]: cali0f6fd710712: Gained IPv6LL May 14 05:03:44.490413 kubelet[2713]: E0514 05:03:44.490361 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:44.490649 containerd[1601]: time="2025-05-14T05:03:44.490564485Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-vwdck,Uid:f37146d1-125d-4167-a0d9-8ac3149fed78,Namespace:calico-system,Attempt:0,}" May 14 05:03:44.491329 containerd[1601]: time="2025-05-14T05:03:44.491266625Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-gkkg2,Uid:d9aa914a-a52c-4b5f-ad56-47d2689175df,Namespace:kube-system,Attempt:0,}" May 14 05:03:44.604729 systemd-networkd[1502]: cali43561bc3405: Link UP May 14 05:03:44.605286 systemd-networkd[1502]: cali43561bc3405: Gained carrier May 14 05:03:44.617780 containerd[1601]: 2025-05-14 05:03:44.534 [INFO][4512] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--668d6bf9bc--gkkg2-eth0 coredns-668d6bf9bc- kube-system d9aa914a-a52c-4b5f-ad56-47d2689175df 692 0 2025-05-14 05:03:08 +0000 UTC map[k8s-app:kube-dns pod-template-hash:668d6bf9bc projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-668d6bf9bc-gkkg2 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali43561bc3405 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="e466a24ab4b8cff7e881b82e598ca7cd2983680c6371b4c4b77cecd5f4c21f29" Namespace="kube-system" Pod="coredns-668d6bf9bc-gkkg2" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--gkkg2-" May 14 05:03:44.617780 containerd[1601]: 2025-05-14 05:03:44.534 [INFO][4512] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="e466a24ab4b8cff7e881b82e598ca7cd2983680c6371b4c4b77cecd5f4c21f29" Namespace="kube-system" Pod="coredns-668d6bf9bc-gkkg2" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--gkkg2-eth0" May 14 05:03:44.617780 containerd[1601]: 2025-05-14 05:03:44.566 [INFO][4534] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="e466a24ab4b8cff7e881b82e598ca7cd2983680c6371b4c4b77cecd5f4c21f29" HandleID="k8s-pod-network.e466a24ab4b8cff7e881b82e598ca7cd2983680c6371b4c4b77cecd5f4c21f29" Workload="localhost-k8s-coredns--668d6bf9bc--gkkg2-eth0" May 14 05:03:44.617780 containerd[1601]: 2025-05-14 05:03:44.574 [INFO][4534] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="e466a24ab4b8cff7e881b82e598ca7cd2983680c6371b4c4b77cecd5f4c21f29" HandleID="k8s-pod-network.e466a24ab4b8cff7e881b82e598ca7cd2983680c6371b4c4b77cecd5f4c21f29" Workload="localhost-k8s-coredns--668d6bf9bc--gkkg2-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002e4c90), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-668d6bf9bc-gkkg2", "timestamp":"2025-05-14 05:03:44.566435431 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 14 05:03:44.617780 containerd[1601]: 2025-05-14 05:03:44.574 [INFO][4534] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 14 05:03:44.617780 containerd[1601]: 2025-05-14 05:03:44.574 [INFO][4534] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 14 05:03:44.617780 containerd[1601]: 2025-05-14 05:03:44.574 [INFO][4534] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' May 14 05:03:44.617780 containerd[1601]: 2025-05-14 05:03:44.576 [INFO][4534] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.e466a24ab4b8cff7e881b82e598ca7cd2983680c6371b4c4b77cecd5f4c21f29" host="localhost" May 14 05:03:44.617780 containerd[1601]: 2025-05-14 05:03:44.579 [INFO][4534] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" May 14 05:03:44.617780 containerd[1601]: 2025-05-14 05:03:44.583 [INFO][4534] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" May 14 05:03:44.617780 containerd[1601]: 2025-05-14 05:03:44.585 [INFO][4534] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" May 14 05:03:44.617780 containerd[1601]: 2025-05-14 05:03:44.587 [INFO][4534] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" May 14 05:03:44.617780 containerd[1601]: 2025-05-14 05:03:44.587 [INFO][4534] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.e466a24ab4b8cff7e881b82e598ca7cd2983680c6371b4c4b77cecd5f4c21f29" host="localhost" May 14 05:03:44.617780 containerd[1601]: 2025-05-14 05:03:44.589 [INFO][4534] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.e466a24ab4b8cff7e881b82e598ca7cd2983680c6371b4c4b77cecd5f4c21f29 May 14 05:03:44.617780 containerd[1601]: 2025-05-14 05:03:44.592 [INFO][4534] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.e466a24ab4b8cff7e881b82e598ca7cd2983680c6371b4c4b77cecd5f4c21f29" host="localhost" May 14 05:03:44.617780 containerd[1601]: 2025-05-14 05:03:44.599 [INFO][4534] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.133/26] block=192.168.88.128/26 handle="k8s-pod-network.e466a24ab4b8cff7e881b82e598ca7cd2983680c6371b4c4b77cecd5f4c21f29" host="localhost" May 14 05:03:44.617780 containerd[1601]: 2025-05-14 05:03:44.599 [INFO][4534] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.133/26] handle="k8s-pod-network.e466a24ab4b8cff7e881b82e598ca7cd2983680c6371b4c4b77cecd5f4c21f29" host="localhost" May 14 05:03:44.617780 containerd[1601]: 2025-05-14 05:03:44.599 [INFO][4534] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 14 05:03:44.617780 containerd[1601]: 2025-05-14 05:03:44.599 [INFO][4534] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.133/26] IPv6=[] ContainerID="e466a24ab4b8cff7e881b82e598ca7cd2983680c6371b4c4b77cecd5f4c21f29" HandleID="k8s-pod-network.e466a24ab4b8cff7e881b82e598ca7cd2983680c6371b4c4b77cecd5f4c21f29" Workload="localhost-k8s-coredns--668d6bf9bc--gkkg2-eth0" May 14 05:03:44.618541 containerd[1601]: 2025-05-14 05:03:44.601 [INFO][4512] cni-plugin/k8s.go 386: Populated endpoint ContainerID="e466a24ab4b8cff7e881b82e598ca7cd2983680c6371b4c4b77cecd5f4c21f29" Namespace="kube-system" Pod="coredns-668d6bf9bc-gkkg2" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--gkkg2-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--668d6bf9bc--gkkg2-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"d9aa914a-a52c-4b5f-ad56-47d2689175df", ResourceVersion:"692", Generation:0, CreationTimestamp:time.Date(2025, time.May, 14, 5, 3, 8, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-668d6bf9bc-gkkg2", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali43561bc3405", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} May 14 05:03:44.618541 containerd[1601]: 2025-05-14 05:03:44.602 [INFO][4512] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.133/32] ContainerID="e466a24ab4b8cff7e881b82e598ca7cd2983680c6371b4c4b77cecd5f4c21f29" Namespace="kube-system" Pod="coredns-668d6bf9bc-gkkg2" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--gkkg2-eth0" May 14 05:03:44.618541 containerd[1601]: 2025-05-14 05:03:44.602 [INFO][4512] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali43561bc3405 ContainerID="e466a24ab4b8cff7e881b82e598ca7cd2983680c6371b4c4b77cecd5f4c21f29" Namespace="kube-system" Pod="coredns-668d6bf9bc-gkkg2" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--gkkg2-eth0" May 14 05:03:44.618541 containerd[1601]: 2025-05-14 05:03:44.605 [INFO][4512] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="e466a24ab4b8cff7e881b82e598ca7cd2983680c6371b4c4b77cecd5f4c21f29" Namespace="kube-system" Pod="coredns-668d6bf9bc-gkkg2" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--gkkg2-eth0" May 14 05:03:44.618541 containerd[1601]: 2025-05-14 05:03:44.605 [INFO][4512] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="e466a24ab4b8cff7e881b82e598ca7cd2983680c6371b4c4b77cecd5f4c21f29" Namespace="kube-system" Pod="coredns-668d6bf9bc-gkkg2" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--gkkg2-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--668d6bf9bc--gkkg2-eth0", GenerateName:"coredns-668d6bf9bc-", Namespace:"kube-system", SelfLink:"", UID:"d9aa914a-a52c-4b5f-ad56-47d2689175df", ResourceVersion:"692", Generation:0, CreationTimestamp:time.Date(2025, time.May, 14, 5, 3, 8, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"668d6bf9bc", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"e466a24ab4b8cff7e881b82e598ca7cd2983680c6371b4c4b77cecd5f4c21f29", Pod:"coredns-668d6bf9bc-gkkg2", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali43561bc3405", MAC:"2e:47:0f:49:f0:05", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} May 14 05:03:44.618541 containerd[1601]: 2025-05-14 05:03:44.614 [INFO][4512] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="e466a24ab4b8cff7e881b82e598ca7cd2983680c6371b4c4b77cecd5f4c21f29" Namespace="kube-system" Pod="coredns-668d6bf9bc-gkkg2" WorkloadEndpoint="localhost-k8s-coredns--668d6bf9bc--gkkg2-eth0" May 14 05:03:44.673117 containerd[1601]: time="2025-05-14T05:03:44.673070514Z" level=info msg="connecting to shim e466a24ab4b8cff7e881b82e598ca7cd2983680c6371b4c4b77cecd5f4c21f29" address="unix:///run/containerd/s/c7ae3ceff44861ddd66ca3a6cc8234fafae5f4e475a008d080a0ff29a6245a9c" namespace=k8s.io protocol=ttrpc version=3 May 14 05:03:44.699558 systemd[1]: Started cri-containerd-e466a24ab4b8cff7e881b82e598ca7cd2983680c6371b4c4b77cecd5f4c21f29.scope - libcontainer container e466a24ab4b8cff7e881b82e598ca7cd2983680c6371b4c4b77cecd5f4c21f29. May 14 05:03:44.712586 systemd-resolved[1413]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address May 14 05:03:44.717531 systemd-networkd[1502]: cali6b688c1748b: Link UP May 14 05:03:44.719065 systemd-networkd[1502]: cali6b688c1748b: Gained carrier May 14 05:03:44.740925 containerd[1601]: 2025-05-14 05:03:44.533 [INFO][4505] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-csi--node--driver--vwdck-eth0 csi-node-driver- calico-system f37146d1-125d-4167-a0d9-8ac3149fed78 601 0 2025-05-14 05:03:17 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:5b5cc68cd5 k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s localhost csi-node-driver-vwdck eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali6b688c1748b [] []}} ContainerID="c66d1501764dc68679044f77cb61e5bfdd9e97907cca72763368ef158dcbf468" Namespace="calico-system" Pod="csi-node-driver-vwdck" WorkloadEndpoint="localhost-k8s-csi--node--driver--vwdck-" May 14 05:03:44.740925 containerd[1601]: 2025-05-14 05:03:44.533 [INFO][4505] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="c66d1501764dc68679044f77cb61e5bfdd9e97907cca72763368ef158dcbf468" Namespace="calico-system" Pod="csi-node-driver-vwdck" WorkloadEndpoint="localhost-k8s-csi--node--driver--vwdck-eth0" May 14 05:03:44.740925 containerd[1601]: 2025-05-14 05:03:44.565 [INFO][4536] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="c66d1501764dc68679044f77cb61e5bfdd9e97907cca72763368ef158dcbf468" HandleID="k8s-pod-network.c66d1501764dc68679044f77cb61e5bfdd9e97907cca72763368ef158dcbf468" Workload="localhost-k8s-csi--node--driver--vwdck-eth0" May 14 05:03:44.740925 containerd[1601]: 2025-05-14 05:03:44.576 [INFO][4536] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="c66d1501764dc68679044f77cb61e5bfdd9e97907cca72763368ef158dcbf468" HandleID="k8s-pod-network.c66d1501764dc68679044f77cb61e5bfdd9e97907cca72763368ef158dcbf468" Workload="localhost-k8s-csi--node--driver--vwdck-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0003acd80), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"csi-node-driver-vwdck", "timestamp":"2025-05-14 05:03:44.565174981 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} May 14 05:03:44.740925 containerd[1601]: 2025-05-14 05:03:44.576 [INFO][4536] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. May 14 05:03:44.740925 containerd[1601]: 2025-05-14 05:03:44.599 [INFO][4536] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. May 14 05:03:44.740925 containerd[1601]: 2025-05-14 05:03:44.599 [INFO][4536] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' May 14 05:03:44.740925 containerd[1601]: 2025-05-14 05:03:44.678 [INFO][4536] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.c66d1501764dc68679044f77cb61e5bfdd9e97907cca72763368ef158dcbf468" host="localhost" May 14 05:03:44.740925 containerd[1601]: 2025-05-14 05:03:44.685 [INFO][4536] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" May 14 05:03:44.740925 containerd[1601]: 2025-05-14 05:03:44.690 [INFO][4536] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" May 14 05:03:44.740925 containerd[1601]: 2025-05-14 05:03:44.692 [INFO][4536] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" May 14 05:03:44.740925 containerd[1601]: 2025-05-14 05:03:44.694 [INFO][4536] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" May 14 05:03:44.740925 containerd[1601]: 2025-05-14 05:03:44.694 [INFO][4536] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.c66d1501764dc68679044f77cb61e5bfdd9e97907cca72763368ef158dcbf468" host="localhost" May 14 05:03:44.740925 containerd[1601]: 2025-05-14 05:03:44.696 [INFO][4536] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.c66d1501764dc68679044f77cb61e5bfdd9e97907cca72763368ef158dcbf468 May 14 05:03:44.740925 containerd[1601]: 2025-05-14 05:03:44.701 [INFO][4536] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.c66d1501764dc68679044f77cb61e5bfdd9e97907cca72763368ef158dcbf468" host="localhost" May 14 05:03:44.740925 containerd[1601]: 2025-05-14 05:03:44.708 [INFO][4536] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.134/26] block=192.168.88.128/26 handle="k8s-pod-network.c66d1501764dc68679044f77cb61e5bfdd9e97907cca72763368ef158dcbf468" host="localhost" May 14 05:03:44.740925 containerd[1601]: 2025-05-14 05:03:44.708 [INFO][4536] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.134/26] handle="k8s-pod-network.c66d1501764dc68679044f77cb61e5bfdd9e97907cca72763368ef158dcbf468" host="localhost" May 14 05:03:44.740925 containerd[1601]: 2025-05-14 05:03:44.709 [INFO][4536] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. May 14 05:03:44.740925 containerd[1601]: 2025-05-14 05:03:44.709 [INFO][4536] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.134/26] IPv6=[] ContainerID="c66d1501764dc68679044f77cb61e5bfdd9e97907cca72763368ef158dcbf468" HandleID="k8s-pod-network.c66d1501764dc68679044f77cb61e5bfdd9e97907cca72763368ef158dcbf468" Workload="localhost-k8s-csi--node--driver--vwdck-eth0" May 14 05:03:44.742311 containerd[1601]: 2025-05-14 05:03:44.712 [INFO][4505] cni-plugin/k8s.go 386: Populated endpoint ContainerID="c66d1501764dc68679044f77cb61e5bfdd9e97907cca72763368ef158dcbf468" Namespace="calico-system" Pod="csi-node-driver-vwdck" WorkloadEndpoint="localhost-k8s-csi--node--driver--vwdck-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--vwdck-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"f37146d1-125d-4167-a0d9-8ac3149fed78", ResourceVersion:"601", Generation:0, CreationTimestamp:time.Date(2025, time.May, 14, 5, 3, 17, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"5b5cc68cd5", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"csi-node-driver-vwdck", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali6b688c1748b", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 14 05:03:44.742311 containerd[1601]: 2025-05-14 05:03:44.713 [INFO][4505] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.134/32] ContainerID="c66d1501764dc68679044f77cb61e5bfdd9e97907cca72763368ef158dcbf468" Namespace="calico-system" Pod="csi-node-driver-vwdck" WorkloadEndpoint="localhost-k8s-csi--node--driver--vwdck-eth0" May 14 05:03:44.742311 containerd[1601]: 2025-05-14 05:03:44.713 [INFO][4505] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali6b688c1748b ContainerID="c66d1501764dc68679044f77cb61e5bfdd9e97907cca72763368ef158dcbf468" Namespace="calico-system" Pod="csi-node-driver-vwdck" WorkloadEndpoint="localhost-k8s-csi--node--driver--vwdck-eth0" May 14 05:03:44.742311 containerd[1601]: 2025-05-14 05:03:44.719 [INFO][4505] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="c66d1501764dc68679044f77cb61e5bfdd9e97907cca72763368ef158dcbf468" Namespace="calico-system" Pod="csi-node-driver-vwdck" WorkloadEndpoint="localhost-k8s-csi--node--driver--vwdck-eth0" May 14 05:03:44.742311 containerd[1601]: 2025-05-14 05:03:44.720 [INFO][4505] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="c66d1501764dc68679044f77cb61e5bfdd9e97907cca72763368ef158dcbf468" Namespace="calico-system" Pod="csi-node-driver-vwdck" WorkloadEndpoint="localhost-k8s-csi--node--driver--vwdck-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--vwdck-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"f37146d1-125d-4167-a0d9-8ac3149fed78", ResourceVersion:"601", Generation:0, CreationTimestamp:time.Date(2025, time.May, 14, 5, 3, 17, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"5b5cc68cd5", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"c66d1501764dc68679044f77cb61e5bfdd9e97907cca72763368ef158dcbf468", Pod:"csi-node-driver-vwdck", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali6b688c1748b", MAC:"ca:58:e9:6b:a3:0d", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} May 14 05:03:44.742311 containerd[1601]: 2025-05-14 05:03:44.733 [INFO][4505] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="c66d1501764dc68679044f77cb61e5bfdd9e97907cca72763368ef158dcbf468" Namespace="calico-system" Pod="csi-node-driver-vwdck" WorkloadEndpoint="localhost-k8s-csi--node--driver--vwdck-eth0" May 14 05:03:44.752913 containerd[1601]: time="2025-05-14T05:03:44.752870080Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-gkkg2,Uid:d9aa914a-a52c-4b5f-ad56-47d2689175df,Namespace:kube-system,Attempt:0,} returns sandbox id \"e466a24ab4b8cff7e881b82e598ca7cd2983680c6371b4c4b77cecd5f4c21f29\"" May 14 05:03:44.754580 kubelet[2713]: E0514 05:03:44.754368 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:44.758330 containerd[1601]: time="2025-05-14T05:03:44.758266260Z" level=info msg="CreateContainer within sandbox \"e466a24ab4b8cff7e881b82e598ca7cd2983680c6371b4c4b77cecd5f4c21f29\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" May 14 05:03:44.776171 containerd[1601]: time="2025-05-14T05:03:44.776113940Z" level=info msg="Container 57d23e9c00432609d1197d57b86bee48b2e6ba66fc9b56c6f56d95cf2a51501c: CDI devices from CRI Config.CDIDevices: []" May 14 05:03:44.780282 containerd[1601]: time="2025-05-14T05:03:44.780260129Z" level=info msg="connecting to shim c66d1501764dc68679044f77cb61e5bfdd9e97907cca72763368ef158dcbf468" address="unix:///run/containerd/s/d8027fbbbd3a9ebb1eefb9ee09e9bb1c04496910bd5d3000971c0ff5a060a84d" namespace=k8s.io protocol=ttrpc version=3 May 14 05:03:44.784417 containerd[1601]: time="2025-05-14T05:03:44.784350893Z" level=info msg="CreateContainer within sandbox \"e466a24ab4b8cff7e881b82e598ca7cd2983680c6371b4c4b77cecd5f4c21f29\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"57d23e9c00432609d1197d57b86bee48b2e6ba66fc9b56c6f56d95cf2a51501c\"" May 14 05:03:44.785091 containerd[1601]: time="2025-05-14T05:03:44.785052503Z" level=info msg="StartContainer for \"57d23e9c00432609d1197d57b86bee48b2e6ba66fc9b56c6f56d95cf2a51501c\"" May 14 05:03:44.786239 containerd[1601]: time="2025-05-14T05:03:44.786192187Z" level=info msg="connecting to shim 57d23e9c00432609d1197d57b86bee48b2e6ba66fc9b56c6f56d95cf2a51501c" address="unix:///run/containerd/s/c7ae3ceff44861ddd66ca3a6cc8234fafae5f4e475a008d080a0ff29a6245a9c" protocol=ttrpc version=3 May 14 05:03:44.806531 systemd[1]: Started cri-containerd-57d23e9c00432609d1197d57b86bee48b2e6ba66fc9b56c6f56d95cf2a51501c.scope - libcontainer container 57d23e9c00432609d1197d57b86bee48b2e6ba66fc9b56c6f56d95cf2a51501c. May 14 05:03:44.809873 systemd[1]: Started cri-containerd-c66d1501764dc68679044f77cb61e5bfdd9e97907cca72763368ef158dcbf468.scope - libcontainer container c66d1501764dc68679044f77cb61e5bfdd9e97907cca72763368ef158dcbf468. May 14 05:03:44.824617 systemd-resolved[1413]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address May 14 05:03:44.831674 systemd-networkd[1502]: cali09d777db576: Gained IPv6LL May 14 05:03:44.841153 containerd[1601]: time="2025-05-14T05:03:44.841044990Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-vwdck,Uid:f37146d1-125d-4167-a0d9-8ac3149fed78,Namespace:calico-system,Attempt:0,} returns sandbox id \"c66d1501764dc68679044f77cb61e5bfdd9e97907cca72763368ef158dcbf468\"" May 14 05:03:44.852944 containerd[1601]: time="2025-05-14T05:03:44.852903896Z" level=info msg="StartContainer for \"57d23e9c00432609d1197d57b86bee48b2e6ba66fc9b56c6f56d95cf2a51501c\" returns successfully" May 14 05:03:45.004307 kubelet[2713]: E0514 05:03:45.004098 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:45.004307 kubelet[2713]: E0514 05:03:45.004218 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:45.086135 kubelet[2713]: I0514 05:03:45.085956 2713 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-gkkg2" podStartSLOduration=37.085940391 podStartE2EDuration="37.085940391s" podCreationTimestamp="2025-05-14 05:03:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-14 05:03:45.07326316 +0000 UTC m=+41.655750078" watchObservedRunningTime="2025-05-14 05:03:45.085940391 +0000 UTC m=+41.668427309" May 14 05:03:45.442930 containerd[1601]: time="2025-05-14T05:03:45.442851456Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:03:45.443614 containerd[1601]: time="2025-05-14T05:03:45.443582952Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.29.3: active requests=0, bytes read=34789138" May 14 05:03:45.445121 containerd[1601]: time="2025-05-14T05:03:45.445091498Z" level=info msg="ImageCreate event name:\"sha256:4e982138231b3653a012db4f21ed5e7be69afd5f553dba38cf7e88f0ed740b94\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:03:45.447357 containerd[1601]: time="2025-05-14T05:03:45.447308538Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:feaab0197035d474845e0f8137a99a78cab274f0a3cac4d5485cf9b1bdf9ffa9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:03:45.447876 containerd[1601]: time="2025-05-14T05:03:45.447819549Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\" with image id \"sha256:4e982138231b3653a012db4f21ed5e7be69afd5f553dba38cf7e88f0ed740b94\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:feaab0197035d474845e0f8137a99a78cab274f0a3cac4d5485cf9b1bdf9ffa9\", size \"36281728\" in 2.678955921s" May 14 05:03:45.447876 containerd[1601]: time="2025-05-14T05:03:45.447861939Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\" returns image reference \"sha256:4e982138231b3653a012db4f21ed5e7be69afd5f553dba38cf7e88f0ed740b94\"" May 14 05:03:45.448947 containerd[1601]: time="2025-05-14T05:03:45.448859004Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.3\"" May 14 05:03:45.455820 containerd[1601]: time="2025-05-14T05:03:45.455794627Z" level=info msg="CreateContainer within sandbox \"19c60bc325dca1b7b906d84b6bd0ee9fab1743101dd38e240652f54645f5142a\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" May 14 05:03:45.464263 containerd[1601]: time="2025-05-14T05:03:45.464218990Z" level=info msg="Container c384a3b8e16feec06f2be019703423cb2939d797c526cc2963acccf932e3dad8: CDI devices from CRI Config.CDIDevices: []" May 14 05:03:45.471115 containerd[1601]: time="2025-05-14T05:03:45.471082458Z" level=info msg="CreateContainer within sandbox \"19c60bc325dca1b7b906d84b6bd0ee9fab1743101dd38e240652f54645f5142a\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"c384a3b8e16feec06f2be019703423cb2939d797c526cc2963acccf932e3dad8\"" May 14 05:03:45.471440 containerd[1601]: time="2025-05-14T05:03:45.471390527Z" level=info msg="StartContainer for \"c384a3b8e16feec06f2be019703423cb2939d797c526cc2963acccf932e3dad8\"" May 14 05:03:45.472287 containerd[1601]: time="2025-05-14T05:03:45.472262496Z" level=info msg="connecting to shim c384a3b8e16feec06f2be019703423cb2939d797c526cc2963acccf932e3dad8" address="unix:///run/containerd/s/639b5588637fc05f03392a71700e0ed957a7e4b62120821818fecde669678930" protocol=ttrpc version=3 May 14 05:03:45.496530 systemd[1]: Started cri-containerd-c384a3b8e16feec06f2be019703423cb2939d797c526cc2963acccf932e3dad8.scope - libcontainer container c384a3b8e16feec06f2be019703423cb2939d797c526cc2963acccf932e3dad8. May 14 05:03:45.543457 containerd[1601]: time="2025-05-14T05:03:45.543420656Z" level=info msg="StartContainer for \"c384a3b8e16feec06f2be019703423cb2939d797c526cc2963acccf932e3dad8\" returns successfully" May 14 05:03:45.855581 systemd-networkd[1502]: cali6b688c1748b: Gained IPv6LL May 14 05:03:46.007759 kubelet[2713]: E0514 05:03:46.007730 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:46.239623 systemd-networkd[1502]: cali43561bc3405: Gained IPv6LL May 14 05:03:47.009332 kubelet[2713]: I0514 05:03:47.009289 2713 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 14 05:03:47.009808 kubelet[2713]: E0514 05:03:47.009719 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:03:48.391952 containerd[1601]: time="2025-05-14T05:03:48.391887407Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:03:48.392644 containerd[1601]: time="2025-05-14T05:03:48.392589897Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.29.3: active requests=0, bytes read=43021437" May 14 05:03:48.393876 containerd[1601]: time="2025-05-14T05:03:48.393822985Z" level=info msg="ImageCreate event name:\"sha256:b1960e792987d99ee8f3583d7354dcd25a683cf854e8f10322ca7eeb83128532\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:03:48.395695 containerd[1601]: time="2025-05-14T05:03:48.395662482Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:bcb659f25f9aebaa389ed1dbb65edb39478ddf82c57d07d8da474e8cab38d77b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:03:48.396257 containerd[1601]: time="2025-05-14T05:03:48.396211885Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.29.3\" with image id \"sha256:b1960e792987d99ee8f3583d7354dcd25a683cf854e8f10322ca7eeb83128532\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:bcb659f25f9aebaa389ed1dbb65edb39478ddf82c57d07d8da474e8cab38d77b\", size \"44514075\" in 2.947311263s" May 14 05:03:48.396257 containerd[1601]: time="2025-05-14T05:03:48.396254054Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.3\" returns image reference \"sha256:b1960e792987d99ee8f3583d7354dcd25a683cf854e8f10322ca7eeb83128532\"" May 14 05:03:48.399489 containerd[1601]: time="2025-05-14T05:03:48.399128748Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.3\"" May 14 05:03:48.402287 containerd[1601]: time="2025-05-14T05:03:48.402249365Z" level=info msg="CreateContainer within sandbox \"579dde6dd65bb084cc426da55e91b87142056a1acd12ff786be6053f3abc0b02\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" May 14 05:03:48.410952 containerd[1601]: time="2025-05-14T05:03:48.410917869Z" level=info msg="Container 381afb1c398d435da68683be5a968b835ed5cfc9fa02bb7e63b3b48a024897fb: CDI devices from CRI Config.CDIDevices: []" May 14 05:03:48.417852 containerd[1601]: time="2025-05-14T05:03:48.417799324Z" level=info msg="CreateContainer within sandbox \"579dde6dd65bb084cc426da55e91b87142056a1acd12ff786be6053f3abc0b02\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"381afb1c398d435da68683be5a968b835ed5cfc9fa02bb7e63b3b48a024897fb\"" May 14 05:03:48.418412 containerd[1601]: time="2025-05-14T05:03:48.418353235Z" level=info msg="StartContainer for \"381afb1c398d435da68683be5a968b835ed5cfc9fa02bb7e63b3b48a024897fb\"" May 14 05:03:48.419507 containerd[1601]: time="2025-05-14T05:03:48.419481416Z" level=info msg="connecting to shim 381afb1c398d435da68683be5a968b835ed5cfc9fa02bb7e63b3b48a024897fb" address="unix:///run/containerd/s/ad2107fdabc14e9eb59a6aa6e660346b4360665c154b002aef60521db0069244" protocol=ttrpc version=3 May 14 05:03:48.440527 systemd[1]: Started cri-containerd-381afb1c398d435da68683be5a968b835ed5cfc9fa02bb7e63b3b48a024897fb.scope - libcontainer container 381afb1c398d435da68683be5a968b835ed5cfc9fa02bb7e63b3b48a024897fb. May 14 05:03:48.468097 systemd[1]: Started sshd@10-10.0.0.15:22-10.0.0.1:36498.service - OpenSSH per-connection server daemon (10.0.0.1:36498). May 14 05:03:48.493191 containerd[1601]: time="2025-05-14T05:03:48.492616551Z" level=info msg="StartContainer for \"381afb1c398d435da68683be5a968b835ed5cfc9fa02bb7e63b3b48a024897fb\" returns successfully" May 14 05:03:48.528025 sshd[4782]: Accepted publickey for core from 10.0.0.1 port 36498 ssh2: RSA SHA256:9779yoEmBEYtokxLadky4y4rhX8tiwkjz8vWSk0iWXU May 14 05:03:48.529973 sshd-session[4782]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 05:03:48.535450 systemd-logind[1587]: New session 11 of user core. May 14 05:03:48.541594 systemd[1]: Started session-11.scope - Session 11 of User core. May 14 05:03:48.684871 sshd[4799]: Connection closed by 10.0.0.1 port 36498 May 14 05:03:48.685531 sshd-session[4782]: pam_unix(sshd:session): session closed for user core May 14 05:03:48.696091 systemd[1]: sshd@10-10.0.0.15:22-10.0.0.1:36498.service: Deactivated successfully. May 14 05:03:48.698081 systemd[1]: session-11.scope: Deactivated successfully. May 14 05:03:48.699061 systemd-logind[1587]: Session 11 logged out. Waiting for processes to exit. May 14 05:03:48.702742 systemd[1]: Started sshd@11-10.0.0.15:22-10.0.0.1:36500.service - OpenSSH per-connection server daemon (10.0.0.1:36500). May 14 05:03:48.703873 systemd-logind[1587]: Removed session 11. May 14 05:03:48.752156 sshd[4814]: Accepted publickey for core from 10.0.0.1 port 36500 ssh2: RSA SHA256:9779yoEmBEYtokxLadky4y4rhX8tiwkjz8vWSk0iWXU May 14 05:03:48.754272 sshd-session[4814]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 05:03:48.760227 systemd-logind[1587]: New session 12 of user core. May 14 05:03:48.767570 systemd[1]: Started session-12.scope - Session 12 of User core. May 14 05:03:48.904201 containerd[1601]: time="2025-05-14T05:03:48.904126280Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:03:48.905105 containerd[1601]: time="2025-05-14T05:03:48.905000513Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.29.3: active requests=0, bytes read=77" May 14 05:03:48.912812 containerd[1601]: time="2025-05-14T05:03:48.912699986Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.29.3\" with image id \"sha256:b1960e792987d99ee8f3583d7354dcd25a683cf854e8f10322ca7eeb83128532\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:bcb659f25f9aebaa389ed1dbb65edb39478ddf82c57d07d8da474e8cab38d77b\", size \"44514075\" in 513.505404ms" May 14 05:03:48.912812 containerd[1601]: time="2025-05-14T05:03:48.912753086Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.3\" returns image reference \"sha256:b1960e792987d99ee8f3583d7354dcd25a683cf854e8f10322ca7eeb83128532\"" May 14 05:03:48.916680 containerd[1601]: time="2025-05-14T05:03:48.916633460Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.3\"" May 14 05:03:48.920692 containerd[1601]: time="2025-05-14T05:03:48.920044913Z" level=info msg="CreateContainer within sandbox \"a2a9caea92810841b3fc4c0f6a1be62821bd645aa7f46186755f0176c401ea39\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" May 14 05:03:48.972657 sshd[4816]: Connection closed by 10.0.0.1 port 36500 May 14 05:03:48.974059 sshd-session[4814]: pam_unix(sshd:session): session closed for user core May 14 05:03:48.985219 systemd[1]: sshd@11-10.0.0.15:22-10.0.0.1:36500.service: Deactivated successfully. May 14 05:03:48.987476 systemd[1]: session-12.scope: Deactivated successfully. May 14 05:03:48.988324 systemd-logind[1587]: Session 12 logged out. Waiting for processes to exit. May 14 05:03:48.994577 systemd[1]: Started sshd@12-10.0.0.15:22-10.0.0.1:36510.service - OpenSSH per-connection server daemon (10.0.0.1:36510). May 14 05:03:48.996553 systemd-logind[1587]: Removed session 12. May 14 05:03:49.044005 sshd[4827]: Accepted publickey for core from 10.0.0.1 port 36510 ssh2: RSA SHA256:9779yoEmBEYtokxLadky4y4rhX8tiwkjz8vWSk0iWXU May 14 05:03:49.045648 sshd-session[4827]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 05:03:49.049977 systemd-logind[1587]: New session 13 of user core. May 14 05:03:49.061524 systemd[1]: Started session-13.scope - Session 13 of User core. May 14 05:03:49.254937 kubelet[2713]: I0514 05:03:49.253917 2713 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-bbffb6fd-kw4jm" podStartSLOduration=26.831037807 podStartE2EDuration="32.253828144s" podCreationTimestamp="2025-05-14 05:03:17 +0000 UTC" firstStartedPulling="2025-05-14 05:03:42.975853229 +0000 UTC m=+39.558340147" lastFinishedPulling="2025-05-14 05:03:48.398643566 +0000 UTC m=+44.981130484" observedRunningTime="2025-05-14 05:03:49.253555291 +0000 UTC m=+45.836042209" watchObservedRunningTime="2025-05-14 05:03:49.253828144 +0000 UTC m=+45.836315052" May 14 05:03:49.255420 kubelet[2713]: I0514 05:03:49.255003 2713 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-6479477dcb-f59jw" podStartSLOduration=29.57282442 podStartE2EDuration="32.254018952s" podCreationTimestamp="2025-05-14 05:03:17 +0000 UTC" firstStartedPulling="2025-05-14 05:03:42.767522996 +0000 UTC m=+39.350009914" lastFinishedPulling="2025-05-14 05:03:45.448717528 +0000 UTC m=+42.031204446" observedRunningTime="2025-05-14 05:03:46.018117754 +0000 UTC m=+42.600604692" watchObservedRunningTime="2025-05-14 05:03:49.254018952 +0000 UTC m=+45.836505871" May 14 05:03:49.269627 sshd[4829]: Connection closed by 10.0.0.1 port 36510 May 14 05:03:49.269967 sshd-session[4827]: pam_unix(sshd:session): session closed for user core May 14 05:03:49.274776 systemd[1]: sshd@12-10.0.0.15:22-10.0.0.1:36510.service: Deactivated successfully. May 14 05:03:49.277075 systemd[1]: session-13.scope: Deactivated successfully. May 14 05:03:49.279502 systemd-logind[1587]: Session 13 logged out. Waiting for processes to exit. May 14 05:03:49.281423 systemd-logind[1587]: Removed session 13. May 14 05:03:49.296565 containerd[1601]: time="2025-05-14T05:03:49.296515263Z" level=info msg="Container ab2240d8f7c1c08471fe9f8cf2315bdaa7b63806c8e0d9cd6b79f182f171a88b: CDI devices from CRI Config.CDIDevices: []" May 14 05:03:49.501865 containerd[1601]: time="2025-05-14T05:03:49.501822385Z" level=info msg="CreateContainer within sandbox \"a2a9caea92810841b3fc4c0f6a1be62821bd645aa7f46186755f0176c401ea39\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"ab2240d8f7c1c08471fe9f8cf2315bdaa7b63806c8e0d9cd6b79f182f171a88b\"" May 14 05:03:49.502575 containerd[1601]: time="2025-05-14T05:03:49.502562486Z" level=info msg="StartContainer for \"ab2240d8f7c1c08471fe9f8cf2315bdaa7b63806c8e0d9cd6b79f182f171a88b\"" May 14 05:03:49.503862 containerd[1601]: time="2025-05-14T05:03:49.503829849Z" level=info msg="connecting to shim ab2240d8f7c1c08471fe9f8cf2315bdaa7b63806c8e0d9cd6b79f182f171a88b" address="unix:///run/containerd/s/d974314019a4dae73031b984ebb955bb10145480262a9cc997af1ae274ec7a62" protocol=ttrpc version=3 May 14 05:03:49.522573 systemd[1]: Started cri-containerd-ab2240d8f7c1c08471fe9f8cf2315bdaa7b63806c8e0d9cd6b79f182f171a88b.scope - libcontainer container ab2240d8f7c1c08471fe9f8cf2315bdaa7b63806c8e0d9cd6b79f182f171a88b. May 14 05:03:49.611496 containerd[1601]: time="2025-05-14T05:03:49.611454517Z" level=info msg="StartContainer for \"ab2240d8f7c1c08471fe9f8cf2315bdaa7b63806c8e0d9cd6b79f182f171a88b\" returns successfully" May 14 05:03:50.021315 kubelet[2713]: I0514 05:03:50.021275 2713 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 14 05:03:50.032189 kubelet[2713]: I0514 05:03:50.031677 2713 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-bbffb6fd-5t7tc" podStartSLOduration=27.262215773 podStartE2EDuration="33.031657314s" podCreationTimestamp="2025-05-14 05:03:17 +0000 UTC" firstStartedPulling="2025-05-14 05:03:43.1460428 +0000 UTC m=+39.728529718" lastFinishedPulling="2025-05-14 05:03:48.915484341 +0000 UTC m=+45.497971259" observedRunningTime="2025-05-14 05:03:50.031340449 +0000 UTC m=+46.613827387" watchObservedRunningTime="2025-05-14 05:03:50.031657314 +0000 UTC m=+46.614144232" May 14 05:03:50.400200 kubelet[2713]: I0514 05:03:50.399927 2713 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 14 05:03:50.450961 containerd[1601]: time="2025-05-14T05:03:50.450900757Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c384a3b8e16feec06f2be019703423cb2939d797c526cc2963acccf932e3dad8\" id:\"a0824b4ec6835d1b5a3903c6d32cccc53ac8704d331a4e89437dbb2bf2a3c0e7\" pid:4899 exited_at:{seconds:1747199030 nanos:450339763}" May 14 05:03:50.499349 containerd[1601]: time="2025-05-14T05:03:50.499310341Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c384a3b8e16feec06f2be019703423cb2939d797c526cc2963acccf932e3dad8\" id:\"4cdcb3d8e0f3f3db745f8218d96b54209ac31a58828b5df16122a59c55c662b9\" pid:4924 exited_at:{seconds:1747199030 nanos:499086070}" May 14 05:03:50.819439 containerd[1601]: time="2025-05-14T05:03:50.819305481Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:03:50.820274 containerd[1601]: time="2025-05-14T05:03:50.820238385Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.29.3: active requests=0, bytes read=7912898" May 14 05:03:50.821412 containerd[1601]: time="2025-05-14T05:03:50.821364872Z" level=info msg="ImageCreate event name:\"sha256:4c37db5645f4075f8b8170eea8f14e340cb13550e0a392962f1f211ded741505\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:03:50.823367 containerd[1601]: time="2025-05-14T05:03:50.823272486Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:72455a36febc7c56ec8881007f4805caed5764026a0694e4f86a2503209b2d31\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:03:50.823798 containerd[1601]: time="2025-05-14T05:03:50.823751065Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.29.3\" with image id \"sha256:4c37db5645f4075f8b8170eea8f14e340cb13550e0a392962f1f211ded741505\", repo tag \"ghcr.io/flatcar/calico/csi:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:72455a36febc7c56ec8881007f4805caed5764026a0694e4f86a2503209b2d31\", size \"9405520\" in 1.907066009s" May 14 05:03:50.825348 containerd[1601]: time="2025-05-14T05:03:50.823801510Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.3\" returns image reference \"sha256:4c37db5645f4075f8b8170eea8f14e340cb13550e0a392962f1f211ded741505\"" May 14 05:03:50.829276 containerd[1601]: time="2025-05-14T05:03:50.829250501Z" level=info msg="CreateContainer within sandbox \"c66d1501764dc68679044f77cb61e5bfdd9e97907cca72763368ef158dcbf468\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" May 14 05:03:50.842123 containerd[1601]: time="2025-05-14T05:03:50.842081162Z" level=info msg="Container bf83528653687d27fe30f6ee91eb39a7c6b7bfd7618695384f791c3bc30f38bb: CDI devices from CRI Config.CDIDevices: []" May 14 05:03:50.867557 containerd[1601]: time="2025-05-14T05:03:50.867504720Z" level=info msg="CreateContainer within sandbox \"c66d1501764dc68679044f77cb61e5bfdd9e97907cca72763368ef158dcbf468\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"bf83528653687d27fe30f6ee91eb39a7c6b7bfd7618695384f791c3bc30f38bb\"" May 14 05:03:50.867978 containerd[1601]: time="2025-05-14T05:03:50.867951890Z" level=info msg="StartContainer for \"bf83528653687d27fe30f6ee91eb39a7c6b7bfd7618695384f791c3bc30f38bb\"" May 14 05:03:50.869267 containerd[1601]: time="2025-05-14T05:03:50.869240582Z" level=info msg="connecting to shim bf83528653687d27fe30f6ee91eb39a7c6b7bfd7618695384f791c3bc30f38bb" address="unix:///run/containerd/s/d8027fbbbd3a9ebb1eefb9ee09e9bb1c04496910bd5d3000971c0ff5a060a84d" protocol=ttrpc version=3 May 14 05:03:50.907608 systemd[1]: Started cri-containerd-bf83528653687d27fe30f6ee91eb39a7c6b7bfd7618695384f791c3bc30f38bb.scope - libcontainer container bf83528653687d27fe30f6ee91eb39a7c6b7bfd7618695384f791c3bc30f38bb. May 14 05:03:51.036191 containerd[1601]: time="2025-05-14T05:03:51.036126975Z" level=info msg="StartContainer for \"bf83528653687d27fe30f6ee91eb39a7c6b7bfd7618695384f791c3bc30f38bb\" returns successfully" May 14 05:03:51.037624 containerd[1601]: time="2025-05-14T05:03:51.037376513Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\"" May 14 05:03:51.037766 kubelet[2713]: I0514 05:03:51.037619 2713 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 14 05:03:52.767243 containerd[1601]: time="2025-05-14T05:03:52.767187808Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:03:52.767926 containerd[1601]: time="2025-05-14T05:03:52.767900036Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3: active requests=0, bytes read=13991773" May 14 05:03:52.769044 containerd[1601]: time="2025-05-14T05:03:52.769012617Z" level=info msg="ImageCreate event name:\"sha256:e909e2ccf54404290b577fbddd190d036984deed184001767f820b0dddf77fd9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:03:52.770986 containerd[1601]: time="2025-05-14T05:03:52.770932314Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:3f15090a9bb45773d1fd019455ec3d3f3746f3287c35d8013e497b38d8237324\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 14 05:03:52.771353 containerd[1601]: time="2025-05-14T05:03:52.771315844Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\" with image id \"sha256:e909e2ccf54404290b577fbddd190d036984deed184001767f820b0dddf77fd9\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:3f15090a9bb45773d1fd019455ec3d3f3746f3287c35d8013e497b38d8237324\", size \"15484347\" in 1.733876894s" May 14 05:03:52.771388 containerd[1601]: time="2025-05-14T05:03:52.771351251Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\" returns image reference \"sha256:e909e2ccf54404290b577fbddd190d036984deed184001767f820b0dddf77fd9\"" May 14 05:03:52.773330 containerd[1601]: time="2025-05-14T05:03:52.773025486Z" level=info msg="CreateContainer within sandbox \"c66d1501764dc68679044f77cb61e5bfdd9e97907cca72763368ef158dcbf468\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" May 14 05:03:52.780043 containerd[1601]: time="2025-05-14T05:03:52.779994941Z" level=info msg="Container e2508e50c7414ffa2925acc434ebfe0a8283d5e9ad32d2ab81679b430050e71a: CDI devices from CRI Config.CDIDevices: []" May 14 05:03:52.790157 containerd[1601]: time="2025-05-14T05:03:52.790116388Z" level=info msg="CreateContainer within sandbox \"c66d1501764dc68679044f77cb61e5bfdd9e97907cca72763368ef158dcbf468\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"e2508e50c7414ffa2925acc434ebfe0a8283d5e9ad32d2ab81679b430050e71a\"" May 14 05:03:52.790477 containerd[1601]: time="2025-05-14T05:03:52.790428984Z" level=info msg="StartContainer for \"e2508e50c7414ffa2925acc434ebfe0a8283d5e9ad32d2ab81679b430050e71a\"" May 14 05:03:52.791659 containerd[1601]: time="2025-05-14T05:03:52.791634549Z" level=info msg="connecting to shim e2508e50c7414ffa2925acc434ebfe0a8283d5e9ad32d2ab81679b430050e71a" address="unix:///run/containerd/s/d8027fbbbd3a9ebb1eefb9ee09e9bb1c04496910bd5d3000971c0ff5a060a84d" protocol=ttrpc version=3 May 14 05:03:52.825536 systemd[1]: Started cri-containerd-e2508e50c7414ffa2925acc434ebfe0a8283d5e9ad32d2ab81679b430050e71a.scope - libcontainer container e2508e50c7414ffa2925acc434ebfe0a8283d5e9ad32d2ab81679b430050e71a. May 14 05:03:52.866781 containerd[1601]: time="2025-05-14T05:03:52.866737795Z" level=info msg="StartContainer for \"e2508e50c7414ffa2925acc434ebfe0a8283d5e9ad32d2ab81679b430050e71a\" returns successfully" May 14 05:03:52.977647 kubelet[2713]: I0514 05:03:52.977605 2713 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 May 14 05:03:52.977647 kubelet[2713]: I0514 05:03:52.977645 2713 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock May 14 05:03:53.054149 kubelet[2713]: I0514 05:03:53.053723 2713 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-vwdck" podStartSLOduration=28.124177598 podStartE2EDuration="36.053692559s" podCreationTimestamp="2025-05-14 05:03:17 +0000 UTC" firstStartedPulling="2025-05-14 05:03:44.842433532 +0000 UTC m=+41.424920440" lastFinishedPulling="2025-05-14 05:03:52.771948483 +0000 UTC m=+49.354435401" observedRunningTime="2025-05-14 05:03:53.05313412 +0000 UTC m=+49.635621058" watchObservedRunningTime="2025-05-14 05:03:53.053692559 +0000 UTC m=+49.636179477" May 14 05:03:54.285968 systemd[1]: Started sshd@13-10.0.0.15:22-10.0.0.1:36522.service - OpenSSH per-connection server daemon (10.0.0.1:36522). May 14 05:03:54.343132 sshd[5014]: Accepted publickey for core from 10.0.0.1 port 36522 ssh2: RSA SHA256:9779yoEmBEYtokxLadky4y4rhX8tiwkjz8vWSk0iWXU May 14 05:03:54.344987 sshd-session[5014]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 05:03:54.349943 systemd-logind[1587]: New session 14 of user core. May 14 05:03:54.358549 systemd[1]: Started session-14.scope - Session 14 of User core. May 14 05:03:54.480840 sshd[5017]: Connection closed by 10.0.0.1 port 36522 May 14 05:03:54.481153 sshd-session[5014]: pam_unix(sshd:session): session closed for user core May 14 05:03:54.485848 systemd[1]: sshd@13-10.0.0.15:22-10.0.0.1:36522.service: Deactivated successfully. May 14 05:03:54.488143 systemd[1]: session-14.scope: Deactivated successfully. May 14 05:03:54.489036 systemd-logind[1587]: Session 14 logged out. Waiting for processes to exit. May 14 05:03:54.490690 systemd-logind[1587]: Removed session 14. May 14 05:03:59.497579 systemd[1]: Started sshd@14-10.0.0.15:22-10.0.0.1:43118.service - OpenSSH per-connection server daemon (10.0.0.1:43118). May 14 05:03:59.554505 sshd[5032]: Accepted publickey for core from 10.0.0.1 port 43118 ssh2: RSA SHA256:9779yoEmBEYtokxLadky4y4rhX8tiwkjz8vWSk0iWXU May 14 05:03:59.555870 sshd-session[5032]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 05:03:59.559793 systemd-logind[1587]: New session 15 of user core. May 14 05:03:59.568541 systemd[1]: Started session-15.scope - Session 15 of User core. May 14 05:03:59.681950 sshd[5034]: Connection closed by 10.0.0.1 port 43118 May 14 05:03:59.682268 sshd-session[5032]: pam_unix(sshd:session): session closed for user core May 14 05:03:59.686293 systemd[1]: sshd@14-10.0.0.15:22-10.0.0.1:43118.service: Deactivated successfully. May 14 05:03:59.688557 systemd[1]: session-15.scope: Deactivated successfully. May 14 05:03:59.689380 systemd-logind[1587]: Session 15 logged out. Waiting for processes to exit. May 14 05:03:59.690892 systemd-logind[1587]: Removed session 15. May 14 05:04:04.706363 systemd[1]: Started sshd@15-10.0.0.15:22-10.0.0.1:43120.service - OpenSSH per-connection server daemon (10.0.0.1:43120). May 14 05:04:04.760955 sshd[5056]: Accepted publickey for core from 10.0.0.1 port 43120 ssh2: RSA SHA256:9779yoEmBEYtokxLadky4y4rhX8tiwkjz8vWSk0iWXU May 14 05:04:04.762622 sshd-session[5056]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 05:04:04.767642 systemd-logind[1587]: New session 16 of user core. May 14 05:04:04.780542 systemd[1]: Started session-16.scope - Session 16 of User core. May 14 05:04:04.895278 sshd[5060]: Connection closed by 10.0.0.1 port 43120 May 14 05:04:04.895565 sshd-session[5056]: pam_unix(sshd:session): session closed for user core May 14 05:04:04.900652 systemd[1]: sshd@15-10.0.0.15:22-10.0.0.1:43120.service: Deactivated successfully. May 14 05:04:04.904858 systemd[1]: session-16.scope: Deactivated successfully. May 14 05:04:04.907038 systemd-logind[1587]: Session 16 logged out. Waiting for processes to exit. May 14 05:04:04.908279 systemd-logind[1587]: Removed session 16. May 14 05:04:09.913532 systemd[1]: Started sshd@16-10.0.0.15:22-10.0.0.1:56910.service - OpenSSH per-connection server daemon (10.0.0.1:56910). May 14 05:04:09.964501 sshd[5074]: Accepted publickey for core from 10.0.0.1 port 56910 ssh2: RSA SHA256:9779yoEmBEYtokxLadky4y4rhX8tiwkjz8vWSk0iWXU May 14 05:04:09.965973 sshd-session[5074]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 05:04:09.970264 systemd-logind[1587]: New session 17 of user core. May 14 05:04:09.976521 systemd[1]: Started session-17.scope - Session 17 of User core. May 14 05:04:10.089101 sshd[5078]: Connection closed by 10.0.0.1 port 56910 May 14 05:04:10.089522 sshd-session[5074]: pam_unix(sshd:session): session closed for user core May 14 05:04:10.102524 systemd[1]: sshd@16-10.0.0.15:22-10.0.0.1:56910.service: Deactivated successfully. May 14 05:04:10.105070 systemd[1]: session-17.scope: Deactivated successfully. May 14 05:04:10.105886 systemd-logind[1587]: Session 17 logged out. Waiting for processes to exit. May 14 05:04:10.109192 systemd[1]: Started sshd@17-10.0.0.15:22-10.0.0.1:56914.service - OpenSSH per-connection server daemon (10.0.0.1:56914). May 14 05:04:10.110261 systemd-logind[1587]: Removed session 17. May 14 05:04:10.166888 sshd[5091]: Accepted publickey for core from 10.0.0.1 port 56914 ssh2: RSA SHA256:9779yoEmBEYtokxLadky4y4rhX8tiwkjz8vWSk0iWXU May 14 05:04:10.168422 sshd-session[5091]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 05:04:10.173240 systemd-logind[1587]: New session 18 of user core. May 14 05:04:10.180532 systemd[1]: Started session-18.scope - Session 18 of User core. May 14 05:04:10.490360 sshd[5093]: Connection closed by 10.0.0.1 port 56914 May 14 05:04:10.491087 sshd-session[5091]: pam_unix(sshd:session): session closed for user core May 14 05:04:10.499220 systemd[1]: sshd@17-10.0.0.15:22-10.0.0.1:56914.service: Deactivated successfully. May 14 05:04:10.501158 systemd[1]: session-18.scope: Deactivated successfully. May 14 05:04:10.502020 systemd-logind[1587]: Session 18 logged out. Waiting for processes to exit. May 14 05:04:10.504827 systemd[1]: Started sshd@18-10.0.0.15:22-10.0.0.1:56930.service - OpenSSH per-connection server daemon (10.0.0.1:56930). May 14 05:04:10.505525 systemd-logind[1587]: Removed session 18. May 14 05:04:10.555971 sshd[5105]: Accepted publickey for core from 10.0.0.1 port 56930 ssh2: RSA SHA256:9779yoEmBEYtokxLadky4y4rhX8tiwkjz8vWSk0iWXU May 14 05:04:10.557309 sshd-session[5105]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 05:04:10.561679 systemd-logind[1587]: New session 19 of user core. May 14 05:04:10.569556 systemd[1]: Started session-19.scope - Session 19 of User core. May 14 05:04:10.615333 containerd[1601]: time="2025-05-14T05:04:10.615284753Z" level=info msg="TaskExit event in podsandbox handler container_id:\"02ddbce4998bf831dc35ac719c47bc7fdbc2458ddad1416a7fb98143145b3a0e\" id:\"4859dbcd3f62bd4eeb1d832063ae5c8fc5070fdd6fa94459599720f14d68033e\" pid:5121 exited_at:{seconds:1747199050 nanos:614914147}" May 14 05:04:10.629252 kubelet[2713]: E0514 05:04:10.629127 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:04:11.591120 sshd[5117]: Connection closed by 10.0.0.1 port 56930 May 14 05:04:11.591564 sshd-session[5105]: pam_unix(sshd:session): session closed for user core May 14 05:04:11.605068 systemd[1]: sshd@18-10.0.0.15:22-10.0.0.1:56930.service: Deactivated successfully. May 14 05:04:11.607182 systemd[1]: session-19.scope: Deactivated successfully. May 14 05:04:11.607959 systemd-logind[1587]: Session 19 logged out. Waiting for processes to exit. May 14 05:04:11.611572 systemd[1]: Started sshd@19-10.0.0.15:22-10.0.0.1:56940.service - OpenSSH per-connection server daemon (10.0.0.1:56940). May 14 05:04:11.612189 systemd-logind[1587]: Removed session 19. May 14 05:04:11.665224 sshd[5153]: Accepted publickey for core from 10.0.0.1 port 56940 ssh2: RSA SHA256:9779yoEmBEYtokxLadky4y4rhX8tiwkjz8vWSk0iWXU May 14 05:04:11.666935 sshd-session[5153]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 05:04:11.671615 systemd-logind[1587]: New session 20 of user core. May 14 05:04:11.683561 systemd[1]: Started session-20.scope - Session 20 of User core. May 14 05:04:11.924016 sshd[5155]: Connection closed by 10.0.0.1 port 56940 May 14 05:04:11.924302 sshd-session[5153]: pam_unix(sshd:session): session closed for user core May 14 05:04:11.934178 systemd[1]: sshd@19-10.0.0.15:22-10.0.0.1:56940.service: Deactivated successfully. May 14 05:04:11.936253 systemd[1]: session-20.scope: Deactivated successfully. May 14 05:04:11.936973 systemd-logind[1587]: Session 20 logged out. Waiting for processes to exit. May 14 05:04:11.939779 systemd[1]: Started sshd@20-10.0.0.15:22-10.0.0.1:56952.service - OpenSSH per-connection server daemon (10.0.0.1:56952). May 14 05:04:11.940422 systemd-logind[1587]: Removed session 20. May 14 05:04:11.990475 sshd[5166]: Accepted publickey for core from 10.0.0.1 port 56952 ssh2: RSA SHA256:9779yoEmBEYtokxLadky4y4rhX8tiwkjz8vWSk0iWXU May 14 05:04:11.992054 sshd-session[5166]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 05:04:11.996365 systemd-logind[1587]: New session 21 of user core. May 14 05:04:12.007516 systemd[1]: Started session-21.scope - Session 21 of User core. May 14 05:04:12.124957 sshd[5168]: Connection closed by 10.0.0.1 port 56952 May 14 05:04:12.125263 sshd-session[5166]: pam_unix(sshd:session): session closed for user core May 14 05:04:12.129458 systemd[1]: sshd@20-10.0.0.15:22-10.0.0.1:56952.service: Deactivated successfully. May 14 05:04:12.131299 systemd[1]: session-21.scope: Deactivated successfully. May 14 05:04:12.132094 systemd-logind[1587]: Session 21 logged out. Waiting for processes to exit. May 14 05:04:12.133329 systemd-logind[1587]: Removed session 21. May 14 05:04:17.138027 systemd[1]: Started sshd@21-10.0.0.15:22-10.0.0.1:56964.service - OpenSSH per-connection server daemon (10.0.0.1:56964). May 14 05:04:17.190850 sshd[5182]: Accepted publickey for core from 10.0.0.1 port 56964 ssh2: RSA SHA256:9779yoEmBEYtokxLadky4y4rhX8tiwkjz8vWSk0iWXU May 14 05:04:17.192070 sshd-session[5182]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 05:04:17.196086 systemd-logind[1587]: New session 22 of user core. May 14 05:04:17.207505 systemd[1]: Started session-22.scope - Session 22 of User core. May 14 05:04:17.312694 sshd[5184]: Connection closed by 10.0.0.1 port 56964 May 14 05:04:17.313479 sshd-session[5182]: pam_unix(sshd:session): session closed for user core May 14 05:04:17.316383 systemd[1]: sshd@21-10.0.0.15:22-10.0.0.1:56964.service: Deactivated successfully. May 14 05:04:17.318760 systemd[1]: session-22.scope: Deactivated successfully. May 14 05:04:17.320596 systemd-logind[1587]: Session 22 logged out. Waiting for processes to exit. May 14 05:04:17.322256 systemd-logind[1587]: Removed session 22. May 14 05:04:18.489728 kubelet[2713]: E0514 05:04:18.489676 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:04:19.538361 kubelet[2713]: I0514 05:04:19.538299 2713 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 14 05:04:20.496563 containerd[1601]: time="2025-05-14T05:04:20.496505461Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c384a3b8e16feec06f2be019703423cb2939d797c526cc2963acccf932e3dad8\" id:\"dfe4a52de03be114c47fa1b70b3f76ddebbfdcd40a6deb1f990df76e4b6b07c5\" pid:5212 exited_at:{seconds:1747199060 nanos:496194331}" May 14 05:04:22.334945 systemd[1]: Started sshd@22-10.0.0.15:22-10.0.0.1:56446.service - OpenSSH per-connection server daemon (10.0.0.1:56446). May 14 05:04:22.393943 sshd[5229]: Accepted publickey for core from 10.0.0.1 port 56446 ssh2: RSA SHA256:9779yoEmBEYtokxLadky4y4rhX8tiwkjz8vWSk0iWXU May 14 05:04:22.395424 sshd-session[5229]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 05:04:22.399760 systemd-logind[1587]: New session 23 of user core. May 14 05:04:22.409593 systemd[1]: Started session-23.scope - Session 23 of User core. May 14 05:04:22.490032 kubelet[2713]: E0514 05:04:22.489986 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:04:22.530025 sshd[5231]: Connection closed by 10.0.0.1 port 56446 May 14 05:04:22.530317 sshd-session[5229]: pam_unix(sshd:session): session closed for user core May 14 05:04:22.535066 systemd[1]: sshd@22-10.0.0.15:22-10.0.0.1:56446.service: Deactivated successfully. May 14 05:04:22.537168 systemd[1]: session-23.scope: Deactivated successfully. May 14 05:04:22.537890 systemd-logind[1587]: Session 23 logged out. Waiting for processes to exit. May 14 05:04:22.539346 systemd-logind[1587]: Removed session 23. May 14 05:04:23.269677 kubelet[2713]: I0514 05:04:23.269632 2713 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 14 05:04:27.542466 systemd[1]: Started sshd@23-10.0.0.15:22-10.0.0.1:56452.service - OpenSSH per-connection server daemon (10.0.0.1:56452). May 14 05:04:27.600414 sshd[5248]: Accepted publickey for core from 10.0.0.1 port 56452 ssh2: RSA SHA256:9779yoEmBEYtokxLadky4y4rhX8tiwkjz8vWSk0iWXU May 14 05:04:27.602141 sshd-session[5248]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 05:04:27.606840 systemd-logind[1587]: New session 24 of user core. May 14 05:04:27.611519 systemd[1]: Started session-24.scope - Session 24 of User core. May 14 05:04:27.787284 sshd[5250]: Connection closed by 10.0.0.1 port 56452 May 14 05:04:27.787631 sshd-session[5248]: pam_unix(sshd:session): session closed for user core May 14 05:04:27.792082 systemd[1]: sshd@23-10.0.0.15:22-10.0.0.1:56452.service: Deactivated successfully. May 14 05:04:27.793986 systemd[1]: session-24.scope: Deactivated successfully. May 14 05:04:27.794870 systemd-logind[1587]: Session 24 logged out. Waiting for processes to exit. May 14 05:04:27.795922 systemd-logind[1587]: Removed session 24. May 14 05:04:28.490289 kubelet[2713]: E0514 05:04:28.490235 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:04:30.490078 kubelet[2713]: E0514 05:04:30.490033 2713 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" May 14 05:04:32.804795 systemd[1]: Started sshd@24-10.0.0.15:22-10.0.0.1:51198.service - OpenSSH per-connection server daemon (10.0.0.1:51198). May 14 05:04:32.854894 sshd[5263]: Accepted publickey for core from 10.0.0.1 port 51198 ssh2: RSA SHA256:9779yoEmBEYtokxLadky4y4rhX8tiwkjz8vWSk0iWXU May 14 05:04:32.856671 sshd-session[5263]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 14 05:04:32.861617 systemd-logind[1587]: New session 25 of user core. May 14 05:04:32.870606 systemd[1]: Started session-25.scope - Session 25 of User core. May 14 05:04:32.987974 sshd[5265]: Connection closed by 10.0.0.1 port 51198 May 14 05:04:32.988307 sshd-session[5263]: pam_unix(sshd:session): session closed for user core May 14 05:04:32.992892 systemd[1]: sshd@24-10.0.0.15:22-10.0.0.1:51198.service: Deactivated successfully. May 14 05:04:32.995766 systemd[1]: session-25.scope: Deactivated successfully. May 14 05:04:32.996689 systemd-logind[1587]: Session 25 logged out. Waiting for processes to exit. May 14 05:04:32.998517 systemd-logind[1587]: Removed session 25. May 14 05:04:34.009318 containerd[1601]: time="2025-05-14T05:04:34.009269523Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c384a3b8e16feec06f2be019703423cb2939d797c526cc2963acccf932e3dad8\" id:\"8f97d128508136d2c5959f47de889064650fe716261cf7adac654dc195d0becf\" pid:5290 exited_at:{seconds:1747199074 nanos:9036557}"