Jul 10 00:25:36.721021 kernel: Linux version 6.12.36-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.2.1_p20241221 p7) 14.2.1 20241221, GNU ld (Gentoo 2.44 p1) 2.44.0) #1 SMP PREEMPT_DYNAMIC Wed Jul 9 22:15:30 -00 2025 Jul 10 00:25:36.721038 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=vmware flatcar.autologin verity.usrhash=844005237fb9709f65a093d5533c4229fb6c54e8e257736d9c3d041b6d3080ea Jul 10 00:25:36.721044 kernel: Disabled fast string operations Jul 10 00:25:36.721048 kernel: BIOS-provided physical RAM map: Jul 10 00:25:36.721052 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ebff] usable Jul 10 00:25:36.721056 kernel: BIOS-e820: [mem 0x000000000009ec00-0x000000000009ffff] reserved Jul 10 00:25:36.721062 kernel: BIOS-e820: [mem 0x00000000000dc000-0x00000000000fffff] reserved Jul 10 00:25:36.721066 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007fedffff] usable Jul 10 00:25:36.721070 kernel: BIOS-e820: [mem 0x000000007fee0000-0x000000007fefefff] ACPI data Jul 10 00:25:36.721074 kernel: BIOS-e820: [mem 0x000000007feff000-0x000000007fefffff] ACPI NVS Jul 10 00:25:36.721078 kernel: BIOS-e820: [mem 0x000000007ff00000-0x000000007fffffff] usable Jul 10 00:25:36.721083 kernel: BIOS-e820: [mem 0x00000000f0000000-0x00000000f7ffffff] reserved Jul 10 00:25:36.721087 kernel: BIOS-e820: [mem 0x00000000fec00000-0x00000000fec0ffff] reserved Jul 10 00:25:36.721091 kernel: BIOS-e820: [mem 0x00000000fee00000-0x00000000fee00fff] reserved Jul 10 00:25:36.721097 kernel: BIOS-e820: [mem 0x00000000fffe0000-0x00000000ffffffff] reserved Jul 10 00:25:36.721102 kernel: NX (Execute Disable) protection: active Jul 10 00:25:36.721106 kernel: APIC: Static calls initialized Jul 10 00:25:36.721111 kernel: SMBIOS 2.7 present. Jul 10 00:25:36.721116 kernel: DMI: VMware, Inc. VMware Virtual Platform/440BX Desktop Reference Platform, BIOS 6.00 05/28/2020 Jul 10 00:25:36.721121 kernel: DMI: Memory slots populated: 1/128 Jul 10 00:25:36.721126 kernel: vmware: hypercall mode: 0x00 Jul 10 00:25:36.721131 kernel: Hypervisor detected: VMware Jul 10 00:25:36.721135 kernel: vmware: TSC freq read from hypervisor : 3408.000 MHz Jul 10 00:25:36.721140 kernel: vmware: Host bus clock speed read from hypervisor : 66000000 Hz Jul 10 00:25:36.721145 kernel: vmware: using clock offset of 3357061072 ns Jul 10 00:25:36.721149 kernel: tsc: Detected 3408.000 MHz processor Jul 10 00:25:36.721154 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Jul 10 00:25:36.721160 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Jul 10 00:25:36.721165 kernel: last_pfn = 0x80000 max_arch_pfn = 0x400000000 Jul 10 00:25:36.721169 kernel: total RAM covered: 3072M Jul 10 00:25:36.721175 kernel: Found optimal setting for mtrr clean up Jul 10 00:25:36.721182 kernel: gran_size: 64K chunk_size: 64K num_reg: 2 lose cover RAM: 0G Jul 10 00:25:36.721187 kernel: MTRR map: 6 entries (5 fixed + 1 variable; max 21), built from 8 variable MTRRs Jul 10 00:25:36.721192 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Jul 10 00:25:36.721197 kernel: Using GB pages for direct mapping Jul 10 00:25:36.721202 kernel: ACPI: Early table checksum verification disabled Jul 10 00:25:36.721206 kernel: ACPI: RSDP 0x00000000000F6A00 000024 (v02 PTLTD ) Jul 10 00:25:36.721211 kernel: ACPI: XSDT 0x000000007FEE965B 00005C (v01 INTEL 440BX 06040000 VMW 01324272) Jul 10 00:25:36.721216 kernel: ACPI: FACP 0x000000007FEFEE73 0000F4 (v04 INTEL 440BX 06040000 PTL 000F4240) Jul 10 00:25:36.721222 kernel: ACPI: DSDT 0x000000007FEEAD55 01411E (v01 PTLTD Custom 06040000 MSFT 03000001) Jul 10 00:25:36.721233 kernel: ACPI: FACS 0x000000007FEFFFC0 000040 Jul 10 00:25:36.721238 kernel: ACPI: FACS 0x000000007FEFFFC0 000040 Jul 10 00:25:36.721244 kernel: ACPI: BOOT 0x000000007FEEAD2D 000028 (v01 PTLTD $SBFTBL$ 06040000 LTP 00000001) Jul 10 00:25:36.721249 kernel: ACPI: APIC 0x000000007FEEA5EB 000742 (v01 PTLTD ? APIC 06040000 LTP 00000000) Jul 10 00:25:36.721254 kernel: ACPI: MCFG 0x000000007FEEA5AF 00003C (v01 PTLTD $PCITBL$ 06040000 LTP 00000001) Jul 10 00:25:36.721263 kernel: ACPI: SRAT 0x000000007FEE9757 0008A8 (v02 VMWARE MEMPLUG 06040000 VMW 00000001) Jul 10 00:25:36.721269 kernel: ACPI: HPET 0x000000007FEE971F 000038 (v01 VMWARE VMW HPET 06040000 VMW 00000001) Jul 10 00:25:36.721273 kernel: ACPI: WAET 0x000000007FEE96F7 000028 (v01 VMWARE VMW WAET 06040000 VMW 00000001) Jul 10 00:25:36.721278 kernel: ACPI: Reserving FACP table memory at [mem 0x7fefee73-0x7fefef66] Jul 10 00:25:36.721283 kernel: ACPI: Reserving DSDT table memory at [mem 0x7feead55-0x7fefee72] Jul 10 00:25:36.721291 kernel: ACPI: Reserving FACS table memory at [mem 0x7fefffc0-0x7fefffff] Jul 10 00:25:36.721297 kernel: ACPI: Reserving FACS table memory at [mem 0x7fefffc0-0x7fefffff] Jul 10 00:25:36.721302 kernel: ACPI: Reserving BOOT table memory at [mem 0x7feead2d-0x7feead54] Jul 10 00:25:36.721306 kernel: ACPI: Reserving APIC table memory at [mem 0x7feea5eb-0x7feead2c] Jul 10 00:25:36.721313 kernel: ACPI: Reserving MCFG table memory at [mem 0x7feea5af-0x7feea5ea] Jul 10 00:25:36.721321 kernel: ACPI: Reserving SRAT table memory at [mem 0x7fee9757-0x7fee9ffe] Jul 10 00:25:36.721326 kernel: ACPI: Reserving HPET table memory at [mem 0x7fee971f-0x7fee9756] Jul 10 00:25:36.721331 kernel: ACPI: Reserving WAET table memory at [mem 0x7fee96f7-0x7fee971e] Jul 10 00:25:36.721336 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x0009ffff] Jul 10 00:25:36.721341 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00100000-0x7fffffff] Jul 10 00:25:36.721349 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x80000000-0xbfffffff] hotplug Jul 10 00:25:36.721354 kernel: NUMA: Node 0 [mem 0x00001000-0x0009ffff] + [mem 0x00100000-0x7fffffff] -> [mem 0x00001000-0x7fffffff] Jul 10 00:25:36.721359 kernel: NODE_DATA(0) allocated [mem 0x7fff8dc0-0x7fffffff] Jul 10 00:25:36.721366 kernel: Zone ranges: Jul 10 00:25:36.721379 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Jul 10 00:25:36.721385 kernel: DMA32 [mem 0x0000000001000000-0x000000007fffffff] Jul 10 00:25:36.721390 kernel: Normal empty Jul 10 00:25:36.721395 kernel: Device empty Jul 10 00:25:36.721400 kernel: Movable zone start for each node Jul 10 00:25:36.721405 kernel: Early memory node ranges Jul 10 00:25:36.721410 kernel: node 0: [mem 0x0000000000001000-0x000000000009dfff] Jul 10 00:25:36.721414 kernel: node 0: [mem 0x0000000000100000-0x000000007fedffff] Jul 10 00:25:36.721419 kernel: node 0: [mem 0x000000007ff00000-0x000000007fffffff] Jul 10 00:25:36.721431 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007fffffff] Jul 10 00:25:36.721436 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Jul 10 00:25:36.721441 kernel: On node 0, zone DMA: 98 pages in unavailable ranges Jul 10 00:25:36.721446 kernel: On node 0, zone DMA32: 32 pages in unavailable ranges Jul 10 00:25:36.721451 kernel: ACPI: PM-Timer IO Port: 0x1008 Jul 10 00:25:36.721455 kernel: ACPI: LAPIC_NMI (acpi_id[0x00] high edge lint[0x1]) Jul 10 00:25:36.721460 kernel: ACPI: LAPIC_NMI (acpi_id[0x01] high edge lint[0x1]) Jul 10 00:25:36.721465 kernel: ACPI: LAPIC_NMI (acpi_id[0x02] high edge lint[0x1]) Jul 10 00:25:36.721470 kernel: ACPI: LAPIC_NMI (acpi_id[0x03] high edge lint[0x1]) Jul 10 00:25:36.721477 kernel: ACPI: LAPIC_NMI (acpi_id[0x04] high edge lint[0x1]) Jul 10 00:25:36.721482 kernel: ACPI: LAPIC_NMI (acpi_id[0x05] high edge lint[0x1]) Jul 10 00:25:36.721487 kernel: ACPI: LAPIC_NMI (acpi_id[0x06] high edge lint[0x1]) Jul 10 00:25:36.721491 kernel: ACPI: LAPIC_NMI (acpi_id[0x07] high edge lint[0x1]) Jul 10 00:25:36.721496 kernel: ACPI: LAPIC_NMI (acpi_id[0x08] high edge lint[0x1]) Jul 10 00:25:36.721501 kernel: ACPI: LAPIC_NMI (acpi_id[0x09] high edge lint[0x1]) Jul 10 00:25:36.721506 kernel: ACPI: LAPIC_NMI (acpi_id[0x0a] high edge lint[0x1]) Jul 10 00:25:36.721511 kernel: ACPI: LAPIC_NMI (acpi_id[0x0b] high edge lint[0x1]) Jul 10 00:25:36.721516 kernel: ACPI: LAPIC_NMI (acpi_id[0x0c] high edge lint[0x1]) Jul 10 00:25:36.721521 kernel: ACPI: LAPIC_NMI (acpi_id[0x0d] high edge lint[0x1]) Jul 10 00:25:36.721527 kernel: ACPI: LAPIC_NMI (acpi_id[0x0e] high edge lint[0x1]) Jul 10 00:25:36.721535 kernel: ACPI: LAPIC_NMI (acpi_id[0x0f] high edge lint[0x1]) Jul 10 00:25:36.721540 kernel: ACPI: LAPIC_NMI (acpi_id[0x10] high edge lint[0x1]) Jul 10 00:25:36.721545 kernel: ACPI: LAPIC_NMI (acpi_id[0x11] high edge lint[0x1]) Jul 10 00:25:36.721550 kernel: ACPI: LAPIC_NMI (acpi_id[0x12] high edge lint[0x1]) Jul 10 00:25:36.721555 kernel: ACPI: LAPIC_NMI (acpi_id[0x13] high edge lint[0x1]) Jul 10 00:25:36.721562 kernel: ACPI: LAPIC_NMI (acpi_id[0x14] high edge lint[0x1]) Jul 10 00:25:36.721568 kernel: ACPI: LAPIC_NMI (acpi_id[0x15] high edge lint[0x1]) Jul 10 00:25:36.721573 kernel: ACPI: LAPIC_NMI (acpi_id[0x16] high edge lint[0x1]) Jul 10 00:25:36.721578 kernel: ACPI: LAPIC_NMI (acpi_id[0x17] high edge lint[0x1]) Jul 10 00:25:36.721584 kernel: ACPI: LAPIC_NMI (acpi_id[0x18] high edge lint[0x1]) Jul 10 00:25:36.721589 kernel: ACPI: LAPIC_NMI (acpi_id[0x19] high edge lint[0x1]) Jul 10 00:25:36.721594 kernel: ACPI: LAPIC_NMI (acpi_id[0x1a] high edge lint[0x1]) Jul 10 00:25:36.721598 kernel: ACPI: LAPIC_NMI (acpi_id[0x1b] high edge lint[0x1]) Jul 10 00:25:36.721603 kernel: ACPI: LAPIC_NMI (acpi_id[0x1c] high edge lint[0x1]) Jul 10 00:25:36.721608 kernel: ACPI: LAPIC_NMI (acpi_id[0x1d] high edge lint[0x1]) Jul 10 00:25:36.721613 kernel: ACPI: LAPIC_NMI (acpi_id[0x1e] high edge lint[0x1]) Jul 10 00:25:36.721621 kernel: ACPI: LAPIC_NMI (acpi_id[0x1f] high edge lint[0x1]) Jul 10 00:25:36.721626 kernel: ACPI: LAPIC_NMI (acpi_id[0x20] high edge lint[0x1]) Jul 10 00:25:36.721631 kernel: ACPI: LAPIC_NMI (acpi_id[0x21] high edge lint[0x1]) Jul 10 00:25:36.721637 kernel: ACPI: LAPIC_NMI (acpi_id[0x22] high edge lint[0x1]) Jul 10 00:25:36.721641 kernel: ACPI: LAPIC_NMI (acpi_id[0x23] high edge lint[0x1]) Jul 10 00:25:36.721646 kernel: ACPI: LAPIC_NMI (acpi_id[0x24] high edge lint[0x1]) Jul 10 00:25:36.721651 kernel: ACPI: LAPIC_NMI (acpi_id[0x25] high edge lint[0x1]) Jul 10 00:25:36.721656 kernel: ACPI: LAPIC_NMI (acpi_id[0x26] high edge lint[0x1]) Jul 10 00:25:36.721662 kernel: ACPI: LAPIC_NMI (acpi_id[0x27] high edge lint[0x1]) Jul 10 00:25:36.721670 kernel: ACPI: LAPIC_NMI (acpi_id[0x28] high edge lint[0x1]) Jul 10 00:25:36.721676 kernel: ACPI: LAPIC_NMI (acpi_id[0x29] high edge lint[0x1]) Jul 10 00:25:36.721681 kernel: ACPI: LAPIC_NMI (acpi_id[0x2a] high edge lint[0x1]) Jul 10 00:25:36.721687 kernel: ACPI: LAPIC_NMI (acpi_id[0x2b] high edge lint[0x1]) Jul 10 00:25:36.721693 kernel: ACPI: LAPIC_NMI (acpi_id[0x2c] high edge lint[0x1]) Jul 10 00:25:36.721698 kernel: ACPI: LAPIC_NMI (acpi_id[0x2d] high edge lint[0x1]) Jul 10 00:25:36.721703 kernel: ACPI: LAPIC_NMI (acpi_id[0x2e] high edge lint[0x1]) Jul 10 00:25:36.721708 kernel: ACPI: LAPIC_NMI (acpi_id[0x2f] high edge lint[0x1]) Jul 10 00:25:36.721714 kernel: ACPI: LAPIC_NMI (acpi_id[0x30] high edge lint[0x1]) Jul 10 00:25:36.721719 kernel: ACPI: LAPIC_NMI (acpi_id[0x31] high edge lint[0x1]) Jul 10 00:25:36.721724 kernel: ACPI: LAPIC_NMI (acpi_id[0x32] high edge lint[0x1]) Jul 10 00:25:36.721729 kernel: ACPI: LAPIC_NMI (acpi_id[0x33] high edge lint[0x1]) Jul 10 00:25:36.721735 kernel: ACPI: LAPIC_NMI (acpi_id[0x34] high edge lint[0x1]) Jul 10 00:25:36.721740 kernel: ACPI: LAPIC_NMI (acpi_id[0x35] high edge lint[0x1]) Jul 10 00:25:36.721746 kernel: ACPI: LAPIC_NMI (acpi_id[0x36] high edge lint[0x1]) Jul 10 00:25:36.721751 kernel: ACPI: LAPIC_NMI (acpi_id[0x37] high edge lint[0x1]) Jul 10 00:25:36.721756 kernel: ACPI: LAPIC_NMI (acpi_id[0x38] high edge lint[0x1]) Jul 10 00:25:36.721761 kernel: ACPI: LAPIC_NMI (acpi_id[0x39] high edge lint[0x1]) Jul 10 00:25:36.721766 kernel: ACPI: LAPIC_NMI (acpi_id[0x3a] high edge lint[0x1]) Jul 10 00:25:36.721771 kernel: ACPI: LAPIC_NMI (acpi_id[0x3b] high edge lint[0x1]) Jul 10 00:25:36.721777 kernel: ACPI: LAPIC_NMI (acpi_id[0x3c] high edge lint[0x1]) Jul 10 00:25:36.721783 kernel: ACPI: LAPIC_NMI (acpi_id[0x3d] high edge lint[0x1]) Jul 10 00:25:36.721788 kernel: ACPI: LAPIC_NMI (acpi_id[0x3e] high edge lint[0x1]) Jul 10 00:25:36.721793 kernel: ACPI: LAPIC_NMI (acpi_id[0x3f] high edge lint[0x1]) Jul 10 00:25:36.721799 kernel: ACPI: LAPIC_NMI (acpi_id[0x40] high edge lint[0x1]) Jul 10 00:25:36.721804 kernel: ACPI: LAPIC_NMI (acpi_id[0x41] high edge lint[0x1]) Jul 10 00:25:36.721809 kernel: ACPI: LAPIC_NMI (acpi_id[0x42] high edge lint[0x1]) Jul 10 00:25:36.721814 kernel: ACPI: LAPIC_NMI (acpi_id[0x43] high edge lint[0x1]) Jul 10 00:25:36.721819 kernel: ACPI: LAPIC_NMI (acpi_id[0x44] high edge lint[0x1]) Jul 10 00:25:36.721824 kernel: ACPI: LAPIC_NMI (acpi_id[0x45] high edge lint[0x1]) Jul 10 00:25:36.721830 kernel: ACPI: LAPIC_NMI (acpi_id[0x46] high edge lint[0x1]) Jul 10 00:25:36.721836 kernel: ACPI: LAPIC_NMI (acpi_id[0x47] high edge lint[0x1]) Jul 10 00:25:36.721842 kernel: ACPI: LAPIC_NMI (acpi_id[0x48] high edge lint[0x1]) Jul 10 00:25:36.721847 kernel: ACPI: LAPIC_NMI (acpi_id[0x49] high edge lint[0x1]) Jul 10 00:25:36.721852 kernel: ACPI: LAPIC_NMI (acpi_id[0x4a] high edge lint[0x1]) Jul 10 00:25:36.721857 kernel: ACPI: LAPIC_NMI (acpi_id[0x4b] high edge lint[0x1]) Jul 10 00:25:36.721862 kernel: ACPI: LAPIC_NMI (acpi_id[0x4c] high edge lint[0x1]) Jul 10 00:25:36.721867 kernel: ACPI: LAPIC_NMI (acpi_id[0x4d] high edge lint[0x1]) Jul 10 00:25:36.721873 kernel: ACPI: LAPIC_NMI (acpi_id[0x4e] high edge lint[0x1]) Jul 10 00:25:36.721878 kernel: ACPI: LAPIC_NMI (acpi_id[0x4f] high edge lint[0x1]) Jul 10 00:25:36.721883 kernel: ACPI: LAPIC_NMI (acpi_id[0x50] high edge lint[0x1]) Jul 10 00:25:36.721889 kernel: ACPI: LAPIC_NMI (acpi_id[0x51] high edge lint[0x1]) Jul 10 00:25:36.721894 kernel: ACPI: LAPIC_NMI (acpi_id[0x52] high edge lint[0x1]) Jul 10 00:25:36.721900 kernel: ACPI: LAPIC_NMI (acpi_id[0x53] high edge lint[0x1]) Jul 10 00:25:36.721905 kernel: ACPI: LAPIC_NMI (acpi_id[0x54] high edge lint[0x1]) Jul 10 00:25:36.721910 kernel: ACPI: LAPIC_NMI (acpi_id[0x55] high edge lint[0x1]) Jul 10 00:25:36.721915 kernel: ACPI: LAPIC_NMI (acpi_id[0x56] high edge lint[0x1]) Jul 10 00:25:36.721920 kernel: ACPI: LAPIC_NMI (acpi_id[0x57] high edge lint[0x1]) Jul 10 00:25:36.721925 kernel: ACPI: LAPIC_NMI (acpi_id[0x58] high edge lint[0x1]) Jul 10 00:25:36.721930 kernel: ACPI: LAPIC_NMI (acpi_id[0x59] high edge lint[0x1]) Jul 10 00:25:36.721935 kernel: ACPI: LAPIC_NMI (acpi_id[0x5a] high edge lint[0x1]) Jul 10 00:25:36.721942 kernel: ACPI: LAPIC_NMI (acpi_id[0x5b] high edge lint[0x1]) Jul 10 00:25:36.721947 kernel: ACPI: LAPIC_NMI (acpi_id[0x5c] high edge lint[0x1]) Jul 10 00:25:36.721952 kernel: ACPI: LAPIC_NMI (acpi_id[0x5d] high edge lint[0x1]) Jul 10 00:25:36.721957 kernel: ACPI: LAPIC_NMI (acpi_id[0x5e] high edge lint[0x1]) Jul 10 00:25:36.721963 kernel: ACPI: LAPIC_NMI (acpi_id[0x5f] high edge lint[0x1]) Jul 10 00:25:36.721968 kernel: ACPI: LAPIC_NMI (acpi_id[0x60] high edge lint[0x1]) Jul 10 00:25:36.721973 kernel: ACPI: LAPIC_NMI (acpi_id[0x61] high edge lint[0x1]) Jul 10 00:25:36.721978 kernel: ACPI: LAPIC_NMI (acpi_id[0x62] high edge lint[0x1]) Jul 10 00:25:36.721983 kernel: ACPI: LAPIC_NMI (acpi_id[0x63] high edge lint[0x1]) Jul 10 00:25:36.721988 kernel: ACPI: LAPIC_NMI (acpi_id[0x64] high edge lint[0x1]) Jul 10 00:25:36.721995 kernel: ACPI: LAPIC_NMI (acpi_id[0x65] high edge lint[0x1]) Jul 10 00:25:36.722000 kernel: ACPI: LAPIC_NMI (acpi_id[0x66] high edge lint[0x1]) Jul 10 00:25:36.722005 kernel: ACPI: LAPIC_NMI (acpi_id[0x67] high edge lint[0x1]) Jul 10 00:25:36.722011 kernel: ACPI: LAPIC_NMI (acpi_id[0x68] high edge lint[0x1]) Jul 10 00:25:36.722016 kernel: ACPI: LAPIC_NMI (acpi_id[0x69] high edge lint[0x1]) Jul 10 00:25:36.722021 kernel: ACPI: LAPIC_NMI (acpi_id[0x6a] high edge lint[0x1]) Jul 10 00:25:36.722026 kernel: ACPI: LAPIC_NMI (acpi_id[0x6b] high edge lint[0x1]) Jul 10 00:25:36.722031 kernel: ACPI: LAPIC_NMI (acpi_id[0x6c] high edge lint[0x1]) Jul 10 00:25:36.722037 kernel: ACPI: LAPIC_NMI (acpi_id[0x6d] high edge lint[0x1]) Jul 10 00:25:36.722043 kernel: ACPI: LAPIC_NMI (acpi_id[0x6e] high edge lint[0x1]) Jul 10 00:25:36.722048 kernel: ACPI: LAPIC_NMI (acpi_id[0x6f] high edge lint[0x1]) Jul 10 00:25:36.722053 kernel: ACPI: LAPIC_NMI (acpi_id[0x70] high edge lint[0x1]) Jul 10 00:25:36.722058 kernel: ACPI: LAPIC_NMI (acpi_id[0x71] high edge lint[0x1]) Jul 10 00:25:36.722064 kernel: ACPI: LAPIC_NMI (acpi_id[0x72] high edge lint[0x1]) Jul 10 00:25:36.722069 kernel: ACPI: LAPIC_NMI (acpi_id[0x73] high edge lint[0x1]) Jul 10 00:25:36.722074 kernel: ACPI: LAPIC_NMI (acpi_id[0x74] high edge lint[0x1]) Jul 10 00:25:36.722079 kernel: ACPI: LAPIC_NMI (acpi_id[0x75] high edge lint[0x1]) Jul 10 00:25:36.722084 kernel: ACPI: LAPIC_NMI (acpi_id[0x76] high edge lint[0x1]) Jul 10 00:25:36.722090 kernel: ACPI: LAPIC_NMI (acpi_id[0x77] high edge lint[0x1]) Jul 10 00:25:36.722096 kernel: ACPI: LAPIC_NMI (acpi_id[0x78] high edge lint[0x1]) Jul 10 00:25:36.722101 kernel: ACPI: LAPIC_NMI (acpi_id[0x79] high edge lint[0x1]) Jul 10 00:25:36.722106 kernel: ACPI: LAPIC_NMI (acpi_id[0x7a] high edge lint[0x1]) Jul 10 00:25:36.722112 kernel: ACPI: LAPIC_NMI (acpi_id[0x7b] high edge lint[0x1]) Jul 10 00:25:36.722117 kernel: ACPI: LAPIC_NMI (acpi_id[0x7c] high edge lint[0x1]) Jul 10 00:25:36.722126 kernel: ACPI: LAPIC_NMI (acpi_id[0x7d] high edge lint[0x1]) Jul 10 00:25:36.722131 kernel: ACPI: LAPIC_NMI (acpi_id[0x7e] high edge lint[0x1]) Jul 10 00:25:36.722136 kernel: ACPI: LAPIC_NMI (acpi_id[0x7f] high edge lint[0x1]) Jul 10 00:25:36.722141 kernel: IOAPIC[0]: apic_id 1, version 17, address 0xfec00000, GSI 0-23 Jul 10 00:25:36.722147 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 high edge) Jul 10 00:25:36.722154 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Jul 10 00:25:36.722159 kernel: ACPI: HPET id: 0x8086af01 base: 0xfed00000 Jul 10 00:25:36.722164 kernel: TSC deadline timer available Jul 10 00:25:36.722170 kernel: CPU topo: Max. logical packages: 128 Jul 10 00:25:36.722175 kernel: CPU topo: Max. logical dies: 128 Jul 10 00:25:36.722180 kernel: CPU topo: Max. dies per package: 1 Jul 10 00:25:36.722185 kernel: CPU topo: Max. threads per core: 1 Jul 10 00:25:36.722191 kernel: CPU topo: Num. cores per package: 1 Jul 10 00:25:36.722196 kernel: CPU topo: Num. threads per package: 1 Jul 10 00:25:36.722202 kernel: CPU topo: Allowing 2 present CPUs plus 126 hotplug CPUs Jul 10 00:25:36.722208 kernel: [mem 0x80000000-0xefffffff] available for PCI devices Jul 10 00:25:36.722213 kernel: Booting paravirtualized kernel on VMware hypervisor Jul 10 00:25:36.722218 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Jul 10 00:25:36.722224 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:128 nr_cpu_ids:128 nr_node_ids:1 Jul 10 00:25:36.722232 kernel: percpu: Embedded 60 pages/cpu s207832 r8192 d29736 u262144 Jul 10 00:25:36.722238 kernel: pcpu-alloc: s207832 r8192 d29736 u262144 alloc=1*2097152 Jul 10 00:25:36.722243 kernel: pcpu-alloc: [0] 000 001 002 003 004 005 006 007 Jul 10 00:25:36.722248 kernel: pcpu-alloc: [0] 008 009 010 011 012 013 014 015 Jul 10 00:25:36.722255 kernel: pcpu-alloc: [0] 016 017 018 019 020 021 022 023 Jul 10 00:25:36.722260 kernel: pcpu-alloc: [0] 024 025 026 027 028 029 030 031 Jul 10 00:25:36.722265 kernel: pcpu-alloc: [0] 032 033 034 035 036 037 038 039 Jul 10 00:25:36.722270 kernel: pcpu-alloc: [0] 040 041 042 043 044 045 046 047 Jul 10 00:25:36.722279 kernel: pcpu-alloc: [0] 048 049 050 051 052 053 054 055 Jul 10 00:25:36.722285 kernel: pcpu-alloc: [0] 056 057 058 059 060 061 062 063 Jul 10 00:25:36.722290 kernel: pcpu-alloc: [0] 064 065 066 067 068 069 070 071 Jul 10 00:25:36.722295 kernel: pcpu-alloc: [0] 072 073 074 075 076 077 078 079 Jul 10 00:25:36.722300 kernel: pcpu-alloc: [0] 080 081 082 083 084 085 086 087 Jul 10 00:25:36.722307 kernel: pcpu-alloc: [0] 088 089 090 091 092 093 094 095 Jul 10 00:25:36.722312 kernel: pcpu-alloc: [0] 096 097 098 099 100 101 102 103 Jul 10 00:25:36.722318 kernel: pcpu-alloc: [0] 104 105 106 107 108 109 110 111 Jul 10 00:25:36.722323 kernel: pcpu-alloc: [0] 112 113 114 115 116 117 118 119 Jul 10 00:25:36.722328 kernel: pcpu-alloc: [0] 120 121 122 123 124 125 126 127 Jul 10 00:25:36.722334 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=vmware flatcar.autologin verity.usrhash=844005237fb9709f65a093d5533c4229fb6c54e8e257736d9c3d041b6d3080ea Jul 10 00:25:36.722340 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Jul 10 00:25:36.722345 kernel: random: crng init done Jul 10 00:25:36.722351 kernel: printk: log_buf_len individual max cpu contribution: 4096 bytes Jul 10 00:25:36.722357 kernel: printk: log_buf_len total cpu_extra contributions: 520192 bytes Jul 10 00:25:36.722362 kernel: printk: log_buf_len min size: 262144 bytes Jul 10 00:25:36.725375 kernel: printk: log_buf_len: 1048576 bytes Jul 10 00:25:36.725385 kernel: printk: early log buf free: 245576(93%) Jul 10 00:25:36.725390 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Jul 10 00:25:36.725395 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) Jul 10 00:25:36.725401 kernel: Fallback order for Node 0: 0 Jul 10 00:25:36.725406 kernel: Built 1 zonelists, mobility grouping on. Total pages: 524157 Jul 10 00:25:36.725418 kernel: Policy zone: DMA32 Jul 10 00:25:36.725424 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Jul 10 00:25:36.725429 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=128, Nodes=1 Jul 10 00:25:36.725434 kernel: ftrace: allocating 40095 entries in 157 pages Jul 10 00:25:36.725440 kernel: ftrace: allocated 157 pages with 5 groups Jul 10 00:25:36.725448 kernel: Dynamic Preempt: voluntary Jul 10 00:25:36.725454 kernel: rcu: Preemptible hierarchical RCU implementation. Jul 10 00:25:36.725460 kernel: rcu: RCU event tracing is enabled. Jul 10 00:25:36.725465 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=128. Jul 10 00:25:36.725472 kernel: Trampoline variant of Tasks RCU enabled. Jul 10 00:25:36.725477 kernel: Rude variant of Tasks RCU enabled. Jul 10 00:25:36.725483 kernel: Tracing variant of Tasks RCU enabled. Jul 10 00:25:36.725488 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Jul 10 00:25:36.725493 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=128 Jul 10 00:25:36.725499 kernel: RCU Tasks: Setting shift to 7 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=128. Jul 10 00:25:36.725504 kernel: RCU Tasks Rude: Setting shift to 7 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=128. Jul 10 00:25:36.725510 kernel: RCU Tasks Trace: Setting shift to 7 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=128. Jul 10 00:25:36.725515 kernel: NR_IRQS: 33024, nr_irqs: 1448, preallocated irqs: 16 Jul 10 00:25:36.725520 kernel: rcu: srcu_init: Setting srcu_struct sizes to big. Jul 10 00:25:36.725526 kernel: Console: colour VGA+ 80x25 Jul 10 00:25:36.725532 kernel: printk: legacy console [tty0] enabled Jul 10 00:25:36.725537 kernel: printk: legacy console [ttyS0] enabled Jul 10 00:25:36.725542 kernel: ACPI: Core revision 20240827 Jul 10 00:25:36.725548 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 133484882848 ns Jul 10 00:25:36.725553 kernel: APIC: Switch to symmetric I/O mode setup Jul 10 00:25:36.725558 kernel: x2apic enabled Jul 10 00:25:36.725563 kernel: APIC: Switched APIC routing to: physical x2apic Jul 10 00:25:36.725569 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Jul 10 00:25:36.725576 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x311fd3cd494, max_idle_ns: 440795223879 ns Jul 10 00:25:36.725581 kernel: Calibrating delay loop (skipped) preset value.. 6816.00 BogoMIPS (lpj=3408000) Jul 10 00:25:36.725586 kernel: Disabled fast string operations Jul 10 00:25:36.725592 kernel: Last level iTLB entries: 4KB 64, 2MB 8, 4MB 8 Jul 10 00:25:36.725597 kernel: Last level dTLB entries: 4KB 64, 2MB 32, 4MB 32, 1GB 4 Jul 10 00:25:36.725602 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Jul 10 00:25:36.725607 kernel: Spectre V2 : Spectre BHI mitigation: SW BHB clearing on syscall and VM exit Jul 10 00:25:36.725613 kernel: Spectre V2 : Mitigation: Enhanced / Automatic IBRS Jul 10 00:25:36.725618 kernel: Spectre V2 : Spectre v2 / PBRSB-eIBRS: Retire a single CALL on VMEXIT Jul 10 00:25:36.725624 kernel: RETBleed: Mitigation: Enhanced IBRS Jul 10 00:25:36.725630 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Jul 10 00:25:36.725635 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Jul 10 00:25:36.725641 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode Jul 10 00:25:36.725646 kernel: SRBDS: Unknown: Dependent on hypervisor status Jul 10 00:25:36.725651 kernel: GDS: Unknown: Dependent on hypervisor status Jul 10 00:25:36.725657 kernel: ITS: Mitigation: Aligned branch/return thunks Jul 10 00:25:36.725662 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Jul 10 00:25:36.725667 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Jul 10 00:25:36.725674 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Jul 10 00:25:36.725679 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Jul 10 00:25:36.725684 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Jul 10 00:25:36.725689 kernel: Freeing SMP alternatives memory: 32K Jul 10 00:25:36.725695 kernel: pid_max: default: 131072 minimum: 1024 Jul 10 00:25:36.725700 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Jul 10 00:25:36.725705 kernel: landlock: Up and running. Jul 10 00:25:36.725711 kernel: SELinux: Initializing. Jul 10 00:25:36.725716 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Jul 10 00:25:36.725722 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Jul 10 00:25:36.725728 kernel: smpboot: CPU0: Intel(R) Xeon(R) E-2278G CPU @ 3.40GHz (family: 0x6, model: 0x9e, stepping: 0xd) Jul 10 00:25:36.725733 kernel: Performance Events: Skylake events, core PMU driver. Jul 10 00:25:36.725738 kernel: core: CPUID marked event: 'cpu cycles' unavailable Jul 10 00:25:36.725744 kernel: core: CPUID marked event: 'instructions' unavailable Jul 10 00:25:36.725749 kernel: core: CPUID marked event: 'bus cycles' unavailable Jul 10 00:25:36.725754 kernel: core: CPUID marked event: 'cache references' unavailable Jul 10 00:25:36.725759 kernel: core: CPUID marked event: 'cache misses' unavailable Jul 10 00:25:36.725766 kernel: core: CPUID marked event: 'branch instructions' unavailable Jul 10 00:25:36.725771 kernel: core: CPUID marked event: 'branch misses' unavailable Jul 10 00:25:36.725776 kernel: ... version: 1 Jul 10 00:25:36.725781 kernel: ... bit width: 48 Jul 10 00:25:36.725786 kernel: ... generic registers: 4 Jul 10 00:25:36.725792 kernel: ... value mask: 0000ffffffffffff Jul 10 00:25:36.725797 kernel: ... max period: 000000007fffffff Jul 10 00:25:36.725802 kernel: ... fixed-purpose events: 0 Jul 10 00:25:36.725808 kernel: ... event mask: 000000000000000f Jul 10 00:25:36.725814 kernel: signal: max sigframe size: 1776 Jul 10 00:25:36.725819 kernel: rcu: Hierarchical SRCU implementation. Jul 10 00:25:36.725825 kernel: rcu: Max phase no-delay instances is 400. Jul 10 00:25:36.725830 kernel: Timer migration: 3 hierarchy levels; 8 children per group; 3 crossnode level Jul 10 00:25:36.725835 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Jul 10 00:25:36.725840 kernel: smp: Bringing up secondary CPUs ... Jul 10 00:25:36.725846 kernel: smpboot: x86: Booting SMP configuration: Jul 10 00:25:36.725851 kernel: .... node #0, CPUs: #1 Jul 10 00:25:36.725856 kernel: Disabled fast string operations Jul 10 00:25:36.725861 kernel: smp: Brought up 1 node, 2 CPUs Jul 10 00:25:36.725868 kernel: smpboot: Total of 2 processors activated (13632.00 BogoMIPS) Jul 10 00:25:36.725873 kernel: Memory: 1924244K/2096628K available (14336K kernel code, 2430K rwdata, 9956K rodata, 54420K init, 2548K bss, 161000K reserved, 0K cma-reserved) Jul 10 00:25:36.725879 kernel: devtmpfs: initialized Jul 10 00:25:36.725884 kernel: x86/mm: Memory block size: 128MB Jul 10 00:25:36.725889 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x7feff000-0x7fefffff] (4096 bytes) Jul 10 00:25:36.725895 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Jul 10 00:25:36.725900 kernel: futex hash table entries: 32768 (order: 9, 2097152 bytes, linear) Jul 10 00:25:36.725905 kernel: pinctrl core: initialized pinctrl subsystem Jul 10 00:25:36.725912 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Jul 10 00:25:36.725917 kernel: audit: initializing netlink subsys (disabled) Jul 10 00:25:36.725922 kernel: audit: type=2000 audit(1752107133.277:1): state=initialized audit_enabled=0 res=1 Jul 10 00:25:36.725928 kernel: thermal_sys: Registered thermal governor 'step_wise' Jul 10 00:25:36.725933 kernel: thermal_sys: Registered thermal governor 'user_space' Jul 10 00:25:36.725938 kernel: cpuidle: using governor menu Jul 10 00:25:36.725944 kernel: Simple Boot Flag at 0x36 set to 0x80 Jul 10 00:25:36.725949 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Jul 10 00:25:36.725954 kernel: dca service started, version 1.12.1 Jul 10 00:25:36.725961 kernel: PCI: ECAM [mem 0xf0000000-0xf7ffffff] (base 0xf0000000) for domain 0000 [bus 00-7f] Jul 10 00:25:36.725973 kernel: PCI: Using configuration type 1 for base access Jul 10 00:25:36.725980 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Jul 10 00:25:36.725985 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Jul 10 00:25:36.725991 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Jul 10 00:25:36.725997 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Jul 10 00:25:36.726002 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Jul 10 00:25:36.726008 kernel: ACPI: Added _OSI(Module Device) Jul 10 00:25:36.726013 kernel: ACPI: Added _OSI(Processor Device) Jul 10 00:25:36.726020 kernel: ACPI: Added _OSI(Processor Aggregator Device) Jul 10 00:25:36.726025 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Jul 10 00:25:36.726031 kernel: ACPI: [Firmware Bug]: BIOS _OSI(Linux) query ignored Jul 10 00:25:36.726037 kernel: ACPI: Interpreter enabled Jul 10 00:25:36.726042 kernel: ACPI: PM: (supports S0 S1 S5) Jul 10 00:25:36.726048 kernel: ACPI: Using IOAPIC for interrupt routing Jul 10 00:25:36.726053 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Jul 10 00:25:36.726059 kernel: PCI: Using E820 reservations for host bridge windows Jul 10 00:25:36.726064 kernel: ACPI: Enabled 4 GPEs in block 00 to 0F Jul 10 00:25:36.726071 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-7f]) Jul 10 00:25:36.726152 kernel: acpi PNP0A03:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Jul 10 00:25:36.726203 kernel: acpi PNP0A03:00: _OSC: platform does not support [AER LTR] Jul 10 00:25:36.726251 kernel: acpi PNP0A03:00: _OSC: OS now controls [PCIeHotplug PME PCIeCapability] Jul 10 00:25:36.726259 kernel: PCI host bridge to bus 0000:00 Jul 10 00:25:36.726310 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Jul 10 00:25:36.726406 kernel: pci_bus 0000:00: root bus resource [mem 0x000cc000-0x000dbfff window] Jul 10 00:25:36.726471 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] Jul 10 00:25:36.726513 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Jul 10 00:25:36.726555 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xfeff window] Jul 10 00:25:36.726596 kernel: pci_bus 0000:00: root bus resource [bus 00-7f] Jul 10 00:25:36.726653 kernel: pci 0000:00:00.0: [8086:7190] type 00 class 0x060000 conventional PCI endpoint Jul 10 00:25:36.726711 kernel: pci 0000:00:01.0: [8086:7191] type 01 class 0x060400 conventional PCI bridge Jul 10 00:25:36.726763 kernel: pci 0000:00:01.0: PCI bridge to [bus 01] Jul 10 00:25:36.726816 kernel: pci 0000:00:07.0: [8086:7110] type 00 class 0x060100 conventional PCI endpoint Jul 10 00:25:36.726869 kernel: pci 0000:00:07.1: [8086:7111] type 00 class 0x01018a conventional PCI endpoint Jul 10 00:25:36.726920 kernel: pci 0000:00:07.1: BAR 4 [io 0x1060-0x106f] Jul 10 00:25:36.726971 kernel: pci 0000:00:07.1: BAR 0 [io 0x01f0-0x01f7]: legacy IDE quirk Jul 10 00:25:36.727019 kernel: pci 0000:00:07.1: BAR 1 [io 0x03f6]: legacy IDE quirk Jul 10 00:25:36.727067 kernel: pci 0000:00:07.1: BAR 2 [io 0x0170-0x0177]: legacy IDE quirk Jul 10 00:25:36.727114 kernel: pci 0000:00:07.1: BAR 3 [io 0x0376]: legacy IDE quirk Jul 10 00:25:36.727170 kernel: pci 0000:00:07.3: [8086:7113] type 00 class 0x068000 conventional PCI endpoint Jul 10 00:25:36.727219 kernel: pci 0000:00:07.3: quirk: [io 0x1000-0x103f] claimed by PIIX4 ACPI Jul 10 00:25:36.727270 kernel: pci 0000:00:07.3: quirk: [io 0x1040-0x104f] claimed by PIIX4 SMB Jul 10 00:25:36.727323 kernel: pci 0000:00:07.7: [15ad:0740] type 00 class 0x088000 conventional PCI endpoint Jul 10 00:25:36.729188 kernel: pci 0000:00:07.7: BAR 0 [io 0x1080-0x10bf] Jul 10 00:25:36.729245 kernel: pci 0000:00:07.7: BAR 1 [mem 0xfebfe000-0xfebfffff 64bit] Jul 10 00:25:36.729299 kernel: pci 0000:00:0f.0: [15ad:0405] type 00 class 0x030000 conventional PCI endpoint Jul 10 00:25:36.729349 kernel: pci 0000:00:0f.0: BAR 0 [io 0x1070-0x107f] Jul 10 00:25:36.729408 kernel: pci 0000:00:0f.0: BAR 1 [mem 0xe8000000-0xefffffff pref] Jul 10 00:25:36.729460 kernel: pci 0000:00:0f.0: BAR 2 [mem 0xfe000000-0xfe7fffff] Jul 10 00:25:36.729507 kernel: pci 0000:00:0f.0: ROM [mem 0x00000000-0x00007fff pref] Jul 10 00:25:36.729555 kernel: pci 0000:00:0f.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Jul 10 00:25:36.729607 kernel: pci 0000:00:11.0: [15ad:0790] type 01 class 0x060401 conventional PCI bridge Jul 10 00:25:36.729655 kernel: pci 0000:00:11.0: PCI bridge to [bus 02] (subtractive decode) Jul 10 00:25:36.729703 kernel: pci 0000:00:11.0: bridge window [io 0x2000-0x3fff] Jul 10 00:25:36.729749 kernel: pci 0000:00:11.0: bridge window [mem 0xfd600000-0xfdffffff] Jul 10 00:25:36.729799 kernel: pci 0000:00:11.0: bridge window [mem 0xe7b00000-0xe7ffffff 64bit pref] Jul 10 00:25:36.729852 kernel: pci 0000:00:15.0: [15ad:07a0] type 01 class 0x060400 PCIe Root Port Jul 10 00:25:36.729901 kernel: pci 0000:00:15.0: PCI bridge to [bus 03] Jul 10 00:25:36.729950 kernel: pci 0000:00:15.0: bridge window [io 0x4000-0x4fff] Jul 10 00:25:36.729997 kernel: pci 0000:00:15.0: bridge window [mem 0xfd500000-0xfd5fffff] Jul 10 00:25:36.730047 kernel: pci 0000:00:15.0: PME# supported from D0 D3hot D3cold Jul 10 00:25:36.730101 kernel: pci 0000:00:15.1: [15ad:07a0] type 01 class 0x060400 PCIe Root Port Jul 10 00:25:36.730153 kernel: pci 0000:00:15.1: PCI bridge to [bus 04] Jul 10 00:25:36.730202 kernel: pci 0000:00:15.1: bridge window [io 0x8000-0x8fff] Jul 10 00:25:36.730250 kernel: pci 0000:00:15.1: bridge window [mem 0xfd100000-0xfd1fffff] Jul 10 00:25:36.730298 kernel: pci 0000:00:15.1: bridge window [mem 0xe7800000-0xe78fffff 64bit pref] Jul 10 00:25:36.730354 kernel: pci 0000:00:15.1: PME# supported from D0 D3hot D3cold Jul 10 00:25:36.730414 kernel: pci 0000:00:15.2: [15ad:07a0] type 01 class 0x060400 PCIe Root Port Jul 10 00:25:36.730463 kernel: pci 0000:00:15.2: PCI bridge to [bus 05] Jul 10 00:25:36.730514 kernel: pci 0000:00:15.2: bridge window [io 0xc000-0xcfff] Jul 10 00:25:36.730562 kernel: pci 0000:00:15.2: bridge window [mem 0xfcd00000-0xfcdfffff] Jul 10 00:25:36.730611 kernel: pci 0000:00:15.2: bridge window [mem 0xe7400000-0xe74fffff 64bit pref] Jul 10 00:25:36.730668 kernel: pci 0000:00:15.2: PME# supported from D0 D3hot D3cold Jul 10 00:25:36.730736 kernel: pci 0000:00:15.3: [15ad:07a0] type 01 class 0x060400 PCIe Root Port Jul 10 00:25:36.730786 kernel: pci 0000:00:15.3: PCI bridge to [bus 06] Jul 10 00:25:36.730838 kernel: pci 0000:00:15.3: bridge window [mem 0xfc900000-0xfc9fffff] Jul 10 00:25:36.730886 kernel: pci 0000:00:15.3: bridge window [mem 0xe7000000-0xe70fffff 64bit pref] Jul 10 00:25:36.730934 kernel: pci 0000:00:15.3: PME# supported from D0 D3hot D3cold Jul 10 00:25:36.730987 kernel: pci 0000:00:15.4: [15ad:07a0] type 01 class 0x060400 PCIe Root Port Jul 10 00:25:36.731037 kernel: pci 0000:00:15.4: PCI bridge to [bus 07] Jul 10 00:25:36.731085 kernel: pci 0000:00:15.4: bridge window [mem 0xfc500000-0xfc5fffff] Jul 10 00:25:36.731133 kernel: pci 0000:00:15.4: bridge window [mem 0xe6c00000-0xe6cfffff 64bit pref] Jul 10 00:25:36.731183 kernel: pci 0000:00:15.4: PME# supported from D0 D3hot D3cold Jul 10 00:25:36.731239 kernel: pci 0000:00:15.5: [15ad:07a0] type 01 class 0x060400 PCIe Root Port Jul 10 00:25:36.731287 kernel: pci 0000:00:15.5: PCI bridge to [bus 08] Jul 10 00:25:36.731335 kernel: pci 0000:00:15.5: bridge window [mem 0xfc100000-0xfc1fffff] Jul 10 00:25:36.733172 kernel: pci 0000:00:15.5: bridge window [mem 0xe6800000-0xe68fffff 64bit pref] Jul 10 00:25:36.733232 kernel: pci 0000:00:15.5: PME# supported from D0 D3hot D3cold Jul 10 00:25:36.733290 kernel: pci 0000:00:15.6: [15ad:07a0] type 01 class 0x060400 PCIe Root Port Jul 10 00:25:36.733345 kernel: pci 0000:00:15.6: PCI bridge to [bus 09] Jul 10 00:25:36.733421 kernel: pci 0000:00:15.6: bridge window [mem 0xfbd00000-0xfbdfffff] Jul 10 00:25:36.733472 kernel: pci 0000:00:15.6: bridge window [mem 0xe6400000-0xe64fffff 64bit pref] Jul 10 00:25:36.733824 kernel: pci 0000:00:15.6: PME# supported from D0 D3hot D3cold Jul 10 00:25:36.733880 kernel: pci 0000:00:15.7: [15ad:07a0] type 01 class 0x060400 PCIe Root Port Jul 10 00:25:36.733930 kernel: pci 0000:00:15.7: PCI bridge to [bus 0a] Jul 10 00:25:36.733978 kernel: pci 0000:00:15.7: bridge window [mem 0xfb900000-0xfb9fffff] Jul 10 00:25:36.734030 kernel: pci 0000:00:15.7: bridge window [mem 0xe6000000-0xe60fffff 64bit pref] Jul 10 00:25:36.734078 kernel: pci 0000:00:15.7: PME# supported from D0 D3hot D3cold Jul 10 00:25:36.734131 kernel: pci 0000:00:16.0: [15ad:07a0] type 01 class 0x060400 PCIe Root Port Jul 10 00:25:36.734180 kernel: pci 0000:00:16.0: PCI bridge to [bus 0b] Jul 10 00:25:36.734229 kernel: pci 0000:00:16.0: bridge window [io 0x5000-0x5fff] Jul 10 00:25:36.734277 kernel: pci 0000:00:16.0: bridge window [mem 0xfd400000-0xfd4fffff] Jul 10 00:25:36.734324 kernel: pci 0000:00:16.0: PME# supported from D0 D3hot D3cold Jul 10 00:25:36.734434 kernel: pci 0000:00:16.1: [15ad:07a0] type 01 class 0x060400 PCIe Root Port Jul 10 00:25:36.734488 kernel: pci 0000:00:16.1: PCI bridge to [bus 0c] Jul 10 00:25:36.734537 kernel: pci 0000:00:16.1: bridge window [io 0x9000-0x9fff] Jul 10 00:25:36.734585 kernel: pci 0000:00:16.1: bridge window [mem 0xfd000000-0xfd0fffff] Jul 10 00:25:36.734632 kernel: pci 0000:00:16.1: bridge window [mem 0xe7700000-0xe77fffff 64bit pref] Jul 10 00:25:36.734679 kernel: pci 0000:00:16.1: PME# supported from D0 D3hot D3cold Jul 10 00:25:36.734732 kernel: pci 0000:00:16.2: [15ad:07a0] type 01 class 0x060400 PCIe Root Port Jul 10 00:25:36.734781 kernel: pci 0000:00:16.2: PCI bridge to [bus 0d] Jul 10 00:25:36.734831 kernel: pci 0000:00:16.2: bridge window [io 0xd000-0xdfff] Jul 10 00:25:36.734880 kernel: pci 0000:00:16.2: bridge window [mem 0xfcc00000-0xfccfffff] Jul 10 00:25:36.734928 kernel: pci 0000:00:16.2: bridge window [mem 0xe7300000-0xe73fffff 64bit pref] Jul 10 00:25:36.734976 kernel: pci 0000:00:16.2: PME# supported from D0 D3hot D3cold Jul 10 00:25:36.735029 kernel: pci 0000:00:16.3: [15ad:07a0] type 01 class 0x060400 PCIe Root Port Jul 10 00:25:36.735078 kernel: pci 0000:00:16.3: PCI bridge to [bus 0e] Jul 10 00:25:36.735126 kernel: pci 0000:00:16.3: bridge window [mem 0xfc800000-0xfc8fffff] Jul 10 00:25:36.735176 kernel: pci 0000:00:16.3: bridge window [mem 0xe6f00000-0xe6ffffff 64bit pref] Jul 10 00:25:36.735223 kernel: pci 0000:00:16.3: PME# supported from D0 D3hot D3cold Jul 10 00:25:36.735277 kernel: pci 0000:00:16.4: [15ad:07a0] type 01 class 0x060400 PCIe Root Port Jul 10 00:25:36.735326 kernel: pci 0000:00:16.4: PCI bridge to [bus 0f] Jul 10 00:25:36.736570 kernel: pci 0000:00:16.4: bridge window [mem 0xfc400000-0xfc4fffff] Jul 10 00:25:36.736626 kernel: pci 0000:00:16.4: bridge window [mem 0xe6b00000-0xe6bfffff 64bit pref] Jul 10 00:25:36.736676 kernel: pci 0000:00:16.4: PME# supported from D0 D3hot D3cold Jul 10 00:25:36.736749 kernel: pci 0000:00:16.5: [15ad:07a0] type 01 class 0x060400 PCIe Root Port Jul 10 00:25:36.736800 kernel: pci 0000:00:16.5: PCI bridge to [bus 10] Jul 10 00:25:36.736849 kernel: pci 0000:00:16.5: bridge window [mem 0xfc000000-0xfc0fffff] Jul 10 00:25:36.736897 kernel: pci 0000:00:16.5: bridge window [mem 0xe6700000-0xe67fffff 64bit pref] Jul 10 00:25:36.736945 kernel: pci 0000:00:16.5: PME# supported from D0 D3hot D3cold Jul 10 00:25:36.736998 kernel: pci 0000:00:16.6: [15ad:07a0] type 01 class 0x060400 PCIe Root Port Jul 10 00:25:36.737047 kernel: pci 0000:00:16.6: PCI bridge to [bus 11] Jul 10 00:25:36.737098 kernel: pci 0000:00:16.6: bridge window [mem 0xfbc00000-0xfbcfffff] Jul 10 00:25:36.737146 kernel: pci 0000:00:16.6: bridge window [mem 0xe6300000-0xe63fffff 64bit pref] Jul 10 00:25:36.737194 kernel: pci 0000:00:16.6: PME# supported from D0 D3hot D3cold Jul 10 00:25:36.737246 kernel: pci 0000:00:16.7: [15ad:07a0] type 01 class 0x060400 PCIe Root Port Jul 10 00:25:36.737295 kernel: pci 0000:00:16.7: PCI bridge to [bus 12] Jul 10 00:25:36.737344 kernel: pci 0000:00:16.7: bridge window [mem 0xfb800000-0xfb8fffff] Jul 10 00:25:36.737555 kernel: pci 0000:00:16.7: bridge window [mem 0xe5f00000-0xe5ffffff 64bit pref] Jul 10 00:25:36.737610 kernel: pci 0000:00:16.7: PME# supported from D0 D3hot D3cold Jul 10 00:25:36.737665 kernel: pci 0000:00:17.0: [15ad:07a0] type 01 class 0x060400 PCIe Root Port Jul 10 00:25:36.737715 kernel: pci 0000:00:17.0: PCI bridge to [bus 13] Jul 10 00:25:36.737763 kernel: pci 0000:00:17.0: bridge window [io 0x6000-0x6fff] Jul 10 00:25:36.737811 kernel: pci 0000:00:17.0: bridge window [mem 0xfd300000-0xfd3fffff] Jul 10 00:25:36.737860 kernel: pci 0000:00:17.0: bridge window [mem 0xe7a00000-0xe7afffff 64bit pref] Jul 10 00:25:36.737909 kernel: pci 0000:00:17.0: PME# supported from D0 D3hot D3cold Jul 10 00:25:36.737962 kernel: pci 0000:00:17.1: [15ad:07a0] type 01 class 0x060400 PCIe Root Port Jul 10 00:25:36.738015 kernel: pci 0000:00:17.1: PCI bridge to [bus 14] Jul 10 00:25:36.738063 kernel: pci 0000:00:17.1: bridge window [io 0xa000-0xafff] Jul 10 00:25:36.738111 kernel: pci 0000:00:17.1: bridge window [mem 0xfcf00000-0xfcffffff] Jul 10 00:25:36.738162 kernel: pci 0000:00:17.1: bridge window [mem 0xe7600000-0xe76fffff 64bit pref] Jul 10 00:25:36.738210 kernel: pci 0000:00:17.1: PME# supported from D0 D3hot D3cold Jul 10 00:25:36.738266 kernel: pci 0000:00:17.2: [15ad:07a0] type 01 class 0x060400 PCIe Root Port Jul 10 00:25:36.738317 kernel: pci 0000:00:17.2: PCI bridge to [bus 15] Jul 10 00:25:36.738412 kernel: pci 0000:00:17.2: bridge window [io 0xe000-0xefff] Jul 10 00:25:36.738463 kernel: pci 0000:00:17.2: bridge window [mem 0xfcb00000-0xfcbfffff] Jul 10 00:25:36.738512 kernel: pci 0000:00:17.2: bridge window [mem 0xe7200000-0xe72fffff 64bit pref] Jul 10 00:25:36.738564 kernel: pci 0000:00:17.2: PME# supported from D0 D3hot D3cold Jul 10 00:25:36.738619 kernel: pci 0000:00:17.3: [15ad:07a0] type 01 class 0x060400 PCIe Root Port Jul 10 00:25:36.738668 kernel: pci 0000:00:17.3: PCI bridge to [bus 16] Jul 10 00:25:36.738716 kernel: pci 0000:00:17.3: bridge window [mem 0xfc700000-0xfc7fffff] Jul 10 00:25:36.738765 kernel: pci 0000:00:17.3: bridge window [mem 0xe6e00000-0xe6efffff 64bit pref] Jul 10 00:25:36.738813 kernel: pci 0000:00:17.3: PME# supported from D0 D3hot D3cold Jul 10 00:25:36.738866 kernel: pci 0000:00:17.4: [15ad:07a0] type 01 class 0x060400 PCIe Root Port Jul 10 00:25:36.738918 kernel: pci 0000:00:17.4: PCI bridge to [bus 17] Jul 10 00:25:36.738967 kernel: pci 0000:00:17.4: bridge window [mem 0xfc300000-0xfc3fffff] Jul 10 00:25:36.739016 kernel: pci 0000:00:17.4: bridge window [mem 0xe6a00000-0xe6afffff 64bit pref] Jul 10 00:25:36.739064 kernel: pci 0000:00:17.4: PME# supported from D0 D3hot D3cold Jul 10 00:25:36.739138 kernel: pci 0000:00:17.5: [15ad:07a0] type 01 class 0x060400 PCIe Root Port Jul 10 00:25:36.739388 kernel: pci 0000:00:17.5: PCI bridge to [bus 18] Jul 10 00:25:36.739443 kernel: pci 0000:00:17.5: bridge window [mem 0xfbf00000-0xfbffffff] Jul 10 00:25:36.739570 kernel: pci 0000:00:17.5: bridge window [mem 0xe6600000-0xe66fffff 64bit pref] Jul 10 00:25:36.739624 kernel: pci 0000:00:17.5: PME# supported from D0 D3hot D3cold Jul 10 00:25:36.739697 kernel: pci 0000:00:17.6: [15ad:07a0] type 01 class 0x060400 PCIe Root Port Jul 10 00:25:36.739749 kernel: pci 0000:00:17.6: PCI bridge to [bus 19] Jul 10 00:25:36.739799 kernel: pci 0000:00:17.6: bridge window [mem 0xfbb00000-0xfbbfffff] Jul 10 00:25:36.739848 kernel: pci 0000:00:17.6: bridge window [mem 0xe6200000-0xe62fffff 64bit pref] Jul 10 00:25:36.739896 kernel: pci 0000:00:17.6: PME# supported from D0 D3hot D3cold Jul 10 00:25:36.739952 kernel: pci 0000:00:17.7: [15ad:07a0] type 01 class 0x060400 PCIe Root Port Jul 10 00:25:36.740005 kernel: pci 0000:00:17.7: PCI bridge to [bus 1a] Jul 10 00:25:36.740055 kernel: pci 0000:00:17.7: bridge window [mem 0xfb700000-0xfb7fffff] Jul 10 00:25:36.740104 kernel: pci 0000:00:17.7: bridge window [mem 0xe5e00000-0xe5efffff 64bit pref] Jul 10 00:25:36.740152 kernel: pci 0000:00:17.7: PME# supported from D0 D3hot D3cold Jul 10 00:25:36.740207 kernel: pci 0000:00:18.0: [15ad:07a0] type 01 class 0x060400 PCIe Root Port Jul 10 00:25:36.740257 kernel: pci 0000:00:18.0: PCI bridge to [bus 1b] Jul 10 00:25:36.740306 kernel: pci 0000:00:18.0: bridge window [io 0x7000-0x7fff] Jul 10 00:25:36.740402 kernel: pci 0000:00:18.0: bridge window [mem 0xfd200000-0xfd2fffff] Jul 10 00:25:36.740452 kernel: pci 0000:00:18.0: bridge window [mem 0xe7900000-0xe79fffff 64bit pref] Jul 10 00:25:36.740500 kernel: pci 0000:00:18.0: PME# supported from D0 D3hot D3cold Jul 10 00:25:36.740553 kernel: pci 0000:00:18.1: [15ad:07a0] type 01 class 0x060400 PCIe Root Port Jul 10 00:25:36.740602 kernel: pci 0000:00:18.1: PCI bridge to [bus 1c] Jul 10 00:25:36.740652 kernel: pci 0000:00:18.1: bridge window [io 0xb000-0xbfff] Jul 10 00:25:36.740737 kernel: pci 0000:00:18.1: bridge window [mem 0xfce00000-0xfcefffff] Jul 10 00:25:36.740806 kernel: pci 0000:00:18.1: bridge window [mem 0xe7500000-0xe75fffff 64bit pref] Jul 10 00:25:36.740855 kernel: pci 0000:00:18.1: PME# supported from D0 D3hot D3cold Jul 10 00:25:36.740910 kernel: pci 0000:00:18.2: [15ad:07a0] type 01 class 0x060400 PCIe Root Port Jul 10 00:25:36.740960 kernel: pci 0000:00:18.2: PCI bridge to [bus 1d] Jul 10 00:25:36.741008 kernel: pci 0000:00:18.2: bridge window [mem 0xfca00000-0xfcafffff] Jul 10 00:25:36.741058 kernel: pci 0000:00:18.2: bridge window [mem 0xe7100000-0xe71fffff 64bit pref] Jul 10 00:25:36.741107 kernel: pci 0000:00:18.2: PME# supported from D0 D3hot D3cold Jul 10 00:25:36.741166 kernel: pci 0000:00:18.3: [15ad:07a0] type 01 class 0x060400 PCIe Root Port Jul 10 00:25:36.741216 kernel: pci 0000:00:18.3: PCI bridge to [bus 1e] Jul 10 00:25:36.741282 kernel: pci 0000:00:18.3: bridge window [mem 0xfc600000-0xfc6fffff] Jul 10 00:25:36.741348 kernel: pci 0000:00:18.3: bridge window [mem 0xe6d00000-0xe6dfffff 64bit pref] Jul 10 00:25:36.741448 kernel: pci 0000:00:18.3: PME# supported from D0 D3hot D3cold Jul 10 00:25:36.741504 kernel: pci 0000:00:18.4: [15ad:07a0] type 01 class 0x060400 PCIe Root Port Jul 10 00:25:36.741554 kernel: pci 0000:00:18.4: PCI bridge to [bus 1f] Jul 10 00:25:36.741607 kernel: pci 0000:00:18.4: bridge window [mem 0xfc200000-0xfc2fffff] Jul 10 00:25:36.741657 kernel: pci 0000:00:18.4: bridge window [mem 0xe6900000-0xe69fffff 64bit pref] Jul 10 00:25:36.741706 kernel: pci 0000:00:18.4: PME# supported from D0 D3hot D3cold Jul 10 00:25:36.741796 kernel: pci 0000:00:18.5: [15ad:07a0] type 01 class 0x060400 PCIe Root Port Jul 10 00:25:36.741862 kernel: pci 0000:00:18.5: PCI bridge to [bus 20] Jul 10 00:25:36.741911 kernel: pci 0000:00:18.5: bridge window [mem 0xfbe00000-0xfbefffff] Jul 10 00:25:36.741960 kernel: pci 0000:00:18.5: bridge window [mem 0xe6500000-0xe65fffff 64bit pref] Jul 10 00:25:36.742011 kernel: pci 0000:00:18.5: PME# supported from D0 D3hot D3cold Jul 10 00:25:36.742064 kernel: pci 0000:00:18.6: [15ad:07a0] type 01 class 0x060400 PCIe Root Port Jul 10 00:25:36.742113 kernel: pci 0000:00:18.6: PCI bridge to [bus 21] Jul 10 00:25:36.742161 kernel: pci 0000:00:18.6: bridge window [mem 0xfba00000-0xfbafffff] Jul 10 00:25:36.742210 kernel: pci 0000:00:18.6: bridge window [mem 0xe6100000-0xe61fffff 64bit pref] Jul 10 00:25:36.742259 kernel: pci 0000:00:18.6: PME# supported from D0 D3hot D3cold Jul 10 00:25:36.742313 kernel: pci 0000:00:18.7: [15ad:07a0] type 01 class 0x060400 PCIe Root Port Jul 10 00:25:36.742377 kernel: pci 0000:00:18.7: PCI bridge to [bus 22] Jul 10 00:25:36.742429 kernel: pci 0000:00:18.7: bridge window [mem 0xfb600000-0xfb6fffff] Jul 10 00:25:36.742478 kernel: pci 0000:00:18.7: bridge window [mem 0xe5d00000-0xe5dfffff 64bit pref] Jul 10 00:25:36.742531 kernel: pci 0000:00:18.7: PME# supported from D0 D3hot D3cold Jul 10 00:25:36.742593 kernel: pci_bus 0000:01: extended config space not accessible Jul 10 00:25:36.742645 kernel: pci 0000:00:01.0: PCI bridge to [bus 01] Jul 10 00:25:36.742700 kernel: pci_bus 0000:02: extended config space not accessible Jul 10 00:25:36.742731 kernel: acpiphp: Slot [32] registered Jul 10 00:25:36.742738 kernel: acpiphp: Slot [33] registered Jul 10 00:25:36.742744 kernel: acpiphp: Slot [34] registered Jul 10 00:25:36.742750 kernel: acpiphp: Slot [35] registered Jul 10 00:25:36.742755 kernel: acpiphp: Slot [36] registered Jul 10 00:25:36.742776 kernel: acpiphp: Slot [37] registered Jul 10 00:25:36.742782 kernel: acpiphp: Slot [38] registered Jul 10 00:25:36.742788 kernel: acpiphp: Slot [39] registered Jul 10 00:25:36.742793 kernel: acpiphp: Slot [40] registered Jul 10 00:25:36.742799 kernel: acpiphp: Slot [41] registered Jul 10 00:25:36.742805 kernel: acpiphp: Slot [42] registered Jul 10 00:25:36.742811 kernel: acpiphp: Slot [43] registered Jul 10 00:25:36.742817 kernel: acpiphp: Slot [44] registered Jul 10 00:25:36.742823 kernel: acpiphp: Slot [45] registered Jul 10 00:25:36.742828 kernel: acpiphp: Slot [46] registered Jul 10 00:25:36.742834 kernel: acpiphp: Slot [47] registered Jul 10 00:25:36.742840 kernel: acpiphp: Slot [48] registered Jul 10 00:25:36.742845 kernel: acpiphp: Slot [49] registered Jul 10 00:25:36.742851 kernel: acpiphp: Slot [50] registered Jul 10 00:25:36.742864 kernel: acpiphp: Slot [51] registered Jul 10 00:25:36.742871 kernel: acpiphp: Slot [52] registered Jul 10 00:25:36.742877 kernel: acpiphp: Slot [53] registered Jul 10 00:25:36.742882 kernel: acpiphp: Slot [54] registered Jul 10 00:25:36.742891 kernel: acpiphp: Slot [55] registered Jul 10 00:25:36.742897 kernel: acpiphp: Slot [56] registered Jul 10 00:25:36.742903 kernel: acpiphp: Slot [57] registered Jul 10 00:25:36.742909 kernel: acpiphp: Slot [58] registered Jul 10 00:25:36.742914 kernel: acpiphp: Slot [59] registered Jul 10 00:25:36.742920 kernel: acpiphp: Slot [60] registered Jul 10 00:25:36.742927 kernel: acpiphp: Slot [61] registered Jul 10 00:25:36.742933 kernel: acpiphp: Slot [62] registered Jul 10 00:25:36.742938 kernel: acpiphp: Slot [63] registered Jul 10 00:25:36.742992 kernel: pci 0000:00:11.0: PCI bridge to [bus 02] (subtractive decode) Jul 10 00:25:36.743041 kernel: pci 0000:00:11.0: bridge window [mem 0x000a0000-0x000bffff window] (subtractive decode) Jul 10 00:25:36.743089 kernel: pci 0000:00:11.0: bridge window [mem 0x000cc000-0x000dbfff window] (subtractive decode) Jul 10 00:25:36.743137 kernel: pci 0000:00:11.0: bridge window [mem 0xc0000000-0xfebfffff window] (subtractive decode) Jul 10 00:25:36.743185 kernel: pci 0000:00:11.0: bridge window [io 0x0000-0x0cf7 window] (subtractive decode) Jul 10 00:25:36.743235 kernel: pci 0000:00:11.0: bridge window [io 0x0d00-0xfeff window] (subtractive decode) Jul 10 00:25:36.743289 kernel: pci 0000:03:00.0: [15ad:07c0] type 00 class 0x010700 PCIe Endpoint Jul 10 00:25:36.743357 kernel: pci 0000:03:00.0: BAR 0 [io 0x4000-0x4007] Jul 10 00:25:36.743430 kernel: pci 0000:03:00.0: BAR 1 [mem 0xfd5f8000-0xfd5fffff 64bit] Jul 10 00:25:36.743495 kernel: pci 0000:03:00.0: ROM [mem 0x00000000-0x0000ffff pref] Jul 10 00:25:36.743550 kernel: pci 0000:03:00.0: PME# supported from D0 D3hot D3cold Jul 10 00:25:36.743601 kernel: pci 0000:03:00.0: disabling ASPM on pre-1.1 PCIe device. You can enable it with 'pcie_aspm=force' Jul 10 00:25:36.743654 kernel: pci 0000:00:15.0: PCI bridge to [bus 03] Jul 10 00:25:36.743704 kernel: pci 0000:00:15.1: PCI bridge to [bus 04] Jul 10 00:25:36.743875 kernel: pci 0000:00:15.2: PCI bridge to [bus 05] Jul 10 00:25:36.743934 kernel: pci 0000:00:15.3: PCI bridge to [bus 06] Jul 10 00:25:36.743986 kernel: pci 0000:00:15.4: PCI bridge to [bus 07] Jul 10 00:25:36.744036 kernel: pci 0000:00:15.5: PCI bridge to [bus 08] Jul 10 00:25:36.744086 kernel: pci 0000:00:15.6: PCI bridge to [bus 09] Jul 10 00:25:36.744139 kernel: pci 0000:00:15.7: PCI bridge to [bus 0a] Jul 10 00:25:36.744195 kernel: pci 0000:0b:00.0: [15ad:07b0] type 00 class 0x020000 PCIe Endpoint Jul 10 00:25:36.744247 kernel: pci 0000:0b:00.0: BAR 0 [mem 0xfd4fc000-0xfd4fcfff] Jul 10 00:25:36.744298 kernel: pci 0000:0b:00.0: BAR 1 [mem 0xfd4fd000-0xfd4fdfff] Jul 10 00:25:36.744416 kernel: pci 0000:0b:00.0: BAR 2 [mem 0xfd4fe000-0xfd4fffff] Jul 10 00:25:36.744470 kernel: pci 0000:0b:00.0: BAR 3 [io 0x5000-0x500f] Jul 10 00:25:36.744519 kernel: pci 0000:0b:00.0: ROM [mem 0x00000000-0x0000ffff pref] Jul 10 00:25:36.744572 kernel: pci 0000:0b:00.0: supports D1 D2 Jul 10 00:25:36.744621 kernel: pci 0000:0b:00.0: PME# supported from D0 D1 D2 D3hot D3cold Jul 10 00:25:36.744671 kernel: pci 0000:0b:00.0: disabling ASPM on pre-1.1 PCIe device. You can enable it with 'pcie_aspm=force' Jul 10 00:25:36.744720 kernel: pci 0000:00:16.0: PCI bridge to [bus 0b] Jul 10 00:25:36.744769 kernel: pci 0000:00:16.1: PCI bridge to [bus 0c] Jul 10 00:25:36.744819 kernel: pci 0000:00:16.2: PCI bridge to [bus 0d] Jul 10 00:25:36.744869 kernel: pci 0000:00:16.3: PCI bridge to [bus 0e] Jul 10 00:25:36.744918 kernel: pci 0000:00:16.4: PCI bridge to [bus 0f] Jul 10 00:25:36.744970 kernel: pci 0000:00:16.5: PCI bridge to [bus 10] Jul 10 00:25:36.745020 kernel: pci 0000:00:16.6: PCI bridge to [bus 11] Jul 10 00:25:36.745069 kernel: pci 0000:00:16.7: PCI bridge to [bus 12] Jul 10 00:25:36.745117 kernel: pci 0000:00:17.0: PCI bridge to [bus 13] Jul 10 00:25:36.745167 kernel: pci 0000:00:17.1: PCI bridge to [bus 14] Jul 10 00:25:36.745216 kernel: pci 0000:00:17.2: PCI bridge to [bus 15] Jul 10 00:25:36.745265 kernel: pci 0000:00:17.3: PCI bridge to [bus 16] Jul 10 00:25:36.745314 kernel: pci 0000:00:17.4: PCI bridge to [bus 17] Jul 10 00:25:36.745365 kernel: pci 0000:00:17.5: PCI bridge to [bus 18] Jul 10 00:25:36.745439 kernel: pci 0000:00:17.6: PCI bridge to [bus 19] Jul 10 00:25:36.745490 kernel: pci 0000:00:17.7: PCI bridge to [bus 1a] Jul 10 00:25:36.745539 kernel: pci 0000:00:18.0: PCI bridge to [bus 1b] Jul 10 00:25:36.745589 kernel: pci 0000:00:18.1: PCI bridge to [bus 1c] Jul 10 00:25:36.745639 kernel: pci 0000:00:18.2: PCI bridge to [bus 1d] Jul 10 00:25:36.745689 kernel: pci 0000:00:18.3: PCI bridge to [bus 1e] Jul 10 00:25:36.745741 kernel: pci 0000:00:18.4: PCI bridge to [bus 1f] Jul 10 00:25:36.745790 kernel: pci 0000:00:18.5: PCI bridge to [bus 20] Jul 10 00:25:36.745840 kernel: pci 0000:00:18.6: PCI bridge to [bus 21] Jul 10 00:25:36.745889 kernel: pci 0000:00:18.7: PCI bridge to [bus 22] Jul 10 00:25:36.745897 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 9 Jul 10 00:25:36.745903 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 0 Jul 10 00:25:36.745909 kernel: ACPI: PCI: Interrupt link LNKB disabled Jul 10 00:25:36.745915 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Jul 10 00:25:36.745923 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 10 Jul 10 00:25:36.745928 kernel: iommu: Default domain type: Translated Jul 10 00:25:36.745934 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Jul 10 00:25:36.745940 kernel: PCI: Using ACPI for IRQ routing Jul 10 00:25:36.745945 kernel: PCI: pci_cache_line_size set to 64 bytes Jul 10 00:25:36.745952 kernel: e820: reserve RAM buffer [mem 0x0009ec00-0x0009ffff] Jul 10 00:25:36.745957 kernel: e820: reserve RAM buffer [mem 0x7fee0000-0x7fffffff] Jul 10 00:25:36.746004 kernel: pci 0000:00:0f.0: vgaarb: setting as boot VGA device Jul 10 00:25:36.746055 kernel: pci 0000:00:0f.0: vgaarb: bridge control possible Jul 10 00:25:36.746106 kernel: pci 0000:00:0f.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Jul 10 00:25:36.746115 kernel: vgaarb: loaded Jul 10 00:25:36.746121 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 Jul 10 00:25:36.746126 kernel: hpet0: 16 comparators, 64-bit 14.318180 MHz counter Jul 10 00:25:36.746132 kernel: clocksource: Switched to clocksource tsc-early Jul 10 00:25:36.746138 kernel: VFS: Disk quotas dquot_6.6.0 Jul 10 00:25:36.746144 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Jul 10 00:25:36.746149 kernel: pnp: PnP ACPI init Jul 10 00:25:36.746202 kernel: system 00:00: [io 0x1000-0x103f] has been reserved Jul 10 00:25:36.746251 kernel: system 00:00: [io 0x1040-0x104f] has been reserved Jul 10 00:25:36.746295 kernel: system 00:00: [io 0x0cf0-0x0cf1] has been reserved Jul 10 00:25:36.746345 kernel: system 00:04: [mem 0xfed00000-0xfed003ff] has been reserved Jul 10 00:25:36.746476 kernel: pnp 00:06: [dma 2] Jul 10 00:25:36.746530 kernel: system 00:07: [io 0xfce0-0xfcff] has been reserved Jul 10 00:25:36.746576 kernel: system 00:07: [mem 0xf0000000-0xf7ffffff] has been reserved Jul 10 00:25:36.746622 kernel: system 00:07: [mem 0xfe800000-0xfe9fffff] has been reserved Jul 10 00:25:36.746630 kernel: pnp: PnP ACPI: found 8 devices Jul 10 00:25:36.746636 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Jul 10 00:25:36.746642 kernel: NET: Registered PF_INET protocol family Jul 10 00:25:36.746648 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) Jul 10 00:25:36.746654 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) Jul 10 00:25:36.746660 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Jul 10 00:25:36.746665 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) Jul 10 00:25:36.746673 kernel: TCP bind hash table entries: 16384 (order: 7, 524288 bytes, linear) Jul 10 00:25:36.746679 kernel: TCP: Hash tables configured (established 16384 bind 16384) Jul 10 00:25:36.746685 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) Jul 10 00:25:36.746691 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) Jul 10 00:25:36.746696 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Jul 10 00:25:36.746702 kernel: NET: Registered PF_XDP protocol family Jul 10 00:25:36.746751 kernel: pci 0000:00:15.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 03] add_size 200000 add_align 100000 Jul 10 00:25:36.746802 kernel: pci 0000:00:15.3: bridge window [io 0x1000-0x0fff] to [bus 06] add_size 1000 Jul 10 00:25:36.746854 kernel: pci 0000:00:15.4: bridge window [io 0x1000-0x0fff] to [bus 07] add_size 1000 Jul 10 00:25:36.746903 kernel: pci 0000:00:15.5: bridge window [io 0x1000-0x0fff] to [bus 08] add_size 1000 Jul 10 00:25:36.746952 kernel: pci 0000:00:15.6: bridge window [io 0x1000-0x0fff] to [bus 09] add_size 1000 Jul 10 00:25:36.747001 kernel: pci 0000:00:15.7: bridge window [io 0x1000-0x0fff] to [bus 0a] add_size 1000 Jul 10 00:25:36.747050 kernel: pci 0000:00:16.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 0b] add_size 200000 add_align 100000 Jul 10 00:25:36.747099 kernel: pci 0000:00:16.3: bridge window [io 0x1000-0x0fff] to [bus 0e] add_size 1000 Jul 10 00:25:36.747148 kernel: pci 0000:00:16.4: bridge window [io 0x1000-0x0fff] to [bus 0f] add_size 1000 Jul 10 00:25:36.747196 kernel: pci 0000:00:16.5: bridge window [io 0x1000-0x0fff] to [bus 10] add_size 1000 Jul 10 00:25:36.747246 kernel: pci 0000:00:16.6: bridge window [io 0x1000-0x0fff] to [bus 11] add_size 1000 Jul 10 00:25:36.747295 kernel: pci 0000:00:16.7: bridge window [io 0x1000-0x0fff] to [bus 12] add_size 1000 Jul 10 00:25:36.747346 kernel: pci 0000:00:17.3: bridge window [io 0x1000-0x0fff] to [bus 16] add_size 1000 Jul 10 00:25:36.747409 kernel: pci 0000:00:17.4: bridge window [io 0x1000-0x0fff] to [bus 17] add_size 1000 Jul 10 00:25:36.747474 kernel: pci 0000:00:17.5: bridge window [io 0x1000-0x0fff] to [bus 18] add_size 1000 Jul 10 00:25:36.747524 kernel: pci 0000:00:17.6: bridge window [io 0x1000-0x0fff] to [bus 19] add_size 1000 Jul 10 00:25:36.747573 kernel: pci 0000:00:17.7: bridge window [io 0x1000-0x0fff] to [bus 1a] add_size 1000 Jul 10 00:25:36.747621 kernel: pci 0000:00:18.2: bridge window [io 0x1000-0x0fff] to [bus 1d] add_size 1000 Jul 10 00:25:36.747672 kernel: pci 0000:00:18.3: bridge window [io 0x1000-0x0fff] to [bus 1e] add_size 1000 Jul 10 00:25:36.747721 kernel: pci 0000:00:18.4: bridge window [io 0x1000-0x0fff] to [bus 1f] add_size 1000 Jul 10 00:25:36.747770 kernel: pci 0000:00:18.5: bridge window [io 0x1000-0x0fff] to [bus 20] add_size 1000 Jul 10 00:25:36.747818 kernel: pci 0000:00:18.6: bridge window [io 0x1000-0x0fff] to [bus 21] add_size 1000 Jul 10 00:25:36.747866 kernel: pci 0000:00:18.7: bridge window [io 0x1000-0x0fff] to [bus 22] add_size 1000 Jul 10 00:25:36.747914 kernel: pci 0000:00:15.0: bridge window [mem 0xc0000000-0xc01fffff 64bit pref]: assigned Jul 10 00:25:36.747962 kernel: pci 0000:00:16.0: bridge window [mem 0xc0200000-0xc03fffff 64bit pref]: assigned Jul 10 00:25:36.748010 kernel: pci 0000:00:15.3: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.748063 kernel: pci 0000:00:15.3: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.748111 kernel: pci 0000:00:15.4: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.748159 kernel: pci 0000:00:15.4: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.748207 kernel: pci 0000:00:15.5: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.748255 kernel: pci 0000:00:15.5: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.748303 kernel: pci 0000:00:15.6: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.748352 kernel: pci 0000:00:15.6: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.748432 kernel: pci 0000:00:15.7: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.748482 kernel: pci 0000:00:15.7: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.748531 kernel: pci 0000:00:16.3: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.748579 kernel: pci 0000:00:16.3: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.748628 kernel: pci 0000:00:16.4: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.748676 kernel: pci 0000:00:16.4: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.748724 kernel: pci 0000:00:16.5: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.748772 kernel: pci 0000:00:16.5: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.748823 kernel: pci 0000:00:16.6: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.748871 kernel: pci 0000:00:16.6: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.748920 kernel: pci 0000:00:16.7: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.748968 kernel: pci 0000:00:16.7: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.749016 kernel: pci 0000:00:17.3: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.749064 kernel: pci 0000:00:17.3: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.749113 kernel: pci 0000:00:17.4: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.749161 kernel: pci 0000:00:17.4: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.749212 kernel: pci 0000:00:17.5: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.749260 kernel: pci 0000:00:17.5: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.749309 kernel: pci 0000:00:17.6: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.749357 kernel: pci 0000:00:17.6: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.749604 kernel: pci 0000:00:17.7: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.749657 kernel: pci 0000:00:17.7: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.749706 kernel: pci 0000:00:18.2: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.749755 kernel: pci 0000:00:18.2: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.749807 kernel: pci 0000:00:18.3: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.749856 kernel: pci 0000:00:18.3: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.749905 kernel: pci 0000:00:18.4: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.749953 kernel: pci 0000:00:18.4: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.750001 kernel: pci 0000:00:18.5: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.750050 kernel: pci 0000:00:18.5: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.750098 kernel: pci 0000:00:18.6: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.750146 kernel: pci 0000:00:18.6: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.750198 kernel: pci 0000:00:18.7: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.750246 kernel: pci 0000:00:18.7: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.750293 kernel: pci 0000:00:18.7: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.750345 kernel: pci 0000:00:18.7: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.750404 kernel: pci 0000:00:18.6: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.750463 kernel: pci 0000:00:18.6: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.750525 kernel: pci 0000:00:18.5: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.750575 kernel: pci 0000:00:18.5: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.750624 kernel: pci 0000:00:18.4: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.750677 kernel: pci 0000:00:18.4: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.750741 kernel: pci 0000:00:18.3: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.750946 kernel: pci 0000:00:18.3: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.751002 kernel: pci 0000:00:18.2: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.751052 kernel: pci 0000:00:18.2: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.751101 kernel: pci 0000:00:17.7: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.751152 kernel: pci 0000:00:17.7: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.751201 kernel: pci 0000:00:17.6: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.751390 kernel: pci 0000:00:17.6: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.751445 kernel: pci 0000:00:17.5: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.751495 kernel: pci 0000:00:17.5: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.751543 kernel: pci 0000:00:17.4: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.751592 kernel: pci 0000:00:17.4: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.751640 kernel: pci 0000:00:17.3: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.751689 kernel: pci 0000:00:17.3: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.751738 kernel: pci 0000:00:16.7: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.751786 kernel: pci 0000:00:16.7: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.751834 kernel: pci 0000:00:16.6: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.751883 kernel: pci 0000:00:16.6: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.751933 kernel: pci 0000:00:16.5: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.751981 kernel: pci 0000:00:16.5: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.752030 kernel: pci 0000:00:16.4: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.752078 kernel: pci 0000:00:16.4: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.752126 kernel: pci 0000:00:16.3: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.752366 kernel: pci 0000:00:16.3: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.752437 kernel: pci 0000:00:15.7: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.752488 kernel: pci 0000:00:15.7: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.752538 kernel: pci 0000:00:15.6: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.752587 kernel: pci 0000:00:15.6: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.752636 kernel: pci 0000:00:15.5: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.752685 kernel: pci 0000:00:15.5: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.752734 kernel: pci 0000:00:15.4: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.752783 kernel: pci 0000:00:15.4: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.752831 kernel: pci 0000:00:15.3: bridge window [io size 0x1000]: can't assign; no space Jul 10 00:25:36.752883 kernel: pci 0000:00:15.3: bridge window [io size 0x1000]: failed to assign Jul 10 00:25:36.753016 kernel: pci 0000:00:01.0: PCI bridge to [bus 01] Jul 10 00:25:36.753067 kernel: pci 0000:00:11.0: PCI bridge to [bus 02] Jul 10 00:25:36.753115 kernel: pci 0000:00:11.0: bridge window [io 0x2000-0x3fff] Jul 10 00:25:36.753163 kernel: pci 0000:00:11.0: bridge window [mem 0xfd600000-0xfdffffff] Jul 10 00:25:36.753212 kernel: pci 0000:00:11.0: bridge window [mem 0xe7b00000-0xe7ffffff 64bit pref] Jul 10 00:25:36.753263 kernel: pci 0000:03:00.0: ROM [mem 0xfd500000-0xfd50ffff pref]: assigned Jul 10 00:25:36.753312 kernel: pci 0000:00:15.0: PCI bridge to [bus 03] Jul 10 00:25:36.753572 kernel: pci 0000:00:15.0: bridge window [io 0x4000-0x4fff] Jul 10 00:25:36.753623 kernel: pci 0000:00:15.0: bridge window [mem 0xfd500000-0xfd5fffff] Jul 10 00:25:36.753674 kernel: pci 0000:00:15.0: bridge window [mem 0xc0000000-0xc01fffff 64bit pref] Jul 10 00:25:36.753724 kernel: pci 0000:00:15.1: PCI bridge to [bus 04] Jul 10 00:25:36.753772 kernel: pci 0000:00:15.1: bridge window [io 0x8000-0x8fff] Jul 10 00:25:36.753820 kernel: pci 0000:00:15.1: bridge window [mem 0xfd100000-0xfd1fffff] Jul 10 00:25:36.753869 kernel: pci 0000:00:15.1: bridge window [mem 0xe7800000-0xe78fffff 64bit pref] Jul 10 00:25:36.753919 kernel: pci 0000:00:15.2: PCI bridge to [bus 05] Jul 10 00:25:36.753968 kernel: pci 0000:00:15.2: bridge window [io 0xc000-0xcfff] Jul 10 00:25:36.754016 kernel: pci 0000:00:15.2: bridge window [mem 0xfcd00000-0xfcdfffff] Jul 10 00:25:36.754068 kernel: pci 0000:00:15.2: bridge window [mem 0xe7400000-0xe74fffff 64bit pref] Jul 10 00:25:36.754116 kernel: pci 0000:00:15.3: PCI bridge to [bus 06] Jul 10 00:25:36.754164 kernel: pci 0000:00:15.3: bridge window [mem 0xfc900000-0xfc9fffff] Jul 10 00:25:36.754213 kernel: pci 0000:00:15.3: bridge window [mem 0xe7000000-0xe70fffff 64bit pref] Jul 10 00:25:36.754261 kernel: pci 0000:00:15.4: PCI bridge to [bus 07] Jul 10 00:25:36.754310 kernel: pci 0000:00:15.4: bridge window [mem 0xfc500000-0xfc5fffff] Jul 10 00:25:36.754413 kernel: pci 0000:00:15.4: bridge window [mem 0xe6c00000-0xe6cfffff 64bit pref] Jul 10 00:25:36.754466 kernel: pci 0000:00:15.5: PCI bridge to [bus 08] Jul 10 00:25:36.754517 kernel: pci 0000:00:15.5: bridge window [mem 0xfc100000-0xfc1fffff] Jul 10 00:25:36.754566 kernel: pci 0000:00:15.5: bridge window [mem 0xe6800000-0xe68fffff 64bit pref] Jul 10 00:25:36.754614 kernel: pci 0000:00:15.6: PCI bridge to [bus 09] Jul 10 00:25:36.754943 kernel: pci 0000:00:15.6: bridge window [mem 0xfbd00000-0xfbdfffff] Jul 10 00:25:36.755002 kernel: pci 0000:00:15.6: bridge window [mem 0xe6400000-0xe64fffff 64bit pref] Jul 10 00:25:36.755053 kernel: pci 0000:00:15.7: PCI bridge to [bus 0a] Jul 10 00:25:36.755102 kernel: pci 0000:00:15.7: bridge window [mem 0xfb900000-0xfb9fffff] Jul 10 00:25:36.755154 kernel: pci 0000:00:15.7: bridge window [mem 0xe6000000-0xe60fffff 64bit pref] Jul 10 00:25:36.755205 kernel: pci 0000:0b:00.0: ROM [mem 0xfd400000-0xfd40ffff pref]: assigned Jul 10 00:25:36.755254 kernel: pci 0000:00:16.0: PCI bridge to [bus 0b] Jul 10 00:25:36.755302 kernel: pci 0000:00:16.0: bridge window [io 0x5000-0x5fff] Jul 10 00:25:36.755350 kernel: pci 0000:00:16.0: bridge window [mem 0xfd400000-0xfd4fffff] Jul 10 00:25:36.755407 kernel: pci 0000:00:16.0: bridge window [mem 0xc0200000-0xc03fffff 64bit pref] Jul 10 00:25:36.755457 kernel: pci 0000:00:16.1: PCI bridge to [bus 0c] Jul 10 00:25:36.755505 kernel: pci 0000:00:16.1: bridge window [io 0x9000-0x9fff] Jul 10 00:25:36.755553 kernel: pci 0000:00:16.1: bridge window [mem 0xfd000000-0xfd0fffff] Jul 10 00:25:36.755604 kernel: pci 0000:00:16.1: bridge window [mem 0xe7700000-0xe77fffff 64bit pref] Jul 10 00:25:36.755652 kernel: pci 0000:00:16.2: PCI bridge to [bus 0d] Jul 10 00:25:36.755700 kernel: pci 0000:00:16.2: bridge window [io 0xd000-0xdfff] Jul 10 00:25:36.755748 kernel: pci 0000:00:16.2: bridge window [mem 0xfcc00000-0xfccfffff] Jul 10 00:25:36.755796 kernel: pci 0000:00:16.2: bridge window [mem 0xe7300000-0xe73fffff 64bit pref] Jul 10 00:25:36.755844 kernel: pci 0000:00:16.3: PCI bridge to [bus 0e] Jul 10 00:25:36.755893 kernel: pci 0000:00:16.3: bridge window [mem 0xfc800000-0xfc8fffff] Jul 10 00:25:36.755942 kernel: pci 0000:00:16.3: bridge window [mem 0xe6f00000-0xe6ffffff 64bit pref] Jul 10 00:25:36.755992 kernel: pci 0000:00:16.4: PCI bridge to [bus 0f] Jul 10 00:25:36.756078 kernel: pci 0000:00:16.4: bridge window [mem 0xfc400000-0xfc4fffff] Jul 10 00:25:36.756128 kernel: pci 0000:00:16.4: bridge window [mem 0xe6b00000-0xe6bfffff 64bit pref] Jul 10 00:25:36.756177 kernel: pci 0000:00:16.5: PCI bridge to [bus 10] Jul 10 00:25:36.756225 kernel: pci 0000:00:16.5: bridge window [mem 0xfc000000-0xfc0fffff] Jul 10 00:25:36.756273 kernel: pci 0000:00:16.5: bridge window [mem 0xe6700000-0xe67fffff 64bit pref] Jul 10 00:25:36.756321 kernel: pci 0000:00:16.6: PCI bridge to [bus 11] Jul 10 00:25:36.756384 kernel: pci 0000:00:16.6: bridge window [mem 0xfbc00000-0xfbcfffff] Jul 10 00:25:36.756441 kernel: pci 0000:00:16.6: bridge window [mem 0xe6300000-0xe63fffff 64bit pref] Jul 10 00:25:36.756490 kernel: pci 0000:00:16.7: PCI bridge to [bus 12] Jul 10 00:25:36.756538 kernel: pci 0000:00:16.7: bridge window [mem 0xfb800000-0xfb8fffff] Jul 10 00:25:36.756586 kernel: pci 0000:00:16.7: bridge window [mem 0xe5f00000-0xe5ffffff 64bit pref] Jul 10 00:25:36.756636 kernel: pci 0000:00:17.0: PCI bridge to [bus 13] Jul 10 00:25:36.756684 kernel: pci 0000:00:17.0: bridge window [io 0x6000-0x6fff] Jul 10 00:25:36.756732 kernel: pci 0000:00:17.0: bridge window [mem 0xfd300000-0xfd3fffff] Jul 10 00:25:36.756782 kernel: pci 0000:00:17.0: bridge window [mem 0xe7a00000-0xe7afffff 64bit pref] Jul 10 00:25:36.756831 kernel: pci 0000:00:17.1: PCI bridge to [bus 14] Jul 10 00:25:36.756880 kernel: pci 0000:00:17.1: bridge window [io 0xa000-0xafff] Jul 10 00:25:36.758438 kernel: pci 0000:00:17.1: bridge window [mem 0xfcf00000-0xfcffffff] Jul 10 00:25:36.758493 kernel: pci 0000:00:17.1: bridge window [mem 0xe7600000-0xe76fffff 64bit pref] Jul 10 00:25:36.758545 kernel: pci 0000:00:17.2: PCI bridge to [bus 15] Jul 10 00:25:36.758595 kernel: pci 0000:00:17.2: bridge window [io 0xe000-0xefff] Jul 10 00:25:36.758645 kernel: pci 0000:00:17.2: bridge window [mem 0xfcb00000-0xfcbfffff] Jul 10 00:25:36.758693 kernel: pci 0000:00:17.2: bridge window [mem 0xe7200000-0xe72fffff 64bit pref] Jul 10 00:25:36.758742 kernel: pci 0000:00:17.3: PCI bridge to [bus 16] Jul 10 00:25:36.758793 kernel: pci 0000:00:17.3: bridge window [mem 0xfc700000-0xfc7fffff] Jul 10 00:25:36.758842 kernel: pci 0000:00:17.3: bridge window [mem 0xe6e00000-0xe6efffff 64bit pref] Jul 10 00:25:36.758890 kernel: pci 0000:00:17.4: PCI bridge to [bus 17] Jul 10 00:25:36.758938 kernel: pci 0000:00:17.4: bridge window [mem 0xfc300000-0xfc3fffff] Jul 10 00:25:36.758987 kernel: pci 0000:00:17.4: bridge window [mem 0xe6a00000-0xe6afffff 64bit pref] Jul 10 00:25:36.759036 kernel: pci 0000:00:17.5: PCI bridge to [bus 18] Jul 10 00:25:36.759084 kernel: pci 0000:00:17.5: bridge window [mem 0xfbf00000-0xfbffffff] Jul 10 00:25:36.759132 kernel: pci 0000:00:17.5: bridge window [mem 0xe6600000-0xe66fffff 64bit pref] Jul 10 00:25:36.759184 kernel: pci 0000:00:17.6: PCI bridge to [bus 19] Jul 10 00:25:36.759232 kernel: pci 0000:00:17.6: bridge window [mem 0xfbb00000-0xfbbfffff] Jul 10 00:25:36.759280 kernel: pci 0000:00:17.6: bridge window [mem 0xe6200000-0xe62fffff 64bit pref] Jul 10 00:25:36.759329 kernel: pci 0000:00:17.7: PCI bridge to [bus 1a] Jul 10 00:25:36.759396 kernel: pci 0000:00:17.7: bridge window [mem 0xfb700000-0xfb7fffff] Jul 10 00:25:36.759448 kernel: pci 0000:00:17.7: bridge window [mem 0xe5e00000-0xe5efffff 64bit pref] Jul 10 00:25:36.759501 kernel: pci 0000:00:18.0: PCI bridge to [bus 1b] Jul 10 00:25:36.759550 kernel: pci 0000:00:18.0: bridge window [io 0x7000-0x7fff] Jul 10 00:25:36.759598 kernel: pci 0000:00:18.0: bridge window [mem 0xfd200000-0xfd2fffff] Jul 10 00:25:36.759647 kernel: pci 0000:00:18.0: bridge window [mem 0xe7900000-0xe79fffff 64bit pref] Jul 10 00:25:36.759695 kernel: pci 0000:00:18.1: PCI bridge to [bus 1c] Jul 10 00:25:36.759743 kernel: pci 0000:00:18.1: bridge window [io 0xb000-0xbfff] Jul 10 00:25:36.759791 kernel: pci 0000:00:18.1: bridge window [mem 0xfce00000-0xfcefffff] Jul 10 00:25:36.759839 kernel: pci 0000:00:18.1: bridge window [mem 0xe7500000-0xe75fffff 64bit pref] Jul 10 00:25:36.759887 kernel: pci 0000:00:18.2: PCI bridge to [bus 1d] Jul 10 00:25:36.759934 kernel: pci 0000:00:18.2: bridge window [mem 0xfca00000-0xfcafffff] Jul 10 00:25:36.759985 kernel: pci 0000:00:18.2: bridge window [mem 0xe7100000-0xe71fffff 64bit pref] Jul 10 00:25:36.760034 kernel: pci 0000:00:18.3: PCI bridge to [bus 1e] Jul 10 00:25:36.760082 kernel: pci 0000:00:18.3: bridge window [mem 0xfc600000-0xfc6fffff] Jul 10 00:25:36.760130 kernel: pci 0000:00:18.3: bridge window [mem 0xe6d00000-0xe6dfffff 64bit pref] Jul 10 00:25:36.760179 kernel: pci 0000:00:18.4: PCI bridge to [bus 1f] Jul 10 00:25:36.760227 kernel: pci 0000:00:18.4: bridge window [mem 0xfc200000-0xfc2fffff] Jul 10 00:25:36.760275 kernel: pci 0000:00:18.4: bridge window [mem 0xe6900000-0xe69fffff 64bit pref] Jul 10 00:25:36.760326 kernel: pci 0000:00:18.5: PCI bridge to [bus 20] Jul 10 00:25:36.760649 kernel: pci 0000:00:18.5: bridge window [mem 0xfbe00000-0xfbefffff] Jul 10 00:25:36.760707 kernel: pci 0000:00:18.5: bridge window [mem 0xe6500000-0xe65fffff 64bit pref] Jul 10 00:25:36.760759 kernel: pci 0000:00:18.6: PCI bridge to [bus 21] Jul 10 00:25:36.760809 kernel: pci 0000:00:18.6: bridge window [mem 0xfba00000-0xfbafffff] Jul 10 00:25:36.760857 kernel: pci 0000:00:18.6: bridge window [mem 0xe6100000-0xe61fffff 64bit pref] Jul 10 00:25:36.760906 kernel: pci 0000:00:18.7: PCI bridge to [bus 22] Jul 10 00:25:36.760954 kernel: pci 0000:00:18.7: bridge window [mem 0xfb600000-0xfb6fffff] Jul 10 00:25:36.761005 kernel: pci 0000:00:18.7: bridge window [mem 0xe5d00000-0xe5dfffff 64bit pref] Jul 10 00:25:36.761052 kernel: pci_bus 0000:00: resource 4 [mem 0x000a0000-0x000bffff window] Jul 10 00:25:36.761095 kernel: pci_bus 0000:00: resource 5 [mem 0x000cc000-0x000dbfff window] Jul 10 00:25:36.761424 kernel: pci_bus 0000:00: resource 6 [mem 0xc0000000-0xfebfffff window] Jul 10 00:25:36.761471 kernel: pci_bus 0000:00: resource 7 [io 0x0000-0x0cf7 window] Jul 10 00:25:36.761514 kernel: pci_bus 0000:00: resource 8 [io 0x0d00-0xfeff window] Jul 10 00:25:36.761561 kernel: pci_bus 0000:02: resource 0 [io 0x2000-0x3fff] Jul 10 00:25:36.761609 kernel: pci_bus 0000:02: resource 1 [mem 0xfd600000-0xfdffffff] Jul 10 00:25:36.761653 kernel: pci_bus 0000:02: resource 2 [mem 0xe7b00000-0xe7ffffff 64bit pref] Jul 10 00:25:36.761698 kernel: pci_bus 0000:02: resource 4 [mem 0x000a0000-0x000bffff window] Jul 10 00:25:36.761742 kernel: pci_bus 0000:02: resource 5 [mem 0x000cc000-0x000dbfff window] Jul 10 00:25:36.761789 kernel: pci_bus 0000:02: resource 6 [mem 0xc0000000-0xfebfffff window] Jul 10 00:25:36.761833 kernel: pci_bus 0000:02: resource 7 [io 0x0000-0x0cf7 window] Jul 10 00:25:36.761877 kernel: pci_bus 0000:02: resource 8 [io 0x0d00-0xfeff window] Jul 10 00:25:36.761928 kernel: pci_bus 0000:03: resource 0 [io 0x4000-0x4fff] Jul 10 00:25:36.761973 kernel: pci_bus 0000:03: resource 1 [mem 0xfd500000-0xfd5fffff] Jul 10 00:25:36.762017 kernel: pci_bus 0000:03: resource 2 [mem 0xc0000000-0xc01fffff 64bit pref] Jul 10 00:25:36.762065 kernel: pci_bus 0000:04: resource 0 [io 0x8000-0x8fff] Jul 10 00:25:36.762109 kernel: pci_bus 0000:04: resource 1 [mem 0xfd100000-0xfd1fffff] Jul 10 00:25:36.762153 kernel: pci_bus 0000:04: resource 2 [mem 0xe7800000-0xe78fffff 64bit pref] Jul 10 00:25:36.762201 kernel: pci_bus 0000:05: resource 0 [io 0xc000-0xcfff] Jul 10 00:25:36.762248 kernel: pci_bus 0000:05: resource 1 [mem 0xfcd00000-0xfcdfffff] Jul 10 00:25:36.762292 kernel: pci_bus 0000:05: resource 2 [mem 0xe7400000-0xe74fffff 64bit pref] Jul 10 00:25:36.762340 kernel: pci_bus 0000:06: resource 1 [mem 0xfc900000-0xfc9fffff] Jul 10 00:25:36.762399 kernel: pci_bus 0000:06: resource 2 [mem 0xe7000000-0xe70fffff 64bit pref] Jul 10 00:25:36.762451 kernel: pci_bus 0000:07: resource 1 [mem 0xfc500000-0xfc5fffff] Jul 10 00:25:36.762496 kernel: pci_bus 0000:07: resource 2 [mem 0xe6c00000-0xe6cfffff 64bit pref] Jul 10 00:25:36.762547 kernel: pci_bus 0000:08: resource 1 [mem 0xfc100000-0xfc1fffff] Jul 10 00:25:36.762592 kernel: pci_bus 0000:08: resource 2 [mem 0xe6800000-0xe68fffff 64bit pref] Jul 10 00:25:36.762639 kernel: pci_bus 0000:09: resource 1 [mem 0xfbd00000-0xfbdfffff] Jul 10 00:25:36.762683 kernel: pci_bus 0000:09: resource 2 [mem 0xe6400000-0xe64fffff 64bit pref] Jul 10 00:25:36.762730 kernel: pci_bus 0000:0a: resource 1 [mem 0xfb900000-0xfb9fffff] Jul 10 00:25:36.762774 kernel: pci_bus 0000:0a: resource 2 [mem 0xe6000000-0xe60fffff 64bit pref] Jul 10 00:25:36.762826 kernel: pci_bus 0000:0b: resource 0 [io 0x5000-0x5fff] Jul 10 00:25:36.762871 kernel: pci_bus 0000:0b: resource 1 [mem 0xfd400000-0xfd4fffff] Jul 10 00:25:36.762914 kernel: pci_bus 0000:0b: resource 2 [mem 0xc0200000-0xc03fffff 64bit pref] Jul 10 00:25:36.762962 kernel: pci_bus 0000:0c: resource 0 [io 0x9000-0x9fff] Jul 10 00:25:36.763006 kernel: pci_bus 0000:0c: resource 1 [mem 0xfd000000-0xfd0fffff] Jul 10 00:25:36.763050 kernel: pci_bus 0000:0c: resource 2 [mem 0xe7700000-0xe77fffff 64bit pref] Jul 10 00:25:36.763101 kernel: pci_bus 0000:0d: resource 0 [io 0xd000-0xdfff] Jul 10 00:25:36.763146 kernel: pci_bus 0000:0d: resource 1 [mem 0xfcc00000-0xfccfffff] Jul 10 00:25:36.763189 kernel: pci_bus 0000:0d: resource 2 [mem 0xe7300000-0xe73fffff 64bit pref] Jul 10 00:25:36.763237 kernel: pci_bus 0000:0e: resource 1 [mem 0xfc800000-0xfc8fffff] Jul 10 00:25:36.763281 kernel: pci_bus 0000:0e: resource 2 [mem 0xe6f00000-0xe6ffffff 64bit pref] Jul 10 00:25:36.763329 kernel: pci_bus 0000:0f: resource 1 [mem 0xfc400000-0xfc4fffff] Jul 10 00:25:36.763392 kernel: pci_bus 0000:0f: resource 2 [mem 0xe6b00000-0xe6bfffff 64bit pref] Jul 10 00:25:36.763448 kernel: pci_bus 0000:10: resource 1 [mem 0xfc000000-0xfc0fffff] Jul 10 00:25:36.763494 kernel: pci_bus 0000:10: resource 2 [mem 0xe6700000-0xe67fffff 64bit pref] Jul 10 00:25:36.763542 kernel: pci_bus 0000:11: resource 1 [mem 0xfbc00000-0xfbcfffff] Jul 10 00:25:36.763587 kernel: pci_bus 0000:11: resource 2 [mem 0xe6300000-0xe63fffff 64bit pref] Jul 10 00:25:36.763635 kernel: pci_bus 0000:12: resource 1 [mem 0xfb800000-0xfb8fffff] Jul 10 00:25:36.763679 kernel: pci_bus 0000:12: resource 2 [mem 0xe5f00000-0xe5ffffff 64bit pref] Jul 10 00:25:36.763729 kernel: pci_bus 0000:13: resource 0 [io 0x6000-0x6fff] Jul 10 00:25:36.763773 kernel: pci_bus 0000:13: resource 1 [mem 0xfd300000-0xfd3fffff] Jul 10 00:25:36.763817 kernel: pci_bus 0000:13: resource 2 [mem 0xe7a00000-0xe7afffff 64bit pref] Jul 10 00:25:36.763866 kernel: pci_bus 0000:14: resource 0 [io 0xa000-0xafff] Jul 10 00:25:36.763911 kernel: pci_bus 0000:14: resource 1 [mem 0xfcf00000-0xfcffffff] Jul 10 00:25:36.763956 kernel: pci_bus 0000:14: resource 2 [mem 0xe7600000-0xe76fffff 64bit pref] Jul 10 00:25:36.764002 kernel: pci_bus 0000:15: resource 0 [io 0xe000-0xefff] Jul 10 00:25:36.764050 kernel: pci_bus 0000:15: resource 1 [mem 0xfcb00000-0xfcbfffff] Jul 10 00:25:36.764094 kernel: pci_bus 0000:15: resource 2 [mem 0xe7200000-0xe72fffff 64bit pref] Jul 10 00:25:36.764141 kernel: pci_bus 0000:16: resource 1 [mem 0xfc700000-0xfc7fffff] Jul 10 00:25:36.764186 kernel: pci_bus 0000:16: resource 2 [mem 0xe6e00000-0xe6efffff 64bit pref] Jul 10 00:25:36.764233 kernel: pci_bus 0000:17: resource 1 [mem 0xfc300000-0xfc3fffff] Jul 10 00:25:36.764277 kernel: pci_bus 0000:17: resource 2 [mem 0xe6a00000-0xe6afffff 64bit pref] Jul 10 00:25:36.764331 kernel: pci_bus 0000:18: resource 1 [mem 0xfbf00000-0xfbffffff] Jul 10 00:25:36.764457 kernel: pci_bus 0000:18: resource 2 [mem 0xe6600000-0xe66fffff 64bit pref] Jul 10 00:25:36.764521 kernel: pci_bus 0000:19: resource 1 [mem 0xfbb00000-0xfbbfffff] Jul 10 00:25:36.764565 kernel: pci_bus 0000:19: resource 2 [mem 0xe6200000-0xe62fffff 64bit pref] Jul 10 00:25:36.764612 kernel: pci_bus 0000:1a: resource 1 [mem 0xfb700000-0xfb7fffff] Jul 10 00:25:36.764657 kernel: pci_bus 0000:1a: resource 2 [mem 0xe5e00000-0xe5efffff 64bit pref] Jul 10 00:25:36.764708 kernel: pci_bus 0000:1b: resource 0 [io 0x7000-0x7fff] Jul 10 00:25:36.764753 kernel: pci_bus 0000:1b: resource 1 [mem 0xfd200000-0xfd2fffff] Jul 10 00:25:36.764796 kernel: pci_bus 0000:1b: resource 2 [mem 0xe7900000-0xe79fffff 64bit pref] Jul 10 00:25:36.764845 kernel: pci_bus 0000:1c: resource 0 [io 0xb000-0xbfff] Jul 10 00:25:36.764889 kernel: pci_bus 0000:1c: resource 1 [mem 0xfce00000-0xfcefffff] Jul 10 00:25:36.764932 kernel: pci_bus 0000:1c: resource 2 [mem 0xe7500000-0xe75fffff 64bit pref] Jul 10 00:25:36.764982 kernel: pci_bus 0000:1d: resource 1 [mem 0xfca00000-0xfcafffff] Jul 10 00:25:36.765029 kernel: pci_bus 0000:1d: resource 2 [mem 0xe7100000-0xe71fffff 64bit pref] Jul 10 00:25:36.765077 kernel: pci_bus 0000:1e: resource 1 [mem 0xfc600000-0xfc6fffff] Jul 10 00:25:36.765121 kernel: pci_bus 0000:1e: resource 2 [mem 0xe6d00000-0xe6dfffff 64bit pref] Jul 10 00:25:36.765169 kernel: pci_bus 0000:1f: resource 1 [mem 0xfc200000-0xfc2fffff] Jul 10 00:25:36.765213 kernel: pci_bus 0000:1f: resource 2 [mem 0xe6900000-0xe69fffff 64bit pref] Jul 10 00:25:36.765261 kernel: pci_bus 0000:20: resource 1 [mem 0xfbe00000-0xfbefffff] Jul 10 00:25:36.765308 kernel: pci_bus 0000:20: resource 2 [mem 0xe6500000-0xe65fffff 64bit pref] Jul 10 00:25:36.765358 kernel: pci_bus 0000:21: resource 1 [mem 0xfba00000-0xfbafffff] Jul 10 00:25:36.765420 kernel: pci_bus 0000:21: resource 2 [mem 0xe6100000-0xe61fffff 64bit pref] Jul 10 00:25:36.765472 kernel: pci_bus 0000:22: resource 1 [mem 0xfb600000-0xfb6fffff] Jul 10 00:25:36.765516 kernel: pci_bus 0000:22: resource 2 [mem 0xe5d00000-0xe5dfffff 64bit pref] Jul 10 00:25:36.765570 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers Jul 10 00:25:36.765581 kernel: PCI: CLS 32 bytes, default 64 Jul 10 00:25:36.765587 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Jul 10 00:25:36.765593 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x311fd3cd494, max_idle_ns: 440795223879 ns Jul 10 00:25:36.765599 kernel: clocksource: Switched to clocksource tsc Jul 10 00:25:36.765605 kernel: Initialise system trusted keyrings Jul 10 00:25:36.765611 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 Jul 10 00:25:36.765617 kernel: Key type asymmetric registered Jul 10 00:25:36.765622 kernel: Asymmetric key parser 'x509' registered Jul 10 00:25:36.765628 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Jul 10 00:25:36.765635 kernel: io scheduler mq-deadline registered Jul 10 00:25:36.765641 kernel: io scheduler kyber registered Jul 10 00:25:36.765646 kernel: io scheduler bfq registered Jul 10 00:25:36.765696 kernel: pcieport 0000:00:15.0: PME: Signaling with IRQ 24 Jul 10 00:25:36.765746 kernel: pcieport 0000:00:15.0: pciehp: Slot #160 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 10 00:25:36.765795 kernel: pcieport 0000:00:15.1: PME: Signaling with IRQ 25 Jul 10 00:25:36.765845 kernel: pcieport 0000:00:15.1: pciehp: Slot #161 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 10 00:25:36.765893 kernel: pcieport 0000:00:15.2: PME: Signaling with IRQ 26 Jul 10 00:25:36.765944 kernel: pcieport 0000:00:15.2: pciehp: Slot #162 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 10 00:25:36.765994 kernel: pcieport 0000:00:15.3: PME: Signaling with IRQ 27 Jul 10 00:25:36.766043 kernel: pcieport 0000:00:15.3: pciehp: Slot #163 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 10 00:25:36.766092 kernel: pcieport 0000:00:15.4: PME: Signaling with IRQ 28 Jul 10 00:25:36.766141 kernel: pcieport 0000:00:15.4: pciehp: Slot #164 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 10 00:25:36.766189 kernel: pcieport 0000:00:15.5: PME: Signaling with IRQ 29 Jul 10 00:25:36.766238 kernel: pcieport 0000:00:15.5: pciehp: Slot #165 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 10 00:25:36.766289 kernel: pcieport 0000:00:15.6: PME: Signaling with IRQ 30 Jul 10 00:25:36.766339 kernel: pcieport 0000:00:15.6: pciehp: Slot #166 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 10 00:25:36.766460 kernel: pcieport 0000:00:15.7: PME: Signaling with IRQ 31 Jul 10 00:25:36.766510 kernel: pcieport 0000:00:15.7: pciehp: Slot #167 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 10 00:25:36.766559 kernel: pcieport 0000:00:16.0: PME: Signaling with IRQ 32 Jul 10 00:25:36.766611 kernel: pcieport 0000:00:16.0: pciehp: Slot #192 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 10 00:25:36.766660 kernel: pcieport 0000:00:16.1: PME: Signaling with IRQ 33 Jul 10 00:25:36.766711 kernel: pcieport 0000:00:16.1: pciehp: Slot #193 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 10 00:25:36.766761 kernel: pcieport 0000:00:16.2: PME: Signaling with IRQ 34 Jul 10 00:25:36.766809 kernel: pcieport 0000:00:16.2: pciehp: Slot #194 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 10 00:25:36.766857 kernel: pcieport 0000:00:16.3: PME: Signaling with IRQ 35 Jul 10 00:25:36.766906 kernel: pcieport 0000:00:16.3: pciehp: Slot #195 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 10 00:25:36.766963 kernel: pcieport 0000:00:16.4: PME: Signaling with IRQ 36 Jul 10 00:25:36.767014 kernel: pcieport 0000:00:16.4: pciehp: Slot #196 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 10 00:25:36.767063 kernel: pcieport 0000:00:16.5: PME: Signaling with IRQ 37 Jul 10 00:25:36.767115 kernel: pcieport 0000:00:16.5: pciehp: Slot #197 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 10 00:25:36.767164 kernel: pcieport 0000:00:16.6: PME: Signaling with IRQ 38 Jul 10 00:25:36.767213 kernel: pcieport 0000:00:16.6: pciehp: Slot #198 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 10 00:25:36.767262 kernel: pcieport 0000:00:16.7: PME: Signaling with IRQ 39 Jul 10 00:25:36.767312 kernel: pcieport 0000:00:16.7: pciehp: Slot #199 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 10 00:25:36.767360 kernel: pcieport 0000:00:17.0: PME: Signaling with IRQ 40 Jul 10 00:25:36.767425 kernel: pcieport 0000:00:17.0: pciehp: Slot #224 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 10 00:25:36.767477 kernel: pcieport 0000:00:17.1: PME: Signaling with IRQ 41 Jul 10 00:25:36.767526 kernel: pcieport 0000:00:17.1: pciehp: Slot #225 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 10 00:25:36.767575 kernel: pcieport 0000:00:17.2: PME: Signaling with IRQ 42 Jul 10 00:25:36.767624 kernel: pcieport 0000:00:17.2: pciehp: Slot #226 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 10 00:25:36.767673 kernel: pcieport 0000:00:17.3: PME: Signaling with IRQ 43 Jul 10 00:25:36.767723 kernel: pcieport 0000:00:17.3: pciehp: Slot #227 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 10 00:25:36.767771 kernel: pcieport 0000:00:17.4: PME: Signaling with IRQ 44 Jul 10 00:25:36.767821 kernel: pcieport 0000:00:17.4: pciehp: Slot #228 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 10 00:25:36.767872 kernel: pcieport 0000:00:17.5: PME: Signaling with IRQ 45 Jul 10 00:25:36.767921 kernel: pcieport 0000:00:17.5: pciehp: Slot #229 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 10 00:25:36.767970 kernel: pcieport 0000:00:17.6: PME: Signaling with IRQ 46 Jul 10 00:25:36.768019 kernel: pcieport 0000:00:17.6: pciehp: Slot #230 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 10 00:25:36.768068 kernel: pcieport 0000:00:17.7: PME: Signaling with IRQ 47 Jul 10 00:25:36.768117 kernel: pcieport 0000:00:17.7: pciehp: Slot #231 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 10 00:25:36.768166 kernel: pcieport 0000:00:18.0: PME: Signaling with IRQ 48 Jul 10 00:25:36.768217 kernel: pcieport 0000:00:18.0: pciehp: Slot #256 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 10 00:25:36.768266 kernel: pcieport 0000:00:18.1: PME: Signaling with IRQ 49 Jul 10 00:25:36.768315 kernel: pcieport 0000:00:18.1: pciehp: Slot #257 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 10 00:25:36.768375 kernel: pcieport 0000:00:18.2: PME: Signaling with IRQ 50 Jul 10 00:25:36.768428 kernel: pcieport 0000:00:18.2: pciehp: Slot #258 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 10 00:25:36.768478 kernel: pcieport 0000:00:18.3: PME: Signaling with IRQ 51 Jul 10 00:25:36.768526 kernel: pcieport 0000:00:18.3: pciehp: Slot #259 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 10 00:25:36.768578 kernel: pcieport 0000:00:18.4: PME: Signaling with IRQ 52 Jul 10 00:25:36.768627 kernel: pcieport 0000:00:18.4: pciehp: Slot #260 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 10 00:25:36.768675 kernel: pcieport 0000:00:18.5: PME: Signaling with IRQ 53 Jul 10 00:25:36.768724 kernel: pcieport 0000:00:18.5: pciehp: Slot #261 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 10 00:25:36.768772 kernel: pcieport 0000:00:18.6: PME: Signaling with IRQ 54 Jul 10 00:25:36.768821 kernel: pcieport 0000:00:18.6: pciehp: Slot #262 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 10 00:25:36.768869 kernel: pcieport 0000:00:18.7: PME: Signaling with IRQ 55 Jul 10 00:25:36.768918 kernel: pcieport 0000:00:18.7: pciehp: Slot #263 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 10 00:25:36.768929 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Jul 10 00:25:36.768936 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Jul 10 00:25:36.768943 kernel: 00:05: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Jul 10 00:25:36.768949 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBC,PNP0f13:MOUS] at 0x60,0x64 irq 1,12 Jul 10 00:25:36.768955 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Jul 10 00:25:36.768961 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Jul 10 00:25:36.769010 kernel: rtc_cmos 00:01: registered as rtc0 Jul 10 00:25:36.769059 kernel: rtc_cmos 00:01: setting system clock to 2025-07-10T00:25:36 UTC (1752107136) Jul 10 00:25:36.769103 kernel: rtc_cmos 00:01: alarms up to one month, y3k, 114 bytes nvram Jul 10 00:25:36.769112 kernel: intel_pstate: CPU model not supported Jul 10 00:25:36.769118 kernel: NET: Registered PF_INET6 protocol family Jul 10 00:25:36.769124 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Jul 10 00:25:36.769132 kernel: Segment Routing with IPv6 Jul 10 00:25:36.769138 kernel: In-situ OAM (IOAM) with IPv6 Jul 10 00:25:36.769144 kernel: NET: Registered PF_PACKET protocol family Jul 10 00:25:36.769151 kernel: Key type dns_resolver registered Jul 10 00:25:36.769158 kernel: IPI shorthand broadcast: enabled Jul 10 00:25:36.769164 kernel: sched_clock: Marking stable (2666451972, 170443250)->(2851784368, -14889146) Jul 10 00:25:36.769170 kernel: registered taskstats version 1 Jul 10 00:25:36.769176 kernel: Loading compiled-in X.509 certificates Jul 10 00:25:36.769182 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.36-flatcar: f515550de55d4e43b2ea11ae212aa0cb3a4e55cf' Jul 10 00:25:36.769188 kernel: Demotion targets for Node 0: null Jul 10 00:25:36.769194 kernel: Key type .fscrypt registered Jul 10 00:25:36.769200 kernel: Key type fscrypt-provisioning registered Jul 10 00:25:36.769207 kernel: ima: No TPM chip found, activating TPM-bypass! Jul 10 00:25:36.769213 kernel: ima: Allocated hash algorithm: sha1 Jul 10 00:25:36.769219 kernel: ima: No architecture policies found Jul 10 00:25:36.769225 kernel: clk: Disabling unused clocks Jul 10 00:25:36.769231 kernel: Warning: unable to open an initial console. Jul 10 00:25:36.769238 kernel: Freeing unused kernel image (initmem) memory: 54420K Jul 10 00:25:36.769243 kernel: Write protecting the kernel read-only data: 24576k Jul 10 00:25:36.769249 kernel: Freeing unused kernel image (rodata/data gap) memory: 284K Jul 10 00:25:36.769255 kernel: Run /init as init process Jul 10 00:25:36.769263 kernel: with arguments: Jul 10 00:25:36.769269 kernel: /init Jul 10 00:25:36.769275 kernel: with environment: Jul 10 00:25:36.769281 kernel: HOME=/ Jul 10 00:25:36.769286 kernel: TERM=linux Jul 10 00:25:36.769292 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Jul 10 00:25:36.769299 systemd[1]: Successfully made /usr/ read-only. Jul 10 00:25:36.769307 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Jul 10 00:25:36.769315 systemd[1]: Detected virtualization vmware. Jul 10 00:25:36.769321 systemd[1]: Detected architecture x86-64. Jul 10 00:25:36.769327 systemd[1]: Running in initrd. Jul 10 00:25:36.769333 systemd[1]: No hostname configured, using default hostname. Jul 10 00:25:36.769339 systemd[1]: Hostname set to . Jul 10 00:25:36.769345 systemd[1]: Initializing machine ID from random generator. Jul 10 00:25:36.769351 systemd[1]: Queued start job for default target initrd.target. Jul 10 00:25:36.769357 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jul 10 00:25:36.769363 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jul 10 00:25:36.769559 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Jul 10 00:25:36.769567 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jul 10 00:25:36.769573 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Jul 10 00:25:36.769580 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Jul 10 00:25:36.769587 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Jul 10 00:25:36.769593 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Jul 10 00:25:36.769602 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jul 10 00:25:36.769608 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jul 10 00:25:36.769614 systemd[1]: Reached target paths.target - Path Units. Jul 10 00:25:36.769620 systemd[1]: Reached target slices.target - Slice Units. Jul 10 00:25:36.769627 systemd[1]: Reached target swap.target - Swaps. Jul 10 00:25:36.769633 systemd[1]: Reached target timers.target - Timer Units. Jul 10 00:25:36.769639 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Jul 10 00:25:36.769645 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jul 10 00:25:36.769651 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Jul 10 00:25:36.769659 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Jul 10 00:25:36.769665 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jul 10 00:25:36.769671 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jul 10 00:25:36.769678 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jul 10 00:25:36.769684 systemd[1]: Reached target sockets.target - Socket Units. Jul 10 00:25:36.769690 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Jul 10 00:25:36.769696 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jul 10 00:25:36.769703 systemd[1]: Finished network-cleanup.service - Network Cleanup. Jul 10 00:25:36.769710 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Jul 10 00:25:36.769717 systemd[1]: Starting systemd-fsck-usr.service... Jul 10 00:25:36.769723 systemd[1]: Starting systemd-journald.service - Journal Service... Jul 10 00:25:36.769729 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jul 10 00:25:36.769735 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jul 10 00:25:36.769742 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Jul 10 00:25:36.769750 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jul 10 00:25:36.769756 systemd[1]: Finished systemd-fsck-usr.service. Jul 10 00:25:36.769775 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Jul 10 00:25:36.769782 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jul 10 00:25:36.769788 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jul 10 00:25:36.769807 systemd-journald[242]: Collecting audit messages is disabled. Jul 10 00:25:36.769825 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jul 10 00:25:36.769981 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jul 10 00:25:36.769989 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jul 10 00:25:36.769996 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Jul 10 00:25:36.770002 kernel: Bridge firewalling registered Jul 10 00:25:36.770010 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jul 10 00:25:36.770017 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jul 10 00:25:36.770023 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jul 10 00:25:36.770030 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Jul 10 00:25:36.770036 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jul 10 00:25:36.770044 systemd-journald[242]: Journal started Jul 10 00:25:36.770061 systemd-journald[242]: Runtime Journal (/run/log/journal/92785cf8c59642e8b4517f8631ec12e6) is 4.8M, max 38.8M, 34M free. Jul 10 00:25:36.720520 systemd-modules-load[244]: Inserted module 'overlay' Jul 10 00:25:36.743521 systemd-modules-load[244]: Inserted module 'br_netfilter' Jul 10 00:25:36.771387 systemd[1]: Started systemd-journald.service - Journal Service. Jul 10 00:25:36.773293 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jul 10 00:25:36.777577 dracut-cmdline[268]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=vmware flatcar.autologin verity.usrhash=844005237fb9709f65a093d5533c4229fb6c54e8e257736d9c3d041b6d3080ea Jul 10 00:25:36.782027 systemd-tmpfiles[273]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Jul 10 00:25:36.784593 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jul 10 00:25:36.785510 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jul 10 00:25:36.809717 systemd-resolved[310]: Positive Trust Anchors: Jul 10 00:25:36.809905 systemd-resolved[310]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jul 10 00:25:36.810049 systemd-resolved[310]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jul 10 00:25:36.812217 systemd-resolved[310]: Defaulting to hostname 'linux'. Jul 10 00:25:36.812898 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jul 10 00:25:36.813145 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jul 10 00:25:36.825380 kernel: SCSI subsystem initialized Jul 10 00:25:36.840381 kernel: Loading iSCSI transport class v2.0-870. Jul 10 00:25:36.848380 kernel: iscsi: registered transport (tcp) Jul 10 00:25:36.870392 kernel: iscsi: registered transport (qla4xxx) Jul 10 00:25:36.870411 kernel: QLogic iSCSI HBA Driver Jul 10 00:25:36.880039 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Jul 10 00:25:36.892860 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Jul 10 00:25:36.893788 systemd[1]: Reached target network-pre.target - Preparation for Network. Jul 10 00:25:36.916235 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Jul 10 00:25:36.917446 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Jul 10 00:25:36.957400 kernel: raid6: avx2x4 gen() 46491 MB/s Jul 10 00:25:36.974380 kernel: raid6: avx2x2 gen() 53226 MB/s Jul 10 00:25:36.991502 kernel: raid6: avx2x1 gen() 45669 MB/s Jul 10 00:25:36.991518 kernel: raid6: using algorithm avx2x2 gen() 53226 MB/s Jul 10 00:25:37.009569 kernel: raid6: .... xor() 32518 MB/s, rmw enabled Jul 10 00:25:37.009584 kernel: raid6: using avx2x2 recovery algorithm Jul 10 00:25:37.023382 kernel: xor: automatically using best checksumming function avx Jul 10 00:25:37.125386 kernel: Btrfs loaded, zoned=no, fsverity=no Jul 10 00:25:37.129108 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Jul 10 00:25:37.130039 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jul 10 00:25:37.152964 systemd-udevd[490]: Using default interface naming scheme 'v255'. Jul 10 00:25:37.156815 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jul 10 00:25:37.158092 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Jul 10 00:25:37.184051 dracut-pre-trigger[496]: rd.md=0: removing MD RAID activation Jul 10 00:25:37.201289 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Jul 10 00:25:37.202485 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jul 10 00:25:37.278733 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jul 10 00:25:37.280195 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Jul 10 00:25:37.341383 kernel: VMware PVSCSI driver - version 1.0.7.0-k Jul 10 00:25:37.343273 kernel: vmw_pvscsi: using 64bit dma Jul 10 00:25:37.343294 kernel: vmw_pvscsi: max_id: 16 Jul 10 00:25:37.343302 kernel: vmw_pvscsi: setting ring_pages to 8 Jul 10 00:25:37.349386 kernel: VMware vmxnet3 virtual NIC driver - version 1.9.0.0-k-NAPI Jul 10 00:25:37.356707 kernel: vmw_pvscsi: enabling reqCallThreshold Jul 10 00:25:37.356724 kernel: vmw_pvscsi: driver-based request coalescing enabled Jul 10 00:25:37.356736 kernel: vmw_pvscsi: using MSI-X Jul 10 00:25:37.359440 kernel: vmxnet3 0000:0b:00.0: # of Tx queues : 2, # of Rx queues : 2 Jul 10 00:25:37.363473 kernel: scsi host0: VMware PVSCSI storage adapter rev 2, req/cmp/msg rings: 8/8/1 pages, cmd_per_lun=254 Jul 10 00:25:37.363507 kernel: vmxnet3 0000:0b:00.0 eth0: NIC Link is Up 10000 Mbps Jul 10 00:25:37.369388 kernel: vmw_pvscsi 0000:03:00.0: VMware PVSCSI rev 2 host #0 Jul 10 00:25:37.373388 kernel: scsi 0:0:0:0: Direct-Access VMware Virtual disk 2.0 PQ: 0 ANSI: 6 Jul 10 00:25:37.380121 (udev-worker)[533]: id: Truncating stdout of 'dmi_memory_id' up to 16384 byte. Jul 10 00:25:37.380382 kernel: vmxnet3 0000:0b:00.0 ens192: renamed from eth0 Jul 10 00:25:37.384524 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jul 10 00:25:37.384598 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jul 10 00:25:37.388909 kernel: cryptd: max_cpu_qlen set to 1000 Jul 10 00:25:37.388931 kernel: sd 0:0:0:0: [sda] 17805312 512-byte logical blocks: (9.12 GB/8.49 GiB) Jul 10 00:25:37.389639 kernel: sd 0:0:0:0: [sda] Write Protect is off Jul 10 00:25:37.389715 kernel: sd 0:0:0:0: [sda] Mode Sense: 31 00 00 00 Jul 10 00:25:37.389790 kernel: sd 0:0:0:0: [sda] Cache data unavailable Jul 10 00:25:37.391139 kernel: sd 0:0:0:0: [sda] Assuming drive cache: write through Jul 10 00:25:37.392431 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Jul 10 00:25:37.394554 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jul 10 00:25:37.395386 kernel: libata version 3.00 loaded. Jul 10 00:25:37.398395 kernel: input: ImPS/2 Generic Wheel Mouse as /devices/platform/i8042/serio1/input/input2 Jul 10 00:25:37.403403 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jul 10 00:25:37.404420 kernel: sd 0:0:0:0: [sda] Attached SCSI disk Jul 10 00:25:37.407452 kernel: ata_piix 0000:00:07.1: version 2.13 Jul 10 00:25:37.407542 kernel: scsi host1: ata_piix Jul 10 00:25:37.407608 kernel: scsi host2: ata_piix Jul 10 00:25:37.407672 kernel: ata1: PATA max UDMA/33 cmd 0x1f0 ctl 0x3f6 bmdma 0x1060 irq 14 lpm-pol 0 Jul 10 00:25:37.410343 kernel: ata2: PATA max UDMA/33 cmd 0x170 ctl 0x376 bmdma 0x1068 irq 15 lpm-pol 0 Jul 10 00:25:37.415383 kernel: AES CTR mode by8 optimization enabled Jul 10 00:25:37.430793 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jul 10 00:25:37.456198 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Virtual_disk ROOT. Jul 10 00:25:37.461492 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Virtual_disk EFI-SYSTEM. Jul 10 00:25:37.466808 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Virtual_disk OEM. Jul 10 00:25:37.471023 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Virtual_disk USR-A. Jul 10 00:25:37.471264 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Virtual_disk USR-A. Jul 10 00:25:37.471937 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Jul 10 00:25:37.510387 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jul 10 00:25:37.524385 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jul 10 00:25:37.580682 kernel: ata2.00: ATAPI: VMware Virtual IDE CDROM Drive, 00000001, max UDMA/33 Jul 10 00:25:37.586406 kernel: scsi 2:0:0:0: CD-ROM NECVMWar VMware IDE CDR10 1.00 PQ: 0 ANSI: 5 Jul 10 00:25:37.613477 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 1x/1x writer dvd-ram cd/rw xa/form2 cdda tray Jul 10 00:25:37.613601 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Jul 10 00:25:37.626385 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Jul 10 00:25:37.986997 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Jul 10 00:25:37.987697 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Jul 10 00:25:37.988035 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jul 10 00:25:37.988348 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jul 10 00:25:37.989261 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Jul 10 00:25:38.006467 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Jul 10 00:25:38.525418 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jul 10 00:25:38.526954 disk-uuid[635]: The operation has completed successfully. Jul 10 00:25:38.564187 systemd[1]: disk-uuid.service: Deactivated successfully. Jul 10 00:25:38.564483 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Jul 10 00:25:38.580348 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Jul 10 00:25:38.593303 sh[673]: Success Jul 10 00:25:38.608131 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Jul 10 00:25:38.608165 kernel: device-mapper: uevent: version 1.0.3 Jul 10 00:25:38.608174 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Jul 10 00:25:38.615386 kernel: device-mapper: verity: sha256 using shash "sha256-avx2" Jul 10 00:25:38.682157 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Jul 10 00:25:38.684413 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Jul 10 00:25:38.694004 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Jul 10 00:25:38.713582 kernel: BTRFS info: 'norecovery' is for compatibility only, recommended to use 'rescue=nologreplay' Jul 10 00:25:38.713614 kernel: BTRFS: device fsid c4cb30b0-bb74-4f98-aab6-7a1c6f47edee devid 1 transid 36 /dev/mapper/usr (254:0) scanned by mount (685) Jul 10 00:25:38.715384 kernel: BTRFS info (device dm-0): first mount of filesystem c4cb30b0-bb74-4f98-aab6-7a1c6f47edee Jul 10 00:25:38.716723 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Jul 10 00:25:38.716741 kernel: BTRFS info (device dm-0): using free-space-tree Jul 10 00:25:38.731911 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Jul 10 00:25:38.732324 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Jul 10 00:25:38.733104 systemd[1]: Starting afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments... Jul 10 00:25:38.733707 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Jul 10 00:25:38.753379 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/sda6 (8:6) scanned by mount (708) Jul 10 00:25:38.756099 kernel: BTRFS info (device sda6): first mount of filesystem 66535909-6865-4f30-ad42-a3000fffd5f6 Jul 10 00:25:38.756118 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Jul 10 00:25:38.756130 kernel: BTRFS info (device sda6): using free-space-tree Jul 10 00:25:38.806685 kernel: BTRFS info (device sda6): last unmount of filesystem 66535909-6865-4f30-ad42-a3000fffd5f6 Jul 10 00:25:38.808115 systemd[1]: Finished ignition-setup.service - Ignition (setup). Jul 10 00:25:38.809452 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Jul 10 00:25:38.839620 systemd[1]: Finished afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments. Jul 10 00:25:38.840258 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Jul 10 00:25:38.918071 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jul 10 00:25:38.919069 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jul 10 00:25:38.925214 ignition[727]: Ignition 2.21.0 Jul 10 00:25:38.925456 ignition[727]: Stage: fetch-offline Jul 10 00:25:38.925571 ignition[727]: no configs at "/usr/lib/ignition/base.d" Jul 10 00:25:38.925681 ignition[727]: no config dir at "/usr/lib/ignition/base.platform.d/vmware" Jul 10 00:25:38.925830 ignition[727]: parsed url from cmdline: "" Jul 10 00:25:38.925832 ignition[727]: no config URL provided Jul 10 00:25:38.925835 ignition[727]: reading system config file "/usr/lib/ignition/user.ign" Jul 10 00:25:38.925839 ignition[727]: no config at "/usr/lib/ignition/user.ign" Jul 10 00:25:38.926199 ignition[727]: config successfully fetched Jul 10 00:25:38.926218 ignition[727]: parsing config with SHA512: fcd230554d0f94093e193e8bb61841514a0e5e12275651467045e5e19d2a4419c80b51a133b4de56158080e4bb15846016537d2cce0be6d323b6c503f5feea49 Jul 10 00:25:38.931532 unknown[727]: fetched base config from "system" Jul 10 00:25:38.931538 unknown[727]: fetched user config from "vmware" Jul 10 00:25:38.931735 ignition[727]: fetch-offline: fetch-offline passed Jul 10 00:25:38.931766 ignition[727]: Ignition finished successfully Jul 10 00:25:38.933914 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Jul 10 00:25:38.944187 systemd-networkd[864]: lo: Link UP Jul 10 00:25:38.944194 systemd-networkd[864]: lo: Gained carrier Jul 10 00:25:38.944963 systemd-networkd[864]: Enumeration completed Jul 10 00:25:38.945230 systemd[1]: Started systemd-networkd.service - Network Configuration. Jul 10 00:25:38.945423 systemd-networkd[864]: ens192: Configuring with /etc/systemd/network/10-dracut-cmdline-99.network. Jul 10 00:25:38.946480 systemd[1]: Reached target network.target - Network. Jul 10 00:25:38.949287 kernel: vmxnet3 0000:0b:00.0 ens192: intr type 3, mode 0, 3 vectors allocated Jul 10 00:25:38.949411 kernel: vmxnet3 0000:0b:00.0 ens192: NIC Link is Up 10000 Mbps Jul 10 00:25:38.947783 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Jul 10 00:25:38.948263 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Jul 10 00:25:38.948836 systemd-networkd[864]: ens192: Link UP Jul 10 00:25:38.948839 systemd-networkd[864]: ens192: Gained carrier Jul 10 00:25:38.961564 ignition[869]: Ignition 2.21.0 Jul 10 00:25:38.961792 ignition[869]: Stage: kargs Jul 10 00:25:38.961868 ignition[869]: no configs at "/usr/lib/ignition/base.d" Jul 10 00:25:38.961873 ignition[869]: no config dir at "/usr/lib/ignition/base.platform.d/vmware" Jul 10 00:25:38.962567 ignition[869]: kargs: kargs passed Jul 10 00:25:38.962600 ignition[869]: Ignition finished successfully Jul 10 00:25:38.964241 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Jul 10 00:25:38.964929 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Jul 10 00:25:38.982139 ignition[877]: Ignition 2.21.0 Jul 10 00:25:38.982146 ignition[877]: Stage: disks Jul 10 00:25:38.982238 ignition[877]: no configs at "/usr/lib/ignition/base.d" Jul 10 00:25:38.982243 ignition[877]: no config dir at "/usr/lib/ignition/base.platform.d/vmware" Jul 10 00:25:38.982737 ignition[877]: disks: disks passed Jul 10 00:25:38.982763 ignition[877]: Ignition finished successfully Jul 10 00:25:38.983550 systemd[1]: Finished ignition-disks.service - Ignition (disks). Jul 10 00:25:38.983900 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Jul 10 00:25:38.984032 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Jul 10 00:25:38.984214 systemd[1]: Reached target local-fs.target - Local File Systems. Jul 10 00:25:38.984445 systemd[1]: Reached target sysinit.target - System Initialization. Jul 10 00:25:38.984607 systemd[1]: Reached target basic.target - Basic System. Jul 10 00:25:38.985246 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Jul 10 00:25:39.002602 systemd-fsck[885]: ROOT: clean, 15/1628000 files, 120826/1617920 blocks Jul 10 00:25:39.003409 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Jul 10 00:25:39.005067 systemd[1]: Mounting sysroot.mount - /sysroot... Jul 10 00:25:39.083382 kernel: EXT4-fs (sda9): mounted filesystem a310c019-7915-47f5-9fce-db4a09ac26c2 r/w with ordered data mode. Quota mode: none. Jul 10 00:25:39.083476 systemd[1]: Mounted sysroot.mount - /sysroot. Jul 10 00:25:39.083933 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Jul 10 00:25:39.084962 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jul 10 00:25:39.086404 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Jul 10 00:25:39.086805 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Jul 10 00:25:39.086966 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Jul 10 00:25:39.086981 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Jul 10 00:25:39.096055 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Jul 10 00:25:39.096825 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Jul 10 00:25:39.101408 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/sda6 (8:6) scanned by mount (893) Jul 10 00:25:39.104173 kernel: BTRFS info (device sda6): first mount of filesystem 66535909-6865-4f30-ad42-a3000fffd5f6 Jul 10 00:25:39.104191 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Jul 10 00:25:39.104200 kernel: BTRFS info (device sda6): using free-space-tree Jul 10 00:25:39.108736 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jul 10 00:25:39.122330 initrd-setup-root[917]: cut: /sysroot/etc/passwd: No such file or directory Jul 10 00:25:39.125027 initrd-setup-root[924]: cut: /sysroot/etc/group: No such file or directory Jul 10 00:25:39.127086 initrd-setup-root[931]: cut: /sysroot/etc/shadow: No such file or directory Jul 10 00:25:39.129053 initrd-setup-root[938]: cut: /sysroot/etc/gshadow: No such file or directory Jul 10 00:25:39.215487 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Jul 10 00:25:39.216293 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Jul 10 00:25:39.217445 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Jul 10 00:25:39.237390 kernel: BTRFS info (device sda6): last unmount of filesystem 66535909-6865-4f30-ad42-a3000fffd5f6 Jul 10 00:25:39.252794 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Jul 10 00:25:39.254486 ignition[1006]: INFO : Ignition 2.21.0 Jul 10 00:25:39.254709 ignition[1006]: INFO : Stage: mount Jul 10 00:25:39.254898 ignition[1006]: INFO : no configs at "/usr/lib/ignition/base.d" Jul 10 00:25:39.255020 ignition[1006]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/vmware" Jul 10 00:25:39.255587 ignition[1006]: INFO : mount: mount passed Jul 10 00:25:39.255711 ignition[1006]: INFO : Ignition finished successfully Jul 10 00:25:39.256569 systemd[1]: Finished ignition-mount.service - Ignition (mount). Jul 10 00:25:39.257443 systemd[1]: Starting ignition-files.service - Ignition (files)... Jul 10 00:25:39.710977 systemd[1]: sysroot-oem.mount: Deactivated successfully. Jul 10 00:25:39.711938 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jul 10 00:25:39.737420 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/sda6 (8:6) scanned by mount (1018) Jul 10 00:25:39.740245 kernel: BTRFS info (device sda6): first mount of filesystem 66535909-6865-4f30-ad42-a3000fffd5f6 Jul 10 00:25:39.740263 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Jul 10 00:25:39.740271 kernel: BTRFS info (device sda6): using free-space-tree Jul 10 00:25:39.744593 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jul 10 00:25:39.762767 ignition[1035]: INFO : Ignition 2.21.0 Jul 10 00:25:39.762767 ignition[1035]: INFO : Stage: files Jul 10 00:25:39.763542 ignition[1035]: INFO : no configs at "/usr/lib/ignition/base.d" Jul 10 00:25:39.763542 ignition[1035]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/vmware" Jul 10 00:25:39.764019 ignition[1035]: DEBUG : files: compiled without relabeling support, skipping Jul 10 00:25:39.764830 ignition[1035]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Jul 10 00:25:39.764830 ignition[1035]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Jul 10 00:25:39.767025 ignition[1035]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Jul 10 00:25:39.767281 ignition[1035]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Jul 10 00:25:39.767606 unknown[1035]: wrote ssh authorized keys file for user: core Jul 10 00:25:39.767888 ignition[1035]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Jul 10 00:25:39.771049 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Jul 10 00:25:39.771049 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-amd64.tar.gz: attempt #1 Jul 10 00:25:39.806625 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Jul 10 00:25:39.878236 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Jul 10 00:25:39.878512 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Jul 10 00:25:39.878512 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Jul 10 00:25:39.878512 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Jul 10 00:25:39.878512 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Jul 10 00:25:39.878512 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Jul 10 00:25:39.879523 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Jul 10 00:25:39.879523 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Jul 10 00:25:39.879523 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Jul 10 00:25:39.880082 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Jul 10 00:25:39.880082 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Jul 10 00:25:39.880082 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Jul 10 00:25:39.882551 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Jul 10 00:25:39.882772 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Jul 10 00:25:39.882772 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.31.8-x86-64.raw: attempt #1 Jul 10 00:25:40.173636 systemd-networkd[864]: ens192: Gained IPv6LL Jul 10 00:25:40.574858 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Jul 10 00:25:40.807449 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Jul 10 00:25:40.807449 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/etc/systemd/network/00-vmware.network" Jul 10 00:25:40.808547 ignition[1035]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/etc/systemd/network/00-vmware.network" Jul 10 00:25:40.808547 ignition[1035]: INFO : files: op(c): [started] processing unit "prepare-helm.service" Jul 10 00:25:40.809110 ignition[1035]: INFO : files: op(c): op(d): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jul 10 00:25:40.809660 ignition[1035]: INFO : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jul 10 00:25:40.809660 ignition[1035]: INFO : files: op(c): [finished] processing unit "prepare-helm.service" Jul 10 00:25:40.809660 ignition[1035]: INFO : files: op(e): [started] processing unit "coreos-metadata.service" Jul 10 00:25:40.809660 ignition[1035]: INFO : files: op(e): op(f): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Jul 10 00:25:40.809660 ignition[1035]: INFO : files: op(e): op(f): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Jul 10 00:25:40.809660 ignition[1035]: INFO : files: op(e): [finished] processing unit "coreos-metadata.service" Jul 10 00:25:40.809660 ignition[1035]: INFO : files: op(10): [started] setting preset to disabled for "coreos-metadata.service" Jul 10 00:25:40.833659 ignition[1035]: INFO : files: op(10): op(11): [started] removing enablement symlink(s) for "coreos-metadata.service" Jul 10 00:25:40.835978 ignition[1035]: INFO : files: op(10): op(11): [finished] removing enablement symlink(s) for "coreos-metadata.service" Jul 10 00:25:40.836214 ignition[1035]: INFO : files: op(10): [finished] setting preset to disabled for "coreos-metadata.service" Jul 10 00:25:40.836214 ignition[1035]: INFO : files: op(12): [started] setting preset to enabled for "prepare-helm.service" Jul 10 00:25:40.836214 ignition[1035]: INFO : files: op(12): [finished] setting preset to enabled for "prepare-helm.service" Jul 10 00:25:40.836214 ignition[1035]: INFO : files: createResultFile: createFiles: op(13): [started] writing file "/sysroot/etc/.ignition-result.json" Jul 10 00:25:40.837500 ignition[1035]: INFO : files: createResultFile: createFiles: op(13): [finished] writing file "/sysroot/etc/.ignition-result.json" Jul 10 00:25:40.837500 ignition[1035]: INFO : files: files passed Jul 10 00:25:40.837500 ignition[1035]: INFO : Ignition finished successfully Jul 10 00:25:40.838061 systemd[1]: Finished ignition-files.service - Ignition (files). Jul 10 00:25:40.840459 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Jul 10 00:25:40.841025 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Jul 10 00:25:40.850115 initrd-setup-root-after-ignition[1066]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jul 10 00:25:40.850115 initrd-setup-root-after-ignition[1066]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Jul 10 00:25:40.851140 systemd[1]: ignition-quench.service: Deactivated successfully. Jul 10 00:25:40.851613 initrd-setup-root-after-ignition[1071]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jul 10 00:25:40.851796 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Jul 10 00:25:40.852991 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Jul 10 00:25:40.853529 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Jul 10 00:25:40.854225 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Jul 10 00:25:40.890637 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Jul 10 00:25:40.890906 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Jul 10 00:25:40.891444 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Jul 10 00:25:40.891890 systemd[1]: Reached target initrd.target - Initrd Default Target. Jul 10 00:25:40.892170 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Jul 10 00:25:40.892901 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Jul 10 00:25:40.908285 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jul 10 00:25:40.909148 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Jul 10 00:25:40.926619 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Jul 10 00:25:40.927014 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Jul 10 00:25:40.927351 systemd[1]: Stopped target timers.target - Timer Units. Jul 10 00:25:40.927663 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Jul 10 00:25:40.927864 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jul 10 00:25:40.928304 systemd[1]: Stopped target initrd.target - Initrd Default Target. Jul 10 00:25:40.928626 systemd[1]: Stopped target basic.target - Basic System. Jul 10 00:25:40.928925 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Jul 10 00:25:40.929278 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Jul 10 00:25:40.929643 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Jul 10 00:25:40.930001 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Jul 10 00:25:40.930363 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Jul 10 00:25:40.930715 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Jul 10 00:25:40.931071 systemd[1]: Stopped target sysinit.target - System Initialization. Jul 10 00:25:40.931346 systemd[1]: Stopped target local-fs.target - Local File Systems. Jul 10 00:25:40.931670 systemd[1]: Stopped target swap.target - Swaps. Jul 10 00:25:40.931896 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Jul 10 00:25:40.931980 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Jul 10 00:25:40.932519 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Jul 10 00:25:40.932826 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jul 10 00:25:40.933119 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Jul 10 00:25:40.933282 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jul 10 00:25:40.933597 systemd[1]: dracut-initqueue.service: Deactivated successfully. Jul 10 00:25:40.933667 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Jul 10 00:25:40.934137 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Jul 10 00:25:40.934208 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Jul 10 00:25:40.934703 systemd[1]: Stopped target paths.target - Path Units. Jul 10 00:25:40.934954 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Jul 10 00:25:40.935149 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jul 10 00:25:40.935500 systemd[1]: Stopped target slices.target - Slice Units. Jul 10 00:25:40.935776 systemd[1]: Stopped target sockets.target - Socket Units. Jul 10 00:25:40.936047 systemd[1]: iscsid.socket: Deactivated successfully. Jul 10 00:25:40.936106 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Jul 10 00:25:40.936522 systemd[1]: iscsiuio.socket: Deactivated successfully. Jul 10 00:25:40.936580 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jul 10 00:25:40.936889 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Jul 10 00:25:40.936967 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Jul 10 00:25:40.937582 systemd[1]: ignition-files.service: Deactivated successfully. Jul 10 00:25:40.937651 systemd[1]: Stopped ignition-files.service - Ignition (files). Jul 10 00:25:40.938433 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Jul 10 00:25:40.938550 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Jul 10 00:25:40.938616 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Jul 10 00:25:40.939193 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Jul 10 00:25:40.940418 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Jul 10 00:25:40.940518 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Jul 10 00:25:40.940733 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Jul 10 00:25:40.940796 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Jul 10 00:25:40.943134 systemd[1]: initrd-cleanup.service: Deactivated successfully. Jul 10 00:25:40.951003 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Jul 10 00:25:40.959077 systemd[1]: sysroot-boot.mount: Deactivated successfully. Jul 10 00:25:40.964814 ignition[1091]: INFO : Ignition 2.21.0 Jul 10 00:25:40.964814 ignition[1091]: INFO : Stage: umount Jul 10 00:25:40.965618 ignition[1091]: INFO : no configs at "/usr/lib/ignition/base.d" Jul 10 00:25:40.965618 ignition[1091]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/vmware" Jul 10 00:25:40.966288 ignition[1091]: INFO : umount: umount passed Jul 10 00:25:40.966952 ignition[1091]: INFO : Ignition finished successfully Jul 10 00:25:40.967853 systemd[1]: ignition-mount.service: Deactivated successfully. Jul 10 00:25:40.967934 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Jul 10 00:25:40.968191 systemd[1]: Stopped target network.target - Network. Jul 10 00:25:40.968297 systemd[1]: ignition-disks.service: Deactivated successfully. Jul 10 00:25:40.968328 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Jul 10 00:25:40.968481 systemd[1]: ignition-kargs.service: Deactivated successfully. Jul 10 00:25:40.968504 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Jul 10 00:25:40.968659 systemd[1]: ignition-setup.service: Deactivated successfully. Jul 10 00:25:40.968681 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Jul 10 00:25:40.968833 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Jul 10 00:25:40.968855 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Jul 10 00:25:40.969065 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Jul 10 00:25:40.969575 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Jul 10 00:25:40.974447 systemd[1]: systemd-resolved.service: Deactivated successfully. Jul 10 00:25:40.974513 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Jul 10 00:25:40.975972 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Jul 10 00:25:40.976112 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Jul 10 00:25:40.976138 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Jul 10 00:25:40.977296 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Jul 10 00:25:40.977674 systemd[1]: systemd-networkd.service: Deactivated successfully. Jul 10 00:25:40.977738 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Jul 10 00:25:40.978702 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Jul 10 00:25:40.978883 systemd[1]: Stopped target network-pre.target - Preparation for Network. Jul 10 00:25:40.979026 systemd[1]: systemd-networkd.socket: Deactivated successfully. Jul 10 00:25:40.979047 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Jul 10 00:25:40.979839 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Jul 10 00:25:40.979930 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Jul 10 00:25:40.979957 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jul 10 00:25:40.980081 systemd[1]: afterburn-network-kargs.service: Deactivated successfully. Jul 10 00:25:40.980104 systemd[1]: Stopped afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments. Jul 10 00:25:40.980219 systemd[1]: systemd-sysctl.service: Deactivated successfully. Jul 10 00:25:40.980242 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Jul 10 00:25:40.980448 systemd[1]: systemd-modules-load.service: Deactivated successfully. Jul 10 00:25:40.980471 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Jul 10 00:25:40.980589 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Jul 10 00:25:40.981414 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Jul 10 00:25:40.991214 systemd[1]: network-cleanup.service: Deactivated successfully. Jul 10 00:25:40.991293 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Jul 10 00:25:40.993739 systemd[1]: systemd-udevd.service: Deactivated successfully. Jul 10 00:25:40.993822 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Jul 10 00:25:40.994074 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Jul 10 00:25:40.994097 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Jul 10 00:25:40.994216 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Jul 10 00:25:40.994231 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Jul 10 00:25:40.994344 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Jul 10 00:25:40.995315 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Jul 10 00:25:40.995628 systemd[1]: dracut-cmdline.service: Deactivated successfully. Jul 10 00:25:40.995763 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Jul 10 00:25:40.996040 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jul 10 00:25:40.996063 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jul 10 00:25:40.996948 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Jul 10 00:25:40.997211 systemd[1]: systemd-network-generator.service: Deactivated successfully. Jul 10 00:25:40.997435 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Jul 10 00:25:40.998453 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Jul 10 00:25:40.998481 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jul 10 00:25:40.998944 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jul 10 00:25:40.998970 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jul 10 00:25:41.005539 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Jul 10 00:25:41.005597 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Jul 10 00:25:41.020813 systemd[1]: sysroot-boot.service: Deactivated successfully. Jul 10 00:25:41.020890 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Jul 10 00:25:41.021333 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Jul 10 00:25:41.021541 systemd[1]: initrd-setup-root.service: Deactivated successfully. Jul 10 00:25:41.021581 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Jul 10 00:25:41.022365 systemd[1]: Starting initrd-switch-root.service - Switch Root... Jul 10 00:25:41.040995 systemd[1]: Switching root. Jul 10 00:25:41.079067 systemd-journald[242]: Journal stopped Jul 10 00:25:42.307504 systemd-journald[242]: Received SIGTERM from PID 1 (systemd). Jul 10 00:25:42.307524 kernel: SELinux: policy capability network_peer_controls=1 Jul 10 00:25:42.307532 kernel: SELinux: policy capability open_perms=1 Jul 10 00:25:42.307538 kernel: SELinux: policy capability extended_socket_class=1 Jul 10 00:25:42.307543 kernel: SELinux: policy capability always_check_network=0 Jul 10 00:25:42.307550 kernel: SELinux: policy capability cgroup_seclabel=1 Jul 10 00:25:42.307556 kernel: SELinux: policy capability nnp_nosuid_transition=1 Jul 10 00:25:42.307562 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Jul 10 00:25:42.307567 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Jul 10 00:25:42.307573 kernel: SELinux: policy capability userspace_initial_context=0 Jul 10 00:25:42.307579 kernel: audit: type=1403 audit(1752107141.678:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Jul 10 00:25:42.307586 systemd[1]: Successfully loaded SELinux policy in 49.054ms. Jul 10 00:25:42.307594 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 7.548ms. Jul 10 00:25:42.307601 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Jul 10 00:25:42.307608 systemd[1]: Detected virtualization vmware. Jul 10 00:25:42.307614 systemd[1]: Detected architecture x86-64. Jul 10 00:25:42.307621 systemd[1]: Detected first boot. Jul 10 00:25:42.307628 systemd[1]: Initializing machine ID from random generator. Jul 10 00:25:42.307634 zram_generator::config[1135]: No configuration found. Jul 10 00:25:42.307716 kernel: vmw_vmci 0000:00:07.7: Using capabilities 0xc Jul 10 00:25:42.307726 kernel: Guest personality initialized and is active Jul 10 00:25:42.307732 kernel: VMCI host device registered (name=vmci, major=10, minor=125) Jul 10 00:25:42.307738 kernel: Initialized host personality Jul 10 00:25:42.307746 kernel: NET: Registered PF_VSOCK protocol family Jul 10 00:25:42.307752 systemd[1]: Populated /etc with preset unit settings. Jul 10 00:25:42.307760 systemd[1]: /etc/systemd/system/coreos-metadata.service:11: Ignoring unknown escape sequences: "echo "COREOS_CUSTOM_PRIVATE_IPV4=$(ip addr show ens192 | grep "inet 10." | grep -Po "inet \K[\d.]+") Jul 10 00:25:42.307767 systemd[1]: COREOS_CUSTOM_PUBLIC_IPV4=$(ip addr show ens192 | grep -v "inet 10." | grep -Po "inet \K[\d.]+")" > ${OUTPUT}" Jul 10 00:25:42.307774 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Jul 10 00:25:42.307780 systemd[1]: initrd-switch-root.service: Deactivated successfully. Jul 10 00:25:42.307786 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Jul 10 00:25:42.307793 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Jul 10 00:25:42.307800 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Jul 10 00:25:42.307807 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Jul 10 00:25:42.307814 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Jul 10 00:25:42.307820 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Jul 10 00:25:42.307827 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Jul 10 00:25:42.307834 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Jul 10 00:25:42.307841 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Jul 10 00:25:42.307848 systemd[1]: Created slice user.slice - User and Session Slice. Jul 10 00:25:42.307855 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jul 10 00:25:42.307863 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jul 10 00:25:42.307870 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Jul 10 00:25:42.307877 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Jul 10 00:25:42.307884 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Jul 10 00:25:42.307891 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jul 10 00:25:42.307898 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Jul 10 00:25:42.307905 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jul 10 00:25:42.307911 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jul 10 00:25:42.307918 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Jul 10 00:25:42.307925 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Jul 10 00:25:42.307931 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Jul 10 00:25:42.307938 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Jul 10 00:25:42.307944 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jul 10 00:25:42.307952 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jul 10 00:25:42.307958 systemd[1]: Reached target slices.target - Slice Units. Jul 10 00:25:42.307965 systemd[1]: Reached target swap.target - Swaps. Jul 10 00:25:42.307971 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Jul 10 00:25:42.307978 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Jul 10 00:25:42.307986 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Jul 10 00:25:42.307993 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jul 10 00:25:42.308000 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jul 10 00:25:42.308006 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jul 10 00:25:42.308013 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Jul 10 00:25:42.308020 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Jul 10 00:25:42.308026 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Jul 10 00:25:42.308033 systemd[1]: Mounting media.mount - External Media Directory... Jul 10 00:25:42.308041 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jul 10 00:25:42.308047 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Jul 10 00:25:42.308054 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Jul 10 00:25:42.308061 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Jul 10 00:25:42.308068 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Jul 10 00:25:42.308074 systemd[1]: Reached target machines.target - Containers. Jul 10 00:25:42.308081 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Jul 10 00:25:42.308089 systemd[1]: Starting ignition-delete-config.service - Ignition (delete config)... Jul 10 00:25:42.308097 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jul 10 00:25:42.308103 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Jul 10 00:25:42.308110 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jul 10 00:25:42.308117 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jul 10 00:25:42.308124 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jul 10 00:25:42.308131 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Jul 10 00:25:42.308137 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jul 10 00:25:42.308144 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Jul 10 00:25:42.308152 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Jul 10 00:25:42.308159 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Jul 10 00:25:42.308165 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Jul 10 00:25:42.308172 systemd[1]: Stopped systemd-fsck-usr.service. Jul 10 00:25:42.308179 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Jul 10 00:25:42.308186 systemd[1]: Starting systemd-journald.service - Journal Service... Jul 10 00:25:42.308193 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jul 10 00:25:42.308199 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Jul 10 00:25:42.308206 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Jul 10 00:25:42.308214 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Jul 10 00:25:42.308221 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jul 10 00:25:42.308227 systemd[1]: verity-setup.service: Deactivated successfully. Jul 10 00:25:42.308234 systemd[1]: Stopped verity-setup.service. Jul 10 00:25:42.308240 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jul 10 00:25:42.308247 kernel: loop: module loaded Jul 10 00:25:42.308253 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Jul 10 00:25:42.308260 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Jul 10 00:25:42.308267 systemd[1]: Mounted media.mount - External Media Directory. Jul 10 00:25:42.308274 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Jul 10 00:25:42.308281 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Jul 10 00:25:42.308288 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Jul 10 00:25:42.308294 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jul 10 00:25:42.308301 systemd[1]: modprobe@configfs.service: Deactivated successfully. Jul 10 00:25:42.308308 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Jul 10 00:25:42.308314 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jul 10 00:25:42.308321 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jul 10 00:25:42.308329 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jul 10 00:25:42.308339 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jul 10 00:25:42.308347 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jul 10 00:25:42.308354 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Jul 10 00:25:42.308360 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Jul 10 00:25:42.308375 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Jul 10 00:25:42.308427 systemd[1]: Reached target local-fs.target - Local File Systems. Jul 10 00:25:42.308434 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Jul 10 00:25:42.308443 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Jul 10 00:25:42.308450 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jul 10 00:25:42.308464 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Jul 10 00:25:42.308487 systemd-journald[1222]: Collecting audit messages is disabled. Jul 10 00:25:42.308503 systemd-journald[1222]: Journal started Jul 10 00:25:42.308519 systemd-journald[1222]: Runtime Journal (/run/log/journal/4e590c1d797f4b36b65dc6bd552c21f6) is 4.8M, max 38.8M, 34M free. Jul 10 00:25:42.118779 systemd[1]: Queued start job for default target multi-user.target. Jul 10 00:25:42.140880 systemd[1]: Unnecessary job was removed for dev-sda6.device - /dev/sda6. Jul 10 00:25:42.141172 systemd[1]: systemd-journald.service: Deactivated successfully. Jul 10 00:25:42.308990 jq[1206]: true Jul 10 00:25:42.310632 kernel: fuse: init (API version 7.41) Jul 10 00:25:42.315381 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jul 10 00:25:42.318378 jq[1241]: true Jul 10 00:25:42.325215 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Jul 10 00:25:42.327382 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jul 10 00:25:42.334677 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Jul 10 00:25:42.334704 systemd[1]: Started systemd-journald.service - Journal Service. Jul 10 00:25:42.342463 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Jul 10 00:25:42.342755 systemd[1]: modprobe@fuse.service: Deactivated successfully. Jul 10 00:25:42.342862 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Jul 10 00:25:42.343087 systemd[1]: modprobe@loop.service: Deactivated successfully. Jul 10 00:25:42.343185 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jul 10 00:25:42.343643 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Jul 10 00:25:42.344156 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Jul 10 00:25:42.346542 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Jul 10 00:25:42.352450 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Jul 10 00:25:42.359079 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Jul 10 00:25:42.360522 systemd[1]: Reached target network-pre.target - Preparation for Network. Jul 10 00:25:42.370178 kernel: loop0: detected capacity change from 0 to 221472 Jul 10 00:25:42.369430 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Jul 10 00:25:42.371474 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Jul 10 00:25:42.384235 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Jul 10 00:25:42.384445 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jul 10 00:25:42.388435 systemd[1]: Starting systemd-sysusers.service - Create System Users... Jul 10 00:25:42.389292 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Jul 10 00:25:42.391488 kernel: ACPI: bus type drm_connector registered Jul 10 00:25:42.392593 systemd[1]: modprobe@drm.service: Deactivated successfully. Jul 10 00:25:42.402194 systemd-journald[1222]: Time spent on flushing to /var/log/journal/4e590c1d797f4b36b65dc6bd552c21f6 is 62.251ms for 1758 entries. Jul 10 00:25:42.402194 systemd-journald[1222]: System Journal (/var/log/journal/4e590c1d797f4b36b65dc6bd552c21f6) is 8M, max 584.8M, 576.8M free. Jul 10 00:25:42.467698 systemd-journald[1222]: Received client request to flush runtime journal. Jul 10 00:25:42.467726 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Jul 10 00:25:42.424767 ignition[1259]: Ignition 2.21.0 Jul 10 00:25:42.403110 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jul 10 00:25:42.424922 ignition[1259]: deleting config from guestinfo properties Jul 10 00:25:42.403908 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jul 10 00:25:42.441652 ignition[1259]: Successfully deleted config Jul 10 00:25:42.445509 systemd[1]: Finished ignition-delete-config.service - Ignition (delete config). Jul 10 00:25:42.469073 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Jul 10 00:25:42.487608 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jul 10 00:25:42.516083 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Jul 10 00:25:42.525660 systemd[1]: Finished systemd-sysusers.service - Create System Users. Jul 10 00:25:42.528618 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jul 10 00:25:42.531445 kernel: loop1: detected capacity change from 0 to 113872 Jul 10 00:25:42.557994 kernel: loop2: detected capacity change from 0 to 146240 Jul 10 00:25:42.562787 systemd-tmpfiles[1304]: ACLs are not supported, ignoring. Jul 10 00:25:42.562799 systemd-tmpfiles[1304]: ACLs are not supported, ignoring. Jul 10 00:25:42.570768 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jul 10 00:25:42.624389 kernel: loop3: detected capacity change from 0 to 2960 Jul 10 00:25:42.669392 kernel: loop4: detected capacity change from 0 to 221472 Jul 10 00:25:42.701394 kernel: loop5: detected capacity change from 0 to 113872 Jul 10 00:25:42.716539 kernel: loop6: detected capacity change from 0 to 146240 Jul 10 00:25:42.748397 kernel: loop7: detected capacity change from 0 to 2960 Jul 10 00:25:42.757149 (sd-merge)[1310]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-vmware'. Jul 10 00:25:42.757803 (sd-merge)[1310]: Merged extensions into '/usr'. Jul 10 00:25:42.762404 systemd[1]: Reload requested from client PID 1258 ('systemd-sysext') (unit systemd-sysext.service)... Jul 10 00:25:42.762412 systemd[1]: Reloading... Jul 10 00:25:42.822407 zram_generator::config[1333]: No configuration found. Jul 10 00:25:42.923847 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jul 10 00:25:42.933967 systemd[1]: /etc/systemd/system/coreos-metadata.service:11: Ignoring unknown escape sequences: "echo "COREOS_CUSTOM_PRIVATE_IPV4=$(ip addr show ens192 | grep "inet 10." | grep -Po "inet \K[\d.]+") Jul 10 00:25:42.982747 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Jul 10 00:25:42.982931 systemd[1]: Reloading finished in 220 ms. Jul 10 00:25:42.997321 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Jul 10 00:25:43.006461 systemd[1]: Starting ensure-sysext.service... Jul 10 00:25:43.007641 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jul 10 00:25:43.015652 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Jul 10 00:25:43.017756 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jul 10 00:25:43.024041 systemd-tmpfiles[1393]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Jul 10 00:25:43.024064 systemd-tmpfiles[1393]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Jul 10 00:25:43.024219 systemd-tmpfiles[1393]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Jul 10 00:25:43.024407 systemd-tmpfiles[1393]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Jul 10 00:25:43.024922 systemd-tmpfiles[1393]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Jul 10 00:25:43.025096 systemd-tmpfiles[1393]: ACLs are not supported, ignoring. Jul 10 00:25:43.025132 systemd-tmpfiles[1393]: ACLs are not supported, ignoring. Jul 10 00:25:43.036915 systemd-tmpfiles[1393]: Detected autofs mount point /boot during canonicalization of boot. Jul 10 00:25:43.036921 systemd-tmpfiles[1393]: Skipping /boot Jul 10 00:25:43.043427 systemd[1]: Reload requested from client PID 1392 ('systemctl') (unit ensure-sysext.service)... Jul 10 00:25:43.043437 systemd[1]: Reloading... Jul 10 00:25:43.046307 systemd-tmpfiles[1393]: Detected autofs mount point /boot during canonicalization of boot. Jul 10 00:25:43.046790 systemd-tmpfiles[1393]: Skipping /boot Jul 10 00:25:43.049208 systemd-udevd[1395]: Using default interface naming scheme 'v255'. Jul 10 00:25:43.070093 ldconfig[1247]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Jul 10 00:25:43.113489 zram_generator::config[1445]: No configuration found. Jul 10 00:25:43.211170 kernel: mousedev: PS/2 mouse device common for all mice Jul 10 00:25:43.222387 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Jul 10 00:25:43.233396 kernel: ACPI: button: Power Button [PWRF] Jul 10 00:25:43.239300 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jul 10 00:25:43.248712 systemd[1]: /etc/systemd/system/coreos-metadata.service:11: Ignoring unknown escape sequences: "echo "COREOS_CUSTOM_PRIVATE_IPV4=$(ip addr show ens192 | grep "inet 10." | grep -Po "inet \K[\d.]+") Jul 10 00:25:43.305461 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Jul 10 00:25:43.305823 systemd[1]: Reloading finished in 262 ms. Jul 10 00:25:43.311308 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jul 10 00:25:43.312651 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Jul 10 00:25:43.316411 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jul 10 00:25:43.338813 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Virtual_disk OEM. Jul 10 00:25:43.342294 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jul 10 00:25:43.345501 systemd[1]: Starting audit-rules.service - Load Audit Rules... Jul 10 00:25:43.348004 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Jul 10 00:25:43.349868 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jul 10 00:25:43.355652 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jul 10 00:25:43.357971 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jul 10 00:25:43.358197 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jul 10 00:25:43.359567 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Jul 10 00:25:43.359758 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Jul 10 00:25:43.362513 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Jul 10 00:25:43.370681 kernel: piix4_smbus 0000:00:07.3: SMBus Host Controller not enabled! Jul 10 00:25:43.369423 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jul 10 00:25:43.371607 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jul 10 00:25:43.373589 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Jul 10 00:25:43.373717 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jul 10 00:25:43.375764 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jul 10 00:25:43.375907 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jul 10 00:25:43.380311 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jul 10 00:25:43.382433 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jul 10 00:25:43.383785 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jul 10 00:25:43.383968 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jul 10 00:25:43.384031 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Jul 10 00:25:43.384123 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jul 10 00:25:43.384826 systemd[1]: modprobe@loop.service: Deactivated successfully. Jul 10 00:25:43.385410 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jul 10 00:25:43.393928 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Jul 10 00:25:43.394956 systemd[1]: Finished ensure-sysext.service. Jul 10 00:25:43.395240 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jul 10 00:25:43.395353 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jul 10 00:25:43.398782 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jul 10 00:25:43.403468 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Jul 10 00:25:43.406322 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Jul 10 00:25:43.406638 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Jul 10 00:25:43.415021 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jul 10 00:25:43.415161 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jul 10 00:25:43.415365 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jul 10 00:25:43.415702 systemd[1]: modprobe@drm.service: Deactivated successfully. Jul 10 00:25:43.415811 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jul 10 00:25:43.425507 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Jul 10 00:25:43.431944 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Jul 10 00:25:43.452483 (udev-worker)[1417]: id: Truncating stdout of 'dmi_memory_id' up to 16384 byte. Jul 10 00:25:43.456180 augenrules[1574]: No rules Jul 10 00:25:43.458347 systemd[1]: audit-rules.service: Deactivated successfully. Jul 10 00:25:43.459434 systemd[1]: Finished audit-rules.service - Load Audit Rules. Jul 10 00:25:43.461600 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Jul 10 00:25:43.464747 systemd[1]: Starting systemd-update-done.service - Update is Completed... Jul 10 00:25:43.475905 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jul 10 00:25:43.491418 systemd[1]: Finished systemd-update-done.service - Update is Completed. Jul 10 00:25:43.530124 systemd[1]: Started systemd-userdbd.service - User Database Manager. Jul 10 00:25:43.565302 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jul 10 00:25:43.596191 systemd-networkd[1524]: lo: Link UP Jul 10 00:25:43.596196 systemd-networkd[1524]: lo: Gained carrier Jul 10 00:25:43.597316 systemd-networkd[1524]: Enumeration completed Jul 10 00:25:43.597364 systemd[1]: Started systemd-networkd.service - Network Configuration. Jul 10 00:25:43.601488 kernel: vmxnet3 0000:0b:00.0 ens192: intr type 3, mode 0, 3 vectors allocated Jul 10 00:25:43.601611 kernel: vmxnet3 0000:0b:00.0 ens192: NIC Link is Up 10000 Mbps Jul 10 00:25:43.598616 systemd-networkd[1524]: ens192: Configuring with /etc/systemd/network/00-vmware.network. Jul 10 00:25:43.600642 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Jul 10 00:25:43.602013 systemd-networkd[1524]: ens192: Link UP Jul 10 00:25:43.602139 systemd-networkd[1524]: ens192: Gained carrier Jul 10 00:25:43.605448 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Jul 10 00:25:43.615455 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Jul 10 00:25:43.615653 systemd[1]: Reached target time-set.target - System Time Set. Jul 10 00:25:43.617647 systemd-resolved[1528]: Positive Trust Anchors: Jul 10 00:25:43.617803 systemd-resolved[1528]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jul 10 00:25:43.617861 systemd-resolved[1528]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jul 10 00:25:43.619480 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Jul 10 00:25:43.621195 systemd-resolved[1528]: Defaulting to hostname 'linux'. Jul 10 00:25:43.622213 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jul 10 00:25:43.622448 systemd[1]: Reached target network.target - Network. Jul 10 00:25:43.622555 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jul 10 00:25:43.622681 systemd[1]: Reached target sysinit.target - System Initialization. Jul 10 00:25:43.622845 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Jul 10 00:25:43.622982 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Jul 10 00:25:43.623108 systemd[1]: Started google-oslogin-cache.timer - NSS cache refresh timer. Jul 10 00:25:43.623302 systemd[1]: Started logrotate.timer - Daily rotation of log files. Jul 10 00:25:43.623499 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Jul 10 00:25:43.623624 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Jul 10 00:25:43.623743 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Jul 10 00:25:43.623767 systemd[1]: Reached target paths.target - Path Units. Jul 10 00:25:43.623865 systemd[1]: Reached target timers.target - Timer Units. Jul 10 00:27:10.209226 systemd-resolved[1528]: Clock change detected. Flushing caches. Jul 10 00:27:10.209258 systemd-timesyncd[1540]: Contacted time server 155.248.196.28:123 (0.flatcar.pool.ntp.org). Jul 10 00:27:10.209283 systemd-timesyncd[1540]: Initial clock synchronization to Thu 2025-07-10 00:27:10.209207 UTC. Jul 10 00:27:10.216353 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Jul 10 00:27:10.217467 systemd[1]: Starting docker.socket - Docker Socket for the API... Jul 10 00:27:10.219425 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Jul 10 00:27:10.219713 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Jul 10 00:27:10.219909 systemd[1]: Reached target ssh-access.target - SSH Access Available. Jul 10 00:27:10.223342 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Jul 10 00:27:10.223711 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Jul 10 00:27:10.224296 systemd[1]: Listening on docker.socket - Docker Socket for the API. Jul 10 00:27:10.224952 systemd[1]: Reached target sockets.target - Socket Units. Jul 10 00:27:10.225113 systemd[1]: Reached target basic.target - Basic System. Jul 10 00:27:10.225301 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Jul 10 00:27:10.225361 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Jul 10 00:27:10.226088 systemd[1]: Starting containerd.service - containerd container runtime... Jul 10 00:27:10.228618 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Jul 10 00:27:10.229444 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Jul 10 00:27:10.230264 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Jul 10 00:27:10.232494 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Jul 10 00:27:10.232664 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Jul 10 00:27:10.233939 systemd[1]: Starting google-oslogin-cache.service - NSS cache refresh... Jul 10 00:27:10.234778 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Jul 10 00:27:10.244232 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Jul 10 00:27:10.245620 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Jul 10 00:27:10.249292 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Jul 10 00:27:10.252657 jq[1607]: false Jul 10 00:27:10.253825 google_oslogin_nss_cache[1609]: oslogin_cache_refresh[1609]: Refreshing passwd entry cache Jul 10 00:27:10.253408 oslogin_cache_refresh[1609]: Refreshing passwd entry cache Jul 10 00:27:10.256223 systemd[1]: Starting systemd-logind.service - User Login Management... Jul 10 00:27:10.256820 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Jul 10 00:27:10.257708 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Jul 10 00:27:10.258488 systemd[1]: Starting update-engine.service - Update Engine... Jul 10 00:27:10.259058 oslogin_cache_refresh[1609]: Failure getting users, quitting Jul 10 00:27:10.259910 google_oslogin_nss_cache[1609]: oslogin_cache_refresh[1609]: Failure getting users, quitting Jul 10 00:27:10.259910 google_oslogin_nss_cache[1609]: oslogin_cache_refresh[1609]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Jul 10 00:27:10.259910 google_oslogin_nss_cache[1609]: oslogin_cache_refresh[1609]: Refreshing group entry cache Jul 10 00:27:10.259068 oslogin_cache_refresh[1609]: Produced empty passwd cache file, removing /etc/oslogin_passwd.cache.bak. Jul 10 00:27:10.259095 oslogin_cache_refresh[1609]: Refreshing group entry cache Jul 10 00:27:10.260868 extend-filesystems[1608]: Found /dev/sda6 Jul 10 00:27:10.262965 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Jul 10 00:27:10.263859 extend-filesystems[1608]: Found /dev/sda9 Jul 10 00:27:10.267915 google_oslogin_nss_cache[1609]: oslogin_cache_refresh[1609]: Failure getting groups, quitting Jul 10 00:27:10.267915 google_oslogin_nss_cache[1609]: oslogin_cache_refresh[1609]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Jul 10 00:27:10.265648 oslogin_cache_refresh[1609]: Failure getting groups, quitting Jul 10 00:27:10.266504 systemd[1]: Starting vgauthd.service - VGAuth Service for open-vm-tools... Jul 10 00:27:10.268029 extend-filesystems[1608]: Checking size of /dev/sda9 Jul 10 00:27:10.265655 oslogin_cache_refresh[1609]: Produced empty group cache file, removing /etc/oslogin_group.cache.bak. Jul 10 00:27:10.270522 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Jul 10 00:27:10.273262 jq[1622]: true Jul 10 00:27:10.270760 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Jul 10 00:27:10.272204 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Jul 10 00:27:10.272362 systemd[1]: google-oslogin-cache.service: Deactivated successfully. Jul 10 00:27:10.272469 systemd[1]: Finished google-oslogin-cache.service - NSS cache refresh. Jul 10 00:27:10.275716 extend-filesystems[1608]: Old size kept for /dev/sda9 Jul 10 00:27:10.283375 systemd[1]: extend-filesystems.service: Deactivated successfully. Jul 10 00:27:10.287037 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Jul 10 00:27:10.287606 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Jul 10 00:27:10.287874 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Jul 10 00:27:10.288981 systemd[1]: motdgen.service: Deactivated successfully. Jul 10 00:27:10.289094 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Jul 10 00:27:10.291514 update_engine[1619]: I20250710 00:27:10.290129 1619 main.cc:92] Flatcar Update Engine starting Jul 10 00:27:10.302022 jq[1631]: true Jul 10 00:27:10.311302 tar[1629]: linux-amd64/helm Jul 10 00:27:10.311784 (ntainerd)[1640]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Jul 10 00:27:10.320633 systemd[1]: Started vgauthd.service - VGAuth Service for open-vm-tools. Jul 10 00:27:10.321839 systemd[1]: Starting vmtoolsd.service - Service for virtual machines hosted on VMware... Jul 10 00:27:10.341188 dbus-daemon[1605]: [system] SELinux support is enabled Jul 10 00:27:10.341283 systemd[1]: Started dbus.service - D-Bus System Message Bus. Jul 10 00:27:10.346398 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Jul 10 00:27:10.346415 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Jul 10 00:27:10.346562 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Jul 10 00:27:10.346570 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Jul 10 00:27:10.351946 bash[1672]: Updated "/home/core/.ssh/authorized_keys" Jul 10 00:27:10.352302 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Jul 10 00:27:10.353334 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Jul 10 00:27:10.355823 systemd[1]: Started update-engine.service - Update Engine. Jul 10 00:27:10.357392 update_engine[1619]: I20250710 00:27:10.357285 1619 update_check_scheduler.cc:74] Next update check in 3m26s Jul 10 00:27:10.370853 systemd-logind[1614]: Watching system buttons on /dev/input/event2 (Power Button) Jul 10 00:27:10.371029 systemd-logind[1614]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Jul 10 00:27:10.372207 systemd-logind[1614]: New seat seat0. Jul 10 00:27:10.374638 systemd[1]: Started locksmithd.service - Cluster reboot manager. Jul 10 00:27:10.374870 systemd[1]: Started systemd-logind.service - User Login Management. Jul 10 00:27:10.376230 systemd[1]: Started vmtoolsd.service - Service for virtual machines hosted on VMware. Jul 10 00:27:10.388383 unknown[1661]: Pref_Init: Using '/etc/vmware-tools/vgauth.conf' as preferences filepath Jul 10 00:27:10.389451 unknown[1661]: Core dump limit set to -1 Jul 10 00:27:10.517400 locksmithd[1675]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Jul 10 00:27:10.617210 containerd[1640]: time="2025-07-10T00:27:10Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Jul 10 00:27:10.618460 containerd[1640]: time="2025-07-10T00:27:10.618444754Z" level=info msg="starting containerd" revision=06b99ca80cdbfbc6cc8bd567021738c9af2b36ce version=v2.0.4 Jul 10 00:27:10.627774 containerd[1640]: time="2025-07-10T00:27:10.627746836Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="6.763µs" Jul 10 00:27:10.627774 containerd[1640]: time="2025-07-10T00:27:10.627771729Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Jul 10 00:27:10.627842 containerd[1640]: time="2025-07-10T00:27:10.627786518Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Jul 10 00:27:10.627885 containerd[1640]: time="2025-07-10T00:27:10.627874849Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Jul 10 00:27:10.627902 containerd[1640]: time="2025-07-10T00:27:10.627888016Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Jul 10 00:27:10.627915 containerd[1640]: time="2025-07-10T00:27:10.627905292Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Jul 10 00:27:10.627952 containerd[1640]: time="2025-07-10T00:27:10.627942206Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Jul 10 00:27:10.627967 containerd[1640]: time="2025-07-10T00:27:10.627954519Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Jul 10 00:27:10.628113 containerd[1640]: time="2025-07-10T00:27:10.628099465Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Jul 10 00:27:10.628137 containerd[1640]: time="2025-07-10T00:27:10.628110322Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Jul 10 00:27:10.628137 containerd[1640]: time="2025-07-10T00:27:10.628119070Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Jul 10 00:27:10.628137 containerd[1640]: time="2025-07-10T00:27:10.628126516Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Jul 10 00:27:10.628197 containerd[1640]: time="2025-07-10T00:27:10.628182929Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Jul 10 00:27:10.628370 containerd[1640]: time="2025-07-10T00:27:10.628359873Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Jul 10 00:27:10.628393 containerd[1640]: time="2025-07-10T00:27:10.628381239Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Jul 10 00:27:10.628393 containerd[1640]: time="2025-07-10T00:27:10.628389954Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Jul 10 00:27:10.628423 containerd[1640]: time="2025-07-10T00:27:10.628403410Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Jul 10 00:27:10.630826 containerd[1640]: time="2025-07-10T00:27:10.630516078Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Jul 10 00:27:10.630826 containerd[1640]: time="2025-07-10T00:27:10.630570396Z" level=info msg="metadata content store policy set" policy=shared Jul 10 00:27:10.633055 containerd[1640]: time="2025-07-10T00:27:10.633039967Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Jul 10 00:27:10.633080 containerd[1640]: time="2025-07-10T00:27:10.633067133Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Jul 10 00:27:10.633080 containerd[1640]: time="2025-07-10T00:27:10.633076486Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Jul 10 00:27:10.633122 containerd[1640]: time="2025-07-10T00:27:10.633083188Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Jul 10 00:27:10.633122 containerd[1640]: time="2025-07-10T00:27:10.633089709Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Jul 10 00:27:10.633122 containerd[1640]: time="2025-07-10T00:27:10.633095987Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Jul 10 00:27:10.633122 containerd[1640]: time="2025-07-10T00:27:10.633102390Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Jul 10 00:27:10.633122 containerd[1640]: time="2025-07-10T00:27:10.633108669Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Jul 10 00:27:10.633216 containerd[1640]: time="2025-07-10T00:27:10.633114471Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Jul 10 00:27:10.633216 containerd[1640]: time="2025-07-10T00:27:10.633141016Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Jul 10 00:27:10.633216 containerd[1640]: time="2025-07-10T00:27:10.633147346Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Jul 10 00:27:10.633216 containerd[1640]: time="2025-07-10T00:27:10.633154140Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Jul 10 00:27:10.633300 containerd[1640]: time="2025-07-10T00:27:10.633289127Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Jul 10 00:27:10.633318 containerd[1640]: time="2025-07-10T00:27:10.633305772Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Jul 10 00:27:10.633318 containerd[1640]: time="2025-07-10T00:27:10.633314606Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Jul 10 00:27:10.633344 containerd[1640]: time="2025-07-10T00:27:10.633320597Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Jul 10 00:27:10.633344 containerd[1640]: time="2025-07-10T00:27:10.633326408Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Jul 10 00:27:10.633344 containerd[1640]: time="2025-07-10T00:27:10.633334800Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Jul 10 00:27:10.633344 containerd[1640]: time="2025-07-10T00:27:10.633340795Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Jul 10 00:27:10.633399 containerd[1640]: time="2025-07-10T00:27:10.633345926Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Jul 10 00:27:10.633399 containerd[1640]: time="2025-07-10T00:27:10.633352516Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Jul 10 00:27:10.633399 containerd[1640]: time="2025-07-10T00:27:10.633360163Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Jul 10 00:27:10.633399 containerd[1640]: time="2025-07-10T00:27:10.633366422Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Jul 10 00:27:10.633450 containerd[1640]: time="2025-07-10T00:27:10.633402152Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Jul 10 00:27:10.633450 containerd[1640]: time="2025-07-10T00:27:10.633410482Z" level=info msg="Start snapshots syncer" Jul 10 00:27:10.635177 containerd[1640]: time="2025-07-10T00:27:10.633747580Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Jul 10 00:27:10.635177 containerd[1640]: time="2025-07-10T00:27:10.634108398Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Jul 10 00:27:10.635263 containerd[1640]: time="2025-07-10T00:27:10.634138042Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Jul 10 00:27:10.635263 containerd[1640]: time="2025-07-10T00:27:10.635205491Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Jul 10 00:27:10.635263 containerd[1640]: time="2025-07-10T00:27:10.635259071Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Jul 10 00:27:10.635309 containerd[1640]: time="2025-07-10T00:27:10.635272340Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Jul 10 00:27:10.635309 containerd[1640]: time="2025-07-10T00:27:10.635278503Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Jul 10 00:27:10.635309 containerd[1640]: time="2025-07-10T00:27:10.635284615Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Jul 10 00:27:10.635309 containerd[1640]: time="2025-07-10T00:27:10.635291645Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Jul 10 00:27:10.635309 containerd[1640]: time="2025-07-10T00:27:10.635297192Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Jul 10 00:27:10.635309 containerd[1640]: time="2025-07-10T00:27:10.635302553Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Jul 10 00:27:10.635385 containerd[1640]: time="2025-07-10T00:27:10.635319692Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Jul 10 00:27:10.635385 containerd[1640]: time="2025-07-10T00:27:10.635326857Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Jul 10 00:27:10.635385 containerd[1640]: time="2025-07-10T00:27:10.635332577Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Jul 10 00:27:10.636469 containerd[1640]: time="2025-07-10T00:27:10.636453169Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Jul 10 00:27:10.636495 containerd[1640]: time="2025-07-10T00:27:10.636469234Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Jul 10 00:27:10.636495 containerd[1640]: time="2025-07-10T00:27:10.636475606Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Jul 10 00:27:10.636495 containerd[1640]: time="2025-07-10T00:27:10.636481892Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Jul 10 00:27:10.636495 containerd[1640]: time="2025-07-10T00:27:10.636487106Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Jul 10 00:27:10.636495 containerd[1640]: time="2025-07-10T00:27:10.636492393Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Jul 10 00:27:10.636582 containerd[1640]: time="2025-07-10T00:27:10.636498035Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Jul 10 00:27:10.636582 containerd[1640]: time="2025-07-10T00:27:10.636535843Z" level=info msg="runtime interface created" Jul 10 00:27:10.636582 containerd[1640]: time="2025-07-10T00:27:10.636540264Z" level=info msg="created NRI interface" Jul 10 00:27:10.636582 containerd[1640]: time="2025-07-10T00:27:10.636545224Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Jul 10 00:27:10.636582 containerd[1640]: time="2025-07-10T00:27:10.636551491Z" level=info msg="Connect containerd service" Jul 10 00:27:10.636582 containerd[1640]: time="2025-07-10T00:27:10.636565456Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Jul 10 00:27:10.637892 containerd[1640]: time="2025-07-10T00:27:10.637876810Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Jul 10 00:27:10.665800 sshd_keygen[1653]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Jul 10 00:27:10.684007 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Jul 10 00:27:10.685368 systemd[1]: Starting issuegen.service - Generate /run/issue... Jul 10 00:27:10.697530 systemd[1]: issuegen.service: Deactivated successfully. Jul 10 00:27:10.697738 systemd[1]: Finished issuegen.service - Generate /run/issue. Jul 10 00:27:10.698982 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Jul 10 00:27:10.720116 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Jul 10 00:27:10.724352 systemd[1]: Started getty@tty1.service - Getty on tty1. Jul 10 00:27:10.725888 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Jul 10 00:27:10.727876 systemd[1]: Reached target getty.target - Login Prompts. Jul 10 00:27:10.764670 containerd[1640]: time="2025-07-10T00:27:10.764642934Z" level=info msg="Start subscribing containerd event" Jul 10 00:27:10.764734 containerd[1640]: time="2025-07-10T00:27:10.764673219Z" level=info msg="Start recovering state" Jul 10 00:27:10.764734 containerd[1640]: time="2025-07-10T00:27:10.764730117Z" level=info msg="Start event monitor" Jul 10 00:27:10.764773 containerd[1640]: time="2025-07-10T00:27:10.764739244Z" level=info msg="Start cni network conf syncer for default" Jul 10 00:27:10.764773 containerd[1640]: time="2025-07-10T00:27:10.764744313Z" level=info msg="Start streaming server" Jul 10 00:27:10.764773 containerd[1640]: time="2025-07-10T00:27:10.764757618Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Jul 10 00:27:10.764773 containerd[1640]: time="2025-07-10T00:27:10.764762462Z" level=info msg="runtime interface starting up..." Jul 10 00:27:10.764773 containerd[1640]: time="2025-07-10T00:27:10.764765467Z" level=info msg="starting plugins..." Jul 10 00:27:10.764773 containerd[1640]: time="2025-07-10T00:27:10.764772652Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Jul 10 00:27:10.765261 containerd[1640]: time="2025-07-10T00:27:10.765248803Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Jul 10 00:27:10.765288 containerd[1640]: time="2025-07-10T00:27:10.765276174Z" level=info msg=serving... address=/run/containerd/containerd.sock Jul 10 00:27:10.765375 systemd[1]: Started containerd.service - containerd container runtime. Jul 10 00:27:10.766073 containerd[1640]: time="2025-07-10T00:27:10.766062533Z" level=info msg="containerd successfully booted in 0.150348s" Jul 10 00:27:10.808062 tar[1629]: linux-amd64/LICENSE Jul 10 00:27:10.808131 tar[1629]: linux-amd64/README.md Jul 10 00:27:10.815282 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Jul 10 00:27:11.684312 systemd-networkd[1524]: ens192: Gained IPv6LL Jul 10 00:27:11.685795 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Jul 10 00:27:11.686232 systemd[1]: Reached target network-online.target - Network is Online. Jul 10 00:27:11.687431 systemd[1]: Starting coreos-metadata.service - VMware metadata agent... Jul 10 00:27:11.690001 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 10 00:27:11.693293 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Jul 10 00:27:11.717004 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Jul 10 00:27:11.728293 systemd[1]: coreos-metadata.service: Deactivated successfully. Jul 10 00:27:11.728587 systemd[1]: Finished coreos-metadata.service - VMware metadata agent. Jul 10 00:27:11.729551 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Jul 10 00:27:12.681795 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jul 10 00:27:12.682514 systemd[1]: Reached target multi-user.target - Multi-User System. Jul 10 00:27:12.682713 systemd[1]: Startup finished in 2.699s (kernel) + 5.089s (initrd) + 4.469s (userspace) = 12.258s. Jul 10 00:27:12.697297 (kubelet)[1804]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jul 10 00:27:12.725601 login[1763]: pam_unix(login:session): session opened for user core(uid=500) by LOGIN(uid=0) Jul 10 00:27:12.727059 login[1764]: pam_unix(login:session): session opened for user core(uid=500) by LOGIN(uid=0) Jul 10 00:27:12.733698 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Jul 10 00:27:12.734780 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Jul 10 00:27:12.737076 systemd-logind[1614]: New session 1 of user core. Jul 10 00:27:12.740214 systemd-logind[1614]: New session 2 of user core. Jul 10 00:27:12.749905 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Jul 10 00:27:12.751749 systemd[1]: Starting user@500.service - User Manager for UID 500... Jul 10 00:27:12.759422 (systemd)[1811]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Jul 10 00:27:12.760792 systemd-logind[1614]: New session c1 of user core. Jul 10 00:27:12.845184 systemd[1811]: Queued start job for default target default.target. Jul 10 00:27:12.855221 systemd[1811]: Created slice app.slice - User Application Slice. Jul 10 00:27:12.855241 systemd[1811]: Reached target paths.target - Paths. Jul 10 00:27:12.855267 systemd[1811]: Reached target timers.target - Timers. Jul 10 00:27:12.855973 systemd[1811]: Starting dbus.socket - D-Bus User Message Bus Socket... Jul 10 00:27:12.862374 systemd[1811]: Listening on dbus.socket - D-Bus User Message Bus Socket. Jul 10 00:27:12.862405 systemd[1811]: Reached target sockets.target - Sockets. Jul 10 00:27:12.862432 systemd[1811]: Reached target basic.target - Basic System. Jul 10 00:27:12.862453 systemd[1811]: Reached target default.target - Main User Target. Jul 10 00:27:12.862469 systemd[1811]: Startup finished in 98ms. Jul 10 00:27:12.862500 systemd[1]: Started user@500.service - User Manager for UID 500. Jul 10 00:27:12.863638 systemd[1]: Started session-1.scope - Session 1 of User core. Jul 10 00:27:12.864161 systemd[1]: Started session-2.scope - Session 2 of User core. Jul 10 00:27:13.232438 kubelet[1804]: E0710 00:27:13.232400 1804 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jul 10 00:27:13.233718 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jul 10 00:27:13.233804 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jul 10 00:27:13.234113 systemd[1]: kubelet.service: Consumed 663ms CPU time, 265.2M memory peak. Jul 10 00:27:23.484310 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Jul 10 00:27:23.485896 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 10 00:27:23.986966 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jul 10 00:27:23.989832 (kubelet)[1855]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jul 10 00:27:24.019511 kubelet[1855]: E0710 00:27:24.019471 1855 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jul 10 00:27:24.022143 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jul 10 00:27:24.022366 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jul 10 00:27:24.022832 systemd[1]: kubelet.service: Consumed 103ms CPU time, 108.2M memory peak. Jul 10 00:27:26.959094 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Jul 10 00:27:26.961021 systemd[1]: Started sshd@0-139.178.70.108:22-185.156.73.234:41322.service - OpenSSH per-connection server daemon (185.156.73.234:41322). Jul 10 00:27:29.157564 sshd[1863]: Connection closed by authenticating user root 185.156.73.234 port 41322 [preauth] Jul 10 00:27:29.158619 systemd[1]: sshd@0-139.178.70.108:22-185.156.73.234:41322.service: Deactivated successfully. Jul 10 00:27:34.272614 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Jul 10 00:27:34.273736 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 10 00:27:34.539176 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jul 10 00:27:34.541648 (kubelet)[1875]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jul 10 00:27:34.595103 kubelet[1875]: E0710 00:27:34.595068 1875 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jul 10 00:27:34.596533 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jul 10 00:27:34.596617 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jul 10 00:27:34.596818 systemd[1]: kubelet.service: Consumed 92ms CPU time, 108.9M memory peak. Jul 10 00:27:40.519821 systemd[1]: Started sshd@1-139.178.70.108:22-139.178.68.195:56786.service - OpenSSH per-connection server daemon (139.178.68.195:56786). Jul 10 00:27:40.563025 sshd[1883]: Accepted publickey for core from 139.178.68.195 port 56786 ssh2: RSA SHA256:4dbLs3K8zeCUdpJVvc+oLD6Wxu1uro36XJoOlJl6xXw Jul 10 00:27:40.563850 sshd-session[1883]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:27:40.566725 systemd-logind[1614]: New session 3 of user core. Jul 10 00:27:40.575314 systemd[1]: Started session-3.scope - Session 3 of User core. Jul 10 00:27:40.628341 systemd[1]: Started sshd@2-139.178.70.108:22-139.178.68.195:56788.service - OpenSSH per-connection server daemon (139.178.68.195:56788). Jul 10 00:27:40.674122 sshd[1888]: Accepted publickey for core from 139.178.68.195 port 56788 ssh2: RSA SHA256:4dbLs3K8zeCUdpJVvc+oLD6Wxu1uro36XJoOlJl6xXw Jul 10 00:27:40.675017 sshd-session[1888]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:27:40.678438 systemd-logind[1614]: New session 4 of user core. Jul 10 00:27:40.686315 systemd[1]: Started session-4.scope - Session 4 of User core. Jul 10 00:27:40.735677 sshd[1890]: Connection closed by 139.178.68.195 port 56788 Jul 10 00:27:40.735979 sshd-session[1888]: pam_unix(sshd:session): session closed for user core Jul 10 00:27:40.744500 systemd[1]: sshd@2-139.178.70.108:22-139.178.68.195:56788.service: Deactivated successfully. Jul 10 00:27:40.745590 systemd[1]: session-4.scope: Deactivated successfully. Jul 10 00:27:40.746065 systemd-logind[1614]: Session 4 logged out. Waiting for processes to exit. Jul 10 00:27:40.747994 systemd[1]: Started sshd@3-139.178.70.108:22-139.178.68.195:56800.service - OpenSSH per-connection server daemon (139.178.68.195:56800). Jul 10 00:27:40.748627 systemd-logind[1614]: Removed session 4. Jul 10 00:27:40.793884 sshd[1896]: Accepted publickey for core from 139.178.68.195 port 56800 ssh2: RSA SHA256:4dbLs3K8zeCUdpJVvc+oLD6Wxu1uro36XJoOlJl6xXw Jul 10 00:27:40.794867 sshd-session[1896]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:27:40.797861 systemd-logind[1614]: New session 5 of user core. Jul 10 00:27:40.805318 systemd[1]: Started session-5.scope - Session 5 of User core. Jul 10 00:27:40.852414 sshd[1898]: Connection closed by 139.178.68.195 port 56800 Jul 10 00:27:40.853239 sshd-session[1896]: pam_unix(sshd:session): session closed for user core Jul 10 00:27:40.865317 systemd[1]: sshd@3-139.178.70.108:22-139.178.68.195:56800.service: Deactivated successfully. Jul 10 00:27:40.866542 systemd[1]: session-5.scope: Deactivated successfully. Jul 10 00:27:40.867909 systemd-logind[1614]: Session 5 logged out. Waiting for processes to exit. Jul 10 00:27:40.868611 systemd[1]: Started sshd@4-139.178.70.108:22-139.178.68.195:56810.service - OpenSSH per-connection server daemon (139.178.68.195:56810). Jul 10 00:27:40.869540 systemd-logind[1614]: Removed session 5. Jul 10 00:27:40.908493 sshd[1904]: Accepted publickey for core from 139.178.68.195 port 56810 ssh2: RSA SHA256:4dbLs3K8zeCUdpJVvc+oLD6Wxu1uro36XJoOlJl6xXw Jul 10 00:27:40.909322 sshd-session[1904]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:27:40.912978 systemd-logind[1614]: New session 6 of user core. Jul 10 00:27:40.922324 systemd[1]: Started session-6.scope - Session 6 of User core. Jul 10 00:27:40.971284 sshd[1906]: Connection closed by 139.178.68.195 port 56810 Jul 10 00:27:40.970674 sshd-session[1904]: pam_unix(sshd:session): session closed for user core Jul 10 00:27:40.980986 systemd[1]: sshd@4-139.178.70.108:22-139.178.68.195:56810.service: Deactivated successfully. Jul 10 00:27:40.982434 systemd[1]: session-6.scope: Deactivated successfully. Jul 10 00:27:40.983015 systemd-logind[1614]: Session 6 logged out. Waiting for processes to exit. Jul 10 00:27:40.984157 systemd-logind[1614]: Removed session 6. Jul 10 00:27:40.985441 systemd[1]: Started sshd@5-139.178.70.108:22-139.178.68.195:56816.service - OpenSSH per-connection server daemon (139.178.68.195:56816). Jul 10 00:27:41.026098 sshd[1912]: Accepted publickey for core from 139.178.68.195 port 56816 ssh2: RSA SHA256:4dbLs3K8zeCUdpJVvc+oLD6Wxu1uro36XJoOlJl6xXw Jul 10 00:27:41.026959 sshd-session[1912]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:27:41.029852 systemd-logind[1614]: New session 7 of user core. Jul 10 00:27:41.038559 systemd[1]: Started session-7.scope - Session 7 of User core. Jul 10 00:27:41.097242 sudo[1915]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Jul 10 00:27:41.097411 sudo[1915]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jul 10 00:27:41.106630 sudo[1915]: pam_unix(sudo:session): session closed for user root Jul 10 00:27:41.108361 sshd[1914]: Connection closed by 139.178.68.195 port 56816 Jul 10 00:27:41.107806 sshd-session[1912]: pam_unix(sshd:session): session closed for user core Jul 10 00:27:41.115046 systemd[1]: sshd@5-139.178.70.108:22-139.178.68.195:56816.service: Deactivated successfully. Jul 10 00:27:41.116296 systemd[1]: session-7.scope: Deactivated successfully. Jul 10 00:27:41.116974 systemd-logind[1614]: Session 7 logged out. Waiting for processes to exit. Jul 10 00:27:41.118632 systemd[1]: Started sshd@6-139.178.70.108:22-139.178.68.195:56824.service - OpenSSH per-connection server daemon (139.178.68.195:56824). Jul 10 00:27:41.119382 systemd-logind[1614]: Removed session 7. Jul 10 00:27:41.162704 sshd[1921]: Accepted publickey for core from 139.178.68.195 port 56824 ssh2: RSA SHA256:4dbLs3K8zeCUdpJVvc+oLD6Wxu1uro36XJoOlJl6xXw Jul 10 00:27:41.163490 sshd-session[1921]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:27:41.166204 systemd-logind[1614]: New session 8 of user core. Jul 10 00:27:41.178297 systemd[1]: Started session-8.scope - Session 8 of User core. Jul 10 00:27:41.227456 sudo[1925]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Jul 10 00:27:41.227609 sudo[1925]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jul 10 00:27:41.230234 sudo[1925]: pam_unix(sudo:session): session closed for user root Jul 10 00:27:41.233270 sudo[1924]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Jul 10 00:27:41.233420 sudo[1924]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jul 10 00:27:41.239928 systemd[1]: Starting audit-rules.service - Load Audit Rules... Jul 10 00:27:41.268190 augenrules[1947]: No rules Jul 10 00:27:41.268842 systemd[1]: audit-rules.service: Deactivated successfully. Jul 10 00:27:41.268993 systemd[1]: Finished audit-rules.service - Load Audit Rules. Jul 10 00:27:41.269752 sudo[1924]: pam_unix(sudo:session): session closed for user root Jul 10 00:27:41.271189 sshd[1923]: Connection closed by 139.178.68.195 port 56824 Jul 10 00:27:41.270814 sshd-session[1921]: pam_unix(sshd:session): session closed for user core Jul 10 00:27:41.276636 systemd[1]: sshd@6-139.178.70.108:22-139.178.68.195:56824.service: Deactivated successfully. Jul 10 00:27:41.278423 systemd[1]: session-8.scope: Deactivated successfully. Jul 10 00:27:41.279421 systemd-logind[1614]: Session 8 logged out. Waiting for processes to exit. Jul 10 00:27:41.280737 systemd[1]: Started sshd@7-139.178.70.108:22-139.178.68.195:56826.service - OpenSSH per-connection server daemon (139.178.68.195:56826). Jul 10 00:27:41.281580 systemd-logind[1614]: Removed session 8. Jul 10 00:27:41.320083 sshd[1956]: Accepted publickey for core from 139.178.68.195 port 56826 ssh2: RSA SHA256:4dbLs3K8zeCUdpJVvc+oLD6Wxu1uro36XJoOlJl6xXw Jul 10 00:27:41.320905 sshd-session[1956]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:27:41.324208 systemd-logind[1614]: New session 9 of user core. Jul 10 00:27:41.333325 systemd[1]: Started session-9.scope - Session 9 of User core. Jul 10 00:27:41.381542 sudo[1959]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Jul 10 00:27:41.381710 sudo[1959]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jul 10 00:27:41.796020 systemd[1]: Starting docker.service - Docker Application Container Engine... Jul 10 00:27:41.808509 (dockerd)[1977]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Jul 10 00:27:42.048931 dockerd[1977]: time="2025-07-10T00:27:42.048620192Z" level=info msg="Starting up" Jul 10 00:27:42.049494 dockerd[1977]: time="2025-07-10T00:27:42.049476218Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Jul 10 00:27:42.083482 dockerd[1977]: time="2025-07-10T00:27:42.083316117Z" level=info msg="Loading containers: start." Jul 10 00:27:42.098237 kernel: Initializing XFRM netlink socket Jul 10 00:27:42.260819 systemd-networkd[1524]: docker0: Link UP Jul 10 00:27:42.262095 dockerd[1977]: time="2025-07-10T00:27:42.262074264Z" level=info msg="Loading containers: done." Jul 10 00:27:42.269911 dockerd[1977]: time="2025-07-10T00:27:42.269879819Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Jul 10 00:27:42.269996 dockerd[1977]: time="2025-07-10T00:27:42.269941765Z" level=info msg="Docker daemon" commit=bbd0a17ccc67e48d4a69393287b7fcc4f0578683 containerd-snapshotter=false storage-driver=overlay2 version=28.0.1 Jul 10 00:27:42.270019 dockerd[1977]: time="2025-07-10T00:27:42.270004633Z" level=info msg="Initializing buildkit" Jul 10 00:27:42.283778 dockerd[1977]: time="2025-07-10T00:27:42.283712066Z" level=info msg="Completed buildkit initialization" Jul 10 00:27:42.286267 dockerd[1977]: time="2025-07-10T00:27:42.286247115Z" level=info msg="Daemon has completed initialization" Jul 10 00:27:42.286405 dockerd[1977]: time="2025-07-10T00:27:42.286319507Z" level=info msg="API listen on /run/docker.sock" Jul 10 00:27:42.286420 systemd[1]: Started docker.service - Docker Application Container Engine. Jul 10 00:27:43.020181 containerd[1640]: time="2025-07-10T00:27:43.020129618Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.10\"" Jul 10 00:27:43.626602 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount77661011.mount: Deactivated successfully. Jul 10 00:27:44.509112 containerd[1640]: time="2025-07-10T00:27:44.508611297Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.31.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:27:44.509112 containerd[1640]: time="2025-07-10T00:27:44.508971500Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.31.10: active requests=0, bytes read=28077744" Jul 10 00:27:44.509112 containerd[1640]: time="2025-07-10T00:27:44.509088949Z" level=info msg="ImageCreate event name:\"sha256:74c5154ea84d9a53c406e6c00e53cf66145cce821fd80e3c74e2e1bf312f3977\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:27:44.510519 containerd[1640]: time="2025-07-10T00:27:44.510508513Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:083d7d64af31cd090f870eb49fb815e6bb42c175fc602ee9dae2f28f082bd4dc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:27:44.510992 containerd[1640]: time="2025-07-10T00:27:44.510977383Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.31.10\" with image id \"sha256:74c5154ea84d9a53c406e6c00e53cf66145cce821fd80e3c74e2e1bf312f3977\", repo tag \"registry.k8s.io/kube-apiserver:v1.31.10\", repo digest \"registry.k8s.io/kube-apiserver@sha256:083d7d64af31cd090f870eb49fb815e6bb42c175fc602ee9dae2f28f082bd4dc\", size \"28074544\" in 1.490821188s" Jul 10 00:27:44.511020 containerd[1640]: time="2025-07-10T00:27:44.510999736Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.10\" returns image reference \"sha256:74c5154ea84d9a53c406e6c00e53cf66145cce821fd80e3c74e2e1bf312f3977\"" Jul 10 00:27:44.511472 containerd[1640]: time="2025-07-10T00:27:44.511401611Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.10\"" Jul 10 00:27:44.847027 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Jul 10 00:27:44.849237 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 10 00:27:45.148831 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jul 10 00:27:45.151412 (kubelet)[2241]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jul 10 00:27:45.182499 kubelet[2241]: E0710 00:27:45.182461 2241 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jul 10 00:27:45.184117 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jul 10 00:27:45.184313 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jul 10 00:27:45.184727 systemd[1]: kubelet.service: Consumed 106ms CPU time, 108.7M memory peak. Jul 10 00:27:45.952909 containerd[1640]: time="2025-07-10T00:27:45.952869324Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.31.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:27:45.957873 containerd[1640]: time="2025-07-10T00:27:45.957845263Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.31.10: active requests=0, bytes read=24713294" Jul 10 00:27:45.964121 containerd[1640]: time="2025-07-10T00:27:45.964093366Z" level=info msg="ImageCreate event name:\"sha256:c285c4e62c91c434e9928bee7063b361509f43f43faa31641b626d6eff97616d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:27:45.971873 containerd[1640]: time="2025-07-10T00:27:45.971833592Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:3c67387d023c6114879f1e817669fd641797d30f117230682faf3930ecaaf0fe\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:27:45.972574 containerd[1640]: time="2025-07-10T00:27:45.972437080Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.31.10\" with image id \"sha256:c285c4e62c91c434e9928bee7063b361509f43f43faa31641b626d6eff97616d\", repo tag \"registry.k8s.io/kube-controller-manager:v1.31.10\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:3c67387d023c6114879f1e817669fd641797d30f117230682faf3930ecaaf0fe\", size \"26315128\" in 1.461019631s" Jul 10 00:27:45.972574 containerd[1640]: time="2025-07-10T00:27:45.972470191Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.10\" returns image reference \"sha256:c285c4e62c91c434e9928bee7063b361509f43f43faa31641b626d6eff97616d\"" Jul 10 00:27:45.972783 containerd[1640]: time="2025-07-10T00:27:45.972766770Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.10\"" Jul 10 00:27:47.108790 containerd[1640]: time="2025-07-10T00:27:47.108756669Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.31.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:27:47.109505 containerd[1640]: time="2025-07-10T00:27:47.109486835Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.31.10: active requests=0, bytes read=18783671" Jul 10 00:27:47.109860 containerd[1640]: time="2025-07-10T00:27:47.109846906Z" level=info msg="ImageCreate event name:\"sha256:61daeb7d112d9547792027cb16242b1d131f357f511545477381457fff5a69e2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:27:47.111123 containerd[1640]: time="2025-07-10T00:27:47.111093022Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:284dc2a5cf6afc9b76e39ad4b79c680c23d289488517643b28784a06d0141272\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:27:47.112187 containerd[1640]: time="2025-07-10T00:27:47.112172295Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.31.10\" with image id \"sha256:61daeb7d112d9547792027cb16242b1d131f357f511545477381457fff5a69e2\", repo tag \"registry.k8s.io/kube-scheduler:v1.31.10\", repo digest \"registry.k8s.io/kube-scheduler@sha256:284dc2a5cf6afc9b76e39ad4b79c680c23d289488517643b28784a06d0141272\", size \"20385523\" in 1.139387633s" Jul 10 00:27:47.112215 containerd[1640]: time="2025-07-10T00:27:47.112188956Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.10\" returns image reference \"sha256:61daeb7d112d9547792027cb16242b1d131f357f511545477381457fff5a69e2\"" Jul 10 00:27:47.112612 containerd[1640]: time="2025-07-10T00:27:47.112470365Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.10\"" Jul 10 00:27:47.959971 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount822496606.mount: Deactivated successfully. Jul 10 00:27:48.305852 containerd[1640]: time="2025-07-10T00:27:48.305318052Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.31.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:27:48.312518 containerd[1640]: time="2025-07-10T00:27:48.312491631Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.31.10: active requests=0, bytes read=30383943" Jul 10 00:27:48.317989 containerd[1640]: time="2025-07-10T00:27:48.317967163Z" level=info msg="ImageCreate event name:\"sha256:3ed600862d3e69931e0f9f4dbf5c2b46343af40aa079772434f13de771bdc30c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:27:48.324535 containerd[1640]: time="2025-07-10T00:27:48.324512172Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:bcbb293812bdf587b28ea98369a8c347ca84884160046296761acdf12b27029d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:27:48.325199 containerd[1640]: time="2025-07-10T00:27:48.325180540Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.31.10\" with image id \"sha256:3ed600862d3e69931e0f9f4dbf5c2b46343af40aa079772434f13de771bdc30c\", repo tag \"registry.k8s.io/kube-proxy:v1.31.10\", repo digest \"registry.k8s.io/kube-proxy@sha256:bcbb293812bdf587b28ea98369a8c347ca84884160046296761acdf12b27029d\", size \"30382962\" in 1.212676729s" Jul 10 00:27:48.325265 containerd[1640]: time="2025-07-10T00:27:48.325253298Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.10\" returns image reference \"sha256:3ed600862d3e69931e0f9f4dbf5c2b46343af40aa079772434f13de771bdc30c\"" Jul 10 00:27:48.327786 containerd[1640]: time="2025-07-10T00:27:48.327767925Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Jul 10 00:27:48.952604 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2230832402.mount: Deactivated successfully. Jul 10 00:27:49.722191 containerd[1640]: time="2025-07-10T00:27:49.721966464Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:27:49.726830 containerd[1640]: time="2025-07-10T00:27:49.726803908Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=18565241" Jul 10 00:27:49.736766 containerd[1640]: time="2025-07-10T00:27:49.736742083Z" level=info msg="ImageCreate event name:\"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:27:49.741695 containerd[1640]: time="2025-07-10T00:27:49.741664203Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:27:49.742495 containerd[1640]: time="2025-07-10T00:27:49.742279358Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"18562039\" in 1.414490624s" Jul 10 00:27:49.742495 containerd[1640]: time="2025-07-10T00:27:49.742301581Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\"" Jul 10 00:27:49.742705 containerd[1640]: time="2025-07-10T00:27:49.742686769Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Jul 10 00:27:50.224021 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4097700473.mount: Deactivated successfully. Jul 10 00:27:50.226173 containerd[1640]: time="2025-07-10T00:27:50.226146175Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jul 10 00:27:50.226751 containerd[1640]: time="2025-07-10T00:27:50.226553108Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jul 10 00:27:50.226751 containerd[1640]: time="2025-07-10T00:27:50.226714445Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Jul 10 00:27:50.229360 containerd[1640]: time="2025-07-10T00:27:50.229107856Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jul 10 00:27:50.229410 containerd[1640]: time="2025-07-10T00:27:50.229380761Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 486.659389ms" Jul 10 00:27:50.229410 containerd[1640]: time="2025-07-10T00:27:50.229394154Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Jul 10 00:27:50.229702 containerd[1640]: time="2025-07-10T00:27:50.229673783Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\"" Jul 10 00:27:50.748667 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1213127193.mount: Deactivated successfully. Jul 10 00:27:53.344045 containerd[1640]: time="2025-07-10T00:27:53.343703835Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.15-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:27:53.344980 containerd[1640]: time="2025-07-10T00:27:53.344457234Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.15-0: active requests=0, bytes read=56780013" Jul 10 00:27:53.344980 containerd[1640]: time="2025-07-10T00:27:53.344606354Z" level=info msg="ImageCreate event name:\"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:27:53.346664 containerd[1640]: time="2025-07-10T00:27:53.346635591Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:27:53.347766 containerd[1640]: time="2025-07-10T00:27:53.347478390Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.15-0\" with image id \"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\", repo tag \"registry.k8s.io/etcd:3.5.15-0\", repo digest \"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\", size \"56909194\" in 3.117364794s" Jul 10 00:27:53.347766 containerd[1640]: time="2025-07-10T00:27:53.347504271Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\" returns image reference \"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\"" Jul 10 00:27:55.429899 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 4. Jul 10 00:27:55.431108 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 10 00:27:55.437264 update_engine[1619]: I20250710 00:27:55.437238 1619 update_attempter.cc:509] Updating boot flags... Jul 10 00:27:55.438509 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Jul 10 00:27:55.438555 systemd[1]: kubelet.service: Failed with result 'signal'. Jul 10 00:27:55.438848 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jul 10 00:27:55.441468 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 10 00:27:55.461161 systemd[1]: Reload requested from client PID 2411 ('systemctl') (unit session-9.scope)... Jul 10 00:27:55.461253 systemd[1]: Reloading... Jul 10 00:27:55.537202 zram_generator::config[2467]: No configuration found. Jul 10 00:27:55.600403 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jul 10 00:27:55.613464 systemd[1]: /etc/systemd/system/coreos-metadata.service:11: Ignoring unknown escape sequences: "echo "COREOS_CUSTOM_PRIVATE_IPV4=$(ip addr show ens192 | grep "inet 10." | grep -Po "inet \K[\d.]+") Jul 10 00:27:55.688934 systemd[1]: Reloading finished in 227 ms. Jul 10 00:27:55.730718 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Jul 10 00:27:55.730764 systemd[1]: kubelet.service: Failed with result 'signal'. Jul 10 00:27:55.730929 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jul 10 00:27:55.730955 systemd[1]: kubelet.service: Consumed 40ms CPU time, 71.9M memory peak. Jul 10 00:27:55.734330 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 10 00:27:56.197962 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jul 10 00:27:56.208599 (kubelet)[2535]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jul 10 00:27:56.260124 kubelet[2535]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jul 10 00:27:56.260124 kubelet[2535]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Jul 10 00:27:56.260124 kubelet[2535]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jul 10 00:27:56.260124 kubelet[2535]: I0710 00:27:56.259759 2535 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jul 10 00:27:56.487998 kubelet[2535]: I0710 00:27:56.487782 2535 server.go:491] "Kubelet version" kubeletVersion="v1.31.8" Jul 10 00:27:56.489519 kubelet[2535]: I0710 00:27:56.489506 2535 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jul 10 00:27:56.489750 kubelet[2535]: I0710 00:27:56.489740 2535 server.go:934] "Client rotation is on, will bootstrap in background" Jul 10 00:27:56.511014 kubelet[2535]: I0710 00:27:56.511004 2535 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jul 10 00:27:56.513980 kubelet[2535]: E0710 00:27:56.513960 2535 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://139.178.70.108:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 139.178.70.108:6443: connect: connection refused" logger="UnhandledError" Jul 10 00:27:56.518911 kubelet[2535]: I0710 00:27:56.518903 2535 server.go:1431] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Jul 10 00:27:56.521824 kubelet[2535]: I0710 00:27:56.521785 2535 server.go:749] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jul 10 00:27:56.523180 kubelet[2535]: I0710 00:27:56.523154 2535 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Jul 10 00:27:56.523304 kubelet[2535]: I0710 00:27:56.523289 2535 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jul 10 00:27:56.523430 kubelet[2535]: I0710 00:27:56.523340 2535 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jul 10 00:27:56.523649 kubelet[2535]: I0710 00:27:56.523512 2535 topology_manager.go:138] "Creating topology manager with none policy" Jul 10 00:27:56.523649 kubelet[2535]: I0710 00:27:56.523521 2535 container_manager_linux.go:300] "Creating device plugin manager" Jul 10 00:27:56.523649 kubelet[2535]: I0710 00:27:56.523571 2535 state_mem.go:36] "Initialized new in-memory state store" Jul 10 00:27:56.527693 kubelet[2535]: I0710 00:27:56.527685 2535 kubelet.go:408] "Attempting to sync node with API server" Jul 10 00:27:56.527743 kubelet[2535]: I0710 00:27:56.527737 2535 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Jul 10 00:27:56.527785 kubelet[2535]: I0710 00:27:56.527780 2535 kubelet.go:314] "Adding apiserver pod source" Jul 10 00:27:56.527817 kubelet[2535]: I0710 00:27:56.527813 2535 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jul 10 00:27:56.530171 kubelet[2535]: W0710 00:27:56.529962 2535 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://139.178.70.108:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 139.178.70.108:6443: connect: connection refused Jul 10 00:27:56.530171 kubelet[2535]: E0710 00:27:56.530000 2535 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://139.178.70.108:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 139.178.70.108:6443: connect: connection refused" logger="UnhandledError" Jul 10 00:27:56.530171 kubelet[2535]: I0710 00:27:56.530067 2535 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v2.0.4" apiVersion="v1" Jul 10 00:27:56.532605 kubelet[2535]: I0710 00:27:56.532318 2535 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jul 10 00:27:56.532605 kubelet[2535]: W0710 00:27:56.532348 2535 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Jul 10 00:27:56.533365 kubelet[2535]: I0710 00:27:56.533357 2535 server.go:1274] "Started kubelet" Jul 10 00:27:56.541105 kubelet[2535]: E0710 00:27:56.538106 2535 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://139.178.70.108:6443/api/v1/namespaces/default/events\": dial tcp 139.178.70.108:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.1850bc48f225b160 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-07-10 00:27:56.533346656 +0000 UTC m=+0.322253059,LastTimestamp:2025-07-10 00:27:56.533346656 +0000 UTC m=+0.322253059,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Jul 10 00:27:56.543168 kubelet[2535]: I0710 00:27:56.542033 2535 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jul 10 00:27:56.543168 kubelet[2535]: W0710 00:27:56.542276 2535 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://139.178.70.108:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 139.178.70.108:6443: connect: connection refused Jul 10 00:27:56.543168 kubelet[2535]: E0710 00:27:56.542308 2535 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://139.178.70.108:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 139.178.70.108:6443: connect: connection refused" logger="UnhandledError" Jul 10 00:27:56.544970 kubelet[2535]: I0710 00:27:56.544019 2535 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Jul 10 00:27:56.546596 kubelet[2535]: I0710 00:27:56.546406 2535 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jul 10 00:27:56.546596 kubelet[2535]: I0710 00:27:56.546540 2535 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jul 10 00:27:56.546698 kubelet[2535]: I0710 00:27:56.546686 2535 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Jul 10 00:27:56.546877 kubelet[2535]: I0710 00:27:56.546869 2535 volume_manager.go:289] "Starting Kubelet Volume Manager" Jul 10 00:27:56.547007 kubelet[2535]: E0710 00:27:56.546998 2535 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Jul 10 00:27:56.549068 kubelet[2535]: I0710 00:27:56.549001 2535 server.go:449] "Adding debug handlers to kubelet server" Jul 10 00:27:56.550915 kubelet[2535]: E0710 00:27:56.550895 2535 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://139.178.70.108:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 139.178.70.108:6443: connect: connection refused" interval="200ms" Jul 10 00:27:56.551799 kubelet[2535]: I0710 00:27:56.551791 2535 desired_state_of_world_populator.go:147] "Desired state populator starts to run" Jul 10 00:27:56.551979 kubelet[2535]: I0710 00:27:56.551971 2535 factory.go:221] Registration of the systemd container factory successfully Jul 10 00:27:56.552051 kubelet[2535]: I0710 00:27:56.552042 2535 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jul 10 00:27:56.552179 kubelet[2535]: I0710 00:27:56.551808 2535 reconciler.go:26] "Reconciler: start to sync state" Jul 10 00:27:56.554075 kubelet[2535]: W0710 00:27:56.554057 2535 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://139.178.70.108:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 139.178.70.108:6443: connect: connection refused Jul 10 00:27:56.554131 kubelet[2535]: E0710 00:27:56.554123 2535 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://139.178.70.108:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 139.178.70.108:6443: connect: connection refused" logger="UnhandledError" Jul 10 00:27:56.554358 kubelet[2535]: I0710 00:27:56.554349 2535 factory.go:221] Registration of the containerd container factory successfully Jul 10 00:27:56.557831 kubelet[2535]: I0710 00:27:56.557814 2535 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jul 10 00:27:56.558447 kubelet[2535]: I0710 00:27:56.558439 2535 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jul 10 00:27:56.558487 kubelet[2535]: I0710 00:27:56.558483 2535 status_manager.go:217] "Starting to sync pod status with apiserver" Jul 10 00:27:56.558523 kubelet[2535]: I0710 00:27:56.558518 2535 kubelet.go:2321] "Starting kubelet main sync loop" Jul 10 00:27:56.558576 kubelet[2535]: E0710 00:27:56.558565 2535 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jul 10 00:27:56.558735 kubelet[2535]: E0710 00:27:56.558728 2535 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jul 10 00:27:56.563608 kubelet[2535]: W0710 00:27:56.563558 2535 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://139.178.70.108:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 139.178.70.108:6443: connect: connection refused Jul 10 00:27:56.563608 kubelet[2535]: E0710 00:27:56.563586 2535 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://139.178.70.108:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 139.178.70.108:6443: connect: connection refused" logger="UnhandledError" Jul 10 00:27:56.579361 kubelet[2535]: I0710 00:27:56.579341 2535 cpu_manager.go:214] "Starting CPU manager" policy="none" Jul 10 00:27:56.579361 kubelet[2535]: I0710 00:27:56.579349 2535 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Jul 10 00:27:56.579361 kubelet[2535]: I0710 00:27:56.579358 2535 state_mem.go:36] "Initialized new in-memory state store" Jul 10 00:27:56.580264 kubelet[2535]: I0710 00:27:56.580246 2535 policy_none.go:49] "None policy: Start" Jul 10 00:27:56.580562 kubelet[2535]: I0710 00:27:56.580552 2535 memory_manager.go:170] "Starting memorymanager" policy="None" Jul 10 00:27:56.580675 kubelet[2535]: I0710 00:27:56.580633 2535 state_mem.go:35] "Initializing new in-memory state store" Jul 10 00:27:56.586436 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Jul 10 00:27:56.599088 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Jul 10 00:27:56.612123 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Jul 10 00:27:56.614242 kubelet[2535]: I0710 00:27:56.614220 2535 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jul 10 00:27:56.614918 kubelet[2535]: I0710 00:27:56.614347 2535 eviction_manager.go:189] "Eviction manager: starting control loop" Jul 10 00:27:56.614918 kubelet[2535]: I0710 00:27:56.614358 2535 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jul 10 00:27:56.614918 kubelet[2535]: I0710 00:27:56.614595 2535 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jul 10 00:27:56.615965 kubelet[2535]: E0710 00:27:56.615921 2535 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Jul 10 00:27:56.668386 systemd[1]: Created slice kubepods-burstable-podb50796046d5aecf337e2bf42d9faa7d3.slice - libcontainer container kubepods-burstable-podb50796046d5aecf337e2bf42d9faa7d3.slice. Jul 10 00:27:56.680859 systemd[1]: Created slice kubepods-burstable-pod3f04709fe51ae4ab5abd58e8da771b74.slice - libcontainer container kubepods-burstable-pod3f04709fe51ae4ab5abd58e8da771b74.slice. Jul 10 00:27:56.691762 systemd[1]: Created slice kubepods-burstable-podb35b56493416c25588cb530e37ffc065.slice - libcontainer container kubepods-burstable-podb35b56493416c25588cb530e37ffc065.slice. Jul 10 00:27:56.715307 kubelet[2535]: I0710 00:27:56.715277 2535 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Jul 10 00:27:56.715525 kubelet[2535]: E0710 00:27:56.715506 2535 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://139.178.70.108:6443/api/v1/nodes\": dial tcp 139.178.70.108:6443: connect: connection refused" node="localhost" Jul 10 00:27:56.751963 kubelet[2535]: E0710 00:27:56.751896 2535 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://139.178.70.108:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 139.178.70.108:6443: connect: connection refused" interval="400ms" Jul 10 00:27:56.753320 kubelet[2535]: I0710 00:27:56.753160 2535 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/b35b56493416c25588cb530e37ffc065-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"b35b56493416c25588cb530e37ffc065\") " pod="kube-system/kube-scheduler-localhost" Jul 10 00:27:56.753320 kubelet[2535]: I0710 00:27:56.753196 2535 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/b50796046d5aecf337e2bf42d9faa7d3-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"b50796046d5aecf337e2bf42d9faa7d3\") " pod="kube-system/kube-apiserver-localhost" Jul 10 00:27:56.753320 kubelet[2535]: I0710 00:27:56.753209 2535 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/b50796046d5aecf337e2bf42d9faa7d3-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"b50796046d5aecf337e2bf42d9faa7d3\") " pod="kube-system/kube-apiserver-localhost" Jul 10 00:27:56.753320 kubelet[2535]: I0710 00:27:56.753221 2535 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b50796046d5aecf337e2bf42d9faa7d3-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"b50796046d5aecf337e2bf42d9faa7d3\") " pod="kube-system/kube-apiserver-localhost" Jul 10 00:27:56.753320 kubelet[2535]: I0710 00:27:56.753233 2535 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/3f04709fe51ae4ab5abd58e8da771b74-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"3f04709fe51ae4ab5abd58e8da771b74\") " pod="kube-system/kube-controller-manager-localhost" Jul 10 00:27:56.753448 kubelet[2535]: I0710 00:27:56.753244 2535 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/3f04709fe51ae4ab5abd58e8da771b74-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"3f04709fe51ae4ab5abd58e8da771b74\") " pod="kube-system/kube-controller-manager-localhost" Jul 10 00:27:56.753448 kubelet[2535]: I0710 00:27:56.753254 2535 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/3f04709fe51ae4ab5abd58e8da771b74-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"3f04709fe51ae4ab5abd58e8da771b74\") " pod="kube-system/kube-controller-manager-localhost" Jul 10 00:27:56.753448 kubelet[2535]: I0710 00:27:56.753264 2535 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/3f04709fe51ae4ab5abd58e8da771b74-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"3f04709fe51ae4ab5abd58e8da771b74\") " pod="kube-system/kube-controller-manager-localhost" Jul 10 00:27:56.753448 kubelet[2535]: I0710 00:27:56.753276 2535 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/3f04709fe51ae4ab5abd58e8da771b74-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"3f04709fe51ae4ab5abd58e8da771b74\") " pod="kube-system/kube-controller-manager-localhost" Jul 10 00:27:56.917006 kubelet[2535]: I0710 00:27:56.916985 2535 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Jul 10 00:27:56.917217 kubelet[2535]: E0710 00:27:56.917197 2535 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://139.178.70.108:6443/api/v1/nodes\": dial tcp 139.178.70.108:6443: connect: connection refused" node="localhost" Jul 10 00:27:56.980888 containerd[1640]: time="2025-07-10T00:27:56.980852780Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:b50796046d5aecf337e2bf42d9faa7d3,Namespace:kube-system,Attempt:0,}" Jul 10 00:27:57.003817 containerd[1640]: time="2025-07-10T00:27:57.003602510Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:3f04709fe51ae4ab5abd58e8da771b74,Namespace:kube-system,Attempt:0,}" Jul 10 00:27:57.003968 containerd[1640]: time="2025-07-10T00:27:57.003907042Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:b35b56493416c25588cb530e37ffc065,Namespace:kube-system,Attempt:0,}" Jul 10 00:27:57.152598 kubelet[2535]: E0710 00:27:57.152560 2535 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://139.178.70.108:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 139.178.70.108:6443: connect: connection refused" interval="800ms" Jul 10 00:27:57.320103 kubelet[2535]: I0710 00:27:57.319314 2535 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Jul 10 00:27:57.320103 kubelet[2535]: E0710 00:27:57.319517 2535 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://139.178.70.108:6443/api/v1/nodes\": dial tcp 139.178.70.108:6443: connect: connection refused" node="localhost" Jul 10 00:27:57.369699 kubelet[2535]: W0710 00:27:57.369663 2535 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://139.178.70.108:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 139.178.70.108:6443: connect: connection refused Jul 10 00:27:57.369815 kubelet[2535]: E0710 00:27:57.369805 2535 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://139.178.70.108:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 139.178.70.108:6443: connect: connection refused" logger="UnhandledError" Jul 10 00:27:57.374984 containerd[1640]: time="2025-07-10T00:27:57.374958045Z" level=info msg="connecting to shim fad7119aa8f71f8f9ab3f773ab56fefdf50297ad560b99cdeb21b9ff0cef04d5" address="unix:///run/containerd/s/993ff83631a2680f321124d0efa48d58372b1e752472d8655f4b5c2749b832b6" namespace=k8s.io protocol=ttrpc version=3 Jul 10 00:27:57.375596 containerd[1640]: time="2025-07-10T00:27:57.374989161Z" level=info msg="connecting to shim 0941e4b37eb892624c61e352b9064a016958c0e3ebe08cfdc54ba16c6baf3fe1" address="unix:///run/containerd/s/19c77d82ac7dc781085e26dfe5abbcd59e7e6f84e9018d7094ae6f9077dc076e" namespace=k8s.io protocol=ttrpc version=3 Jul 10 00:27:57.380142 containerd[1640]: time="2025-07-10T00:27:57.375106705Z" level=info msg="connecting to shim 747cb9e24cedf175b29190930b69089d51a34a19137413397ebba8f8d1ef6eb9" address="unix:///run/containerd/s/9b34e586d5c6bef285f02462971d32c6c6309bca653fa2b8971b83bdb0de9283" namespace=k8s.io protocol=ttrpc version=3 Jul 10 00:27:57.458738 systemd[1]: Started cri-containerd-0941e4b37eb892624c61e352b9064a016958c0e3ebe08cfdc54ba16c6baf3fe1.scope - libcontainer container 0941e4b37eb892624c61e352b9064a016958c0e3ebe08cfdc54ba16c6baf3fe1. Jul 10 00:27:57.460224 systemd[1]: Started cri-containerd-747cb9e24cedf175b29190930b69089d51a34a19137413397ebba8f8d1ef6eb9.scope - libcontainer container 747cb9e24cedf175b29190930b69089d51a34a19137413397ebba8f8d1ef6eb9. Jul 10 00:27:57.461613 systemd[1]: Started cri-containerd-fad7119aa8f71f8f9ab3f773ab56fefdf50297ad560b99cdeb21b9ff0cef04d5.scope - libcontainer container fad7119aa8f71f8f9ab3f773ab56fefdf50297ad560b99cdeb21b9ff0cef04d5. Jul 10 00:27:57.473329 kubelet[2535]: W0710 00:27:57.473250 2535 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://139.178.70.108:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 139.178.70.108:6443: connect: connection refused Jul 10 00:27:57.473329 kubelet[2535]: E0710 00:27:57.473293 2535 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://139.178.70.108:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 139.178.70.108:6443: connect: connection refused" logger="UnhandledError" Jul 10 00:27:57.512098 containerd[1640]: time="2025-07-10T00:27:57.512019671Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:3f04709fe51ae4ab5abd58e8da771b74,Namespace:kube-system,Attempt:0,} returns sandbox id \"0941e4b37eb892624c61e352b9064a016958c0e3ebe08cfdc54ba16c6baf3fe1\"" Jul 10 00:27:57.514276 containerd[1640]: time="2025-07-10T00:27:57.514231542Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:b50796046d5aecf337e2bf42d9faa7d3,Namespace:kube-system,Attempt:0,} returns sandbox id \"747cb9e24cedf175b29190930b69089d51a34a19137413397ebba8f8d1ef6eb9\"" Jul 10 00:27:57.515359 containerd[1640]: time="2025-07-10T00:27:57.515326559Z" level=info msg="CreateContainer within sandbox \"747cb9e24cedf175b29190930b69089d51a34a19137413397ebba8f8d1ef6eb9\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Jul 10 00:27:57.515866 containerd[1640]: time="2025-07-10T00:27:57.515853277Z" level=info msg="CreateContainer within sandbox \"0941e4b37eb892624c61e352b9064a016958c0e3ebe08cfdc54ba16c6baf3fe1\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Jul 10 00:27:57.524442 containerd[1640]: time="2025-07-10T00:27:57.524144532Z" level=info msg="Container 80fe289bbee02ec00bdc7488d4ba0d22aeddfe219d299d77fbc334ff9db84d4c: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:27:57.525654 containerd[1640]: time="2025-07-10T00:27:57.525497586Z" level=info msg="Container 2ba8c31e5342f5d7a222fff9d55b49e0f976512b8070e12d1cd67e31167d9317: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:27:57.527282 containerd[1640]: time="2025-07-10T00:27:57.527254227Z" level=info msg="CreateContainer within sandbox \"0941e4b37eb892624c61e352b9064a016958c0e3ebe08cfdc54ba16c6baf3fe1\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"80fe289bbee02ec00bdc7488d4ba0d22aeddfe219d299d77fbc334ff9db84d4c\"" Jul 10 00:27:57.528357 containerd[1640]: time="2025-07-10T00:27:57.528347709Z" level=info msg="StartContainer for \"80fe289bbee02ec00bdc7488d4ba0d22aeddfe219d299d77fbc334ff9db84d4c\"" Jul 10 00:27:57.529219 containerd[1640]: time="2025-07-10T00:27:57.529188691Z" level=info msg="connecting to shim 80fe289bbee02ec00bdc7488d4ba0d22aeddfe219d299d77fbc334ff9db84d4c" address="unix:///run/containerd/s/19c77d82ac7dc781085e26dfe5abbcd59e7e6f84e9018d7094ae6f9077dc076e" protocol=ttrpc version=3 Jul 10 00:27:57.530260 containerd[1640]: time="2025-07-10T00:27:57.530207240Z" level=info msg="CreateContainer within sandbox \"747cb9e24cedf175b29190930b69089d51a34a19137413397ebba8f8d1ef6eb9\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"2ba8c31e5342f5d7a222fff9d55b49e0f976512b8070e12d1cd67e31167d9317\"" Jul 10 00:27:57.530781 containerd[1640]: time="2025-07-10T00:27:57.530729966Z" level=info msg="StartContainer for \"2ba8c31e5342f5d7a222fff9d55b49e0f976512b8070e12d1cd67e31167d9317\"" Jul 10 00:27:57.531770 containerd[1640]: time="2025-07-10T00:27:57.531741288Z" level=info msg="connecting to shim 2ba8c31e5342f5d7a222fff9d55b49e0f976512b8070e12d1cd67e31167d9317" address="unix:///run/containerd/s/9b34e586d5c6bef285f02462971d32c6c6309bca653fa2b8971b83bdb0de9283" protocol=ttrpc version=3 Jul 10 00:27:57.532331 containerd[1640]: time="2025-07-10T00:27:57.532296423Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:b35b56493416c25588cb530e37ffc065,Namespace:kube-system,Attempt:0,} returns sandbox id \"fad7119aa8f71f8f9ab3f773ab56fefdf50297ad560b99cdeb21b9ff0cef04d5\"" Jul 10 00:27:57.533701 containerd[1640]: time="2025-07-10T00:27:57.533591772Z" level=info msg="CreateContainer within sandbox \"fad7119aa8f71f8f9ab3f773ab56fefdf50297ad560b99cdeb21b9ff0cef04d5\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Jul 10 00:27:57.536997 containerd[1640]: time="2025-07-10T00:27:57.536984999Z" level=info msg="Container 157bd5039f7b3e4e3e56100b06c7bd3fb905bf67539d4e504bdde866fca77db8: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:27:57.540716 containerd[1640]: time="2025-07-10T00:27:57.540614964Z" level=info msg="CreateContainer within sandbox \"fad7119aa8f71f8f9ab3f773ab56fefdf50297ad560b99cdeb21b9ff0cef04d5\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"157bd5039f7b3e4e3e56100b06c7bd3fb905bf67539d4e504bdde866fca77db8\"" Jul 10 00:27:57.541178 containerd[1640]: time="2025-07-10T00:27:57.540978459Z" level=info msg="StartContainer for \"157bd5039f7b3e4e3e56100b06c7bd3fb905bf67539d4e504bdde866fca77db8\"" Jul 10 00:27:57.541879 containerd[1640]: time="2025-07-10T00:27:57.541864821Z" level=info msg="connecting to shim 157bd5039f7b3e4e3e56100b06c7bd3fb905bf67539d4e504bdde866fca77db8" address="unix:///run/containerd/s/993ff83631a2680f321124d0efa48d58372b1e752472d8655f4b5c2749b832b6" protocol=ttrpc version=3 Jul 10 00:27:57.549364 systemd[1]: Started cri-containerd-80fe289bbee02ec00bdc7488d4ba0d22aeddfe219d299d77fbc334ff9db84d4c.scope - libcontainer container 80fe289bbee02ec00bdc7488d4ba0d22aeddfe219d299d77fbc334ff9db84d4c. Jul 10 00:27:57.556257 systemd[1]: Started cri-containerd-2ba8c31e5342f5d7a222fff9d55b49e0f976512b8070e12d1cd67e31167d9317.scope - libcontainer container 2ba8c31e5342f5d7a222fff9d55b49e0f976512b8070e12d1cd67e31167d9317. Jul 10 00:27:57.558942 systemd[1]: Started cri-containerd-157bd5039f7b3e4e3e56100b06c7bd3fb905bf67539d4e504bdde866fca77db8.scope - libcontainer container 157bd5039f7b3e4e3e56100b06c7bd3fb905bf67539d4e504bdde866fca77db8. Jul 10 00:27:57.603877 containerd[1640]: time="2025-07-10T00:27:57.603434850Z" level=info msg="StartContainer for \"80fe289bbee02ec00bdc7488d4ba0d22aeddfe219d299d77fbc334ff9db84d4c\" returns successfully" Jul 10 00:27:57.613036 containerd[1640]: time="2025-07-10T00:27:57.612970245Z" level=info msg="StartContainer for \"2ba8c31e5342f5d7a222fff9d55b49e0f976512b8070e12d1cd67e31167d9317\" returns successfully" Jul 10 00:27:57.618385 containerd[1640]: time="2025-07-10T00:27:57.618339482Z" level=info msg="StartContainer for \"157bd5039f7b3e4e3e56100b06c7bd3fb905bf67539d4e504bdde866fca77db8\" returns successfully" Jul 10 00:27:57.883190 kubelet[2535]: W0710 00:27:57.882936 2535 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://139.178.70.108:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 139.178.70.108:6443: connect: connection refused Jul 10 00:27:57.883190 kubelet[2535]: E0710 00:27:57.882979 2535 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://139.178.70.108:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 139.178.70.108:6443: connect: connection refused" logger="UnhandledError" Jul 10 00:27:57.953482 kubelet[2535]: E0710 00:27:57.953453 2535 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://139.178.70.108:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 139.178.70.108:6443: connect: connection refused" interval="1.6s" Jul 10 00:27:57.992958 kubelet[2535]: W0710 00:27:57.992921 2535 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://139.178.70.108:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 139.178.70.108:6443: connect: connection refused Jul 10 00:27:57.992958 kubelet[2535]: E0710 00:27:57.992961 2535 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://139.178.70.108:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 139.178.70.108:6443: connect: connection refused" logger="UnhandledError" Jul 10 00:27:58.120564 kubelet[2535]: I0710 00:27:58.120546 2535 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Jul 10 00:27:58.920159 kubelet[2535]: I0710 00:27:58.920075 2535 kubelet_node_status.go:75] "Successfully registered node" node="localhost" Jul 10 00:27:58.920159 kubelet[2535]: E0710 00:27:58.920099 2535 kubelet_node_status.go:535] "Error updating node status, will retry" err="error getting node \"localhost\": node \"localhost\" not found" Jul 10 00:27:58.924954 kubelet[2535]: E0710 00:27:58.924903 2535 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Jul 10 00:27:59.025257 kubelet[2535]: E0710 00:27:59.025233 2535 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Jul 10 00:27:59.126289 kubelet[2535]: E0710 00:27:59.126255 2535 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Jul 10 00:27:59.226476 kubelet[2535]: E0710 00:27:59.226324 2535 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Jul 10 00:27:59.327101 kubelet[2535]: E0710 00:27:59.327064 2535 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Jul 10 00:27:59.427742 kubelet[2535]: E0710 00:27:59.427707 2535 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Jul 10 00:27:59.528790 kubelet[2535]: E0710 00:27:59.528716 2535 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Jul 10 00:27:59.629799 kubelet[2535]: E0710 00:27:59.629776 2535 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Jul 10 00:27:59.730583 kubelet[2535]: E0710 00:27:59.730552 2535 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Jul 10 00:27:59.831529 kubelet[2535]: E0710 00:27:59.831352 2535 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Jul 10 00:27:59.931853 kubelet[2535]: E0710 00:27:59.931829 2535 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Jul 10 00:28:00.032736 kubelet[2535]: E0710 00:28:00.032704 2535 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Jul 10 00:28:00.133850 kubelet[2535]: E0710 00:28:00.133785 2535 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Jul 10 00:28:00.234376 kubelet[2535]: E0710 00:28:00.234329 2535 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Jul 10 00:28:00.335402 kubelet[2535]: E0710 00:28:00.335376 2535 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Jul 10 00:28:00.435735 kubelet[2535]: E0710 00:28:00.435657 2535 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Jul 10 00:28:00.535631 kubelet[2535]: I0710 00:28:00.535569 2535 apiserver.go:52] "Watching apiserver" Jul 10 00:28:00.552074 kubelet[2535]: I0710 00:28:00.552039 2535 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" Jul 10 00:28:00.716561 systemd[1]: Reload requested from client PID 2804 ('systemctl') (unit session-9.scope)... Jul 10 00:28:00.716572 systemd[1]: Reloading... Jul 10 00:28:00.776181 zram_generator::config[2848]: No configuration found. Jul 10 00:28:00.845685 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jul 10 00:28:00.853761 systemd[1]: /etc/systemd/system/coreos-metadata.service:11: Ignoring unknown escape sequences: "echo "COREOS_CUSTOM_PRIVATE_IPV4=$(ip addr show ens192 | grep "inet 10." | grep -Po "inet \K[\d.]+") Jul 10 00:28:00.926364 systemd[1]: Reloading finished in 209 ms. Jul 10 00:28:00.956782 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Jul 10 00:28:00.968523 systemd[1]: kubelet.service: Deactivated successfully. Jul 10 00:28:00.968779 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jul 10 00:28:00.968867 systemd[1]: kubelet.service: Consumed 483ms CPU time, 127.2M memory peak. Jul 10 00:28:00.972493 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 10 00:28:01.232071 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jul 10 00:28:01.241479 (kubelet)[2915]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jul 10 00:28:01.341182 kubelet[2915]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jul 10 00:28:01.341182 kubelet[2915]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Jul 10 00:28:01.341182 kubelet[2915]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jul 10 00:28:01.341182 kubelet[2915]: I0710 00:28:01.340918 2915 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jul 10 00:28:01.358305 kubelet[2915]: I0710 00:28:01.358294 2915 server.go:491] "Kubelet version" kubeletVersion="v1.31.8" Jul 10 00:28:01.358360 kubelet[2915]: I0710 00:28:01.358354 2915 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jul 10 00:28:01.358517 kubelet[2915]: I0710 00:28:01.358509 2915 server.go:934] "Client rotation is on, will bootstrap in background" Jul 10 00:28:01.359267 kubelet[2915]: I0710 00:28:01.359258 2915 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Jul 10 00:28:01.360370 kubelet[2915]: I0710 00:28:01.360361 2915 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jul 10 00:28:01.362837 kubelet[2915]: I0710 00:28:01.362824 2915 server.go:1431] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Jul 10 00:28:01.364896 kubelet[2915]: I0710 00:28:01.364865 2915 server.go:749] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jul 10 00:28:01.364937 kubelet[2915]: I0710 00:28:01.364932 2915 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Jul 10 00:28:01.365179 kubelet[2915]: I0710 00:28:01.364987 2915 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jul 10 00:28:01.365245 kubelet[2915]: I0710 00:28:01.365028 2915 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jul 10 00:28:01.365319 kubelet[2915]: I0710 00:28:01.365313 2915 topology_manager.go:138] "Creating topology manager with none policy" Jul 10 00:28:01.365354 kubelet[2915]: I0710 00:28:01.365350 2915 container_manager_linux.go:300] "Creating device plugin manager" Jul 10 00:28:01.365396 kubelet[2915]: I0710 00:28:01.365391 2915 state_mem.go:36] "Initialized new in-memory state store" Jul 10 00:28:01.365484 kubelet[2915]: I0710 00:28:01.365478 2915 kubelet.go:408] "Attempting to sync node with API server" Jul 10 00:28:01.365519 kubelet[2915]: I0710 00:28:01.365514 2915 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Jul 10 00:28:01.365560 kubelet[2915]: I0710 00:28:01.365556 2915 kubelet.go:314] "Adding apiserver pod source" Jul 10 00:28:01.365595 kubelet[2915]: I0710 00:28:01.365590 2915 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jul 10 00:28:01.366761 kubelet[2915]: I0710 00:28:01.366738 2915 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v2.0.4" apiVersion="v1" Jul 10 00:28:01.367024 kubelet[2915]: I0710 00:28:01.367013 2915 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jul 10 00:28:01.367718 kubelet[2915]: I0710 00:28:01.367708 2915 server.go:1274] "Started kubelet" Jul 10 00:28:01.375023 kubelet[2915]: I0710 00:28:01.375009 2915 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jul 10 00:28:01.387538 kubelet[2915]: I0710 00:28:01.387399 2915 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Jul 10 00:28:01.388081 kubelet[2915]: I0710 00:28:01.388069 2915 server.go:449] "Adding debug handlers to kubelet server" Jul 10 00:28:01.388691 kubelet[2915]: I0710 00:28:01.388671 2915 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jul 10 00:28:01.388786 kubelet[2915]: I0710 00:28:01.388776 2915 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jul 10 00:28:01.388940 kubelet[2915]: I0710 00:28:01.388926 2915 volume_manager.go:289] "Starting Kubelet Volume Manager" Jul 10 00:28:01.389207 kubelet[2915]: I0710 00:28:01.389195 2915 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Jul 10 00:28:01.389688 kubelet[2915]: E0710 00:28:01.389680 2915 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jul 10 00:28:01.390740 kubelet[2915]: I0710 00:28:01.390115 2915 factory.go:221] Registration of the systemd container factory successfully Jul 10 00:28:01.390740 kubelet[2915]: I0710 00:28:01.390188 2915 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jul 10 00:28:01.391343 kubelet[2915]: I0710 00:28:01.391194 2915 factory.go:221] Registration of the containerd container factory successfully Jul 10 00:28:01.391943 kubelet[2915]: I0710 00:28:01.391749 2915 desired_state_of_world_populator.go:147] "Desired state populator starts to run" Jul 10 00:28:01.391943 kubelet[2915]: I0710 00:28:01.391817 2915 reconciler.go:26] "Reconciler: start to sync state" Jul 10 00:28:01.393210 kubelet[2915]: I0710 00:28:01.393191 2915 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jul 10 00:28:01.394100 kubelet[2915]: I0710 00:28:01.393778 2915 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jul 10 00:28:01.394100 kubelet[2915]: I0710 00:28:01.393791 2915 status_manager.go:217] "Starting to sync pod status with apiserver" Jul 10 00:28:01.394100 kubelet[2915]: I0710 00:28:01.393801 2915 kubelet.go:2321] "Starting kubelet main sync loop" Jul 10 00:28:01.394100 kubelet[2915]: E0710 00:28:01.393822 2915 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jul 10 00:28:01.427086 kubelet[2915]: I0710 00:28:01.427067 2915 cpu_manager.go:214] "Starting CPU manager" policy="none" Jul 10 00:28:01.427185 kubelet[2915]: I0710 00:28:01.427178 2915 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Jul 10 00:28:01.427268 kubelet[2915]: I0710 00:28:01.427263 2915 state_mem.go:36] "Initialized new in-memory state store" Jul 10 00:28:01.427386 kubelet[2915]: I0710 00:28:01.427379 2915 state_mem.go:88] "Updated default CPUSet" cpuSet="" Jul 10 00:28:01.427433 kubelet[2915]: I0710 00:28:01.427416 2915 state_mem.go:96] "Updated CPUSet assignments" assignments={} Jul 10 00:28:01.427512 kubelet[2915]: I0710 00:28:01.427507 2915 policy_none.go:49] "None policy: Start" Jul 10 00:28:01.427855 kubelet[2915]: I0710 00:28:01.427848 2915 memory_manager.go:170] "Starting memorymanager" policy="None" Jul 10 00:28:01.427930 kubelet[2915]: I0710 00:28:01.427925 2915 state_mem.go:35] "Initializing new in-memory state store" Jul 10 00:28:01.428182 kubelet[2915]: I0710 00:28:01.428055 2915 state_mem.go:75] "Updated machine memory state" Jul 10 00:28:01.430770 kubelet[2915]: I0710 00:28:01.430697 2915 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jul 10 00:28:01.430894 kubelet[2915]: I0710 00:28:01.430887 2915 eviction_manager.go:189] "Eviction manager: starting control loop" Jul 10 00:28:01.430975 kubelet[2915]: I0710 00:28:01.430961 2915 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jul 10 00:28:01.431665 kubelet[2915]: I0710 00:28:01.431658 2915 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jul 10 00:28:01.498461 kubelet[2915]: E0710 00:28:01.498396 2915 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Jul 10 00:28:01.535517 kubelet[2915]: I0710 00:28:01.535498 2915 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Jul 10 00:28:01.539845 kubelet[2915]: I0710 00:28:01.539821 2915 kubelet_node_status.go:111] "Node was previously registered" node="localhost" Jul 10 00:28:01.540186 kubelet[2915]: I0710 00:28:01.539994 2915 kubelet_node_status.go:75] "Successfully registered node" node="localhost" Jul 10 00:28:01.692728 kubelet[2915]: I0710 00:28:01.692698 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/3f04709fe51ae4ab5abd58e8da771b74-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"3f04709fe51ae4ab5abd58e8da771b74\") " pod="kube-system/kube-controller-manager-localhost" Jul 10 00:28:01.692728 kubelet[2915]: I0710 00:28:01.692724 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/3f04709fe51ae4ab5abd58e8da771b74-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"3f04709fe51ae4ab5abd58e8da771b74\") " pod="kube-system/kube-controller-manager-localhost" Jul 10 00:28:01.692850 kubelet[2915]: I0710 00:28:01.692739 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/b35b56493416c25588cb530e37ffc065-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"b35b56493416c25588cb530e37ffc065\") " pod="kube-system/kube-scheduler-localhost" Jul 10 00:28:01.692850 kubelet[2915]: I0710 00:28:01.692750 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/3f04709fe51ae4ab5abd58e8da771b74-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"3f04709fe51ae4ab5abd58e8da771b74\") " pod="kube-system/kube-controller-manager-localhost" Jul 10 00:28:01.692850 kubelet[2915]: I0710 00:28:01.692762 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/b50796046d5aecf337e2bf42d9faa7d3-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"b50796046d5aecf337e2bf42d9faa7d3\") " pod="kube-system/kube-apiserver-localhost" Jul 10 00:28:01.692850 kubelet[2915]: I0710 00:28:01.692773 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b50796046d5aecf337e2bf42d9faa7d3-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"b50796046d5aecf337e2bf42d9faa7d3\") " pod="kube-system/kube-apiserver-localhost" Jul 10 00:28:01.692850 kubelet[2915]: I0710 00:28:01.692790 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/3f04709fe51ae4ab5abd58e8da771b74-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"3f04709fe51ae4ab5abd58e8da771b74\") " pod="kube-system/kube-controller-manager-localhost" Jul 10 00:28:01.692998 kubelet[2915]: I0710 00:28:01.692801 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/3f04709fe51ae4ab5abd58e8da771b74-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"3f04709fe51ae4ab5abd58e8da771b74\") " pod="kube-system/kube-controller-manager-localhost" Jul 10 00:28:01.692998 kubelet[2915]: I0710 00:28:01.692812 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/b50796046d5aecf337e2bf42d9faa7d3-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"b50796046d5aecf337e2bf42d9faa7d3\") " pod="kube-system/kube-apiserver-localhost" Jul 10 00:28:02.368291 kubelet[2915]: I0710 00:28:02.368265 2915 apiserver.go:52] "Watching apiserver" Jul 10 00:28:02.392248 kubelet[2915]: I0710 00:28:02.392226 2915 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" Jul 10 00:28:02.435270 kubelet[2915]: I0710 00:28:02.435237 2915 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=2.435218182 podStartE2EDuration="2.435218182s" podCreationTimestamp="2025-07-10 00:28:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-10 00:28:02.434752997 +0000 UTC m=+1.179533219" watchObservedRunningTime="2025-07-10 00:28:02.435218182 +0000 UTC m=+1.179998398" Jul 10 00:28:02.435407 kubelet[2915]: I0710 00:28:02.435301 2915 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.435295945 podStartE2EDuration="1.435295945s" podCreationTimestamp="2025-07-10 00:28:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-10 00:28:02.431170735 +0000 UTC m=+1.175950971" watchObservedRunningTime="2025-07-10 00:28:02.435295945 +0000 UTC m=+1.180076182" Jul 10 00:28:02.442487 kubelet[2915]: I0710 00:28:02.442432 2915 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.442421221 podStartE2EDuration="1.442421221s" podCreationTimestamp="2025-07-10 00:28:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-10 00:28:02.438693054 +0000 UTC m=+1.183473276" watchObservedRunningTime="2025-07-10 00:28:02.442421221 +0000 UTC m=+1.187201442" Jul 10 00:28:06.641872 kubelet[2915]: I0710 00:28:06.641851 2915 kuberuntime_manager.go:1635] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Jul 10 00:28:06.643210 containerd[1640]: time="2025-07-10T00:28:06.642749290Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Jul 10 00:28:06.643350 kubelet[2915]: I0710 00:28:06.642851 2915 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Jul 10 00:28:07.658582 systemd[1]: Created slice kubepods-besteffort-pod4efc45ba_61d2_4850_aeb3_da0039199776.slice - libcontainer container kubepods-besteffort-pod4efc45ba_61d2_4850_aeb3_da0039199776.slice. Jul 10 00:28:07.731511 kubelet[2915]: I0710 00:28:07.731407 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/4efc45ba-61d2-4850-aeb3-da0039199776-kube-proxy\") pod \"kube-proxy-h845m\" (UID: \"4efc45ba-61d2-4850-aeb3-da0039199776\") " pod="kube-system/kube-proxy-h845m" Jul 10 00:28:07.731511 kubelet[2915]: I0710 00:28:07.731440 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/4efc45ba-61d2-4850-aeb3-da0039199776-xtables-lock\") pod \"kube-proxy-h845m\" (UID: \"4efc45ba-61d2-4850-aeb3-da0039199776\") " pod="kube-system/kube-proxy-h845m" Jul 10 00:28:07.731511 kubelet[2915]: I0710 00:28:07.731453 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4efc45ba-61d2-4850-aeb3-da0039199776-lib-modules\") pod \"kube-proxy-h845m\" (UID: \"4efc45ba-61d2-4850-aeb3-da0039199776\") " pod="kube-system/kube-proxy-h845m" Jul 10 00:28:07.731511 kubelet[2915]: I0710 00:28:07.731467 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xw9np\" (UniqueName: \"kubernetes.io/projected/4efc45ba-61d2-4850-aeb3-da0039199776-kube-api-access-xw9np\") pod \"kube-proxy-h845m\" (UID: \"4efc45ba-61d2-4850-aeb3-da0039199776\") " pod="kube-system/kube-proxy-h845m" Jul 10 00:28:07.819790 systemd[1]: Created slice kubepods-besteffort-pode7a3d508_9c3a_4eb2_8b3c_53aa67c4c3fc.slice - libcontainer container kubepods-besteffort-pode7a3d508_9c3a_4eb2_8b3c_53aa67c4c3fc.slice. Jul 10 00:28:07.832160 kubelet[2915]: I0710 00:28:07.832126 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/e7a3d508-9c3a-4eb2-8b3c-53aa67c4c3fc-var-lib-calico\") pod \"tigera-operator-5bf8dfcb4-qbslx\" (UID: \"e7a3d508-9c3a-4eb2-8b3c-53aa67c4c3fc\") " pod="tigera-operator/tigera-operator-5bf8dfcb4-qbslx" Jul 10 00:28:07.832420 kubelet[2915]: I0710 00:28:07.832393 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xsn2\" (UniqueName: \"kubernetes.io/projected/e7a3d508-9c3a-4eb2-8b3c-53aa67c4c3fc-kube-api-access-6xsn2\") pod \"tigera-operator-5bf8dfcb4-qbslx\" (UID: \"e7a3d508-9c3a-4eb2-8b3c-53aa67c4c3fc\") " pod="tigera-operator/tigera-operator-5bf8dfcb4-qbslx" Jul 10 00:28:07.967727 containerd[1640]: time="2025-07-10T00:28:07.966996879Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-h845m,Uid:4efc45ba-61d2-4850-aeb3-da0039199776,Namespace:kube-system,Attempt:0,}" Jul 10 00:28:07.979548 containerd[1640]: time="2025-07-10T00:28:07.979518557Z" level=info msg="connecting to shim 36237808221e22244b24df476fb1e96b0dcb0c5aa5692b938d292beb7560ac56" address="unix:///run/containerd/s/b06380b4b70e8f66c55a1eacb4cf7e519d41b79b303e5132b2cd1b9dce2db69f" namespace=k8s.io protocol=ttrpc version=3 Jul 10 00:28:08.003255 systemd[1]: Started cri-containerd-36237808221e22244b24df476fb1e96b0dcb0c5aa5692b938d292beb7560ac56.scope - libcontainer container 36237808221e22244b24df476fb1e96b0dcb0c5aa5692b938d292beb7560ac56. Jul 10 00:28:08.016882 containerd[1640]: time="2025-07-10T00:28:08.016863000Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-h845m,Uid:4efc45ba-61d2-4850-aeb3-da0039199776,Namespace:kube-system,Attempt:0,} returns sandbox id \"36237808221e22244b24df476fb1e96b0dcb0c5aa5692b938d292beb7560ac56\"" Jul 10 00:28:08.018769 containerd[1640]: time="2025-07-10T00:28:08.018671071Z" level=info msg="CreateContainer within sandbox \"36237808221e22244b24df476fb1e96b0dcb0c5aa5692b938d292beb7560ac56\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Jul 10 00:28:08.024794 containerd[1640]: time="2025-07-10T00:28:08.024562198Z" level=info msg="Container 8bc75db8c7814149e3a7903697ccb7347af6d88713ae75d1eb429c7adf6fb606: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:28:08.028336 containerd[1640]: time="2025-07-10T00:28:08.028319673Z" level=info msg="CreateContainer within sandbox \"36237808221e22244b24df476fb1e96b0dcb0c5aa5692b938d292beb7560ac56\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"8bc75db8c7814149e3a7903697ccb7347af6d88713ae75d1eb429c7adf6fb606\"" Jul 10 00:28:08.028863 containerd[1640]: time="2025-07-10T00:28:08.028826588Z" level=info msg="StartContainer for \"8bc75db8c7814149e3a7903697ccb7347af6d88713ae75d1eb429c7adf6fb606\"" Jul 10 00:28:08.029770 containerd[1640]: time="2025-07-10T00:28:08.029734524Z" level=info msg="connecting to shim 8bc75db8c7814149e3a7903697ccb7347af6d88713ae75d1eb429c7adf6fb606" address="unix:///run/containerd/s/b06380b4b70e8f66c55a1eacb4cf7e519d41b79b303e5132b2cd1b9dce2db69f" protocol=ttrpc version=3 Jul 10 00:28:08.048285 systemd[1]: Started cri-containerd-8bc75db8c7814149e3a7903697ccb7347af6d88713ae75d1eb429c7adf6fb606.scope - libcontainer container 8bc75db8c7814149e3a7903697ccb7347af6d88713ae75d1eb429c7adf6fb606. Jul 10 00:28:08.070196 containerd[1640]: time="2025-07-10T00:28:08.070174158Z" level=info msg="StartContainer for \"8bc75db8c7814149e3a7903697ccb7347af6d88713ae75d1eb429c7adf6fb606\" returns successfully" Jul 10 00:28:08.121850 containerd[1640]: time="2025-07-10T00:28:08.121824385Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-5bf8dfcb4-qbslx,Uid:e7a3d508-9c3a-4eb2-8b3c-53aa67c4c3fc,Namespace:tigera-operator,Attempt:0,}" Jul 10 00:28:08.131896 containerd[1640]: time="2025-07-10T00:28:08.131819677Z" level=info msg="connecting to shim ec7c5d2911897dd45c3b93893331774ea4892b6c0cc1c8bc650b20f551561f0c" address="unix:///run/containerd/s/0d1372dc2c5c19554bb7a77830f6dd3c38c20cafa220353bd1b98c57721103bb" namespace=k8s.io protocol=ttrpc version=3 Jul 10 00:28:08.149278 systemd[1]: Started cri-containerd-ec7c5d2911897dd45c3b93893331774ea4892b6c0cc1c8bc650b20f551561f0c.scope - libcontainer container ec7c5d2911897dd45c3b93893331774ea4892b6c0cc1c8bc650b20f551561f0c. Jul 10 00:28:08.179308 containerd[1640]: time="2025-07-10T00:28:08.179282698Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-5bf8dfcb4-qbslx,Uid:e7a3d508-9c3a-4eb2-8b3c-53aa67c4c3fc,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"ec7c5d2911897dd45c3b93893331774ea4892b6c0cc1c8bc650b20f551561f0c\"" Jul 10 00:28:08.180796 containerd[1640]: time="2025-07-10T00:28:08.180766763Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.3\"" Jul 10 00:28:08.437297 kubelet[2915]: I0710 00:28:08.436957 2915 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-h845m" podStartSLOduration=1.4369452360000001 podStartE2EDuration="1.436945236s" podCreationTimestamp="2025-07-10 00:28:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-10 00:28:08.436819089 +0000 UTC m=+7.181599313" watchObservedRunningTime="2025-07-10 00:28:08.436945236 +0000 UTC m=+7.181725454" Jul 10 00:28:08.843355 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3753256291.mount: Deactivated successfully. Jul 10 00:28:09.567656 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount85081215.mount: Deactivated successfully. Jul 10 00:28:10.068470 containerd[1640]: time="2025-07-10T00:28:10.068336954Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.38.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:10.068470 containerd[1640]: time="2025-07-10T00:28:10.068442534Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.38.3: active requests=0, bytes read=25056543" Jul 10 00:28:10.068911 containerd[1640]: time="2025-07-10T00:28:10.068888661Z" level=info msg="ImageCreate event name:\"sha256:8bde16470b09d1963e19456806d73180c9778a6c2b3c1fda2335c67c1cd4ce93\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:10.069971 containerd[1640]: time="2025-07-10T00:28:10.069949190Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:dbf1bad0def7b5955dc8e4aeee96e23ead0bc5822f6872518e685cd0ed484121\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:10.070523 containerd[1640]: time="2025-07-10T00:28:10.070369762Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.38.3\" with image id \"sha256:8bde16470b09d1963e19456806d73180c9778a6c2b3c1fda2335c67c1cd4ce93\", repo tag \"quay.io/tigera/operator:v1.38.3\", repo digest \"quay.io/tigera/operator@sha256:dbf1bad0def7b5955dc8e4aeee96e23ead0bc5822f6872518e685cd0ed484121\", size \"25052538\" in 1.889481803s" Jul 10 00:28:10.070523 containerd[1640]: time="2025-07-10T00:28:10.070387113Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.3\" returns image reference \"sha256:8bde16470b09d1963e19456806d73180c9778a6c2b3c1fda2335c67c1cd4ce93\"" Jul 10 00:28:10.071474 containerd[1640]: time="2025-07-10T00:28:10.071458089Z" level=info msg="CreateContainer within sandbox \"ec7c5d2911897dd45c3b93893331774ea4892b6c0cc1c8bc650b20f551561f0c\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Jul 10 00:28:10.077514 containerd[1640]: time="2025-07-10T00:28:10.077222392Z" level=info msg="Container 342f2a7fc1ca5576fdfc538caef4f924c649a58c4702f7481ca80d17870f8940: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:28:10.079249 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1247040480.mount: Deactivated successfully. Jul 10 00:28:10.087357 containerd[1640]: time="2025-07-10T00:28:10.087339802Z" level=info msg="CreateContainer within sandbox \"ec7c5d2911897dd45c3b93893331774ea4892b6c0cc1c8bc650b20f551561f0c\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"342f2a7fc1ca5576fdfc538caef4f924c649a58c4702f7481ca80d17870f8940\"" Jul 10 00:28:10.087834 containerd[1640]: time="2025-07-10T00:28:10.087816351Z" level=info msg="StartContainer for \"342f2a7fc1ca5576fdfc538caef4f924c649a58c4702f7481ca80d17870f8940\"" Jul 10 00:28:10.089153 containerd[1640]: time="2025-07-10T00:28:10.089120663Z" level=info msg="connecting to shim 342f2a7fc1ca5576fdfc538caef4f924c649a58c4702f7481ca80d17870f8940" address="unix:///run/containerd/s/0d1372dc2c5c19554bb7a77830f6dd3c38c20cafa220353bd1b98c57721103bb" protocol=ttrpc version=3 Jul 10 00:28:10.106376 systemd[1]: Started cri-containerd-342f2a7fc1ca5576fdfc538caef4f924c649a58c4702f7481ca80d17870f8940.scope - libcontainer container 342f2a7fc1ca5576fdfc538caef4f924c649a58c4702f7481ca80d17870f8940. Jul 10 00:28:10.123451 containerd[1640]: time="2025-07-10T00:28:10.123431982Z" level=info msg="StartContainer for \"342f2a7fc1ca5576fdfc538caef4f924c649a58c4702f7481ca80d17870f8940\" returns successfully" Jul 10 00:28:14.575477 kubelet[2915]: I0710 00:28:14.575427 2915 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-5bf8dfcb4-qbslx" podStartSLOduration=5.684743022 podStartE2EDuration="7.575413197s" podCreationTimestamp="2025-07-10 00:28:07 +0000 UTC" firstStartedPulling="2025-07-10 00:28:08.18018606 +0000 UTC m=+6.924966272" lastFinishedPulling="2025-07-10 00:28:10.070856235 +0000 UTC m=+8.815636447" observedRunningTime="2025-07-10 00:28:10.440467278 +0000 UTC m=+9.185247494" watchObservedRunningTime="2025-07-10 00:28:14.575413197 +0000 UTC m=+13.320193411" Jul 10 00:28:15.124280 sudo[1959]: pam_unix(sudo:session): session closed for user root Jul 10 00:28:15.126651 sshd[1958]: Connection closed by 139.178.68.195 port 56826 Jul 10 00:28:15.128552 sshd-session[1956]: pam_unix(sshd:session): session closed for user core Jul 10 00:28:15.131353 systemd[1]: sshd@7-139.178.70.108:22-139.178.68.195:56826.service: Deactivated successfully. Jul 10 00:28:15.133306 systemd[1]: session-9.scope: Deactivated successfully. Jul 10 00:28:15.133528 systemd[1]: session-9.scope: Consumed 3.094s CPU time, 153.6M memory peak. Jul 10 00:28:15.136601 systemd-logind[1614]: Session 9 logged out. Waiting for processes to exit. Jul 10 00:28:15.138940 systemd-logind[1614]: Removed session 9. Jul 10 00:28:17.660183 systemd[1]: Created slice kubepods-besteffort-pod93ae8767_c2ff_4aef_aacd_fac06dda5c94.slice - libcontainer container kubepods-besteffort-pod93ae8767_c2ff_4aef_aacd_fac06dda5c94.slice. Jul 10 00:28:17.693279 kubelet[2915]: I0710 00:28:17.693251 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2sdm\" (UniqueName: \"kubernetes.io/projected/93ae8767-c2ff-4aef-aacd-fac06dda5c94-kube-api-access-k2sdm\") pod \"calico-typha-6d8c447964-6hgkr\" (UID: \"93ae8767-c2ff-4aef-aacd-fac06dda5c94\") " pod="calico-system/calico-typha-6d8c447964-6hgkr" Jul 10 00:28:17.693279 kubelet[2915]: I0710 00:28:17.693281 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/93ae8767-c2ff-4aef-aacd-fac06dda5c94-tigera-ca-bundle\") pod \"calico-typha-6d8c447964-6hgkr\" (UID: \"93ae8767-c2ff-4aef-aacd-fac06dda5c94\") " pod="calico-system/calico-typha-6d8c447964-6hgkr" Jul 10 00:28:17.693638 kubelet[2915]: I0710 00:28:17.693294 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/93ae8767-c2ff-4aef-aacd-fac06dda5c94-typha-certs\") pod \"calico-typha-6d8c447964-6hgkr\" (UID: \"93ae8767-c2ff-4aef-aacd-fac06dda5c94\") " pod="calico-system/calico-typha-6d8c447964-6hgkr" Jul 10 00:28:17.985617 containerd[1640]: time="2025-07-10T00:28:17.984780791Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-6d8c447964-6hgkr,Uid:93ae8767-c2ff-4aef-aacd-fac06dda5c94,Namespace:calico-system,Attempt:0,}" Jul 10 00:28:17.995593 kubelet[2915]: I0710 00:28:17.995330 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/afc60f75-eb3c-47ec-9d9f-1bcd50e60c76-policysync\") pod \"calico-node-f5gtj\" (UID: \"afc60f75-eb3c-47ec-9d9f-1bcd50e60c76\") " pod="calico-system/calico-node-f5gtj" Jul 10 00:28:17.995593 kubelet[2915]: I0710 00:28:17.995362 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsltl\" (UniqueName: \"kubernetes.io/projected/afc60f75-eb3c-47ec-9d9f-1bcd50e60c76-kube-api-access-dsltl\") pod \"calico-node-f5gtj\" (UID: \"afc60f75-eb3c-47ec-9d9f-1bcd50e60c76\") " pod="calico-system/calico-node-f5gtj" Jul 10 00:28:17.995593 kubelet[2915]: I0710 00:28:17.995413 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/afc60f75-eb3c-47ec-9d9f-1bcd50e60c76-tigera-ca-bundle\") pod \"calico-node-f5gtj\" (UID: \"afc60f75-eb3c-47ec-9d9f-1bcd50e60c76\") " pod="calico-system/calico-node-f5gtj" Jul 10 00:28:17.995593 kubelet[2915]: I0710 00:28:17.995429 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/afc60f75-eb3c-47ec-9d9f-1bcd50e60c76-cni-net-dir\") pod \"calico-node-f5gtj\" (UID: \"afc60f75-eb3c-47ec-9d9f-1bcd50e60c76\") " pod="calico-system/calico-node-f5gtj" Jul 10 00:28:17.995593 kubelet[2915]: I0710 00:28:17.995442 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/afc60f75-eb3c-47ec-9d9f-1bcd50e60c76-cni-bin-dir\") pod \"calico-node-f5gtj\" (UID: \"afc60f75-eb3c-47ec-9d9f-1bcd50e60c76\") " pod="calico-system/calico-node-f5gtj" Jul 10 00:28:17.995776 kubelet[2915]: I0710 00:28:17.995453 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/afc60f75-eb3c-47ec-9d9f-1bcd50e60c76-var-lib-calico\") pod \"calico-node-f5gtj\" (UID: \"afc60f75-eb3c-47ec-9d9f-1bcd50e60c76\") " pod="calico-system/calico-node-f5gtj" Jul 10 00:28:17.997177 kubelet[2915]: I0710 00:28:17.995945 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/afc60f75-eb3c-47ec-9d9f-1bcd50e60c76-cni-log-dir\") pod \"calico-node-f5gtj\" (UID: \"afc60f75-eb3c-47ec-9d9f-1bcd50e60c76\") " pod="calico-system/calico-node-f5gtj" Jul 10 00:28:17.997177 kubelet[2915]: I0710 00:28:17.995963 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/afc60f75-eb3c-47ec-9d9f-1bcd50e60c76-lib-modules\") pod \"calico-node-f5gtj\" (UID: \"afc60f75-eb3c-47ec-9d9f-1bcd50e60c76\") " pod="calico-system/calico-node-f5gtj" Jul 10 00:28:17.997177 kubelet[2915]: I0710 00:28:17.995975 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/afc60f75-eb3c-47ec-9d9f-1bcd50e60c76-node-certs\") pod \"calico-node-f5gtj\" (UID: \"afc60f75-eb3c-47ec-9d9f-1bcd50e60c76\") " pod="calico-system/calico-node-f5gtj" Jul 10 00:28:17.997177 kubelet[2915]: I0710 00:28:17.995987 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/afc60f75-eb3c-47ec-9d9f-1bcd50e60c76-flexvol-driver-host\") pod \"calico-node-f5gtj\" (UID: \"afc60f75-eb3c-47ec-9d9f-1bcd50e60c76\") " pod="calico-system/calico-node-f5gtj" Jul 10 00:28:17.997177 kubelet[2915]: I0710 00:28:17.996008 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/afc60f75-eb3c-47ec-9d9f-1bcd50e60c76-var-run-calico\") pod \"calico-node-f5gtj\" (UID: \"afc60f75-eb3c-47ec-9d9f-1bcd50e60c76\") " pod="calico-system/calico-node-f5gtj" Jul 10 00:28:17.997313 kubelet[2915]: I0710 00:28:17.996018 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/afc60f75-eb3c-47ec-9d9f-1bcd50e60c76-xtables-lock\") pod \"calico-node-f5gtj\" (UID: \"afc60f75-eb3c-47ec-9d9f-1bcd50e60c76\") " pod="calico-system/calico-node-f5gtj" Jul 10 00:28:17.997846 systemd[1]: Created slice kubepods-besteffort-podafc60f75_eb3c_47ec_9d9f_1bcd50e60c76.slice - libcontainer container kubepods-besteffort-podafc60f75_eb3c_47ec_9d9f_1bcd50e60c76.slice. Jul 10 00:28:18.014324 containerd[1640]: time="2025-07-10T00:28:18.014262779Z" level=info msg="connecting to shim da2a207ed1967602a04d05efe4fd08c7da3fc5ab0a11ec590304f96ad2ffdd31" address="unix:///run/containerd/s/4d2a34df27b4ab99216220f4876ed4eb22ce011575ae113778495c45ce91fa32" namespace=k8s.io protocol=ttrpc version=3 Jul 10 00:28:18.046334 systemd[1]: Started cri-containerd-da2a207ed1967602a04d05efe4fd08c7da3fc5ab0a11ec590304f96ad2ffdd31.scope - libcontainer container da2a207ed1967602a04d05efe4fd08c7da3fc5ab0a11ec590304f96ad2ffdd31. Jul 10 00:28:18.098759 containerd[1640]: time="2025-07-10T00:28:18.098708630Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-6d8c447964-6hgkr,Uid:93ae8767-c2ff-4aef-aacd-fac06dda5c94,Namespace:calico-system,Attempt:0,} returns sandbox id \"da2a207ed1967602a04d05efe4fd08c7da3fc5ab0a11ec590304f96ad2ffdd31\"" Jul 10 00:28:18.103013 containerd[1640]: time="2025-07-10T00:28:18.102811211Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.2\"" Jul 10 00:28:18.107310 kubelet[2915]: E0710 00:28:18.107277 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.107536 kubelet[2915]: W0710 00:28:18.107343 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.107536 kubelet[2915]: E0710 00:28:18.107362 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.240258 kubelet[2915]: E0710 00:28:18.240171 2915 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-pzv7w" podUID="2587365c-21ee-44fd-91df-0ab3a24762fc" Jul 10 00:28:18.293247 kubelet[2915]: E0710 00:28:18.293138 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.293247 kubelet[2915]: W0710 00:28:18.293153 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.293247 kubelet[2915]: E0710 00:28:18.293185 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.293377 kubelet[2915]: E0710 00:28:18.293290 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.293377 kubelet[2915]: W0710 00:28:18.293295 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.293377 kubelet[2915]: E0710 00:28:18.293301 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.293377 kubelet[2915]: E0710 00:28:18.293370 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.293377 kubelet[2915]: W0710 00:28:18.293375 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.293466 kubelet[2915]: E0710 00:28:18.293380 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.293466 kubelet[2915]: E0710 00:28:18.293453 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.293466 kubelet[2915]: W0710 00:28:18.293457 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.293466 kubelet[2915]: E0710 00:28:18.293462 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.293548 kubelet[2915]: E0710 00:28:18.293537 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.293548 kubelet[2915]: W0710 00:28:18.293543 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.293588 kubelet[2915]: E0710 00:28:18.293548 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.293625 kubelet[2915]: E0710 00:28:18.293622 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.293649 kubelet[2915]: W0710 00:28:18.293626 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.293649 kubelet[2915]: E0710 00:28:18.293631 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.293710 kubelet[2915]: E0710 00:28:18.293698 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.293710 kubelet[2915]: W0710 00:28:18.293707 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.293752 kubelet[2915]: E0710 00:28:18.293712 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.293793 kubelet[2915]: E0710 00:28:18.293781 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.293815 kubelet[2915]: W0710 00:28:18.293791 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.293815 kubelet[2915]: E0710 00:28:18.293800 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.300591 kubelet[2915]: E0710 00:28:18.293886 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.300591 kubelet[2915]: W0710 00:28:18.293890 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.300591 kubelet[2915]: E0710 00:28:18.293895 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.300591 kubelet[2915]: E0710 00:28:18.293960 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.300591 kubelet[2915]: W0710 00:28:18.293965 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.300591 kubelet[2915]: E0710 00:28:18.293969 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.300591 kubelet[2915]: E0710 00:28:18.294060 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.300591 kubelet[2915]: W0710 00:28:18.294065 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.300591 kubelet[2915]: E0710 00:28:18.294070 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.300591 kubelet[2915]: E0710 00:28:18.294172 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.300756 kubelet[2915]: W0710 00:28:18.294178 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.300756 kubelet[2915]: E0710 00:28:18.294183 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.300756 kubelet[2915]: E0710 00:28:18.294273 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.300756 kubelet[2915]: W0710 00:28:18.294278 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.300756 kubelet[2915]: E0710 00:28:18.294283 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.300756 kubelet[2915]: E0710 00:28:18.294373 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.300756 kubelet[2915]: W0710 00:28:18.294380 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.300756 kubelet[2915]: E0710 00:28:18.294384 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.300756 kubelet[2915]: E0710 00:28:18.294477 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.300756 kubelet[2915]: W0710 00:28:18.294481 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.305174 kubelet[2915]: E0710 00:28:18.294486 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.305174 kubelet[2915]: E0710 00:28:18.294574 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.305174 kubelet[2915]: W0710 00:28:18.294579 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.305174 kubelet[2915]: E0710 00:28:18.294584 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.305174 kubelet[2915]: E0710 00:28:18.294680 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.305174 kubelet[2915]: W0710 00:28:18.294684 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.305174 kubelet[2915]: E0710 00:28:18.294690 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.305174 kubelet[2915]: E0710 00:28:18.294783 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.305174 kubelet[2915]: W0710 00:28:18.294787 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.305174 kubelet[2915]: E0710 00:28:18.294791 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.305335 kubelet[2915]: E0710 00:28:18.294885 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.305335 kubelet[2915]: W0710 00:28:18.294892 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.305335 kubelet[2915]: E0710 00:28:18.294897 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.305335 kubelet[2915]: E0710 00:28:18.294982 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.305335 kubelet[2915]: W0710 00:28:18.294987 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.305335 kubelet[2915]: E0710 00:28:18.294991 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.305335 kubelet[2915]: E0710 00:28:18.298307 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.305335 kubelet[2915]: W0710 00:28:18.298315 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.305335 kubelet[2915]: E0710 00:28:18.298324 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.305472 kubelet[2915]: I0710 00:28:18.298347 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/2587365c-21ee-44fd-91df-0ab3a24762fc-varrun\") pod \"csi-node-driver-pzv7w\" (UID: \"2587365c-21ee-44fd-91df-0ab3a24762fc\") " pod="calico-system/csi-node-driver-pzv7w" Jul 10 00:28:18.305472 kubelet[2915]: E0710 00:28:18.298462 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.305472 kubelet[2915]: W0710 00:28:18.298471 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.305472 kubelet[2915]: E0710 00:28:18.298483 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.305472 kubelet[2915]: E0710 00:28:18.298586 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.305472 kubelet[2915]: W0710 00:28:18.298593 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.305472 kubelet[2915]: E0710 00:28:18.298611 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.305472 kubelet[2915]: E0710 00:28:18.298707 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.305472 kubelet[2915]: W0710 00:28:18.298712 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.305623 kubelet[2915]: E0710 00:28:18.298717 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.305623 kubelet[2915]: I0710 00:28:18.298731 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54b4h\" (UniqueName: \"kubernetes.io/projected/2587365c-21ee-44fd-91df-0ab3a24762fc-kube-api-access-54b4h\") pod \"csi-node-driver-pzv7w\" (UID: \"2587365c-21ee-44fd-91df-0ab3a24762fc\") " pod="calico-system/csi-node-driver-pzv7w" Jul 10 00:28:18.305623 kubelet[2915]: E0710 00:28:18.298841 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.305623 kubelet[2915]: W0710 00:28:18.298847 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.305623 kubelet[2915]: E0710 00:28:18.298859 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.305623 kubelet[2915]: I0710 00:28:18.298868 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2587365c-21ee-44fd-91df-0ab3a24762fc-kubelet-dir\") pod \"csi-node-driver-pzv7w\" (UID: \"2587365c-21ee-44fd-91df-0ab3a24762fc\") " pod="calico-system/csi-node-driver-pzv7w" Jul 10 00:28:18.305623 kubelet[2915]: E0710 00:28:18.298957 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.305623 kubelet[2915]: W0710 00:28:18.298964 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.305758 kubelet[2915]: E0710 00:28:18.298969 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.305758 kubelet[2915]: I0710 00:28:18.299026 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2587365c-21ee-44fd-91df-0ab3a24762fc-registration-dir\") pod \"csi-node-driver-pzv7w\" (UID: \"2587365c-21ee-44fd-91df-0ab3a24762fc\") " pod="calico-system/csi-node-driver-pzv7w" Jul 10 00:28:18.305758 kubelet[2915]: E0710 00:28:18.299062 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.305758 kubelet[2915]: W0710 00:28:18.299067 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.305758 kubelet[2915]: E0710 00:28:18.299077 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.305758 kubelet[2915]: E0710 00:28:18.299240 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.305758 kubelet[2915]: W0710 00:28:18.299247 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.305758 kubelet[2915]: E0710 00:28:18.299266 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.305758 kubelet[2915]: E0710 00:28:18.299421 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.305891 kubelet[2915]: W0710 00:28:18.299427 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.305891 kubelet[2915]: E0710 00:28:18.299439 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.305891 kubelet[2915]: E0710 00:28:18.299553 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.305891 kubelet[2915]: W0710 00:28:18.299558 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.305891 kubelet[2915]: E0710 00:28:18.299570 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.305891 kubelet[2915]: E0710 00:28:18.299655 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.305891 kubelet[2915]: W0710 00:28:18.299660 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.305891 kubelet[2915]: E0710 00:28:18.299669 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.305891 kubelet[2915]: I0710 00:28:18.299679 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2587365c-21ee-44fd-91df-0ab3a24762fc-socket-dir\") pod \"csi-node-driver-pzv7w\" (UID: \"2587365c-21ee-44fd-91df-0ab3a24762fc\") " pod="calico-system/csi-node-driver-pzv7w" Jul 10 00:28:18.306024 kubelet[2915]: E0710 00:28:18.299742 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.306024 kubelet[2915]: W0710 00:28:18.299747 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.306024 kubelet[2915]: E0710 00:28:18.299758 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.306024 kubelet[2915]: E0710 00:28:18.299841 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.306024 kubelet[2915]: W0710 00:28:18.299846 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.306024 kubelet[2915]: E0710 00:28:18.299855 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.306024 kubelet[2915]: E0710 00:28:18.299944 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.306024 kubelet[2915]: W0710 00:28:18.299948 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.306024 kubelet[2915]: E0710 00:28:18.299953 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.306024 kubelet[2915]: E0710 00:28:18.300053 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.306186 kubelet[2915]: W0710 00:28:18.300058 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.306186 kubelet[2915]: E0710 00:28:18.300063 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.308077 containerd[1640]: time="2025-07-10T00:28:18.308059030Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-f5gtj,Uid:afc60f75-eb3c-47ec-9d9f-1bcd50e60c76,Namespace:calico-system,Attempt:0,}" Jul 10 00:28:18.388121 containerd[1640]: time="2025-07-10T00:28:18.388028123Z" level=info msg="connecting to shim 8b90456f67aa597456857686a92371d19ec3f2e1bcb861c2abc6c4ba132c72b4" address="unix:///run/containerd/s/f79aefe1967c5a13065a2b4085ecc3678090c4bd55fc4472652f6b875f2abf6a" namespace=k8s.io protocol=ttrpc version=3 Jul 10 00:28:18.401486 kubelet[2915]: E0710 00:28:18.401398 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.401486 kubelet[2915]: W0710 00:28:18.401412 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.401486 kubelet[2915]: E0710 00:28:18.401425 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.402475 kubelet[2915]: E0710 00:28:18.402409 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.402475 kubelet[2915]: W0710 00:28:18.402417 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.402475 kubelet[2915]: E0710 00:28:18.402424 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.402623 kubelet[2915]: E0710 00:28:18.402561 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.402623 kubelet[2915]: W0710 00:28:18.402567 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.402623 kubelet[2915]: E0710 00:28:18.402572 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.402757 kubelet[2915]: E0710 00:28:18.402705 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.402757 kubelet[2915]: W0710 00:28:18.402711 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.402757 kubelet[2915]: E0710 00:28:18.402716 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.402892 kubelet[2915]: E0710 00:28:18.402833 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.402892 kubelet[2915]: W0710 00:28:18.402839 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.402892 kubelet[2915]: E0710 00:28:18.402843 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.403027 kubelet[2915]: E0710 00:28:18.402971 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.403027 kubelet[2915]: W0710 00:28:18.402977 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.403027 kubelet[2915]: E0710 00:28:18.402982 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.403843 kubelet[2915]: E0710 00:28:18.403836 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.403882 kubelet[2915]: W0710 00:28:18.403876 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.403919 kubelet[2915]: E0710 00:28:18.403913 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.404029 kubelet[2915]: E0710 00:28:18.404024 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.404066 kubelet[2915]: W0710 00:28:18.404060 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.404111 kubelet[2915]: E0710 00:28:18.404105 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.404273 kubelet[2915]: E0710 00:28:18.404267 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.404313 kubelet[2915]: W0710 00:28:18.404307 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.404358 kubelet[2915]: E0710 00:28:18.404352 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.404787 kubelet[2915]: E0710 00:28:18.404763 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.404900 kubelet[2915]: W0710 00:28:18.404892 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.404980 kubelet[2915]: E0710 00:28:18.404971 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.405084 kubelet[2915]: E0710 00:28:18.405075 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.405084 kubelet[2915]: W0710 00:28:18.405082 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.405254 kubelet[2915]: E0710 00:28:18.405092 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.405254 kubelet[2915]: E0710 00:28:18.405171 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.405254 kubelet[2915]: W0710 00:28:18.405176 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.405254 kubelet[2915]: E0710 00:28:18.405182 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.405254 kubelet[2915]: E0710 00:28:18.405246 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.405254 kubelet[2915]: W0710 00:28:18.405251 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.405254 kubelet[2915]: E0710 00:28:18.405259 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.405643 kubelet[2915]: E0710 00:28:18.405353 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.405643 kubelet[2915]: W0710 00:28:18.405358 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.405643 kubelet[2915]: E0710 00:28:18.405363 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.405643 kubelet[2915]: E0710 00:28:18.405434 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.405643 kubelet[2915]: W0710 00:28:18.405439 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.405643 kubelet[2915]: E0710 00:28:18.405447 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.405643 kubelet[2915]: E0710 00:28:18.405506 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.405643 kubelet[2915]: W0710 00:28:18.405510 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.405643 kubelet[2915]: E0710 00:28:18.405514 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.406001 kubelet[2915]: E0710 00:28:18.405683 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.406001 kubelet[2915]: W0710 00:28:18.405687 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.406001 kubelet[2915]: E0710 00:28:18.405695 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.406446 kubelet[2915]: E0710 00:28:18.406233 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.406446 kubelet[2915]: W0710 00:28:18.406241 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.406584 kubelet[2915]: E0710 00:28:18.406247 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.407011 kubelet[2915]: E0710 00:28:18.406879 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.407011 kubelet[2915]: W0710 00:28:18.406887 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.407011 kubelet[2915]: E0710 00:28:18.406905 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.407312 kubelet[2915]: E0710 00:28:18.407271 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.407312 kubelet[2915]: W0710 00:28:18.407279 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.407312 kubelet[2915]: E0710 00:28:18.407289 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.407677 kubelet[2915]: E0710 00:28:18.407547 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.407677 kubelet[2915]: W0710 00:28:18.407554 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.407677 kubelet[2915]: E0710 00:28:18.407560 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.408069 systemd[1]: Started cri-containerd-8b90456f67aa597456857686a92371d19ec3f2e1bcb861c2abc6c4ba132c72b4.scope - libcontainer container 8b90456f67aa597456857686a92371d19ec3f2e1bcb861c2abc6c4ba132c72b4. Jul 10 00:28:18.408275 kubelet[2915]: E0710 00:28:18.408168 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.408275 kubelet[2915]: W0710 00:28:18.408185 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.408275 kubelet[2915]: E0710 00:28:18.408234 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.408518 kubelet[2915]: E0710 00:28:18.408411 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.408518 kubelet[2915]: W0710 00:28:18.408417 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.408518 kubelet[2915]: E0710 00:28:18.408432 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.409083 kubelet[2915]: E0710 00:28:18.408920 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.409083 kubelet[2915]: W0710 00:28:18.408927 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.409083 kubelet[2915]: E0710 00:28:18.408936 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.410387 kubelet[2915]: E0710 00:28:18.410329 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.410387 kubelet[2915]: W0710 00:28:18.410338 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.410387 kubelet[2915]: E0710 00:28:18.410346 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.420325 kubelet[2915]: E0710 00:28:18.420304 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:18.420325 kubelet[2915]: W0710 00:28:18.420317 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:18.420325 kubelet[2915]: E0710 00:28:18.420329 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:18.473877 containerd[1640]: time="2025-07-10T00:28:18.473851389Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-f5gtj,Uid:afc60f75-eb3c-47ec-9d9f-1bcd50e60c76,Namespace:calico-system,Attempt:0,} returns sandbox id \"8b90456f67aa597456857686a92371d19ec3f2e1bcb861c2abc6c4ba132c72b4\"" Jul 10 00:28:19.395109 kubelet[2915]: E0710 00:28:19.394772 2915 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-pzv7w" podUID="2587365c-21ee-44fd-91df-0ab3a24762fc" Jul 10 00:28:19.610905 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1821636468.mount: Deactivated successfully. Jul 10 00:28:21.003375 containerd[1640]: time="2025-07-10T00:28:21.003012271Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:21.004056 containerd[1640]: time="2025-07-10T00:28:21.004045305Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.30.2: active requests=0, bytes read=35233364" Jul 10 00:28:21.005016 containerd[1640]: time="2025-07-10T00:28:21.004455230Z" level=info msg="ImageCreate event name:\"sha256:b3baa600c7ff9cd50dc12f2529ef263aaa346dbeca13c77c6553d661fd216b54\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:21.005607 containerd[1640]: time="2025-07-10T00:28:21.005595165Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:da29d745efe5eb7d25f765d3aa439f3fe60710a458efe39c285e58b02bd961af\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:21.006270 containerd[1640]: time="2025-07-10T00:28:21.006255170Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.30.2\" with image id \"sha256:b3baa600c7ff9cd50dc12f2529ef263aaa346dbeca13c77c6553d661fd216b54\", repo tag \"ghcr.io/flatcar/calico/typha:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:da29d745efe5eb7d25f765d3aa439f3fe60710a458efe39c285e58b02bd961af\", size \"35233218\" in 2.903241441s" Jul 10 00:28:21.006327 containerd[1640]: time="2025-07-10T00:28:21.006317795Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.2\" returns image reference \"sha256:b3baa600c7ff9cd50dc12f2529ef263aaa346dbeca13c77c6553d661fd216b54\"" Jul 10 00:28:21.007048 containerd[1640]: time="2025-07-10T00:28:21.007010540Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\"" Jul 10 00:28:21.017809 containerd[1640]: time="2025-07-10T00:28:21.017788327Z" level=info msg="CreateContainer within sandbox \"da2a207ed1967602a04d05efe4fd08c7da3fc5ab0a11ec590304f96ad2ffdd31\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Jul 10 00:28:21.027341 containerd[1640]: time="2025-07-10T00:28:21.027316615Z" level=info msg="Container e0689694875fa02dd8bd48d407627812298bccb104948e9bb7cfdd2e994205b6: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:28:21.031776 containerd[1640]: time="2025-07-10T00:28:21.031750211Z" level=info msg="CreateContainer within sandbox \"da2a207ed1967602a04d05efe4fd08c7da3fc5ab0a11ec590304f96ad2ffdd31\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"e0689694875fa02dd8bd48d407627812298bccb104948e9bb7cfdd2e994205b6\"" Jul 10 00:28:21.033295 containerd[1640]: time="2025-07-10T00:28:21.032661132Z" level=info msg="StartContainer for \"e0689694875fa02dd8bd48d407627812298bccb104948e9bb7cfdd2e994205b6\"" Jul 10 00:28:21.033905 containerd[1640]: time="2025-07-10T00:28:21.033887335Z" level=info msg="connecting to shim e0689694875fa02dd8bd48d407627812298bccb104948e9bb7cfdd2e994205b6" address="unix:///run/containerd/s/4d2a34df27b4ab99216220f4876ed4eb22ce011575ae113778495c45ce91fa32" protocol=ttrpc version=3 Jul 10 00:28:21.051506 systemd[1]: Started cri-containerd-e0689694875fa02dd8bd48d407627812298bccb104948e9bb7cfdd2e994205b6.scope - libcontainer container e0689694875fa02dd8bd48d407627812298bccb104948e9bb7cfdd2e994205b6. Jul 10 00:28:21.098775 containerd[1640]: time="2025-07-10T00:28:21.098754510Z" level=info msg="StartContainer for \"e0689694875fa02dd8bd48d407627812298bccb104948e9bb7cfdd2e994205b6\" returns successfully" Jul 10 00:28:21.396450 kubelet[2915]: E0710 00:28:21.396376 2915 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-pzv7w" podUID="2587365c-21ee-44fd-91df-0ab3a24762fc" Jul 10 00:28:21.500901 kubelet[2915]: I0710 00:28:21.500471 2915 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-6d8c447964-6hgkr" podStartSLOduration=1.59578356 podStartE2EDuration="4.50046057s" podCreationTimestamp="2025-07-10 00:28:17 +0000 UTC" firstStartedPulling="2025-07-10 00:28:18.102277141 +0000 UTC m=+16.847057354" lastFinishedPulling="2025-07-10 00:28:21.006954147 +0000 UTC m=+19.751734364" observedRunningTime="2025-07-10 00:28:21.466353579 +0000 UTC m=+20.211133810" watchObservedRunningTime="2025-07-10 00:28:21.50046057 +0000 UTC m=+20.245240787" Jul 10 00:28:21.517668 kubelet[2915]: E0710 00:28:21.517604 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:21.517668 kubelet[2915]: W0710 00:28:21.517620 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:21.517668 kubelet[2915]: E0710 00:28:21.517634 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:21.518158 kubelet[2915]: E0710 00:28:21.518150 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:21.518244 kubelet[2915]: W0710 00:28:21.518237 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:21.518280 kubelet[2915]: E0710 00:28:21.518275 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:21.518870 kubelet[2915]: E0710 00:28:21.518832 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:21.518870 kubelet[2915]: W0710 00:28:21.518840 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:21.518870 kubelet[2915]: E0710 00:28:21.518847 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:21.519051 kubelet[2915]: E0710 00:28:21.519015 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:21.519051 kubelet[2915]: W0710 00:28:21.519021 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:21.519051 kubelet[2915]: E0710 00:28:21.519026 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:21.520178 kubelet[2915]: E0710 00:28:21.519193 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:21.520178 kubelet[2915]: W0710 00:28:21.519199 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:21.520178 kubelet[2915]: E0710 00:28:21.519205 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:21.520377 kubelet[2915]: E0710 00:28:21.520338 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:21.520377 kubelet[2915]: W0710 00:28:21.520345 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:21.520377 kubelet[2915]: E0710 00:28:21.520351 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:21.520547 kubelet[2915]: E0710 00:28:21.520516 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:21.520547 kubelet[2915]: W0710 00:28:21.520522 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:21.520547 kubelet[2915]: E0710 00:28:21.520527 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:21.520702 kubelet[2915]: E0710 00:28:21.520667 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:21.520702 kubelet[2915]: W0710 00:28:21.520673 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:21.520702 kubelet[2915]: E0710 00:28:21.520678 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:21.520861 kubelet[2915]: E0710 00:28:21.520833 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:21.520861 kubelet[2915]: W0710 00:28:21.520839 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:21.520861 kubelet[2915]: E0710 00:28:21.520844 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:21.521032 kubelet[2915]: E0710 00:28:21.521002 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:21.521032 kubelet[2915]: W0710 00:28:21.521009 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:21.521032 kubelet[2915]: E0710 00:28:21.521015 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:21.521202 kubelet[2915]: E0710 00:28:21.521161 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:21.521202 kubelet[2915]: W0710 00:28:21.521178 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:21.521202 kubelet[2915]: E0710 00:28:21.521183 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:21.521365 kubelet[2915]: E0710 00:28:21.521333 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:21.521365 kubelet[2915]: W0710 00:28:21.521338 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:21.521365 kubelet[2915]: E0710 00:28:21.521343 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:21.521524 kubelet[2915]: E0710 00:28:21.521496 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:21.521524 kubelet[2915]: W0710 00:28:21.521502 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:21.521524 kubelet[2915]: E0710 00:28:21.521506 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:21.521673 kubelet[2915]: E0710 00:28:21.521645 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:21.521673 kubelet[2915]: W0710 00:28:21.521650 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:21.521673 kubelet[2915]: E0710 00:28:21.521655 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:21.522231 kubelet[2915]: E0710 00:28:21.522187 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:21.522231 kubelet[2915]: W0710 00:28:21.522194 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:21.522231 kubelet[2915]: E0710 00:28:21.522200 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:21.523391 kubelet[2915]: E0710 00:28:21.523367 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:21.523391 kubelet[2915]: W0710 00:28:21.523374 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:21.523391 kubelet[2915]: E0710 00:28:21.523381 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:21.523591 kubelet[2915]: E0710 00:28:21.523578 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:21.523591 kubelet[2915]: W0710 00:28:21.523584 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:21.523671 kubelet[2915]: E0710 00:28:21.523641 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:21.523788 kubelet[2915]: E0710 00:28:21.523776 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:21.523788 kubelet[2915]: W0710 00:28:21.523782 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:21.524222 kubelet[2915]: E0710 00:28:21.524182 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:21.524354 kubelet[2915]: E0710 00:28:21.524341 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:21.524354 kubelet[2915]: W0710 00:28:21.524347 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:21.524429 kubelet[2915]: E0710 00:28:21.524404 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:21.524540 kubelet[2915]: E0710 00:28:21.524527 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:21.524540 kubelet[2915]: W0710 00:28:21.524533 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:21.524614 kubelet[2915]: E0710 00:28:21.524594 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:21.524730 kubelet[2915]: E0710 00:28:21.524717 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:21.524730 kubelet[2915]: W0710 00:28:21.524724 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:21.524855 kubelet[2915]: E0710 00:28:21.524839 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:21.525211 kubelet[2915]: E0710 00:28:21.525205 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:21.525251 kubelet[2915]: W0710 00:28:21.525245 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:21.525364 kubelet[2915]: E0710 00:28:21.525350 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:21.525459 kubelet[2915]: E0710 00:28:21.525406 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:21.525459 kubelet[2915]: W0710 00:28:21.525411 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:21.525517 kubelet[2915]: E0710 00:28:21.525511 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:21.525571 kubelet[2915]: E0710 00:28:21.525567 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:21.525667 kubelet[2915]: W0710 00:28:21.525599 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:21.525667 kubelet[2915]: E0710 00:28:21.525612 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:21.527267 kubelet[2915]: E0710 00:28:21.527259 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:21.527344 kubelet[2915]: W0710 00:28:21.527335 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:21.527396 kubelet[2915]: E0710 00:28:21.527376 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:21.527508 kubelet[2915]: E0710 00:28:21.527495 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:21.527508 kubelet[2915]: W0710 00:28:21.527501 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:21.527633 kubelet[2915]: E0710 00:28:21.527619 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:21.527732 kubelet[2915]: E0710 00:28:21.527686 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:21.527732 kubelet[2915]: W0710 00:28:21.527692 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:21.527785 kubelet[2915]: E0710 00:28:21.527779 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:21.527845 kubelet[2915]: E0710 00:28:21.527834 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:21.527845 kubelet[2915]: W0710 00:28:21.527839 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:21.527930 kubelet[2915]: E0710 00:28:21.527890 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:21.528040 kubelet[2915]: E0710 00:28:21.528035 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:21.528080 kubelet[2915]: W0710 00:28:21.528075 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:21.528116 kubelet[2915]: E0710 00:28:21.528112 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:21.528313 kubelet[2915]: E0710 00:28:21.528250 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:21.528348 kubelet[2915]: W0710 00:28:21.528342 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:21.528385 kubelet[2915]: E0710 00:28:21.528380 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:21.528489 kubelet[2915]: E0710 00:28:21.528484 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:21.528560 kubelet[2915]: W0710 00:28:21.528518 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:21.528602 kubelet[2915]: E0710 00:28:21.528589 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:21.528879 kubelet[2915]: E0710 00:28:21.528688 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:21.528879 kubelet[2915]: W0710 00:28:21.528693 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:21.528879 kubelet[2915]: E0710 00:28:21.528699 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:21.529027 kubelet[2915]: E0710 00:28:21.529021 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:21.529072 kubelet[2915]: W0710 00:28:21.529066 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:21.529104 kubelet[2915]: E0710 00:28:21.529099 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:22.528375 kubelet[2915]: E0710 00:28:22.528312 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:22.528375 kubelet[2915]: W0710 00:28:22.528328 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:22.528375 kubelet[2915]: E0710 00:28:22.528342 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:22.528843 kubelet[2915]: E0710 00:28:22.528797 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:22.528843 kubelet[2915]: W0710 00:28:22.528807 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:22.528843 kubelet[2915]: E0710 00:28:22.528816 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:22.529056 kubelet[2915]: E0710 00:28:22.529013 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:22.529056 kubelet[2915]: W0710 00:28:22.529022 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:22.529056 kubelet[2915]: E0710 00:28:22.529029 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:22.529312 kubelet[2915]: E0710 00:28:22.529265 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:22.529312 kubelet[2915]: W0710 00:28:22.529273 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:22.529312 kubelet[2915]: E0710 00:28:22.529281 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:22.529518 kubelet[2915]: E0710 00:28:22.529481 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:22.529518 kubelet[2915]: W0710 00:28:22.529488 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:22.529518 kubelet[2915]: E0710 00:28:22.529495 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:22.529727 kubelet[2915]: E0710 00:28:22.529690 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:22.529727 kubelet[2915]: W0710 00:28:22.529697 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:22.529727 kubelet[2915]: E0710 00:28:22.529704 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:22.529974 kubelet[2915]: E0710 00:28:22.529967 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:22.530047 kubelet[2915]: W0710 00:28:22.530012 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:22.530047 kubelet[2915]: E0710 00:28:22.530021 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:22.530243 kubelet[2915]: E0710 00:28:22.530218 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:22.530243 kubelet[2915]: W0710 00:28:22.530226 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:22.530243 kubelet[2915]: E0710 00:28:22.530233 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:22.530543 kubelet[2915]: E0710 00:28:22.530504 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:22.530543 kubelet[2915]: W0710 00:28:22.530513 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:22.530543 kubelet[2915]: E0710 00:28:22.530520 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:22.530759 kubelet[2915]: E0710 00:28:22.530752 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:22.530874 kubelet[2915]: W0710 00:28:22.530838 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:22.530874 kubelet[2915]: E0710 00:28:22.530849 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:22.531083 kubelet[2915]: E0710 00:28:22.531039 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:22.531083 kubelet[2915]: W0710 00:28:22.531049 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:22.531083 kubelet[2915]: E0710 00:28:22.531056 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:22.531298 kubelet[2915]: E0710 00:28:22.531273 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:22.531298 kubelet[2915]: W0710 00:28:22.531281 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:22.531298 kubelet[2915]: E0710 00:28:22.531288 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:22.531527 kubelet[2915]: E0710 00:28:22.531488 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:22.531527 kubelet[2915]: W0710 00:28:22.531496 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:22.531527 kubelet[2915]: E0710 00:28:22.531503 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:22.531731 kubelet[2915]: E0710 00:28:22.531693 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:22.531731 kubelet[2915]: W0710 00:28:22.531700 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:22.531731 kubelet[2915]: E0710 00:28:22.531707 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:22.531919 kubelet[2915]: E0710 00:28:22.531895 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:22.531919 kubelet[2915]: W0710 00:28:22.531903 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:22.531919 kubelet[2915]: E0710 00:28:22.531911 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:22.532233 kubelet[2915]: E0710 00:28:22.532195 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:22.532233 kubelet[2915]: W0710 00:28:22.532218 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:22.532233 kubelet[2915]: E0710 00:28:22.532224 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:22.532401 kubelet[2915]: E0710 00:28:22.532389 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:22.532401 kubelet[2915]: W0710 00:28:22.532395 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:22.532507 kubelet[2915]: E0710 00:28:22.532451 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:22.532567 kubelet[2915]: E0710 00:28:22.532561 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:22.532605 kubelet[2915]: W0710 00:28:22.532595 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:22.532639 kubelet[2915]: E0710 00:28:22.532634 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:22.533014 kubelet[2915]: E0710 00:28:22.533000 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:22.533014 kubelet[2915]: W0710 00:28:22.533007 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:22.533145 kubelet[2915]: E0710 00:28:22.533067 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:22.533789 kubelet[2915]: E0710 00:28:22.533627 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:22.533789 kubelet[2915]: W0710 00:28:22.533640 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:22.533868 kubelet[2915]: E0710 00:28:22.533833 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:22.533893 kubelet[2915]: E0710 00:28:22.533879 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:22.533893 kubelet[2915]: W0710 00:28:22.533884 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:22.534001 kubelet[2915]: E0710 00:28:22.533980 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:22.534041 kubelet[2915]: E0710 00:28:22.534023 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:22.534041 kubelet[2915]: W0710 00:28:22.534028 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:22.534216 kubelet[2915]: E0710 00:28:22.534176 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:22.534240 kubelet[2915]: E0710 00:28:22.534230 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:22.534240 kubelet[2915]: W0710 00:28:22.534235 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:22.534530 kubelet[2915]: E0710 00:28:22.534327 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:22.534530 kubelet[2915]: E0710 00:28:22.534487 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:22.534530 kubelet[2915]: W0710 00:28:22.534492 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:22.534530 kubelet[2915]: E0710 00:28:22.534502 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:22.534773 kubelet[2915]: E0710 00:28:22.534762 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:22.534773 kubelet[2915]: W0710 00:28:22.534769 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:22.534817 kubelet[2915]: E0710 00:28:22.534776 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:22.534852 kubelet[2915]: E0710 00:28:22.534844 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:22.534852 kubelet[2915]: W0710 00:28:22.534851 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:22.534892 kubelet[2915]: E0710 00:28:22.534856 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:22.535063 kubelet[2915]: E0710 00:28:22.534939 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:22.535063 kubelet[2915]: W0710 00:28:22.534945 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:22.535063 kubelet[2915]: E0710 00:28:22.534956 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:22.535123 kubelet[2915]: E0710 00:28:22.535074 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:22.535123 kubelet[2915]: W0710 00:28:22.535078 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:22.535123 kubelet[2915]: E0710 00:28:22.535083 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:22.535194 kubelet[2915]: E0710 00:28:22.535147 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:22.535194 kubelet[2915]: W0710 00:28:22.535151 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:22.535194 kubelet[2915]: E0710 00:28:22.535156 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:22.535562 kubelet[2915]: E0710 00:28:22.535300 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:22.535562 kubelet[2915]: W0710 00:28:22.535307 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:22.535562 kubelet[2915]: E0710 00:28:22.535312 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:22.535562 kubelet[2915]: E0710 00:28:22.535493 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:22.535562 kubelet[2915]: W0710 00:28:22.535498 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:22.535562 kubelet[2915]: E0710 00:28:22.535503 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:22.535690 kubelet[2915]: E0710 00:28:22.535618 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:22.535690 kubelet[2915]: W0710 00:28:22.535623 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:22.535690 kubelet[2915]: E0710 00:28:22.535633 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:22.535738 kubelet[2915]: E0710 00:28:22.535714 2915 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jul 10 00:28:22.535738 kubelet[2915]: W0710 00:28:22.535718 2915 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jul 10 00:28:22.535738 kubelet[2915]: E0710 00:28:22.535723 2915 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jul 10 00:28:22.599952 containerd[1640]: time="2025-07-10T00:28:22.599921031Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:22.610659 containerd[1640]: time="2025-07-10T00:28:22.602513936Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2: active requests=0, bytes read=4446956" Jul 10 00:28:22.610659 containerd[1640]: time="2025-07-10T00:28:22.606685837Z" level=info msg="ImageCreate event name:\"sha256:639615519fa6f7bc4b4756066ba9780068fd291eacc36c120f6c555e62f2b00e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:22.611177 containerd[1640]: time="2025-07-10T00:28:22.611155285Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:972be127eaecd7d1a2d5393b8d14f1ae8f88550bee83e0519e9590c7e15eb41b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:22.611517 containerd[1640]: time="2025-07-10T00:28:22.611494572Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\" with image id \"sha256:639615519fa6f7bc4b4756066ba9780068fd291eacc36c120f6c555e62f2b00e\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:972be127eaecd7d1a2d5393b8d14f1ae8f88550bee83e0519e9590c7e15eb41b\", size \"5939619\" in 1.604375928s" Jul 10 00:28:22.611612 containerd[1640]: time="2025-07-10T00:28:22.611557931Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\" returns image reference \"sha256:639615519fa6f7bc4b4756066ba9780068fd291eacc36c120f6c555e62f2b00e\"" Jul 10 00:28:22.617443 containerd[1640]: time="2025-07-10T00:28:22.617427766Z" level=info msg="CreateContainer within sandbox \"8b90456f67aa597456857686a92371d19ec3f2e1bcb861c2abc6c4ba132c72b4\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Jul 10 00:28:22.656554 containerd[1640]: time="2025-07-10T00:28:22.656294255Z" level=info msg="Container e496cf02b15d49965412476881f0991daa94aaf41c0205b47de82ee3dfcf2f2e: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:28:22.659197 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount856343847.mount: Deactivated successfully. Jul 10 00:28:22.726100 containerd[1640]: time="2025-07-10T00:28:22.725890159Z" level=info msg="CreateContainer within sandbox \"8b90456f67aa597456857686a92371d19ec3f2e1bcb861c2abc6c4ba132c72b4\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"e496cf02b15d49965412476881f0991daa94aaf41c0205b47de82ee3dfcf2f2e\"" Jul 10 00:28:22.727329 containerd[1640]: time="2025-07-10T00:28:22.726750262Z" level=info msg="StartContainer for \"e496cf02b15d49965412476881f0991daa94aaf41c0205b47de82ee3dfcf2f2e\"" Jul 10 00:28:22.731178 containerd[1640]: time="2025-07-10T00:28:22.731087173Z" level=info msg="connecting to shim e496cf02b15d49965412476881f0991daa94aaf41c0205b47de82ee3dfcf2f2e" address="unix:///run/containerd/s/f79aefe1967c5a13065a2b4085ecc3678090c4bd55fc4472652f6b875f2abf6a" protocol=ttrpc version=3 Jul 10 00:28:22.750287 systemd[1]: Started cri-containerd-e496cf02b15d49965412476881f0991daa94aaf41c0205b47de82ee3dfcf2f2e.scope - libcontainer container e496cf02b15d49965412476881f0991daa94aaf41c0205b47de82ee3dfcf2f2e. Jul 10 00:28:22.786075 containerd[1640]: time="2025-07-10T00:28:22.785995191Z" level=info msg="StartContainer for \"e496cf02b15d49965412476881f0991daa94aaf41c0205b47de82ee3dfcf2f2e\" returns successfully" Jul 10 00:28:22.789383 systemd[1]: cri-containerd-e496cf02b15d49965412476881f0991daa94aaf41c0205b47de82ee3dfcf2f2e.scope: Deactivated successfully. Jul 10 00:28:22.824881 containerd[1640]: time="2025-07-10T00:28:22.824848291Z" level=info msg="TaskExit event in podsandbox handler container_id:\"e496cf02b15d49965412476881f0991daa94aaf41c0205b47de82ee3dfcf2f2e\" id:\"e496cf02b15d49965412476881f0991daa94aaf41c0205b47de82ee3dfcf2f2e\" pid:3614 exited_at:{seconds:1752107302 nanos:790234221}" Jul 10 00:28:22.825268 containerd[1640]: time="2025-07-10T00:28:22.825248885Z" level=info msg="received exit event container_id:\"e496cf02b15d49965412476881f0991daa94aaf41c0205b47de82ee3dfcf2f2e\" id:\"e496cf02b15d49965412476881f0991daa94aaf41c0205b47de82ee3dfcf2f2e\" pid:3614 exited_at:{seconds:1752107302 nanos:790234221}" Jul 10 00:28:22.840819 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e496cf02b15d49965412476881f0991daa94aaf41c0205b47de82ee3dfcf2f2e-rootfs.mount: Deactivated successfully. Jul 10 00:28:23.394652 kubelet[2915]: E0710 00:28:23.394355 2915 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-pzv7w" podUID="2587365c-21ee-44fd-91df-0ab3a24762fc" Jul 10 00:28:23.466645 containerd[1640]: time="2025-07-10T00:28:23.466609769Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.2\"" Jul 10 00:28:25.398904 kubelet[2915]: E0710 00:28:25.398799 2915 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-pzv7w" podUID="2587365c-21ee-44fd-91df-0ab3a24762fc" Jul 10 00:28:26.145129 containerd[1640]: time="2025-07-10T00:28:26.145083638Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:26.145693 containerd[1640]: time="2025-07-10T00:28:26.145645886Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.30.2: active requests=0, bytes read=70436221" Jul 10 00:28:26.146891 containerd[1640]: time="2025-07-10T00:28:26.145990818Z" level=info msg="ImageCreate event name:\"sha256:77a357d0d33e3016e61153f7d2b7de72371579c4aaeb767fb7ef0af606fe1630\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:26.147431 containerd[1640]: time="2025-07-10T00:28:26.147394706Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:50686775cc60acb78bd92a66fa2d84e1700b2d8e43a718fbadbf35e59baefb4d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:26.148155 containerd[1640]: time="2025-07-10T00:28:26.148135907Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.30.2\" with image id \"sha256:77a357d0d33e3016e61153f7d2b7de72371579c4aaeb767fb7ef0af606fe1630\", repo tag \"ghcr.io/flatcar/calico/cni:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:50686775cc60acb78bd92a66fa2d84e1700b2d8e43a718fbadbf35e59baefb4d\", size \"71928924\" in 2.681500237s" Jul 10 00:28:26.148248 containerd[1640]: time="2025-07-10T00:28:26.148234634Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.2\" returns image reference \"sha256:77a357d0d33e3016e61153f7d2b7de72371579c4aaeb767fb7ef0af606fe1630\"" Jul 10 00:28:26.150686 containerd[1640]: time="2025-07-10T00:28:26.150664005Z" level=info msg="CreateContainer within sandbox \"8b90456f67aa597456857686a92371d19ec3f2e1bcb861c2abc6c4ba132c72b4\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Jul 10 00:28:26.158410 containerd[1640]: time="2025-07-10T00:28:26.158388335Z" level=info msg="Container 1615ad8139738c2858ed0c9f07bf37e47a43a39d01d94538df2a131f0d426451: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:28:26.161783 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount885191540.mount: Deactivated successfully. Jul 10 00:28:26.166220 containerd[1640]: time="2025-07-10T00:28:26.166190730Z" level=info msg="CreateContainer within sandbox \"8b90456f67aa597456857686a92371d19ec3f2e1bcb861c2abc6c4ba132c72b4\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"1615ad8139738c2858ed0c9f07bf37e47a43a39d01d94538df2a131f0d426451\"" Jul 10 00:28:26.166702 containerd[1640]: time="2025-07-10T00:28:26.166666373Z" level=info msg="StartContainer for \"1615ad8139738c2858ed0c9f07bf37e47a43a39d01d94538df2a131f0d426451\"" Jul 10 00:28:26.167803 containerd[1640]: time="2025-07-10T00:28:26.167752763Z" level=info msg="connecting to shim 1615ad8139738c2858ed0c9f07bf37e47a43a39d01d94538df2a131f0d426451" address="unix:///run/containerd/s/f79aefe1967c5a13065a2b4085ecc3678090c4bd55fc4472652f6b875f2abf6a" protocol=ttrpc version=3 Jul 10 00:28:26.186266 systemd[1]: Started cri-containerd-1615ad8139738c2858ed0c9f07bf37e47a43a39d01d94538df2a131f0d426451.scope - libcontainer container 1615ad8139738c2858ed0c9f07bf37e47a43a39d01d94538df2a131f0d426451. Jul 10 00:28:26.209394 containerd[1640]: time="2025-07-10T00:28:26.209370519Z" level=info msg="StartContainer for \"1615ad8139738c2858ed0c9f07bf37e47a43a39d01d94538df2a131f0d426451\" returns successfully" Jul 10 00:28:27.558300 kubelet[2915]: E0710 00:28:27.558149 2915 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-pzv7w" podUID="2587365c-21ee-44fd-91df-0ab3a24762fc" Jul 10 00:28:27.712793 systemd[1]: cri-containerd-1615ad8139738c2858ed0c9f07bf37e47a43a39d01d94538df2a131f0d426451.scope: Deactivated successfully. Jul 10 00:28:27.713457 systemd[1]: cri-containerd-1615ad8139738c2858ed0c9f07bf37e47a43a39d01d94538df2a131f0d426451.scope: Consumed 292ms CPU time, 156.4M memory peak, 1.1M read from disk, 171.2M written to disk. Jul 10 00:28:27.746175 containerd[1640]: time="2025-07-10T00:28:27.746034597Z" level=info msg="received exit event container_id:\"1615ad8139738c2858ed0c9f07bf37e47a43a39d01d94538df2a131f0d426451\" id:\"1615ad8139738c2858ed0c9f07bf37e47a43a39d01d94538df2a131f0d426451\" pid:3671 exited_at:{seconds:1752107307 nanos:745888424}" Jul 10 00:28:27.748231 containerd[1640]: time="2025-07-10T00:28:27.748210334Z" level=info msg="TaskExit event in podsandbox handler container_id:\"1615ad8139738c2858ed0c9f07bf37e47a43a39d01d94538df2a131f0d426451\" id:\"1615ad8139738c2858ed0c9f07bf37e47a43a39d01d94538df2a131f0d426451\" pid:3671 exited_at:{seconds:1752107307 nanos:745888424}" Jul 10 00:28:27.805447 kubelet[2915]: I0710 00:28:27.805419 2915 kubelet_node_status.go:488] "Fast updating node status as it just became ready" Jul 10 00:28:27.808859 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-1615ad8139738c2858ed0c9f07bf37e47a43a39d01d94538df2a131f0d426451-rootfs.mount: Deactivated successfully. Jul 10 00:28:27.853590 systemd[1]: Created slice kubepods-burstable-pod783faa36_cde4_4aa6_9139_b51d9b52af99.slice - libcontainer container kubepods-burstable-pod783faa36_cde4_4aa6_9139_b51d9b52af99.slice. Jul 10 00:28:27.860187 kubelet[2915]: I0710 00:28:27.859847 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7ph7\" (UniqueName: \"kubernetes.io/projected/5e5edbe5-bd6e-4f88-b5f5-53f7df248f0b-kube-api-access-z7ph7\") pod \"whisker-b7d48c89d-nd884\" (UID: \"5e5edbe5-bd6e-4f88-b5f5-53f7df248f0b\") " pod="calico-system/whisker-b7d48c89d-nd884" Jul 10 00:28:27.860187 kubelet[2915]: I0710 00:28:27.859868 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/5e5edbe5-bd6e-4f88-b5f5-53f7df248f0b-whisker-backend-key-pair\") pod \"whisker-b7d48c89d-nd884\" (UID: \"5e5edbe5-bd6e-4f88-b5f5-53f7df248f0b\") " pod="calico-system/whisker-b7d48c89d-nd884" Jul 10 00:28:27.860187 kubelet[2915]: I0710 00:28:27.859879 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svwt2\" (UniqueName: \"kubernetes.io/projected/783faa36-cde4-4aa6-9139-b51d9b52af99-kube-api-access-svwt2\") pod \"coredns-7c65d6cfc9-g5zz7\" (UID: \"783faa36-cde4-4aa6-9139-b51d9b52af99\") " pod="kube-system/coredns-7c65d6cfc9-g5zz7" Jul 10 00:28:27.860187 kubelet[2915]: I0710 00:28:27.859888 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e5edbe5-bd6e-4f88-b5f5-53f7df248f0b-whisker-ca-bundle\") pod \"whisker-b7d48c89d-nd884\" (UID: \"5e5edbe5-bd6e-4f88-b5f5-53f7df248f0b\") " pod="calico-system/whisker-b7d48c89d-nd884" Jul 10 00:28:27.860187 kubelet[2915]: I0710 00:28:27.859899 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0841b7eb-4ad6-4003-b535-92aae1f1d1f2-config\") pod \"goldmane-58fd7646b9-52ptt\" (UID: \"0841b7eb-4ad6-4003-b535-92aae1f1d1f2\") " pod="calico-system/goldmane-58fd7646b9-52ptt" Jul 10 00:28:27.860670 kubelet[2915]: I0710 00:28:27.859908 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0841b7eb-4ad6-4003-b535-92aae1f1d1f2-goldmane-ca-bundle\") pod \"goldmane-58fd7646b9-52ptt\" (UID: \"0841b7eb-4ad6-4003-b535-92aae1f1d1f2\") " pod="calico-system/goldmane-58fd7646b9-52ptt" Jul 10 00:28:27.860670 kubelet[2915]: I0710 00:28:27.859917 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-526gg\" (UniqueName: \"kubernetes.io/projected/082d696e-d4fa-497e-ab24-928a088c4f21-kube-api-access-526gg\") pod \"calico-apiserver-7c75d6c687-8jqvg\" (UID: \"082d696e-d4fa-497e-ab24-928a088c4f21\") " pod="calico-apiserver/calico-apiserver-7c75d6c687-8jqvg" Jul 10 00:28:27.860670 kubelet[2915]: I0710 00:28:27.859926 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0a9d7ca6-fb8c-4239-ae97-5a87dba79acf-tigera-ca-bundle\") pod \"calico-kube-controllers-84bd499cb-6t459\" (UID: \"0a9d7ca6-fb8c-4239-ae97-5a87dba79acf\") " pod="calico-system/calico-kube-controllers-84bd499cb-6t459" Jul 10 00:28:27.860670 kubelet[2915]: I0710 00:28:27.859936 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/0841b7eb-4ad6-4003-b535-92aae1f1d1f2-goldmane-key-pair\") pod \"goldmane-58fd7646b9-52ptt\" (UID: \"0841b7eb-4ad6-4003-b535-92aae1f1d1f2\") " pod="calico-system/goldmane-58fd7646b9-52ptt" Jul 10 00:28:27.860670 kubelet[2915]: I0710 00:28:27.859945 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gkcc\" (UniqueName: \"kubernetes.io/projected/0841b7eb-4ad6-4003-b535-92aae1f1d1f2-kube-api-access-8gkcc\") pod \"goldmane-58fd7646b9-52ptt\" (UID: \"0841b7eb-4ad6-4003-b535-92aae1f1d1f2\") " pod="calico-system/goldmane-58fd7646b9-52ptt" Jul 10 00:28:27.860581 systemd[1]: Created slice kubepods-burstable-pod64915bcf_34a2_483d_ae5f_7e4c4ea7d4d2.slice - libcontainer container kubepods-burstable-pod64915bcf_34a2_483d_ae5f_7e4c4ea7d4d2.slice. Jul 10 00:28:27.860809 kubelet[2915]: I0710 00:28:27.859953 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/64915bcf-34a2-483d-ae5f-7e4c4ea7d4d2-config-volume\") pod \"coredns-7c65d6cfc9-8h9ft\" (UID: \"64915bcf-34a2-483d-ae5f-7e4c4ea7d4d2\") " pod="kube-system/coredns-7c65d6cfc9-8h9ft" Jul 10 00:28:27.860809 kubelet[2915]: I0710 00:28:27.859965 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhkc2\" (UniqueName: \"kubernetes.io/projected/0a9d7ca6-fb8c-4239-ae97-5a87dba79acf-kube-api-access-rhkc2\") pod \"calico-kube-controllers-84bd499cb-6t459\" (UID: \"0a9d7ca6-fb8c-4239-ae97-5a87dba79acf\") " pod="calico-system/calico-kube-controllers-84bd499cb-6t459" Jul 10 00:28:27.860809 kubelet[2915]: I0710 00:28:27.859977 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pgwp\" (UniqueName: \"kubernetes.io/projected/9271cfd0-877e-47f5-b015-c6ecba29e40b-kube-api-access-2pgwp\") pod \"calico-apiserver-7c75d6c687-wlhzf\" (UID: \"9271cfd0-877e-47f5-b015-c6ecba29e40b\") " pod="calico-apiserver/calico-apiserver-7c75d6c687-wlhzf" Jul 10 00:28:27.860809 kubelet[2915]: I0710 00:28:27.859987 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/9271cfd0-877e-47f5-b015-c6ecba29e40b-calico-apiserver-certs\") pod \"calico-apiserver-7c75d6c687-wlhzf\" (UID: \"9271cfd0-877e-47f5-b015-c6ecba29e40b\") " pod="calico-apiserver/calico-apiserver-7c75d6c687-wlhzf" Jul 10 00:28:27.860809 kubelet[2915]: I0710 00:28:27.860002 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jftpc\" (UniqueName: \"kubernetes.io/projected/64915bcf-34a2-483d-ae5f-7e4c4ea7d4d2-kube-api-access-jftpc\") pod \"coredns-7c65d6cfc9-8h9ft\" (UID: \"64915bcf-34a2-483d-ae5f-7e4c4ea7d4d2\") " pod="kube-system/coredns-7c65d6cfc9-8h9ft" Jul 10 00:28:27.860899 kubelet[2915]: I0710 00:28:27.860011 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/082d696e-d4fa-497e-ab24-928a088c4f21-calico-apiserver-certs\") pod \"calico-apiserver-7c75d6c687-8jqvg\" (UID: \"082d696e-d4fa-497e-ab24-928a088c4f21\") " pod="calico-apiserver/calico-apiserver-7c75d6c687-8jqvg" Jul 10 00:28:27.860899 kubelet[2915]: I0710 00:28:27.860020 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/783faa36-cde4-4aa6-9139-b51d9b52af99-config-volume\") pod \"coredns-7c65d6cfc9-g5zz7\" (UID: \"783faa36-cde4-4aa6-9139-b51d9b52af99\") " pod="kube-system/coredns-7c65d6cfc9-g5zz7" Jul 10 00:28:27.866833 systemd[1]: Created slice kubepods-besteffort-pod0a9d7ca6_fb8c_4239_ae97_5a87dba79acf.slice - libcontainer container kubepods-besteffort-pod0a9d7ca6_fb8c_4239_ae97_5a87dba79acf.slice. Jul 10 00:28:27.873177 systemd[1]: Created slice kubepods-besteffort-pod0841b7eb_4ad6_4003_b535_92aae1f1d1f2.slice - libcontainer container kubepods-besteffort-pod0841b7eb_4ad6_4003_b535_92aae1f1d1f2.slice. Jul 10 00:28:27.880229 systemd[1]: Created slice kubepods-besteffort-pod5e5edbe5_bd6e_4f88_b5f5_53f7df248f0b.slice - libcontainer container kubepods-besteffort-pod5e5edbe5_bd6e_4f88_b5f5_53f7df248f0b.slice. Jul 10 00:28:27.885094 systemd[1]: Created slice kubepods-besteffort-pod9271cfd0_877e_47f5_b015_c6ecba29e40b.slice - libcontainer container kubepods-besteffort-pod9271cfd0_877e_47f5_b015_c6ecba29e40b.slice. Jul 10 00:28:27.890396 systemd[1]: Created slice kubepods-besteffort-pod082d696e_d4fa_497e_ab24_928a088c4f21.slice - libcontainer container kubepods-besteffort-pod082d696e_d4fa_497e_ab24_928a088c4f21.slice. Jul 10 00:28:28.159517 containerd[1640]: time="2025-07-10T00:28:28.159448449Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-g5zz7,Uid:783faa36-cde4-4aa6-9139-b51d9b52af99,Namespace:kube-system,Attempt:0,}" Jul 10 00:28:28.165104 containerd[1640]: time="2025-07-10T00:28:28.165003334Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-8h9ft,Uid:64915bcf-34a2-483d-ae5f-7e4c4ea7d4d2,Namespace:kube-system,Attempt:0,}" Jul 10 00:28:28.186343 containerd[1640]: time="2025-07-10T00:28:28.186316492Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-b7d48c89d-nd884,Uid:5e5edbe5-bd6e-4f88-b5f5-53f7df248f0b,Namespace:calico-system,Attempt:0,}" Jul 10 00:28:28.187376 containerd[1640]: time="2025-07-10T00:28:28.187355752Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-84bd499cb-6t459,Uid:0a9d7ca6-fb8c-4239-ae97-5a87dba79acf,Namespace:calico-system,Attempt:0,}" Jul 10 00:28:28.188138 containerd[1640]: time="2025-07-10T00:28:28.188122923Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-58fd7646b9-52ptt,Uid:0841b7eb-4ad6-4003-b535-92aae1f1d1f2,Namespace:calico-system,Attempt:0,}" Jul 10 00:28:28.190007 containerd[1640]: time="2025-07-10T00:28:28.189995311Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7c75d6c687-wlhzf,Uid:9271cfd0-877e-47f5-b015-c6ecba29e40b,Namespace:calico-apiserver,Attempt:0,}" Jul 10 00:28:28.196007 containerd[1640]: time="2025-07-10T00:28:28.195991666Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7c75d6c687-8jqvg,Uid:082d696e-d4fa-497e-ab24-928a088c4f21,Namespace:calico-apiserver,Attempt:0,}" Jul 10 00:28:28.450198 containerd[1640]: time="2025-07-10T00:28:28.450088450Z" level=error msg="Failed to destroy network for sandbox \"b9f32a3c5c5c5c4b180eadbe1a964eca6a609f2b702b535de28e5c9e56b839c1\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 10 00:28:28.450968 containerd[1640]: time="2025-07-10T00:28:28.450950918Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-b7d48c89d-nd884,Uid:5e5edbe5-bd6e-4f88-b5f5-53f7df248f0b,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"b9f32a3c5c5c5c4b180eadbe1a964eca6a609f2b702b535de28e5c9e56b839c1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 10 00:28:28.451677 kubelet[2915]: E0710 00:28:28.451647 2915 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b9f32a3c5c5c5c4b180eadbe1a964eca6a609f2b702b535de28e5c9e56b839c1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 10 00:28:28.452194 kubelet[2915]: E0710 00:28:28.452180 2915 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b9f32a3c5c5c5c4b180eadbe1a964eca6a609f2b702b535de28e5c9e56b839c1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-b7d48c89d-nd884" Jul 10 00:28:28.452260 kubelet[2915]: E0710 00:28:28.452241 2915 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b9f32a3c5c5c5c4b180eadbe1a964eca6a609f2b702b535de28e5c9e56b839c1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-b7d48c89d-nd884" Jul 10 00:28:28.452346 kubelet[2915]: E0710 00:28:28.452330 2915 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-b7d48c89d-nd884_calico-system(5e5edbe5-bd6e-4f88-b5f5-53f7df248f0b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-b7d48c89d-nd884_calico-system(5e5edbe5-bd6e-4f88-b5f5-53f7df248f0b)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"b9f32a3c5c5c5c4b180eadbe1a964eca6a609f2b702b535de28e5c9e56b839c1\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-b7d48c89d-nd884" podUID="5e5edbe5-bd6e-4f88-b5f5-53f7df248f0b" Jul 10 00:28:28.456316 containerd[1640]: time="2025-07-10T00:28:28.456289518Z" level=error msg="Failed to destroy network for sandbox \"0f2ffb1653f7d15d5a3940db5da958a95de0d1be64f92ed134fb1f2faec15383\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 10 00:28:28.456886 containerd[1640]: time="2025-07-10T00:28:28.456870569Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-84bd499cb-6t459,Uid:0a9d7ca6-fb8c-4239-ae97-5a87dba79acf,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"0f2ffb1653f7d15d5a3940db5da958a95de0d1be64f92ed134fb1f2faec15383\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 10 00:28:28.457262 containerd[1640]: time="2025-07-10T00:28:28.457030129Z" level=error msg="Failed to destroy network for sandbox \"b66ef0f6490e993557ea53e3b75f249c5a30c9de493d1aac3f3d86a2568b35d0\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 10 00:28:28.459214 kubelet[2915]: E0710 00:28:28.457877 2915 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0f2ffb1653f7d15d5a3940db5da958a95de0d1be64f92ed134fb1f2faec15383\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 10 00:28:28.459214 kubelet[2915]: E0710 00:28:28.457914 2915 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0f2ffb1653f7d15d5a3940db5da958a95de0d1be64f92ed134fb1f2faec15383\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-84bd499cb-6t459" Jul 10 00:28:28.459214 kubelet[2915]: E0710 00:28:28.457928 2915 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0f2ffb1653f7d15d5a3940db5da958a95de0d1be64f92ed134fb1f2faec15383\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-84bd499cb-6t459" Jul 10 00:28:28.459308 kubelet[2915]: E0710 00:28:28.457955 2915 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-84bd499cb-6t459_calico-system(0a9d7ca6-fb8c-4239-ae97-5a87dba79acf)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-84bd499cb-6t459_calico-system(0a9d7ca6-fb8c-4239-ae97-5a87dba79acf)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"0f2ffb1653f7d15d5a3940db5da958a95de0d1be64f92ed134fb1f2faec15383\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-84bd499cb-6t459" podUID="0a9d7ca6-fb8c-4239-ae97-5a87dba79acf" Jul 10 00:28:28.460608 containerd[1640]: time="2025-07-10T00:28:28.460584046Z" level=error msg="Failed to destroy network for sandbox \"d936d73b434a37ef5094d247460c34d2b7ef3ad1e838f2f2064c071426e3bccc\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 10 00:28:28.462019 containerd[1640]: time="2025-07-10T00:28:28.461928317Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-8h9ft,Uid:64915bcf-34a2-483d-ae5f-7e4c4ea7d4d2,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"b66ef0f6490e993557ea53e3b75f249c5a30c9de493d1aac3f3d86a2568b35d0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 10 00:28:28.462771 containerd[1640]: time="2025-07-10T00:28:28.462229866Z" level=error msg="Failed to destroy network for sandbox \"03f2ae8f436de53a3ec9cdc033753aa16bbc220ec3b00a227e25cae6837ebbc1\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 10 00:28:28.463275 kubelet[2915]: E0710 00:28:28.462839 2915 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b66ef0f6490e993557ea53e3b75f249c5a30c9de493d1aac3f3d86a2568b35d0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 10 00:28:28.463275 kubelet[2915]: E0710 00:28:28.462874 2915 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b66ef0f6490e993557ea53e3b75f249c5a30c9de493d1aac3f3d86a2568b35d0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7c65d6cfc9-8h9ft" Jul 10 00:28:28.463275 kubelet[2915]: E0710 00:28:28.462886 2915 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b66ef0f6490e993557ea53e3b75f249c5a30c9de493d1aac3f3d86a2568b35d0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7c65d6cfc9-8h9ft" Jul 10 00:28:28.463365 containerd[1640]: time="2025-07-10T00:28:28.463263238Z" level=error msg="Failed to destroy network for sandbox \"c69226c3af5ec3ee967dede1c7c5607e3c9c2537cf3cf6d48209d9e4ade61aa0\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 10 00:28:28.463387 kubelet[2915]: E0710 00:28:28.462924 2915 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7c65d6cfc9-8h9ft_kube-system(64915bcf-34a2-483d-ae5f-7e4c4ea7d4d2)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7c65d6cfc9-8h9ft_kube-system(64915bcf-34a2-483d-ae5f-7e4c4ea7d4d2)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"b66ef0f6490e993557ea53e3b75f249c5a30c9de493d1aac3f3d86a2568b35d0\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7c65d6cfc9-8h9ft" podUID="64915bcf-34a2-483d-ae5f-7e4c4ea7d4d2" Jul 10 00:28:28.463815 containerd[1640]: time="2025-07-10T00:28:28.463709434Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-58fd7646b9-52ptt,Uid:0841b7eb-4ad6-4003-b535-92aae1f1d1f2,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"03f2ae8f436de53a3ec9cdc033753aa16bbc220ec3b00a227e25cae6837ebbc1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 10 00:28:28.464088 containerd[1640]: time="2025-07-10T00:28:28.463973898Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7c75d6c687-wlhzf,Uid:9271cfd0-877e-47f5-b015-c6ecba29e40b,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"c69226c3af5ec3ee967dede1c7c5607e3c9c2537cf3cf6d48209d9e4ade61aa0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 10 00:28:28.464247 containerd[1640]: time="2025-07-10T00:28:28.464216420Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7c75d6c687-8jqvg,Uid:082d696e-d4fa-497e-ab24-928a088c4f21,Namespace:calico-apiserver,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"d936d73b434a37ef5094d247460c34d2b7ef3ad1e838f2f2064c071426e3bccc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 10 00:28:28.464609 kubelet[2915]: E0710 00:28:28.464554 2915 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d936d73b434a37ef5094d247460c34d2b7ef3ad1e838f2f2064c071426e3bccc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 10 00:28:28.465087 kubelet[2915]: E0710 00:28:28.464974 2915 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d936d73b434a37ef5094d247460c34d2b7ef3ad1e838f2f2064c071426e3bccc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7c75d6c687-8jqvg" Jul 10 00:28:28.465087 kubelet[2915]: E0710 00:28:28.464990 2915 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d936d73b434a37ef5094d247460c34d2b7ef3ad1e838f2f2064c071426e3bccc\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7c75d6c687-8jqvg" Jul 10 00:28:28.465087 kubelet[2915]: E0710 00:28:28.465011 2915 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-7c75d6c687-8jqvg_calico-apiserver(082d696e-d4fa-497e-ab24-928a088c4f21)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-7c75d6c687-8jqvg_calico-apiserver(082d696e-d4fa-497e-ab24-928a088c4f21)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"d936d73b434a37ef5094d247460c34d2b7ef3ad1e838f2f2064c071426e3bccc\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-7c75d6c687-8jqvg" podUID="082d696e-d4fa-497e-ab24-928a088c4f21" Jul 10 00:28:28.465199 kubelet[2915]: E0710 00:28:28.464934 2915 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"03f2ae8f436de53a3ec9cdc033753aa16bbc220ec3b00a227e25cae6837ebbc1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 10 00:28:28.465199 kubelet[2915]: E0710 00:28:28.464945 2915 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c69226c3af5ec3ee967dede1c7c5607e3c9c2537cf3cf6d48209d9e4ade61aa0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 10 00:28:28.465199 kubelet[2915]: E0710 00:28:28.465042 2915 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"03f2ae8f436de53a3ec9cdc033753aa16bbc220ec3b00a227e25cae6837ebbc1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-58fd7646b9-52ptt" Jul 10 00:28:28.465199 kubelet[2915]: E0710 00:28:28.465051 2915 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"03f2ae8f436de53a3ec9cdc033753aa16bbc220ec3b00a227e25cae6837ebbc1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-58fd7646b9-52ptt" Jul 10 00:28:28.465276 containerd[1640]: time="2025-07-10T00:28:28.465136475Z" level=error msg="Failed to destroy network for sandbox \"168e6d2521fb0e15b26ca412d3cf64b718df7ea3cae97c185bca362b91703fa0\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 10 00:28:28.465297 kubelet[2915]: E0710 00:28:28.465055 2915 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c69226c3af5ec3ee967dede1c7c5607e3c9c2537cf3cf6d48209d9e4ade61aa0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7c75d6c687-wlhzf" Jul 10 00:28:28.465297 kubelet[2915]: E0710 00:28:28.465066 2915 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c69226c3af5ec3ee967dede1c7c5607e3c9c2537cf3cf6d48209d9e4ade61aa0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7c75d6c687-wlhzf" Jul 10 00:28:28.465297 kubelet[2915]: E0710 00:28:28.465066 2915 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-58fd7646b9-52ptt_calico-system(0841b7eb-4ad6-4003-b535-92aae1f1d1f2)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-58fd7646b9-52ptt_calico-system(0841b7eb-4ad6-4003-b535-92aae1f1d1f2)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"03f2ae8f436de53a3ec9cdc033753aa16bbc220ec3b00a227e25cae6837ebbc1\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-58fd7646b9-52ptt" podUID="0841b7eb-4ad6-4003-b535-92aae1f1d1f2" Jul 10 00:28:28.465404 kubelet[2915]: E0710 00:28:28.465084 2915 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-7c75d6c687-wlhzf_calico-apiserver(9271cfd0-877e-47f5-b015-c6ecba29e40b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-7c75d6c687-wlhzf_calico-apiserver(9271cfd0-877e-47f5-b015-c6ecba29e40b)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"c69226c3af5ec3ee967dede1c7c5607e3c9c2537cf3cf6d48209d9e4ade61aa0\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-7c75d6c687-wlhzf" podUID="9271cfd0-877e-47f5-b015-c6ecba29e40b" Jul 10 00:28:28.465615 containerd[1640]: time="2025-07-10T00:28:28.465596880Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-g5zz7,Uid:783faa36-cde4-4aa6-9139-b51d9b52af99,Namespace:kube-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"168e6d2521fb0e15b26ca412d3cf64b718df7ea3cae97c185bca362b91703fa0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 10 00:28:28.466195 kubelet[2915]: E0710 00:28:28.465757 2915 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"168e6d2521fb0e15b26ca412d3cf64b718df7ea3cae97c185bca362b91703fa0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 10 00:28:28.466195 kubelet[2915]: E0710 00:28:28.465773 2915 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"168e6d2521fb0e15b26ca412d3cf64b718df7ea3cae97c185bca362b91703fa0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7c65d6cfc9-g5zz7" Jul 10 00:28:28.466195 kubelet[2915]: E0710 00:28:28.465786 2915 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"168e6d2521fb0e15b26ca412d3cf64b718df7ea3cae97c185bca362b91703fa0\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7c65d6cfc9-g5zz7" Jul 10 00:28:28.466284 kubelet[2915]: E0710 00:28:28.465823 2915 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7c65d6cfc9-g5zz7_kube-system(783faa36-cde4-4aa6-9139-b51d9b52af99)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7c65d6cfc9-g5zz7_kube-system(783faa36-cde4-4aa6-9139-b51d9b52af99)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"168e6d2521fb0e15b26ca412d3cf64b718df7ea3cae97c185bca362b91703fa0\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7c65d6cfc9-g5zz7" podUID="783faa36-cde4-4aa6-9139-b51d9b52af99" Jul 10 00:28:28.602181 containerd[1640]: time="2025-07-10T00:28:28.602146724Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.2\"" Jul 10 00:28:29.398420 systemd[1]: Created slice kubepods-besteffort-pod2587365c_21ee_44fd_91df_0ab3a24762fc.slice - libcontainer container kubepods-besteffort-pod2587365c_21ee_44fd_91df_0ab3a24762fc.slice. Jul 10 00:28:29.399651 containerd[1640]: time="2025-07-10T00:28:29.399624164Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-pzv7w,Uid:2587365c-21ee-44fd-91df-0ab3a24762fc,Namespace:calico-system,Attempt:0,}" Jul 10 00:28:29.430888 containerd[1640]: time="2025-07-10T00:28:29.430850692Z" level=error msg="Failed to destroy network for sandbox \"8e6dc63a0a12c84d133f3b6adaac05d5375d68c019d28f1582ad5d5b78f1defe\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 10 00:28:29.432182 systemd[1]: run-netns-cni\x2d76f6bdb1\x2d9e79\x2da776\x2d45a9\x2dcf1cba2f7e86.mount: Deactivated successfully. Jul 10 00:28:29.432543 containerd[1640]: time="2025-07-10T00:28:29.432517303Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-pzv7w,Uid:2587365c-21ee-44fd-91df-0ab3a24762fc,Namespace:calico-system,Attempt:0,} failed, error" error="rpc error: code = Unknown desc = failed to setup network for sandbox \"8e6dc63a0a12c84d133f3b6adaac05d5375d68c019d28f1582ad5d5b78f1defe\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 10 00:28:29.432811 kubelet[2915]: E0710 00:28:29.432780 2915 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8e6dc63a0a12c84d133f3b6adaac05d5375d68c019d28f1582ad5d5b78f1defe\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jul 10 00:28:29.433232 kubelet[2915]: E0710 00:28:29.432825 2915 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8e6dc63a0a12c84d133f3b6adaac05d5375d68c019d28f1582ad5d5b78f1defe\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-pzv7w" Jul 10 00:28:29.433232 kubelet[2915]: E0710 00:28:29.432837 2915 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8e6dc63a0a12c84d133f3b6adaac05d5375d68c019d28f1582ad5d5b78f1defe\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-pzv7w" Jul 10 00:28:29.433232 kubelet[2915]: E0710 00:28:29.432865 2915 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-pzv7w_calico-system(2587365c-21ee-44fd-91df-0ab3a24762fc)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-pzv7w_calico-system(2587365c-21ee-44fd-91df-0ab3a24762fc)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"8e6dc63a0a12c84d133f3b6adaac05d5375d68c019d28f1582ad5d5b78f1defe\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-pzv7w" podUID="2587365c-21ee-44fd-91df-0ab3a24762fc" Jul 10 00:28:32.789537 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2063199592.mount: Deactivated successfully. Jul 10 00:28:32.949867 containerd[1640]: time="2025-07-10T00:28:32.943046732Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.30.2: active requests=0, bytes read=158500163" Jul 10 00:28:32.953386 containerd[1640]: time="2025-07-10T00:28:32.953342093Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:32.961155 containerd[1640]: time="2025-07-10T00:28:32.961120202Z" level=info msg="ImageCreate event name:\"sha256:cc52550d767f73458fee2ee68db9db5de30d175e8fa4569ebdb43610127b6d20\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:32.961930 containerd[1640]: time="2025-07-10T00:28:32.961700369Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:e94d49349cc361ef2216d27dda4a097278984d778279f66e79b0616c827c6760\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:32.962507 containerd[1640]: time="2025-07-10T00:28:32.962486055Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.30.2\" with image id \"sha256:cc52550d767f73458fee2ee68db9db5de30d175e8fa4569ebdb43610127b6d20\", repo tag \"ghcr.io/flatcar/calico/node:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/node@sha256:e94d49349cc361ef2216d27dda4a097278984d778279f66e79b0616c827c6760\", size \"158500025\" in 4.359141714s" Jul 10 00:28:32.962542 containerd[1640]: time="2025-07-10T00:28:32.962510354Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.2\" returns image reference \"sha256:cc52550d767f73458fee2ee68db9db5de30d175e8fa4569ebdb43610127b6d20\"" Jul 10 00:28:32.990131 containerd[1640]: time="2025-07-10T00:28:32.990100427Z" level=info msg="CreateContainer within sandbox \"8b90456f67aa597456857686a92371d19ec3f2e1bcb861c2abc6c4ba132c72b4\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Jul 10 00:28:33.019184 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2664488181.mount: Deactivated successfully. Jul 10 00:28:33.019434 containerd[1640]: time="2025-07-10T00:28:33.019347527Z" level=info msg="Container 7342408a31cb64c6221572622b4fe6b7e010c365050de1a3cf3bc111989059fe: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:28:33.044357 containerd[1640]: time="2025-07-10T00:28:33.044314994Z" level=info msg="CreateContainer within sandbox \"8b90456f67aa597456857686a92371d19ec3f2e1bcb861c2abc6c4ba132c72b4\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"7342408a31cb64c6221572622b4fe6b7e010c365050de1a3cf3bc111989059fe\"" Jul 10 00:28:33.045720 containerd[1640]: time="2025-07-10T00:28:33.045626952Z" level=info msg="StartContainer for \"7342408a31cb64c6221572622b4fe6b7e010c365050de1a3cf3bc111989059fe\"" Jul 10 00:28:33.050123 containerd[1640]: time="2025-07-10T00:28:33.050108693Z" level=info msg="connecting to shim 7342408a31cb64c6221572622b4fe6b7e010c365050de1a3cf3bc111989059fe" address="unix:///run/containerd/s/f79aefe1967c5a13065a2b4085ecc3678090c4bd55fc4472652f6b875f2abf6a" protocol=ttrpc version=3 Jul 10 00:28:33.105300 systemd[1]: Started cri-containerd-7342408a31cb64c6221572622b4fe6b7e010c365050de1a3cf3bc111989059fe.scope - libcontainer container 7342408a31cb64c6221572622b4fe6b7e010c365050de1a3cf3bc111989059fe. Jul 10 00:28:33.150452 containerd[1640]: time="2025-07-10T00:28:33.150428946Z" level=info msg="StartContainer for \"7342408a31cb64c6221572622b4fe6b7e010c365050de1a3cf3bc111989059fe\" returns successfully" Jul 10 00:28:33.233493 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Jul 10 00:28:33.235385 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Jul 10 00:28:33.593092 kubelet[2915]: I0710 00:28:33.593045 2915 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e5edbe5-bd6e-4f88-b5f5-53f7df248f0b-whisker-ca-bundle\") pod \"5e5edbe5-bd6e-4f88-b5f5-53f7df248f0b\" (UID: \"5e5edbe5-bd6e-4f88-b5f5-53f7df248f0b\") " Jul 10 00:28:33.593092 kubelet[2915]: I0710 00:28:33.593078 2915 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7ph7\" (UniqueName: \"kubernetes.io/projected/5e5edbe5-bd6e-4f88-b5f5-53f7df248f0b-kube-api-access-z7ph7\") pod \"5e5edbe5-bd6e-4f88-b5f5-53f7df248f0b\" (UID: \"5e5edbe5-bd6e-4f88-b5f5-53f7df248f0b\") " Jul 10 00:28:33.593092 kubelet[2915]: I0710 00:28:33.593091 2915 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/5e5edbe5-bd6e-4f88-b5f5-53f7df248f0b-whisker-backend-key-pair\") pod \"5e5edbe5-bd6e-4f88-b5f5-53f7df248f0b\" (UID: \"5e5edbe5-bd6e-4f88-b5f5-53f7df248f0b\") " Jul 10 00:28:33.598601 kubelet[2915]: I0710 00:28:33.598447 2915 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e5edbe5-bd6e-4f88-b5f5-53f7df248f0b-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "5e5edbe5-bd6e-4f88-b5f5-53f7df248f0b" (UID: "5e5edbe5-bd6e-4f88-b5f5-53f7df248f0b"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jul 10 00:28:33.603950 kubelet[2915]: I0710 00:28:33.603919 2915 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e5edbe5-bd6e-4f88-b5f5-53f7df248f0b-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "5e5edbe5-bd6e-4f88-b5f5-53f7df248f0b" (UID: "5e5edbe5-bd6e-4f88-b5f5-53f7df248f0b"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGidValue "" Jul 10 00:28:33.604064 kubelet[2915]: I0710 00:28:33.604050 2915 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e5edbe5-bd6e-4f88-b5f5-53f7df248f0b-kube-api-access-z7ph7" (OuterVolumeSpecName: "kube-api-access-z7ph7") pod "5e5edbe5-bd6e-4f88-b5f5-53f7df248f0b" (UID: "5e5edbe5-bd6e-4f88-b5f5-53f7df248f0b"). InnerVolumeSpecName "kube-api-access-z7ph7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jul 10 00:28:33.617236 systemd[1]: Removed slice kubepods-besteffort-pod5e5edbe5_bd6e_4f88_b5f5_53f7df248f0b.slice - libcontainer container kubepods-besteffort-pod5e5edbe5_bd6e_4f88_b5f5_53f7df248f0b.slice. Jul 10 00:28:33.633431 kubelet[2915]: I0710 00:28:33.633330 2915 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-f5gtj" podStartSLOduration=2.145173301 podStartE2EDuration="16.633317713s" podCreationTimestamp="2025-07-10 00:28:17 +0000 UTC" firstStartedPulling="2025-07-10 00:28:18.475046766 +0000 UTC m=+17.219826978" lastFinishedPulling="2025-07-10 00:28:32.963191177 +0000 UTC m=+31.707971390" observedRunningTime="2025-07-10 00:28:33.625986188 +0000 UTC m=+32.370766411" watchObservedRunningTime="2025-07-10 00:28:33.633317713 +0000 UTC m=+32.378097935" Jul 10 00:28:33.694454 kubelet[2915]: I0710 00:28:33.694282 2915 reconciler_common.go:293] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e5edbe5-bd6e-4f88-b5f5-53f7df248f0b-whisker-ca-bundle\") on node \"localhost\" DevicePath \"\"" Jul 10 00:28:33.694454 kubelet[2915]: I0710 00:28:33.694323 2915 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7ph7\" (UniqueName: \"kubernetes.io/projected/5e5edbe5-bd6e-4f88-b5f5-53f7df248f0b-kube-api-access-z7ph7\") on node \"localhost\" DevicePath \"\"" Jul 10 00:28:33.694454 kubelet[2915]: I0710 00:28:33.694340 2915 reconciler_common.go:293] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/5e5edbe5-bd6e-4f88-b5f5-53f7df248f0b-whisker-backend-key-pair\") on node \"localhost\" DevicePath \"\"" Jul 10 00:28:33.710910 systemd[1]: Created slice kubepods-besteffort-podb3328a36_298d_4c0b_8e63_7cff7da58b7b.slice - libcontainer container kubepods-besteffort-podb3328a36_298d_4c0b_8e63_7cff7da58b7b.slice. Jul 10 00:28:33.757897 containerd[1640]: time="2025-07-10T00:28:33.757864197Z" level=info msg="TaskExit event in podsandbox handler container_id:\"7342408a31cb64c6221572622b4fe6b7e010c365050de1a3cf3bc111989059fe\" id:\"44f8ba29f5e82ae6f76d3d40748b41e5cdf0d56f70644add039311fdad0545ab\" pid:4022 exit_status:1 exited_at:{seconds:1752107313 nanos:753870802}" Jul 10 00:28:33.790118 systemd[1]: var-lib-kubelet-pods-5e5edbe5\x2dbd6e\x2d4f88\x2db5f5\x2d53f7df248f0b-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Jul 10 00:28:33.790365 systemd[1]: var-lib-kubelet-pods-5e5edbe5\x2dbd6e\x2d4f88\x2db5f5\x2d53f7df248f0b-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dz7ph7.mount: Deactivated successfully. Jul 10 00:28:33.895731 kubelet[2915]: I0710 00:28:33.895656 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/b3328a36-298d-4c0b-8e63-7cff7da58b7b-whisker-backend-key-pair\") pod \"whisker-7bf9c88658-hfwwb\" (UID: \"b3328a36-298d-4c0b-8e63-7cff7da58b7b\") " pod="calico-system/whisker-7bf9c88658-hfwwb" Jul 10 00:28:33.895731 kubelet[2915]: I0710 00:28:33.895685 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rccrf\" (UniqueName: \"kubernetes.io/projected/b3328a36-298d-4c0b-8e63-7cff7da58b7b-kube-api-access-rccrf\") pod \"whisker-7bf9c88658-hfwwb\" (UID: \"b3328a36-298d-4c0b-8e63-7cff7da58b7b\") " pod="calico-system/whisker-7bf9c88658-hfwwb" Jul 10 00:28:33.895731 kubelet[2915]: I0710 00:28:33.895704 2915 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b3328a36-298d-4c0b-8e63-7cff7da58b7b-whisker-ca-bundle\") pod \"whisker-7bf9c88658-hfwwb\" (UID: \"b3328a36-298d-4c0b-8e63-7cff7da58b7b\") " pod="calico-system/whisker-7bf9c88658-hfwwb" Jul 10 00:28:34.014173 containerd[1640]: time="2025-07-10T00:28:34.014140345Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-7bf9c88658-hfwwb,Uid:b3328a36-298d-4c0b-8e63-7cff7da58b7b,Namespace:calico-system,Attempt:0,}" Jul 10 00:28:34.371568 systemd-networkd[1524]: cali8fc8c348088: Link UP Jul 10 00:28:34.371732 systemd-networkd[1524]: cali8fc8c348088: Gained carrier Jul 10 00:28:34.382484 containerd[1640]: 2025-07-10 00:28:34.042 [INFO][4038] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Jul 10 00:28:34.382484 containerd[1640]: 2025-07-10 00:28:34.081 [INFO][4038] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-whisker--7bf9c88658--hfwwb-eth0 whisker-7bf9c88658- calico-system b3328a36-298d-4c0b-8e63-7cff7da58b7b 872 0 2025-07-10 00:28:33 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:7bf9c88658 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s localhost whisker-7bf9c88658-hfwwb eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] cali8fc8c348088 [] [] }} ContainerID="5d391391a4ed8f8df19521dedccf29836078c765bd4036bf2e8414fbd49e2ea3" Namespace="calico-system" Pod="whisker-7bf9c88658-hfwwb" WorkloadEndpoint="localhost-k8s-whisker--7bf9c88658--hfwwb-" Jul 10 00:28:34.382484 containerd[1640]: 2025-07-10 00:28:34.081 [INFO][4038] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="5d391391a4ed8f8df19521dedccf29836078c765bd4036bf2e8414fbd49e2ea3" Namespace="calico-system" Pod="whisker-7bf9c88658-hfwwb" WorkloadEndpoint="localhost-k8s-whisker--7bf9c88658--hfwwb-eth0" Jul 10 00:28:34.382484 containerd[1640]: 2025-07-10 00:28:34.324 [INFO][4049] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="5d391391a4ed8f8df19521dedccf29836078c765bd4036bf2e8414fbd49e2ea3" HandleID="k8s-pod-network.5d391391a4ed8f8df19521dedccf29836078c765bd4036bf2e8414fbd49e2ea3" Workload="localhost-k8s-whisker--7bf9c88658--hfwwb-eth0" Jul 10 00:28:34.382656 containerd[1640]: 2025-07-10 00:28:34.325 [INFO][4049] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="5d391391a4ed8f8df19521dedccf29836078c765bd4036bf2e8414fbd49e2ea3" HandleID="k8s-pod-network.5d391391a4ed8f8df19521dedccf29836078c765bd4036bf2e8414fbd49e2ea3" Workload="localhost-k8s-whisker--7bf9c88658--hfwwb-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004f9c0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"whisker-7bf9c88658-hfwwb", "timestamp":"2025-07-10 00:28:34.324140589 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jul 10 00:28:34.382656 containerd[1640]: 2025-07-10 00:28:34.325 [INFO][4049] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 10 00:28:34.382656 containerd[1640]: 2025-07-10 00:28:34.325 [INFO][4049] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 10 00:28:34.382656 containerd[1640]: 2025-07-10 00:28:34.326 [INFO][4049] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jul 10 00:28:34.382656 containerd[1640]: 2025-07-10 00:28:34.344 [INFO][4049] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.5d391391a4ed8f8df19521dedccf29836078c765bd4036bf2e8414fbd49e2ea3" host="localhost" Jul 10 00:28:34.382656 containerd[1640]: 2025-07-10 00:28:34.351 [INFO][4049] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jul 10 00:28:34.382656 containerd[1640]: 2025-07-10 00:28:34.353 [INFO][4049] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jul 10 00:28:34.382656 containerd[1640]: 2025-07-10 00:28:34.354 [INFO][4049] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jul 10 00:28:34.382656 containerd[1640]: 2025-07-10 00:28:34.355 [INFO][4049] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jul 10 00:28:34.382656 containerd[1640]: 2025-07-10 00:28:34.355 [INFO][4049] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.5d391391a4ed8f8df19521dedccf29836078c765bd4036bf2e8414fbd49e2ea3" host="localhost" Jul 10 00:28:34.384239 containerd[1640]: 2025-07-10 00:28:34.356 [INFO][4049] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.5d391391a4ed8f8df19521dedccf29836078c765bd4036bf2e8414fbd49e2ea3 Jul 10 00:28:34.384239 containerd[1640]: 2025-07-10 00:28:34.358 [INFO][4049] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.5d391391a4ed8f8df19521dedccf29836078c765bd4036bf2e8414fbd49e2ea3" host="localhost" Jul 10 00:28:34.384239 containerd[1640]: 2025-07-10 00:28:34.360 [INFO][4049] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.129/26] block=192.168.88.128/26 handle="k8s-pod-network.5d391391a4ed8f8df19521dedccf29836078c765bd4036bf2e8414fbd49e2ea3" host="localhost" Jul 10 00:28:34.384239 containerd[1640]: 2025-07-10 00:28:34.360 [INFO][4049] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.129/26] handle="k8s-pod-network.5d391391a4ed8f8df19521dedccf29836078c765bd4036bf2e8414fbd49e2ea3" host="localhost" Jul 10 00:28:34.384239 containerd[1640]: 2025-07-10 00:28:34.360 [INFO][4049] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 10 00:28:34.384239 containerd[1640]: 2025-07-10 00:28:34.360 [INFO][4049] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.129/26] IPv6=[] ContainerID="5d391391a4ed8f8df19521dedccf29836078c765bd4036bf2e8414fbd49e2ea3" HandleID="k8s-pod-network.5d391391a4ed8f8df19521dedccf29836078c765bd4036bf2e8414fbd49e2ea3" Workload="localhost-k8s-whisker--7bf9c88658--hfwwb-eth0" Jul 10 00:28:34.384342 containerd[1640]: 2025-07-10 00:28:34.362 [INFO][4038] cni-plugin/k8s.go 418: Populated endpoint ContainerID="5d391391a4ed8f8df19521dedccf29836078c765bd4036bf2e8414fbd49e2ea3" Namespace="calico-system" Pod="whisker-7bf9c88658-hfwwb" WorkloadEndpoint="localhost-k8s-whisker--7bf9c88658--hfwwb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-whisker--7bf9c88658--hfwwb-eth0", GenerateName:"whisker-7bf9c88658-", Namespace:"calico-system", SelfLink:"", UID:"b3328a36-298d-4c0b-8e63-7cff7da58b7b", ResourceVersion:"872", Generation:0, CreationTimestamp:time.Date(2025, time.July, 10, 0, 28, 33, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"7bf9c88658", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"whisker-7bf9c88658-hfwwb", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali8fc8c348088", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 10 00:28:34.384342 containerd[1640]: 2025-07-10 00:28:34.362 [INFO][4038] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.129/32] ContainerID="5d391391a4ed8f8df19521dedccf29836078c765bd4036bf2e8414fbd49e2ea3" Namespace="calico-system" Pod="whisker-7bf9c88658-hfwwb" WorkloadEndpoint="localhost-k8s-whisker--7bf9c88658--hfwwb-eth0" Jul 10 00:28:34.384408 containerd[1640]: 2025-07-10 00:28:34.362 [INFO][4038] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali8fc8c348088 ContainerID="5d391391a4ed8f8df19521dedccf29836078c765bd4036bf2e8414fbd49e2ea3" Namespace="calico-system" Pod="whisker-7bf9c88658-hfwwb" WorkloadEndpoint="localhost-k8s-whisker--7bf9c88658--hfwwb-eth0" Jul 10 00:28:34.384408 containerd[1640]: 2025-07-10 00:28:34.372 [INFO][4038] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="5d391391a4ed8f8df19521dedccf29836078c765bd4036bf2e8414fbd49e2ea3" Namespace="calico-system" Pod="whisker-7bf9c88658-hfwwb" WorkloadEndpoint="localhost-k8s-whisker--7bf9c88658--hfwwb-eth0" Jul 10 00:28:34.384438 containerd[1640]: 2025-07-10 00:28:34.372 [INFO][4038] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="5d391391a4ed8f8df19521dedccf29836078c765bd4036bf2e8414fbd49e2ea3" Namespace="calico-system" Pod="whisker-7bf9c88658-hfwwb" WorkloadEndpoint="localhost-k8s-whisker--7bf9c88658--hfwwb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-whisker--7bf9c88658--hfwwb-eth0", GenerateName:"whisker-7bf9c88658-", Namespace:"calico-system", SelfLink:"", UID:"b3328a36-298d-4c0b-8e63-7cff7da58b7b", ResourceVersion:"872", Generation:0, CreationTimestamp:time.Date(2025, time.July, 10, 0, 28, 33, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"7bf9c88658", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"5d391391a4ed8f8df19521dedccf29836078c765bd4036bf2e8414fbd49e2ea3", Pod:"whisker-7bf9c88658-hfwwb", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"cali8fc8c348088", MAC:"22:41:63:6c:a2:ab", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 10 00:28:34.384481 containerd[1640]: 2025-07-10 00:28:34.379 [INFO][4038] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="5d391391a4ed8f8df19521dedccf29836078c765bd4036bf2e8414fbd49e2ea3" Namespace="calico-system" Pod="whisker-7bf9c88658-hfwwb" WorkloadEndpoint="localhost-k8s-whisker--7bf9c88658--hfwwb-eth0" Jul 10 00:28:34.510461 containerd[1640]: time="2025-07-10T00:28:34.510291267Z" level=info msg="connecting to shim 5d391391a4ed8f8df19521dedccf29836078c765bd4036bf2e8414fbd49e2ea3" address="unix:///run/containerd/s/d953a3414be4054b79bd174328bebca8f8e0bb59f4bb11a032c0e8605a7ea7e4" namespace=k8s.io protocol=ttrpc version=3 Jul 10 00:28:34.525326 systemd[1]: Started cri-containerd-5d391391a4ed8f8df19521dedccf29836078c765bd4036bf2e8414fbd49e2ea3.scope - libcontainer container 5d391391a4ed8f8df19521dedccf29836078c765bd4036bf2e8414fbd49e2ea3. Jul 10 00:28:34.532765 systemd-resolved[1528]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jul 10 00:28:34.572233 containerd[1640]: time="2025-07-10T00:28:34.572194303Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-7bf9c88658-hfwwb,Uid:b3328a36-298d-4c0b-8e63-7cff7da58b7b,Namespace:calico-system,Attempt:0,} returns sandbox id \"5d391391a4ed8f8df19521dedccf29836078c765bd4036bf2e8414fbd49e2ea3\"" Jul 10 00:28:34.573312 containerd[1640]: time="2025-07-10T00:28:34.573261939Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.2\"" Jul 10 00:28:34.844270 containerd[1640]: time="2025-07-10T00:28:34.844243830Z" level=info msg="TaskExit event in podsandbox handler container_id:\"7342408a31cb64c6221572622b4fe6b7e010c365050de1a3cf3bc111989059fe\" id:\"00f440a4fa4101c0b189458ed8534b91923e4c3b24d91a7394ac6c271ca187bc\" pid:4179 exit_status:1 exited_at:{seconds:1752107314 nanos:844069294}" Jul 10 00:28:35.111696 systemd-networkd[1524]: vxlan.calico: Link UP Jul 10 00:28:35.111702 systemd-networkd[1524]: vxlan.calico: Gained carrier Jul 10 00:28:35.414847 kubelet[2915]: I0710 00:28:35.414726 2915 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e5edbe5-bd6e-4f88-b5f5-53f7df248f0b" path="/var/lib/kubelet/pods/5e5edbe5-bd6e-4f88-b5f5-53f7df248f0b/volumes" Jul 10 00:28:36.256876 containerd[1640]: time="2025-07-10T00:28:36.256847584Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:36.257299 containerd[1640]: time="2025-07-10T00:28:36.257273847Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.2: active requests=0, bytes read=4661207" Jul 10 00:28:36.258007 containerd[1640]: time="2025-07-10T00:28:36.257612505Z" level=info msg="ImageCreate event name:\"sha256:eb8f512acf9402730da120a7b0d47d3d9d451b56e6e5eb8bad53ab24f926f954\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:36.258635 containerd[1640]: time="2025-07-10T00:28:36.258612576Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker@sha256:31346d4524252a3b0d2a1d289c4985b8402b498b5ce82a12e682096ab7446678\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:36.259183 containerd[1640]: time="2025-07-10T00:28:36.259081629Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker:v3.30.2\" with image id \"sha256:eb8f512acf9402730da120a7b0d47d3d9d451b56e6e5eb8bad53ab24f926f954\", repo tag \"ghcr.io/flatcar/calico/whisker:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/whisker@sha256:31346d4524252a3b0d2a1d289c4985b8402b498b5ce82a12e682096ab7446678\", size \"6153902\" in 1.685803178s" Jul 10 00:28:36.259183 containerd[1640]: time="2025-07-10T00:28:36.259105984Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.2\" returns image reference \"sha256:eb8f512acf9402730da120a7b0d47d3d9d451b56e6e5eb8bad53ab24f926f954\"" Jul 10 00:28:36.280080 containerd[1640]: time="2025-07-10T00:28:36.280033638Z" level=info msg="CreateContainer within sandbox \"5d391391a4ed8f8df19521dedccf29836078c765bd4036bf2e8414fbd49e2ea3\" for container &ContainerMetadata{Name:whisker,Attempt:0,}" Jul 10 00:28:36.284061 containerd[1640]: time="2025-07-10T00:28:36.282831251Z" level=info msg="Container 92d96cdfbd134da79156d738692a3e3b1832e008ea2344a9f910a8042da57b92: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:28:36.289011 containerd[1640]: time="2025-07-10T00:28:36.288966722Z" level=info msg="CreateContainer within sandbox \"5d391391a4ed8f8df19521dedccf29836078c765bd4036bf2e8414fbd49e2ea3\" for &ContainerMetadata{Name:whisker,Attempt:0,} returns container id \"92d96cdfbd134da79156d738692a3e3b1832e008ea2344a9f910a8042da57b92\"" Jul 10 00:28:36.292321 containerd[1640]: time="2025-07-10T00:28:36.289491470Z" level=info msg="StartContainer for \"92d96cdfbd134da79156d738692a3e3b1832e008ea2344a9f910a8042da57b92\"" Jul 10 00:28:36.292321 containerd[1640]: time="2025-07-10T00:28:36.290085184Z" level=info msg="connecting to shim 92d96cdfbd134da79156d738692a3e3b1832e008ea2344a9f910a8042da57b92" address="unix:///run/containerd/s/d953a3414be4054b79bd174328bebca8f8e0bb59f4bb11a032c0e8605a7ea7e4" protocol=ttrpc version=3 Jul 10 00:28:36.308536 systemd[1]: Started cri-containerd-92d96cdfbd134da79156d738692a3e3b1832e008ea2344a9f910a8042da57b92.scope - libcontainer container 92d96cdfbd134da79156d738692a3e3b1832e008ea2344a9f910a8042da57b92. Jul 10 00:28:36.348535 containerd[1640]: time="2025-07-10T00:28:36.348509411Z" level=info msg="StartContainer for \"92d96cdfbd134da79156d738692a3e3b1832e008ea2344a9f910a8042da57b92\" returns successfully" Jul 10 00:28:36.351317 containerd[1640]: time="2025-07-10T00:28:36.351268448Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\"" Jul 10 00:28:36.356274 systemd-networkd[1524]: cali8fc8c348088: Gained IPv6LL Jul 10 00:28:36.996355 systemd-networkd[1524]: vxlan.calico: Gained IPv6LL Jul 10 00:28:38.197924 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3377027865.mount: Deactivated successfully. Jul 10 00:28:38.409793 containerd[1640]: time="2025-07-10T00:28:38.409753902Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:38.413466 containerd[1640]: time="2025-07-10T00:28:38.413426451Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.2: active requests=0, bytes read=33083477" Jul 10 00:28:38.414849 containerd[1640]: time="2025-07-10T00:28:38.414814266Z" level=info msg="ImageCreate event name:\"sha256:6ba7e39edcd8be6d32dfccbfdb65533a727b14a19173515e91607d4259f8ee7f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:38.417176 containerd[1640]: time="2025-07-10T00:28:38.416860585Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend@sha256:fbf7f21f5aba95930803ad7e7dea8b083220854eae72c2a7c51681c09c5614b5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:38.417232 containerd[1640]: time="2025-07-10T00:28:38.417202563Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\" with image id \"sha256:6ba7e39edcd8be6d32dfccbfdb65533a727b14a19173515e91607d4259f8ee7f\", repo tag \"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/whisker-backend@sha256:fbf7f21f5aba95930803ad7e7dea8b083220854eae72c2a7c51681c09c5614b5\", size \"33083307\" in 2.065899997s" Jul 10 00:28:38.417232 containerd[1640]: time="2025-07-10T00:28:38.417222838Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\" returns image reference \"sha256:6ba7e39edcd8be6d32dfccbfdb65533a727b14a19173515e91607d4259f8ee7f\"" Jul 10 00:28:38.419154 containerd[1640]: time="2025-07-10T00:28:38.419130581Z" level=info msg="CreateContainer within sandbox \"5d391391a4ed8f8df19521dedccf29836078c765bd4036bf2e8414fbd49e2ea3\" for container &ContainerMetadata{Name:whisker-backend,Attempt:0,}" Jul 10 00:28:38.425947 containerd[1640]: time="2025-07-10T00:28:38.425910357Z" level=info msg="Container 1329130cd22069932c8da7dba78c4bf0fdc09f0075f9bc01fe54e8dcc39b58bf: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:28:38.439542 containerd[1640]: time="2025-07-10T00:28:38.439516493Z" level=info msg="CreateContainer within sandbox \"5d391391a4ed8f8df19521dedccf29836078c765bd4036bf2e8414fbd49e2ea3\" for &ContainerMetadata{Name:whisker-backend,Attempt:0,} returns container id \"1329130cd22069932c8da7dba78c4bf0fdc09f0075f9bc01fe54e8dcc39b58bf\"" Jul 10 00:28:38.440129 containerd[1640]: time="2025-07-10T00:28:38.440111332Z" level=info msg="StartContainer for \"1329130cd22069932c8da7dba78c4bf0fdc09f0075f9bc01fe54e8dcc39b58bf\"" Jul 10 00:28:38.440958 containerd[1640]: time="2025-07-10T00:28:38.440941064Z" level=info msg="connecting to shim 1329130cd22069932c8da7dba78c4bf0fdc09f0075f9bc01fe54e8dcc39b58bf" address="unix:///run/containerd/s/d953a3414be4054b79bd174328bebca8f8e0bb59f4bb11a032c0e8605a7ea7e4" protocol=ttrpc version=3 Jul 10 00:28:38.462276 systemd[1]: Started cri-containerd-1329130cd22069932c8da7dba78c4bf0fdc09f0075f9bc01fe54e8dcc39b58bf.scope - libcontainer container 1329130cd22069932c8da7dba78c4bf0fdc09f0075f9bc01fe54e8dcc39b58bf. Jul 10 00:28:38.505973 containerd[1640]: time="2025-07-10T00:28:38.505944064Z" level=info msg="StartContainer for \"1329130cd22069932c8da7dba78c4bf0fdc09f0075f9bc01fe54e8dcc39b58bf\" returns successfully" Jul 10 00:28:38.729147 kubelet[2915]: I0710 00:28:38.723343 2915 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/whisker-7bf9c88658-hfwwb" podStartSLOduration=1.878593226 podStartE2EDuration="5.723327986s" podCreationTimestamp="2025-07-10 00:28:33 +0000 UTC" firstStartedPulling="2025-07-10 00:28:34.57294098 +0000 UTC m=+33.317721193" lastFinishedPulling="2025-07-10 00:28:38.41767574 +0000 UTC m=+37.162455953" observedRunningTime="2025-07-10 00:28:38.700904105 +0000 UTC m=+37.445684329" watchObservedRunningTime="2025-07-10 00:28:38.723327986 +0000 UTC m=+37.468108203" Jul 10 00:28:40.394984 containerd[1640]: time="2025-07-10T00:28:40.394827563Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-g5zz7,Uid:783faa36-cde4-4aa6-9139-b51d9b52af99,Namespace:kube-system,Attempt:0,}" Jul 10 00:28:40.394984 containerd[1640]: time="2025-07-10T00:28:40.394872868Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7c75d6c687-wlhzf,Uid:9271cfd0-877e-47f5-b015-c6ecba29e40b,Namespace:calico-apiserver,Attempt:0,}" Jul 10 00:28:40.395399 containerd[1640]: time="2025-07-10T00:28:40.394827505Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-8h9ft,Uid:64915bcf-34a2-483d-ae5f-7e4c4ea7d4d2,Namespace:kube-system,Attempt:0,}" Jul 10 00:28:40.783897 systemd-networkd[1524]: cali59a605a182f: Link UP Jul 10 00:28:40.784271 systemd-networkd[1524]: cali59a605a182f: Gained carrier Jul 10 00:28:40.800819 containerd[1640]: 2025-07-10 00:28:40.461 [INFO][4424] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--7c75d6c687--wlhzf-eth0 calico-apiserver-7c75d6c687- calico-apiserver 9271cfd0-877e-47f5-b015-c6ecba29e40b 802 0 2025-07-10 00:28:15 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:7c75d6c687 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-7c75d6c687-wlhzf eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali59a605a182f [] [] }} ContainerID="6cd8f5071bad46e9132abaa80bad3e69ea82733afa4e3b4d244053a865a88c26" Namespace="calico-apiserver" Pod="calico-apiserver-7c75d6c687-wlhzf" WorkloadEndpoint="localhost-k8s-calico--apiserver--7c75d6c687--wlhzf-" Jul 10 00:28:40.800819 containerd[1640]: 2025-07-10 00:28:40.461 [INFO][4424] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="6cd8f5071bad46e9132abaa80bad3e69ea82733afa4e3b4d244053a865a88c26" Namespace="calico-apiserver" Pod="calico-apiserver-7c75d6c687-wlhzf" WorkloadEndpoint="localhost-k8s-calico--apiserver--7c75d6c687--wlhzf-eth0" Jul 10 00:28:40.800819 containerd[1640]: 2025-07-10 00:28:40.676 [INFO][4456] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="6cd8f5071bad46e9132abaa80bad3e69ea82733afa4e3b4d244053a865a88c26" HandleID="k8s-pod-network.6cd8f5071bad46e9132abaa80bad3e69ea82733afa4e3b4d244053a865a88c26" Workload="localhost-k8s-calico--apiserver--7c75d6c687--wlhzf-eth0" Jul 10 00:28:40.801134 containerd[1640]: 2025-07-10 00:28:40.703 [INFO][4456] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="6cd8f5071bad46e9132abaa80bad3e69ea82733afa4e3b4d244053a865a88c26" HandleID="k8s-pod-network.6cd8f5071bad46e9132abaa80bad3e69ea82733afa4e3b4d244053a865a88c26" Workload="localhost-k8s-calico--apiserver--7c75d6c687--wlhzf-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000326230), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-7c75d6c687-wlhzf", "timestamp":"2025-07-10 00:28:40.676104301 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jul 10 00:28:40.801134 containerd[1640]: 2025-07-10 00:28:40.703 [INFO][4456] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 10 00:28:40.801134 containerd[1640]: 2025-07-10 00:28:40.703 [INFO][4456] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 10 00:28:40.801134 containerd[1640]: 2025-07-10 00:28:40.703 [INFO][4456] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jul 10 00:28:40.801134 containerd[1640]: 2025-07-10 00:28:40.754 [INFO][4456] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.6cd8f5071bad46e9132abaa80bad3e69ea82733afa4e3b4d244053a865a88c26" host="localhost" Jul 10 00:28:40.801134 containerd[1640]: 2025-07-10 00:28:40.761 [INFO][4456] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jul 10 00:28:40.801134 containerd[1640]: 2025-07-10 00:28:40.765 [INFO][4456] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jul 10 00:28:40.801134 containerd[1640]: 2025-07-10 00:28:40.766 [INFO][4456] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jul 10 00:28:40.801134 containerd[1640]: 2025-07-10 00:28:40.767 [INFO][4456] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jul 10 00:28:40.801134 containerd[1640]: 2025-07-10 00:28:40.767 [INFO][4456] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.6cd8f5071bad46e9132abaa80bad3e69ea82733afa4e3b4d244053a865a88c26" host="localhost" Jul 10 00:28:40.802027 containerd[1640]: 2025-07-10 00:28:40.768 [INFO][4456] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.6cd8f5071bad46e9132abaa80bad3e69ea82733afa4e3b4d244053a865a88c26 Jul 10 00:28:40.802027 containerd[1640]: 2025-07-10 00:28:40.771 [INFO][4456] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.6cd8f5071bad46e9132abaa80bad3e69ea82733afa4e3b4d244053a865a88c26" host="localhost" Jul 10 00:28:40.802027 containerd[1640]: 2025-07-10 00:28:40.774 [INFO][4456] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.130/26] block=192.168.88.128/26 handle="k8s-pod-network.6cd8f5071bad46e9132abaa80bad3e69ea82733afa4e3b4d244053a865a88c26" host="localhost" Jul 10 00:28:40.802027 containerd[1640]: 2025-07-10 00:28:40.774 [INFO][4456] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.130/26] handle="k8s-pod-network.6cd8f5071bad46e9132abaa80bad3e69ea82733afa4e3b4d244053a865a88c26" host="localhost" Jul 10 00:28:40.802027 containerd[1640]: 2025-07-10 00:28:40.774 [INFO][4456] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 10 00:28:40.802027 containerd[1640]: 2025-07-10 00:28:40.774 [INFO][4456] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.130/26] IPv6=[] ContainerID="6cd8f5071bad46e9132abaa80bad3e69ea82733afa4e3b4d244053a865a88c26" HandleID="k8s-pod-network.6cd8f5071bad46e9132abaa80bad3e69ea82733afa4e3b4d244053a865a88c26" Workload="localhost-k8s-calico--apiserver--7c75d6c687--wlhzf-eth0" Jul 10 00:28:40.802127 containerd[1640]: 2025-07-10 00:28:40.776 [INFO][4424] cni-plugin/k8s.go 418: Populated endpoint ContainerID="6cd8f5071bad46e9132abaa80bad3e69ea82733afa4e3b4d244053a865a88c26" Namespace="calico-apiserver" Pod="calico-apiserver-7c75d6c687-wlhzf" WorkloadEndpoint="localhost-k8s-calico--apiserver--7c75d6c687--wlhzf-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--7c75d6c687--wlhzf-eth0", GenerateName:"calico-apiserver-7c75d6c687-", Namespace:"calico-apiserver", SelfLink:"", UID:"9271cfd0-877e-47f5-b015-c6ecba29e40b", ResourceVersion:"802", Generation:0, CreationTimestamp:time.Date(2025, time.July, 10, 0, 28, 15, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7c75d6c687", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-7c75d6c687-wlhzf", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali59a605a182f", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 10 00:28:40.802181 containerd[1640]: 2025-07-10 00:28:40.777 [INFO][4424] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.130/32] ContainerID="6cd8f5071bad46e9132abaa80bad3e69ea82733afa4e3b4d244053a865a88c26" Namespace="calico-apiserver" Pod="calico-apiserver-7c75d6c687-wlhzf" WorkloadEndpoint="localhost-k8s-calico--apiserver--7c75d6c687--wlhzf-eth0" Jul 10 00:28:40.802181 containerd[1640]: 2025-07-10 00:28:40.777 [INFO][4424] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali59a605a182f ContainerID="6cd8f5071bad46e9132abaa80bad3e69ea82733afa4e3b4d244053a865a88c26" Namespace="calico-apiserver" Pod="calico-apiserver-7c75d6c687-wlhzf" WorkloadEndpoint="localhost-k8s-calico--apiserver--7c75d6c687--wlhzf-eth0" Jul 10 00:28:40.802181 containerd[1640]: 2025-07-10 00:28:40.785 [INFO][4424] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="6cd8f5071bad46e9132abaa80bad3e69ea82733afa4e3b4d244053a865a88c26" Namespace="calico-apiserver" Pod="calico-apiserver-7c75d6c687-wlhzf" WorkloadEndpoint="localhost-k8s-calico--apiserver--7c75d6c687--wlhzf-eth0" Jul 10 00:28:40.802246 containerd[1640]: 2025-07-10 00:28:40.785 [INFO][4424] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="6cd8f5071bad46e9132abaa80bad3e69ea82733afa4e3b4d244053a865a88c26" Namespace="calico-apiserver" Pod="calico-apiserver-7c75d6c687-wlhzf" WorkloadEndpoint="localhost-k8s-calico--apiserver--7c75d6c687--wlhzf-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--7c75d6c687--wlhzf-eth0", GenerateName:"calico-apiserver-7c75d6c687-", Namespace:"calico-apiserver", SelfLink:"", UID:"9271cfd0-877e-47f5-b015-c6ecba29e40b", ResourceVersion:"802", Generation:0, CreationTimestamp:time.Date(2025, time.July, 10, 0, 28, 15, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7c75d6c687", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"6cd8f5071bad46e9132abaa80bad3e69ea82733afa4e3b4d244053a865a88c26", Pod:"calico-apiserver-7c75d6c687-wlhzf", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali59a605a182f", MAC:"ea:ab:5c:d0:32:53", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 10 00:28:40.803491 containerd[1640]: 2025-07-10 00:28:40.795 [INFO][4424] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="6cd8f5071bad46e9132abaa80bad3e69ea82733afa4e3b4d244053a865a88c26" Namespace="calico-apiserver" Pod="calico-apiserver-7c75d6c687-wlhzf" WorkloadEndpoint="localhost-k8s-calico--apiserver--7c75d6c687--wlhzf-eth0" Jul 10 00:28:40.841436 containerd[1640]: time="2025-07-10T00:28:40.841408482Z" level=info msg="connecting to shim 6cd8f5071bad46e9132abaa80bad3e69ea82733afa4e3b4d244053a865a88c26" address="unix:///run/containerd/s/404b43027837d720f6c87db0b40705472b8e94d4d358c43d18e5089c28c4cb1f" namespace=k8s.io protocol=ttrpc version=3 Jul 10 00:28:40.861301 systemd[1]: Started cri-containerd-6cd8f5071bad46e9132abaa80bad3e69ea82733afa4e3b4d244053a865a88c26.scope - libcontainer container 6cd8f5071bad46e9132abaa80bad3e69ea82733afa4e3b4d244053a865a88c26. Jul 10 00:28:40.875881 systemd-resolved[1528]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jul 10 00:28:40.885839 systemd-networkd[1524]: cali86ba49b9c17: Link UP Jul 10 00:28:40.890391 systemd-networkd[1524]: cali86ba49b9c17: Gained carrier Jul 10 00:28:40.905519 containerd[1640]: 2025-07-10 00:28:40.461 [INFO][4432] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--7c65d6cfc9--g5zz7-eth0 coredns-7c65d6cfc9- kube-system 783faa36-cde4-4aa6-9139-b51d9b52af99 798 0 2025-07-10 00:28:07 +0000 UTC map[k8s-app:kube-dns pod-template-hash:7c65d6cfc9 projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-7c65d6cfc9-g5zz7 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali86ba49b9c17 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="788f636fa0d3cd779a0a6e7f05844516cd5429d8dbd96516da4359085a33e220" Namespace="kube-system" Pod="coredns-7c65d6cfc9-g5zz7" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--g5zz7-" Jul 10 00:28:40.905519 containerd[1640]: 2025-07-10 00:28:40.461 [INFO][4432] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="788f636fa0d3cd779a0a6e7f05844516cd5429d8dbd96516da4359085a33e220" Namespace="kube-system" Pod="coredns-7c65d6cfc9-g5zz7" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--g5zz7-eth0" Jul 10 00:28:40.905519 containerd[1640]: 2025-07-10 00:28:40.676 [INFO][4460] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="788f636fa0d3cd779a0a6e7f05844516cd5429d8dbd96516da4359085a33e220" HandleID="k8s-pod-network.788f636fa0d3cd779a0a6e7f05844516cd5429d8dbd96516da4359085a33e220" Workload="localhost-k8s-coredns--7c65d6cfc9--g5zz7-eth0" Jul 10 00:28:40.905706 containerd[1640]: 2025-07-10 00:28:40.703 [INFO][4460] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="788f636fa0d3cd779a0a6e7f05844516cd5429d8dbd96516da4359085a33e220" HandleID="k8s-pod-network.788f636fa0d3cd779a0a6e7f05844516cd5429d8dbd96516da4359085a33e220" Workload="localhost-k8s-coredns--7c65d6cfc9--g5zz7-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002d5850), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-7c65d6cfc9-g5zz7", "timestamp":"2025-07-10 00:28:40.676107201 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jul 10 00:28:40.905706 containerd[1640]: 2025-07-10 00:28:40.703 [INFO][4460] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 10 00:28:40.905706 containerd[1640]: 2025-07-10 00:28:40.774 [INFO][4460] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 10 00:28:40.905706 containerd[1640]: 2025-07-10 00:28:40.775 [INFO][4460] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jul 10 00:28:40.905706 containerd[1640]: 2025-07-10 00:28:40.854 [INFO][4460] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.788f636fa0d3cd779a0a6e7f05844516cd5429d8dbd96516da4359085a33e220" host="localhost" Jul 10 00:28:40.905706 containerd[1640]: 2025-07-10 00:28:40.863 [INFO][4460] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jul 10 00:28:40.905706 containerd[1640]: 2025-07-10 00:28:40.868 [INFO][4460] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jul 10 00:28:40.905706 containerd[1640]: 2025-07-10 00:28:40.870 [INFO][4460] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jul 10 00:28:40.905706 containerd[1640]: 2025-07-10 00:28:40.871 [INFO][4460] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jul 10 00:28:40.905706 containerd[1640]: 2025-07-10 00:28:40.871 [INFO][4460] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.788f636fa0d3cd779a0a6e7f05844516cd5429d8dbd96516da4359085a33e220" host="localhost" Jul 10 00:28:40.906115 containerd[1640]: 2025-07-10 00:28:40.872 [INFO][4460] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.788f636fa0d3cd779a0a6e7f05844516cd5429d8dbd96516da4359085a33e220 Jul 10 00:28:40.906115 containerd[1640]: 2025-07-10 00:28:40.875 [INFO][4460] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.788f636fa0d3cd779a0a6e7f05844516cd5429d8dbd96516da4359085a33e220" host="localhost" Jul 10 00:28:40.906115 containerd[1640]: 2025-07-10 00:28:40.879 [INFO][4460] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.131/26] block=192.168.88.128/26 handle="k8s-pod-network.788f636fa0d3cd779a0a6e7f05844516cd5429d8dbd96516da4359085a33e220" host="localhost" Jul 10 00:28:40.906115 containerd[1640]: 2025-07-10 00:28:40.879 [INFO][4460] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.131/26] handle="k8s-pod-network.788f636fa0d3cd779a0a6e7f05844516cd5429d8dbd96516da4359085a33e220" host="localhost" Jul 10 00:28:40.906115 containerd[1640]: 2025-07-10 00:28:40.879 [INFO][4460] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 10 00:28:40.906115 containerd[1640]: 2025-07-10 00:28:40.879 [INFO][4460] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.131/26] IPv6=[] ContainerID="788f636fa0d3cd779a0a6e7f05844516cd5429d8dbd96516da4359085a33e220" HandleID="k8s-pod-network.788f636fa0d3cd779a0a6e7f05844516cd5429d8dbd96516da4359085a33e220" Workload="localhost-k8s-coredns--7c65d6cfc9--g5zz7-eth0" Jul 10 00:28:40.906363 containerd[1640]: 2025-07-10 00:28:40.882 [INFO][4432] cni-plugin/k8s.go 418: Populated endpoint ContainerID="788f636fa0d3cd779a0a6e7f05844516cd5429d8dbd96516da4359085a33e220" Namespace="kube-system" Pod="coredns-7c65d6cfc9-g5zz7" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--g5zz7-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7c65d6cfc9--g5zz7-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"783faa36-cde4-4aa6-9139-b51d9b52af99", ResourceVersion:"798", Generation:0, CreationTimestamp:time.Date(2025, time.July, 10, 0, 28, 7, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-7c65d6cfc9-g5zz7", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali86ba49b9c17", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 10 00:28:40.906816 containerd[1640]: 2025-07-10 00:28:40.882 [INFO][4432] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.131/32] ContainerID="788f636fa0d3cd779a0a6e7f05844516cd5429d8dbd96516da4359085a33e220" Namespace="kube-system" Pod="coredns-7c65d6cfc9-g5zz7" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--g5zz7-eth0" Jul 10 00:28:40.906816 containerd[1640]: 2025-07-10 00:28:40.882 [INFO][4432] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali86ba49b9c17 ContainerID="788f636fa0d3cd779a0a6e7f05844516cd5429d8dbd96516da4359085a33e220" Namespace="kube-system" Pod="coredns-7c65d6cfc9-g5zz7" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--g5zz7-eth0" Jul 10 00:28:40.906816 containerd[1640]: 2025-07-10 00:28:40.888 [INFO][4432] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="788f636fa0d3cd779a0a6e7f05844516cd5429d8dbd96516da4359085a33e220" Namespace="kube-system" Pod="coredns-7c65d6cfc9-g5zz7" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--g5zz7-eth0" Jul 10 00:28:40.907496 containerd[1640]: 2025-07-10 00:28:40.889 [INFO][4432] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="788f636fa0d3cd779a0a6e7f05844516cd5429d8dbd96516da4359085a33e220" Namespace="kube-system" Pod="coredns-7c65d6cfc9-g5zz7" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--g5zz7-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7c65d6cfc9--g5zz7-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"783faa36-cde4-4aa6-9139-b51d9b52af99", ResourceVersion:"798", Generation:0, CreationTimestamp:time.Date(2025, time.July, 10, 0, 28, 7, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"788f636fa0d3cd779a0a6e7f05844516cd5429d8dbd96516da4359085a33e220", Pod:"coredns-7c65d6cfc9-g5zz7", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali86ba49b9c17", MAC:"8a:b6:02:a4:04:93", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 10 00:28:40.907496 containerd[1640]: 2025-07-10 00:28:40.900 [INFO][4432] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="788f636fa0d3cd779a0a6e7f05844516cd5429d8dbd96516da4359085a33e220" Namespace="kube-system" Pod="coredns-7c65d6cfc9-g5zz7" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--g5zz7-eth0" Jul 10 00:28:40.925115 containerd[1640]: time="2025-07-10T00:28:40.925090660Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7c75d6c687-wlhzf,Uid:9271cfd0-877e-47f5-b015-c6ecba29e40b,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"6cd8f5071bad46e9132abaa80bad3e69ea82733afa4e3b4d244053a865a88c26\"" Jul 10 00:28:40.926825 containerd[1640]: time="2025-07-10T00:28:40.926813427Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.2\"" Jul 10 00:28:40.928319 containerd[1640]: time="2025-07-10T00:28:40.928209721Z" level=info msg="connecting to shim 788f636fa0d3cd779a0a6e7f05844516cd5429d8dbd96516da4359085a33e220" address="unix:///run/containerd/s/520b024473eb82a1bc949ce66a56a89bd2feeadf5496cdc2cb2785354b177464" namespace=k8s.io protocol=ttrpc version=3 Jul 10 00:28:40.946294 systemd[1]: Started cri-containerd-788f636fa0d3cd779a0a6e7f05844516cd5429d8dbd96516da4359085a33e220.scope - libcontainer container 788f636fa0d3cd779a0a6e7f05844516cd5429d8dbd96516da4359085a33e220. Jul 10 00:28:40.956352 systemd-resolved[1528]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jul 10 00:28:40.980914 systemd-networkd[1524]: cali065cf7a8b4e: Link UP Jul 10 00:28:40.981844 systemd-networkd[1524]: cali065cf7a8b4e: Gained carrier Jul 10 00:28:41.001179 containerd[1640]: 2025-07-10 00:28:40.461 [INFO][4423] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--7c65d6cfc9--8h9ft-eth0 coredns-7c65d6cfc9- kube-system 64915bcf-34a2-483d-ae5f-7e4c4ea7d4d2 799 0 2025-07-10 00:28:07 +0000 UTC map[k8s-app:kube-dns pod-template-hash:7c65d6cfc9 projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-7c65d6cfc9-8h9ft eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali065cf7a8b4e [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="a8cc3889bf4caa9e7a98165f9f0d129fdab2a2b7fb5d587fa84d6fa5792623a2" Namespace="kube-system" Pod="coredns-7c65d6cfc9-8h9ft" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--8h9ft-" Jul 10 00:28:41.001179 containerd[1640]: 2025-07-10 00:28:40.461 [INFO][4423] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="a8cc3889bf4caa9e7a98165f9f0d129fdab2a2b7fb5d587fa84d6fa5792623a2" Namespace="kube-system" Pod="coredns-7c65d6cfc9-8h9ft" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--8h9ft-eth0" Jul 10 00:28:41.001179 containerd[1640]: 2025-07-10 00:28:40.676 [INFO][4459] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="a8cc3889bf4caa9e7a98165f9f0d129fdab2a2b7fb5d587fa84d6fa5792623a2" HandleID="k8s-pod-network.a8cc3889bf4caa9e7a98165f9f0d129fdab2a2b7fb5d587fa84d6fa5792623a2" Workload="localhost-k8s-coredns--7c65d6cfc9--8h9ft-eth0" Jul 10 00:28:41.001179 containerd[1640]: 2025-07-10 00:28:40.703 [INFO][4459] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="a8cc3889bf4caa9e7a98165f9f0d129fdab2a2b7fb5d587fa84d6fa5792623a2" HandleID="k8s-pod-network.a8cc3889bf4caa9e7a98165f9f0d129fdab2a2b7fb5d587fa84d6fa5792623a2" Workload="localhost-k8s-coredns--7c65d6cfc9--8h9ft-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000402200), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-7c65d6cfc9-8h9ft", "timestamp":"2025-07-10 00:28:40.676470005 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jul 10 00:28:41.001179 containerd[1640]: 2025-07-10 00:28:40.703 [INFO][4459] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 10 00:28:41.001179 containerd[1640]: 2025-07-10 00:28:40.879 [INFO][4459] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 10 00:28:41.001179 containerd[1640]: 2025-07-10 00:28:40.879 [INFO][4459] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jul 10 00:28:41.001179 containerd[1640]: 2025-07-10 00:28:40.955 [INFO][4459] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.a8cc3889bf4caa9e7a98165f9f0d129fdab2a2b7fb5d587fa84d6fa5792623a2" host="localhost" Jul 10 00:28:41.001179 containerd[1640]: 2025-07-10 00:28:40.963 [INFO][4459] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jul 10 00:28:41.001179 containerd[1640]: 2025-07-10 00:28:40.966 [INFO][4459] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jul 10 00:28:41.001179 containerd[1640]: 2025-07-10 00:28:40.967 [INFO][4459] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jul 10 00:28:41.001179 containerd[1640]: 2025-07-10 00:28:40.968 [INFO][4459] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jul 10 00:28:41.001179 containerd[1640]: 2025-07-10 00:28:40.968 [INFO][4459] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.a8cc3889bf4caa9e7a98165f9f0d129fdab2a2b7fb5d587fa84d6fa5792623a2" host="localhost" Jul 10 00:28:41.001179 containerd[1640]: 2025-07-10 00:28:40.969 [INFO][4459] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.a8cc3889bf4caa9e7a98165f9f0d129fdab2a2b7fb5d587fa84d6fa5792623a2 Jul 10 00:28:41.001179 containerd[1640]: 2025-07-10 00:28:40.970 [INFO][4459] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.a8cc3889bf4caa9e7a98165f9f0d129fdab2a2b7fb5d587fa84d6fa5792623a2" host="localhost" Jul 10 00:28:41.001179 containerd[1640]: 2025-07-10 00:28:40.974 [INFO][4459] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.132/26] block=192.168.88.128/26 handle="k8s-pod-network.a8cc3889bf4caa9e7a98165f9f0d129fdab2a2b7fb5d587fa84d6fa5792623a2" host="localhost" Jul 10 00:28:41.001179 containerd[1640]: 2025-07-10 00:28:40.975 [INFO][4459] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.132/26] handle="k8s-pod-network.a8cc3889bf4caa9e7a98165f9f0d129fdab2a2b7fb5d587fa84d6fa5792623a2" host="localhost" Jul 10 00:28:41.001179 containerd[1640]: 2025-07-10 00:28:40.975 [INFO][4459] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 10 00:28:41.001179 containerd[1640]: 2025-07-10 00:28:40.975 [INFO][4459] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.132/26] IPv6=[] ContainerID="a8cc3889bf4caa9e7a98165f9f0d129fdab2a2b7fb5d587fa84d6fa5792623a2" HandleID="k8s-pod-network.a8cc3889bf4caa9e7a98165f9f0d129fdab2a2b7fb5d587fa84d6fa5792623a2" Workload="localhost-k8s-coredns--7c65d6cfc9--8h9ft-eth0" Jul 10 00:28:41.001591 containerd[1640]: 2025-07-10 00:28:40.978 [INFO][4423] cni-plugin/k8s.go 418: Populated endpoint ContainerID="a8cc3889bf4caa9e7a98165f9f0d129fdab2a2b7fb5d587fa84d6fa5792623a2" Namespace="kube-system" Pod="coredns-7c65d6cfc9-8h9ft" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--8h9ft-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7c65d6cfc9--8h9ft-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"64915bcf-34a2-483d-ae5f-7e4c4ea7d4d2", ResourceVersion:"799", Generation:0, CreationTimestamp:time.Date(2025, time.July, 10, 0, 28, 7, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-7c65d6cfc9-8h9ft", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali065cf7a8b4e", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 10 00:28:41.001591 containerd[1640]: 2025-07-10 00:28:40.978 [INFO][4423] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.132/32] ContainerID="a8cc3889bf4caa9e7a98165f9f0d129fdab2a2b7fb5d587fa84d6fa5792623a2" Namespace="kube-system" Pod="coredns-7c65d6cfc9-8h9ft" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--8h9ft-eth0" Jul 10 00:28:41.001591 containerd[1640]: 2025-07-10 00:28:40.978 [INFO][4423] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali065cf7a8b4e ContainerID="a8cc3889bf4caa9e7a98165f9f0d129fdab2a2b7fb5d587fa84d6fa5792623a2" Namespace="kube-system" Pod="coredns-7c65d6cfc9-8h9ft" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--8h9ft-eth0" Jul 10 00:28:41.001591 containerd[1640]: 2025-07-10 00:28:40.981 [INFO][4423] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="a8cc3889bf4caa9e7a98165f9f0d129fdab2a2b7fb5d587fa84d6fa5792623a2" Namespace="kube-system" Pod="coredns-7c65d6cfc9-8h9ft" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--8h9ft-eth0" Jul 10 00:28:41.001591 containerd[1640]: 2025-07-10 00:28:40.982 [INFO][4423] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="a8cc3889bf4caa9e7a98165f9f0d129fdab2a2b7fb5d587fa84d6fa5792623a2" Namespace="kube-system" Pod="coredns-7c65d6cfc9-8h9ft" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--8h9ft-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--7c65d6cfc9--8h9ft-eth0", GenerateName:"coredns-7c65d6cfc9-", Namespace:"kube-system", SelfLink:"", UID:"64915bcf-34a2-483d-ae5f-7e4c4ea7d4d2", ResourceVersion:"799", Generation:0, CreationTimestamp:time.Date(2025, time.July, 10, 0, 28, 7, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7c65d6cfc9", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"a8cc3889bf4caa9e7a98165f9f0d129fdab2a2b7fb5d587fa84d6fa5792623a2", Pod:"coredns-7c65d6cfc9-8h9ft", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali065cf7a8b4e", MAC:"42:72:66:59:5d:8a", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 10 00:28:41.001591 containerd[1640]: 2025-07-10 00:28:40.992 [INFO][4423] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="a8cc3889bf4caa9e7a98165f9f0d129fdab2a2b7fb5d587fa84d6fa5792623a2" Namespace="kube-system" Pod="coredns-7c65d6cfc9-8h9ft" WorkloadEndpoint="localhost-k8s-coredns--7c65d6cfc9--8h9ft-eth0" Jul 10 00:28:41.005140 containerd[1640]: time="2025-07-10T00:28:41.005106015Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-g5zz7,Uid:783faa36-cde4-4aa6-9139-b51d9b52af99,Namespace:kube-system,Attempt:0,} returns sandbox id \"788f636fa0d3cd779a0a6e7f05844516cd5429d8dbd96516da4359085a33e220\"" Jul 10 00:28:41.016833 containerd[1640]: time="2025-07-10T00:28:41.016755937Z" level=info msg="connecting to shim a8cc3889bf4caa9e7a98165f9f0d129fdab2a2b7fb5d587fa84d6fa5792623a2" address="unix:///run/containerd/s/6c427cdd960f9bee627c5ed25b79e3026961b06e14b781304817bf6a733cd6bd" namespace=k8s.io protocol=ttrpc version=3 Jul 10 00:28:41.020925 containerd[1640]: time="2025-07-10T00:28:41.020886225Z" level=info msg="CreateContainer within sandbox \"788f636fa0d3cd779a0a6e7f05844516cd5429d8dbd96516da4359085a33e220\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jul 10 00:28:41.036333 systemd[1]: Started cri-containerd-a8cc3889bf4caa9e7a98165f9f0d129fdab2a2b7fb5d587fa84d6fa5792623a2.scope - libcontainer container a8cc3889bf4caa9e7a98165f9f0d129fdab2a2b7fb5d587fa84d6fa5792623a2. Jul 10 00:28:41.044800 systemd-resolved[1528]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jul 10 00:28:41.047248 containerd[1640]: time="2025-07-10T00:28:41.047230565Z" level=info msg="Container debbb8ef3b6d0a1ac96f1be0fc8410ae0918602cc6407f7ef9d8f751dfa5eb70: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:28:41.049710 containerd[1640]: time="2025-07-10T00:28:41.049681420Z" level=info msg="CreateContainer within sandbox \"788f636fa0d3cd779a0a6e7f05844516cd5429d8dbd96516da4359085a33e220\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"debbb8ef3b6d0a1ac96f1be0fc8410ae0918602cc6407f7ef9d8f751dfa5eb70\"" Jul 10 00:28:41.050336 containerd[1640]: time="2025-07-10T00:28:41.050230933Z" level=info msg="StartContainer for \"debbb8ef3b6d0a1ac96f1be0fc8410ae0918602cc6407f7ef9d8f751dfa5eb70\"" Jul 10 00:28:41.051394 containerd[1640]: time="2025-07-10T00:28:41.051297683Z" level=info msg="connecting to shim debbb8ef3b6d0a1ac96f1be0fc8410ae0918602cc6407f7ef9d8f751dfa5eb70" address="unix:///run/containerd/s/520b024473eb82a1bc949ce66a56a89bd2feeadf5496cdc2cb2785354b177464" protocol=ttrpc version=3 Jul 10 00:28:41.065249 systemd[1]: Started cri-containerd-debbb8ef3b6d0a1ac96f1be0fc8410ae0918602cc6407f7ef9d8f751dfa5eb70.scope - libcontainer container debbb8ef3b6d0a1ac96f1be0fc8410ae0918602cc6407f7ef9d8f751dfa5eb70. Jul 10 00:28:41.079433 containerd[1640]: time="2025-07-10T00:28:41.079314850Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-8h9ft,Uid:64915bcf-34a2-483d-ae5f-7e4c4ea7d4d2,Namespace:kube-system,Attempt:0,} returns sandbox id \"a8cc3889bf4caa9e7a98165f9f0d129fdab2a2b7fb5d587fa84d6fa5792623a2\"" Jul 10 00:28:41.082117 containerd[1640]: time="2025-07-10T00:28:41.082098333Z" level=info msg="CreateContainer within sandbox \"a8cc3889bf4caa9e7a98165f9f0d129fdab2a2b7fb5d587fa84d6fa5792623a2\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jul 10 00:28:41.134819 containerd[1640]: time="2025-07-10T00:28:41.134601190Z" level=info msg="Container c0c685e0501de854a2aeb174dd9ba94c22deafc1d16af950f58ab6c269f09d72: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:28:41.135188 containerd[1640]: time="2025-07-10T00:28:41.135142080Z" level=info msg="StartContainer for \"debbb8ef3b6d0a1ac96f1be0fc8410ae0918602cc6407f7ef9d8f751dfa5eb70\" returns successfully" Jul 10 00:28:41.141605 containerd[1640]: time="2025-07-10T00:28:41.141564787Z" level=info msg="CreateContainer within sandbox \"a8cc3889bf4caa9e7a98165f9f0d129fdab2a2b7fb5d587fa84d6fa5792623a2\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"c0c685e0501de854a2aeb174dd9ba94c22deafc1d16af950f58ab6c269f09d72\"" Jul 10 00:28:41.142182 containerd[1640]: time="2025-07-10T00:28:41.141882145Z" level=info msg="StartContainer for \"c0c685e0501de854a2aeb174dd9ba94c22deafc1d16af950f58ab6c269f09d72\"" Jul 10 00:28:41.143525 containerd[1640]: time="2025-07-10T00:28:41.143470260Z" level=info msg="connecting to shim c0c685e0501de854a2aeb174dd9ba94c22deafc1d16af950f58ab6c269f09d72" address="unix:///run/containerd/s/6c427cdd960f9bee627c5ed25b79e3026961b06e14b781304817bf6a733cd6bd" protocol=ttrpc version=3 Jul 10 00:28:41.160273 systemd[1]: Started cri-containerd-c0c685e0501de854a2aeb174dd9ba94c22deafc1d16af950f58ab6c269f09d72.scope - libcontainer container c0c685e0501de854a2aeb174dd9ba94c22deafc1d16af950f58ab6c269f09d72. Jul 10 00:28:41.180352 containerd[1640]: time="2025-07-10T00:28:41.180327214Z" level=info msg="StartContainer for \"c0c685e0501de854a2aeb174dd9ba94c22deafc1d16af950f58ab6c269f09d72\" returns successfully" Jul 10 00:28:41.435990 containerd[1640]: time="2025-07-10T00:28:41.435117918Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7c75d6c687-8jqvg,Uid:082d696e-d4fa-497e-ab24-928a088c4f21,Namespace:calico-apiserver,Attempt:0,}" Jul 10 00:28:41.435990 containerd[1640]: time="2025-07-10T00:28:41.435758158Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-pzv7w,Uid:2587365c-21ee-44fd-91df-0ab3a24762fc,Namespace:calico-system,Attempt:0,}" Jul 10 00:28:41.435990 containerd[1640]: time="2025-07-10T00:28:41.435786177Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-84bd499cb-6t459,Uid:0a9d7ca6-fb8c-4239-ae97-5a87dba79acf,Namespace:calico-system,Attempt:0,}" Jul 10 00:28:41.593874 systemd-networkd[1524]: calib45b0a49d9b: Link UP Jul 10 00:28:41.594496 systemd-networkd[1524]: calib45b0a49d9b: Gained carrier Jul 10 00:28:41.605135 containerd[1640]: 2025-07-10 00:28:41.508 [INFO][4722] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--kube--controllers--84bd499cb--6t459-eth0 calico-kube-controllers-84bd499cb- calico-system 0a9d7ca6-fb8c-4239-ae97-5a87dba79acf 795 0 2025-07-10 00:28:18 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:84bd499cb projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s localhost calico-kube-controllers-84bd499cb-6t459 eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] calib45b0a49d9b [] [] }} ContainerID="53803ab3ac7d3053c9196e001fef8e60712767b15aed9d310d5fc83900f11072" Namespace="calico-system" Pod="calico-kube-controllers-84bd499cb-6t459" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--84bd499cb--6t459-" Jul 10 00:28:41.605135 containerd[1640]: 2025-07-10 00:28:41.508 [INFO][4722] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="53803ab3ac7d3053c9196e001fef8e60712767b15aed9d310d5fc83900f11072" Namespace="calico-system" Pod="calico-kube-controllers-84bd499cb-6t459" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--84bd499cb--6t459-eth0" Jul 10 00:28:41.605135 containerd[1640]: 2025-07-10 00:28:41.564 [INFO][4753] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="53803ab3ac7d3053c9196e001fef8e60712767b15aed9d310d5fc83900f11072" HandleID="k8s-pod-network.53803ab3ac7d3053c9196e001fef8e60712767b15aed9d310d5fc83900f11072" Workload="localhost-k8s-calico--kube--controllers--84bd499cb--6t459-eth0" Jul 10 00:28:41.605135 containerd[1640]: 2025-07-10 00:28:41.565 [INFO][4753] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="53803ab3ac7d3053c9196e001fef8e60712767b15aed9d310d5fc83900f11072" HandleID="k8s-pod-network.53803ab3ac7d3053c9196e001fef8e60712767b15aed9d310d5fc83900f11072" Workload="localhost-k8s-calico--kube--controllers--84bd499cb--6t459-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002cd640), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"calico-kube-controllers-84bd499cb-6t459", "timestamp":"2025-07-10 00:28:41.564919864 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jul 10 00:28:41.605135 containerd[1640]: 2025-07-10 00:28:41.565 [INFO][4753] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 10 00:28:41.605135 containerd[1640]: 2025-07-10 00:28:41.565 [INFO][4753] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 10 00:28:41.605135 containerd[1640]: 2025-07-10 00:28:41.568 [INFO][4753] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jul 10 00:28:41.605135 containerd[1640]: 2025-07-10 00:28:41.573 [INFO][4753] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.53803ab3ac7d3053c9196e001fef8e60712767b15aed9d310d5fc83900f11072" host="localhost" Jul 10 00:28:41.605135 containerd[1640]: 2025-07-10 00:28:41.575 [INFO][4753] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jul 10 00:28:41.605135 containerd[1640]: 2025-07-10 00:28:41.577 [INFO][4753] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jul 10 00:28:41.605135 containerd[1640]: 2025-07-10 00:28:41.578 [INFO][4753] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jul 10 00:28:41.605135 containerd[1640]: 2025-07-10 00:28:41.579 [INFO][4753] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jul 10 00:28:41.605135 containerd[1640]: 2025-07-10 00:28:41.579 [INFO][4753] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.53803ab3ac7d3053c9196e001fef8e60712767b15aed9d310d5fc83900f11072" host="localhost" Jul 10 00:28:41.605135 containerd[1640]: 2025-07-10 00:28:41.580 [INFO][4753] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.53803ab3ac7d3053c9196e001fef8e60712767b15aed9d310d5fc83900f11072 Jul 10 00:28:41.605135 containerd[1640]: 2025-07-10 00:28:41.582 [INFO][4753] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.53803ab3ac7d3053c9196e001fef8e60712767b15aed9d310d5fc83900f11072" host="localhost" Jul 10 00:28:41.605135 containerd[1640]: 2025-07-10 00:28:41.588 [INFO][4753] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.133/26] block=192.168.88.128/26 handle="k8s-pod-network.53803ab3ac7d3053c9196e001fef8e60712767b15aed9d310d5fc83900f11072" host="localhost" Jul 10 00:28:41.605135 containerd[1640]: 2025-07-10 00:28:41.588 [INFO][4753] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.133/26] handle="k8s-pod-network.53803ab3ac7d3053c9196e001fef8e60712767b15aed9d310d5fc83900f11072" host="localhost" Jul 10 00:28:41.605135 containerd[1640]: 2025-07-10 00:28:41.588 [INFO][4753] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 10 00:28:41.605135 containerd[1640]: 2025-07-10 00:28:41.588 [INFO][4753] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.133/26] IPv6=[] ContainerID="53803ab3ac7d3053c9196e001fef8e60712767b15aed9d310d5fc83900f11072" HandleID="k8s-pod-network.53803ab3ac7d3053c9196e001fef8e60712767b15aed9d310d5fc83900f11072" Workload="localhost-k8s-calico--kube--controllers--84bd499cb--6t459-eth0" Jul 10 00:28:41.606763 containerd[1640]: 2025-07-10 00:28:41.590 [INFO][4722] cni-plugin/k8s.go 418: Populated endpoint ContainerID="53803ab3ac7d3053c9196e001fef8e60712767b15aed9d310d5fc83900f11072" Namespace="calico-system" Pod="calico-kube-controllers-84bd499cb-6t459" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--84bd499cb--6t459-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--84bd499cb--6t459-eth0", GenerateName:"calico-kube-controllers-84bd499cb-", Namespace:"calico-system", SelfLink:"", UID:"0a9d7ca6-fb8c-4239-ae97-5a87dba79acf", ResourceVersion:"795", Generation:0, CreationTimestamp:time.Date(2025, time.July, 10, 0, 28, 18, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"84bd499cb", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-kube-controllers-84bd499cb-6t459", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calib45b0a49d9b", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 10 00:28:41.606763 containerd[1640]: 2025-07-10 00:28:41.592 [INFO][4722] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.133/32] ContainerID="53803ab3ac7d3053c9196e001fef8e60712767b15aed9d310d5fc83900f11072" Namespace="calico-system" Pod="calico-kube-controllers-84bd499cb-6t459" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--84bd499cb--6t459-eth0" Jul 10 00:28:41.606763 containerd[1640]: 2025-07-10 00:28:41.592 [INFO][4722] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calib45b0a49d9b ContainerID="53803ab3ac7d3053c9196e001fef8e60712767b15aed9d310d5fc83900f11072" Namespace="calico-system" Pod="calico-kube-controllers-84bd499cb-6t459" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--84bd499cb--6t459-eth0" Jul 10 00:28:41.606763 containerd[1640]: 2025-07-10 00:28:41.594 [INFO][4722] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="53803ab3ac7d3053c9196e001fef8e60712767b15aed9d310d5fc83900f11072" Namespace="calico-system" Pod="calico-kube-controllers-84bd499cb-6t459" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--84bd499cb--6t459-eth0" Jul 10 00:28:41.606763 containerd[1640]: 2025-07-10 00:28:41.595 [INFO][4722] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="53803ab3ac7d3053c9196e001fef8e60712767b15aed9d310d5fc83900f11072" Namespace="calico-system" Pod="calico-kube-controllers-84bd499cb-6t459" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--84bd499cb--6t459-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--84bd499cb--6t459-eth0", GenerateName:"calico-kube-controllers-84bd499cb-", Namespace:"calico-system", SelfLink:"", UID:"0a9d7ca6-fb8c-4239-ae97-5a87dba79acf", ResourceVersion:"795", Generation:0, CreationTimestamp:time.Date(2025, time.July, 10, 0, 28, 18, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"84bd499cb", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"53803ab3ac7d3053c9196e001fef8e60712767b15aed9d310d5fc83900f11072", Pod:"calico-kube-controllers-84bd499cb-6t459", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calib45b0a49d9b", MAC:"1e:7f:07:56:4a:8a", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 10 00:28:41.606763 containerd[1640]: 2025-07-10 00:28:41.603 [INFO][4722] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="53803ab3ac7d3053c9196e001fef8e60712767b15aed9d310d5fc83900f11072" Namespace="calico-system" Pod="calico-kube-controllers-84bd499cb-6t459" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--84bd499cb--6t459-eth0" Jul 10 00:28:41.639044 containerd[1640]: time="2025-07-10T00:28:41.638012465Z" level=info msg="connecting to shim 53803ab3ac7d3053c9196e001fef8e60712767b15aed9d310d5fc83900f11072" address="unix:///run/containerd/s/aba001255bed2b76ccc0f0c8e61b2fe915144f94fedbc3b51613214e6ef11a1b" namespace=k8s.io protocol=ttrpc version=3 Jul 10 00:28:41.677426 kubelet[2915]: I0710 00:28:41.677279 2915 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7c65d6cfc9-8h9ft" podStartSLOduration=34.677264386 podStartE2EDuration="34.677264386s" podCreationTimestamp="2025-07-10 00:28:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-10 00:28:41.660796946 +0000 UTC m=+40.405577178" watchObservedRunningTime="2025-07-10 00:28:41.677264386 +0000 UTC m=+40.422044602" Jul 10 00:28:41.689470 kubelet[2915]: I0710 00:28:41.689257 2915 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7c65d6cfc9-g5zz7" podStartSLOduration=34.689242487 podStartE2EDuration="34.689242487s" podCreationTimestamp="2025-07-10 00:28:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-10 00:28:41.675938201 +0000 UTC m=+40.420718423" watchObservedRunningTime="2025-07-10 00:28:41.689242487 +0000 UTC m=+40.434022703" Jul 10 00:28:41.732324 systemd[1]: Started cri-containerd-53803ab3ac7d3053c9196e001fef8e60712767b15aed9d310d5fc83900f11072.scope - libcontainer container 53803ab3ac7d3053c9196e001fef8e60712767b15aed9d310d5fc83900f11072. Jul 10 00:28:41.749251 systemd-resolved[1528]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jul 10 00:28:41.749670 systemd-networkd[1524]: cali2f531b22198: Link UP Jul 10 00:28:41.750223 systemd-networkd[1524]: cali2f531b22198: Gained carrier Jul 10 00:28:41.761028 containerd[1640]: 2025-07-10 00:28:41.501 [INFO][4715] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-csi--node--driver--pzv7w-eth0 csi-node-driver- calico-system 2587365c-21ee-44fd-91df-0ab3a24762fc 689 0 2025-07-10 00:28:18 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:57bd658777 k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s localhost csi-node-driver-pzv7w eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali2f531b22198 [] [] }} ContainerID="d12f1801782c97a0a45de4f7a5eb33bafeee4bce3484d0804d324fea61756a57" Namespace="calico-system" Pod="csi-node-driver-pzv7w" WorkloadEndpoint="localhost-k8s-csi--node--driver--pzv7w-" Jul 10 00:28:41.761028 containerd[1640]: 2025-07-10 00:28:41.501 [INFO][4715] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="d12f1801782c97a0a45de4f7a5eb33bafeee4bce3484d0804d324fea61756a57" Namespace="calico-system" Pod="csi-node-driver-pzv7w" WorkloadEndpoint="localhost-k8s-csi--node--driver--pzv7w-eth0" Jul 10 00:28:41.761028 containerd[1640]: 2025-07-10 00:28:41.570 [INFO][4755] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="d12f1801782c97a0a45de4f7a5eb33bafeee4bce3484d0804d324fea61756a57" HandleID="k8s-pod-network.d12f1801782c97a0a45de4f7a5eb33bafeee4bce3484d0804d324fea61756a57" Workload="localhost-k8s-csi--node--driver--pzv7w-eth0" Jul 10 00:28:41.761028 containerd[1640]: 2025-07-10 00:28:41.571 [INFO][4755] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="d12f1801782c97a0a45de4f7a5eb33bafeee4bce3484d0804d324fea61756a57" HandleID="k8s-pod-network.d12f1801782c97a0a45de4f7a5eb33bafeee4bce3484d0804d324fea61756a57" Workload="localhost-k8s-csi--node--driver--pzv7w-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002d5190), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"csi-node-driver-pzv7w", "timestamp":"2025-07-10 00:28:41.57086136 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jul 10 00:28:41.761028 containerd[1640]: 2025-07-10 00:28:41.571 [INFO][4755] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 10 00:28:41.761028 containerd[1640]: 2025-07-10 00:28:41.588 [INFO][4755] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 10 00:28:41.761028 containerd[1640]: 2025-07-10 00:28:41.588 [INFO][4755] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jul 10 00:28:41.761028 containerd[1640]: 2025-07-10 00:28:41.678 [INFO][4755] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.d12f1801782c97a0a45de4f7a5eb33bafeee4bce3484d0804d324fea61756a57" host="localhost" Jul 10 00:28:41.761028 containerd[1640]: 2025-07-10 00:28:41.703 [INFO][4755] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jul 10 00:28:41.761028 containerd[1640]: 2025-07-10 00:28:41.716 [INFO][4755] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jul 10 00:28:41.761028 containerd[1640]: 2025-07-10 00:28:41.721 [INFO][4755] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jul 10 00:28:41.761028 containerd[1640]: 2025-07-10 00:28:41.728 [INFO][4755] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jul 10 00:28:41.761028 containerd[1640]: 2025-07-10 00:28:41.728 [INFO][4755] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.d12f1801782c97a0a45de4f7a5eb33bafeee4bce3484d0804d324fea61756a57" host="localhost" Jul 10 00:28:41.761028 containerd[1640]: 2025-07-10 00:28:41.732 [INFO][4755] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.d12f1801782c97a0a45de4f7a5eb33bafeee4bce3484d0804d324fea61756a57 Jul 10 00:28:41.761028 containerd[1640]: 2025-07-10 00:28:41.735 [INFO][4755] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.d12f1801782c97a0a45de4f7a5eb33bafeee4bce3484d0804d324fea61756a57" host="localhost" Jul 10 00:28:41.761028 containerd[1640]: 2025-07-10 00:28:41.739 [INFO][4755] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.134/26] block=192.168.88.128/26 handle="k8s-pod-network.d12f1801782c97a0a45de4f7a5eb33bafeee4bce3484d0804d324fea61756a57" host="localhost" Jul 10 00:28:41.761028 containerd[1640]: 2025-07-10 00:28:41.740 [INFO][4755] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.134/26] handle="k8s-pod-network.d12f1801782c97a0a45de4f7a5eb33bafeee4bce3484d0804d324fea61756a57" host="localhost" Jul 10 00:28:41.761028 containerd[1640]: 2025-07-10 00:28:41.740 [INFO][4755] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 10 00:28:41.761028 containerd[1640]: 2025-07-10 00:28:41.740 [INFO][4755] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.134/26] IPv6=[] ContainerID="d12f1801782c97a0a45de4f7a5eb33bafeee4bce3484d0804d324fea61756a57" HandleID="k8s-pod-network.d12f1801782c97a0a45de4f7a5eb33bafeee4bce3484d0804d324fea61756a57" Workload="localhost-k8s-csi--node--driver--pzv7w-eth0" Jul 10 00:28:41.761799 containerd[1640]: 2025-07-10 00:28:41.746 [INFO][4715] cni-plugin/k8s.go 418: Populated endpoint ContainerID="d12f1801782c97a0a45de4f7a5eb33bafeee4bce3484d0804d324fea61756a57" Namespace="calico-system" Pod="csi-node-driver-pzv7w" WorkloadEndpoint="localhost-k8s-csi--node--driver--pzv7w-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--pzv7w-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"2587365c-21ee-44fd-91df-0ab3a24762fc", ResourceVersion:"689", Generation:0, CreationTimestamp:time.Date(2025, time.July, 10, 0, 28, 18, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"57bd658777", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"csi-node-driver-pzv7w", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali2f531b22198", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 10 00:28:41.761799 containerd[1640]: 2025-07-10 00:28:41.746 [INFO][4715] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.134/32] ContainerID="d12f1801782c97a0a45de4f7a5eb33bafeee4bce3484d0804d324fea61756a57" Namespace="calico-system" Pod="csi-node-driver-pzv7w" WorkloadEndpoint="localhost-k8s-csi--node--driver--pzv7w-eth0" Jul 10 00:28:41.761799 containerd[1640]: 2025-07-10 00:28:41.746 [INFO][4715] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali2f531b22198 ContainerID="d12f1801782c97a0a45de4f7a5eb33bafeee4bce3484d0804d324fea61756a57" Namespace="calico-system" Pod="csi-node-driver-pzv7w" WorkloadEndpoint="localhost-k8s-csi--node--driver--pzv7w-eth0" Jul 10 00:28:41.761799 containerd[1640]: 2025-07-10 00:28:41.750 [INFO][4715] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="d12f1801782c97a0a45de4f7a5eb33bafeee4bce3484d0804d324fea61756a57" Namespace="calico-system" Pod="csi-node-driver-pzv7w" WorkloadEndpoint="localhost-k8s-csi--node--driver--pzv7w-eth0" Jul 10 00:28:41.761799 containerd[1640]: 2025-07-10 00:28:41.751 [INFO][4715] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="d12f1801782c97a0a45de4f7a5eb33bafeee4bce3484d0804d324fea61756a57" Namespace="calico-system" Pod="csi-node-driver-pzv7w" WorkloadEndpoint="localhost-k8s-csi--node--driver--pzv7w-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--pzv7w-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"2587365c-21ee-44fd-91df-0ab3a24762fc", ResourceVersion:"689", Generation:0, CreationTimestamp:time.Date(2025, time.July, 10, 0, 28, 18, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"57bd658777", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"d12f1801782c97a0a45de4f7a5eb33bafeee4bce3484d0804d324fea61756a57", Pod:"csi-node-driver-pzv7w", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali2f531b22198", MAC:"0e:9f:a2:d0:f8:ce", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 10 00:28:41.761799 containerd[1640]: 2025-07-10 00:28:41.758 [INFO][4715] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="d12f1801782c97a0a45de4f7a5eb33bafeee4bce3484d0804d324fea61756a57" Namespace="calico-system" Pod="csi-node-driver-pzv7w" WorkloadEndpoint="localhost-k8s-csi--node--driver--pzv7w-eth0" Jul 10 00:28:41.776434 containerd[1640]: time="2025-07-10T00:28:41.776404436Z" level=info msg="connecting to shim d12f1801782c97a0a45de4f7a5eb33bafeee4bce3484d0804d324fea61756a57" address="unix:///run/containerd/s/b8239e5da780f64771a109b606a715fb5b26db6b0cbe0ac05bcab5f0a7ded206" namespace=k8s.io protocol=ttrpc version=3 Jul 10 00:28:41.795262 systemd[1]: Started cri-containerd-d12f1801782c97a0a45de4f7a5eb33bafeee4bce3484d0804d324fea61756a57.scope - libcontainer container d12f1801782c97a0a45de4f7a5eb33bafeee4bce3484d0804d324fea61756a57. Jul 10 00:28:41.801118 containerd[1640]: time="2025-07-10T00:28:41.800728381Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-84bd499cb-6t459,Uid:0a9d7ca6-fb8c-4239-ae97-5a87dba79acf,Namespace:calico-system,Attempt:0,} returns sandbox id \"53803ab3ac7d3053c9196e001fef8e60712767b15aed9d310d5fc83900f11072\"" Jul 10 00:28:41.810970 systemd-resolved[1528]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jul 10 00:28:41.826258 containerd[1640]: time="2025-07-10T00:28:41.826222677Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-pzv7w,Uid:2587365c-21ee-44fd-91df-0ab3a24762fc,Namespace:calico-system,Attempt:0,} returns sandbox id \"d12f1801782c97a0a45de4f7a5eb33bafeee4bce3484d0804d324fea61756a57\"" Jul 10 00:28:41.830619 systemd-networkd[1524]: cali2866707eac1: Link UP Jul 10 00:28:41.831399 systemd-networkd[1524]: cali2866707eac1: Gained carrier Jul 10 00:28:41.840467 containerd[1640]: 2025-07-10 00:28:41.511 [INFO][4710] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--7c75d6c687--8jqvg-eth0 calico-apiserver-7c75d6c687- calico-apiserver 082d696e-d4fa-497e-ab24-928a088c4f21 801 0 2025-07-10 00:28:15 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:7c75d6c687 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-7c75d6c687-8jqvg eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali2866707eac1 [] [] }} ContainerID="b00bb58c3c648815d4d8eb8c2dc340380592eb466066c9c6444054b407de8c7d" Namespace="calico-apiserver" Pod="calico-apiserver-7c75d6c687-8jqvg" WorkloadEndpoint="localhost-k8s-calico--apiserver--7c75d6c687--8jqvg-" Jul 10 00:28:41.840467 containerd[1640]: 2025-07-10 00:28:41.511 [INFO][4710] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="b00bb58c3c648815d4d8eb8c2dc340380592eb466066c9c6444054b407de8c7d" Namespace="calico-apiserver" Pod="calico-apiserver-7c75d6c687-8jqvg" WorkloadEndpoint="localhost-k8s-calico--apiserver--7c75d6c687--8jqvg-eth0" Jul 10 00:28:41.840467 containerd[1640]: 2025-07-10 00:28:41.571 [INFO][4763] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="b00bb58c3c648815d4d8eb8c2dc340380592eb466066c9c6444054b407de8c7d" HandleID="k8s-pod-network.b00bb58c3c648815d4d8eb8c2dc340380592eb466066c9c6444054b407de8c7d" Workload="localhost-k8s-calico--apiserver--7c75d6c687--8jqvg-eth0" Jul 10 00:28:41.840467 containerd[1640]: 2025-07-10 00:28:41.571 [INFO][4763] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="b00bb58c3c648815d4d8eb8c2dc340380592eb466066c9c6444054b407de8c7d" HandleID="k8s-pod-network.b00bb58c3c648815d4d8eb8c2dc340380592eb466066c9c6444054b407de8c7d" Workload="localhost-k8s-calico--apiserver--7c75d6c687--8jqvg-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00004f770), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-7c75d6c687-8jqvg", "timestamp":"2025-07-10 00:28:41.571003873 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jul 10 00:28:41.840467 containerd[1640]: 2025-07-10 00:28:41.571 [INFO][4763] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 10 00:28:41.840467 containerd[1640]: 2025-07-10 00:28:41.740 [INFO][4763] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 10 00:28:41.840467 containerd[1640]: 2025-07-10 00:28:41.740 [INFO][4763] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jul 10 00:28:41.840467 containerd[1640]: 2025-07-10 00:28:41.776 [INFO][4763] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.b00bb58c3c648815d4d8eb8c2dc340380592eb466066c9c6444054b407de8c7d" host="localhost" Jul 10 00:28:41.840467 containerd[1640]: 2025-07-10 00:28:41.800 [INFO][4763] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jul 10 00:28:41.840467 containerd[1640]: 2025-07-10 00:28:41.807 [INFO][4763] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jul 10 00:28:41.840467 containerd[1640]: 2025-07-10 00:28:41.809 [INFO][4763] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jul 10 00:28:41.840467 containerd[1640]: 2025-07-10 00:28:41.811 [INFO][4763] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jul 10 00:28:41.840467 containerd[1640]: 2025-07-10 00:28:41.811 [INFO][4763] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.b00bb58c3c648815d4d8eb8c2dc340380592eb466066c9c6444054b407de8c7d" host="localhost" Jul 10 00:28:41.840467 containerd[1640]: 2025-07-10 00:28:41.812 [INFO][4763] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.b00bb58c3c648815d4d8eb8c2dc340380592eb466066c9c6444054b407de8c7d Jul 10 00:28:41.840467 containerd[1640]: 2025-07-10 00:28:41.815 [INFO][4763] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.b00bb58c3c648815d4d8eb8c2dc340380592eb466066c9c6444054b407de8c7d" host="localhost" Jul 10 00:28:41.840467 containerd[1640]: 2025-07-10 00:28:41.822 [INFO][4763] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.135/26] block=192.168.88.128/26 handle="k8s-pod-network.b00bb58c3c648815d4d8eb8c2dc340380592eb466066c9c6444054b407de8c7d" host="localhost" Jul 10 00:28:41.840467 containerd[1640]: 2025-07-10 00:28:41.822 [INFO][4763] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.135/26] handle="k8s-pod-network.b00bb58c3c648815d4d8eb8c2dc340380592eb466066c9c6444054b407de8c7d" host="localhost" Jul 10 00:28:41.840467 containerd[1640]: 2025-07-10 00:28:41.822 [INFO][4763] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 10 00:28:41.840467 containerd[1640]: 2025-07-10 00:28:41.822 [INFO][4763] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.135/26] IPv6=[] ContainerID="b00bb58c3c648815d4d8eb8c2dc340380592eb466066c9c6444054b407de8c7d" HandleID="k8s-pod-network.b00bb58c3c648815d4d8eb8c2dc340380592eb466066c9c6444054b407de8c7d" Workload="localhost-k8s-calico--apiserver--7c75d6c687--8jqvg-eth0" Jul 10 00:28:41.843652 containerd[1640]: 2025-07-10 00:28:41.826 [INFO][4710] cni-plugin/k8s.go 418: Populated endpoint ContainerID="b00bb58c3c648815d4d8eb8c2dc340380592eb466066c9c6444054b407de8c7d" Namespace="calico-apiserver" Pod="calico-apiserver-7c75d6c687-8jqvg" WorkloadEndpoint="localhost-k8s-calico--apiserver--7c75d6c687--8jqvg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--7c75d6c687--8jqvg-eth0", GenerateName:"calico-apiserver-7c75d6c687-", Namespace:"calico-apiserver", SelfLink:"", UID:"082d696e-d4fa-497e-ab24-928a088c4f21", ResourceVersion:"801", Generation:0, CreationTimestamp:time.Date(2025, time.July, 10, 0, 28, 15, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7c75d6c687", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-7c75d6c687-8jqvg", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali2866707eac1", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 10 00:28:41.843652 containerd[1640]: 2025-07-10 00:28:41.826 [INFO][4710] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.135/32] ContainerID="b00bb58c3c648815d4d8eb8c2dc340380592eb466066c9c6444054b407de8c7d" Namespace="calico-apiserver" Pod="calico-apiserver-7c75d6c687-8jqvg" WorkloadEndpoint="localhost-k8s-calico--apiserver--7c75d6c687--8jqvg-eth0" Jul 10 00:28:41.843652 containerd[1640]: 2025-07-10 00:28:41.826 [INFO][4710] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali2866707eac1 ContainerID="b00bb58c3c648815d4d8eb8c2dc340380592eb466066c9c6444054b407de8c7d" Namespace="calico-apiserver" Pod="calico-apiserver-7c75d6c687-8jqvg" WorkloadEndpoint="localhost-k8s-calico--apiserver--7c75d6c687--8jqvg-eth0" Jul 10 00:28:41.843652 containerd[1640]: 2025-07-10 00:28:41.831 [INFO][4710] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="b00bb58c3c648815d4d8eb8c2dc340380592eb466066c9c6444054b407de8c7d" Namespace="calico-apiserver" Pod="calico-apiserver-7c75d6c687-8jqvg" WorkloadEndpoint="localhost-k8s-calico--apiserver--7c75d6c687--8jqvg-eth0" Jul 10 00:28:41.843652 containerd[1640]: 2025-07-10 00:28:41.832 [INFO][4710] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="b00bb58c3c648815d4d8eb8c2dc340380592eb466066c9c6444054b407de8c7d" Namespace="calico-apiserver" Pod="calico-apiserver-7c75d6c687-8jqvg" WorkloadEndpoint="localhost-k8s-calico--apiserver--7c75d6c687--8jqvg-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--7c75d6c687--8jqvg-eth0", GenerateName:"calico-apiserver-7c75d6c687-", Namespace:"calico-apiserver", SelfLink:"", UID:"082d696e-d4fa-497e-ab24-928a088c4f21", ResourceVersion:"801", Generation:0, CreationTimestamp:time.Date(2025, time.July, 10, 0, 28, 15, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7c75d6c687", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"b00bb58c3c648815d4d8eb8c2dc340380592eb466066c9c6444054b407de8c7d", Pod:"calico-apiserver-7c75d6c687-8jqvg", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.135/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali2866707eac1", MAC:"ea:cd:b5:8c:86:15", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 10 00:28:41.843652 containerd[1640]: 2025-07-10 00:28:41.838 [INFO][4710] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="b00bb58c3c648815d4d8eb8c2dc340380592eb466066c9c6444054b407de8c7d" Namespace="calico-apiserver" Pod="calico-apiserver-7c75d6c687-8jqvg" WorkloadEndpoint="localhost-k8s-calico--apiserver--7c75d6c687--8jqvg-eth0" Jul 10 00:28:41.852487 containerd[1640]: time="2025-07-10T00:28:41.852207259Z" level=info msg="connecting to shim b00bb58c3c648815d4d8eb8c2dc340380592eb466066c9c6444054b407de8c7d" address="unix:///run/containerd/s/afb4684c686eed7e2609914aa4a51f4fd0713a8a5fae1ce685f003dcfa9104c7" namespace=k8s.io protocol=ttrpc version=3 Jul 10 00:28:41.869269 systemd[1]: Started cri-containerd-b00bb58c3c648815d4d8eb8c2dc340380592eb466066c9c6444054b407de8c7d.scope - libcontainer container b00bb58c3c648815d4d8eb8c2dc340380592eb466066c9c6444054b407de8c7d. Jul 10 00:28:41.877895 systemd-resolved[1528]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jul 10 00:28:41.907198 containerd[1640]: time="2025-07-10T00:28:41.907168194Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7c75d6c687-8jqvg,Uid:082d696e-d4fa-497e-ab24-928a088c4f21,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"b00bb58c3c648815d4d8eb8c2dc340380592eb466066c9c6444054b407de8c7d\"" Jul 10 00:28:42.245312 systemd-networkd[1524]: cali86ba49b9c17: Gained IPv6LL Jul 10 00:28:42.308266 systemd-networkd[1524]: cali065cf7a8b4e: Gained IPv6LL Jul 10 00:28:42.395880 containerd[1640]: time="2025-07-10T00:28:42.395855937Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-58fd7646b9-52ptt,Uid:0841b7eb-4ad6-4003-b535-92aae1f1d1f2,Namespace:calico-system,Attempt:0,}" Jul 10 00:28:42.629306 systemd-networkd[1524]: cali59a605a182f: Gained IPv6LL Jul 10 00:28:42.779441 systemd-networkd[1524]: calief9531ad153: Link UP Jul 10 00:28:42.780258 systemd-networkd[1524]: calief9531ad153: Gained carrier Jul 10 00:28:42.798008 containerd[1640]: 2025-07-10 00:28:42.649 [INFO][4942] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-goldmane--58fd7646b9--52ptt-eth0 goldmane-58fd7646b9- calico-system 0841b7eb-4ad6-4003-b535-92aae1f1d1f2 797 0 2025-07-10 00:28:17 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:58fd7646b9 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s localhost goldmane-58fd7646b9-52ptt eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] calief9531ad153 [] [] }} ContainerID="a406a4e994665bce6898b2f2c15a7d025b0c29070bfc5b3b5417aa375134d460" Namespace="calico-system" Pod="goldmane-58fd7646b9-52ptt" WorkloadEndpoint="localhost-k8s-goldmane--58fd7646b9--52ptt-" Jul 10 00:28:42.798008 containerd[1640]: 2025-07-10 00:28:42.649 [INFO][4942] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="a406a4e994665bce6898b2f2c15a7d025b0c29070bfc5b3b5417aa375134d460" Namespace="calico-system" Pod="goldmane-58fd7646b9-52ptt" WorkloadEndpoint="localhost-k8s-goldmane--58fd7646b9--52ptt-eth0" Jul 10 00:28:42.798008 containerd[1640]: 2025-07-10 00:28:42.677 [INFO][4957] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="a406a4e994665bce6898b2f2c15a7d025b0c29070bfc5b3b5417aa375134d460" HandleID="k8s-pod-network.a406a4e994665bce6898b2f2c15a7d025b0c29070bfc5b3b5417aa375134d460" Workload="localhost-k8s-goldmane--58fd7646b9--52ptt-eth0" Jul 10 00:28:42.798008 containerd[1640]: 2025-07-10 00:28:42.677 [INFO][4957] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="a406a4e994665bce6898b2f2c15a7d025b0c29070bfc5b3b5417aa375134d460" HandleID="k8s-pod-network.a406a4e994665bce6898b2f2c15a7d025b0c29070bfc5b3b5417aa375134d460" Workload="localhost-k8s-goldmane--58fd7646b9--52ptt-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002cd730), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"goldmane-58fd7646b9-52ptt", "timestamp":"2025-07-10 00:28:42.677865962 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jul 10 00:28:42.798008 containerd[1640]: 2025-07-10 00:28:42.677 [INFO][4957] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jul 10 00:28:42.798008 containerd[1640]: 2025-07-10 00:28:42.678 [INFO][4957] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jul 10 00:28:42.798008 containerd[1640]: 2025-07-10 00:28:42.678 [INFO][4957] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jul 10 00:28:42.798008 containerd[1640]: 2025-07-10 00:28:42.684 [INFO][4957] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.a406a4e994665bce6898b2f2c15a7d025b0c29070bfc5b3b5417aa375134d460" host="localhost" Jul 10 00:28:42.798008 containerd[1640]: 2025-07-10 00:28:42.700 [INFO][4957] ipam/ipam.go 394: Looking up existing affinities for host host="localhost" Jul 10 00:28:42.798008 containerd[1640]: 2025-07-10 00:28:42.710 [INFO][4957] ipam/ipam.go 511: Trying affinity for 192.168.88.128/26 host="localhost" Jul 10 00:28:42.798008 containerd[1640]: 2025-07-10 00:28:42.711 [INFO][4957] ipam/ipam.go 158: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jul 10 00:28:42.798008 containerd[1640]: 2025-07-10 00:28:42.730 [INFO][4957] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jul 10 00:28:42.798008 containerd[1640]: 2025-07-10 00:28:42.730 [INFO][4957] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.a406a4e994665bce6898b2f2c15a7d025b0c29070bfc5b3b5417aa375134d460" host="localhost" Jul 10 00:28:42.798008 containerd[1640]: 2025-07-10 00:28:42.737 [INFO][4957] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.a406a4e994665bce6898b2f2c15a7d025b0c29070bfc5b3b5417aa375134d460 Jul 10 00:28:42.798008 containerd[1640]: 2025-07-10 00:28:42.752 [INFO][4957] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.a406a4e994665bce6898b2f2c15a7d025b0c29070bfc5b3b5417aa375134d460" host="localhost" Jul 10 00:28:42.798008 containerd[1640]: 2025-07-10 00:28:42.772 [INFO][4957] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.88.136/26] block=192.168.88.128/26 handle="k8s-pod-network.a406a4e994665bce6898b2f2c15a7d025b0c29070bfc5b3b5417aa375134d460" host="localhost" Jul 10 00:28:42.798008 containerd[1640]: 2025-07-10 00:28:42.772 [INFO][4957] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.88.136/26] handle="k8s-pod-network.a406a4e994665bce6898b2f2c15a7d025b0c29070bfc5b3b5417aa375134d460" host="localhost" Jul 10 00:28:42.798008 containerd[1640]: 2025-07-10 00:28:42.773 [INFO][4957] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jul 10 00:28:42.798008 containerd[1640]: 2025-07-10 00:28:42.773 [INFO][4957] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.136/26] IPv6=[] ContainerID="a406a4e994665bce6898b2f2c15a7d025b0c29070bfc5b3b5417aa375134d460" HandleID="k8s-pod-network.a406a4e994665bce6898b2f2c15a7d025b0c29070bfc5b3b5417aa375134d460" Workload="localhost-k8s-goldmane--58fd7646b9--52ptt-eth0" Jul 10 00:28:42.806787 containerd[1640]: 2025-07-10 00:28:42.774 [INFO][4942] cni-plugin/k8s.go 418: Populated endpoint ContainerID="a406a4e994665bce6898b2f2c15a7d025b0c29070bfc5b3b5417aa375134d460" Namespace="calico-system" Pod="goldmane-58fd7646b9-52ptt" WorkloadEndpoint="localhost-k8s-goldmane--58fd7646b9--52ptt-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--58fd7646b9--52ptt-eth0", GenerateName:"goldmane-58fd7646b9-", Namespace:"calico-system", SelfLink:"", UID:"0841b7eb-4ad6-4003-b535-92aae1f1d1f2", ResourceVersion:"797", Generation:0, CreationTimestamp:time.Date(2025, time.July, 10, 0, 28, 17, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"58fd7646b9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"goldmane-58fd7646b9-52ptt", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"calief9531ad153", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 10 00:28:42.806787 containerd[1640]: 2025-07-10 00:28:42.775 [INFO][4942] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.88.136/32] ContainerID="a406a4e994665bce6898b2f2c15a7d025b0c29070bfc5b3b5417aa375134d460" Namespace="calico-system" Pod="goldmane-58fd7646b9-52ptt" WorkloadEndpoint="localhost-k8s-goldmane--58fd7646b9--52ptt-eth0" Jul 10 00:28:42.806787 containerd[1640]: 2025-07-10 00:28:42.775 [INFO][4942] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calief9531ad153 ContainerID="a406a4e994665bce6898b2f2c15a7d025b0c29070bfc5b3b5417aa375134d460" Namespace="calico-system" Pod="goldmane-58fd7646b9-52ptt" WorkloadEndpoint="localhost-k8s-goldmane--58fd7646b9--52ptt-eth0" Jul 10 00:28:42.806787 containerd[1640]: 2025-07-10 00:28:42.780 [INFO][4942] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="a406a4e994665bce6898b2f2c15a7d025b0c29070bfc5b3b5417aa375134d460" Namespace="calico-system" Pod="goldmane-58fd7646b9-52ptt" WorkloadEndpoint="localhost-k8s-goldmane--58fd7646b9--52ptt-eth0" Jul 10 00:28:42.806787 containerd[1640]: 2025-07-10 00:28:42.781 [INFO][4942] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="a406a4e994665bce6898b2f2c15a7d025b0c29070bfc5b3b5417aa375134d460" Namespace="calico-system" Pod="goldmane-58fd7646b9-52ptt" WorkloadEndpoint="localhost-k8s-goldmane--58fd7646b9--52ptt-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-goldmane--58fd7646b9--52ptt-eth0", GenerateName:"goldmane-58fd7646b9-", Namespace:"calico-system", SelfLink:"", UID:"0841b7eb-4ad6-4003-b535-92aae1f1d1f2", ResourceVersion:"797", Generation:0, CreationTimestamp:time.Date(2025, time.July, 10, 0, 28, 17, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"58fd7646b9", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"a406a4e994665bce6898b2f2c15a7d025b0c29070bfc5b3b5417aa375134d460", Pod:"goldmane-58fd7646b9-52ptt", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.88.136/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"calief9531ad153", MAC:"ba:a0:7f:fb:e5:e5", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Jul 10 00:28:42.806787 containerd[1640]: 2025-07-10 00:28:42.793 [INFO][4942] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="a406a4e994665bce6898b2f2c15a7d025b0c29070bfc5b3b5417aa375134d460" Namespace="calico-system" Pod="goldmane-58fd7646b9-52ptt" WorkloadEndpoint="localhost-k8s-goldmane--58fd7646b9--52ptt-eth0" Jul 10 00:28:42.853703 containerd[1640]: time="2025-07-10T00:28:42.853399303Z" level=info msg="connecting to shim a406a4e994665bce6898b2f2c15a7d025b0c29070bfc5b3b5417aa375134d460" address="unix:///run/containerd/s/598468a299605de5131a3b6a3c2504616ebe6b04ba7a5fe4513b7453548db158" namespace=k8s.io protocol=ttrpc version=3 Jul 10 00:28:42.875268 systemd[1]: Started cri-containerd-a406a4e994665bce6898b2f2c15a7d025b0c29070bfc5b3b5417aa375134d460.scope - libcontainer container a406a4e994665bce6898b2f2c15a7d025b0c29070bfc5b3b5417aa375134d460. Jul 10 00:28:42.887067 systemd-resolved[1528]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jul 10 00:28:42.944826 containerd[1640]: time="2025-07-10T00:28:42.944804700Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-58fd7646b9-52ptt,Uid:0841b7eb-4ad6-4003-b535-92aae1f1d1f2,Namespace:calico-system,Attempt:0,} returns sandbox id \"a406a4e994665bce6898b2f2c15a7d025b0c29070bfc5b3b5417aa375134d460\"" Jul 10 00:28:43.012263 systemd-networkd[1524]: calib45b0a49d9b: Gained IPv6LL Jul 10 00:28:43.012763 systemd-networkd[1524]: cali2866707eac1: Gained IPv6LL Jul 10 00:28:43.293606 containerd[1640]: time="2025-07-10T00:28:43.293540566Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:43.294843 containerd[1640]: time="2025-07-10T00:28:43.294829278Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.2: active requests=0, bytes read=47317977" Jul 10 00:28:43.295088 containerd[1640]: time="2025-07-10T00:28:43.295067883Z" level=info msg="ImageCreate event name:\"sha256:5509118eed617ef04ca00f5a095bfd0a4cd1cf69edcfcf9bedf0edb641be51dd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:43.297094 containerd[1640]: time="2025-07-10T00:28:43.296541448Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:ec6b10660962e7caad70c47755049fad68f9fc2f7064e8bc7cb862583e02cc2b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:43.297094 containerd[1640]: time="2025-07-10T00:28:43.296896530Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.30.2\" with image id \"sha256:5509118eed617ef04ca00f5a095bfd0a4cd1cf69edcfcf9bedf0edb641be51dd\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:ec6b10660962e7caad70c47755049fad68f9fc2f7064e8bc7cb862583e02cc2b\", size \"48810696\" in 2.369995119s" Jul 10 00:28:43.297094 containerd[1640]: time="2025-07-10T00:28:43.296910444Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.2\" returns image reference \"sha256:5509118eed617ef04ca00f5a095bfd0a4cd1cf69edcfcf9bedf0edb641be51dd\"" Jul 10 00:28:43.298541 containerd[1640]: time="2025-07-10T00:28:43.298474493Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\"" Jul 10 00:28:43.301587 containerd[1640]: time="2025-07-10T00:28:43.301575527Z" level=info msg="CreateContainer within sandbox \"6cd8f5071bad46e9132abaa80bad3e69ea82733afa4e3b4d244053a865a88c26\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Jul 10 00:28:43.315432 containerd[1640]: time="2025-07-10T00:28:43.315415213Z" level=info msg="Container 184c6b50649ac2f1226c04737b9a476d13bfd7fe6fae93cf518c25716b9b1900: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:28:43.332236 containerd[1640]: time="2025-07-10T00:28:43.332222633Z" level=info msg="CreateContainer within sandbox \"6cd8f5071bad46e9132abaa80bad3e69ea82733afa4e3b4d244053a865a88c26\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"184c6b50649ac2f1226c04737b9a476d13bfd7fe6fae93cf518c25716b9b1900\"" Jul 10 00:28:43.332642 containerd[1640]: time="2025-07-10T00:28:43.332632997Z" level=info msg="StartContainer for \"184c6b50649ac2f1226c04737b9a476d13bfd7fe6fae93cf518c25716b9b1900\"" Jul 10 00:28:43.334858 containerd[1640]: time="2025-07-10T00:28:43.334844763Z" level=info msg="connecting to shim 184c6b50649ac2f1226c04737b9a476d13bfd7fe6fae93cf518c25716b9b1900" address="unix:///run/containerd/s/404b43027837d720f6c87db0b40705472b8e94d4d358c43d18e5089c28c4cb1f" protocol=ttrpc version=3 Jul 10 00:28:43.351248 systemd[1]: Started cri-containerd-184c6b50649ac2f1226c04737b9a476d13bfd7fe6fae93cf518c25716b9b1900.scope - libcontainer container 184c6b50649ac2f1226c04737b9a476d13bfd7fe6fae93cf518c25716b9b1900. Jul 10 00:28:43.394289 containerd[1640]: time="2025-07-10T00:28:43.394268370Z" level=info msg="StartContainer for \"184c6b50649ac2f1226c04737b9a476d13bfd7fe6fae93cf518c25716b9b1900\" returns successfully" Jul 10 00:28:43.399289 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3637967305.mount: Deactivated successfully. Jul 10 00:28:43.716407 systemd-networkd[1524]: cali2f531b22198: Gained IPv6LL Jul 10 00:28:44.484684 systemd-networkd[1524]: calief9531ad153: Gained IPv6LL Jul 10 00:28:44.676880 kubelet[2915]: I0710 00:28:44.676838 2915 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jul 10 00:28:46.934639 containerd[1640]: time="2025-07-10T00:28:46.934594182Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:46.946969 containerd[1640]: time="2025-07-10T00:28:46.943568686Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.2: active requests=0, bytes read=51276688" Jul 10 00:28:46.954278 containerd[1640]: time="2025-07-10T00:28:46.954235743Z" level=info msg="ImageCreate event name:\"sha256:761b294e26556b58aabc85094a3d465389e6b141b7400aee732bd13400a6124a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:46.972794 containerd[1640]: time="2025-07-10T00:28:46.972753427Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:5d3ecdec3cbbe8f7009077102e35e8a2141161b59c548cf3f97829177677cbce\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:46.979221 containerd[1640]: time="2025-07-10T00:28:46.973228983Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\" with image id \"sha256:761b294e26556b58aabc85094a3d465389e6b141b7400aee732bd13400a6124a\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:5d3ecdec3cbbe8f7009077102e35e8a2141161b59c548cf3f97829177677cbce\", size \"52769359\" in 3.674337458s" Jul 10 00:28:46.979221 containerd[1640]: time="2025-07-10T00:28:46.973248088Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\" returns image reference \"sha256:761b294e26556b58aabc85094a3d465389e6b141b7400aee732bd13400a6124a\"" Jul 10 00:28:46.979221 containerd[1640]: time="2025-07-10T00:28:46.974047131Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.2\"" Jul 10 00:28:47.233111 containerd[1640]: time="2025-07-10T00:28:47.233083109Z" level=info msg="CreateContainer within sandbox \"53803ab3ac7d3053c9196e001fef8e60712767b15aed9d310d5fc83900f11072\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Jul 10 00:28:47.421430 containerd[1640]: time="2025-07-10T00:28:47.421391069Z" level=info msg="Container 042fac553317ab0650808e9c78a425b2f8c7da3e6297971f1beddfaeeb2a360e: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:28:47.426761 containerd[1640]: time="2025-07-10T00:28:47.426743082Z" level=info msg="CreateContainer within sandbox \"53803ab3ac7d3053c9196e001fef8e60712767b15aed9d310d5fc83900f11072\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"042fac553317ab0650808e9c78a425b2f8c7da3e6297971f1beddfaeeb2a360e\"" Jul 10 00:28:47.430944 containerd[1640]: time="2025-07-10T00:28:47.430354055Z" level=info msg="StartContainer for \"042fac553317ab0650808e9c78a425b2f8c7da3e6297971f1beddfaeeb2a360e\"" Jul 10 00:28:47.430944 containerd[1640]: time="2025-07-10T00:28:47.430891906Z" level=info msg="connecting to shim 042fac553317ab0650808e9c78a425b2f8c7da3e6297971f1beddfaeeb2a360e" address="unix:///run/containerd/s/aba001255bed2b76ccc0f0c8e61b2fe915144f94fedbc3b51613214e6ef11a1b" protocol=ttrpc version=3 Jul 10 00:28:47.448247 systemd[1]: Started cri-containerd-042fac553317ab0650808e9c78a425b2f8c7da3e6297971f1beddfaeeb2a360e.scope - libcontainer container 042fac553317ab0650808e9c78a425b2f8c7da3e6297971f1beddfaeeb2a360e. Jul 10 00:28:47.486128 containerd[1640]: time="2025-07-10T00:28:47.486013803Z" level=info msg="StartContainer for \"042fac553317ab0650808e9c78a425b2f8c7da3e6297971f1beddfaeeb2a360e\" returns successfully" Jul 10 00:28:47.683522 kubelet[2915]: I0710 00:28:47.683145 2915 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-84bd499cb-6t459" podStartSLOduration=24.505580674 podStartE2EDuration="29.676614722s" podCreationTimestamp="2025-07-10 00:28:18 +0000 UTC" firstStartedPulling="2025-07-10 00:28:41.802877403 +0000 UTC m=+40.547657616" lastFinishedPulling="2025-07-10 00:28:46.973911442 +0000 UTC m=+45.718691664" observedRunningTime="2025-07-10 00:28:47.674810547 +0000 UTC m=+46.419590777" watchObservedRunningTime="2025-07-10 00:28:47.676614722 +0000 UTC m=+46.421394952" Jul 10 00:28:47.683522 kubelet[2915]: I0710 00:28:47.683327 2915 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-7c75d6c687-wlhzf" podStartSLOduration=30.312002536 podStartE2EDuration="32.683320404s" podCreationTimestamp="2025-07-10 00:28:15 +0000 UTC" firstStartedPulling="2025-07-10 00:28:40.926699454 +0000 UTC m=+39.671479666" lastFinishedPulling="2025-07-10 00:28:43.298017321 +0000 UTC m=+42.042797534" observedRunningTime="2025-07-10 00:28:43.690311599 +0000 UTC m=+42.435091821" watchObservedRunningTime="2025-07-10 00:28:47.683320404 +0000 UTC m=+46.428100619" Jul 10 00:28:47.875140 containerd[1640]: time="2025-07-10T00:28:47.875057116Z" level=info msg="TaskExit event in podsandbox handler container_id:\"042fac553317ab0650808e9c78a425b2f8c7da3e6297971f1beddfaeeb2a360e\" id:\"72fdb220b8754e8b0f58954a9af83be1e8854ddf23b2ad877226c48cffaa213d\" pid:5129 exited_at:{seconds:1752107327 nanos:797160830}" Jul 10 00:28:48.383712 containerd[1640]: time="2025-07-10T00:28:48.383685750Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:48.384247 containerd[1640]: time="2025-07-10T00:28:48.384230682Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.2: active requests=0, bytes read=8759190" Jul 10 00:28:48.384281 containerd[1640]: time="2025-07-10T00:28:48.384266363Z" level=info msg="ImageCreate event name:\"sha256:c7fd1cc652979d89a51bbcc125e28e90c9815c0bd8f922a5bd36eed4e1927c6d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:48.387570 containerd[1640]: time="2025-07-10T00:28:48.385369896Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:e570128aa8067a2f06b96d3cc98afa2e0a4b9790b435ee36ca051c8e72aeb8d0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:48.387570 containerd[1640]: time="2025-07-10T00:28:48.385648065Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.30.2\" with image id \"sha256:c7fd1cc652979d89a51bbcc125e28e90c9815c0bd8f922a5bd36eed4e1927c6d\", repo tag \"ghcr.io/flatcar/calico/csi:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:e570128aa8067a2f06b96d3cc98afa2e0a4b9790b435ee36ca051c8e72aeb8d0\", size \"10251893\" in 1.411585381s" Jul 10 00:28:48.387570 containerd[1640]: time="2025-07-10T00:28:48.385662293Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.2\" returns image reference \"sha256:c7fd1cc652979d89a51bbcc125e28e90c9815c0bd8f922a5bd36eed4e1927c6d\"" Jul 10 00:28:48.398716 containerd[1640]: time="2025-07-10T00:28:48.397218447Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.2\"" Jul 10 00:28:48.414761 containerd[1640]: time="2025-07-10T00:28:48.414742964Z" level=info msg="CreateContainer within sandbox \"d12f1801782c97a0a45de4f7a5eb33bafeee4bce3484d0804d324fea61756a57\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Jul 10 00:28:48.427021 containerd[1640]: time="2025-07-10T00:28:48.426999861Z" level=info msg="Container 294c7f86d7b6333fc28e099644cea33a834ed11b41d79c2c76ddfd7985d9aa04: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:28:48.430191 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3678658506.mount: Deactivated successfully. Jul 10 00:28:48.445465 containerd[1640]: time="2025-07-10T00:28:48.445440469Z" level=info msg="CreateContainer within sandbox \"d12f1801782c97a0a45de4f7a5eb33bafeee4bce3484d0804d324fea61756a57\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"294c7f86d7b6333fc28e099644cea33a834ed11b41d79c2c76ddfd7985d9aa04\"" Jul 10 00:28:48.445995 containerd[1640]: time="2025-07-10T00:28:48.445759846Z" level=info msg="StartContainer for \"294c7f86d7b6333fc28e099644cea33a834ed11b41d79c2c76ddfd7985d9aa04\"" Jul 10 00:28:48.446784 containerd[1640]: time="2025-07-10T00:28:48.446769135Z" level=info msg="connecting to shim 294c7f86d7b6333fc28e099644cea33a834ed11b41d79c2c76ddfd7985d9aa04" address="unix:///run/containerd/s/b8239e5da780f64771a109b606a715fb5b26db6b0cbe0ac05bcab5f0a7ded206" protocol=ttrpc version=3 Jul 10 00:28:48.462643 systemd[1]: Started cri-containerd-294c7f86d7b6333fc28e099644cea33a834ed11b41d79c2c76ddfd7985d9aa04.scope - libcontainer container 294c7f86d7b6333fc28e099644cea33a834ed11b41d79c2c76ddfd7985d9aa04. Jul 10 00:28:48.488230 containerd[1640]: time="2025-07-10T00:28:48.488178233Z" level=info msg="StartContainer for \"294c7f86d7b6333fc28e099644cea33a834ed11b41d79c2c76ddfd7985d9aa04\" returns successfully" Jul 10 00:28:48.879808 containerd[1640]: time="2025-07-10T00:28:48.879741462Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:48.880019 containerd[1640]: time="2025-07-10T00:28:48.880007854Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.2: active requests=0, bytes read=77" Jul 10 00:28:48.881096 containerd[1640]: time="2025-07-10T00:28:48.881078847Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.30.2\" with image id \"sha256:5509118eed617ef04ca00f5a095bfd0a4cd1cf69edcfcf9bedf0edb641be51dd\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:ec6b10660962e7caad70c47755049fad68f9fc2f7064e8bc7cb862583e02cc2b\", size \"48810696\" in 483.83172ms" Jul 10 00:28:48.881129 containerd[1640]: time="2025-07-10T00:28:48.881097848Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.2\" returns image reference \"sha256:5509118eed617ef04ca00f5a095bfd0a4cd1cf69edcfcf9bedf0edb641be51dd\"" Jul 10 00:28:48.888549 containerd[1640]: time="2025-07-10T00:28:48.888532195Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.2\"" Jul 10 00:28:48.890369 containerd[1640]: time="2025-07-10T00:28:48.890348955Z" level=info msg="CreateContainer within sandbox \"b00bb58c3c648815d4d8eb8c2dc340380592eb466066c9c6444054b407de8c7d\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Jul 10 00:28:48.896180 containerd[1640]: time="2025-07-10T00:28:48.896018088Z" level=info msg="Container 7c751499fef0f6bd1442da32abf5e6cf6a67644a9ccc9814e5985db75ffdd84f: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:28:48.908398 containerd[1640]: time="2025-07-10T00:28:48.908380605Z" level=info msg="CreateContainer within sandbox \"b00bb58c3c648815d4d8eb8c2dc340380592eb466066c9c6444054b407de8c7d\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"7c751499fef0f6bd1442da32abf5e6cf6a67644a9ccc9814e5985db75ffdd84f\"" Jul 10 00:28:48.908681 containerd[1640]: time="2025-07-10T00:28:48.908628928Z" level=info msg="StartContainer for \"7c751499fef0f6bd1442da32abf5e6cf6a67644a9ccc9814e5985db75ffdd84f\"" Jul 10 00:28:48.909454 containerd[1640]: time="2025-07-10T00:28:48.909441416Z" level=info msg="connecting to shim 7c751499fef0f6bd1442da32abf5e6cf6a67644a9ccc9814e5985db75ffdd84f" address="unix:///run/containerd/s/afb4684c686eed7e2609914aa4a51f4fd0713a8a5fae1ce685f003dcfa9104c7" protocol=ttrpc version=3 Jul 10 00:28:48.925261 systemd[1]: Started cri-containerd-7c751499fef0f6bd1442da32abf5e6cf6a67644a9ccc9814e5985db75ffdd84f.scope - libcontainer container 7c751499fef0f6bd1442da32abf5e6cf6a67644a9ccc9814e5985db75ffdd84f. Jul 10 00:28:48.962850 containerd[1640]: time="2025-07-10T00:28:48.962802357Z" level=info msg="StartContainer for \"7c751499fef0f6bd1442da32abf5e6cf6a67644a9ccc9814e5985db75ffdd84f\" returns successfully" Jul 10 00:28:49.705214 kubelet[2915]: I0710 00:28:49.705142 2915 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-7c75d6c687-8jqvg" podStartSLOduration=27.723824462 podStartE2EDuration="34.703641682s" podCreationTimestamp="2025-07-10 00:28:15 +0000 UTC" firstStartedPulling="2025-07-10 00:28:41.907875062 +0000 UTC m=+40.652655275" lastFinishedPulling="2025-07-10 00:28:48.887692282 +0000 UTC m=+47.632472495" observedRunningTime="2025-07-10 00:28:49.699237291 +0000 UTC m=+48.444017520" watchObservedRunningTime="2025-07-10 00:28:49.703641682 +0000 UTC m=+48.448421913" Jul 10 00:28:50.693894 kubelet[2915]: I0710 00:28:50.693865 2915 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jul 10 00:28:53.964133 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3789322077.mount: Deactivated successfully. Jul 10 00:28:54.581583 containerd[1640]: time="2025-07-10T00:28:54.581552248Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:54.584241 containerd[1640]: time="2025-07-10T00:28:54.584224154Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.2: active requests=0, bytes read=66352308" Jul 10 00:28:54.585660 containerd[1640]: time="2025-07-10T00:28:54.585242090Z" level=info msg="ImageCreate event name:\"sha256:dc4ea8b409b85d2f118bb4677ad3d34b57e7b01d488c9f019f7073bb58b2162b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:54.586525 containerd[1640]: time="2025-07-10T00:28:54.586488760Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane@sha256:a2b761fd93d824431ad93e59e8e670cdf00b478f4b532145297e1e67f2768305\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:54.587059 containerd[1640]: time="2025-07-10T00:28:54.586974890Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/goldmane:v3.30.2\" with image id \"sha256:dc4ea8b409b85d2f118bb4677ad3d34b57e7b01d488c9f019f7073bb58b2162b\", repo tag \"ghcr.io/flatcar/calico/goldmane:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/goldmane@sha256:a2b761fd93d824431ad93e59e8e670cdf00b478f4b532145297e1e67f2768305\", size \"66352154\" in 5.698255989s" Jul 10 00:28:54.587059 containerd[1640]: time="2025-07-10T00:28:54.586991629Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.2\" returns image reference \"sha256:dc4ea8b409b85d2f118bb4677ad3d34b57e7b01d488c9f019f7073bb58b2162b\"" Jul 10 00:28:54.589758 containerd[1640]: time="2025-07-10T00:28:54.587860077Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\"" Jul 10 00:28:54.595576 containerd[1640]: time="2025-07-10T00:28:54.595515850Z" level=info msg="CreateContainer within sandbox \"a406a4e994665bce6898b2f2c15a7d025b0c29070bfc5b3b5417aa375134d460\" for container &ContainerMetadata{Name:goldmane,Attempt:0,}" Jul 10 00:28:54.603936 containerd[1640]: time="2025-07-10T00:28:54.603913580Z" level=info msg="Container b18a9b96dead6a2bda009c3f88ab746b32b9bb7e00cd928ce07a6747bb638694: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:28:54.626060 containerd[1640]: time="2025-07-10T00:28:54.626025108Z" level=info msg="CreateContainer within sandbox \"a406a4e994665bce6898b2f2c15a7d025b0c29070bfc5b3b5417aa375134d460\" for &ContainerMetadata{Name:goldmane,Attempt:0,} returns container id \"b18a9b96dead6a2bda009c3f88ab746b32b9bb7e00cd928ce07a6747bb638694\"" Jul 10 00:28:54.626360 containerd[1640]: time="2025-07-10T00:28:54.626349203Z" level=info msg="StartContainer for \"b18a9b96dead6a2bda009c3f88ab746b32b9bb7e00cd928ce07a6747bb638694\"" Jul 10 00:28:54.627547 containerd[1640]: time="2025-07-10T00:28:54.627437717Z" level=info msg="connecting to shim b18a9b96dead6a2bda009c3f88ab746b32b9bb7e00cd928ce07a6747bb638694" address="unix:///run/containerd/s/598468a299605de5131a3b6a3c2504616ebe6b04ba7a5fe4513b7453548db158" protocol=ttrpc version=3 Jul 10 00:28:54.654278 systemd[1]: Started cri-containerd-b18a9b96dead6a2bda009c3f88ab746b32b9bb7e00cd928ce07a6747bb638694.scope - libcontainer container b18a9b96dead6a2bda009c3f88ab746b32b9bb7e00cd928ce07a6747bb638694. Jul 10 00:28:54.764679 containerd[1640]: time="2025-07-10T00:28:54.764650739Z" level=info msg="StartContainer for \"b18a9b96dead6a2bda009c3f88ab746b32b9bb7e00cd928ce07a6747bb638694\" returns successfully" Jul 10 00:28:54.927291 containerd[1640]: time="2025-07-10T00:28:54.926932364Z" level=info msg="TaskExit event in podsandbox handler container_id:\"7342408a31cb64c6221572622b4fe6b7e010c365050de1a3cf3bc111989059fe\" id:\"c57d795cd93e130ee81121fdd2c91f3044f8b08e499ada418699862e0dfafa2b\" pid:5237 exited_at:{seconds:1752107334 nanos:895622283}" Jul 10 00:28:55.969151 kubelet[2915]: I0710 00:28:55.969118 2915 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jul 10 00:28:56.197571 containerd[1640]: time="2025-07-10T00:28:56.197542871Z" level=info msg="TaskExit event in podsandbox handler container_id:\"b18a9b96dead6a2bda009c3f88ab746b32b9bb7e00cd928ce07a6747bb638694\" id:\"4635bf621d1f05afcfc2234c8fed3f039eea98245682959687fa3651c39a7b80\" pid:5307 exit_status:1 exited_at:{seconds:1752107336 nanos:197201161}" Jul 10 00:28:56.250708 kubelet[2915]: I0710 00:28:56.247929 2915 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/goldmane-58fd7646b9-52ptt" podStartSLOduration=27.584229356 podStartE2EDuration="39.22635206s" podCreationTimestamp="2025-07-10 00:28:17 +0000 UTC" firstStartedPulling="2025-07-10 00:28:42.945620646 +0000 UTC m=+41.690400859" lastFinishedPulling="2025-07-10 00:28:54.58774335 +0000 UTC m=+53.332523563" observedRunningTime="2025-07-10 00:28:55.731180791 +0000 UTC m=+54.475961008" watchObservedRunningTime="2025-07-10 00:28:56.22635206 +0000 UTC m=+54.971132277" Jul 10 00:28:56.267870 containerd[1640]: time="2025-07-10T00:28:56.267490347Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:56.267959 containerd[1640]: time="2025-07-10T00:28:56.267948707Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2: active requests=0, bytes read=14703784" Jul 10 00:28:56.269175 containerd[1640]: time="2025-07-10T00:28:56.269159869Z" level=info msg="ImageCreate event name:\"sha256:9e48822a4fe26f4ed9231b361fdd1357ea3567f1fc0a8db4d616622fe570a866\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:56.269998 containerd[1640]: time="2025-07-10T00:28:56.269986709Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:8fec2de12dfa51bae89d941938a07af2598eb8bfcab55d0dded1d9c193d7b99f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 10 00:28:56.270389 containerd[1640]: time="2025-07-10T00:28:56.270374018Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\" with image id \"sha256:9e48822a4fe26f4ed9231b361fdd1357ea3567f1fc0a8db4d616622fe570a866\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:8fec2de12dfa51bae89d941938a07af2598eb8bfcab55d0dded1d9c193d7b99f\", size \"16196439\" in 1.68249756s" Jul 10 00:28:56.270420 containerd[1640]: time="2025-07-10T00:28:56.270390745Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\" returns image reference \"sha256:9e48822a4fe26f4ed9231b361fdd1357ea3567f1fc0a8db4d616622fe570a866\"" Jul 10 00:28:56.317382 containerd[1640]: time="2025-07-10T00:28:56.317357343Z" level=info msg="CreateContainer within sandbox \"d12f1801782c97a0a45de4f7a5eb33bafeee4bce3484d0804d324fea61756a57\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Jul 10 00:28:56.326364 containerd[1640]: time="2025-07-10T00:28:56.326348073Z" level=info msg="Container 68ad8d726feb0d7592f9f1c723cf4627a1e0f4d7329d483a6527b94d0e25350a: CDI devices from CRI Config.CDIDevices: []" Jul 10 00:28:56.329058 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2409373052.mount: Deactivated successfully. Jul 10 00:28:56.333901 containerd[1640]: time="2025-07-10T00:28:56.333885066Z" level=info msg="CreateContainer within sandbox \"d12f1801782c97a0a45de4f7a5eb33bafeee4bce3484d0804d324fea61756a57\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"68ad8d726feb0d7592f9f1c723cf4627a1e0f4d7329d483a6527b94d0e25350a\"" Jul 10 00:28:56.335287 containerd[1640]: time="2025-07-10T00:28:56.335261391Z" level=info msg="StartContainer for \"68ad8d726feb0d7592f9f1c723cf4627a1e0f4d7329d483a6527b94d0e25350a\"" Jul 10 00:28:56.336193 containerd[1640]: time="2025-07-10T00:28:56.336159453Z" level=info msg="connecting to shim 68ad8d726feb0d7592f9f1c723cf4627a1e0f4d7329d483a6527b94d0e25350a" address="unix:///run/containerd/s/b8239e5da780f64771a109b606a715fb5b26db6b0cbe0ac05bcab5f0a7ded206" protocol=ttrpc version=3 Jul 10 00:28:56.393240 systemd[1]: Started cri-containerd-68ad8d726feb0d7592f9f1c723cf4627a1e0f4d7329d483a6527b94d0e25350a.scope - libcontainer container 68ad8d726feb0d7592f9f1c723cf4627a1e0f4d7329d483a6527b94d0e25350a. Jul 10 00:28:56.429104 containerd[1640]: time="2025-07-10T00:28:56.428737354Z" level=info msg="StartContainer for \"68ad8d726feb0d7592f9f1c723cf4627a1e0f4d7329d483a6527b94d0e25350a\" returns successfully" Jul 10 00:28:56.738500 kubelet[2915]: I0710 00:28:56.738354 2915 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-pzv7w" podStartSLOduration=24.271991454 podStartE2EDuration="38.738339881s" podCreationTimestamp="2025-07-10 00:28:18 +0000 UTC" firstStartedPulling="2025-07-10 00:28:41.827078273 +0000 UTC m=+40.571858485" lastFinishedPulling="2025-07-10 00:28:56.293426696 +0000 UTC m=+55.038206912" observedRunningTime="2025-07-10 00:28:56.73258231 +0000 UTC m=+55.477362532" watchObservedRunningTime="2025-07-10 00:28:56.738339881 +0000 UTC m=+55.483120097" Jul 10 00:28:56.752098 kubelet[2915]: I0710 00:28:56.747647 2915 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Jul 10 00:28:56.753723 kubelet[2915]: I0710 00:28:56.753706 2915 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Jul 10 00:28:56.853540 containerd[1640]: time="2025-07-10T00:28:56.852948029Z" level=info msg="TaskExit event in podsandbox handler container_id:\"b18a9b96dead6a2bda009c3f88ab746b32b9bb7e00cd928ce07a6747bb638694\" id:\"d3a0a1ee84b153ea9a1d39f2dd77eb7f17247c83280d954a85cdecfe3203a3e2\" pid:5375 exit_status:1 exited_at:{seconds:1752107336 nanos:852747734}" Jul 10 00:28:57.785150 containerd[1640]: time="2025-07-10T00:28:57.785110827Z" level=info msg="TaskExit event in podsandbox handler container_id:\"b18a9b96dead6a2bda009c3f88ab746b32b9bb7e00cd928ce07a6747bb638694\" id:\"7194730d329eaca792c2251dbaf4169428f9c7945c549dc3f3f89e979323b8d8\" pid:5398 exit_status:1 exited_at:{seconds:1752107337 nanos:784906851}" Jul 10 00:28:59.623732 kubelet[2915]: I0710 00:28:59.623698 2915 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jul 10 00:29:12.117012 systemd[1]: Started sshd@8-139.178.70.108:22-139.178.68.195:50818.service - OpenSSH per-connection server daemon (139.178.68.195:50818). Jul 10 00:29:12.225884 sshd[5431]: Accepted publickey for core from 139.178.68.195 port 50818 ssh2: RSA SHA256:4dbLs3K8zeCUdpJVvc+oLD6Wxu1uro36XJoOlJl6xXw Jul 10 00:29:12.230587 sshd-session[5431]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:29:12.241823 systemd-logind[1614]: New session 10 of user core. Jul 10 00:29:12.254311 systemd[1]: Started session-10.scope - Session 10 of User core. Jul 10 00:29:12.748546 sshd[5433]: Connection closed by 139.178.68.195 port 50818 Jul 10 00:29:12.748869 sshd-session[5431]: pam_unix(sshd:session): session closed for user core Jul 10 00:29:12.753550 systemd[1]: sshd@8-139.178.70.108:22-139.178.68.195:50818.service: Deactivated successfully. Jul 10 00:29:12.754784 systemd[1]: session-10.scope: Deactivated successfully. Jul 10 00:29:12.755832 systemd-logind[1614]: Session 10 logged out. Waiting for processes to exit. Jul 10 00:29:12.756850 systemd-logind[1614]: Removed session 10. Jul 10 00:29:13.491868 containerd[1640]: time="2025-07-10T00:29:13.491837311Z" level=info msg="TaskExit event in podsandbox handler container_id:\"042fac553317ab0650808e9c78a425b2f8c7da3e6297971f1beddfaeeb2a360e\" id:\"b1466f065998d6d0a8e196dc299780a6922fe42f9daf2a9304be427492522bb8\" pid:5456 exited_at:{seconds:1752107353 nanos:472326995}" Jul 10 00:29:17.760623 systemd[1]: Started sshd@9-139.178.70.108:22-139.178.68.195:50834.service - OpenSSH per-connection server daemon (139.178.68.195:50834). Jul 10 00:29:17.883040 sshd[5474]: Accepted publickey for core from 139.178.68.195 port 50834 ssh2: RSA SHA256:4dbLs3K8zeCUdpJVvc+oLD6Wxu1uro36XJoOlJl6xXw Jul 10 00:29:17.885973 sshd-session[5474]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:29:17.890384 systemd-logind[1614]: New session 11 of user core. Jul 10 00:29:17.896288 systemd[1]: Started session-11.scope - Session 11 of User core. Jul 10 00:29:18.237997 sshd[5476]: Connection closed by 139.178.68.195 port 50834 Jul 10 00:29:18.239042 sshd-session[5474]: pam_unix(sshd:session): session closed for user core Jul 10 00:29:18.242930 systemd[1]: sshd@9-139.178.70.108:22-139.178.68.195:50834.service: Deactivated successfully. Jul 10 00:29:18.246656 systemd[1]: session-11.scope: Deactivated successfully. Jul 10 00:29:18.248075 systemd-logind[1614]: Session 11 logged out. Waiting for processes to exit. Jul 10 00:29:18.249361 systemd-logind[1614]: Removed session 11. Jul 10 00:29:21.667567 containerd[1640]: time="2025-07-10T00:29:21.667531925Z" level=info msg="TaskExit event in podsandbox handler container_id:\"b18a9b96dead6a2bda009c3f88ab746b32b9bb7e00cd928ce07a6747bb638694\" id:\"93f7bee6137b25ceec013754d99f4d1ee79d1fe56e66c4eee5f7f4e72f5601d4\" pid:5500 exited_at:{seconds:1752107361 nanos:667115796}" Jul 10 00:29:23.253161 systemd[1]: Started sshd@10-139.178.70.108:22-139.178.68.195:60526.service - OpenSSH per-connection server daemon (139.178.68.195:60526). Jul 10 00:29:23.380006 sshd[5515]: Accepted publickey for core from 139.178.68.195 port 60526 ssh2: RSA SHA256:4dbLs3K8zeCUdpJVvc+oLD6Wxu1uro36XJoOlJl6xXw Jul 10 00:29:23.381018 sshd-session[5515]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:29:23.384341 systemd-logind[1614]: New session 12 of user core. Jul 10 00:29:23.391271 systemd[1]: Started session-12.scope - Session 12 of User core. Jul 10 00:29:23.742815 sshd[5517]: Connection closed by 139.178.68.195 port 60526 Jul 10 00:29:23.745545 sshd-session[5515]: pam_unix(sshd:session): session closed for user core Jul 10 00:29:23.753696 systemd[1]: sshd@10-139.178.70.108:22-139.178.68.195:60526.service: Deactivated successfully. Jul 10 00:29:23.755511 systemd[1]: session-12.scope: Deactivated successfully. Jul 10 00:29:23.756866 systemd-logind[1614]: Session 12 logged out. Waiting for processes to exit. Jul 10 00:29:23.758340 systemd[1]: Started sshd@11-139.178.70.108:22-139.178.68.195:60536.service - OpenSSH per-connection server daemon (139.178.68.195:60536). Jul 10 00:29:23.759508 systemd-logind[1614]: Removed session 12. Jul 10 00:29:23.808847 sshd[5530]: Accepted publickey for core from 139.178.68.195 port 60536 ssh2: RSA SHA256:4dbLs3K8zeCUdpJVvc+oLD6Wxu1uro36XJoOlJl6xXw Jul 10 00:29:23.810148 sshd-session[5530]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:29:23.813246 systemd-logind[1614]: New session 13 of user core. Jul 10 00:29:23.821276 systemd[1]: Started session-13.scope - Session 13 of User core. Jul 10 00:29:23.965133 sshd[5532]: Connection closed by 139.178.68.195 port 60536 Jul 10 00:29:23.966135 sshd-session[5530]: pam_unix(sshd:session): session closed for user core Jul 10 00:29:23.973541 systemd[1]: sshd@11-139.178.70.108:22-139.178.68.195:60536.service: Deactivated successfully. Jul 10 00:29:23.975617 systemd[1]: session-13.scope: Deactivated successfully. Jul 10 00:29:23.976251 systemd-logind[1614]: Session 13 logged out. Waiting for processes to exit. Jul 10 00:29:23.979695 systemd[1]: Started sshd@12-139.178.70.108:22-139.178.68.195:60546.service - OpenSSH per-connection server daemon (139.178.68.195:60546). Jul 10 00:29:23.981150 systemd-logind[1614]: Removed session 13. Jul 10 00:29:24.040137 sshd[5542]: Accepted publickey for core from 139.178.68.195 port 60546 ssh2: RSA SHA256:4dbLs3K8zeCUdpJVvc+oLD6Wxu1uro36XJoOlJl6xXw Jul 10 00:29:24.040473 sshd-session[5542]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:29:24.045004 systemd-logind[1614]: New session 14 of user core. Jul 10 00:29:24.050389 systemd[1]: Started session-14.scope - Session 14 of User core. Jul 10 00:29:24.490929 sshd[5544]: Connection closed by 139.178.68.195 port 60546 Jul 10 00:29:24.494011 systemd[1]: sshd@12-139.178.70.108:22-139.178.68.195:60546.service: Deactivated successfully. Jul 10 00:29:24.491317 sshd-session[5542]: pam_unix(sshd:session): session closed for user core Jul 10 00:29:24.495709 systemd[1]: session-14.scope: Deactivated successfully. Jul 10 00:29:24.496616 systemd-logind[1614]: Session 14 logged out. Waiting for processes to exit. Jul 10 00:29:24.499675 systemd-logind[1614]: Removed session 14. Jul 10 00:29:25.098985 containerd[1640]: time="2025-07-10T00:29:25.098954074Z" level=info msg="TaskExit event in podsandbox handler container_id:\"7342408a31cb64c6221572622b4fe6b7e010c365050de1a3cf3bc111989059fe\" id:\"8d0943fed9908fdce21c3088b9eaa43c4992e6322b6f56df3e297cbd88850dac\" pid:5563 exited_at:{seconds:1752107365 nanos:98630902}" Jul 10 00:29:27.805937 containerd[1640]: time="2025-07-10T00:29:27.805911349Z" level=info msg="TaskExit event in podsandbox handler container_id:\"b18a9b96dead6a2bda009c3f88ab746b32b9bb7e00cd928ce07a6747bb638694\" id:\"5fcf5e7869e05727a6ec9e2565f58cfffb5bb271b838b8ac87877cf42c47c08e\" pid:5590 exited_at:{seconds:1752107367 nanos:805688884}" Jul 10 00:29:29.502007 systemd[1]: Started sshd@13-139.178.70.108:22-139.178.68.195:37436.service - OpenSSH per-connection server daemon (139.178.68.195:37436). Jul 10 00:29:29.647884 sshd[5605]: Accepted publickey for core from 139.178.68.195 port 37436 ssh2: RSA SHA256:4dbLs3K8zeCUdpJVvc+oLD6Wxu1uro36XJoOlJl6xXw Jul 10 00:29:29.650105 sshd-session[5605]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:29:29.654859 systemd-logind[1614]: New session 15 of user core. Jul 10 00:29:29.659275 systemd[1]: Started session-15.scope - Session 15 of User core. Jul 10 00:29:30.545621 sshd[5607]: Connection closed by 139.178.68.195 port 37436 Jul 10 00:29:30.545254 sshd-session[5605]: pam_unix(sshd:session): session closed for user core Jul 10 00:29:30.550821 systemd[1]: sshd@13-139.178.70.108:22-139.178.68.195:37436.service: Deactivated successfully. Jul 10 00:29:30.552705 systemd[1]: session-15.scope: Deactivated successfully. Jul 10 00:29:30.553578 systemd-logind[1614]: Session 15 logged out. Waiting for processes to exit. Jul 10 00:29:30.554839 systemd-logind[1614]: Removed session 15. Jul 10 00:29:33.924402 containerd[1640]: time="2025-07-10T00:29:33.924346774Z" level=info msg="TaskExit event in podsandbox handler container_id:\"042fac553317ab0650808e9c78a425b2f8c7da3e6297971f1beddfaeeb2a360e\" id:\"4a167edcd904d4afd49e731bf3639b0e0632682ce20873f390342f05056e5801\" pid:5631 exited_at:{seconds:1752107373 nanos:923969991}" Jul 10 00:29:35.553662 systemd[1]: Started sshd@14-139.178.70.108:22-139.178.68.195:37444.service - OpenSSH per-connection server daemon (139.178.68.195:37444). Jul 10 00:29:35.628071 sshd[5641]: Accepted publickey for core from 139.178.68.195 port 37444 ssh2: RSA SHA256:4dbLs3K8zeCUdpJVvc+oLD6Wxu1uro36XJoOlJl6xXw Jul 10 00:29:35.628740 sshd-session[5641]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:29:35.633815 systemd-logind[1614]: New session 16 of user core. Jul 10 00:29:35.637257 systemd[1]: Started session-16.scope - Session 16 of User core. Jul 10 00:29:35.816749 sshd[5643]: Connection closed by 139.178.68.195 port 37444 Jul 10 00:29:35.817295 sshd-session[5641]: pam_unix(sshd:session): session closed for user core Jul 10 00:29:35.819373 systemd[1]: sshd@14-139.178.70.108:22-139.178.68.195:37444.service: Deactivated successfully. Jul 10 00:29:35.820522 systemd[1]: session-16.scope: Deactivated successfully. Jul 10 00:29:35.821342 systemd-logind[1614]: Session 16 logged out. Waiting for processes to exit. Jul 10 00:29:35.822115 systemd-logind[1614]: Removed session 16. Jul 10 00:29:40.827143 systemd[1]: Started sshd@15-139.178.70.108:22-139.178.68.195:46036.service - OpenSSH per-connection server daemon (139.178.68.195:46036). Jul 10 00:29:40.897152 sshd[5659]: Accepted publickey for core from 139.178.68.195 port 46036 ssh2: RSA SHA256:4dbLs3K8zeCUdpJVvc+oLD6Wxu1uro36XJoOlJl6xXw Jul 10 00:29:40.898254 sshd-session[5659]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:29:40.901434 systemd-logind[1614]: New session 17 of user core. Jul 10 00:29:40.910247 systemd[1]: Started session-17.scope - Session 17 of User core. Jul 10 00:29:41.085441 sshd[5662]: Connection closed by 139.178.68.195 port 46036 Jul 10 00:29:41.086004 sshd-session[5659]: pam_unix(sshd:session): session closed for user core Jul 10 00:29:41.091493 systemd[1]: sshd@15-139.178.70.108:22-139.178.68.195:46036.service: Deactivated successfully. Jul 10 00:29:41.092851 systemd[1]: session-17.scope: Deactivated successfully. Jul 10 00:29:41.093555 systemd-logind[1614]: Session 17 logged out. Waiting for processes to exit. Jul 10 00:29:41.094329 systemd-logind[1614]: Removed session 17. Jul 10 00:29:43.308499 containerd[1640]: time="2025-07-10T00:29:43.291348359Z" level=info msg="TaskExit event in podsandbox handler container_id:\"042fac553317ab0650808e9c78a425b2f8c7da3e6297971f1beddfaeeb2a360e\" id:\"c1ff07162ca9ceb3e704592e1be176f77dc4bb5a7424fac36e746d538d3ccc17\" pid:5689 exited_at:{seconds:1752107383 nanos:287681677}" Jul 10 00:29:46.100534 systemd[1]: Started sshd@16-139.178.70.108:22-139.178.68.195:46050.service - OpenSSH per-connection server daemon (139.178.68.195:46050). Jul 10 00:29:46.246681 sshd[5701]: Accepted publickey for core from 139.178.68.195 port 46050 ssh2: RSA SHA256:4dbLs3K8zeCUdpJVvc+oLD6Wxu1uro36XJoOlJl6xXw Jul 10 00:29:46.248820 sshd-session[5701]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:29:46.253341 systemd-logind[1614]: New session 18 of user core. Jul 10 00:29:46.257269 systemd[1]: Started session-18.scope - Session 18 of User core. Jul 10 00:29:46.729039 sshd[5703]: Connection closed by 139.178.68.195 port 46050 Jul 10 00:29:46.729659 sshd-session[5701]: pam_unix(sshd:session): session closed for user core Jul 10 00:29:46.737681 systemd[1]: sshd@16-139.178.70.108:22-139.178.68.195:46050.service: Deactivated successfully. Jul 10 00:29:46.739065 systemd[1]: session-18.scope: Deactivated successfully. Jul 10 00:29:46.741249 systemd-logind[1614]: Session 18 logged out. Waiting for processes to exit. Jul 10 00:29:46.743655 systemd[1]: Started sshd@17-139.178.70.108:22-139.178.68.195:46062.service - OpenSSH per-connection server daemon (139.178.68.195:46062). Jul 10 00:29:46.745320 systemd-logind[1614]: Removed session 18. Jul 10 00:29:46.795842 sshd[5715]: Accepted publickey for core from 139.178.68.195 port 46062 ssh2: RSA SHA256:4dbLs3K8zeCUdpJVvc+oLD6Wxu1uro36XJoOlJl6xXw Jul 10 00:29:46.796672 sshd-session[5715]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:29:46.801874 systemd-logind[1614]: New session 19 of user core. Jul 10 00:29:46.804240 systemd[1]: Started session-19.scope - Session 19 of User core. Jul 10 00:29:47.151176 sshd[5717]: Connection closed by 139.178.68.195 port 46062 Jul 10 00:29:47.152286 sshd-session[5715]: pam_unix(sshd:session): session closed for user core Jul 10 00:29:47.159056 systemd[1]: sshd@17-139.178.70.108:22-139.178.68.195:46062.service: Deactivated successfully. Jul 10 00:29:47.160393 systemd[1]: session-19.scope: Deactivated successfully. Jul 10 00:29:47.161377 systemd-logind[1614]: Session 19 logged out. Waiting for processes to exit. Jul 10 00:29:47.162826 systemd[1]: Started sshd@18-139.178.70.108:22-139.178.68.195:46078.service - OpenSSH per-connection server daemon (139.178.68.195:46078). Jul 10 00:29:47.163865 systemd-logind[1614]: Removed session 19. Jul 10 00:29:47.236972 sshd[5727]: Accepted publickey for core from 139.178.68.195 port 46078 ssh2: RSA SHA256:4dbLs3K8zeCUdpJVvc+oLD6Wxu1uro36XJoOlJl6xXw Jul 10 00:29:47.237940 sshd-session[5727]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:29:47.242081 systemd-logind[1614]: New session 20 of user core. Jul 10 00:29:47.246293 systemd[1]: Started session-20.scope - Session 20 of User core. Jul 10 00:29:49.796575 sshd[5729]: Connection closed by 139.178.68.195 port 46078 Jul 10 00:29:49.890704 systemd[1]: Started sshd@19-139.178.70.108:22-139.178.68.195:60628.service - OpenSSH per-connection server daemon (139.178.68.195:60628). Jul 10 00:29:49.813400 sshd-session[5727]: pam_unix(sshd:session): session closed for user core Jul 10 00:29:49.891026 systemd[1]: sshd@18-139.178.70.108:22-139.178.68.195:46078.service: Deactivated successfully. Jul 10 00:29:49.893430 systemd[1]: session-20.scope: Deactivated successfully. Jul 10 00:29:49.893577 systemd[1]: session-20.scope: Consumed 362ms CPU time, 86.9M memory peak. Jul 10 00:29:49.894136 systemd-logind[1614]: Session 20 logged out. Waiting for processes to exit. Jul 10 00:29:49.896113 systemd-logind[1614]: Removed session 20. Jul 10 00:29:50.069883 sshd[5745]: Accepted publickey for core from 139.178.68.195 port 60628 ssh2: RSA SHA256:4dbLs3K8zeCUdpJVvc+oLD6Wxu1uro36XJoOlJl6xXw Jul 10 00:29:50.071726 sshd-session[5745]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:29:50.086214 systemd-logind[1614]: New session 21 of user core. Jul 10 00:29:50.092268 systemd[1]: Started session-21.scope - Session 21 of User core. Jul 10 00:29:52.594481 kubelet[2915]: E0710 00:29:52.587806 2915 kubelet.go:2512] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="2.597s" Jul 10 00:29:53.389082 containerd[1640]: time="2025-07-10T00:29:53.379972405Z" level=info msg="TaskExit event in podsandbox handler container_id:\"b18a9b96dead6a2bda009c3f88ab746b32b9bb7e00cd928ce07a6747bb638694\" id:\"13f7d72c2735371ee1abe46628d1a9cc0b6b84eb0d2a684c7c9cc10283932a8a\" pid:5783 exited_at:{seconds:1752107393 nanos:328786197}" Jul 10 00:29:55.351977 sshd[5750]: Connection closed by 139.178.68.195 port 60628 Jul 10 00:29:55.403065 sshd-session[5745]: pam_unix(sshd:session): session closed for user core Jul 10 00:29:55.479105 systemd[1]: sshd@19-139.178.70.108:22-139.178.68.195:60628.service: Deactivated successfully. Jul 10 00:29:55.480959 systemd[1]: session-21.scope: Deactivated successfully. Jul 10 00:29:55.482400 systemd[1]: session-21.scope: Consumed 1.041s CPU time, 55.1M memory peak. Jul 10 00:29:55.498133 systemd[1]: Started sshd@20-139.178.70.108:22-139.178.68.195:60632.service - OpenSSH per-connection server daemon (139.178.68.195:60632). Jul 10 00:29:55.498972 systemd-logind[1614]: Session 21 logged out. Waiting for processes to exit. Jul 10 00:29:55.503180 systemd-logind[1614]: Removed session 21. Jul 10 00:29:55.511996 containerd[1640]: time="2025-07-10T00:29:55.511971114Z" level=info msg="TaskExit event in podsandbox handler container_id:\"7342408a31cb64c6221572622b4fe6b7e010c365050de1a3cf3bc111989059fe\" id:\"fa1cd0b79fae7de85f394104d793970a91544ecd6ce3cc87da89b47be7d67d77\" pid:5810 exited_at:{seconds:1752107395 nanos:451220638}" Jul 10 00:29:55.727298 sshd[5830]: Accepted publickey for core from 139.178.68.195 port 60632 ssh2: RSA SHA256:4dbLs3K8zeCUdpJVvc+oLD6Wxu1uro36XJoOlJl6xXw Jul 10 00:29:55.728450 sshd-session[5830]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:29:55.735130 systemd-logind[1614]: New session 22 of user core. Jul 10 00:29:55.740248 systemd[1]: Started session-22.scope - Session 22 of User core. Jul 10 00:29:56.718826 sshd[5837]: Connection closed by 139.178.68.195 port 60632 Jul 10 00:29:56.731107 systemd[1]: sshd@20-139.178.70.108:22-139.178.68.195:60632.service: Deactivated successfully. Jul 10 00:29:56.723811 sshd-session[5830]: pam_unix(sshd:session): session closed for user core Jul 10 00:29:56.733622 systemd[1]: session-22.scope: Deactivated successfully. Jul 10 00:29:56.735062 systemd-logind[1614]: Session 22 logged out. Waiting for processes to exit. Jul 10 00:29:56.739270 systemd-logind[1614]: Removed session 22. Jul 10 00:30:01.738492 systemd[1]: Started sshd@21-139.178.70.108:22-139.178.68.195:47074.service - OpenSSH per-connection server daemon (139.178.68.195:47074). Jul 10 00:30:01.886797 sshd[5853]: Accepted publickey for core from 139.178.68.195 port 47074 ssh2: RSA SHA256:4dbLs3K8zeCUdpJVvc+oLD6Wxu1uro36XJoOlJl6xXw Jul 10 00:30:01.888777 sshd-session[5853]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:30:01.893584 systemd-logind[1614]: New session 23 of user core. Jul 10 00:30:01.899555 systemd[1]: Started session-23.scope - Session 23 of User core. Jul 10 00:30:02.689510 sshd[5855]: Connection closed by 139.178.68.195 port 47074 Jul 10 00:30:02.689713 sshd-session[5853]: pam_unix(sshd:session): session closed for user core Jul 10 00:30:02.696471 systemd[1]: sshd@21-139.178.70.108:22-139.178.68.195:47074.service: Deactivated successfully. Jul 10 00:30:02.697568 systemd[1]: session-23.scope: Deactivated successfully. Jul 10 00:30:02.698301 systemd-logind[1614]: Session 23 logged out. Waiting for processes to exit. Jul 10 00:30:02.699131 systemd-logind[1614]: Removed session 23. Jul 10 00:30:07.723455 systemd[1]: Started sshd@22-139.178.70.108:22-139.178.68.195:47084.service - OpenSSH per-connection server daemon (139.178.68.195:47084). Jul 10 00:30:07.840300 sshd[5869]: Accepted publickey for core from 139.178.68.195 port 47084 ssh2: RSA SHA256:4dbLs3K8zeCUdpJVvc+oLD6Wxu1uro36XJoOlJl6xXw Jul 10 00:30:07.843701 sshd-session[5869]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:30:07.849478 systemd-logind[1614]: New session 24 of user core. Jul 10 00:30:07.853252 systemd[1]: Started session-24.scope - Session 24 of User core. Jul 10 00:30:08.543842 sshd[5871]: Connection closed by 139.178.68.195 port 47084 Jul 10 00:30:08.544296 sshd-session[5869]: pam_unix(sshd:session): session closed for user core Jul 10 00:30:08.546541 systemd-logind[1614]: Session 24 logged out. Waiting for processes to exit. Jul 10 00:30:08.547561 systemd[1]: sshd@22-139.178.70.108:22-139.178.68.195:47084.service: Deactivated successfully. Jul 10 00:30:08.548780 systemd[1]: session-24.scope: Deactivated successfully. Jul 10 00:30:08.550286 systemd-logind[1614]: Removed session 24.