Jan 17 12:18:03.728273 kernel: Linux version 6.6.71-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Fri Jan 17 10:39:07 -00 2025 Jan 17 12:18:03.728290 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=vmware flatcar.autologin verity.usrhash=bf1e0d81a0170850ab02d370c1a7c7a3f5983c980b3730f748240a3bda2dbb2e Jan 17 12:18:03.728297 kernel: Disabled fast string operations Jan 17 12:18:03.728301 kernel: BIOS-provided physical RAM map: Jan 17 12:18:03.728305 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ebff] usable Jan 17 12:18:03.728308 kernel: BIOS-e820: [mem 0x000000000009ec00-0x000000000009ffff] reserved Jan 17 12:18:03.728314 kernel: BIOS-e820: [mem 0x00000000000dc000-0x00000000000fffff] reserved Jan 17 12:18:03.728318 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007fedffff] usable Jan 17 12:18:03.728322 kernel: BIOS-e820: [mem 0x000000007fee0000-0x000000007fefefff] ACPI data Jan 17 12:18:03.728326 kernel: BIOS-e820: [mem 0x000000007feff000-0x000000007fefffff] ACPI NVS Jan 17 12:18:03.728330 kernel: BIOS-e820: [mem 0x000000007ff00000-0x000000007fffffff] usable Jan 17 12:18:03.728334 kernel: BIOS-e820: [mem 0x00000000f0000000-0x00000000f7ffffff] reserved Jan 17 12:18:03.728338 kernel: BIOS-e820: [mem 0x00000000fec00000-0x00000000fec0ffff] reserved Jan 17 12:18:03.728342 kernel: BIOS-e820: [mem 0x00000000fee00000-0x00000000fee00fff] reserved Jan 17 12:18:03.728348 kernel: BIOS-e820: [mem 0x00000000fffe0000-0x00000000ffffffff] reserved Jan 17 12:18:03.728353 kernel: NX (Execute Disable) protection: active Jan 17 12:18:03.728357 kernel: APIC: Static calls initialized Jan 17 12:18:03.728362 kernel: SMBIOS 2.7 present. Jan 17 12:18:03.728367 kernel: DMI: VMware, Inc. VMware Virtual Platform/440BX Desktop Reference Platform, BIOS 6.00 05/28/2020 Jan 17 12:18:03.728371 kernel: vmware: hypercall mode: 0x00 Jan 17 12:18:03.728376 kernel: Hypervisor detected: VMware Jan 17 12:18:03.728380 kernel: vmware: TSC freq read from hypervisor : 3408.000 MHz Jan 17 12:18:03.728409 kernel: vmware: Host bus clock speed read from hypervisor : 66000000 Hz Jan 17 12:18:03.728416 kernel: vmware: using clock offset of 2537798063 ns Jan 17 12:18:03.728421 kernel: tsc: Detected 3408.000 MHz processor Jan 17 12:18:03.728426 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Jan 17 12:18:03.728448 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Jan 17 12:18:03.728452 kernel: last_pfn = 0x80000 max_arch_pfn = 0x400000000 Jan 17 12:18:03.728457 kernel: total RAM covered: 3072M Jan 17 12:18:03.728463 kernel: Found optimal setting for mtrr clean up Jan 17 12:18:03.728471 kernel: gran_size: 64K chunk_size: 64K num_reg: 2 lose cover RAM: 0G Jan 17 12:18:03.728481 kernel: MTRR map: 6 entries (5 fixed + 1 variable; max 21), built from 8 variable MTRRs Jan 17 12:18:03.728489 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Jan 17 12:18:03.728494 kernel: Using GB pages for direct mapping Jan 17 12:18:03.728499 kernel: ACPI: Early table checksum verification disabled Jan 17 12:18:03.728504 kernel: ACPI: RSDP 0x00000000000F6A00 000024 (v02 PTLTD ) Jan 17 12:18:03.728508 kernel: ACPI: XSDT 0x000000007FEE965B 00005C (v01 INTEL 440BX 06040000 VMW 01324272) Jan 17 12:18:03.728513 kernel: ACPI: FACP 0x000000007FEFEE73 0000F4 (v04 INTEL 440BX 06040000 PTL 000F4240) Jan 17 12:18:03.728518 kernel: ACPI: DSDT 0x000000007FEEAD55 01411E (v01 PTLTD Custom 06040000 MSFT 03000001) Jan 17 12:18:03.728523 kernel: ACPI: FACS 0x000000007FEFFFC0 000040 Jan 17 12:18:03.728530 kernel: ACPI: FACS 0x000000007FEFFFC0 000040 Jan 17 12:18:03.728535 kernel: ACPI: BOOT 0x000000007FEEAD2D 000028 (v01 PTLTD $SBFTBL$ 06040000 LTP 00000001) Jan 17 12:18:03.728540 kernel: ACPI: APIC 0x000000007FEEA5EB 000742 (v01 PTLTD ? APIC 06040000 LTP 00000000) Jan 17 12:18:03.728545 kernel: ACPI: MCFG 0x000000007FEEA5AF 00003C (v01 PTLTD $PCITBL$ 06040000 LTP 00000001) Jan 17 12:18:03.728550 kernel: ACPI: SRAT 0x000000007FEE9757 0008A8 (v02 VMWARE MEMPLUG 06040000 VMW 00000001) Jan 17 12:18:03.728556 kernel: ACPI: HPET 0x000000007FEE971F 000038 (v01 VMWARE VMW HPET 06040000 VMW 00000001) Jan 17 12:18:03.728561 kernel: ACPI: WAET 0x000000007FEE96F7 000028 (v01 VMWARE VMW WAET 06040000 VMW 00000001) Jan 17 12:18:03.728566 kernel: ACPI: Reserving FACP table memory at [mem 0x7fefee73-0x7fefef66] Jan 17 12:18:03.728571 kernel: ACPI: Reserving DSDT table memory at [mem 0x7feead55-0x7fefee72] Jan 17 12:18:03.728576 kernel: ACPI: Reserving FACS table memory at [mem 0x7fefffc0-0x7fefffff] Jan 17 12:18:03.728581 kernel: ACPI: Reserving FACS table memory at [mem 0x7fefffc0-0x7fefffff] Jan 17 12:18:03.728586 kernel: ACPI: Reserving BOOT table memory at [mem 0x7feead2d-0x7feead54] Jan 17 12:18:03.728591 kernel: ACPI: Reserving APIC table memory at [mem 0x7feea5eb-0x7feead2c] Jan 17 12:18:03.728596 kernel: ACPI: Reserving MCFG table memory at [mem 0x7feea5af-0x7feea5ea] Jan 17 12:18:03.728600 kernel: ACPI: Reserving SRAT table memory at [mem 0x7fee9757-0x7fee9ffe] Jan 17 12:18:03.728606 kernel: ACPI: Reserving HPET table memory at [mem 0x7fee971f-0x7fee9756] Jan 17 12:18:03.728611 kernel: ACPI: Reserving WAET table memory at [mem 0x7fee96f7-0x7fee971e] Jan 17 12:18:03.728616 kernel: system APIC only can use physical flat Jan 17 12:18:03.728621 kernel: APIC: Switched APIC routing to: physical flat Jan 17 12:18:03.728626 kernel: SRAT: PXM 0 -> APIC 0x00 -> Node 0 Jan 17 12:18:03.728631 kernel: SRAT: PXM 0 -> APIC 0x02 -> Node 0 Jan 17 12:18:03.728635 kernel: SRAT: PXM 0 -> APIC 0x04 -> Node 0 Jan 17 12:18:03.728640 kernel: SRAT: PXM 0 -> APIC 0x06 -> Node 0 Jan 17 12:18:03.728645 kernel: SRAT: PXM 0 -> APIC 0x08 -> Node 0 Jan 17 12:18:03.728651 kernel: SRAT: PXM 0 -> APIC 0x0a -> Node 0 Jan 17 12:18:03.728656 kernel: SRAT: PXM 0 -> APIC 0x0c -> Node 0 Jan 17 12:18:03.728660 kernel: SRAT: PXM 0 -> APIC 0x0e -> Node 0 Jan 17 12:18:03.728665 kernel: SRAT: PXM 0 -> APIC 0x10 -> Node 0 Jan 17 12:18:03.728670 kernel: SRAT: PXM 0 -> APIC 0x12 -> Node 0 Jan 17 12:18:03.728674 kernel: SRAT: PXM 0 -> APIC 0x14 -> Node 0 Jan 17 12:18:03.728679 kernel: SRAT: PXM 0 -> APIC 0x16 -> Node 0 Jan 17 12:18:03.728684 kernel: SRAT: PXM 0 -> APIC 0x18 -> Node 0 Jan 17 12:18:03.728689 kernel: SRAT: PXM 0 -> APIC 0x1a -> Node 0 Jan 17 12:18:03.728694 kernel: SRAT: PXM 0 -> APIC 0x1c -> Node 0 Jan 17 12:18:03.728698 kernel: SRAT: PXM 0 -> APIC 0x1e -> Node 0 Jan 17 12:18:03.728704 kernel: SRAT: PXM 0 -> APIC 0x20 -> Node 0 Jan 17 12:18:03.728709 kernel: SRAT: PXM 0 -> APIC 0x22 -> Node 0 Jan 17 12:18:03.728714 kernel: SRAT: PXM 0 -> APIC 0x24 -> Node 0 Jan 17 12:18:03.728718 kernel: SRAT: PXM 0 -> APIC 0x26 -> Node 0 Jan 17 12:18:03.728723 kernel: SRAT: PXM 0 -> APIC 0x28 -> Node 0 Jan 17 12:18:03.728728 kernel: SRAT: PXM 0 -> APIC 0x2a -> Node 0 Jan 17 12:18:03.728733 kernel: SRAT: PXM 0 -> APIC 0x2c -> Node 0 Jan 17 12:18:03.728737 kernel: SRAT: PXM 0 -> APIC 0x2e -> Node 0 Jan 17 12:18:03.728742 kernel: SRAT: PXM 0 -> APIC 0x30 -> Node 0 Jan 17 12:18:03.728747 kernel: SRAT: PXM 0 -> APIC 0x32 -> Node 0 Jan 17 12:18:03.728753 kernel: SRAT: PXM 0 -> APIC 0x34 -> Node 0 Jan 17 12:18:03.728758 kernel: SRAT: PXM 0 -> APIC 0x36 -> Node 0 Jan 17 12:18:03.728762 kernel: SRAT: PXM 0 -> APIC 0x38 -> Node 0 Jan 17 12:18:03.728767 kernel: SRAT: PXM 0 -> APIC 0x3a -> Node 0 Jan 17 12:18:03.728772 kernel: SRAT: PXM 0 -> APIC 0x3c -> Node 0 Jan 17 12:18:03.728777 kernel: SRAT: PXM 0 -> APIC 0x3e -> Node 0 Jan 17 12:18:03.728782 kernel: SRAT: PXM 0 -> APIC 0x40 -> Node 0 Jan 17 12:18:03.728787 kernel: SRAT: PXM 0 -> APIC 0x42 -> Node 0 Jan 17 12:18:03.728792 kernel: SRAT: PXM 0 -> APIC 0x44 -> Node 0 Jan 17 12:18:03.728796 kernel: SRAT: PXM 0 -> APIC 0x46 -> Node 0 Jan 17 12:18:03.728802 kernel: SRAT: PXM 0 -> APIC 0x48 -> Node 0 Jan 17 12:18:03.728807 kernel: SRAT: PXM 0 -> APIC 0x4a -> Node 0 Jan 17 12:18:03.728812 kernel: SRAT: PXM 0 -> APIC 0x4c -> Node 0 Jan 17 12:18:03.728816 kernel: SRAT: PXM 0 -> APIC 0x4e -> Node 0 Jan 17 12:18:03.728821 kernel: SRAT: PXM 0 -> APIC 0x50 -> Node 0 Jan 17 12:18:03.728826 kernel: SRAT: PXM 0 -> APIC 0x52 -> Node 0 Jan 17 12:18:03.728831 kernel: SRAT: PXM 0 -> APIC 0x54 -> Node 0 Jan 17 12:18:03.728836 kernel: SRAT: PXM 0 -> APIC 0x56 -> Node 0 Jan 17 12:18:03.728840 kernel: SRAT: PXM 0 -> APIC 0x58 -> Node 0 Jan 17 12:18:03.728845 kernel: SRAT: PXM 0 -> APIC 0x5a -> Node 0 Jan 17 12:18:03.728851 kernel: SRAT: PXM 0 -> APIC 0x5c -> Node 0 Jan 17 12:18:03.728856 kernel: SRAT: PXM 0 -> APIC 0x5e -> Node 0 Jan 17 12:18:03.728861 kernel: SRAT: PXM 0 -> APIC 0x60 -> Node 0 Jan 17 12:18:03.728865 kernel: SRAT: PXM 0 -> APIC 0x62 -> Node 0 Jan 17 12:18:03.728870 kernel: SRAT: PXM 0 -> APIC 0x64 -> Node 0 Jan 17 12:18:03.728875 kernel: SRAT: PXM 0 -> APIC 0x66 -> Node 0 Jan 17 12:18:03.728879 kernel: SRAT: PXM 0 -> APIC 0x68 -> Node 0 Jan 17 12:18:03.728884 kernel: SRAT: PXM 0 -> APIC 0x6a -> Node 0 Jan 17 12:18:03.728889 kernel: SRAT: PXM 0 -> APIC 0x6c -> Node 0 Jan 17 12:18:03.728894 kernel: SRAT: PXM 0 -> APIC 0x6e -> Node 0 Jan 17 12:18:03.728900 kernel: SRAT: PXM 0 -> APIC 0x70 -> Node 0 Jan 17 12:18:03.728905 kernel: SRAT: PXM 0 -> APIC 0x72 -> Node 0 Jan 17 12:18:03.728910 kernel: SRAT: PXM 0 -> APIC 0x74 -> Node 0 Jan 17 12:18:03.728918 kernel: SRAT: PXM 0 -> APIC 0x76 -> Node 0 Jan 17 12:18:03.728924 kernel: SRAT: PXM 0 -> APIC 0x78 -> Node 0 Jan 17 12:18:03.728929 kernel: SRAT: PXM 0 -> APIC 0x7a -> Node 0 Jan 17 12:18:03.728934 kernel: SRAT: PXM 0 -> APIC 0x7c -> Node 0 Jan 17 12:18:03.728939 kernel: SRAT: PXM 0 -> APIC 0x7e -> Node 0 Jan 17 12:18:03.728944 kernel: SRAT: PXM 0 -> APIC 0x80 -> Node 0 Jan 17 12:18:03.728951 kernel: SRAT: PXM 0 -> APIC 0x82 -> Node 0 Jan 17 12:18:03.728955 kernel: SRAT: PXM 0 -> APIC 0x84 -> Node 0 Jan 17 12:18:03.728961 kernel: SRAT: PXM 0 -> APIC 0x86 -> Node 0 Jan 17 12:18:03.728966 kernel: SRAT: PXM 0 -> APIC 0x88 -> Node 0 Jan 17 12:18:03.728971 kernel: SRAT: PXM 0 -> APIC 0x8a -> Node 0 Jan 17 12:18:03.728976 kernel: SRAT: PXM 0 -> APIC 0x8c -> Node 0 Jan 17 12:18:03.728981 kernel: SRAT: PXM 0 -> APIC 0x8e -> Node 0 Jan 17 12:18:03.728986 kernel: SRAT: PXM 0 -> APIC 0x90 -> Node 0 Jan 17 12:18:03.728991 kernel: SRAT: PXM 0 -> APIC 0x92 -> Node 0 Jan 17 12:18:03.728996 kernel: SRAT: PXM 0 -> APIC 0x94 -> Node 0 Jan 17 12:18:03.729003 kernel: SRAT: PXM 0 -> APIC 0x96 -> Node 0 Jan 17 12:18:03.729008 kernel: SRAT: PXM 0 -> APIC 0x98 -> Node 0 Jan 17 12:18:03.729013 kernel: SRAT: PXM 0 -> APIC 0x9a -> Node 0 Jan 17 12:18:03.729018 kernel: SRAT: PXM 0 -> APIC 0x9c -> Node 0 Jan 17 12:18:03.729023 kernel: SRAT: PXM 0 -> APIC 0x9e -> Node 0 Jan 17 12:18:03.729029 kernel: SRAT: PXM 0 -> APIC 0xa0 -> Node 0 Jan 17 12:18:03.729034 kernel: SRAT: PXM 0 -> APIC 0xa2 -> Node 0 Jan 17 12:18:03.729039 kernel: SRAT: PXM 0 -> APIC 0xa4 -> Node 0 Jan 17 12:18:03.729044 kernel: SRAT: PXM 0 -> APIC 0xa6 -> Node 0 Jan 17 12:18:03.729049 kernel: SRAT: PXM 0 -> APIC 0xa8 -> Node 0 Jan 17 12:18:03.729055 kernel: SRAT: PXM 0 -> APIC 0xaa -> Node 0 Jan 17 12:18:03.729060 kernel: SRAT: PXM 0 -> APIC 0xac -> Node 0 Jan 17 12:18:03.729065 kernel: SRAT: PXM 0 -> APIC 0xae -> Node 0 Jan 17 12:18:03.729070 kernel: SRAT: PXM 0 -> APIC 0xb0 -> Node 0 Jan 17 12:18:03.729075 kernel: SRAT: PXM 0 -> APIC 0xb2 -> Node 0 Jan 17 12:18:03.729080 kernel: SRAT: PXM 0 -> APIC 0xb4 -> Node 0 Jan 17 12:18:03.729086 kernel: SRAT: PXM 0 -> APIC 0xb6 -> Node 0 Jan 17 12:18:03.729091 kernel: SRAT: PXM 0 -> APIC 0xb8 -> Node 0 Jan 17 12:18:03.729096 kernel: SRAT: PXM 0 -> APIC 0xba -> Node 0 Jan 17 12:18:03.729110 kernel: SRAT: PXM 0 -> APIC 0xbc -> Node 0 Jan 17 12:18:03.729119 kernel: SRAT: PXM 0 -> APIC 0xbe -> Node 0 Jan 17 12:18:03.729124 kernel: SRAT: PXM 0 -> APIC 0xc0 -> Node 0 Jan 17 12:18:03.729129 kernel: SRAT: PXM 0 -> APIC 0xc2 -> Node 0 Jan 17 12:18:03.729134 kernel: SRAT: PXM 0 -> APIC 0xc4 -> Node 0 Jan 17 12:18:03.729139 kernel: SRAT: PXM 0 -> APIC 0xc6 -> Node 0 Jan 17 12:18:03.729145 kernel: SRAT: PXM 0 -> APIC 0xc8 -> Node 0 Jan 17 12:18:03.729150 kernel: SRAT: PXM 0 -> APIC 0xca -> Node 0 Jan 17 12:18:03.729155 kernel: SRAT: PXM 0 -> APIC 0xcc -> Node 0 Jan 17 12:18:03.729160 kernel: SRAT: PXM 0 -> APIC 0xce -> Node 0 Jan 17 12:18:03.729165 kernel: SRAT: PXM 0 -> APIC 0xd0 -> Node 0 Jan 17 12:18:03.729170 kernel: SRAT: PXM 0 -> APIC 0xd2 -> Node 0 Jan 17 12:18:03.729176 kernel: SRAT: PXM 0 -> APIC 0xd4 -> Node 0 Jan 17 12:18:03.729181 kernel: SRAT: PXM 0 -> APIC 0xd6 -> Node 0 Jan 17 12:18:03.729187 kernel: SRAT: PXM 0 -> APIC 0xd8 -> Node 0 Jan 17 12:18:03.729192 kernel: SRAT: PXM 0 -> APIC 0xda -> Node 0 Jan 17 12:18:03.729197 kernel: SRAT: PXM 0 -> APIC 0xdc -> Node 0 Jan 17 12:18:03.729202 kernel: SRAT: PXM 0 -> APIC 0xde -> Node 0 Jan 17 12:18:03.729207 kernel: SRAT: PXM 0 -> APIC 0xe0 -> Node 0 Jan 17 12:18:03.729212 kernel: SRAT: PXM 0 -> APIC 0xe2 -> Node 0 Jan 17 12:18:03.729217 kernel: SRAT: PXM 0 -> APIC 0xe4 -> Node 0 Jan 17 12:18:03.729223 kernel: SRAT: PXM 0 -> APIC 0xe6 -> Node 0 Jan 17 12:18:03.729229 kernel: SRAT: PXM 0 -> APIC 0xe8 -> Node 0 Jan 17 12:18:03.729234 kernel: SRAT: PXM 0 -> APIC 0xea -> Node 0 Jan 17 12:18:03.729239 kernel: SRAT: PXM 0 -> APIC 0xec -> Node 0 Jan 17 12:18:03.729244 kernel: SRAT: PXM 0 -> APIC 0xee -> Node 0 Jan 17 12:18:03.729249 kernel: SRAT: PXM 0 -> APIC 0xf0 -> Node 0 Jan 17 12:18:03.729254 kernel: SRAT: PXM 0 -> APIC 0xf2 -> Node 0 Jan 17 12:18:03.729259 kernel: SRAT: PXM 0 -> APIC 0xf4 -> Node 0 Jan 17 12:18:03.729264 kernel: SRAT: PXM 0 -> APIC 0xf6 -> Node 0 Jan 17 12:18:03.729269 kernel: SRAT: PXM 0 -> APIC 0xf8 -> Node 0 Jan 17 12:18:03.729275 kernel: SRAT: PXM 0 -> APIC 0xfa -> Node 0 Jan 17 12:18:03.729281 kernel: SRAT: PXM 0 -> APIC 0xfc -> Node 0 Jan 17 12:18:03.729286 kernel: SRAT: PXM 0 -> APIC 0xfe -> Node 0 Jan 17 12:18:03.729291 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x0009ffff] Jan 17 12:18:03.729297 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00100000-0x7fffffff] Jan 17 12:18:03.729302 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x80000000-0xbfffffff] hotplug Jan 17 12:18:03.729307 kernel: NUMA: Node 0 [mem 0x00000000-0x0009ffff] + [mem 0x00100000-0x7fffffff] -> [mem 0x00000000-0x7fffffff] Jan 17 12:18:03.729313 kernel: NODE_DATA(0) allocated [mem 0x7fffa000-0x7fffffff] Jan 17 12:18:03.729318 kernel: Zone ranges: Jan 17 12:18:03.729323 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Jan 17 12:18:03.729330 kernel: DMA32 [mem 0x0000000001000000-0x000000007fffffff] Jan 17 12:18:03.729335 kernel: Normal empty Jan 17 12:18:03.729341 kernel: Movable zone start for each node Jan 17 12:18:03.729346 kernel: Early memory node ranges Jan 17 12:18:03.729351 kernel: node 0: [mem 0x0000000000001000-0x000000000009dfff] Jan 17 12:18:03.729356 kernel: node 0: [mem 0x0000000000100000-0x000000007fedffff] Jan 17 12:18:03.729361 kernel: node 0: [mem 0x000000007ff00000-0x000000007fffffff] Jan 17 12:18:03.729367 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007fffffff] Jan 17 12:18:03.729372 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Jan 17 12:18:03.729377 kernel: On node 0, zone DMA: 98 pages in unavailable ranges Jan 17 12:18:03.729383 kernel: On node 0, zone DMA32: 32 pages in unavailable ranges Jan 17 12:18:03.729389 kernel: ACPI: PM-Timer IO Port: 0x1008 Jan 17 12:18:03.729412 kernel: system APIC only can use physical flat Jan 17 12:18:03.729417 kernel: ACPI: LAPIC_NMI (acpi_id[0x00] high edge lint[0x1]) Jan 17 12:18:03.729423 kernel: ACPI: LAPIC_NMI (acpi_id[0x01] high edge lint[0x1]) Jan 17 12:18:03.729428 kernel: ACPI: LAPIC_NMI (acpi_id[0x02] high edge lint[0x1]) Jan 17 12:18:03.729433 kernel: ACPI: LAPIC_NMI (acpi_id[0x03] high edge lint[0x1]) Jan 17 12:18:03.729455 kernel: ACPI: LAPIC_NMI (acpi_id[0x04] high edge lint[0x1]) Jan 17 12:18:03.729460 kernel: ACPI: LAPIC_NMI (acpi_id[0x05] high edge lint[0x1]) Jan 17 12:18:03.729465 kernel: ACPI: LAPIC_NMI (acpi_id[0x06] high edge lint[0x1]) Jan 17 12:18:03.729471 kernel: ACPI: LAPIC_NMI (acpi_id[0x07] high edge lint[0x1]) Jan 17 12:18:03.729476 kernel: ACPI: LAPIC_NMI (acpi_id[0x08] high edge lint[0x1]) Jan 17 12:18:03.729481 kernel: ACPI: LAPIC_NMI (acpi_id[0x09] high edge lint[0x1]) Jan 17 12:18:03.729487 kernel: ACPI: LAPIC_NMI (acpi_id[0x0a] high edge lint[0x1]) Jan 17 12:18:03.729492 kernel: ACPI: LAPIC_NMI (acpi_id[0x0b] high edge lint[0x1]) Jan 17 12:18:03.729497 kernel: ACPI: LAPIC_NMI (acpi_id[0x0c] high edge lint[0x1]) Jan 17 12:18:03.729502 kernel: ACPI: LAPIC_NMI (acpi_id[0x0d] high edge lint[0x1]) Jan 17 12:18:03.729507 kernel: ACPI: LAPIC_NMI (acpi_id[0x0e] high edge lint[0x1]) Jan 17 12:18:03.729512 kernel: ACPI: LAPIC_NMI (acpi_id[0x0f] high edge lint[0x1]) Jan 17 12:18:03.729519 kernel: ACPI: LAPIC_NMI (acpi_id[0x10] high edge lint[0x1]) Jan 17 12:18:03.729524 kernel: ACPI: LAPIC_NMI (acpi_id[0x11] high edge lint[0x1]) Jan 17 12:18:03.729529 kernel: ACPI: LAPIC_NMI (acpi_id[0x12] high edge lint[0x1]) Jan 17 12:18:03.729534 kernel: ACPI: LAPIC_NMI (acpi_id[0x13] high edge lint[0x1]) Jan 17 12:18:03.729540 kernel: ACPI: LAPIC_NMI (acpi_id[0x14] high edge lint[0x1]) Jan 17 12:18:03.729545 kernel: ACPI: LAPIC_NMI (acpi_id[0x15] high edge lint[0x1]) Jan 17 12:18:03.729550 kernel: ACPI: LAPIC_NMI (acpi_id[0x16] high edge lint[0x1]) Jan 17 12:18:03.729555 kernel: ACPI: LAPIC_NMI (acpi_id[0x17] high edge lint[0x1]) Jan 17 12:18:03.729560 kernel: ACPI: LAPIC_NMI (acpi_id[0x18] high edge lint[0x1]) Jan 17 12:18:03.729566 kernel: ACPI: LAPIC_NMI (acpi_id[0x19] high edge lint[0x1]) Jan 17 12:18:03.729572 kernel: ACPI: LAPIC_NMI (acpi_id[0x1a] high edge lint[0x1]) Jan 17 12:18:03.729577 kernel: ACPI: LAPIC_NMI (acpi_id[0x1b] high edge lint[0x1]) Jan 17 12:18:03.729582 kernel: ACPI: LAPIC_NMI (acpi_id[0x1c] high edge lint[0x1]) Jan 17 12:18:03.729587 kernel: ACPI: LAPIC_NMI (acpi_id[0x1d] high edge lint[0x1]) Jan 17 12:18:03.729593 kernel: ACPI: LAPIC_NMI (acpi_id[0x1e] high edge lint[0x1]) Jan 17 12:18:03.729598 kernel: ACPI: LAPIC_NMI (acpi_id[0x1f] high edge lint[0x1]) Jan 17 12:18:03.729603 kernel: ACPI: LAPIC_NMI (acpi_id[0x20] high edge lint[0x1]) Jan 17 12:18:03.729608 kernel: ACPI: LAPIC_NMI (acpi_id[0x21] high edge lint[0x1]) Jan 17 12:18:03.729613 kernel: ACPI: LAPIC_NMI (acpi_id[0x22] high edge lint[0x1]) Jan 17 12:18:03.729618 kernel: ACPI: LAPIC_NMI (acpi_id[0x23] high edge lint[0x1]) Jan 17 12:18:03.729624 kernel: ACPI: LAPIC_NMI (acpi_id[0x24] high edge lint[0x1]) Jan 17 12:18:03.729630 kernel: ACPI: LAPIC_NMI (acpi_id[0x25] high edge lint[0x1]) Jan 17 12:18:03.729635 kernel: ACPI: LAPIC_NMI (acpi_id[0x26] high edge lint[0x1]) Jan 17 12:18:03.729640 kernel: ACPI: LAPIC_NMI (acpi_id[0x27] high edge lint[0x1]) Jan 17 12:18:03.729645 kernel: ACPI: LAPIC_NMI (acpi_id[0x28] high edge lint[0x1]) Jan 17 12:18:03.729651 kernel: ACPI: LAPIC_NMI (acpi_id[0x29] high edge lint[0x1]) Jan 17 12:18:03.729656 kernel: ACPI: LAPIC_NMI (acpi_id[0x2a] high edge lint[0x1]) Jan 17 12:18:03.729661 kernel: ACPI: LAPIC_NMI (acpi_id[0x2b] high edge lint[0x1]) Jan 17 12:18:03.729666 kernel: ACPI: LAPIC_NMI (acpi_id[0x2c] high edge lint[0x1]) Jan 17 12:18:03.729672 kernel: ACPI: LAPIC_NMI (acpi_id[0x2d] high edge lint[0x1]) Jan 17 12:18:03.729678 kernel: ACPI: LAPIC_NMI (acpi_id[0x2e] high edge lint[0x1]) Jan 17 12:18:03.729683 kernel: ACPI: LAPIC_NMI (acpi_id[0x2f] high edge lint[0x1]) Jan 17 12:18:03.729688 kernel: ACPI: LAPIC_NMI (acpi_id[0x30] high edge lint[0x1]) Jan 17 12:18:03.729693 kernel: ACPI: LAPIC_NMI (acpi_id[0x31] high edge lint[0x1]) Jan 17 12:18:03.729698 kernel: ACPI: LAPIC_NMI (acpi_id[0x32] high edge lint[0x1]) Jan 17 12:18:03.729704 kernel: ACPI: LAPIC_NMI (acpi_id[0x33] high edge lint[0x1]) Jan 17 12:18:03.729709 kernel: ACPI: LAPIC_NMI (acpi_id[0x34] high edge lint[0x1]) Jan 17 12:18:03.729714 kernel: ACPI: LAPIC_NMI (acpi_id[0x35] high edge lint[0x1]) Jan 17 12:18:03.729719 kernel: ACPI: LAPIC_NMI (acpi_id[0x36] high edge lint[0x1]) Jan 17 12:18:03.729725 kernel: ACPI: LAPIC_NMI (acpi_id[0x37] high edge lint[0x1]) Jan 17 12:18:03.729730 kernel: ACPI: LAPIC_NMI (acpi_id[0x38] high edge lint[0x1]) Jan 17 12:18:03.729736 kernel: ACPI: LAPIC_NMI (acpi_id[0x39] high edge lint[0x1]) Jan 17 12:18:03.729741 kernel: ACPI: LAPIC_NMI (acpi_id[0x3a] high edge lint[0x1]) Jan 17 12:18:03.729746 kernel: ACPI: LAPIC_NMI (acpi_id[0x3b] high edge lint[0x1]) Jan 17 12:18:03.729751 kernel: ACPI: LAPIC_NMI (acpi_id[0x3c] high edge lint[0x1]) Jan 17 12:18:03.729757 kernel: ACPI: LAPIC_NMI (acpi_id[0x3d] high edge lint[0x1]) Jan 17 12:18:03.729762 kernel: ACPI: LAPIC_NMI (acpi_id[0x3e] high edge lint[0x1]) Jan 17 12:18:03.729767 kernel: ACPI: LAPIC_NMI (acpi_id[0x3f] high edge lint[0x1]) Jan 17 12:18:03.729772 kernel: ACPI: LAPIC_NMI (acpi_id[0x40] high edge lint[0x1]) Jan 17 12:18:03.729778 kernel: ACPI: LAPIC_NMI (acpi_id[0x41] high edge lint[0x1]) Jan 17 12:18:03.729783 kernel: ACPI: LAPIC_NMI (acpi_id[0x42] high edge lint[0x1]) Jan 17 12:18:03.729789 kernel: ACPI: LAPIC_NMI (acpi_id[0x43] high edge lint[0x1]) Jan 17 12:18:03.729794 kernel: ACPI: LAPIC_NMI (acpi_id[0x44] high edge lint[0x1]) Jan 17 12:18:03.729799 kernel: ACPI: LAPIC_NMI (acpi_id[0x45] high edge lint[0x1]) Jan 17 12:18:03.729804 kernel: ACPI: LAPIC_NMI (acpi_id[0x46] high edge lint[0x1]) Jan 17 12:18:03.729810 kernel: ACPI: LAPIC_NMI (acpi_id[0x47] high edge lint[0x1]) Jan 17 12:18:03.729815 kernel: ACPI: LAPIC_NMI (acpi_id[0x48] high edge lint[0x1]) Jan 17 12:18:03.729820 kernel: ACPI: LAPIC_NMI (acpi_id[0x49] high edge lint[0x1]) Jan 17 12:18:03.729826 kernel: ACPI: LAPIC_NMI (acpi_id[0x4a] high edge lint[0x1]) Jan 17 12:18:03.729831 kernel: ACPI: LAPIC_NMI (acpi_id[0x4b] high edge lint[0x1]) Jan 17 12:18:03.729837 kernel: ACPI: LAPIC_NMI (acpi_id[0x4c] high edge lint[0x1]) Jan 17 12:18:03.729842 kernel: ACPI: LAPIC_NMI (acpi_id[0x4d] high edge lint[0x1]) Jan 17 12:18:03.729847 kernel: ACPI: LAPIC_NMI (acpi_id[0x4e] high edge lint[0x1]) Jan 17 12:18:03.729852 kernel: ACPI: LAPIC_NMI (acpi_id[0x4f] high edge lint[0x1]) Jan 17 12:18:03.729858 kernel: ACPI: LAPIC_NMI (acpi_id[0x50] high edge lint[0x1]) Jan 17 12:18:03.729863 kernel: ACPI: LAPIC_NMI (acpi_id[0x51] high edge lint[0x1]) Jan 17 12:18:03.729868 kernel: ACPI: LAPIC_NMI (acpi_id[0x52] high edge lint[0x1]) Jan 17 12:18:03.729873 kernel: ACPI: LAPIC_NMI (acpi_id[0x53] high edge lint[0x1]) Jan 17 12:18:03.729879 kernel: ACPI: LAPIC_NMI (acpi_id[0x54] high edge lint[0x1]) Jan 17 12:18:03.729885 kernel: ACPI: LAPIC_NMI (acpi_id[0x55] high edge lint[0x1]) Jan 17 12:18:03.729890 kernel: ACPI: LAPIC_NMI (acpi_id[0x56] high edge lint[0x1]) Jan 17 12:18:03.729895 kernel: ACPI: LAPIC_NMI (acpi_id[0x57] high edge lint[0x1]) Jan 17 12:18:03.729900 kernel: ACPI: LAPIC_NMI (acpi_id[0x58] high edge lint[0x1]) Jan 17 12:18:03.729906 kernel: ACPI: LAPIC_NMI (acpi_id[0x59] high edge lint[0x1]) Jan 17 12:18:03.729911 kernel: ACPI: LAPIC_NMI (acpi_id[0x5a] high edge lint[0x1]) Jan 17 12:18:03.729916 kernel: ACPI: LAPIC_NMI (acpi_id[0x5b] high edge lint[0x1]) Jan 17 12:18:03.729921 kernel: ACPI: LAPIC_NMI (acpi_id[0x5c] high edge lint[0x1]) Jan 17 12:18:03.729927 kernel: ACPI: LAPIC_NMI (acpi_id[0x5d] high edge lint[0x1]) Jan 17 12:18:03.729933 kernel: ACPI: LAPIC_NMI (acpi_id[0x5e] high edge lint[0x1]) Jan 17 12:18:03.729938 kernel: ACPI: LAPIC_NMI (acpi_id[0x5f] high edge lint[0x1]) Jan 17 12:18:03.729943 kernel: ACPI: LAPIC_NMI (acpi_id[0x60] high edge lint[0x1]) Jan 17 12:18:03.729948 kernel: ACPI: LAPIC_NMI (acpi_id[0x61] high edge lint[0x1]) Jan 17 12:18:03.729954 kernel: ACPI: LAPIC_NMI (acpi_id[0x62] high edge lint[0x1]) Jan 17 12:18:03.729959 kernel: ACPI: LAPIC_NMI (acpi_id[0x63] high edge lint[0x1]) Jan 17 12:18:03.729964 kernel: ACPI: LAPIC_NMI (acpi_id[0x64] high edge lint[0x1]) Jan 17 12:18:03.729969 kernel: ACPI: LAPIC_NMI (acpi_id[0x65] high edge lint[0x1]) Jan 17 12:18:03.729974 kernel: ACPI: LAPIC_NMI (acpi_id[0x66] high edge lint[0x1]) Jan 17 12:18:03.729980 kernel: ACPI: LAPIC_NMI (acpi_id[0x67] high edge lint[0x1]) Jan 17 12:18:03.729986 kernel: ACPI: LAPIC_NMI (acpi_id[0x68] high edge lint[0x1]) Jan 17 12:18:03.729991 kernel: ACPI: LAPIC_NMI (acpi_id[0x69] high edge lint[0x1]) Jan 17 12:18:03.729996 kernel: ACPI: LAPIC_NMI (acpi_id[0x6a] high edge lint[0x1]) Jan 17 12:18:03.730001 kernel: ACPI: LAPIC_NMI (acpi_id[0x6b] high edge lint[0x1]) Jan 17 12:18:03.730007 kernel: ACPI: LAPIC_NMI (acpi_id[0x6c] high edge lint[0x1]) Jan 17 12:18:03.730012 kernel: ACPI: LAPIC_NMI (acpi_id[0x6d] high edge lint[0x1]) Jan 17 12:18:03.730017 kernel: ACPI: LAPIC_NMI (acpi_id[0x6e] high edge lint[0x1]) Jan 17 12:18:03.730022 kernel: ACPI: LAPIC_NMI (acpi_id[0x6f] high edge lint[0x1]) Jan 17 12:18:03.730027 kernel: ACPI: LAPIC_NMI (acpi_id[0x70] high edge lint[0x1]) Jan 17 12:18:03.730034 kernel: ACPI: LAPIC_NMI (acpi_id[0x71] high edge lint[0x1]) Jan 17 12:18:03.730039 kernel: ACPI: LAPIC_NMI (acpi_id[0x72] high edge lint[0x1]) Jan 17 12:18:03.730044 kernel: ACPI: LAPIC_NMI (acpi_id[0x73] high edge lint[0x1]) Jan 17 12:18:03.730049 kernel: ACPI: LAPIC_NMI (acpi_id[0x74] high edge lint[0x1]) Jan 17 12:18:03.730054 kernel: ACPI: LAPIC_NMI (acpi_id[0x75] high edge lint[0x1]) Jan 17 12:18:03.730059 kernel: ACPI: LAPIC_NMI (acpi_id[0x76] high edge lint[0x1]) Jan 17 12:18:03.730065 kernel: ACPI: LAPIC_NMI (acpi_id[0x77] high edge lint[0x1]) Jan 17 12:18:03.730070 kernel: ACPI: LAPIC_NMI (acpi_id[0x78] high edge lint[0x1]) Jan 17 12:18:03.730075 kernel: ACPI: LAPIC_NMI (acpi_id[0x79] high edge lint[0x1]) Jan 17 12:18:03.730080 kernel: ACPI: LAPIC_NMI (acpi_id[0x7a] high edge lint[0x1]) Jan 17 12:18:03.730086 kernel: ACPI: LAPIC_NMI (acpi_id[0x7b] high edge lint[0x1]) Jan 17 12:18:03.730092 kernel: ACPI: LAPIC_NMI (acpi_id[0x7c] high edge lint[0x1]) Jan 17 12:18:03.730097 kernel: ACPI: LAPIC_NMI (acpi_id[0x7d] high edge lint[0x1]) Jan 17 12:18:03.730112 kernel: ACPI: LAPIC_NMI (acpi_id[0x7e] high edge lint[0x1]) Jan 17 12:18:03.730117 kernel: ACPI: LAPIC_NMI (acpi_id[0x7f] high edge lint[0x1]) Jan 17 12:18:03.730122 kernel: IOAPIC[0]: apic_id 1, version 17, address 0xfec00000, GSI 0-23 Jan 17 12:18:03.730128 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 high edge) Jan 17 12:18:03.730133 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Jan 17 12:18:03.730138 kernel: ACPI: HPET id: 0x8086af01 base: 0xfed00000 Jan 17 12:18:03.730145 kernel: TSC deadline timer available Jan 17 12:18:03.730150 kernel: smpboot: Allowing 128 CPUs, 126 hotplug CPUs Jan 17 12:18:03.730155 kernel: [mem 0x80000000-0xefffffff] available for PCI devices Jan 17 12:18:03.730161 kernel: Booting paravirtualized kernel on VMware hypervisor Jan 17 12:18:03.730166 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Jan 17 12:18:03.730171 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:128 nr_cpu_ids:128 nr_node_ids:1 Jan 17 12:18:03.730177 kernel: percpu: Embedded 58 pages/cpu s197032 r8192 d32344 u262144 Jan 17 12:18:03.730182 kernel: pcpu-alloc: s197032 r8192 d32344 u262144 alloc=1*2097152 Jan 17 12:18:03.730187 kernel: pcpu-alloc: [0] 000 001 002 003 004 005 006 007 Jan 17 12:18:03.730194 kernel: pcpu-alloc: [0] 008 009 010 011 012 013 014 015 Jan 17 12:18:03.730199 kernel: pcpu-alloc: [0] 016 017 018 019 020 021 022 023 Jan 17 12:18:03.730204 kernel: pcpu-alloc: [0] 024 025 026 027 028 029 030 031 Jan 17 12:18:03.730209 kernel: pcpu-alloc: [0] 032 033 034 035 036 037 038 039 Jan 17 12:18:03.730221 kernel: pcpu-alloc: [0] 040 041 042 043 044 045 046 047 Jan 17 12:18:03.730228 kernel: pcpu-alloc: [0] 048 049 050 051 052 053 054 055 Jan 17 12:18:03.730233 kernel: pcpu-alloc: [0] 056 057 058 059 060 061 062 063 Jan 17 12:18:03.730239 kernel: pcpu-alloc: [0] 064 065 066 067 068 069 070 071 Jan 17 12:18:03.730245 kernel: pcpu-alloc: [0] 072 073 074 075 076 077 078 079 Jan 17 12:18:03.730251 kernel: pcpu-alloc: [0] 080 081 082 083 084 085 086 087 Jan 17 12:18:03.730257 kernel: pcpu-alloc: [0] 088 089 090 091 092 093 094 095 Jan 17 12:18:03.730262 kernel: pcpu-alloc: [0] 096 097 098 099 100 101 102 103 Jan 17 12:18:03.730268 kernel: pcpu-alloc: [0] 104 105 106 107 108 109 110 111 Jan 17 12:18:03.730273 kernel: pcpu-alloc: [0] 112 113 114 115 116 117 118 119 Jan 17 12:18:03.730278 kernel: pcpu-alloc: [0] 120 121 122 123 124 125 126 127 Jan 17 12:18:03.730284 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=vmware flatcar.autologin verity.usrhash=bf1e0d81a0170850ab02d370c1a7c7a3f5983c980b3730f748240a3bda2dbb2e Jan 17 12:18:03.730290 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Jan 17 12:18:03.730297 kernel: random: crng init done Jan 17 12:18:03.730302 kernel: printk: log_buf_len individual max cpu contribution: 4096 bytes Jan 17 12:18:03.730308 kernel: printk: log_buf_len total cpu_extra contributions: 520192 bytes Jan 17 12:18:03.730314 kernel: printk: log_buf_len min size: 262144 bytes Jan 17 12:18:03.730319 kernel: printk: log_buf_len: 1048576 bytes Jan 17 12:18:03.730325 kernel: printk: early log buf free: 239648(91%) Jan 17 12:18:03.730330 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Jan 17 12:18:03.730336 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) Jan 17 12:18:03.730342 kernel: Fallback order for Node 0: 0 Jan 17 12:18:03.730348 kernel: Built 1 zonelists, mobility grouping on. Total pages: 515808 Jan 17 12:18:03.730354 kernel: Policy zone: DMA32 Jan 17 12:18:03.730359 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Jan 17 12:18:03.730365 kernel: Memory: 1936372K/2096628K available (12288K kernel code, 2299K rwdata, 22728K rodata, 42848K init, 2344K bss, 159996K reserved, 0K cma-reserved) Jan 17 12:18:03.730372 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=128, Nodes=1 Jan 17 12:18:03.730378 kernel: ftrace: allocating 37918 entries in 149 pages Jan 17 12:18:03.730384 kernel: ftrace: allocated 149 pages with 4 groups Jan 17 12:18:03.730411 kernel: Dynamic Preempt: voluntary Jan 17 12:18:03.730417 kernel: rcu: Preemptible hierarchical RCU implementation. Jan 17 12:18:03.730423 kernel: rcu: RCU event tracing is enabled. Jan 17 12:18:03.730429 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=128. Jan 17 12:18:03.730435 kernel: Trampoline variant of Tasks RCU enabled. Jan 17 12:18:03.730457 kernel: Rude variant of Tasks RCU enabled. Jan 17 12:18:03.730463 kernel: Tracing variant of Tasks RCU enabled. Jan 17 12:18:03.730483 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Jan 17 12:18:03.730489 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=128 Jan 17 12:18:03.730495 kernel: NR_IRQS: 33024, nr_irqs: 1448, preallocated irqs: 16 Jan 17 12:18:03.730500 kernel: rcu: srcu_init: Setting srcu_struct sizes to big. Jan 17 12:18:03.730506 kernel: Console: colour VGA+ 80x25 Jan 17 12:18:03.730512 kernel: printk: console [tty0] enabled Jan 17 12:18:03.730517 kernel: printk: console [ttyS0] enabled Jan 17 12:18:03.730523 kernel: ACPI: Core revision 20230628 Jan 17 12:18:03.730529 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 133484882848 ns Jan 17 12:18:03.730534 kernel: APIC: Switch to symmetric I/O mode setup Jan 17 12:18:03.730541 kernel: x2apic enabled Jan 17 12:18:03.730547 kernel: APIC: Switched APIC routing to: physical x2apic Jan 17 12:18:03.730552 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Jan 17 12:18:03.730558 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x311fd3cd494, max_idle_ns: 440795223879 ns Jan 17 12:18:03.730563 kernel: Calibrating delay loop (skipped) preset value.. 6816.00 BogoMIPS (lpj=3408000) Jan 17 12:18:03.730569 kernel: Disabled fast string operations Jan 17 12:18:03.730574 kernel: Last level iTLB entries: 4KB 64, 2MB 8, 4MB 8 Jan 17 12:18:03.730580 kernel: Last level dTLB entries: 4KB 64, 2MB 0, 4MB 0, 1GB 4 Jan 17 12:18:03.730586 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Jan 17 12:18:03.730592 kernel: Spectre V2 : Spectre BHI mitigation: SW BHB clearing on vm exit Jan 17 12:18:03.730598 kernel: Spectre V2 : Spectre BHI mitigation: SW BHB clearing on syscall Jan 17 12:18:03.730603 kernel: Spectre V2 : Mitigation: Enhanced / Automatic IBRS Jan 17 12:18:03.730609 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch Jan 17 12:18:03.730615 kernel: Spectre V2 : Spectre v2 / PBRSB-eIBRS: Retire a single CALL on VMEXIT Jan 17 12:18:03.730620 kernel: RETBleed: Mitigation: Enhanced IBRS Jan 17 12:18:03.730627 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Jan 17 12:18:03.730633 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Jan 17 12:18:03.730638 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode Jan 17 12:18:03.730645 kernel: SRBDS: Unknown: Dependent on hypervisor status Jan 17 12:18:03.730651 kernel: GDS: Unknown: Dependent on hypervisor status Jan 17 12:18:03.730656 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Jan 17 12:18:03.730662 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Jan 17 12:18:03.730667 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Jan 17 12:18:03.730673 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Jan 17 12:18:03.730679 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Jan 17 12:18:03.730684 kernel: Freeing SMP alternatives memory: 32K Jan 17 12:18:03.730690 kernel: pid_max: default: 131072 minimum: 1024 Jan 17 12:18:03.730696 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Jan 17 12:18:03.730702 kernel: landlock: Up and running. Jan 17 12:18:03.730708 kernel: SELinux: Initializing. Jan 17 12:18:03.730713 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Jan 17 12:18:03.730719 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Jan 17 12:18:03.730725 kernel: smpboot: CPU0: Intel(R) Xeon(R) E-2278G CPU @ 3.40GHz (family: 0x6, model: 0x9e, stepping: 0xd) Jan 17 12:18:03.730731 kernel: RCU Tasks: Setting shift to 7 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=128. Jan 17 12:18:03.730736 kernel: RCU Tasks Rude: Setting shift to 7 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=128. Jan 17 12:18:03.730742 kernel: RCU Tasks Trace: Setting shift to 7 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=128. Jan 17 12:18:03.730749 kernel: Performance Events: Skylake events, core PMU driver. Jan 17 12:18:03.730754 kernel: core: CPUID marked event: 'cpu cycles' unavailable Jan 17 12:18:03.730760 kernel: core: CPUID marked event: 'instructions' unavailable Jan 17 12:18:03.730765 kernel: core: CPUID marked event: 'bus cycles' unavailable Jan 17 12:18:03.730771 kernel: core: CPUID marked event: 'cache references' unavailable Jan 17 12:18:03.730776 kernel: core: CPUID marked event: 'cache misses' unavailable Jan 17 12:18:03.730782 kernel: core: CPUID marked event: 'branch instructions' unavailable Jan 17 12:18:03.730787 kernel: core: CPUID marked event: 'branch misses' unavailable Jan 17 12:18:03.730794 kernel: ... version: 1 Jan 17 12:18:03.730799 kernel: ... bit width: 48 Jan 17 12:18:03.730805 kernel: ... generic registers: 4 Jan 17 12:18:03.730810 kernel: ... value mask: 0000ffffffffffff Jan 17 12:18:03.730816 kernel: ... max period: 000000007fffffff Jan 17 12:18:03.730822 kernel: ... fixed-purpose events: 0 Jan 17 12:18:03.730827 kernel: ... event mask: 000000000000000f Jan 17 12:18:03.730833 kernel: signal: max sigframe size: 1776 Jan 17 12:18:03.730838 kernel: rcu: Hierarchical SRCU implementation. Jan 17 12:18:03.730845 kernel: rcu: Max phase no-delay instances is 400. Jan 17 12:18:03.730851 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Jan 17 12:18:03.730856 kernel: smp: Bringing up secondary CPUs ... Jan 17 12:18:03.730862 kernel: smpboot: x86: Booting SMP configuration: Jan 17 12:18:03.730867 kernel: .... node #0, CPUs: #1 Jan 17 12:18:03.730873 kernel: Disabled fast string operations Jan 17 12:18:03.730878 kernel: smpboot: CPU 1 Converting physical 2 to logical package 1 Jan 17 12:18:03.730884 kernel: smpboot: CPU 1 Converting physical 0 to logical die 1 Jan 17 12:18:03.730889 kernel: smp: Brought up 1 node, 2 CPUs Jan 17 12:18:03.730895 kernel: smpboot: Max logical packages: 128 Jan 17 12:18:03.730902 kernel: smpboot: Total of 2 processors activated (13632.00 BogoMIPS) Jan 17 12:18:03.730907 kernel: devtmpfs: initialized Jan 17 12:18:03.730913 kernel: x86/mm: Memory block size: 128MB Jan 17 12:18:03.730918 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x7feff000-0x7fefffff] (4096 bytes) Jan 17 12:18:03.730924 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Jan 17 12:18:03.730930 kernel: futex hash table entries: 32768 (order: 9, 2097152 bytes, linear) Jan 17 12:18:03.730935 kernel: pinctrl core: initialized pinctrl subsystem Jan 17 12:18:03.730941 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Jan 17 12:18:03.730946 kernel: audit: initializing netlink subsys (disabled) Jan 17 12:18:03.730953 kernel: audit: type=2000 audit(1737116282.066:1): state=initialized audit_enabled=0 res=1 Jan 17 12:18:03.730959 kernel: thermal_sys: Registered thermal governor 'step_wise' Jan 17 12:18:03.730965 kernel: thermal_sys: Registered thermal governor 'user_space' Jan 17 12:18:03.730971 kernel: cpuidle: using governor menu Jan 17 12:18:03.730976 kernel: Simple Boot Flag at 0x36 set to 0x80 Jan 17 12:18:03.730982 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Jan 17 12:18:03.730987 kernel: dca service started, version 1.12.1 Jan 17 12:18:03.730993 kernel: PCI: MMCONFIG for domain 0000 [bus 00-7f] at [mem 0xf0000000-0xf7ffffff] (base 0xf0000000) Jan 17 12:18:03.730999 kernel: PCI: Using configuration type 1 for base access Jan 17 12:18:03.731006 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Jan 17 12:18:03.731012 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Jan 17 12:18:03.731017 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Jan 17 12:18:03.731023 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Jan 17 12:18:03.731028 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Jan 17 12:18:03.731034 kernel: ACPI: Added _OSI(Module Device) Jan 17 12:18:03.731039 kernel: ACPI: Added _OSI(Processor Device) Jan 17 12:18:03.731045 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Jan 17 12:18:03.731051 kernel: ACPI: Added _OSI(Processor Aggregator Device) Jan 17 12:18:03.731057 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Jan 17 12:18:03.731063 kernel: ACPI: [Firmware Bug]: BIOS _OSI(Linux) query ignored Jan 17 12:18:03.731068 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Jan 17 12:18:03.731074 kernel: ACPI: Interpreter enabled Jan 17 12:18:03.731079 kernel: ACPI: PM: (supports S0 S1 S5) Jan 17 12:18:03.731085 kernel: ACPI: Using IOAPIC for interrupt routing Jan 17 12:18:03.731090 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Jan 17 12:18:03.731096 kernel: PCI: Using E820 reservations for host bridge windows Jan 17 12:18:03.731114 kernel: ACPI: Enabled 4 GPEs in block 00 to 0F Jan 17 12:18:03.731123 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-7f]) Jan 17 12:18:03.731203 kernel: acpi PNP0A03:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Jan 17 12:18:03.731255 kernel: acpi PNP0A03:00: _OSC: platform does not support [AER LTR] Jan 17 12:18:03.731302 kernel: acpi PNP0A03:00: _OSC: OS now controls [PCIeHotplug PME PCIeCapability] Jan 17 12:18:03.731310 kernel: PCI host bridge to bus 0000:00 Jan 17 12:18:03.731358 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Jan 17 12:18:03.731408 kernel: pci_bus 0000:00: root bus resource [mem 0x000cc000-0x000dbfff window] Jan 17 12:18:03.731450 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] Jan 17 12:18:03.731492 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Jan 17 12:18:03.731533 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xfeff window] Jan 17 12:18:03.731575 kernel: pci_bus 0000:00: root bus resource [bus 00-7f] Jan 17 12:18:03.731631 kernel: pci 0000:00:00.0: [8086:7190] type 00 class 0x060000 Jan 17 12:18:03.731684 kernel: pci 0000:00:01.0: [8086:7191] type 01 class 0x060400 Jan 17 12:18:03.731738 kernel: pci 0000:00:07.0: [8086:7110] type 00 class 0x060100 Jan 17 12:18:03.731789 kernel: pci 0000:00:07.1: [8086:7111] type 00 class 0x01018a Jan 17 12:18:03.731836 kernel: pci 0000:00:07.1: reg 0x20: [io 0x1060-0x106f] Jan 17 12:18:03.731883 kernel: pci 0000:00:07.1: legacy IDE quirk: reg 0x10: [io 0x01f0-0x01f7] Jan 17 12:18:03.731929 kernel: pci 0000:00:07.1: legacy IDE quirk: reg 0x14: [io 0x03f6] Jan 17 12:18:03.731976 kernel: pci 0000:00:07.1: legacy IDE quirk: reg 0x18: [io 0x0170-0x0177] Jan 17 12:18:03.732025 kernel: pci 0000:00:07.1: legacy IDE quirk: reg 0x1c: [io 0x0376] Jan 17 12:18:03.732076 kernel: pci 0000:00:07.3: [8086:7113] type 00 class 0x068000 Jan 17 12:18:03.732172 kernel: pci 0000:00:07.3: quirk: [io 0x1000-0x103f] claimed by PIIX4 ACPI Jan 17 12:18:03.732220 kernel: pci 0000:00:07.3: quirk: [io 0x1040-0x104f] claimed by PIIX4 SMB Jan 17 12:18:03.732270 kernel: pci 0000:00:07.7: [15ad:0740] type 00 class 0x088000 Jan 17 12:18:03.732317 kernel: pci 0000:00:07.7: reg 0x10: [io 0x1080-0x10bf] Jan 17 12:18:03.732363 kernel: pci 0000:00:07.7: reg 0x14: [mem 0xfebfe000-0xfebfffff 64bit] Jan 17 12:18:03.732458 kernel: pci 0000:00:0f.0: [15ad:0405] type 00 class 0x030000 Jan 17 12:18:03.732506 kernel: pci 0000:00:0f.0: reg 0x10: [io 0x1070-0x107f] Jan 17 12:18:03.732553 kernel: pci 0000:00:0f.0: reg 0x14: [mem 0xe8000000-0xefffffff pref] Jan 17 12:18:03.732600 kernel: pci 0000:00:0f.0: reg 0x18: [mem 0xfe000000-0xfe7fffff] Jan 17 12:18:03.732647 kernel: pci 0000:00:0f.0: reg 0x30: [mem 0x00000000-0x00007fff pref] Jan 17 12:18:03.732694 kernel: pci 0000:00:0f.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Jan 17 12:18:03.732744 kernel: pci 0000:00:11.0: [15ad:0790] type 01 class 0x060401 Jan 17 12:18:03.732800 kernel: pci 0000:00:15.0: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.732849 kernel: pci 0000:00:15.0: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.732901 kernel: pci 0000:00:15.1: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.732950 kernel: pci 0000:00:15.1: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.733003 kernel: pci 0000:00:15.2: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.733052 kernel: pci 0000:00:15.2: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.733136 kernel: pci 0000:00:15.3: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.733186 kernel: pci 0000:00:15.3: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.733237 kernel: pci 0000:00:15.4: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.733287 kernel: pci 0000:00:15.4: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.733338 kernel: pci 0000:00:15.5: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.733387 kernel: pci 0000:00:15.5: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.733442 kernel: pci 0000:00:15.6: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.733490 kernel: pci 0000:00:15.6: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.733543 kernel: pci 0000:00:15.7: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.733590 kernel: pci 0000:00:15.7: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.733641 kernel: pci 0000:00:16.0: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.733689 kernel: pci 0000:00:16.0: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.733743 kernel: pci 0000:00:16.1: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.733791 kernel: pci 0000:00:16.1: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.733841 kernel: pci 0000:00:16.2: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.733888 kernel: pci 0000:00:16.2: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.733940 kernel: pci 0000:00:16.3: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.733990 kernel: pci 0000:00:16.3: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.734042 kernel: pci 0000:00:16.4: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.734090 kernel: pci 0000:00:16.4: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.734160 kernel: pci 0000:00:16.5: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.734210 kernel: pci 0000:00:16.5: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.734262 kernel: pci 0000:00:16.6: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.734310 kernel: pci 0000:00:16.6: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.734364 kernel: pci 0000:00:16.7: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.734417 kernel: pci 0000:00:16.7: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.734468 kernel: pci 0000:00:17.0: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.734515 kernel: pci 0000:00:17.0: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.734566 kernel: pci 0000:00:17.1: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.734614 kernel: pci 0000:00:17.1: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.734667 kernel: pci 0000:00:17.2: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.734715 kernel: pci 0000:00:17.2: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.734766 kernel: pci 0000:00:17.3: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.734814 kernel: pci 0000:00:17.3: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.734865 kernel: pci 0000:00:17.4: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.734913 kernel: pci 0000:00:17.4: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.734966 kernel: pci 0000:00:17.5: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.735013 kernel: pci 0000:00:17.5: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.735064 kernel: pci 0000:00:17.6: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.735147 kernel: pci 0000:00:17.6: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.735204 kernel: pci 0000:00:17.7: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.735252 kernel: pci 0000:00:17.7: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.735306 kernel: pci 0000:00:18.0: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.735354 kernel: pci 0000:00:18.0: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.735422 kernel: pci 0000:00:18.1: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.735484 kernel: pci 0000:00:18.1: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.735536 kernel: pci 0000:00:18.2: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.735584 kernel: pci 0000:00:18.2: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.735636 kernel: pci 0000:00:18.3: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.735686 kernel: pci 0000:00:18.3: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.735738 kernel: pci 0000:00:18.4: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.735787 kernel: pci 0000:00:18.4: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.735838 kernel: pci 0000:00:18.5: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.735886 kernel: pci 0000:00:18.5: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.735936 kernel: pci 0000:00:18.6: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.735986 kernel: pci 0000:00:18.6: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.736039 kernel: pci 0000:00:18.7: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.736087 kernel: pci 0000:00:18.7: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.736190 kernel: pci_bus 0000:01: extended config space not accessible Jan 17 12:18:03.736240 kernel: pci 0000:00:01.0: PCI bridge to [bus 01] Jan 17 12:18:03.736287 kernel: pci_bus 0000:02: extended config space not accessible Jan 17 12:18:03.736298 kernel: acpiphp: Slot [32] registered Jan 17 12:18:03.736304 kernel: acpiphp: Slot [33] registered Jan 17 12:18:03.736309 kernel: acpiphp: Slot [34] registered Jan 17 12:18:03.736315 kernel: acpiphp: Slot [35] registered Jan 17 12:18:03.736321 kernel: acpiphp: Slot [36] registered Jan 17 12:18:03.736326 kernel: acpiphp: Slot [37] registered Jan 17 12:18:03.736332 kernel: acpiphp: Slot [38] registered Jan 17 12:18:03.736338 kernel: acpiphp: Slot [39] registered Jan 17 12:18:03.736344 kernel: acpiphp: Slot [40] registered Jan 17 12:18:03.736351 kernel: acpiphp: Slot [41] registered Jan 17 12:18:03.736356 kernel: acpiphp: Slot [42] registered Jan 17 12:18:03.736362 kernel: acpiphp: Slot [43] registered Jan 17 12:18:03.736367 kernel: acpiphp: Slot [44] registered Jan 17 12:18:03.736373 kernel: acpiphp: Slot [45] registered Jan 17 12:18:03.736378 kernel: acpiphp: Slot [46] registered Jan 17 12:18:03.736384 kernel: acpiphp: Slot [47] registered Jan 17 12:18:03.736390 kernel: acpiphp: Slot [48] registered Jan 17 12:18:03.736395 kernel: acpiphp: Slot [49] registered Jan 17 12:18:03.736401 kernel: acpiphp: Slot [50] registered Jan 17 12:18:03.736407 kernel: acpiphp: Slot [51] registered Jan 17 12:18:03.736413 kernel: acpiphp: Slot [52] registered Jan 17 12:18:03.736419 kernel: acpiphp: Slot [53] registered Jan 17 12:18:03.736424 kernel: acpiphp: Slot [54] registered Jan 17 12:18:03.736430 kernel: acpiphp: Slot [55] registered Jan 17 12:18:03.736435 kernel: acpiphp: Slot [56] registered Jan 17 12:18:03.736441 kernel: acpiphp: Slot [57] registered Jan 17 12:18:03.736446 kernel: acpiphp: Slot [58] registered Jan 17 12:18:03.736452 kernel: acpiphp: Slot [59] registered Jan 17 12:18:03.736458 kernel: acpiphp: Slot [60] registered Jan 17 12:18:03.736464 kernel: acpiphp: Slot [61] registered Jan 17 12:18:03.736470 kernel: acpiphp: Slot [62] registered Jan 17 12:18:03.736475 kernel: acpiphp: Slot [63] registered Jan 17 12:18:03.736522 kernel: pci 0000:00:11.0: PCI bridge to [bus 02] (subtractive decode) Jan 17 12:18:03.736570 kernel: pci 0000:00:11.0: bridge window [io 0x2000-0x3fff] Jan 17 12:18:03.736616 kernel: pci 0000:00:11.0: bridge window [mem 0xfd600000-0xfdffffff] Jan 17 12:18:03.736662 kernel: pci 0000:00:11.0: bridge window [mem 0xe7b00000-0xe7ffffff 64bit pref] Jan 17 12:18:03.736709 kernel: pci 0000:00:11.0: bridge window [mem 0x000a0000-0x000bffff window] (subtractive decode) Jan 17 12:18:03.736758 kernel: pci 0000:00:11.0: bridge window [mem 0x000cc000-0x000dbfff window] (subtractive decode) Jan 17 12:18:03.736804 kernel: pci 0000:00:11.0: bridge window [mem 0xc0000000-0xfebfffff window] (subtractive decode) Jan 17 12:18:03.736850 kernel: pci 0000:00:11.0: bridge window [io 0x0000-0x0cf7 window] (subtractive decode) Jan 17 12:18:03.736896 kernel: pci 0000:00:11.0: bridge window [io 0x0d00-0xfeff window] (subtractive decode) Jan 17 12:18:03.736949 kernel: pci 0000:03:00.0: [15ad:07c0] type 00 class 0x010700 Jan 17 12:18:03.736997 kernel: pci 0000:03:00.0: reg 0x10: [io 0x4000-0x4007] Jan 17 12:18:03.737045 kernel: pci 0000:03:00.0: reg 0x14: [mem 0xfd5f8000-0xfd5fffff 64bit] Jan 17 12:18:03.737096 kernel: pci 0000:03:00.0: reg 0x30: [mem 0x00000000-0x0000ffff pref] Jan 17 12:18:03.737237 kernel: pci 0000:03:00.0: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.737285 kernel: pci 0000:03:00.0: disabling ASPM on pre-1.1 PCIe device. You can enable it with 'pcie_aspm=force' Jan 17 12:18:03.737334 kernel: pci 0000:00:15.0: PCI bridge to [bus 03] Jan 17 12:18:03.737381 kernel: pci 0000:00:15.0: bridge window [io 0x4000-0x4fff] Jan 17 12:18:03.737432 kernel: pci 0000:00:15.0: bridge window [mem 0xfd500000-0xfd5fffff] Jan 17 12:18:03.737511 kernel: pci 0000:00:15.1: PCI bridge to [bus 04] Jan 17 12:18:03.737563 kernel: pci 0000:00:15.1: bridge window [io 0x8000-0x8fff] Jan 17 12:18:03.737611 kernel: pci 0000:00:15.1: bridge window [mem 0xfd100000-0xfd1fffff] Jan 17 12:18:03.737676 kernel: pci 0000:00:15.1: bridge window [mem 0xe7800000-0xe78fffff 64bit pref] Jan 17 12:18:03.737726 kernel: pci 0000:00:15.2: PCI bridge to [bus 05] Jan 17 12:18:03.737774 kernel: pci 0000:00:15.2: bridge window [io 0xc000-0xcfff] Jan 17 12:18:03.737822 kernel: pci 0000:00:15.2: bridge window [mem 0xfcd00000-0xfcdfffff] Jan 17 12:18:03.737871 kernel: pci 0000:00:15.2: bridge window [mem 0xe7400000-0xe74fffff 64bit pref] Jan 17 12:18:03.737919 kernel: pci 0000:00:15.3: PCI bridge to [bus 06] Jan 17 12:18:03.737970 kernel: pci 0000:00:15.3: bridge window [mem 0xfc900000-0xfc9fffff] Jan 17 12:18:03.738018 kernel: pci 0000:00:15.3: bridge window [mem 0xe7000000-0xe70fffff 64bit pref] Jan 17 12:18:03.738067 kernel: pci 0000:00:15.4: PCI bridge to [bus 07] Jan 17 12:18:03.738133 kernel: pci 0000:00:15.4: bridge window [mem 0xfc500000-0xfc5fffff] Jan 17 12:18:03.738185 kernel: pci 0000:00:15.4: bridge window [mem 0xe6c00000-0xe6cfffff 64bit pref] Jan 17 12:18:03.738237 kernel: pci 0000:00:15.5: PCI bridge to [bus 08] Jan 17 12:18:03.738286 kernel: pci 0000:00:15.5: bridge window [mem 0xfc100000-0xfc1fffff] Jan 17 12:18:03.738334 kernel: pci 0000:00:15.5: bridge window [mem 0xe6800000-0xe68fffff 64bit pref] Jan 17 12:18:03.738383 kernel: pci 0000:00:15.6: PCI bridge to [bus 09] Jan 17 12:18:03.738447 kernel: pci 0000:00:15.6: bridge window [mem 0xfbd00000-0xfbdfffff] Jan 17 12:18:03.738495 kernel: pci 0000:00:15.6: bridge window [mem 0xe6400000-0xe64fffff 64bit pref] Jan 17 12:18:03.738544 kernel: pci 0000:00:15.7: PCI bridge to [bus 0a] Jan 17 12:18:03.738642 kernel: pci 0000:00:15.7: bridge window [mem 0xfb900000-0xfb9fffff] Jan 17 12:18:03.738694 kernel: pci 0000:00:15.7: bridge window [mem 0xe6000000-0xe60fffff 64bit pref] Jan 17 12:18:03.738749 kernel: pci 0000:0b:00.0: [15ad:07b0] type 00 class 0x020000 Jan 17 12:18:03.738799 kernel: pci 0000:0b:00.0: reg 0x10: [mem 0xfd4fc000-0xfd4fcfff] Jan 17 12:18:03.738848 kernel: pci 0000:0b:00.0: reg 0x14: [mem 0xfd4fd000-0xfd4fdfff] Jan 17 12:18:03.738898 kernel: pci 0000:0b:00.0: reg 0x18: [mem 0xfd4fe000-0xfd4fffff] Jan 17 12:18:03.738947 kernel: pci 0000:0b:00.0: reg 0x1c: [io 0x5000-0x500f] Jan 17 12:18:03.738996 kernel: pci 0000:0b:00.0: reg 0x30: [mem 0x00000000-0x0000ffff pref] Jan 17 12:18:03.739050 kernel: pci 0000:0b:00.0: supports D1 D2 Jan 17 12:18:03.739124 kernel: pci 0000:0b:00.0: PME# supported from D0 D1 D2 D3hot D3cold Jan 17 12:18:03.739182 kernel: pci 0000:0b:00.0: disabling ASPM on pre-1.1 PCIe device. You can enable it with 'pcie_aspm=force' Jan 17 12:18:03.739231 kernel: pci 0000:00:16.0: PCI bridge to [bus 0b] Jan 17 12:18:03.739279 kernel: pci 0000:00:16.0: bridge window [io 0x5000-0x5fff] Jan 17 12:18:03.739327 kernel: pci 0000:00:16.0: bridge window [mem 0xfd400000-0xfd4fffff] Jan 17 12:18:03.739375 kernel: pci 0000:00:16.1: PCI bridge to [bus 0c] Jan 17 12:18:03.739454 kernel: pci 0000:00:16.1: bridge window [io 0x9000-0x9fff] Jan 17 12:18:03.739503 kernel: pci 0000:00:16.1: bridge window [mem 0xfd000000-0xfd0fffff] Jan 17 12:18:03.739552 kernel: pci 0000:00:16.1: bridge window [mem 0xe7700000-0xe77fffff 64bit pref] Jan 17 12:18:03.739601 kernel: pci 0000:00:16.2: PCI bridge to [bus 0d] Jan 17 12:18:03.739652 kernel: pci 0000:00:16.2: bridge window [io 0xd000-0xdfff] Jan 17 12:18:03.739701 kernel: pci 0000:00:16.2: bridge window [mem 0xfcc00000-0xfccfffff] Jan 17 12:18:03.739750 kernel: pci 0000:00:16.2: bridge window [mem 0xe7300000-0xe73fffff 64bit pref] Jan 17 12:18:03.739800 kernel: pci 0000:00:16.3: PCI bridge to [bus 0e] Jan 17 12:18:03.739851 kernel: pci 0000:00:16.3: bridge window [mem 0xfc800000-0xfc8fffff] Jan 17 12:18:03.739900 kernel: pci 0000:00:16.3: bridge window [mem 0xe6f00000-0xe6ffffff 64bit pref] Jan 17 12:18:03.739949 kernel: pci 0000:00:16.4: PCI bridge to [bus 0f] Jan 17 12:18:03.739998 kernel: pci 0000:00:16.4: bridge window [mem 0xfc400000-0xfc4fffff] Jan 17 12:18:03.740047 kernel: pci 0000:00:16.4: bridge window [mem 0xe6b00000-0xe6bfffff 64bit pref] Jan 17 12:18:03.740097 kernel: pci 0000:00:16.5: PCI bridge to [bus 10] Jan 17 12:18:03.740323 kernel: pci 0000:00:16.5: bridge window [mem 0xfc000000-0xfc0fffff] Jan 17 12:18:03.740373 kernel: pci 0000:00:16.5: bridge window [mem 0xe6700000-0xe67fffff 64bit pref] Jan 17 12:18:03.740426 kernel: pci 0000:00:16.6: PCI bridge to [bus 11] Jan 17 12:18:03.740477 kernel: pci 0000:00:16.6: bridge window [mem 0xfbc00000-0xfbcfffff] Jan 17 12:18:03.740526 kernel: pci 0000:00:16.6: bridge window [mem 0xe6300000-0xe63fffff 64bit pref] Jan 17 12:18:03.740575 kernel: pci 0000:00:16.7: PCI bridge to [bus 12] Jan 17 12:18:03.740624 kernel: pci 0000:00:16.7: bridge window [mem 0xfb800000-0xfb8fffff] Jan 17 12:18:03.740673 kernel: pci 0000:00:16.7: bridge window [mem 0xe5f00000-0xe5ffffff 64bit pref] Jan 17 12:18:03.740722 kernel: pci 0000:00:17.0: PCI bridge to [bus 13] Jan 17 12:18:03.740770 kernel: pci 0000:00:17.0: bridge window [io 0x6000-0x6fff] Jan 17 12:18:03.740821 kernel: pci 0000:00:17.0: bridge window [mem 0xfd300000-0xfd3fffff] Jan 17 12:18:03.740869 kernel: pci 0000:00:17.0: bridge window [mem 0xe7a00000-0xe7afffff 64bit pref] Jan 17 12:18:03.740918 kernel: pci 0000:00:17.1: PCI bridge to [bus 14] Jan 17 12:18:03.740968 kernel: pci 0000:00:17.1: bridge window [io 0xa000-0xafff] Jan 17 12:18:03.741016 kernel: pci 0000:00:17.1: bridge window [mem 0xfcf00000-0xfcffffff] Jan 17 12:18:03.741065 kernel: pci 0000:00:17.1: bridge window [mem 0xe7600000-0xe76fffff 64bit pref] Jan 17 12:18:03.741149 kernel: pci 0000:00:17.2: PCI bridge to [bus 15] Jan 17 12:18:03.741203 kernel: pci 0000:00:17.2: bridge window [io 0xe000-0xefff] Jan 17 12:18:03.741252 kernel: pci 0000:00:17.2: bridge window [mem 0xfcb00000-0xfcbfffff] Jan 17 12:18:03.741300 kernel: pci 0000:00:17.2: bridge window [mem 0xe7200000-0xe72fffff 64bit pref] Jan 17 12:18:03.741350 kernel: pci 0000:00:17.3: PCI bridge to [bus 16] Jan 17 12:18:03.741402 kernel: pci 0000:00:17.3: bridge window [mem 0xfc700000-0xfc7fffff] Jan 17 12:18:03.741451 kernel: pci 0000:00:17.3: bridge window [mem 0xe6e00000-0xe6efffff 64bit pref] Jan 17 12:18:03.741501 kernel: pci 0000:00:17.4: PCI bridge to [bus 17] Jan 17 12:18:03.741849 kernel: pci 0000:00:17.4: bridge window [mem 0xfc300000-0xfc3fffff] Jan 17 12:18:03.743843 kernel: pci 0000:00:17.4: bridge window [mem 0xe6a00000-0xe6afffff 64bit pref] Jan 17 12:18:03.743906 kernel: pci 0000:00:17.5: PCI bridge to [bus 18] Jan 17 12:18:03.743959 kernel: pci 0000:00:17.5: bridge window [mem 0xfbf00000-0xfbffffff] Jan 17 12:18:03.744009 kernel: pci 0000:00:17.5: bridge window [mem 0xe6600000-0xe66fffff 64bit pref] Jan 17 12:18:03.744061 kernel: pci 0000:00:17.6: PCI bridge to [bus 19] Jan 17 12:18:03.744137 kernel: pci 0000:00:17.6: bridge window [mem 0xfbb00000-0xfbbfffff] Jan 17 12:18:03.744186 kernel: pci 0000:00:17.6: bridge window [mem 0xe6200000-0xe62fffff 64bit pref] Jan 17 12:18:03.744236 kernel: pci 0000:00:17.7: PCI bridge to [bus 1a] Jan 17 12:18:03.744287 kernel: pci 0000:00:17.7: bridge window [mem 0xfb700000-0xfb7fffff] Jan 17 12:18:03.744335 kernel: pci 0000:00:17.7: bridge window [mem 0xe5e00000-0xe5efffff 64bit pref] Jan 17 12:18:03.744384 kernel: pci 0000:00:18.0: PCI bridge to [bus 1b] Jan 17 12:18:03.744437 kernel: pci 0000:00:18.0: bridge window [io 0x7000-0x7fff] Jan 17 12:18:03.744484 kernel: pci 0000:00:18.0: bridge window [mem 0xfd200000-0xfd2fffff] Jan 17 12:18:03.744532 kernel: pci 0000:00:18.0: bridge window [mem 0xe7900000-0xe79fffff 64bit pref] Jan 17 12:18:03.744583 kernel: pci 0000:00:18.1: PCI bridge to [bus 1c] Jan 17 12:18:03.744631 kernel: pci 0000:00:18.1: bridge window [io 0xb000-0xbfff] Jan 17 12:18:03.744681 kernel: pci 0000:00:18.1: bridge window [mem 0xfce00000-0xfcefffff] Jan 17 12:18:03.744729 kernel: pci 0000:00:18.1: bridge window [mem 0xe7500000-0xe75fffff 64bit pref] Jan 17 12:18:03.744779 kernel: pci 0000:00:18.2: PCI bridge to [bus 1d] Jan 17 12:18:03.744827 kernel: pci 0000:00:18.2: bridge window [mem 0xfca00000-0xfcafffff] Jan 17 12:18:03.744875 kernel: pci 0000:00:18.2: bridge window [mem 0xe7100000-0xe71fffff 64bit pref] Jan 17 12:18:03.744925 kernel: pci 0000:00:18.3: PCI bridge to [bus 1e] Jan 17 12:18:03.744973 kernel: pci 0000:00:18.3: bridge window [mem 0xfc600000-0xfc6fffff] Jan 17 12:18:03.745021 kernel: pci 0000:00:18.3: bridge window [mem 0xe6d00000-0xe6dfffff 64bit pref] Jan 17 12:18:03.745073 kernel: pci 0000:00:18.4: PCI bridge to [bus 1f] Jan 17 12:18:03.746200 kernel: pci 0000:00:18.4: bridge window [mem 0xfc200000-0xfc2fffff] Jan 17 12:18:03.746258 kernel: pci 0000:00:18.4: bridge window [mem 0xe6900000-0xe69fffff 64bit pref] Jan 17 12:18:03.746329 kernel: pci 0000:00:18.5: PCI bridge to [bus 20] Jan 17 12:18:03.746382 kernel: pci 0000:00:18.5: bridge window [mem 0xfbe00000-0xfbefffff] Jan 17 12:18:03.746431 kernel: pci 0000:00:18.5: bridge window [mem 0xe6500000-0xe65fffff 64bit pref] Jan 17 12:18:03.746481 kernel: pci 0000:00:18.6: PCI bridge to [bus 21] Jan 17 12:18:03.746530 kernel: pci 0000:00:18.6: bridge window [mem 0xfba00000-0xfbafffff] Jan 17 12:18:03.746582 kernel: pci 0000:00:18.6: bridge window [mem 0xe6100000-0xe61fffff 64bit pref] Jan 17 12:18:03.746632 kernel: pci 0000:00:18.7: PCI bridge to [bus 22] Jan 17 12:18:03.746681 kernel: pci 0000:00:18.7: bridge window [mem 0xfb600000-0xfb6fffff] Jan 17 12:18:03.746729 kernel: pci 0000:00:18.7: bridge window [mem 0xe5d00000-0xe5dfffff 64bit pref] Jan 17 12:18:03.746737 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 9 Jan 17 12:18:03.746743 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 0 Jan 17 12:18:03.746749 kernel: ACPI: PCI: Interrupt link LNKB disabled Jan 17 12:18:03.746755 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Jan 17 12:18:03.746763 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 10 Jan 17 12:18:03.746768 kernel: iommu: Default domain type: Translated Jan 17 12:18:03.746774 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Jan 17 12:18:03.746780 kernel: PCI: Using ACPI for IRQ routing Jan 17 12:18:03.746786 kernel: PCI: pci_cache_line_size set to 64 bytes Jan 17 12:18:03.746792 kernel: e820: reserve RAM buffer [mem 0x0009ec00-0x0009ffff] Jan 17 12:18:03.746797 kernel: e820: reserve RAM buffer [mem 0x7fee0000-0x7fffffff] Jan 17 12:18:03.746845 kernel: pci 0000:00:0f.0: vgaarb: setting as boot VGA device Jan 17 12:18:03.746894 kernel: pci 0000:00:0f.0: vgaarb: bridge control possible Jan 17 12:18:03.746944 kernel: pci 0000:00:0f.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Jan 17 12:18:03.746952 kernel: vgaarb: loaded Jan 17 12:18:03.746958 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 Jan 17 12:18:03.746964 kernel: hpet0: 16 comparators, 64-bit 14.318180 MHz counter Jan 17 12:18:03.746970 kernel: clocksource: Switched to clocksource tsc-early Jan 17 12:18:03.746976 kernel: VFS: Disk quotas dquot_6.6.0 Jan 17 12:18:03.746982 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Jan 17 12:18:03.746987 kernel: pnp: PnP ACPI init Jan 17 12:18:03.747037 kernel: system 00:00: [io 0x1000-0x103f] has been reserved Jan 17 12:18:03.747085 kernel: system 00:00: [io 0x1040-0x104f] has been reserved Jan 17 12:18:03.748141 kernel: system 00:00: [io 0x0cf0-0x0cf1] has been reserved Jan 17 12:18:03.748193 kernel: system 00:04: [mem 0xfed00000-0xfed003ff] has been reserved Jan 17 12:18:03.748262 kernel: pnp 00:06: [dma 2] Jan 17 12:18:03.748310 kernel: system 00:07: [io 0xfce0-0xfcff] has been reserved Jan 17 12:18:03.748355 kernel: system 00:07: [mem 0xf0000000-0xf7ffffff] has been reserved Jan 17 12:18:03.748448 kernel: system 00:07: [mem 0xfe800000-0xfe9fffff] has been reserved Jan 17 12:18:03.748458 kernel: pnp: PnP ACPI: found 8 devices Jan 17 12:18:03.748464 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Jan 17 12:18:03.748470 kernel: NET: Registered PF_INET protocol family Jan 17 12:18:03.748476 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) Jan 17 12:18:03.748482 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) Jan 17 12:18:03.748488 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Jan 17 12:18:03.748494 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) Jan 17 12:18:03.748502 kernel: TCP bind hash table entries: 16384 (order: 7, 524288 bytes, linear) Jan 17 12:18:03.748508 kernel: TCP: Hash tables configured (established 16384 bind 16384) Jan 17 12:18:03.748514 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) Jan 17 12:18:03.748519 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) Jan 17 12:18:03.748525 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Jan 17 12:18:03.748531 kernel: NET: Registered PF_XDP protocol family Jan 17 12:18:03.748582 kernel: pci 0000:00:15.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 03] add_size 200000 add_align 100000 Jan 17 12:18:03.748632 kernel: pci 0000:00:15.3: bridge window [io 0x1000-0x0fff] to [bus 06] add_size 1000 Jan 17 12:18:03.748684 kernel: pci 0000:00:15.4: bridge window [io 0x1000-0x0fff] to [bus 07] add_size 1000 Jan 17 12:18:03.748735 kernel: pci 0000:00:15.5: bridge window [io 0x1000-0x0fff] to [bus 08] add_size 1000 Jan 17 12:18:03.748784 kernel: pci 0000:00:15.6: bridge window [io 0x1000-0x0fff] to [bus 09] add_size 1000 Jan 17 12:18:03.748833 kernel: pci 0000:00:15.7: bridge window [io 0x1000-0x0fff] to [bus 0a] add_size 1000 Jan 17 12:18:03.748883 kernel: pci 0000:00:16.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 0b] add_size 200000 add_align 100000 Jan 17 12:18:03.748932 kernel: pci 0000:00:16.3: bridge window [io 0x1000-0x0fff] to [bus 0e] add_size 1000 Jan 17 12:18:03.748984 kernel: pci 0000:00:16.4: bridge window [io 0x1000-0x0fff] to [bus 0f] add_size 1000 Jan 17 12:18:03.749033 kernel: pci 0000:00:16.5: bridge window [io 0x1000-0x0fff] to [bus 10] add_size 1000 Jan 17 12:18:03.749097 kernel: pci 0000:00:16.6: bridge window [io 0x1000-0x0fff] to [bus 11] add_size 1000 Jan 17 12:18:03.750174 kernel: pci 0000:00:16.7: bridge window [io 0x1000-0x0fff] to [bus 12] add_size 1000 Jan 17 12:18:03.750224 kernel: pci 0000:00:17.3: bridge window [io 0x1000-0x0fff] to [bus 16] add_size 1000 Jan 17 12:18:03.750273 kernel: pci 0000:00:17.4: bridge window [io 0x1000-0x0fff] to [bus 17] add_size 1000 Jan 17 12:18:03.750325 kernel: pci 0000:00:17.5: bridge window [io 0x1000-0x0fff] to [bus 18] add_size 1000 Jan 17 12:18:03.750372 kernel: pci 0000:00:17.6: bridge window [io 0x1000-0x0fff] to [bus 19] add_size 1000 Jan 17 12:18:03.750491 kernel: pci 0000:00:17.7: bridge window [io 0x1000-0x0fff] to [bus 1a] add_size 1000 Jan 17 12:18:03.750539 kernel: pci 0000:00:18.2: bridge window [io 0x1000-0x0fff] to [bus 1d] add_size 1000 Jan 17 12:18:03.750587 kernel: pci 0000:00:18.3: bridge window [io 0x1000-0x0fff] to [bus 1e] add_size 1000 Jan 17 12:18:03.750634 kernel: pci 0000:00:18.4: bridge window [io 0x1000-0x0fff] to [bus 1f] add_size 1000 Jan 17 12:18:03.750684 kernel: pci 0000:00:18.5: bridge window [io 0x1000-0x0fff] to [bus 20] add_size 1000 Jan 17 12:18:03.750731 kernel: pci 0000:00:18.6: bridge window [io 0x1000-0x0fff] to [bus 21] add_size 1000 Jan 17 12:18:03.750779 kernel: pci 0000:00:18.7: bridge window [io 0x1000-0x0fff] to [bus 22] add_size 1000 Jan 17 12:18:03.750826 kernel: pci 0000:00:15.0: BAR 15: assigned [mem 0xc0000000-0xc01fffff 64bit pref] Jan 17 12:18:03.750891 kernel: pci 0000:00:16.0: BAR 15: assigned [mem 0xc0200000-0xc03fffff 64bit pref] Jan 17 12:18:03.750940 kernel: pci 0000:00:15.3: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.751045 kernel: pci 0000:00:15.3: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.752153 kernel: pci 0000:00:15.4: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.752205 kernel: pci 0000:00:15.4: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.752254 kernel: pci 0000:00:15.5: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.752301 kernel: pci 0000:00:15.5: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.752347 kernel: pci 0000:00:15.6: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.752395 kernel: pci 0000:00:15.6: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.752442 kernel: pci 0000:00:15.7: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.752491 kernel: pci 0000:00:15.7: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.752538 kernel: pci 0000:00:16.3: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.752584 kernel: pci 0000:00:16.3: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.752630 kernel: pci 0000:00:16.4: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.752677 kernel: pci 0000:00:16.4: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.752724 kernel: pci 0000:00:16.5: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.752770 kernel: pci 0000:00:16.5: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.752817 kernel: pci 0000:00:16.6: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.752866 kernel: pci 0000:00:16.6: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.752913 kernel: pci 0000:00:16.7: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.752960 kernel: pci 0000:00:16.7: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.753005 kernel: pci 0000:00:17.3: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.753052 kernel: pci 0000:00:17.3: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.754110 kernel: pci 0000:00:17.4: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.754169 kernel: pci 0000:00:17.4: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.754216 kernel: pci 0000:00:17.5: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.754266 kernel: pci 0000:00:17.5: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.754314 kernel: pci 0000:00:17.6: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.754361 kernel: pci 0000:00:17.6: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.754412 kernel: pci 0000:00:17.7: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.754458 kernel: pci 0000:00:17.7: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.754505 kernel: pci 0000:00:18.2: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.754553 kernel: pci 0000:00:18.2: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.754600 kernel: pci 0000:00:18.3: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.754648 kernel: pci 0000:00:18.3: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.754698 kernel: pci 0000:00:18.4: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.754746 kernel: pci 0000:00:18.4: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.754794 kernel: pci 0000:00:18.5: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.754842 kernel: pci 0000:00:18.5: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.754888 kernel: pci 0000:00:18.6: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.754935 kernel: pci 0000:00:18.6: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.754982 kernel: pci 0000:00:18.7: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.755030 kernel: pci 0000:00:18.7: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.755080 kernel: pci 0000:00:18.7: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.755135 kernel: pci 0000:00:18.7: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.755182 kernel: pci 0000:00:18.6: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.755229 kernel: pci 0000:00:18.6: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.755277 kernel: pci 0000:00:18.5: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.755323 kernel: pci 0000:00:18.5: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.755369 kernel: pci 0000:00:18.4: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.755417 kernel: pci 0000:00:18.4: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.755464 kernel: pci 0000:00:18.3: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.755514 kernel: pci 0000:00:18.3: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.755561 kernel: pci 0000:00:18.2: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.755608 kernel: pci 0000:00:18.2: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.755654 kernel: pci 0000:00:17.7: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.755701 kernel: pci 0000:00:17.7: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.755748 kernel: pci 0000:00:17.6: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.755794 kernel: pci 0000:00:17.6: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.755841 kernel: pci 0000:00:17.5: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.755888 kernel: pci 0000:00:17.5: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.755955 kernel: pci 0000:00:17.4: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.756002 kernel: pci 0000:00:17.4: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.756050 kernel: pci 0000:00:17.3: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.756097 kernel: pci 0000:00:17.3: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.758183 kernel: pci 0000:00:16.7: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.758233 kernel: pci 0000:00:16.7: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.758282 kernel: pci 0000:00:16.6: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.758330 kernel: pci 0000:00:16.6: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.758378 kernel: pci 0000:00:16.5: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.758453 kernel: pci 0000:00:16.5: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.758504 kernel: pci 0000:00:16.4: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.758551 kernel: pci 0000:00:16.4: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.758598 kernel: pci 0000:00:16.3: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.758646 kernel: pci 0000:00:16.3: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.758694 kernel: pci 0000:00:15.7: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.758742 kernel: pci 0000:00:15.7: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.758789 kernel: pci 0000:00:15.6: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.758837 kernel: pci 0000:00:15.6: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.758884 kernel: pci 0000:00:15.5: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.758934 kernel: pci 0000:00:15.5: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.758982 kernel: pci 0000:00:15.4: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.759030 kernel: pci 0000:00:15.4: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.759078 kernel: pci 0000:00:15.3: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.759139 kernel: pci 0000:00:15.3: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.759190 kernel: pci 0000:00:01.0: PCI bridge to [bus 01] Jan 17 12:18:03.759239 kernel: pci 0000:00:11.0: PCI bridge to [bus 02] Jan 17 12:18:03.759287 kernel: pci 0000:00:11.0: bridge window [io 0x2000-0x3fff] Jan 17 12:18:03.759335 kernel: pci 0000:00:11.0: bridge window [mem 0xfd600000-0xfdffffff] Jan 17 12:18:03.759382 kernel: pci 0000:00:11.0: bridge window [mem 0xe7b00000-0xe7ffffff 64bit pref] Jan 17 12:18:03.759437 kernel: pci 0000:03:00.0: BAR 6: assigned [mem 0xfd500000-0xfd50ffff pref] Jan 17 12:18:03.759486 kernel: pci 0000:00:15.0: PCI bridge to [bus 03] Jan 17 12:18:03.759535 kernel: pci 0000:00:15.0: bridge window [io 0x4000-0x4fff] Jan 17 12:18:03.759583 kernel: pci 0000:00:15.0: bridge window [mem 0xfd500000-0xfd5fffff] Jan 17 12:18:03.759631 kernel: pci 0000:00:15.0: bridge window [mem 0xc0000000-0xc01fffff 64bit pref] Jan 17 12:18:03.759680 kernel: pci 0000:00:15.1: PCI bridge to [bus 04] Jan 17 12:18:03.759728 kernel: pci 0000:00:15.1: bridge window [io 0x8000-0x8fff] Jan 17 12:18:03.759777 kernel: pci 0000:00:15.1: bridge window [mem 0xfd100000-0xfd1fffff] Jan 17 12:18:03.759829 kernel: pci 0000:00:15.1: bridge window [mem 0xe7800000-0xe78fffff 64bit pref] Jan 17 12:18:03.759877 kernel: pci 0000:00:15.2: PCI bridge to [bus 05] Jan 17 12:18:03.759925 kernel: pci 0000:00:15.2: bridge window [io 0xc000-0xcfff] Jan 17 12:18:03.759974 kernel: pci 0000:00:15.2: bridge window [mem 0xfcd00000-0xfcdfffff] Jan 17 12:18:03.760021 kernel: pci 0000:00:15.2: bridge window [mem 0xe7400000-0xe74fffff 64bit pref] Jan 17 12:18:03.760069 kernel: pci 0000:00:15.3: PCI bridge to [bus 06] Jan 17 12:18:03.761026 kernel: pci 0000:00:15.3: bridge window [mem 0xfc900000-0xfc9fffff] Jan 17 12:18:03.761084 kernel: pci 0000:00:15.3: bridge window [mem 0xe7000000-0xe70fffff 64bit pref] Jan 17 12:18:03.761149 kernel: pci 0000:00:15.4: PCI bridge to [bus 07] Jan 17 12:18:03.761218 kernel: pci 0000:00:15.4: bridge window [mem 0xfc500000-0xfc5fffff] Jan 17 12:18:03.761267 kernel: pci 0000:00:15.4: bridge window [mem 0xe6c00000-0xe6cfffff 64bit pref] Jan 17 12:18:03.761318 kernel: pci 0000:00:15.5: PCI bridge to [bus 08] Jan 17 12:18:03.761366 kernel: pci 0000:00:15.5: bridge window [mem 0xfc100000-0xfc1fffff] Jan 17 12:18:03.761418 kernel: pci 0000:00:15.5: bridge window [mem 0xe6800000-0xe68fffff 64bit pref] Jan 17 12:18:03.761467 kernel: pci 0000:00:15.6: PCI bridge to [bus 09] Jan 17 12:18:03.761518 kernel: pci 0000:00:15.6: bridge window [mem 0xfbd00000-0xfbdfffff] Jan 17 12:18:03.761567 kernel: pci 0000:00:15.6: bridge window [mem 0xe6400000-0xe64fffff 64bit pref] Jan 17 12:18:03.761615 kernel: pci 0000:00:15.7: PCI bridge to [bus 0a] Jan 17 12:18:03.761663 kernel: pci 0000:00:15.7: bridge window [mem 0xfb900000-0xfb9fffff] Jan 17 12:18:03.761711 kernel: pci 0000:00:15.7: bridge window [mem 0xe6000000-0xe60fffff 64bit pref] Jan 17 12:18:03.761763 kernel: pci 0000:0b:00.0: BAR 6: assigned [mem 0xfd400000-0xfd40ffff pref] Jan 17 12:18:03.761811 kernel: pci 0000:00:16.0: PCI bridge to [bus 0b] Jan 17 12:18:03.761859 kernel: pci 0000:00:16.0: bridge window [io 0x5000-0x5fff] Jan 17 12:18:03.761908 kernel: pci 0000:00:16.0: bridge window [mem 0xfd400000-0xfd4fffff] Jan 17 12:18:03.761959 kernel: pci 0000:00:16.0: bridge window [mem 0xc0200000-0xc03fffff 64bit pref] Jan 17 12:18:03.762008 kernel: pci 0000:00:16.1: PCI bridge to [bus 0c] Jan 17 12:18:03.762056 kernel: pci 0000:00:16.1: bridge window [io 0x9000-0x9fff] Jan 17 12:18:03.762114 kernel: pci 0000:00:16.1: bridge window [mem 0xfd000000-0xfd0fffff] Jan 17 12:18:03.762167 kernel: pci 0000:00:16.1: bridge window [mem 0xe7700000-0xe77fffff 64bit pref] Jan 17 12:18:03.762215 kernel: pci 0000:00:16.2: PCI bridge to [bus 0d] Jan 17 12:18:03.762273 kernel: pci 0000:00:16.2: bridge window [io 0xd000-0xdfff] Jan 17 12:18:03.762345 kernel: pci 0000:00:16.2: bridge window [mem 0xfcc00000-0xfccfffff] Jan 17 12:18:03.762396 kernel: pci 0000:00:16.2: bridge window [mem 0xe7300000-0xe73fffff 64bit pref] Jan 17 12:18:03.762444 kernel: pci 0000:00:16.3: PCI bridge to [bus 0e] Jan 17 12:18:03.762496 kernel: pci 0000:00:16.3: bridge window [mem 0xfc800000-0xfc8fffff] Jan 17 12:18:03.762544 kernel: pci 0000:00:16.3: bridge window [mem 0xe6f00000-0xe6ffffff 64bit pref] Jan 17 12:18:03.762592 kernel: pci 0000:00:16.4: PCI bridge to [bus 0f] Jan 17 12:18:03.762641 kernel: pci 0000:00:16.4: bridge window [mem 0xfc400000-0xfc4fffff] Jan 17 12:18:03.762690 kernel: pci 0000:00:16.4: bridge window [mem 0xe6b00000-0xe6bfffff 64bit pref] Jan 17 12:18:03.762737 kernel: pci 0000:00:16.5: PCI bridge to [bus 10] Jan 17 12:18:03.762785 kernel: pci 0000:00:16.5: bridge window [mem 0xfc000000-0xfc0fffff] Jan 17 12:18:03.762834 kernel: pci 0000:00:16.5: bridge window [mem 0xe6700000-0xe67fffff 64bit pref] Jan 17 12:18:03.762882 kernel: pci 0000:00:16.6: PCI bridge to [bus 11] Jan 17 12:18:03.762933 kernel: pci 0000:00:16.6: bridge window [mem 0xfbc00000-0xfbcfffff] Jan 17 12:18:03.762992 kernel: pci 0000:00:16.6: bridge window [mem 0xe6300000-0xe63fffff 64bit pref] Jan 17 12:18:03.763050 kernel: pci 0000:00:16.7: PCI bridge to [bus 12] Jan 17 12:18:03.763149 kernel: pci 0000:00:16.7: bridge window [mem 0xfb800000-0xfb8fffff] Jan 17 12:18:03.763205 kernel: pci 0000:00:16.7: bridge window [mem 0xe5f00000-0xe5ffffff 64bit pref] Jan 17 12:18:03.763254 kernel: pci 0000:00:17.0: PCI bridge to [bus 13] Jan 17 12:18:03.763302 kernel: pci 0000:00:17.0: bridge window [io 0x6000-0x6fff] Jan 17 12:18:03.763350 kernel: pci 0000:00:17.0: bridge window [mem 0xfd300000-0xfd3fffff] Jan 17 12:18:03.763398 kernel: pci 0000:00:17.0: bridge window [mem 0xe7a00000-0xe7afffff 64bit pref] Jan 17 12:18:03.763446 kernel: pci 0000:00:17.1: PCI bridge to [bus 14] Jan 17 12:18:03.763497 kernel: pci 0000:00:17.1: bridge window [io 0xa000-0xafff] Jan 17 12:18:03.763546 kernel: pci 0000:00:17.1: bridge window [mem 0xfcf00000-0xfcffffff] Jan 17 12:18:03.763594 kernel: pci 0000:00:17.1: bridge window [mem 0xe7600000-0xe76fffff 64bit pref] Jan 17 12:18:03.763643 kernel: pci 0000:00:17.2: PCI bridge to [bus 15] Jan 17 12:18:03.763692 kernel: pci 0000:00:17.2: bridge window [io 0xe000-0xefff] Jan 17 12:18:03.763741 kernel: pci 0000:00:17.2: bridge window [mem 0xfcb00000-0xfcbfffff] Jan 17 12:18:03.763789 kernel: pci 0000:00:17.2: bridge window [mem 0xe7200000-0xe72fffff 64bit pref] Jan 17 12:18:03.763849 kernel: pci 0000:00:17.3: PCI bridge to [bus 16] Jan 17 12:18:03.763898 kernel: pci 0000:00:17.3: bridge window [mem 0xfc700000-0xfc7fffff] Jan 17 12:18:03.763948 kernel: pci 0000:00:17.3: bridge window [mem 0xe6e00000-0xe6efffff 64bit pref] Jan 17 12:18:03.763997 kernel: pci 0000:00:17.4: PCI bridge to [bus 17] Jan 17 12:18:03.764047 kernel: pci 0000:00:17.4: bridge window [mem 0xfc300000-0xfc3fffff] Jan 17 12:18:03.764138 kernel: pci 0000:00:17.4: bridge window [mem 0xe6a00000-0xe6afffff 64bit pref] Jan 17 12:18:03.764191 kernel: pci 0000:00:17.5: PCI bridge to [bus 18] Jan 17 12:18:03.764240 kernel: pci 0000:00:17.5: bridge window [mem 0xfbf00000-0xfbffffff] Jan 17 12:18:03.764288 kernel: pci 0000:00:17.5: bridge window [mem 0xe6600000-0xe66fffff 64bit pref] Jan 17 12:18:03.764336 kernel: pci 0000:00:17.6: PCI bridge to [bus 19] Jan 17 12:18:03.764384 kernel: pci 0000:00:17.6: bridge window [mem 0xfbb00000-0xfbbfffff] Jan 17 12:18:03.764437 kernel: pci 0000:00:17.6: bridge window [mem 0xe6200000-0xe62fffff 64bit pref] Jan 17 12:18:03.764488 kernel: pci 0000:00:17.7: PCI bridge to [bus 1a] Jan 17 12:18:03.764535 kernel: pci 0000:00:17.7: bridge window [mem 0xfb700000-0xfb7fffff] Jan 17 12:18:03.764583 kernel: pci 0000:00:17.7: bridge window [mem 0xe5e00000-0xe5efffff 64bit pref] Jan 17 12:18:03.764630 kernel: pci 0000:00:18.0: PCI bridge to [bus 1b] Jan 17 12:18:03.764678 kernel: pci 0000:00:18.0: bridge window [io 0x7000-0x7fff] Jan 17 12:18:03.764726 kernel: pci 0000:00:18.0: bridge window [mem 0xfd200000-0xfd2fffff] Jan 17 12:18:03.764773 kernel: pci 0000:00:18.0: bridge window [mem 0xe7900000-0xe79fffff 64bit pref] Jan 17 12:18:03.764822 kernel: pci 0000:00:18.1: PCI bridge to [bus 1c] Jan 17 12:18:03.764871 kernel: pci 0000:00:18.1: bridge window [io 0xb000-0xbfff] Jan 17 12:18:03.764921 kernel: pci 0000:00:18.1: bridge window [mem 0xfce00000-0xfcefffff] Jan 17 12:18:03.764969 kernel: pci 0000:00:18.1: bridge window [mem 0xe7500000-0xe75fffff 64bit pref] Jan 17 12:18:03.765017 kernel: pci 0000:00:18.2: PCI bridge to [bus 1d] Jan 17 12:18:03.765064 kernel: pci 0000:00:18.2: bridge window [mem 0xfca00000-0xfcafffff] Jan 17 12:18:03.765176 kernel: pci 0000:00:18.2: bridge window [mem 0xe7100000-0xe71fffff 64bit pref] Jan 17 12:18:03.765225 kernel: pci 0000:00:18.3: PCI bridge to [bus 1e] Jan 17 12:18:03.765272 kernel: pci 0000:00:18.3: bridge window [mem 0xfc600000-0xfc6fffff] Jan 17 12:18:03.765320 kernel: pci 0000:00:18.3: bridge window [mem 0xe6d00000-0xe6dfffff 64bit pref] Jan 17 12:18:03.765368 kernel: pci 0000:00:18.4: PCI bridge to [bus 1f] Jan 17 12:18:03.765416 kernel: pci 0000:00:18.4: bridge window [mem 0xfc200000-0xfc2fffff] Jan 17 12:18:03.765473 kernel: pci 0000:00:18.4: bridge window [mem 0xe6900000-0xe69fffff 64bit pref] Jan 17 12:18:03.765526 kernel: pci 0000:00:18.5: PCI bridge to [bus 20] Jan 17 12:18:03.765574 kernel: pci 0000:00:18.5: bridge window [mem 0xfbe00000-0xfbefffff] Jan 17 12:18:03.765621 kernel: pci 0000:00:18.5: bridge window [mem 0xe6500000-0xe65fffff 64bit pref] Jan 17 12:18:03.765668 kernel: pci 0000:00:18.6: PCI bridge to [bus 21] Jan 17 12:18:03.765716 kernel: pci 0000:00:18.6: bridge window [mem 0xfba00000-0xfbafffff] Jan 17 12:18:03.765762 kernel: pci 0000:00:18.6: bridge window [mem 0xe6100000-0xe61fffff 64bit pref] Jan 17 12:18:03.765810 kernel: pci 0000:00:18.7: PCI bridge to [bus 22] Jan 17 12:18:03.765858 kernel: pci 0000:00:18.7: bridge window [mem 0xfb600000-0xfb6fffff] Jan 17 12:18:03.765908 kernel: pci 0000:00:18.7: bridge window [mem 0xe5d00000-0xe5dfffff 64bit pref] Jan 17 12:18:03.765956 kernel: pci_bus 0000:00: resource 4 [mem 0x000a0000-0x000bffff window] Jan 17 12:18:03.766000 kernel: pci_bus 0000:00: resource 5 [mem 0x000cc000-0x000dbfff window] Jan 17 12:18:03.766042 kernel: pci_bus 0000:00: resource 6 [mem 0xc0000000-0xfebfffff window] Jan 17 12:18:03.766084 kernel: pci_bus 0000:00: resource 7 [io 0x0000-0x0cf7 window] Jan 17 12:18:03.767254 kernel: pci_bus 0000:00: resource 8 [io 0x0d00-0xfeff window] Jan 17 12:18:03.767309 kernel: pci_bus 0000:02: resource 0 [io 0x2000-0x3fff] Jan 17 12:18:03.767354 kernel: pci_bus 0000:02: resource 1 [mem 0xfd600000-0xfdffffff] Jan 17 12:18:03.767406 kernel: pci_bus 0000:02: resource 2 [mem 0xe7b00000-0xe7ffffff 64bit pref] Jan 17 12:18:03.767450 kernel: pci_bus 0000:02: resource 4 [mem 0x000a0000-0x000bffff window] Jan 17 12:18:03.767493 kernel: pci_bus 0000:02: resource 5 [mem 0x000cc000-0x000dbfff window] Jan 17 12:18:03.767536 kernel: pci_bus 0000:02: resource 6 [mem 0xc0000000-0xfebfffff window] Jan 17 12:18:03.767579 kernel: pci_bus 0000:02: resource 7 [io 0x0000-0x0cf7 window] Jan 17 12:18:03.767622 kernel: pci_bus 0000:02: resource 8 [io 0x0d00-0xfeff window] Jan 17 12:18:03.767669 kernel: pci_bus 0000:03: resource 0 [io 0x4000-0x4fff] Jan 17 12:18:03.767716 kernel: pci_bus 0000:03: resource 1 [mem 0xfd500000-0xfd5fffff] Jan 17 12:18:03.767760 kernel: pci_bus 0000:03: resource 2 [mem 0xc0000000-0xc01fffff 64bit pref] Jan 17 12:18:03.767807 kernel: pci_bus 0000:04: resource 0 [io 0x8000-0x8fff] Jan 17 12:18:03.767851 kernel: pci_bus 0000:04: resource 1 [mem 0xfd100000-0xfd1fffff] Jan 17 12:18:03.767894 kernel: pci_bus 0000:04: resource 2 [mem 0xe7800000-0xe78fffff 64bit pref] Jan 17 12:18:03.767941 kernel: pci_bus 0000:05: resource 0 [io 0xc000-0xcfff] Jan 17 12:18:03.767985 kernel: pci_bus 0000:05: resource 1 [mem 0xfcd00000-0xfcdfffff] Jan 17 12:18:03.768031 kernel: pci_bus 0000:05: resource 2 [mem 0xe7400000-0xe74fffff 64bit pref] Jan 17 12:18:03.768077 kernel: pci_bus 0000:06: resource 1 [mem 0xfc900000-0xfc9fffff] Jan 17 12:18:03.768134 kernel: pci_bus 0000:06: resource 2 [mem 0xe7000000-0xe70fffff 64bit pref] Jan 17 12:18:03.768197 kernel: pci_bus 0000:07: resource 1 [mem 0xfc500000-0xfc5fffff] Jan 17 12:18:03.768240 kernel: pci_bus 0000:07: resource 2 [mem 0xe6c00000-0xe6cfffff 64bit pref] Jan 17 12:18:03.768290 kernel: pci_bus 0000:08: resource 1 [mem 0xfc100000-0xfc1fffff] Jan 17 12:18:03.768336 kernel: pci_bus 0000:08: resource 2 [mem 0xe6800000-0xe68fffff 64bit pref] Jan 17 12:18:03.768384 kernel: pci_bus 0000:09: resource 1 [mem 0xfbd00000-0xfbdfffff] Jan 17 12:18:03.768457 kernel: pci_bus 0000:09: resource 2 [mem 0xe6400000-0xe64fffff 64bit pref] Jan 17 12:18:03.768507 kernel: pci_bus 0000:0a: resource 1 [mem 0xfb900000-0xfb9fffff] Jan 17 12:18:03.768560 kernel: pci_bus 0000:0a: resource 2 [mem 0xe6000000-0xe60fffff 64bit pref] Jan 17 12:18:03.768610 kernel: pci_bus 0000:0b: resource 0 [io 0x5000-0x5fff] Jan 17 12:18:03.768657 kernel: pci_bus 0000:0b: resource 1 [mem 0xfd400000-0xfd4fffff] Jan 17 12:18:03.768702 kernel: pci_bus 0000:0b: resource 2 [mem 0xc0200000-0xc03fffff 64bit pref] Jan 17 12:18:03.768749 kernel: pci_bus 0000:0c: resource 0 [io 0x9000-0x9fff] Jan 17 12:18:03.768794 kernel: pci_bus 0000:0c: resource 1 [mem 0xfd000000-0xfd0fffff] Jan 17 12:18:03.768839 kernel: pci_bus 0000:0c: resource 2 [mem 0xe7700000-0xe77fffff 64bit pref] Jan 17 12:18:03.768887 kernel: pci_bus 0000:0d: resource 0 [io 0xd000-0xdfff] Jan 17 12:18:03.768935 kernel: pci_bus 0000:0d: resource 1 [mem 0xfcc00000-0xfccfffff] Jan 17 12:18:03.768982 kernel: pci_bus 0000:0d: resource 2 [mem 0xe7300000-0xe73fffff 64bit pref] Jan 17 12:18:03.769030 kernel: pci_bus 0000:0e: resource 1 [mem 0xfc800000-0xfc8fffff] Jan 17 12:18:03.769076 kernel: pci_bus 0000:0e: resource 2 [mem 0xe6f00000-0xe6ffffff 64bit pref] Jan 17 12:18:03.770118 kernel: pci_bus 0000:0f: resource 1 [mem 0xfc400000-0xfc4fffff] Jan 17 12:18:03.770172 kernel: pci_bus 0000:0f: resource 2 [mem 0xe6b00000-0xe6bfffff 64bit pref] Jan 17 12:18:03.770223 kernel: pci_bus 0000:10: resource 1 [mem 0xfc000000-0xfc0fffff] Jan 17 12:18:03.770272 kernel: pci_bus 0000:10: resource 2 [mem 0xe6700000-0xe67fffff 64bit pref] Jan 17 12:18:03.770323 kernel: pci_bus 0000:11: resource 1 [mem 0xfbc00000-0xfbcfffff] Jan 17 12:18:03.770368 kernel: pci_bus 0000:11: resource 2 [mem 0xe6300000-0xe63fffff 64bit pref] Jan 17 12:18:03.770454 kernel: pci_bus 0000:12: resource 1 [mem 0xfb800000-0xfb8fffff] Jan 17 12:18:03.770498 kernel: pci_bus 0000:12: resource 2 [mem 0xe5f00000-0xe5ffffff 64bit pref] Jan 17 12:18:03.770546 kernel: pci_bus 0000:13: resource 0 [io 0x6000-0x6fff] Jan 17 12:18:03.770593 kernel: pci_bus 0000:13: resource 1 [mem 0xfd300000-0xfd3fffff] Jan 17 12:18:03.770637 kernel: pci_bus 0000:13: resource 2 [mem 0xe7a00000-0xe7afffff 64bit pref] Jan 17 12:18:03.770685 kernel: pci_bus 0000:14: resource 0 [io 0xa000-0xafff] Jan 17 12:18:03.770730 kernel: pci_bus 0000:14: resource 1 [mem 0xfcf00000-0xfcffffff] Jan 17 12:18:03.770774 kernel: pci_bus 0000:14: resource 2 [mem 0xe7600000-0xe76fffff 64bit pref] Jan 17 12:18:03.770821 kernel: pci_bus 0000:15: resource 0 [io 0xe000-0xefff] Jan 17 12:18:03.770865 kernel: pci_bus 0000:15: resource 1 [mem 0xfcb00000-0xfcbfffff] Jan 17 12:18:03.770911 kernel: pci_bus 0000:15: resource 2 [mem 0xe7200000-0xe72fffff 64bit pref] Jan 17 12:18:03.770958 kernel: pci_bus 0000:16: resource 1 [mem 0xfc700000-0xfc7fffff] Jan 17 12:18:03.771003 kernel: pci_bus 0000:16: resource 2 [mem 0xe6e00000-0xe6efffff 64bit pref] Jan 17 12:18:03.771051 kernel: pci_bus 0000:17: resource 1 [mem 0xfc300000-0xfc3fffff] Jan 17 12:18:03.771096 kernel: pci_bus 0000:17: resource 2 [mem 0xe6a00000-0xe6afffff 64bit pref] Jan 17 12:18:03.771454 kernel: pci_bus 0000:18: resource 1 [mem 0xfbf00000-0xfbffffff] Jan 17 12:18:03.771506 kernel: pci_bus 0000:18: resource 2 [mem 0xe6600000-0xe66fffff 64bit pref] Jan 17 12:18:03.771560 kernel: pci_bus 0000:19: resource 1 [mem 0xfbb00000-0xfbbfffff] Jan 17 12:18:03.771605 kernel: pci_bus 0000:19: resource 2 [mem 0xe6200000-0xe62fffff 64bit pref] Jan 17 12:18:03.771653 kernel: pci_bus 0000:1a: resource 1 [mem 0xfb700000-0xfb7fffff] Jan 17 12:18:03.771698 kernel: pci_bus 0000:1a: resource 2 [mem 0xe5e00000-0xe5efffff 64bit pref] Jan 17 12:18:03.771750 kernel: pci_bus 0000:1b: resource 0 [io 0x7000-0x7fff] Jan 17 12:18:03.771798 kernel: pci_bus 0000:1b: resource 1 [mem 0xfd200000-0xfd2fffff] Jan 17 12:18:03.771842 kernel: pci_bus 0000:1b: resource 2 [mem 0xe7900000-0xe79fffff 64bit pref] Jan 17 12:18:03.771891 kernel: pci_bus 0000:1c: resource 0 [io 0xb000-0xbfff] Jan 17 12:18:03.771937 kernel: pci_bus 0000:1c: resource 1 [mem 0xfce00000-0xfcefffff] Jan 17 12:18:03.771982 kernel: pci_bus 0000:1c: resource 2 [mem 0xe7500000-0xe75fffff 64bit pref] Jan 17 12:18:03.772030 kernel: pci_bus 0000:1d: resource 1 [mem 0xfca00000-0xfcafffff] Jan 17 12:18:03.772078 kernel: pci_bus 0000:1d: resource 2 [mem 0xe7100000-0xe71fffff 64bit pref] Jan 17 12:18:03.772146 kernel: pci_bus 0000:1e: resource 1 [mem 0xfc600000-0xfc6fffff] Jan 17 12:18:03.772194 kernel: pci_bus 0000:1e: resource 2 [mem 0xe6d00000-0xe6dfffff 64bit pref] Jan 17 12:18:03.772242 kernel: pci_bus 0000:1f: resource 1 [mem 0xfc200000-0xfc2fffff] Jan 17 12:18:03.772287 kernel: pci_bus 0000:1f: resource 2 [mem 0xe6900000-0xe69fffff 64bit pref] Jan 17 12:18:03.772476 kernel: pci_bus 0000:20: resource 1 [mem 0xfbe00000-0xfbefffff] Jan 17 12:18:03.772526 kernel: pci_bus 0000:20: resource 2 [mem 0xe6500000-0xe65fffff 64bit pref] Jan 17 12:18:03.772582 kernel: pci_bus 0000:21: resource 1 [mem 0xfba00000-0xfbafffff] Jan 17 12:18:03.772627 kernel: pci_bus 0000:21: resource 2 [mem 0xe6100000-0xe61fffff 64bit pref] Jan 17 12:18:03.772676 kernel: pci_bus 0000:22: resource 1 [mem 0xfb600000-0xfb6fffff] Jan 17 12:18:03.772721 kernel: pci_bus 0000:22: resource 2 [mem 0xe5d00000-0xe5dfffff 64bit pref] Jan 17 12:18:03.772774 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers Jan 17 12:18:03.772784 kernel: PCI: CLS 32 bytes, default 64 Jan 17 12:18:03.772793 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Jan 17 12:18:03.772800 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x311fd3cd494, max_idle_ns: 440795223879 ns Jan 17 12:18:03.772806 kernel: clocksource: Switched to clocksource tsc Jan 17 12:18:03.772812 kernel: Initialise system trusted keyrings Jan 17 12:18:03.772818 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 Jan 17 12:18:03.772824 kernel: Key type asymmetric registered Jan 17 12:18:03.772830 kernel: Asymmetric key parser 'x509' registered Jan 17 12:18:03.772836 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Jan 17 12:18:03.772842 kernel: io scheduler mq-deadline registered Jan 17 12:18:03.772850 kernel: io scheduler kyber registered Jan 17 12:18:03.772856 kernel: io scheduler bfq registered Jan 17 12:18:03.772907 kernel: pcieport 0000:00:15.0: PME: Signaling with IRQ 24 Jan 17 12:18:03.772958 kernel: pcieport 0000:00:15.0: pciehp: Slot #160 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.773008 kernel: pcieport 0000:00:15.1: PME: Signaling with IRQ 25 Jan 17 12:18:03.773058 kernel: pcieport 0000:00:15.1: pciehp: Slot #161 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.773118 kernel: pcieport 0000:00:15.2: PME: Signaling with IRQ 26 Jan 17 12:18:03.773172 kernel: pcieport 0000:00:15.2: pciehp: Slot #162 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.773225 kernel: pcieport 0000:00:15.3: PME: Signaling with IRQ 27 Jan 17 12:18:03.773275 kernel: pcieport 0000:00:15.3: pciehp: Slot #163 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.773324 kernel: pcieport 0000:00:15.4: PME: Signaling with IRQ 28 Jan 17 12:18:03.773374 kernel: pcieport 0000:00:15.4: pciehp: Slot #164 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.773423 kernel: pcieport 0000:00:15.5: PME: Signaling with IRQ 29 Jan 17 12:18:03.773472 kernel: pcieport 0000:00:15.5: pciehp: Slot #165 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.773523 kernel: pcieport 0000:00:15.6: PME: Signaling with IRQ 30 Jan 17 12:18:03.773573 kernel: pcieport 0000:00:15.6: pciehp: Slot #166 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.773622 kernel: pcieport 0000:00:15.7: PME: Signaling with IRQ 31 Jan 17 12:18:03.773671 kernel: pcieport 0000:00:15.7: pciehp: Slot #167 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.773720 kernel: pcieport 0000:00:16.0: PME: Signaling with IRQ 32 Jan 17 12:18:03.773772 kernel: pcieport 0000:00:16.0: pciehp: Slot #192 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.773822 kernel: pcieport 0000:00:16.1: PME: Signaling with IRQ 33 Jan 17 12:18:03.773880 kernel: pcieport 0000:00:16.1: pciehp: Slot #193 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.773939 kernel: pcieport 0000:00:16.2: PME: Signaling with IRQ 34 Jan 17 12:18:03.773988 kernel: pcieport 0000:00:16.2: pciehp: Slot #194 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.774037 kernel: pcieport 0000:00:16.3: PME: Signaling with IRQ 35 Jan 17 12:18:03.774087 kernel: pcieport 0000:00:16.3: pciehp: Slot #195 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.774215 kernel: pcieport 0000:00:16.4: PME: Signaling with IRQ 36 Jan 17 12:18:03.774266 kernel: pcieport 0000:00:16.4: pciehp: Slot #196 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.774318 kernel: pcieport 0000:00:16.5: PME: Signaling with IRQ 37 Jan 17 12:18:03.774369 kernel: pcieport 0000:00:16.5: pciehp: Slot #197 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.774422 kernel: pcieport 0000:00:16.6: PME: Signaling with IRQ 38 Jan 17 12:18:03.774474 kernel: pcieport 0000:00:16.6: pciehp: Slot #198 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.774522 kernel: pcieport 0000:00:16.7: PME: Signaling with IRQ 39 Jan 17 12:18:03.774571 kernel: pcieport 0000:00:16.7: pciehp: Slot #199 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.774620 kernel: pcieport 0000:00:17.0: PME: Signaling with IRQ 40 Jan 17 12:18:03.774670 kernel: pcieport 0000:00:17.0: pciehp: Slot #224 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.774718 kernel: pcieport 0000:00:17.1: PME: Signaling with IRQ 41 Jan 17 12:18:03.774770 kernel: pcieport 0000:00:17.1: pciehp: Slot #225 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.774819 kernel: pcieport 0000:00:17.2: PME: Signaling with IRQ 42 Jan 17 12:18:03.774868 kernel: pcieport 0000:00:17.2: pciehp: Slot #226 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.774928 kernel: pcieport 0000:00:17.3: PME: Signaling with IRQ 43 Jan 17 12:18:03.774984 kernel: pcieport 0000:00:17.3: pciehp: Slot #227 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.776155 kernel: pcieport 0000:00:17.4: PME: Signaling with IRQ 44 Jan 17 12:18:03.776217 kernel: pcieport 0000:00:17.4: pciehp: Slot #228 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.776269 kernel: pcieport 0000:00:17.5: PME: Signaling with IRQ 45 Jan 17 12:18:03.776320 kernel: pcieport 0000:00:17.5: pciehp: Slot #229 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.776370 kernel: pcieport 0000:00:17.6: PME: Signaling with IRQ 46 Jan 17 12:18:03.776419 kernel: pcieport 0000:00:17.6: pciehp: Slot #230 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.776469 kernel: pcieport 0000:00:17.7: PME: Signaling with IRQ 47 Jan 17 12:18:03.776556 kernel: pcieport 0000:00:17.7: pciehp: Slot #231 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.776604 kernel: pcieport 0000:00:18.0: PME: Signaling with IRQ 48 Jan 17 12:18:03.776652 kernel: pcieport 0000:00:18.0: pciehp: Slot #256 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.776701 kernel: pcieport 0000:00:18.1: PME: Signaling with IRQ 49 Jan 17 12:18:03.776749 kernel: pcieport 0000:00:18.1: pciehp: Slot #257 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.776800 kernel: pcieport 0000:00:18.2: PME: Signaling with IRQ 50 Jan 17 12:18:03.776851 kernel: pcieport 0000:00:18.2: pciehp: Slot #258 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.776900 kernel: pcieport 0000:00:18.3: PME: Signaling with IRQ 51 Jan 17 12:18:03.776950 kernel: pcieport 0000:00:18.3: pciehp: Slot #259 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.777000 kernel: pcieport 0000:00:18.4: PME: Signaling with IRQ 52 Jan 17 12:18:03.777049 kernel: pcieport 0000:00:18.4: pciehp: Slot #260 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.777232 kernel: pcieport 0000:00:18.5: PME: Signaling with IRQ 53 Jan 17 12:18:03.777292 kernel: pcieport 0000:00:18.5: pciehp: Slot #261 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.777342 kernel: pcieport 0000:00:18.6: PME: Signaling with IRQ 54 Jan 17 12:18:03.777397 kernel: pcieport 0000:00:18.6: pciehp: Slot #262 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.777448 kernel: pcieport 0000:00:18.7: PME: Signaling with IRQ 55 Jan 17 12:18:03.777498 kernel: pcieport 0000:00:18.7: pciehp: Slot #263 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.777509 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Jan 17 12:18:03.777516 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Jan 17 12:18:03.777522 kernel: 00:05: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Jan 17 12:18:03.777530 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBC,PNP0f13:MOUS] at 0x60,0x64 irq 1,12 Jan 17 12:18:03.777536 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Jan 17 12:18:03.777542 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Jan 17 12:18:03.777593 kernel: rtc_cmos 00:01: registered as rtc0 Jan 17 12:18:03.777641 kernel: rtc_cmos 00:01: setting system clock to 2025-01-17T12:18:03 UTC (1737116283) Jan 17 12:18:03.777686 kernel: rtc_cmos 00:01: alarms up to one month, y3k, 114 bytes nvram Jan 17 12:18:03.777695 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input1 Jan 17 12:18:03.777701 kernel: intel_pstate: CPU model not supported Jan 17 12:18:03.777708 kernel: NET: Registered PF_INET6 protocol family Jan 17 12:18:03.777714 kernel: Segment Routing with IPv6 Jan 17 12:18:03.777720 kernel: In-situ OAM (IOAM) with IPv6 Jan 17 12:18:03.777726 kernel: NET: Registered PF_PACKET protocol family Jan 17 12:18:03.777732 kernel: Key type dns_resolver registered Jan 17 12:18:03.777740 kernel: IPI shorthand broadcast: enabled Jan 17 12:18:03.777746 kernel: sched_clock: Marking stable (882003491, 224063293)->(1160789631, -54722847) Jan 17 12:18:03.777752 kernel: registered taskstats version 1 Jan 17 12:18:03.777759 kernel: Loading compiled-in X.509 certificates Jan 17 12:18:03.777765 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.71-flatcar: 6baa290b0089ed5c4c5f7248306af816ac8c7f80' Jan 17 12:18:03.777771 kernel: Key type .fscrypt registered Jan 17 12:18:03.777777 kernel: Key type fscrypt-provisioning registered Jan 17 12:18:03.777783 kernel: ima: No TPM chip found, activating TPM-bypass! Jan 17 12:18:03.777790 kernel: ima: Allocated hash algorithm: sha1 Jan 17 12:18:03.777796 kernel: ima: No architecture policies found Jan 17 12:18:03.777803 kernel: clk: Disabling unused clocks Jan 17 12:18:03.777809 kernel: Freeing unused kernel image (initmem) memory: 42848K Jan 17 12:18:03.777815 kernel: Write protecting the kernel read-only data: 36864k Jan 17 12:18:03.777821 kernel: Freeing unused kernel image (rodata/data gap) memory: 1848K Jan 17 12:18:03.777827 kernel: Run /init as init process Jan 17 12:18:03.777833 kernel: with arguments: Jan 17 12:18:03.777839 kernel: /init Jan 17 12:18:03.777845 kernel: with environment: Jan 17 12:18:03.777852 kernel: HOME=/ Jan 17 12:18:03.777858 kernel: TERM=linux Jan 17 12:18:03.777864 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Jan 17 12:18:03.777872 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Jan 17 12:18:03.777879 systemd[1]: Detected virtualization vmware. Jan 17 12:18:03.777886 systemd[1]: Detected architecture x86-64. Jan 17 12:18:03.777892 systemd[1]: Running in initrd. Jan 17 12:18:03.777898 systemd[1]: No hostname configured, using default hostname. Jan 17 12:18:03.777906 systemd[1]: Hostname set to <localhost>. Jan 17 12:18:03.777912 systemd[1]: Initializing machine ID from random generator. Jan 17 12:18:03.777919 systemd[1]: Queued start job for default target initrd.target. Jan 17 12:18:03.777925 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 17 12:18:03.777931 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 17 12:18:03.777938 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Jan 17 12:18:03.777945 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jan 17 12:18:03.777951 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Jan 17 12:18:03.777959 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Jan 17 12:18:03.777966 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Jan 17 12:18:03.777973 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Jan 17 12:18:03.777980 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 17 12:18:03.777986 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jan 17 12:18:03.777992 systemd[1]: Reached target paths.target - Path Units. Jan 17 12:18:03.778000 systemd[1]: Reached target slices.target - Slice Units. Jan 17 12:18:03.778006 systemd[1]: Reached target swap.target - Swaps. Jan 17 12:18:03.778013 systemd[1]: Reached target timers.target - Timer Units. Jan 17 12:18:03.778019 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Jan 17 12:18:03.778025 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jan 17 12:18:03.778032 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Jan 17 12:18:03.778038 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Jan 17 12:18:03.778045 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jan 17 12:18:03.778051 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jan 17 12:18:03.778059 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jan 17 12:18:03.778065 systemd[1]: Reached target sockets.target - Socket Units. Jan 17 12:18:03.778071 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Jan 17 12:18:03.778078 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jan 17 12:18:03.778084 systemd[1]: Finished network-cleanup.service - Network Cleanup. Jan 17 12:18:03.778090 systemd[1]: Starting systemd-fsck-usr.service... Jan 17 12:18:03.778097 systemd[1]: Starting systemd-journald.service - Journal Service... Jan 17 12:18:03.779126 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jan 17 12:18:03.779134 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 17 12:18:03.779158 systemd-journald[216]: Collecting audit messages is disabled. Jan 17 12:18:03.779174 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Jan 17 12:18:03.779181 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jan 17 12:18:03.779187 systemd[1]: Finished systemd-fsck-usr.service. Jan 17 12:18:03.779196 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Jan 17 12:18:03.779203 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jan 17 12:18:03.779210 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Jan 17 12:18:03.779217 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 17 12:18:03.779224 kernel: Bridge firewalling registered Jan 17 12:18:03.779231 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jan 17 12:18:03.779237 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jan 17 12:18:03.779244 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jan 17 12:18:03.779250 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jan 17 12:18:03.779258 systemd-journald[216]: Journal started Jan 17 12:18:03.779273 systemd-journald[216]: Runtime Journal (/run/log/journal/4812979c72774891b284461cf9c463ae) is 4.8M, max 38.6M, 33.8M free. Jan 17 12:18:03.741714 systemd-modules-load[217]: Inserted module 'overlay' Jan 17 12:18:03.782211 systemd[1]: Started systemd-journald.service - Journal Service. Jan 17 12:18:03.767233 systemd-modules-load[217]: Inserted module 'br_netfilter' Jan 17 12:18:03.782030 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 17 12:18:03.784092 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jan 17 12:18:03.784515 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jan 17 12:18:03.788802 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 17 12:18:03.789640 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Jan 17 12:18:03.789918 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 17 12:18:03.791168 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jan 17 12:18:03.797566 dracut-cmdline[248]: dracut-dracut-053 Jan 17 12:18:03.800548 dracut-cmdline[248]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=vmware flatcar.autologin verity.usrhash=bf1e0d81a0170850ab02d370c1a7c7a3f5983c980b3730f748240a3bda2dbb2e Jan 17 12:18:03.815513 systemd-resolved[250]: Positive Trust Anchors: Jan 17 12:18:03.815524 systemd-resolved[250]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jan 17 12:18:03.815545 systemd-resolved[250]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jan 17 12:18:03.818019 systemd-resolved[250]: Defaulting to hostname 'linux'. Jan 17 12:18:03.818582 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jan 17 12:18:03.818713 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jan 17 12:18:03.845112 kernel: SCSI subsystem initialized Jan 17 12:18:03.850109 kernel: Loading iSCSI transport class v2.0-870. Jan 17 12:18:03.857111 kernel: iscsi: registered transport (tcp) Jan 17 12:18:03.869421 kernel: iscsi: registered transport (qla4xxx) Jan 17 12:18:03.869441 kernel: QLogic iSCSI HBA Driver Jan 17 12:18:03.888756 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Jan 17 12:18:03.892181 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Jan 17 12:18:03.906889 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Jan 17 12:18:03.906913 kernel: device-mapper: uevent: version 1.0.3 Jan 17 12:18:03.906922 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Jan 17 12:18:03.937140 kernel: raid6: avx2x4 gen() 52549 MB/s Jan 17 12:18:03.954142 kernel: raid6: avx2x2 gen() 53087 MB/s Jan 17 12:18:03.971415 kernel: raid6: avx2x1 gen() 44554 MB/s Jan 17 12:18:03.971442 kernel: raid6: using algorithm avx2x2 gen() 53087 MB/s Jan 17 12:18:03.989366 kernel: raid6: .... xor() 31005 MB/s, rmw enabled Jan 17 12:18:03.989394 kernel: raid6: using avx2x2 recovery algorithm Jan 17 12:18:04.003117 kernel: xor: automatically using best checksumming function avx Jan 17 12:18:04.099119 kernel: Btrfs loaded, zoned=no, fsverity=no Jan 17 12:18:04.104515 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Jan 17 12:18:04.109204 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 17 12:18:04.116393 systemd-udevd[433]: Using default interface naming scheme 'v255'. Jan 17 12:18:04.118837 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 17 12:18:04.124189 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Jan 17 12:18:04.130752 dracut-pre-trigger[436]: rd.md=0: removing MD RAID activation Jan 17 12:18:04.145339 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Jan 17 12:18:04.150214 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jan 17 12:18:04.220477 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jan 17 12:18:04.226201 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Jan 17 12:18:04.234791 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Jan 17 12:18:04.235748 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Jan 17 12:18:04.236047 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 17 12:18:04.236301 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jan 17 12:18:04.240192 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Jan 17 12:18:04.247523 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Jan 17 12:18:04.291158 kernel: VMware PVSCSI driver - version 1.0.7.0-k Jan 17 12:18:04.292127 kernel: vmw_pvscsi: using 64bit dma Jan 17 12:18:04.296169 kernel: vmw_pvscsi: max_id: 16 Jan 17 12:18:04.296189 kernel: vmw_pvscsi: setting ring_pages to 8 Jan 17 12:18:04.303253 kernel: VMware vmxnet3 virtual NIC driver - version 1.7.0.0-k-NAPI Jan 17 12:18:04.303271 kernel: vmw_pvscsi: enabling reqCallThreshold Jan 17 12:18:04.303279 kernel: vmw_pvscsi: driver-based request coalescing enabled Jan 17 12:18:04.303287 kernel: vmw_pvscsi: using MSI-X Jan 17 12:18:04.304610 kernel: vmxnet3 0000:0b:00.0: # of Tx queues : 2, # of Rx queues : 2 Jan 17 12:18:04.316042 kernel: scsi host0: VMware PVSCSI storage adapter rev 2, req/cmp/msg rings: 8/8/1 pages, cmd_per_lun=254 Jan 17 12:18:04.316068 kernel: cryptd: max_cpu_qlen set to 1000 Jan 17 12:18:04.316077 kernel: libata version 3.00 loaded. Jan 17 12:18:04.316087 kernel: vmxnet3 0000:0b:00.0 eth0: NIC Link is Up 10000 Mbps Jan 17 12:18:04.316192 kernel: ata_piix 0000:00:07.1: version 2.13 Jan 17 12:18:04.327980 kernel: vmw_pvscsi 0000:03:00.0: VMware PVSCSI rev 2 host #0 Jan 17 12:18:04.328065 kernel: scsi 0:0:0:0: Direct-Access VMware Virtual disk 2.0 PQ: 0 ANSI: 6 Jan 17 12:18:04.328160 kernel: scsi host1: ata_piix Jan 17 12:18:04.328221 kernel: scsi host2: ata_piix Jan 17 12:18:04.328281 kernel: ata1: PATA max UDMA/33 cmd 0x1f0 ctl 0x3f6 bmdma 0x1060 irq 14 Jan 17 12:18:04.328289 kernel: vmxnet3 0000:0b:00.0 ens192: renamed from eth0 Jan 17 12:18:04.328355 kernel: ata2: PATA max UDMA/33 cmd 0x170 ctl 0x376 bmdma 0x1068 irq 15 Jan 17 12:18:04.328364 kernel: AVX2 version of gcm_enc/dec engaged. Jan 17 12:18:04.328371 kernel: AES CTR mode by8 optimization enabled Jan 17 12:18:04.330557 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jan 17 12:18:04.330781 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 17 12:18:04.330940 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jan 17 12:18:04.331029 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 17 12:18:04.331093 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 17 12:18:04.331200 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Jan 17 12:18:04.337433 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 17 12:18:04.348443 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 17 12:18:04.349271 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jan 17 12:18:04.361492 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 17 12:18:04.497120 kernel: ata2.00: ATAPI: VMware Virtual IDE CDROM Drive, 00000001, max UDMA/33 Jan 17 12:18:04.504236 kernel: scsi 2:0:0:0: CD-ROM NECVMWar VMware IDE CDR10 1.00 PQ: 0 ANSI: 5 Jan 17 12:18:04.516412 kernel: sd 0:0:0:0: [sda] 17805312 512-byte logical blocks: (9.12 GB/8.49 GiB) Jan 17 12:18:04.522323 kernel: sd 0:0:0:0: [sda] Write Protect is off Jan 17 12:18:04.522391 kernel: sd 0:0:0:0: [sda] Mode Sense: 31 00 00 00 Jan 17 12:18:04.522452 kernel: sd 0:0:0:0: [sda] Cache data unavailable Jan 17 12:18:04.522510 kernel: sd 0:0:0:0: [sda] Assuming drive cache: write through Jan 17 12:18:04.522567 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 1x/1x writer dvd-ram cd/rw xa/form2 cdda tray Jan 17 12:18:04.532814 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Jan 17 12:18:04.532824 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jan 17 12:18:04.532837 kernel: sd 0:0:0:0: [sda] Attached SCSI disk Jan 17 12:18:04.532902 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Jan 17 12:18:04.692609 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/sda6 scanned by (udev-worker) (483) Jan 17 12:18:04.700799 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Virtual_disk OEM. Jan 17 12:18:04.704027 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Virtual_disk EFI-SYSTEM. Jan 17 12:18:04.707186 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Virtual_disk ROOT. Jan 17 12:18:04.713114 kernel: BTRFS: device fsid e459b8ee-f1f7-4c3d-a087-3f1955f52c85 devid 1 transid 36 /dev/sda3 scanned by (udev-worker) (481) Jan 17 12:18:04.718819 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Virtual_disk USR-A. Jan 17 12:18:04.719118 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Virtual_disk USR-A. Jan 17 12:18:04.727281 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Jan 17 12:18:04.757122 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jan 17 12:18:04.762113 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jan 17 12:18:05.770155 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jan 17 12:18:05.771074 disk-uuid[588]: The operation has completed successfully. Jan 17 12:18:05.806248 systemd[1]: disk-uuid.service: Deactivated successfully. Jan 17 12:18:05.806307 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Jan 17 12:18:05.811367 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Jan 17 12:18:05.812997 sh[605]: Success Jan 17 12:18:05.821127 kernel: device-mapper: verity: sha256 using implementation "sha256-avx2" Jan 17 12:18:05.866185 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Jan 17 12:18:05.873530 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Jan 17 12:18:05.873995 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Jan 17 12:18:05.891974 kernel: BTRFS info (device dm-0): first mount of filesystem e459b8ee-f1f7-4c3d-a087-3f1955f52c85 Jan 17 12:18:05.891995 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Jan 17 12:18:05.892004 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Jan 17 12:18:05.892011 kernel: BTRFS info (device dm-0): disabling log replay at mount time Jan 17 12:18:05.892018 kernel: BTRFS info (device dm-0): using free space tree Jan 17 12:18:05.898113 kernel: BTRFS info (device dm-0): enabling ssd optimizations Jan 17 12:18:05.898888 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Jan 17 12:18:05.906187 systemd[1]: Starting afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments... Jan 17 12:18:05.907174 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Jan 17 12:18:05.923506 kernel: BTRFS info (device sda6): first mount of filesystem a70a40d6-5ab2-4665-81b1-b8e9f58c5ff8 Jan 17 12:18:05.923534 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Jan 17 12:18:05.923542 kernel: BTRFS info (device sda6): using free space tree Jan 17 12:18:05.928181 kernel: BTRFS info (device sda6): enabling ssd optimizations Jan 17 12:18:05.934701 systemd[1]: mnt-oem.mount: Deactivated successfully. Jan 17 12:18:05.936426 kernel: BTRFS info (device sda6): last unmount of filesystem a70a40d6-5ab2-4665-81b1-b8e9f58c5ff8 Jan 17 12:18:05.942847 systemd[1]: Finished ignition-setup.service - Ignition (setup). Jan 17 12:18:05.948175 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Jan 17 12:18:05.970317 systemd[1]: Finished afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments. Jan 17 12:18:05.979320 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Jan 17 12:18:06.017350 ignition[664]: Ignition 2.19.0 Jan 17 12:18:06.017358 ignition[664]: Stage: fetch-offline Jan 17 12:18:06.017383 ignition[664]: no configs at "/usr/lib/ignition/base.d" Jan 17 12:18:06.017390 ignition[664]: no config dir at "/usr/lib/ignition/base.platform.d/vmware" Jan 17 12:18:06.017443 ignition[664]: parsed url from cmdline: "" Jan 17 12:18:06.017445 ignition[664]: no config URL provided Jan 17 12:18:06.017448 ignition[664]: reading system config file "/usr/lib/ignition/user.ign" Jan 17 12:18:06.017453 ignition[664]: no config at "/usr/lib/ignition/user.ign" Jan 17 12:18:06.017833 ignition[664]: config successfully fetched Jan 17 12:18:06.017850 ignition[664]: parsing config with SHA512: 2b2b8494d6bfd05329bb7d196c73328bab11e189c274c6ed5a8b4488d854439289a668dac01d59a5ee639bf6f8486a3245506eb1d49630a99d3fdd97bea97388 Jan 17 12:18:06.020578 unknown[664]: fetched base config from "system" Jan 17 12:18:06.020584 unknown[664]: fetched user config from "vmware" Jan 17 12:18:06.022241 ignition[664]: fetch-offline: fetch-offline passed Jan 17 12:18:06.022296 ignition[664]: Ignition finished successfully Jan 17 12:18:06.023023 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Jan 17 12:18:06.045452 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jan 17 12:18:06.050332 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jan 17 12:18:06.061287 systemd-networkd[800]: lo: Link UP Jan 17 12:18:06.061292 systemd-networkd[800]: lo: Gained carrier Jan 17 12:18:06.062133 systemd-networkd[800]: Enumeration completed Jan 17 12:18:06.062262 systemd[1]: Started systemd-networkd.service - Network Configuration. Jan 17 12:18:06.062383 systemd[1]: Reached target network.target - Network. Jan 17 12:18:06.062406 systemd-networkd[800]: ens192: Configuring with /etc/systemd/network/10-dracut-cmdline-99.network. Jan 17 12:18:06.062461 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Jan 17 12:18:06.065947 kernel: vmxnet3 0000:0b:00.0 ens192: intr type 3, mode 0, 3 vectors allocated Jan 17 12:18:06.066059 kernel: vmxnet3 0000:0b:00.0 ens192: NIC Link is Up 10000 Mbps Jan 17 12:18:06.065819 systemd-networkd[800]: ens192: Link UP Jan 17 12:18:06.065821 systemd-networkd[800]: ens192: Gained carrier Jan 17 12:18:06.072147 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Jan 17 12:18:06.078749 ignition[802]: Ignition 2.19.0 Jan 17 12:18:06.078754 ignition[802]: Stage: kargs Jan 17 12:18:06.078859 ignition[802]: no configs at "/usr/lib/ignition/base.d" Jan 17 12:18:06.078865 ignition[802]: no config dir at "/usr/lib/ignition/base.platform.d/vmware" Jan 17 12:18:06.079373 ignition[802]: kargs: kargs passed Jan 17 12:18:06.079400 ignition[802]: Ignition finished successfully Jan 17 12:18:06.080612 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Jan 17 12:18:06.086313 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Jan 17 12:18:06.092991 ignition[809]: Ignition 2.19.0 Jan 17 12:18:06.092999 ignition[809]: Stage: disks Jan 17 12:18:06.093093 ignition[809]: no configs at "/usr/lib/ignition/base.d" Jan 17 12:18:06.093107 ignition[809]: no config dir at "/usr/lib/ignition/base.platform.d/vmware" Jan 17 12:18:06.093590 ignition[809]: disks: disks passed Jan 17 12:18:06.093616 ignition[809]: Ignition finished successfully Jan 17 12:18:06.094207 systemd[1]: Finished ignition-disks.service - Ignition (disks). Jan 17 12:18:06.094583 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Jan 17 12:18:06.094704 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Jan 17 12:18:06.094886 systemd[1]: Reached target local-fs.target - Local File Systems. Jan 17 12:18:06.095065 systemd[1]: Reached target sysinit.target - System Initialization. Jan 17 12:18:06.095266 systemd[1]: Reached target basic.target - Basic System. Jan 17 12:18:06.098183 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Jan 17 12:18:06.140989 systemd-fsck[817]: ROOT: clean, 14/1628000 files, 120691/1617920 blocks Jan 17 12:18:06.141978 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Jan 17 12:18:06.148117 systemd[1]: Mounting sysroot.mount - /sysroot... Jan 17 12:18:06.208810 systemd[1]: Mounted sysroot.mount - /sysroot. Jan 17 12:18:06.209111 kernel: EXT4-fs (sda9): mounted filesystem 0ba4fe0e-76d7-406f-b570-4642d86198f6 r/w with ordered data mode. Quota mode: none. Jan 17 12:18:06.209324 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Jan 17 12:18:06.214278 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jan 17 12:18:06.215147 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Jan 17 12:18:06.215936 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Jan 17 12:18:06.215961 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Jan 17 12:18:06.215974 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Jan 17 12:18:06.219157 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Jan 17 12:18:06.219671 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Jan 17 12:18:06.223108 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/sda6 scanned by mount (825) Jan 17 12:18:06.225280 kernel: BTRFS info (device sda6): first mount of filesystem a70a40d6-5ab2-4665-81b1-b8e9f58c5ff8 Jan 17 12:18:06.225298 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Jan 17 12:18:06.225309 kernel: BTRFS info (device sda6): using free space tree Jan 17 12:18:06.230111 kernel: BTRFS info (device sda6): enabling ssd optimizations Jan 17 12:18:06.230335 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jan 17 12:18:06.250387 initrd-setup-root[849]: cut: /sysroot/etc/passwd: No such file or directory Jan 17 12:18:06.252699 initrd-setup-root[856]: cut: /sysroot/etc/group: No such file or directory Jan 17 12:18:06.254802 initrd-setup-root[863]: cut: /sysroot/etc/shadow: No such file or directory Jan 17 12:18:06.257023 initrd-setup-root[870]: cut: /sysroot/etc/gshadow: No such file or directory Jan 17 12:18:06.306431 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Jan 17 12:18:06.311246 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Jan 17 12:18:06.313587 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Jan 17 12:18:06.318178 kernel: BTRFS info (device sda6): last unmount of filesystem a70a40d6-5ab2-4665-81b1-b8e9f58c5ff8 Jan 17 12:18:06.345767 ignition[938]: INFO : Ignition 2.19.0 Jan 17 12:18:06.345767 ignition[938]: INFO : Stage: mount Jan 17 12:18:06.345767 ignition[938]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 17 12:18:06.345767 ignition[938]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/vmware" Jan 17 12:18:06.345767 ignition[938]: INFO : mount: mount passed Jan 17 12:18:06.345767 ignition[938]: INFO : Ignition finished successfully Jan 17 12:18:06.346545 systemd[1]: Finished ignition-mount.service - Ignition (mount). Jan 17 12:18:06.350312 systemd[1]: Starting ignition-files.service - Ignition (files)... Jan 17 12:18:06.350565 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Jan 17 12:18:06.887776 systemd[1]: sysroot-oem.mount: Deactivated successfully. Jan 17 12:18:06.894324 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jan 17 12:18:06.902116 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/sda6 scanned by mount (951) Jan 17 12:18:06.904965 kernel: BTRFS info (device sda6): first mount of filesystem a70a40d6-5ab2-4665-81b1-b8e9f58c5ff8 Jan 17 12:18:06.904985 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Jan 17 12:18:06.904996 kernel: BTRFS info (device sda6): using free space tree Jan 17 12:18:06.909111 kernel: BTRFS info (device sda6): enabling ssd optimizations Jan 17 12:18:06.909321 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jan 17 12:18:06.927986 ignition[968]: INFO : Ignition 2.19.0 Jan 17 12:18:06.927986 ignition[968]: INFO : Stage: files Jan 17 12:18:06.927986 ignition[968]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 17 12:18:06.927986 ignition[968]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/vmware" Jan 17 12:18:06.927986 ignition[968]: DEBUG : files: compiled without relabeling support, skipping Jan 17 12:18:06.928837 ignition[968]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Jan 17 12:18:06.928980 ignition[968]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Jan 17 12:18:06.931094 ignition[968]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Jan 17 12:18:06.931318 ignition[968]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Jan 17 12:18:06.931593 unknown[968]: wrote ssh authorized keys file for user: core Jan 17 12:18:06.931779 ignition[968]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Jan 17 12:18:06.933013 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Jan 17 12:18:06.933355 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-amd64.tar.gz: attempt #1 Jan 17 12:18:06.978452 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Jan 17 12:18:07.082782 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Jan 17 12:18:07.083555 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Jan 17 12:18:07.083555 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Jan 17 12:18:07.083555 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Jan 17 12:18:07.083555 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Jan 17 12:18:07.083555 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Jan 17 12:18:07.083555 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Jan 17 12:18:07.083555 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Jan 17 12:18:07.083555 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Jan 17 12:18:07.083555 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Jan 17 12:18:07.085157 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Jan 17 12:18:07.085157 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.0-x86-64.raw" Jan 17 12:18:07.085157 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.0-x86-64.raw" Jan 17 12:18:07.085157 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.0-x86-64.raw" Jan 17 12:18:07.085157 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.31.0-x86-64.raw: attempt #1 Jan 17 12:18:07.599089 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Jan 17 12:18:07.782636 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.0-x86-64.raw" Jan 17 12:18:07.782636 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/etc/systemd/network/00-vmware.network" Jan 17 12:18:07.783134 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/etc/systemd/network/00-vmware.network" Jan 17 12:18:07.783134 ignition[968]: INFO : files: op(c): [started] processing unit "prepare-helm.service" Jan 17 12:18:07.786931 ignition[968]: INFO : files: op(c): op(d): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jan 17 12:18:07.787119 ignition[968]: INFO : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jan 17 12:18:07.787119 ignition[968]: INFO : files: op(c): [finished] processing unit "prepare-helm.service" Jan 17 12:18:07.787119 ignition[968]: INFO : files: op(e): [started] processing unit "coreos-metadata.service" Jan 17 12:18:07.787119 ignition[968]: INFO : files: op(e): op(f): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Jan 17 12:18:07.787119 ignition[968]: INFO : files: op(e): op(f): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Jan 17 12:18:07.787119 ignition[968]: INFO : files: op(e): [finished] processing unit "coreos-metadata.service" Jan 17 12:18:07.787119 ignition[968]: INFO : files: op(10): [started] setting preset to disabled for "coreos-metadata.service" Jan 17 12:18:08.130490 systemd-networkd[800]: ens192: Gained IPv6LL Jan 17 12:18:08.196996 ignition[968]: INFO : files: op(10): op(11): [started] removing enablement symlink(s) for "coreos-metadata.service" Jan 17 12:18:08.201122 ignition[968]: INFO : files: op(10): op(11): [finished] removing enablement symlink(s) for "coreos-metadata.service" Jan 17 12:18:08.201122 ignition[968]: INFO : files: op(10): [finished] setting preset to disabled for "coreos-metadata.service" Jan 17 12:18:08.201122 ignition[968]: INFO : files: op(12): [started] setting preset to enabled for "prepare-helm.service" Jan 17 12:18:08.201122 ignition[968]: INFO : files: op(12): [finished] setting preset to enabled for "prepare-helm.service" Jan 17 12:18:08.201905 ignition[968]: INFO : files: createResultFile: createFiles: op(13): [started] writing file "/sysroot/etc/.ignition-result.json" Jan 17 12:18:08.201905 ignition[968]: INFO : files: createResultFile: createFiles: op(13): [finished] writing file "/sysroot/etc/.ignition-result.json" Jan 17 12:18:08.201905 ignition[968]: INFO : files: files passed Jan 17 12:18:08.201905 ignition[968]: INFO : Ignition finished successfully Jan 17 12:18:08.202120 systemd[1]: Finished ignition-files.service - Ignition (files). Jan 17 12:18:08.212325 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Jan 17 12:18:08.213585 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Jan 17 12:18:08.215033 systemd[1]: ignition-quench.service: Deactivated successfully. Jan 17 12:18:08.215084 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Jan 17 12:18:08.220512 initrd-setup-root-after-ignition[998]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jan 17 12:18:08.220512 initrd-setup-root-after-ignition[998]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Jan 17 12:18:08.220933 initrd-setup-root-after-ignition[1002]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jan 17 12:18:08.221920 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Jan 17 12:18:08.222399 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Jan 17 12:18:08.226286 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Jan 17 12:18:08.237566 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Jan 17 12:18:08.237615 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Jan 17 12:18:08.237914 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Jan 17 12:18:08.238018 systemd[1]: Reached target initrd.target - Initrd Default Target. Jan 17 12:18:08.238146 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Jan 17 12:18:08.240177 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Jan 17 12:18:08.247182 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jan 17 12:18:08.251205 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Jan 17 12:18:08.256079 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Jan 17 12:18:08.256358 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 17 12:18:08.256517 systemd[1]: Stopped target timers.target - Timer Units. Jan 17 12:18:08.256651 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Jan 17 12:18:08.256717 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jan 17 12:18:08.256931 systemd[1]: Stopped target initrd.target - Initrd Default Target. Jan 17 12:18:08.257146 systemd[1]: Stopped target basic.target - Basic System. Jan 17 12:18:08.257311 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Jan 17 12:18:08.257504 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Jan 17 12:18:08.257701 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Jan 17 12:18:08.257888 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Jan 17 12:18:08.258059 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Jan 17 12:18:08.258463 systemd[1]: Stopped target sysinit.target - System Initialization. Jan 17 12:18:08.258637 systemd[1]: Stopped target local-fs.target - Local File Systems. Jan 17 12:18:08.258804 systemd[1]: Stopped target swap.target - Swaps. Jan 17 12:18:08.258961 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Jan 17 12:18:08.259018 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Jan 17 12:18:08.259306 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Jan 17 12:18:08.259540 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 17 12:18:08.259685 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Jan 17 12:18:08.259726 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 17 12:18:08.259870 systemd[1]: dracut-initqueue.service: Deactivated successfully. Jan 17 12:18:08.259927 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Jan 17 12:18:08.260177 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Jan 17 12:18:08.260237 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Jan 17 12:18:08.260500 systemd[1]: Stopped target paths.target - Path Units. Jan 17 12:18:08.260623 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Jan 17 12:18:08.263120 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 17 12:18:08.263283 systemd[1]: Stopped target slices.target - Slice Units. Jan 17 12:18:08.263497 systemd[1]: Stopped target sockets.target - Socket Units. Jan 17 12:18:08.263669 systemd[1]: iscsid.socket: Deactivated successfully. Jan 17 12:18:08.263731 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Jan 17 12:18:08.263918 systemd[1]: iscsiuio.socket: Deactivated successfully. Jan 17 12:18:08.263961 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jan 17 12:18:08.264206 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Jan 17 12:18:08.264263 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Jan 17 12:18:08.264496 systemd[1]: ignition-files.service: Deactivated successfully. Jan 17 12:18:08.264548 systemd[1]: Stopped ignition-files.service - Ignition (files). Jan 17 12:18:08.269187 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Jan 17 12:18:08.271218 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Jan 17 12:18:08.271328 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Jan 17 12:18:08.271416 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Jan 17 12:18:08.271667 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Jan 17 12:18:08.271742 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Jan 17 12:18:08.274044 systemd[1]: initrd-cleanup.service: Deactivated successfully. Jan 17 12:18:08.274131 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Jan 17 12:18:08.278107 ignition[1022]: INFO : Ignition 2.19.0 Jan 17 12:18:08.278107 ignition[1022]: INFO : Stage: umount Jan 17 12:18:08.278107 ignition[1022]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 17 12:18:08.278107 ignition[1022]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/vmware" Jan 17 12:18:08.279362 ignition[1022]: INFO : umount: umount passed Jan 17 12:18:08.279362 ignition[1022]: INFO : Ignition finished successfully Jan 17 12:18:08.279997 systemd[1]: ignition-mount.service: Deactivated successfully. Jan 17 12:18:08.280198 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Jan 17 12:18:08.280664 systemd[1]: Stopped target network.target - Network. Jan 17 12:18:08.280880 systemd[1]: ignition-disks.service: Deactivated successfully. Jan 17 12:18:08.281029 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Jan 17 12:18:08.281292 systemd[1]: ignition-kargs.service: Deactivated successfully. Jan 17 12:18:08.281403 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Jan 17 12:18:08.281612 systemd[1]: ignition-setup.service: Deactivated successfully. Jan 17 12:18:08.281652 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Jan 17 12:18:08.281864 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Jan 17 12:18:08.281884 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Jan 17 12:18:08.282207 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Jan 17 12:18:08.282592 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Jan 17 12:18:08.283513 systemd[1]: sysroot-boot.mount: Deactivated successfully. Jan 17 12:18:08.288212 systemd[1]: systemd-resolved.service: Deactivated successfully. Jan 17 12:18:08.288272 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Jan 17 12:18:08.288830 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Jan 17 12:18:08.288862 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 17 12:18:08.290052 systemd[1]: systemd-networkd.service: Deactivated successfully. Jan 17 12:18:08.290118 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Jan 17 12:18:08.290419 systemd[1]: systemd-networkd.socket: Deactivated successfully. Jan 17 12:18:08.290456 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Jan 17 12:18:08.293196 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Jan 17 12:18:08.293284 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Jan 17 12:18:08.293309 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jan 17 12:18:08.293422 systemd[1]: afterburn-network-kargs.service: Deactivated successfully. Jan 17 12:18:08.293442 systemd[1]: Stopped afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments. Jan 17 12:18:08.293567 systemd[1]: systemd-sysctl.service: Deactivated successfully. Jan 17 12:18:08.293602 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Jan 17 12:18:08.294531 systemd[1]: systemd-modules-load.service: Deactivated successfully. Jan 17 12:18:08.294554 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Jan 17 12:18:08.295794 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 17 12:18:08.300258 systemd[1]: network-cleanup.service: Deactivated successfully. Jan 17 12:18:08.300321 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Jan 17 12:18:08.300675 systemd[1]: systemd-udevd.service: Deactivated successfully. Jan 17 12:18:08.300741 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 17 12:18:08.301281 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Jan 17 12:18:08.301308 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Jan 17 12:18:08.301511 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Jan 17 12:18:08.301528 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Jan 17 12:18:08.301675 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Jan 17 12:18:08.301696 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Jan 17 12:18:08.301952 systemd[1]: dracut-cmdline.service: Deactivated successfully. Jan 17 12:18:08.301973 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Jan 17 12:18:08.302244 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jan 17 12:18:08.302264 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 17 12:18:08.306390 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Jan 17 12:18:08.306498 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Jan 17 12:18:08.306525 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 17 12:18:08.306648 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Jan 17 12:18:08.306669 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jan 17 12:18:08.306790 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Jan 17 12:18:08.306811 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Jan 17 12:18:08.306930 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 17 12:18:08.306949 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 17 12:18:08.309117 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Jan 17 12:18:08.309184 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Jan 17 12:18:08.386151 systemd[1]: sysroot-boot.service: Deactivated successfully. Jan 17 12:18:08.386232 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Jan 17 12:18:08.386718 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Jan 17 12:18:08.386888 systemd[1]: initrd-setup-root.service: Deactivated successfully. Jan 17 12:18:08.386926 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Jan 17 12:18:08.389204 systemd[1]: Starting initrd-switch-root.service - Switch Root... Jan 17 12:18:08.402132 systemd[1]: Switching root. Jan 17 12:18:08.430620 systemd-journald[216]: Journal stopped Jan 17 12:18:03.728273 kernel: Linux version 6.6.71-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT_DYNAMIC Fri Jan 17 10:39:07 -00 2025 Jan 17 12:18:03.728290 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=vmware flatcar.autologin verity.usrhash=bf1e0d81a0170850ab02d370c1a7c7a3f5983c980b3730f748240a3bda2dbb2e Jan 17 12:18:03.728297 kernel: Disabled fast string operations Jan 17 12:18:03.728301 kernel: BIOS-provided physical RAM map: Jan 17 12:18:03.728305 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ebff] usable Jan 17 12:18:03.728308 kernel: BIOS-e820: [mem 0x000000000009ec00-0x000000000009ffff] reserved Jan 17 12:18:03.728314 kernel: BIOS-e820: [mem 0x00000000000dc000-0x00000000000fffff] reserved Jan 17 12:18:03.728318 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007fedffff] usable Jan 17 12:18:03.728322 kernel: BIOS-e820: [mem 0x000000007fee0000-0x000000007fefefff] ACPI data Jan 17 12:18:03.728326 kernel: BIOS-e820: [mem 0x000000007feff000-0x000000007fefffff] ACPI NVS Jan 17 12:18:03.728330 kernel: BIOS-e820: [mem 0x000000007ff00000-0x000000007fffffff] usable Jan 17 12:18:03.728334 kernel: BIOS-e820: [mem 0x00000000f0000000-0x00000000f7ffffff] reserved Jan 17 12:18:03.728338 kernel: BIOS-e820: [mem 0x00000000fec00000-0x00000000fec0ffff] reserved Jan 17 12:18:03.728342 kernel: BIOS-e820: [mem 0x00000000fee00000-0x00000000fee00fff] reserved Jan 17 12:18:03.728348 kernel: BIOS-e820: [mem 0x00000000fffe0000-0x00000000ffffffff] reserved Jan 17 12:18:03.728353 kernel: NX (Execute Disable) protection: active Jan 17 12:18:03.728357 kernel: APIC: Static calls initialized Jan 17 12:18:03.728362 kernel: SMBIOS 2.7 present. Jan 17 12:18:03.728367 kernel: DMI: VMware, Inc. VMware Virtual Platform/440BX Desktop Reference Platform, BIOS 6.00 05/28/2020 Jan 17 12:18:03.728371 kernel: vmware: hypercall mode: 0x00 Jan 17 12:18:03.728376 kernel: Hypervisor detected: VMware Jan 17 12:18:03.728380 kernel: vmware: TSC freq read from hypervisor : 3408.000 MHz Jan 17 12:18:03.728409 kernel: vmware: Host bus clock speed read from hypervisor : 66000000 Hz Jan 17 12:18:03.728416 kernel: vmware: using clock offset of 2537798063 ns Jan 17 12:18:03.728421 kernel: tsc: Detected 3408.000 MHz processor Jan 17 12:18:03.728426 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Jan 17 12:18:03.728448 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Jan 17 12:18:03.728452 kernel: last_pfn = 0x80000 max_arch_pfn = 0x400000000 Jan 17 12:18:03.728457 kernel: total RAM covered: 3072M Jan 17 12:18:03.728463 kernel: Found optimal setting for mtrr clean up Jan 17 12:18:03.728471 kernel: gran_size: 64K chunk_size: 64K num_reg: 2 lose cover RAM: 0G Jan 17 12:18:03.728481 kernel: MTRR map: 6 entries (5 fixed + 1 variable; max 21), built from 8 variable MTRRs Jan 17 12:18:03.728489 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Jan 17 12:18:03.728494 kernel: Using GB pages for direct mapping Jan 17 12:18:03.728499 kernel: ACPI: Early table checksum verification disabled Jan 17 12:18:03.728504 kernel: ACPI: RSDP 0x00000000000F6A00 000024 (v02 PTLTD ) Jan 17 12:18:03.728508 kernel: ACPI: XSDT 0x000000007FEE965B 00005C (v01 INTEL 440BX 06040000 VMW 01324272) Jan 17 12:18:03.728513 kernel: ACPI: FACP 0x000000007FEFEE73 0000F4 (v04 INTEL 440BX 06040000 PTL 000F4240) Jan 17 12:18:03.728518 kernel: ACPI: DSDT 0x000000007FEEAD55 01411E (v01 PTLTD Custom 06040000 MSFT 03000001) Jan 17 12:18:03.728523 kernel: ACPI: FACS 0x000000007FEFFFC0 000040 Jan 17 12:18:03.728530 kernel: ACPI: FACS 0x000000007FEFFFC0 000040 Jan 17 12:18:03.728535 kernel: ACPI: BOOT 0x000000007FEEAD2D 000028 (v01 PTLTD $SBFTBL$ 06040000 LTP 00000001) Jan 17 12:18:03.728540 kernel: ACPI: APIC 0x000000007FEEA5EB 000742 (v01 PTLTD ? APIC 06040000 LTP 00000000) Jan 17 12:18:03.728545 kernel: ACPI: MCFG 0x000000007FEEA5AF 00003C (v01 PTLTD $PCITBL$ 06040000 LTP 00000001) Jan 17 12:18:03.728550 kernel: ACPI: SRAT 0x000000007FEE9757 0008A8 (v02 VMWARE MEMPLUG 06040000 VMW 00000001) Jan 17 12:18:03.728556 kernel: ACPI: HPET 0x000000007FEE971F 000038 (v01 VMWARE VMW HPET 06040000 VMW 00000001) Jan 17 12:18:03.728561 kernel: ACPI: WAET 0x000000007FEE96F7 000028 (v01 VMWARE VMW WAET 06040000 VMW 00000001) Jan 17 12:18:03.728566 kernel: ACPI: Reserving FACP table memory at [mem 0x7fefee73-0x7fefef66] Jan 17 12:18:03.728571 kernel: ACPI: Reserving DSDT table memory at [mem 0x7feead55-0x7fefee72] Jan 17 12:18:03.728576 kernel: ACPI: Reserving FACS table memory at [mem 0x7fefffc0-0x7fefffff] Jan 17 12:18:03.728581 kernel: ACPI: Reserving FACS table memory at [mem 0x7fefffc0-0x7fefffff] Jan 17 12:18:03.728586 kernel: ACPI: Reserving BOOT table memory at [mem 0x7feead2d-0x7feead54] Jan 17 12:18:03.728591 kernel: ACPI: Reserving APIC table memory at [mem 0x7feea5eb-0x7feead2c] Jan 17 12:18:03.728596 kernel: ACPI: Reserving MCFG table memory at [mem 0x7feea5af-0x7feea5ea] Jan 17 12:18:03.728600 kernel: ACPI: Reserving SRAT table memory at [mem 0x7fee9757-0x7fee9ffe] Jan 17 12:18:03.728606 kernel: ACPI: Reserving HPET table memory at [mem 0x7fee971f-0x7fee9756] Jan 17 12:18:03.728611 kernel: ACPI: Reserving WAET table memory at [mem 0x7fee96f7-0x7fee971e] Jan 17 12:18:03.728616 kernel: system APIC only can use physical flat Jan 17 12:18:03.728621 kernel: APIC: Switched APIC routing to: physical flat Jan 17 12:18:03.728626 kernel: SRAT: PXM 0 -> APIC 0x00 -> Node 0 Jan 17 12:18:03.728631 kernel: SRAT: PXM 0 -> APIC 0x02 -> Node 0 Jan 17 12:18:03.728635 kernel: SRAT: PXM 0 -> APIC 0x04 -> Node 0 Jan 17 12:18:03.728640 kernel: SRAT: PXM 0 -> APIC 0x06 -> Node 0 Jan 17 12:18:03.728645 kernel: SRAT: PXM 0 -> APIC 0x08 -> Node 0 Jan 17 12:18:03.728651 kernel: SRAT: PXM 0 -> APIC 0x0a -> Node 0 Jan 17 12:18:03.728656 kernel: SRAT: PXM 0 -> APIC 0x0c -> Node 0 Jan 17 12:18:03.728660 kernel: SRAT: PXM 0 -> APIC 0x0e -> Node 0 Jan 17 12:18:03.728665 kernel: SRAT: PXM 0 -> APIC 0x10 -> Node 0 Jan 17 12:18:03.728670 kernel: SRAT: PXM 0 -> APIC 0x12 -> Node 0 Jan 17 12:18:03.728674 kernel: SRAT: PXM 0 -> APIC 0x14 -> Node 0 Jan 17 12:18:03.728679 kernel: SRAT: PXM 0 -> APIC 0x16 -> Node 0 Jan 17 12:18:03.728684 kernel: SRAT: PXM 0 -> APIC 0x18 -> Node 0 Jan 17 12:18:03.728689 kernel: SRAT: PXM 0 -> APIC 0x1a -> Node 0 Jan 17 12:18:03.728694 kernel: SRAT: PXM 0 -> APIC 0x1c -> Node 0 Jan 17 12:18:03.728698 kernel: SRAT: PXM 0 -> APIC 0x1e -> Node 0 Jan 17 12:18:03.728704 kernel: SRAT: PXM 0 -> APIC 0x20 -> Node 0 Jan 17 12:18:03.728709 kernel: SRAT: PXM 0 -> APIC 0x22 -> Node 0 Jan 17 12:18:03.728714 kernel: SRAT: PXM 0 -> APIC 0x24 -> Node 0 Jan 17 12:18:03.728718 kernel: SRAT: PXM 0 -> APIC 0x26 -> Node 0 Jan 17 12:18:03.728723 kernel: SRAT: PXM 0 -> APIC 0x28 -> Node 0 Jan 17 12:18:03.728728 kernel: SRAT: PXM 0 -> APIC 0x2a -> Node 0 Jan 17 12:18:03.728733 kernel: SRAT: PXM 0 -> APIC 0x2c -> Node 0 Jan 17 12:18:03.728737 kernel: SRAT: PXM 0 -> APIC 0x2e -> Node 0 Jan 17 12:18:03.728742 kernel: SRAT: PXM 0 -> APIC 0x30 -> Node 0 Jan 17 12:18:03.728747 kernel: SRAT: PXM 0 -> APIC 0x32 -> Node 0 Jan 17 12:18:03.728753 kernel: SRAT: PXM 0 -> APIC 0x34 -> Node 0 Jan 17 12:18:03.728758 kernel: SRAT: PXM 0 -> APIC 0x36 -> Node 0 Jan 17 12:18:03.728762 kernel: SRAT: PXM 0 -> APIC 0x38 -> Node 0 Jan 17 12:18:03.728767 kernel: SRAT: PXM 0 -> APIC 0x3a -> Node 0 Jan 17 12:18:03.728772 kernel: SRAT: PXM 0 -> APIC 0x3c -> Node 0 Jan 17 12:18:03.728777 kernel: SRAT: PXM 0 -> APIC 0x3e -> Node 0 Jan 17 12:18:03.728782 kernel: SRAT: PXM 0 -> APIC 0x40 -> Node 0 Jan 17 12:18:03.728787 kernel: SRAT: PXM 0 -> APIC 0x42 -> Node 0 Jan 17 12:18:03.728792 kernel: SRAT: PXM 0 -> APIC 0x44 -> Node 0 Jan 17 12:18:03.728796 kernel: SRAT: PXM 0 -> APIC 0x46 -> Node 0 Jan 17 12:18:03.728802 kernel: SRAT: PXM 0 -> APIC 0x48 -> Node 0 Jan 17 12:18:03.728807 kernel: SRAT: PXM 0 -> APIC 0x4a -> Node 0 Jan 17 12:18:03.728812 kernel: SRAT: PXM 0 -> APIC 0x4c -> Node 0 Jan 17 12:18:03.728816 kernel: SRAT: PXM 0 -> APIC 0x4e -> Node 0 Jan 17 12:18:03.728821 kernel: SRAT: PXM 0 -> APIC 0x50 -> Node 0 Jan 17 12:18:03.728826 kernel: SRAT: PXM 0 -> APIC 0x52 -> Node 0 Jan 17 12:18:03.728831 kernel: SRAT: PXM 0 -> APIC 0x54 -> Node 0 Jan 17 12:18:03.728836 kernel: SRAT: PXM 0 -> APIC 0x56 -> Node 0 Jan 17 12:18:03.728840 kernel: SRAT: PXM 0 -> APIC 0x58 -> Node 0 Jan 17 12:18:03.728845 kernel: SRAT: PXM 0 -> APIC 0x5a -> Node 0 Jan 17 12:18:03.728851 kernel: SRAT: PXM 0 -> APIC 0x5c -> Node 0 Jan 17 12:18:03.728856 kernel: SRAT: PXM 0 -> APIC 0x5e -> Node 0 Jan 17 12:18:03.728861 kernel: SRAT: PXM 0 -> APIC 0x60 -> Node 0 Jan 17 12:18:03.728865 kernel: SRAT: PXM 0 -> APIC 0x62 -> Node 0 Jan 17 12:18:03.728870 kernel: SRAT: PXM 0 -> APIC 0x64 -> Node 0 Jan 17 12:18:03.728875 kernel: SRAT: PXM 0 -> APIC 0x66 -> Node 0 Jan 17 12:18:03.728879 kernel: SRAT: PXM 0 -> APIC 0x68 -> Node 0 Jan 17 12:18:03.728884 kernel: SRAT: PXM 0 -> APIC 0x6a -> Node 0 Jan 17 12:18:03.728889 kernel: SRAT: PXM 0 -> APIC 0x6c -> Node 0 Jan 17 12:18:03.728894 kernel: SRAT: PXM 0 -> APIC 0x6e -> Node 0 Jan 17 12:18:03.728900 kernel: SRAT: PXM 0 -> APIC 0x70 -> Node 0 Jan 17 12:18:03.728905 kernel: SRAT: PXM 0 -> APIC 0x72 -> Node 0 Jan 17 12:18:03.728910 kernel: SRAT: PXM 0 -> APIC 0x74 -> Node 0 Jan 17 12:18:03.728918 kernel: SRAT: PXM 0 -> APIC 0x76 -> Node 0 Jan 17 12:18:03.728924 kernel: SRAT: PXM 0 -> APIC 0x78 -> Node 0 Jan 17 12:18:03.728929 kernel: SRAT: PXM 0 -> APIC 0x7a -> Node 0 Jan 17 12:18:03.728934 kernel: SRAT: PXM 0 -> APIC 0x7c -> Node 0 Jan 17 12:18:03.728939 kernel: SRAT: PXM 0 -> APIC 0x7e -> Node 0 Jan 17 12:18:03.728944 kernel: SRAT: PXM 0 -> APIC 0x80 -> Node 0 Jan 17 12:18:03.728951 kernel: SRAT: PXM 0 -> APIC 0x82 -> Node 0 Jan 17 12:18:03.728955 kernel: SRAT: PXM 0 -> APIC 0x84 -> Node 0 Jan 17 12:18:03.728961 kernel: SRAT: PXM 0 -> APIC 0x86 -> Node 0 Jan 17 12:18:03.728966 kernel: SRAT: PXM 0 -> APIC 0x88 -> Node 0 Jan 17 12:18:03.728971 kernel: SRAT: PXM 0 -> APIC 0x8a -> Node 0 Jan 17 12:18:03.728976 kernel: SRAT: PXM 0 -> APIC 0x8c -> Node 0 Jan 17 12:18:03.728981 kernel: SRAT: PXM 0 -> APIC 0x8e -> Node 0 Jan 17 12:18:03.728986 kernel: SRAT: PXM 0 -> APIC 0x90 -> Node 0 Jan 17 12:18:03.728991 kernel: SRAT: PXM 0 -> APIC 0x92 -> Node 0 Jan 17 12:18:03.728996 kernel: SRAT: PXM 0 -> APIC 0x94 -> Node 0 Jan 17 12:18:03.729003 kernel: SRAT: PXM 0 -> APIC 0x96 -> Node 0 Jan 17 12:18:03.729008 kernel: SRAT: PXM 0 -> APIC 0x98 -> Node 0 Jan 17 12:18:03.729013 kernel: SRAT: PXM 0 -> APIC 0x9a -> Node 0 Jan 17 12:18:03.729018 kernel: SRAT: PXM 0 -> APIC 0x9c -> Node 0 Jan 17 12:18:03.729023 kernel: SRAT: PXM 0 -> APIC 0x9e -> Node 0 Jan 17 12:18:03.729029 kernel: SRAT: PXM 0 -> APIC 0xa0 -> Node 0 Jan 17 12:18:03.729034 kernel: SRAT: PXM 0 -> APIC 0xa2 -> Node 0 Jan 17 12:18:03.729039 kernel: SRAT: PXM 0 -> APIC 0xa4 -> Node 0 Jan 17 12:18:03.729044 kernel: SRAT: PXM 0 -> APIC 0xa6 -> Node 0 Jan 17 12:18:03.729049 kernel: SRAT: PXM 0 -> APIC 0xa8 -> Node 0 Jan 17 12:18:03.729055 kernel: SRAT: PXM 0 -> APIC 0xaa -> Node 0 Jan 17 12:18:03.729060 kernel: SRAT: PXM 0 -> APIC 0xac -> Node 0 Jan 17 12:18:03.729065 kernel: SRAT: PXM 0 -> APIC 0xae -> Node 0 Jan 17 12:18:03.729070 kernel: SRAT: PXM 0 -> APIC 0xb0 -> Node 0 Jan 17 12:18:03.729075 kernel: SRAT: PXM 0 -> APIC 0xb2 -> Node 0 Jan 17 12:18:03.729080 kernel: SRAT: PXM 0 -> APIC 0xb4 -> Node 0 Jan 17 12:18:03.729086 kernel: SRAT: PXM 0 -> APIC 0xb6 -> Node 0 Jan 17 12:18:03.729091 kernel: SRAT: PXM 0 -> APIC 0xb8 -> Node 0 Jan 17 12:18:03.729096 kernel: SRAT: PXM 0 -> APIC 0xba -> Node 0 Jan 17 12:18:03.729110 kernel: SRAT: PXM 0 -> APIC 0xbc -> Node 0 Jan 17 12:18:03.729119 kernel: SRAT: PXM 0 -> APIC 0xbe -> Node 0 Jan 17 12:18:03.729124 kernel: SRAT: PXM 0 -> APIC 0xc0 -> Node 0 Jan 17 12:18:03.729129 kernel: SRAT: PXM 0 -> APIC 0xc2 -> Node 0 Jan 17 12:18:03.729134 kernel: SRAT: PXM 0 -> APIC 0xc4 -> Node 0 Jan 17 12:18:03.729139 kernel: SRAT: PXM 0 -> APIC 0xc6 -> Node 0 Jan 17 12:18:03.729145 kernel: SRAT: PXM 0 -> APIC 0xc8 -> Node 0 Jan 17 12:18:03.729150 kernel: SRAT: PXM 0 -> APIC 0xca -> Node 0 Jan 17 12:18:03.729155 kernel: SRAT: PXM 0 -> APIC 0xcc -> Node 0 Jan 17 12:18:03.729160 kernel: SRAT: PXM 0 -> APIC 0xce -> Node 0 Jan 17 12:18:03.729165 kernel: SRAT: PXM 0 -> APIC 0xd0 -> Node 0 Jan 17 12:18:03.729170 kernel: SRAT: PXM 0 -> APIC 0xd2 -> Node 0 Jan 17 12:18:03.729176 kernel: SRAT: PXM 0 -> APIC 0xd4 -> Node 0 Jan 17 12:18:03.729181 kernel: SRAT: PXM 0 -> APIC 0xd6 -> Node 0 Jan 17 12:18:03.729187 kernel: SRAT: PXM 0 -> APIC 0xd8 -> Node 0 Jan 17 12:18:03.729192 kernel: SRAT: PXM 0 -> APIC 0xda -> Node 0 Jan 17 12:18:03.729197 kernel: SRAT: PXM 0 -> APIC 0xdc -> Node 0 Jan 17 12:18:03.729202 kernel: SRAT: PXM 0 -> APIC 0xde -> Node 0 Jan 17 12:18:03.729207 kernel: SRAT: PXM 0 -> APIC 0xe0 -> Node 0 Jan 17 12:18:03.729212 kernel: SRAT: PXM 0 -> APIC 0xe2 -> Node 0 Jan 17 12:18:03.729217 kernel: SRAT: PXM 0 -> APIC 0xe4 -> Node 0 Jan 17 12:18:03.729223 kernel: SRAT: PXM 0 -> APIC 0xe6 -> Node 0 Jan 17 12:18:03.729229 kernel: SRAT: PXM 0 -> APIC 0xe8 -> Node 0 Jan 17 12:18:03.729234 kernel: SRAT: PXM 0 -> APIC 0xea -> Node 0 Jan 17 12:18:03.729239 kernel: SRAT: PXM 0 -> APIC 0xec -> Node 0 Jan 17 12:18:03.729244 kernel: SRAT: PXM 0 -> APIC 0xee -> Node 0 Jan 17 12:18:03.729249 kernel: SRAT: PXM 0 -> APIC 0xf0 -> Node 0 Jan 17 12:18:03.729254 kernel: SRAT: PXM 0 -> APIC 0xf2 -> Node 0 Jan 17 12:18:03.729259 kernel: SRAT: PXM 0 -> APIC 0xf4 -> Node 0 Jan 17 12:18:03.729264 kernel: SRAT: PXM 0 -> APIC 0xf6 -> Node 0 Jan 17 12:18:03.729269 kernel: SRAT: PXM 0 -> APIC 0xf8 -> Node 0 Jan 17 12:18:03.729275 kernel: SRAT: PXM 0 -> APIC 0xfa -> Node 0 Jan 17 12:18:03.729281 kernel: SRAT: PXM 0 -> APIC 0xfc -> Node 0 Jan 17 12:18:03.729286 kernel: SRAT: PXM 0 -> APIC 0xfe -> Node 0 Jan 17 12:18:03.729291 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x0009ffff] Jan 17 12:18:03.729297 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00100000-0x7fffffff] Jan 17 12:18:03.729302 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x80000000-0xbfffffff] hotplug Jan 17 12:18:03.729307 kernel: NUMA: Node 0 [mem 0x00000000-0x0009ffff] + [mem 0x00100000-0x7fffffff] -> [mem 0x00000000-0x7fffffff] Jan 17 12:18:03.729313 kernel: NODE_DATA(0) allocated [mem 0x7fffa000-0x7fffffff] Jan 17 12:18:03.729318 kernel: Zone ranges: Jan 17 12:18:03.729323 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Jan 17 12:18:03.729330 kernel: DMA32 [mem 0x0000000001000000-0x000000007fffffff] Jan 17 12:18:03.729335 kernel: Normal empty Jan 17 12:18:03.729341 kernel: Movable zone start for each node Jan 17 12:18:03.729346 kernel: Early memory node ranges Jan 17 12:18:03.729351 kernel: node 0: [mem 0x0000000000001000-0x000000000009dfff] Jan 17 12:18:03.729356 kernel: node 0: [mem 0x0000000000100000-0x000000007fedffff] Jan 17 12:18:03.729361 kernel: node 0: [mem 0x000000007ff00000-0x000000007fffffff] Jan 17 12:18:03.729367 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007fffffff] Jan 17 12:18:03.729372 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Jan 17 12:18:03.729377 kernel: On node 0, zone DMA: 98 pages in unavailable ranges Jan 17 12:18:03.729383 kernel: On node 0, zone DMA32: 32 pages in unavailable ranges Jan 17 12:18:03.729389 kernel: ACPI: PM-Timer IO Port: 0x1008 Jan 17 12:18:03.729412 kernel: system APIC only can use physical flat Jan 17 12:18:03.729417 kernel: ACPI: LAPIC_NMI (acpi_id[0x00] high edge lint[0x1]) Jan 17 12:18:03.729423 kernel: ACPI: LAPIC_NMI (acpi_id[0x01] high edge lint[0x1]) Jan 17 12:18:03.729428 kernel: ACPI: LAPIC_NMI (acpi_id[0x02] high edge lint[0x1]) Jan 17 12:18:03.729433 kernel: ACPI: LAPIC_NMI (acpi_id[0x03] high edge lint[0x1]) Jan 17 12:18:03.729455 kernel: ACPI: LAPIC_NMI (acpi_id[0x04] high edge lint[0x1]) Jan 17 12:18:03.729460 kernel: ACPI: LAPIC_NMI (acpi_id[0x05] high edge lint[0x1]) Jan 17 12:18:03.729465 kernel: ACPI: LAPIC_NMI (acpi_id[0x06] high edge lint[0x1]) Jan 17 12:18:03.729471 kernel: ACPI: LAPIC_NMI (acpi_id[0x07] high edge lint[0x1]) Jan 17 12:18:03.729476 kernel: ACPI: LAPIC_NMI (acpi_id[0x08] high edge lint[0x1]) Jan 17 12:18:03.729481 kernel: ACPI: LAPIC_NMI (acpi_id[0x09] high edge lint[0x1]) Jan 17 12:18:03.729487 kernel: ACPI: LAPIC_NMI (acpi_id[0x0a] high edge lint[0x1]) Jan 17 12:18:03.729492 kernel: ACPI: LAPIC_NMI (acpi_id[0x0b] high edge lint[0x1]) Jan 17 12:18:03.729497 kernel: ACPI: LAPIC_NMI (acpi_id[0x0c] high edge lint[0x1]) Jan 17 12:18:03.729502 kernel: ACPI: LAPIC_NMI (acpi_id[0x0d] high edge lint[0x1]) Jan 17 12:18:03.729507 kernel: ACPI: LAPIC_NMI (acpi_id[0x0e] high edge lint[0x1]) Jan 17 12:18:03.729512 kernel: ACPI: LAPIC_NMI (acpi_id[0x0f] high edge lint[0x1]) Jan 17 12:18:03.729519 kernel: ACPI: LAPIC_NMI (acpi_id[0x10] high edge lint[0x1]) Jan 17 12:18:03.729524 kernel: ACPI: LAPIC_NMI (acpi_id[0x11] high edge lint[0x1]) Jan 17 12:18:03.729529 kernel: ACPI: LAPIC_NMI (acpi_id[0x12] high edge lint[0x1]) Jan 17 12:18:03.729534 kernel: ACPI: LAPIC_NMI (acpi_id[0x13] high edge lint[0x1]) Jan 17 12:18:03.729540 kernel: ACPI: LAPIC_NMI (acpi_id[0x14] high edge lint[0x1]) Jan 17 12:18:03.729545 kernel: ACPI: LAPIC_NMI (acpi_id[0x15] high edge lint[0x1]) Jan 17 12:18:03.729550 kernel: ACPI: LAPIC_NMI (acpi_id[0x16] high edge lint[0x1]) Jan 17 12:18:03.729555 kernel: ACPI: LAPIC_NMI (acpi_id[0x17] high edge lint[0x1]) Jan 17 12:18:03.729560 kernel: ACPI: LAPIC_NMI (acpi_id[0x18] high edge lint[0x1]) Jan 17 12:18:03.729566 kernel: ACPI: LAPIC_NMI (acpi_id[0x19] high edge lint[0x1]) Jan 17 12:18:03.729572 kernel: ACPI: LAPIC_NMI (acpi_id[0x1a] high edge lint[0x1]) Jan 17 12:18:03.729577 kernel: ACPI: LAPIC_NMI (acpi_id[0x1b] high edge lint[0x1]) Jan 17 12:18:03.729582 kernel: ACPI: LAPIC_NMI (acpi_id[0x1c] high edge lint[0x1]) Jan 17 12:18:03.729587 kernel: ACPI: LAPIC_NMI (acpi_id[0x1d] high edge lint[0x1]) Jan 17 12:18:03.729593 kernel: ACPI: LAPIC_NMI (acpi_id[0x1e] high edge lint[0x1]) Jan 17 12:18:03.729598 kernel: ACPI: LAPIC_NMI (acpi_id[0x1f] high edge lint[0x1]) Jan 17 12:18:03.729603 kernel: ACPI: LAPIC_NMI (acpi_id[0x20] high edge lint[0x1]) Jan 17 12:18:03.729608 kernel: ACPI: LAPIC_NMI (acpi_id[0x21] high edge lint[0x1]) Jan 17 12:18:03.729613 kernel: ACPI: LAPIC_NMI (acpi_id[0x22] high edge lint[0x1]) Jan 17 12:18:03.729618 kernel: ACPI: LAPIC_NMI (acpi_id[0x23] high edge lint[0x1]) Jan 17 12:18:03.729624 kernel: ACPI: LAPIC_NMI (acpi_id[0x24] high edge lint[0x1]) Jan 17 12:18:03.729630 kernel: ACPI: LAPIC_NMI (acpi_id[0x25] high edge lint[0x1]) Jan 17 12:18:03.729635 kernel: ACPI: LAPIC_NMI (acpi_id[0x26] high edge lint[0x1]) Jan 17 12:18:03.729640 kernel: ACPI: LAPIC_NMI (acpi_id[0x27] high edge lint[0x1]) Jan 17 12:18:03.729645 kernel: ACPI: LAPIC_NMI (acpi_id[0x28] high edge lint[0x1]) Jan 17 12:18:03.729651 kernel: ACPI: LAPIC_NMI (acpi_id[0x29] high edge lint[0x1]) Jan 17 12:18:03.729656 kernel: ACPI: LAPIC_NMI (acpi_id[0x2a] high edge lint[0x1]) Jan 17 12:18:03.729661 kernel: ACPI: LAPIC_NMI (acpi_id[0x2b] high edge lint[0x1]) Jan 17 12:18:03.729666 kernel: ACPI: LAPIC_NMI (acpi_id[0x2c] high edge lint[0x1]) Jan 17 12:18:03.729672 kernel: ACPI: LAPIC_NMI (acpi_id[0x2d] high edge lint[0x1]) Jan 17 12:18:03.729678 kernel: ACPI: LAPIC_NMI (acpi_id[0x2e] high edge lint[0x1]) Jan 17 12:18:03.729683 kernel: ACPI: LAPIC_NMI (acpi_id[0x2f] high edge lint[0x1]) Jan 17 12:18:03.729688 kernel: ACPI: LAPIC_NMI (acpi_id[0x30] high edge lint[0x1]) Jan 17 12:18:03.729693 kernel: ACPI: LAPIC_NMI (acpi_id[0x31] high edge lint[0x1]) Jan 17 12:18:03.729698 kernel: ACPI: LAPIC_NMI (acpi_id[0x32] high edge lint[0x1]) Jan 17 12:18:03.729704 kernel: ACPI: LAPIC_NMI (acpi_id[0x33] high edge lint[0x1]) Jan 17 12:18:03.729709 kernel: ACPI: LAPIC_NMI (acpi_id[0x34] high edge lint[0x1]) Jan 17 12:18:03.729714 kernel: ACPI: LAPIC_NMI (acpi_id[0x35] high edge lint[0x1]) Jan 17 12:18:03.729719 kernel: ACPI: LAPIC_NMI (acpi_id[0x36] high edge lint[0x1]) Jan 17 12:18:03.729725 kernel: ACPI: LAPIC_NMI (acpi_id[0x37] high edge lint[0x1]) Jan 17 12:18:03.729730 kernel: ACPI: LAPIC_NMI (acpi_id[0x38] high edge lint[0x1]) Jan 17 12:18:03.729736 kernel: ACPI: LAPIC_NMI (acpi_id[0x39] high edge lint[0x1]) Jan 17 12:18:03.729741 kernel: ACPI: LAPIC_NMI (acpi_id[0x3a] high edge lint[0x1]) Jan 17 12:18:03.729746 kernel: ACPI: LAPIC_NMI (acpi_id[0x3b] high edge lint[0x1]) Jan 17 12:18:03.729751 kernel: ACPI: LAPIC_NMI (acpi_id[0x3c] high edge lint[0x1]) Jan 17 12:18:03.729757 kernel: ACPI: LAPIC_NMI (acpi_id[0x3d] high edge lint[0x1]) Jan 17 12:18:03.729762 kernel: ACPI: LAPIC_NMI (acpi_id[0x3e] high edge lint[0x1]) Jan 17 12:18:03.729767 kernel: ACPI: LAPIC_NMI (acpi_id[0x3f] high edge lint[0x1]) Jan 17 12:18:03.729772 kernel: ACPI: LAPIC_NMI (acpi_id[0x40] high edge lint[0x1]) Jan 17 12:18:03.729778 kernel: ACPI: LAPIC_NMI (acpi_id[0x41] high edge lint[0x1]) Jan 17 12:18:03.729783 kernel: ACPI: LAPIC_NMI (acpi_id[0x42] high edge lint[0x1]) Jan 17 12:18:03.729789 kernel: ACPI: LAPIC_NMI (acpi_id[0x43] high edge lint[0x1]) Jan 17 12:18:03.729794 kernel: ACPI: LAPIC_NMI (acpi_id[0x44] high edge lint[0x1]) Jan 17 12:18:03.729799 kernel: ACPI: LAPIC_NMI (acpi_id[0x45] high edge lint[0x1]) Jan 17 12:18:03.729804 kernel: ACPI: LAPIC_NMI (acpi_id[0x46] high edge lint[0x1]) Jan 17 12:18:03.729810 kernel: ACPI: LAPIC_NMI (acpi_id[0x47] high edge lint[0x1]) Jan 17 12:18:03.729815 kernel: ACPI: LAPIC_NMI (acpi_id[0x48] high edge lint[0x1]) Jan 17 12:18:03.729820 kernel: ACPI: LAPIC_NMI (acpi_id[0x49] high edge lint[0x1]) Jan 17 12:18:03.729826 kernel: ACPI: LAPIC_NMI (acpi_id[0x4a] high edge lint[0x1]) Jan 17 12:18:03.729831 kernel: ACPI: LAPIC_NMI (acpi_id[0x4b] high edge lint[0x1]) Jan 17 12:18:03.729837 kernel: ACPI: LAPIC_NMI (acpi_id[0x4c] high edge lint[0x1]) Jan 17 12:18:03.729842 kernel: ACPI: LAPIC_NMI (acpi_id[0x4d] high edge lint[0x1]) Jan 17 12:18:03.729847 kernel: ACPI: LAPIC_NMI (acpi_id[0x4e] high edge lint[0x1]) Jan 17 12:18:03.729852 kernel: ACPI: LAPIC_NMI (acpi_id[0x4f] high edge lint[0x1]) Jan 17 12:18:03.729858 kernel: ACPI: LAPIC_NMI (acpi_id[0x50] high edge lint[0x1]) Jan 17 12:18:03.729863 kernel: ACPI: LAPIC_NMI (acpi_id[0x51] high edge lint[0x1]) Jan 17 12:18:03.729868 kernel: ACPI: LAPIC_NMI (acpi_id[0x52] high edge lint[0x1]) Jan 17 12:18:03.729873 kernel: ACPI: LAPIC_NMI (acpi_id[0x53] high edge lint[0x1]) Jan 17 12:18:03.729879 kernel: ACPI: LAPIC_NMI (acpi_id[0x54] high edge lint[0x1]) Jan 17 12:18:03.729885 kernel: ACPI: LAPIC_NMI (acpi_id[0x55] high edge lint[0x1]) Jan 17 12:18:03.729890 kernel: ACPI: LAPIC_NMI (acpi_id[0x56] high edge lint[0x1]) Jan 17 12:18:03.729895 kernel: ACPI: LAPIC_NMI (acpi_id[0x57] high edge lint[0x1]) Jan 17 12:18:03.729900 kernel: ACPI: LAPIC_NMI (acpi_id[0x58] high edge lint[0x1]) Jan 17 12:18:03.729906 kernel: ACPI: LAPIC_NMI (acpi_id[0x59] high edge lint[0x1]) Jan 17 12:18:03.729911 kernel: ACPI: LAPIC_NMI (acpi_id[0x5a] high edge lint[0x1]) Jan 17 12:18:03.729916 kernel: ACPI: LAPIC_NMI (acpi_id[0x5b] high edge lint[0x1]) Jan 17 12:18:03.729921 kernel: ACPI: LAPIC_NMI (acpi_id[0x5c] high edge lint[0x1]) Jan 17 12:18:03.729927 kernel: ACPI: LAPIC_NMI (acpi_id[0x5d] high edge lint[0x1]) Jan 17 12:18:03.729933 kernel: ACPI: LAPIC_NMI (acpi_id[0x5e] high edge lint[0x1]) Jan 17 12:18:03.729938 kernel: ACPI: LAPIC_NMI (acpi_id[0x5f] high edge lint[0x1]) Jan 17 12:18:03.729943 kernel: ACPI: LAPIC_NMI (acpi_id[0x60] high edge lint[0x1]) Jan 17 12:18:03.729948 kernel: ACPI: LAPIC_NMI (acpi_id[0x61] high edge lint[0x1]) Jan 17 12:18:03.729954 kernel: ACPI: LAPIC_NMI (acpi_id[0x62] high edge lint[0x1]) Jan 17 12:18:03.729959 kernel: ACPI: LAPIC_NMI (acpi_id[0x63] high edge lint[0x1]) Jan 17 12:18:03.729964 kernel: ACPI: LAPIC_NMI (acpi_id[0x64] high edge lint[0x1]) Jan 17 12:18:03.729969 kernel: ACPI: LAPIC_NMI (acpi_id[0x65] high edge lint[0x1]) Jan 17 12:18:03.729974 kernel: ACPI: LAPIC_NMI (acpi_id[0x66] high edge lint[0x1]) Jan 17 12:18:03.729980 kernel: ACPI: LAPIC_NMI (acpi_id[0x67] high edge lint[0x1]) Jan 17 12:18:03.729986 kernel: ACPI: LAPIC_NMI (acpi_id[0x68] high edge lint[0x1]) Jan 17 12:18:03.729991 kernel: ACPI: LAPIC_NMI (acpi_id[0x69] high edge lint[0x1]) Jan 17 12:18:03.729996 kernel: ACPI: LAPIC_NMI (acpi_id[0x6a] high edge lint[0x1]) Jan 17 12:18:03.730001 kernel: ACPI: LAPIC_NMI (acpi_id[0x6b] high edge lint[0x1]) Jan 17 12:18:03.730007 kernel: ACPI: LAPIC_NMI (acpi_id[0x6c] high edge lint[0x1]) Jan 17 12:18:03.730012 kernel: ACPI: LAPIC_NMI (acpi_id[0x6d] high edge lint[0x1]) Jan 17 12:18:03.730017 kernel: ACPI: LAPIC_NMI (acpi_id[0x6e] high edge lint[0x1]) Jan 17 12:18:03.730022 kernel: ACPI: LAPIC_NMI (acpi_id[0x6f] high edge lint[0x1]) Jan 17 12:18:03.730027 kernel: ACPI: LAPIC_NMI (acpi_id[0x70] high edge lint[0x1]) Jan 17 12:18:03.730034 kernel: ACPI: LAPIC_NMI (acpi_id[0x71] high edge lint[0x1]) Jan 17 12:18:03.730039 kernel: ACPI: LAPIC_NMI (acpi_id[0x72] high edge lint[0x1]) Jan 17 12:18:03.730044 kernel: ACPI: LAPIC_NMI (acpi_id[0x73] high edge lint[0x1]) Jan 17 12:18:03.730049 kernel: ACPI: LAPIC_NMI (acpi_id[0x74] high edge lint[0x1]) Jan 17 12:18:03.730054 kernel: ACPI: LAPIC_NMI (acpi_id[0x75] high edge lint[0x1]) Jan 17 12:18:03.730059 kernel: ACPI: LAPIC_NMI (acpi_id[0x76] high edge lint[0x1]) Jan 17 12:18:03.730065 kernel: ACPI: LAPIC_NMI (acpi_id[0x77] high edge lint[0x1]) Jan 17 12:18:03.730070 kernel: ACPI: LAPIC_NMI (acpi_id[0x78] high edge lint[0x1]) Jan 17 12:18:03.730075 kernel: ACPI: LAPIC_NMI (acpi_id[0x79] high edge lint[0x1]) Jan 17 12:18:03.730080 kernel: ACPI: LAPIC_NMI (acpi_id[0x7a] high edge lint[0x1]) Jan 17 12:18:03.730086 kernel: ACPI: LAPIC_NMI (acpi_id[0x7b] high edge lint[0x1]) Jan 17 12:18:03.730092 kernel: ACPI: LAPIC_NMI (acpi_id[0x7c] high edge lint[0x1]) Jan 17 12:18:03.730097 kernel: ACPI: LAPIC_NMI (acpi_id[0x7d] high edge lint[0x1]) Jan 17 12:18:03.730112 kernel: ACPI: LAPIC_NMI (acpi_id[0x7e] high edge lint[0x1]) Jan 17 12:18:03.730117 kernel: ACPI: LAPIC_NMI (acpi_id[0x7f] high edge lint[0x1]) Jan 17 12:18:03.730122 kernel: IOAPIC[0]: apic_id 1, version 17, address 0xfec00000, GSI 0-23 Jan 17 12:18:03.730128 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 high edge) Jan 17 12:18:03.730133 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Jan 17 12:18:03.730138 kernel: ACPI: HPET id: 0x8086af01 base: 0xfed00000 Jan 17 12:18:03.730145 kernel: TSC deadline timer available Jan 17 12:18:03.730150 kernel: smpboot: Allowing 128 CPUs, 126 hotplug CPUs Jan 17 12:18:03.730155 kernel: [mem 0x80000000-0xefffffff] available for PCI devices Jan 17 12:18:03.730161 kernel: Booting paravirtualized kernel on VMware hypervisor Jan 17 12:18:03.730166 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Jan 17 12:18:03.730171 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:128 nr_cpu_ids:128 nr_node_ids:1 Jan 17 12:18:03.730177 kernel: percpu: Embedded 58 pages/cpu s197032 r8192 d32344 u262144 Jan 17 12:18:03.730182 kernel: pcpu-alloc: s197032 r8192 d32344 u262144 alloc=1*2097152 Jan 17 12:18:03.730187 kernel: pcpu-alloc: [0] 000 001 002 003 004 005 006 007 Jan 17 12:18:03.730194 kernel: pcpu-alloc: [0] 008 009 010 011 012 013 014 015 Jan 17 12:18:03.730199 kernel: pcpu-alloc: [0] 016 017 018 019 020 021 022 023 Jan 17 12:18:03.730204 kernel: pcpu-alloc: [0] 024 025 026 027 028 029 030 031 Jan 17 12:18:03.730209 kernel: pcpu-alloc: [0] 032 033 034 035 036 037 038 039 Jan 17 12:18:03.730221 kernel: pcpu-alloc: [0] 040 041 042 043 044 045 046 047 Jan 17 12:18:03.730228 kernel: pcpu-alloc: [0] 048 049 050 051 052 053 054 055 Jan 17 12:18:03.730233 kernel: pcpu-alloc: [0] 056 057 058 059 060 061 062 063 Jan 17 12:18:03.730239 kernel: pcpu-alloc: [0] 064 065 066 067 068 069 070 071 Jan 17 12:18:03.730245 kernel: pcpu-alloc: [0] 072 073 074 075 076 077 078 079 Jan 17 12:18:03.730251 kernel: pcpu-alloc: [0] 080 081 082 083 084 085 086 087 Jan 17 12:18:03.730257 kernel: pcpu-alloc: [0] 088 089 090 091 092 093 094 095 Jan 17 12:18:03.730262 kernel: pcpu-alloc: [0] 096 097 098 099 100 101 102 103 Jan 17 12:18:03.730268 kernel: pcpu-alloc: [0] 104 105 106 107 108 109 110 111 Jan 17 12:18:03.730273 kernel: pcpu-alloc: [0] 112 113 114 115 116 117 118 119 Jan 17 12:18:03.730278 kernel: pcpu-alloc: [0] 120 121 122 123 124 125 126 127 Jan 17 12:18:03.730284 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=vmware flatcar.autologin verity.usrhash=bf1e0d81a0170850ab02d370c1a7c7a3f5983c980b3730f748240a3bda2dbb2e Jan 17 12:18:03.730290 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Jan 17 12:18:03.730297 kernel: random: crng init done Jan 17 12:18:03.730302 kernel: printk: log_buf_len individual max cpu contribution: 4096 bytes Jan 17 12:18:03.730308 kernel: printk: log_buf_len total cpu_extra contributions: 520192 bytes Jan 17 12:18:03.730314 kernel: printk: log_buf_len min size: 262144 bytes Jan 17 12:18:03.730319 kernel: printk: log_buf_len: 1048576 bytes Jan 17 12:18:03.730325 kernel: printk: early log buf free: 239648(91%) Jan 17 12:18:03.730330 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Jan 17 12:18:03.730336 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) Jan 17 12:18:03.730342 kernel: Fallback order for Node 0: 0 Jan 17 12:18:03.730348 kernel: Built 1 zonelists, mobility grouping on. Total pages: 515808 Jan 17 12:18:03.730354 kernel: Policy zone: DMA32 Jan 17 12:18:03.730359 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Jan 17 12:18:03.730365 kernel: Memory: 1936372K/2096628K available (12288K kernel code, 2299K rwdata, 22728K rodata, 42848K init, 2344K bss, 159996K reserved, 0K cma-reserved) Jan 17 12:18:03.730372 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=128, Nodes=1 Jan 17 12:18:03.730378 kernel: ftrace: allocating 37918 entries in 149 pages Jan 17 12:18:03.730384 kernel: ftrace: allocated 149 pages with 4 groups Jan 17 12:18:03.730411 kernel: Dynamic Preempt: voluntary Jan 17 12:18:03.730417 kernel: rcu: Preemptible hierarchical RCU implementation. Jan 17 12:18:03.730423 kernel: rcu: RCU event tracing is enabled. Jan 17 12:18:03.730429 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=128. Jan 17 12:18:03.730435 kernel: Trampoline variant of Tasks RCU enabled. Jan 17 12:18:03.730457 kernel: Rude variant of Tasks RCU enabled. Jan 17 12:18:03.730463 kernel: Tracing variant of Tasks RCU enabled. Jan 17 12:18:03.730483 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Jan 17 12:18:03.730489 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=128 Jan 17 12:18:03.730495 kernel: NR_IRQS: 33024, nr_irqs: 1448, preallocated irqs: 16 Jan 17 12:18:03.730500 kernel: rcu: srcu_init: Setting srcu_struct sizes to big. Jan 17 12:18:03.730506 kernel: Console: colour VGA+ 80x25 Jan 17 12:18:03.730512 kernel: printk: console [tty0] enabled Jan 17 12:18:03.730517 kernel: printk: console [ttyS0] enabled Jan 17 12:18:03.730523 kernel: ACPI: Core revision 20230628 Jan 17 12:18:03.730529 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 133484882848 ns Jan 17 12:18:03.730534 kernel: APIC: Switch to symmetric I/O mode setup Jan 17 12:18:03.730541 kernel: x2apic enabled Jan 17 12:18:03.730547 kernel: APIC: Switched APIC routing to: physical x2apic Jan 17 12:18:03.730552 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Jan 17 12:18:03.730558 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x311fd3cd494, max_idle_ns: 440795223879 ns Jan 17 12:18:03.730563 kernel: Calibrating delay loop (skipped) preset value.. 6816.00 BogoMIPS (lpj=3408000) Jan 17 12:18:03.730569 kernel: Disabled fast string operations Jan 17 12:18:03.730574 kernel: Last level iTLB entries: 4KB 64, 2MB 8, 4MB 8 Jan 17 12:18:03.730580 kernel: Last level dTLB entries: 4KB 64, 2MB 0, 4MB 0, 1GB 4 Jan 17 12:18:03.730586 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Jan 17 12:18:03.730592 kernel: Spectre V2 : Spectre BHI mitigation: SW BHB clearing on vm exit Jan 17 12:18:03.730598 kernel: Spectre V2 : Spectre BHI mitigation: SW BHB clearing on syscall Jan 17 12:18:03.730603 kernel: Spectre V2 : Mitigation: Enhanced / Automatic IBRS Jan 17 12:18:03.730609 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch Jan 17 12:18:03.730615 kernel: Spectre V2 : Spectre v2 / PBRSB-eIBRS: Retire a single CALL on VMEXIT Jan 17 12:18:03.730620 kernel: RETBleed: Mitigation: Enhanced IBRS Jan 17 12:18:03.730627 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Jan 17 12:18:03.730633 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Jan 17 12:18:03.730638 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode Jan 17 12:18:03.730645 kernel: SRBDS: Unknown: Dependent on hypervisor status Jan 17 12:18:03.730651 kernel: GDS: Unknown: Dependent on hypervisor status Jan 17 12:18:03.730656 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Jan 17 12:18:03.730662 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Jan 17 12:18:03.730667 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Jan 17 12:18:03.730673 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Jan 17 12:18:03.730679 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Jan 17 12:18:03.730684 kernel: Freeing SMP alternatives memory: 32K Jan 17 12:18:03.730690 kernel: pid_max: default: 131072 minimum: 1024 Jan 17 12:18:03.730696 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Jan 17 12:18:03.730702 kernel: landlock: Up and running. Jan 17 12:18:03.730708 kernel: SELinux: Initializing. Jan 17 12:18:03.730713 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Jan 17 12:18:03.730719 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Jan 17 12:18:03.730725 kernel: smpboot: CPU0: Intel(R) Xeon(R) E-2278G CPU @ 3.40GHz (family: 0x6, model: 0x9e, stepping: 0xd) Jan 17 12:18:03.730731 kernel: RCU Tasks: Setting shift to 7 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=128. Jan 17 12:18:03.730736 kernel: RCU Tasks Rude: Setting shift to 7 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=128. Jan 17 12:18:03.730742 kernel: RCU Tasks Trace: Setting shift to 7 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=128. Jan 17 12:18:03.730749 kernel: Performance Events: Skylake events, core PMU driver. Jan 17 12:18:03.730754 kernel: core: CPUID marked event: 'cpu cycles' unavailable Jan 17 12:18:03.730760 kernel: core: CPUID marked event: 'instructions' unavailable Jan 17 12:18:03.730765 kernel: core: CPUID marked event: 'bus cycles' unavailable Jan 17 12:18:03.730771 kernel: core: CPUID marked event: 'cache references' unavailable Jan 17 12:18:03.730776 kernel: core: CPUID marked event: 'cache misses' unavailable Jan 17 12:18:03.730782 kernel: core: CPUID marked event: 'branch instructions' unavailable Jan 17 12:18:03.730787 kernel: core: CPUID marked event: 'branch misses' unavailable Jan 17 12:18:03.730794 kernel: ... version: 1 Jan 17 12:18:03.730799 kernel: ... bit width: 48 Jan 17 12:18:03.730805 kernel: ... generic registers: 4 Jan 17 12:18:03.730810 kernel: ... value mask: 0000ffffffffffff Jan 17 12:18:03.730816 kernel: ... max period: 000000007fffffff Jan 17 12:18:03.730822 kernel: ... fixed-purpose events: 0 Jan 17 12:18:03.730827 kernel: ... event mask: 000000000000000f Jan 17 12:18:03.730833 kernel: signal: max sigframe size: 1776 Jan 17 12:18:03.730838 kernel: rcu: Hierarchical SRCU implementation. Jan 17 12:18:03.730845 kernel: rcu: Max phase no-delay instances is 400. Jan 17 12:18:03.730851 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Jan 17 12:18:03.730856 kernel: smp: Bringing up secondary CPUs ... Jan 17 12:18:03.730862 kernel: smpboot: x86: Booting SMP configuration: Jan 17 12:18:03.730867 kernel: .... node #0, CPUs: #1 Jan 17 12:18:03.730873 kernel: Disabled fast string operations Jan 17 12:18:03.730878 kernel: smpboot: CPU 1 Converting physical 2 to logical package 1 Jan 17 12:18:03.730884 kernel: smpboot: CPU 1 Converting physical 0 to logical die 1 Jan 17 12:18:03.730889 kernel: smp: Brought up 1 node, 2 CPUs Jan 17 12:18:03.730895 kernel: smpboot: Max logical packages: 128 Jan 17 12:18:03.730902 kernel: smpboot: Total of 2 processors activated (13632.00 BogoMIPS) Jan 17 12:18:03.730907 kernel: devtmpfs: initialized Jan 17 12:18:03.730913 kernel: x86/mm: Memory block size: 128MB Jan 17 12:18:03.730918 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x7feff000-0x7fefffff] (4096 bytes) Jan 17 12:18:03.730924 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Jan 17 12:18:03.730930 kernel: futex hash table entries: 32768 (order: 9, 2097152 bytes, linear) Jan 17 12:18:03.730935 kernel: pinctrl core: initialized pinctrl subsystem Jan 17 12:18:03.730941 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Jan 17 12:18:03.730946 kernel: audit: initializing netlink subsys (disabled) Jan 17 12:18:03.730953 kernel: audit: type=2000 audit(1737116282.066:1): state=initialized audit_enabled=0 res=1 Jan 17 12:18:03.730959 kernel: thermal_sys: Registered thermal governor 'step_wise' Jan 17 12:18:03.730965 kernel: thermal_sys: Registered thermal governor 'user_space' Jan 17 12:18:03.730971 kernel: cpuidle: using governor menu Jan 17 12:18:03.730976 kernel: Simple Boot Flag at 0x36 set to 0x80 Jan 17 12:18:03.730982 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Jan 17 12:18:03.730987 kernel: dca service started, version 1.12.1 Jan 17 12:18:03.730993 kernel: PCI: MMCONFIG for domain 0000 [bus 00-7f] at [mem 0xf0000000-0xf7ffffff] (base 0xf0000000) Jan 17 12:18:03.730999 kernel: PCI: Using configuration type 1 for base access Jan 17 12:18:03.731006 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Jan 17 12:18:03.731012 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Jan 17 12:18:03.731017 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Jan 17 12:18:03.731023 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Jan 17 12:18:03.731028 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Jan 17 12:18:03.731034 kernel: ACPI: Added _OSI(Module Device) Jan 17 12:18:03.731039 kernel: ACPI: Added _OSI(Processor Device) Jan 17 12:18:03.731045 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Jan 17 12:18:03.731051 kernel: ACPI: Added _OSI(Processor Aggregator Device) Jan 17 12:18:03.731057 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Jan 17 12:18:03.731063 kernel: ACPI: [Firmware Bug]: BIOS _OSI(Linux) query ignored Jan 17 12:18:03.731068 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Jan 17 12:18:03.731074 kernel: ACPI: Interpreter enabled Jan 17 12:18:03.731079 kernel: ACPI: PM: (supports S0 S1 S5) Jan 17 12:18:03.731085 kernel: ACPI: Using IOAPIC for interrupt routing Jan 17 12:18:03.731090 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Jan 17 12:18:03.731096 kernel: PCI: Using E820 reservations for host bridge windows Jan 17 12:18:03.731114 kernel: ACPI: Enabled 4 GPEs in block 00 to 0F Jan 17 12:18:03.731123 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-7f]) Jan 17 12:18:03.731203 kernel: acpi PNP0A03:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Jan 17 12:18:03.731255 kernel: acpi PNP0A03:00: _OSC: platform does not support [AER LTR] Jan 17 12:18:03.731302 kernel: acpi PNP0A03:00: _OSC: OS now controls [PCIeHotplug PME PCIeCapability] Jan 17 12:18:03.731310 kernel: PCI host bridge to bus 0000:00 Jan 17 12:18:03.731358 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Jan 17 12:18:03.731408 kernel: pci_bus 0000:00: root bus resource [mem 0x000cc000-0x000dbfff window] Jan 17 12:18:03.731450 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] Jan 17 12:18:03.731492 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Jan 17 12:18:03.731533 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xfeff window] Jan 17 12:18:03.731575 kernel: pci_bus 0000:00: root bus resource [bus 00-7f] Jan 17 12:18:03.731631 kernel: pci 0000:00:00.0: [8086:7190] type 00 class 0x060000 Jan 17 12:18:03.731684 kernel: pci 0000:00:01.0: [8086:7191] type 01 class 0x060400 Jan 17 12:18:03.731738 kernel: pci 0000:00:07.0: [8086:7110] type 00 class 0x060100 Jan 17 12:18:03.731789 kernel: pci 0000:00:07.1: [8086:7111] type 00 class 0x01018a Jan 17 12:18:03.731836 kernel: pci 0000:00:07.1: reg 0x20: [io 0x1060-0x106f] Jan 17 12:18:03.731883 kernel: pci 0000:00:07.1: legacy IDE quirk: reg 0x10: [io 0x01f0-0x01f7] Jan 17 12:18:03.731929 kernel: pci 0000:00:07.1: legacy IDE quirk: reg 0x14: [io 0x03f6] Jan 17 12:18:03.731976 kernel: pci 0000:00:07.1: legacy IDE quirk: reg 0x18: [io 0x0170-0x0177] Jan 17 12:18:03.732025 kernel: pci 0000:00:07.1: legacy IDE quirk: reg 0x1c: [io 0x0376] Jan 17 12:18:03.732076 kernel: pci 0000:00:07.3: [8086:7113] type 00 class 0x068000 Jan 17 12:18:03.732172 kernel: pci 0000:00:07.3: quirk: [io 0x1000-0x103f] claimed by PIIX4 ACPI Jan 17 12:18:03.732220 kernel: pci 0000:00:07.3: quirk: [io 0x1040-0x104f] claimed by PIIX4 SMB Jan 17 12:18:03.732270 kernel: pci 0000:00:07.7: [15ad:0740] type 00 class 0x088000 Jan 17 12:18:03.732317 kernel: pci 0000:00:07.7: reg 0x10: [io 0x1080-0x10bf] Jan 17 12:18:03.732363 kernel: pci 0000:00:07.7: reg 0x14: [mem 0xfebfe000-0xfebfffff 64bit] Jan 17 12:18:03.732458 kernel: pci 0000:00:0f.0: [15ad:0405] type 00 class 0x030000 Jan 17 12:18:03.732506 kernel: pci 0000:00:0f.0: reg 0x10: [io 0x1070-0x107f] Jan 17 12:18:03.732553 kernel: pci 0000:00:0f.0: reg 0x14: [mem 0xe8000000-0xefffffff pref] Jan 17 12:18:03.732600 kernel: pci 0000:00:0f.0: reg 0x18: [mem 0xfe000000-0xfe7fffff] Jan 17 12:18:03.732647 kernel: pci 0000:00:0f.0: reg 0x30: [mem 0x00000000-0x00007fff pref] Jan 17 12:18:03.732694 kernel: pci 0000:00:0f.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Jan 17 12:18:03.732744 kernel: pci 0000:00:11.0: [15ad:0790] type 01 class 0x060401 Jan 17 12:18:03.732800 kernel: pci 0000:00:15.0: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.732849 kernel: pci 0000:00:15.0: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.732901 kernel: pci 0000:00:15.1: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.732950 kernel: pci 0000:00:15.1: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.733003 kernel: pci 0000:00:15.2: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.733052 kernel: pci 0000:00:15.2: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.733136 kernel: pci 0000:00:15.3: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.733186 kernel: pci 0000:00:15.3: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.733237 kernel: pci 0000:00:15.4: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.733287 kernel: pci 0000:00:15.4: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.733338 kernel: pci 0000:00:15.5: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.733387 kernel: pci 0000:00:15.5: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.733442 kernel: pci 0000:00:15.6: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.733490 kernel: pci 0000:00:15.6: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.733543 kernel: pci 0000:00:15.7: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.733590 kernel: pci 0000:00:15.7: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.733641 kernel: pci 0000:00:16.0: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.733689 kernel: pci 0000:00:16.0: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.733743 kernel: pci 0000:00:16.1: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.733791 kernel: pci 0000:00:16.1: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.733841 kernel: pci 0000:00:16.2: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.733888 kernel: pci 0000:00:16.2: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.733940 kernel: pci 0000:00:16.3: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.733990 kernel: pci 0000:00:16.3: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.734042 kernel: pci 0000:00:16.4: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.734090 kernel: pci 0000:00:16.4: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.734160 kernel: pci 0000:00:16.5: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.734210 kernel: pci 0000:00:16.5: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.734262 kernel: pci 0000:00:16.6: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.734310 kernel: pci 0000:00:16.6: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.734364 kernel: pci 0000:00:16.7: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.734417 kernel: pci 0000:00:16.7: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.734468 kernel: pci 0000:00:17.0: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.734515 kernel: pci 0000:00:17.0: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.734566 kernel: pci 0000:00:17.1: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.734614 kernel: pci 0000:00:17.1: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.734667 kernel: pci 0000:00:17.2: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.734715 kernel: pci 0000:00:17.2: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.734766 kernel: pci 0000:00:17.3: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.734814 kernel: pci 0000:00:17.3: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.734865 kernel: pci 0000:00:17.4: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.734913 kernel: pci 0000:00:17.4: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.734966 kernel: pci 0000:00:17.5: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.735013 kernel: pci 0000:00:17.5: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.735064 kernel: pci 0000:00:17.6: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.735147 kernel: pci 0000:00:17.6: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.735204 kernel: pci 0000:00:17.7: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.735252 kernel: pci 0000:00:17.7: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.735306 kernel: pci 0000:00:18.0: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.735354 kernel: pci 0000:00:18.0: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.735422 kernel: pci 0000:00:18.1: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.735484 kernel: pci 0000:00:18.1: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.735536 kernel: pci 0000:00:18.2: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.735584 kernel: pci 0000:00:18.2: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.735636 kernel: pci 0000:00:18.3: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.735686 kernel: pci 0000:00:18.3: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.735738 kernel: pci 0000:00:18.4: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.735787 kernel: pci 0000:00:18.4: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.735838 kernel: pci 0000:00:18.5: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.735886 kernel: pci 0000:00:18.5: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.735936 kernel: pci 0000:00:18.6: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.735986 kernel: pci 0000:00:18.6: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.736039 kernel: pci 0000:00:18.7: [15ad:07a0] type 01 class 0x060400 Jan 17 12:18:03.736087 kernel: pci 0000:00:18.7: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.736190 kernel: pci_bus 0000:01: extended config space not accessible Jan 17 12:18:03.736240 kernel: pci 0000:00:01.0: PCI bridge to [bus 01] Jan 17 12:18:03.736287 kernel: pci_bus 0000:02: extended config space not accessible Jan 17 12:18:03.736298 kernel: acpiphp: Slot [32] registered Jan 17 12:18:03.736304 kernel: acpiphp: Slot [33] registered Jan 17 12:18:03.736309 kernel: acpiphp: Slot [34] registered Jan 17 12:18:03.736315 kernel: acpiphp: Slot [35] registered Jan 17 12:18:03.736321 kernel: acpiphp: Slot [36] registered Jan 17 12:18:03.736326 kernel: acpiphp: Slot [37] registered Jan 17 12:18:03.736332 kernel: acpiphp: Slot [38] registered Jan 17 12:18:03.736338 kernel: acpiphp: Slot [39] registered Jan 17 12:18:03.736344 kernel: acpiphp: Slot [40] registered Jan 17 12:18:03.736351 kernel: acpiphp: Slot [41] registered Jan 17 12:18:03.736356 kernel: acpiphp: Slot [42] registered Jan 17 12:18:03.736362 kernel: acpiphp: Slot [43] registered Jan 17 12:18:03.736367 kernel: acpiphp: Slot [44] registered Jan 17 12:18:03.736373 kernel: acpiphp: Slot [45] registered Jan 17 12:18:03.736378 kernel: acpiphp: Slot [46] registered Jan 17 12:18:03.736384 kernel: acpiphp: Slot [47] registered Jan 17 12:18:03.736390 kernel: acpiphp: Slot [48] registered Jan 17 12:18:03.736395 kernel: acpiphp: Slot [49] registered Jan 17 12:18:03.736401 kernel: acpiphp: Slot [50] registered Jan 17 12:18:03.736407 kernel: acpiphp: Slot [51] registered Jan 17 12:18:03.736413 kernel: acpiphp: Slot [52] registered Jan 17 12:18:03.736419 kernel: acpiphp: Slot [53] registered Jan 17 12:18:03.736424 kernel: acpiphp: Slot [54] registered Jan 17 12:18:03.736430 kernel: acpiphp: Slot [55] registered Jan 17 12:18:03.736435 kernel: acpiphp: Slot [56] registered Jan 17 12:18:03.736441 kernel: acpiphp: Slot [57] registered Jan 17 12:18:03.736446 kernel: acpiphp: Slot [58] registered Jan 17 12:18:03.736452 kernel: acpiphp: Slot [59] registered Jan 17 12:18:03.736458 kernel: acpiphp: Slot [60] registered Jan 17 12:18:03.736464 kernel: acpiphp: Slot [61] registered Jan 17 12:18:03.736470 kernel: acpiphp: Slot [62] registered Jan 17 12:18:03.736475 kernel: acpiphp: Slot [63] registered Jan 17 12:18:03.736522 kernel: pci 0000:00:11.0: PCI bridge to [bus 02] (subtractive decode) Jan 17 12:18:03.736570 kernel: pci 0000:00:11.0: bridge window [io 0x2000-0x3fff] Jan 17 12:18:03.736616 kernel: pci 0000:00:11.0: bridge window [mem 0xfd600000-0xfdffffff] Jan 17 12:18:03.736662 kernel: pci 0000:00:11.0: bridge window [mem 0xe7b00000-0xe7ffffff 64bit pref] Jan 17 12:18:03.736709 kernel: pci 0000:00:11.0: bridge window [mem 0x000a0000-0x000bffff window] (subtractive decode) Jan 17 12:18:03.736758 kernel: pci 0000:00:11.0: bridge window [mem 0x000cc000-0x000dbfff window] (subtractive decode) Jan 17 12:18:03.736804 kernel: pci 0000:00:11.0: bridge window [mem 0xc0000000-0xfebfffff window] (subtractive decode) Jan 17 12:18:03.736850 kernel: pci 0000:00:11.0: bridge window [io 0x0000-0x0cf7 window] (subtractive decode) Jan 17 12:18:03.736896 kernel: pci 0000:00:11.0: bridge window [io 0x0d00-0xfeff window] (subtractive decode) Jan 17 12:18:03.736949 kernel: pci 0000:03:00.0: [15ad:07c0] type 00 class 0x010700 Jan 17 12:18:03.736997 kernel: pci 0000:03:00.0: reg 0x10: [io 0x4000-0x4007] Jan 17 12:18:03.737045 kernel: pci 0000:03:00.0: reg 0x14: [mem 0xfd5f8000-0xfd5fffff 64bit] Jan 17 12:18:03.737096 kernel: pci 0000:03:00.0: reg 0x30: [mem 0x00000000-0x0000ffff pref] Jan 17 12:18:03.737237 kernel: pci 0000:03:00.0: PME# supported from D0 D3hot D3cold Jan 17 12:18:03.737285 kernel: pci 0000:03:00.0: disabling ASPM on pre-1.1 PCIe device. You can enable it with 'pcie_aspm=force' Jan 17 12:18:03.737334 kernel: pci 0000:00:15.0: PCI bridge to [bus 03] Jan 17 12:18:03.737381 kernel: pci 0000:00:15.0: bridge window [io 0x4000-0x4fff] Jan 17 12:18:03.737432 kernel: pci 0000:00:15.0: bridge window [mem 0xfd500000-0xfd5fffff] Jan 17 12:18:03.737511 kernel: pci 0000:00:15.1: PCI bridge to [bus 04] Jan 17 12:18:03.737563 kernel: pci 0000:00:15.1: bridge window [io 0x8000-0x8fff] Jan 17 12:18:03.737611 kernel: pci 0000:00:15.1: bridge window [mem 0xfd100000-0xfd1fffff] Jan 17 12:18:03.737676 kernel: pci 0000:00:15.1: bridge window [mem 0xe7800000-0xe78fffff 64bit pref] Jan 17 12:18:03.737726 kernel: pci 0000:00:15.2: PCI bridge to [bus 05] Jan 17 12:18:03.737774 kernel: pci 0000:00:15.2: bridge window [io 0xc000-0xcfff] Jan 17 12:18:03.737822 kernel: pci 0000:00:15.2: bridge window [mem 0xfcd00000-0xfcdfffff] Jan 17 12:18:03.737871 kernel: pci 0000:00:15.2: bridge window [mem 0xe7400000-0xe74fffff 64bit pref] Jan 17 12:18:03.737919 kernel: pci 0000:00:15.3: PCI bridge to [bus 06] Jan 17 12:18:03.737970 kernel: pci 0000:00:15.3: bridge window [mem 0xfc900000-0xfc9fffff] Jan 17 12:18:03.738018 kernel: pci 0000:00:15.3: bridge window [mem 0xe7000000-0xe70fffff 64bit pref] Jan 17 12:18:03.738067 kernel: pci 0000:00:15.4: PCI bridge to [bus 07] Jan 17 12:18:03.738133 kernel: pci 0000:00:15.4: bridge window [mem 0xfc500000-0xfc5fffff] Jan 17 12:18:03.738185 kernel: pci 0000:00:15.4: bridge window [mem 0xe6c00000-0xe6cfffff 64bit pref] Jan 17 12:18:03.738237 kernel: pci 0000:00:15.5: PCI bridge to [bus 08] Jan 17 12:18:03.738286 kernel: pci 0000:00:15.5: bridge window [mem 0xfc100000-0xfc1fffff] Jan 17 12:18:03.738334 kernel: pci 0000:00:15.5: bridge window [mem 0xe6800000-0xe68fffff 64bit pref] Jan 17 12:18:03.738383 kernel: pci 0000:00:15.6: PCI bridge to [bus 09] Jan 17 12:18:03.738447 kernel: pci 0000:00:15.6: bridge window [mem 0xfbd00000-0xfbdfffff] Jan 17 12:18:03.738495 kernel: pci 0000:00:15.6: bridge window [mem 0xe6400000-0xe64fffff 64bit pref] Jan 17 12:18:03.738544 kernel: pci 0000:00:15.7: PCI bridge to [bus 0a] Jan 17 12:18:03.738642 kernel: pci 0000:00:15.7: bridge window [mem 0xfb900000-0xfb9fffff] Jan 17 12:18:03.738694 kernel: pci 0000:00:15.7: bridge window [mem 0xe6000000-0xe60fffff 64bit pref] Jan 17 12:18:03.738749 kernel: pci 0000:0b:00.0: [15ad:07b0] type 00 class 0x020000 Jan 17 12:18:03.738799 kernel: pci 0000:0b:00.0: reg 0x10: [mem 0xfd4fc000-0xfd4fcfff] Jan 17 12:18:03.738848 kernel: pci 0000:0b:00.0: reg 0x14: [mem 0xfd4fd000-0xfd4fdfff] Jan 17 12:18:03.738898 kernel: pci 0000:0b:00.0: reg 0x18: [mem 0xfd4fe000-0xfd4fffff] Jan 17 12:18:03.738947 kernel: pci 0000:0b:00.0: reg 0x1c: [io 0x5000-0x500f] Jan 17 12:18:03.738996 kernel: pci 0000:0b:00.0: reg 0x30: [mem 0x00000000-0x0000ffff pref] Jan 17 12:18:03.739050 kernel: pci 0000:0b:00.0: supports D1 D2 Jan 17 12:18:03.739124 kernel: pci 0000:0b:00.0: PME# supported from D0 D1 D2 D3hot D3cold Jan 17 12:18:03.739182 kernel: pci 0000:0b:00.0: disabling ASPM on pre-1.1 PCIe device. You can enable it with 'pcie_aspm=force' Jan 17 12:18:03.739231 kernel: pci 0000:00:16.0: PCI bridge to [bus 0b] Jan 17 12:18:03.739279 kernel: pci 0000:00:16.0: bridge window [io 0x5000-0x5fff] Jan 17 12:18:03.739327 kernel: pci 0000:00:16.0: bridge window [mem 0xfd400000-0xfd4fffff] Jan 17 12:18:03.739375 kernel: pci 0000:00:16.1: PCI bridge to [bus 0c] Jan 17 12:18:03.739454 kernel: pci 0000:00:16.1: bridge window [io 0x9000-0x9fff] Jan 17 12:18:03.739503 kernel: pci 0000:00:16.1: bridge window [mem 0xfd000000-0xfd0fffff] Jan 17 12:18:03.739552 kernel: pci 0000:00:16.1: bridge window [mem 0xe7700000-0xe77fffff 64bit pref] Jan 17 12:18:03.739601 kernel: pci 0000:00:16.2: PCI bridge to [bus 0d] Jan 17 12:18:03.739652 kernel: pci 0000:00:16.2: bridge window [io 0xd000-0xdfff] Jan 17 12:18:03.739701 kernel: pci 0000:00:16.2: bridge window [mem 0xfcc00000-0xfccfffff] Jan 17 12:18:03.739750 kernel: pci 0000:00:16.2: bridge window [mem 0xe7300000-0xe73fffff 64bit pref] Jan 17 12:18:03.739800 kernel: pci 0000:00:16.3: PCI bridge to [bus 0e] Jan 17 12:18:03.739851 kernel: pci 0000:00:16.3: bridge window [mem 0xfc800000-0xfc8fffff] Jan 17 12:18:03.739900 kernel: pci 0000:00:16.3: bridge window [mem 0xe6f00000-0xe6ffffff 64bit pref] Jan 17 12:18:03.739949 kernel: pci 0000:00:16.4: PCI bridge to [bus 0f] Jan 17 12:18:03.739998 kernel: pci 0000:00:16.4: bridge window [mem 0xfc400000-0xfc4fffff] Jan 17 12:18:03.740047 kernel: pci 0000:00:16.4: bridge window [mem 0xe6b00000-0xe6bfffff 64bit pref] Jan 17 12:18:03.740097 kernel: pci 0000:00:16.5: PCI bridge to [bus 10] Jan 17 12:18:03.740323 kernel: pci 0000:00:16.5: bridge window [mem 0xfc000000-0xfc0fffff] Jan 17 12:18:03.740373 kernel: pci 0000:00:16.5: bridge window [mem 0xe6700000-0xe67fffff 64bit pref] Jan 17 12:18:03.740426 kernel: pci 0000:00:16.6: PCI bridge to [bus 11] Jan 17 12:18:03.740477 kernel: pci 0000:00:16.6: bridge window [mem 0xfbc00000-0xfbcfffff] Jan 17 12:18:03.740526 kernel: pci 0000:00:16.6: bridge window [mem 0xe6300000-0xe63fffff 64bit pref] Jan 17 12:18:03.740575 kernel: pci 0000:00:16.7: PCI bridge to [bus 12] Jan 17 12:18:03.740624 kernel: pci 0000:00:16.7: bridge window [mem 0xfb800000-0xfb8fffff] Jan 17 12:18:03.740673 kernel: pci 0000:00:16.7: bridge window [mem 0xe5f00000-0xe5ffffff 64bit pref] Jan 17 12:18:03.740722 kernel: pci 0000:00:17.0: PCI bridge to [bus 13] Jan 17 12:18:03.740770 kernel: pci 0000:00:17.0: bridge window [io 0x6000-0x6fff] Jan 17 12:18:03.740821 kernel: pci 0000:00:17.0: bridge window [mem 0xfd300000-0xfd3fffff] Jan 17 12:18:03.740869 kernel: pci 0000:00:17.0: bridge window [mem 0xe7a00000-0xe7afffff 64bit pref] Jan 17 12:18:03.740918 kernel: pci 0000:00:17.1: PCI bridge to [bus 14] Jan 17 12:18:03.740968 kernel: pci 0000:00:17.1: bridge window [io 0xa000-0xafff] Jan 17 12:18:03.741016 kernel: pci 0000:00:17.1: bridge window [mem 0xfcf00000-0xfcffffff] Jan 17 12:18:03.741065 kernel: pci 0000:00:17.1: bridge window [mem 0xe7600000-0xe76fffff 64bit pref] Jan 17 12:18:03.741149 kernel: pci 0000:00:17.2: PCI bridge to [bus 15] Jan 17 12:18:03.741203 kernel: pci 0000:00:17.2: bridge window [io 0xe000-0xefff] Jan 17 12:18:03.741252 kernel: pci 0000:00:17.2: bridge window [mem 0xfcb00000-0xfcbfffff] Jan 17 12:18:03.741300 kernel: pci 0000:00:17.2: bridge window [mem 0xe7200000-0xe72fffff 64bit pref] Jan 17 12:18:03.741350 kernel: pci 0000:00:17.3: PCI bridge to [bus 16] Jan 17 12:18:03.741402 kernel: pci 0000:00:17.3: bridge window [mem 0xfc700000-0xfc7fffff] Jan 17 12:18:03.741451 kernel: pci 0000:00:17.3: bridge window [mem 0xe6e00000-0xe6efffff 64bit pref] Jan 17 12:18:03.741501 kernel: pci 0000:00:17.4: PCI bridge to [bus 17] Jan 17 12:18:03.741849 kernel: pci 0000:00:17.4: bridge window [mem 0xfc300000-0xfc3fffff] Jan 17 12:18:03.743843 kernel: pci 0000:00:17.4: bridge window [mem 0xe6a00000-0xe6afffff 64bit pref] Jan 17 12:18:03.743906 kernel: pci 0000:00:17.5: PCI bridge to [bus 18] Jan 17 12:18:03.743959 kernel: pci 0000:00:17.5: bridge window [mem 0xfbf00000-0xfbffffff] Jan 17 12:18:03.744009 kernel: pci 0000:00:17.5: bridge window [mem 0xe6600000-0xe66fffff 64bit pref] Jan 17 12:18:03.744061 kernel: pci 0000:00:17.6: PCI bridge to [bus 19] Jan 17 12:18:03.744137 kernel: pci 0000:00:17.6: bridge window [mem 0xfbb00000-0xfbbfffff] Jan 17 12:18:03.744186 kernel: pci 0000:00:17.6: bridge window [mem 0xe6200000-0xe62fffff 64bit pref] Jan 17 12:18:03.744236 kernel: pci 0000:00:17.7: PCI bridge to [bus 1a] Jan 17 12:18:03.744287 kernel: pci 0000:00:17.7: bridge window [mem 0xfb700000-0xfb7fffff] Jan 17 12:18:03.744335 kernel: pci 0000:00:17.7: bridge window [mem 0xe5e00000-0xe5efffff 64bit pref] Jan 17 12:18:03.744384 kernel: pci 0000:00:18.0: PCI bridge to [bus 1b] Jan 17 12:18:03.744437 kernel: pci 0000:00:18.0: bridge window [io 0x7000-0x7fff] Jan 17 12:18:03.744484 kernel: pci 0000:00:18.0: bridge window [mem 0xfd200000-0xfd2fffff] Jan 17 12:18:03.744532 kernel: pci 0000:00:18.0: bridge window [mem 0xe7900000-0xe79fffff 64bit pref] Jan 17 12:18:03.744583 kernel: pci 0000:00:18.1: PCI bridge to [bus 1c] Jan 17 12:18:03.744631 kernel: pci 0000:00:18.1: bridge window [io 0xb000-0xbfff] Jan 17 12:18:03.744681 kernel: pci 0000:00:18.1: bridge window [mem 0xfce00000-0xfcefffff] Jan 17 12:18:03.744729 kernel: pci 0000:00:18.1: bridge window [mem 0xe7500000-0xe75fffff 64bit pref] Jan 17 12:18:03.744779 kernel: pci 0000:00:18.2: PCI bridge to [bus 1d] Jan 17 12:18:03.744827 kernel: pci 0000:00:18.2: bridge window [mem 0xfca00000-0xfcafffff] Jan 17 12:18:03.744875 kernel: pci 0000:00:18.2: bridge window [mem 0xe7100000-0xe71fffff 64bit pref] Jan 17 12:18:03.744925 kernel: pci 0000:00:18.3: PCI bridge to [bus 1e] Jan 17 12:18:03.744973 kernel: pci 0000:00:18.3: bridge window [mem 0xfc600000-0xfc6fffff] Jan 17 12:18:03.745021 kernel: pci 0000:00:18.3: bridge window [mem 0xe6d00000-0xe6dfffff 64bit pref] Jan 17 12:18:03.745073 kernel: pci 0000:00:18.4: PCI bridge to [bus 1f] Jan 17 12:18:03.746200 kernel: pci 0000:00:18.4: bridge window [mem 0xfc200000-0xfc2fffff] Jan 17 12:18:03.746258 kernel: pci 0000:00:18.4: bridge window [mem 0xe6900000-0xe69fffff 64bit pref] Jan 17 12:18:03.746329 kernel: pci 0000:00:18.5: PCI bridge to [bus 20] Jan 17 12:18:03.746382 kernel: pci 0000:00:18.5: bridge window [mem 0xfbe00000-0xfbefffff] Jan 17 12:18:03.746431 kernel: pci 0000:00:18.5: bridge window [mem 0xe6500000-0xe65fffff 64bit pref] Jan 17 12:18:03.746481 kernel: pci 0000:00:18.6: PCI bridge to [bus 21] Jan 17 12:18:03.746530 kernel: pci 0000:00:18.6: bridge window [mem 0xfba00000-0xfbafffff] Jan 17 12:18:03.746582 kernel: pci 0000:00:18.6: bridge window [mem 0xe6100000-0xe61fffff 64bit pref] Jan 17 12:18:03.746632 kernel: pci 0000:00:18.7: PCI bridge to [bus 22] Jan 17 12:18:03.746681 kernel: pci 0000:00:18.7: bridge window [mem 0xfb600000-0xfb6fffff] Jan 17 12:18:03.746729 kernel: pci 0000:00:18.7: bridge window [mem 0xe5d00000-0xe5dfffff 64bit pref] Jan 17 12:18:03.746737 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 9 Jan 17 12:18:03.746743 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 0 Jan 17 12:18:03.746749 kernel: ACPI: PCI: Interrupt link LNKB disabled Jan 17 12:18:03.746755 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Jan 17 12:18:03.746763 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 10 Jan 17 12:18:03.746768 kernel: iommu: Default domain type: Translated Jan 17 12:18:03.746774 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Jan 17 12:18:03.746780 kernel: PCI: Using ACPI for IRQ routing Jan 17 12:18:03.746786 kernel: PCI: pci_cache_line_size set to 64 bytes Jan 17 12:18:03.746792 kernel: e820: reserve RAM buffer [mem 0x0009ec00-0x0009ffff] Jan 17 12:18:03.746797 kernel: e820: reserve RAM buffer [mem 0x7fee0000-0x7fffffff] Jan 17 12:18:03.746845 kernel: pci 0000:00:0f.0: vgaarb: setting as boot VGA device Jan 17 12:18:03.746894 kernel: pci 0000:00:0f.0: vgaarb: bridge control possible Jan 17 12:18:03.746944 kernel: pci 0000:00:0f.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Jan 17 12:18:03.746952 kernel: vgaarb: loaded Jan 17 12:18:03.746958 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 Jan 17 12:18:03.746964 kernel: hpet0: 16 comparators, 64-bit 14.318180 MHz counter Jan 17 12:18:03.746970 kernel: clocksource: Switched to clocksource tsc-early Jan 17 12:18:03.746976 kernel: VFS: Disk quotas dquot_6.6.0 Jan 17 12:18:03.746982 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Jan 17 12:18:03.746987 kernel: pnp: PnP ACPI init Jan 17 12:18:03.747037 kernel: system 00:00: [io 0x1000-0x103f] has been reserved Jan 17 12:18:03.747085 kernel: system 00:00: [io 0x1040-0x104f] has been reserved Jan 17 12:18:03.748141 kernel: system 00:00: [io 0x0cf0-0x0cf1] has been reserved Jan 17 12:18:03.748193 kernel: system 00:04: [mem 0xfed00000-0xfed003ff] has been reserved Jan 17 12:18:03.748262 kernel: pnp 00:06: [dma 2] Jan 17 12:18:03.748310 kernel: system 00:07: [io 0xfce0-0xfcff] has been reserved Jan 17 12:18:03.748355 kernel: system 00:07: [mem 0xf0000000-0xf7ffffff] has been reserved Jan 17 12:18:03.748448 kernel: system 00:07: [mem 0xfe800000-0xfe9fffff] has been reserved Jan 17 12:18:03.748458 kernel: pnp: PnP ACPI: found 8 devices Jan 17 12:18:03.748464 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Jan 17 12:18:03.748470 kernel: NET: Registered PF_INET protocol family Jan 17 12:18:03.748476 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) Jan 17 12:18:03.748482 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) Jan 17 12:18:03.748488 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Jan 17 12:18:03.748494 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) Jan 17 12:18:03.748502 kernel: TCP bind hash table entries: 16384 (order: 7, 524288 bytes, linear) Jan 17 12:18:03.748508 kernel: TCP: Hash tables configured (established 16384 bind 16384) Jan 17 12:18:03.748514 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) Jan 17 12:18:03.748519 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) Jan 17 12:18:03.748525 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Jan 17 12:18:03.748531 kernel: NET: Registered PF_XDP protocol family Jan 17 12:18:03.748582 kernel: pci 0000:00:15.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 03] add_size 200000 add_align 100000 Jan 17 12:18:03.748632 kernel: pci 0000:00:15.3: bridge window [io 0x1000-0x0fff] to [bus 06] add_size 1000 Jan 17 12:18:03.748684 kernel: pci 0000:00:15.4: bridge window [io 0x1000-0x0fff] to [bus 07] add_size 1000 Jan 17 12:18:03.748735 kernel: pci 0000:00:15.5: bridge window [io 0x1000-0x0fff] to [bus 08] add_size 1000 Jan 17 12:18:03.748784 kernel: pci 0000:00:15.6: bridge window [io 0x1000-0x0fff] to [bus 09] add_size 1000 Jan 17 12:18:03.748833 kernel: pci 0000:00:15.7: bridge window [io 0x1000-0x0fff] to [bus 0a] add_size 1000 Jan 17 12:18:03.748883 kernel: pci 0000:00:16.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 0b] add_size 200000 add_align 100000 Jan 17 12:18:03.748932 kernel: pci 0000:00:16.3: bridge window [io 0x1000-0x0fff] to [bus 0e] add_size 1000 Jan 17 12:18:03.748984 kernel: pci 0000:00:16.4: bridge window [io 0x1000-0x0fff] to [bus 0f] add_size 1000 Jan 17 12:18:03.749033 kernel: pci 0000:00:16.5: bridge window [io 0x1000-0x0fff] to [bus 10] add_size 1000 Jan 17 12:18:03.749097 kernel: pci 0000:00:16.6: bridge window [io 0x1000-0x0fff] to [bus 11] add_size 1000 Jan 17 12:18:03.750174 kernel: pci 0000:00:16.7: bridge window [io 0x1000-0x0fff] to [bus 12] add_size 1000 Jan 17 12:18:03.750224 kernel: pci 0000:00:17.3: bridge window [io 0x1000-0x0fff] to [bus 16] add_size 1000 Jan 17 12:18:03.750273 kernel: pci 0000:00:17.4: bridge window [io 0x1000-0x0fff] to [bus 17] add_size 1000 Jan 17 12:18:03.750325 kernel: pci 0000:00:17.5: bridge window [io 0x1000-0x0fff] to [bus 18] add_size 1000 Jan 17 12:18:03.750372 kernel: pci 0000:00:17.6: bridge window [io 0x1000-0x0fff] to [bus 19] add_size 1000 Jan 17 12:18:03.750491 kernel: pci 0000:00:17.7: bridge window [io 0x1000-0x0fff] to [bus 1a] add_size 1000 Jan 17 12:18:03.750539 kernel: pci 0000:00:18.2: bridge window [io 0x1000-0x0fff] to [bus 1d] add_size 1000 Jan 17 12:18:03.750587 kernel: pci 0000:00:18.3: bridge window [io 0x1000-0x0fff] to [bus 1e] add_size 1000 Jan 17 12:18:03.750634 kernel: pci 0000:00:18.4: bridge window [io 0x1000-0x0fff] to [bus 1f] add_size 1000 Jan 17 12:18:03.750684 kernel: pci 0000:00:18.5: bridge window [io 0x1000-0x0fff] to [bus 20] add_size 1000 Jan 17 12:18:03.750731 kernel: pci 0000:00:18.6: bridge window [io 0x1000-0x0fff] to [bus 21] add_size 1000 Jan 17 12:18:03.750779 kernel: pci 0000:00:18.7: bridge window [io 0x1000-0x0fff] to [bus 22] add_size 1000 Jan 17 12:18:03.750826 kernel: pci 0000:00:15.0: BAR 15: assigned [mem 0xc0000000-0xc01fffff 64bit pref] Jan 17 12:18:03.750891 kernel: pci 0000:00:16.0: BAR 15: assigned [mem 0xc0200000-0xc03fffff 64bit pref] Jan 17 12:18:03.750940 kernel: pci 0000:00:15.3: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.751045 kernel: pci 0000:00:15.3: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.752153 kernel: pci 0000:00:15.4: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.752205 kernel: pci 0000:00:15.4: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.752254 kernel: pci 0000:00:15.5: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.752301 kernel: pci 0000:00:15.5: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.752347 kernel: pci 0000:00:15.6: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.752395 kernel: pci 0000:00:15.6: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.752442 kernel: pci 0000:00:15.7: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.752491 kernel: pci 0000:00:15.7: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.752538 kernel: pci 0000:00:16.3: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.752584 kernel: pci 0000:00:16.3: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.752630 kernel: pci 0000:00:16.4: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.752677 kernel: pci 0000:00:16.4: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.752724 kernel: pci 0000:00:16.5: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.752770 kernel: pci 0000:00:16.5: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.752817 kernel: pci 0000:00:16.6: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.752866 kernel: pci 0000:00:16.6: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.752913 kernel: pci 0000:00:16.7: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.752960 kernel: pci 0000:00:16.7: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.753005 kernel: pci 0000:00:17.3: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.753052 kernel: pci 0000:00:17.3: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.754110 kernel: pci 0000:00:17.4: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.754169 kernel: pci 0000:00:17.4: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.754216 kernel: pci 0000:00:17.5: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.754266 kernel: pci 0000:00:17.5: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.754314 kernel: pci 0000:00:17.6: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.754361 kernel: pci 0000:00:17.6: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.754412 kernel: pci 0000:00:17.7: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.754458 kernel: pci 0000:00:17.7: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.754505 kernel: pci 0000:00:18.2: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.754553 kernel: pci 0000:00:18.2: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.754600 kernel: pci 0000:00:18.3: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.754648 kernel: pci 0000:00:18.3: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.754698 kernel: pci 0000:00:18.4: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.754746 kernel: pci 0000:00:18.4: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.754794 kernel: pci 0000:00:18.5: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.754842 kernel: pci 0000:00:18.5: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.754888 kernel: pci 0000:00:18.6: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.754935 kernel: pci 0000:00:18.6: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.754982 kernel: pci 0000:00:18.7: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.755030 kernel: pci 0000:00:18.7: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.755080 kernel: pci 0000:00:18.7: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.755135 kernel: pci 0000:00:18.7: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.755182 kernel: pci 0000:00:18.6: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.755229 kernel: pci 0000:00:18.6: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.755277 kernel: pci 0000:00:18.5: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.755323 kernel: pci 0000:00:18.5: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.755369 kernel: pci 0000:00:18.4: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.755417 kernel: pci 0000:00:18.4: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.755464 kernel: pci 0000:00:18.3: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.755514 kernel: pci 0000:00:18.3: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.755561 kernel: pci 0000:00:18.2: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.755608 kernel: pci 0000:00:18.2: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.755654 kernel: pci 0000:00:17.7: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.755701 kernel: pci 0000:00:17.7: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.755748 kernel: pci 0000:00:17.6: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.755794 kernel: pci 0000:00:17.6: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.755841 kernel: pci 0000:00:17.5: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.755888 kernel: pci 0000:00:17.5: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.755955 kernel: pci 0000:00:17.4: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.756002 kernel: pci 0000:00:17.4: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.756050 kernel: pci 0000:00:17.3: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.756097 kernel: pci 0000:00:17.3: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.758183 kernel: pci 0000:00:16.7: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.758233 kernel: pci 0000:00:16.7: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.758282 kernel: pci 0000:00:16.6: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.758330 kernel: pci 0000:00:16.6: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.758378 kernel: pci 0000:00:16.5: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.758453 kernel: pci 0000:00:16.5: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.758504 kernel: pci 0000:00:16.4: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.758551 kernel: pci 0000:00:16.4: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.758598 kernel: pci 0000:00:16.3: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.758646 kernel: pci 0000:00:16.3: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.758694 kernel: pci 0000:00:15.7: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.758742 kernel: pci 0000:00:15.7: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.758789 kernel: pci 0000:00:15.6: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.758837 kernel: pci 0000:00:15.6: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.758884 kernel: pci 0000:00:15.5: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.758934 kernel: pci 0000:00:15.5: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.758982 kernel: pci 0000:00:15.4: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.759030 kernel: pci 0000:00:15.4: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.759078 kernel: pci 0000:00:15.3: BAR 13: no space for [io size 0x1000] Jan 17 12:18:03.759139 kernel: pci 0000:00:15.3: BAR 13: failed to assign [io size 0x1000] Jan 17 12:18:03.759190 kernel: pci 0000:00:01.0: PCI bridge to [bus 01] Jan 17 12:18:03.759239 kernel: pci 0000:00:11.0: PCI bridge to [bus 02] Jan 17 12:18:03.759287 kernel: pci 0000:00:11.0: bridge window [io 0x2000-0x3fff] Jan 17 12:18:03.759335 kernel: pci 0000:00:11.0: bridge window [mem 0xfd600000-0xfdffffff] Jan 17 12:18:03.759382 kernel: pci 0000:00:11.0: bridge window [mem 0xe7b00000-0xe7ffffff 64bit pref] Jan 17 12:18:03.759437 kernel: pci 0000:03:00.0: BAR 6: assigned [mem 0xfd500000-0xfd50ffff pref] Jan 17 12:18:03.759486 kernel: pci 0000:00:15.0: PCI bridge to [bus 03] Jan 17 12:18:03.759535 kernel: pci 0000:00:15.0: bridge window [io 0x4000-0x4fff] Jan 17 12:18:03.759583 kernel: pci 0000:00:15.0: bridge window [mem 0xfd500000-0xfd5fffff] Jan 17 12:18:03.759631 kernel: pci 0000:00:15.0: bridge window [mem 0xc0000000-0xc01fffff 64bit pref] Jan 17 12:18:03.759680 kernel: pci 0000:00:15.1: PCI bridge to [bus 04] Jan 17 12:18:03.759728 kernel: pci 0000:00:15.1: bridge window [io 0x8000-0x8fff] Jan 17 12:18:03.759777 kernel: pci 0000:00:15.1: bridge window [mem 0xfd100000-0xfd1fffff] Jan 17 12:18:03.759829 kernel: pci 0000:00:15.1: bridge window [mem 0xe7800000-0xe78fffff 64bit pref] Jan 17 12:18:03.759877 kernel: pci 0000:00:15.2: PCI bridge to [bus 05] Jan 17 12:18:03.759925 kernel: pci 0000:00:15.2: bridge window [io 0xc000-0xcfff] Jan 17 12:18:03.759974 kernel: pci 0000:00:15.2: bridge window [mem 0xfcd00000-0xfcdfffff] Jan 17 12:18:03.760021 kernel: pci 0000:00:15.2: bridge window [mem 0xe7400000-0xe74fffff 64bit pref] Jan 17 12:18:03.760069 kernel: pci 0000:00:15.3: PCI bridge to [bus 06] Jan 17 12:18:03.761026 kernel: pci 0000:00:15.3: bridge window [mem 0xfc900000-0xfc9fffff] Jan 17 12:18:03.761084 kernel: pci 0000:00:15.3: bridge window [mem 0xe7000000-0xe70fffff 64bit pref] Jan 17 12:18:03.761149 kernel: pci 0000:00:15.4: PCI bridge to [bus 07] Jan 17 12:18:03.761218 kernel: pci 0000:00:15.4: bridge window [mem 0xfc500000-0xfc5fffff] Jan 17 12:18:03.761267 kernel: pci 0000:00:15.4: bridge window [mem 0xe6c00000-0xe6cfffff 64bit pref] Jan 17 12:18:03.761318 kernel: pci 0000:00:15.5: PCI bridge to [bus 08] Jan 17 12:18:03.761366 kernel: pci 0000:00:15.5: bridge window [mem 0xfc100000-0xfc1fffff] Jan 17 12:18:03.761418 kernel: pci 0000:00:15.5: bridge window [mem 0xe6800000-0xe68fffff 64bit pref] Jan 17 12:18:03.761467 kernel: pci 0000:00:15.6: PCI bridge to [bus 09] Jan 17 12:18:03.761518 kernel: pci 0000:00:15.6: bridge window [mem 0xfbd00000-0xfbdfffff] Jan 17 12:18:03.761567 kernel: pci 0000:00:15.6: bridge window [mem 0xe6400000-0xe64fffff 64bit pref] Jan 17 12:18:03.761615 kernel: pci 0000:00:15.7: PCI bridge to [bus 0a] Jan 17 12:18:03.761663 kernel: pci 0000:00:15.7: bridge window [mem 0xfb900000-0xfb9fffff] Jan 17 12:18:03.761711 kernel: pci 0000:00:15.7: bridge window [mem 0xe6000000-0xe60fffff 64bit pref] Jan 17 12:18:03.761763 kernel: pci 0000:0b:00.0: BAR 6: assigned [mem 0xfd400000-0xfd40ffff pref] Jan 17 12:18:03.761811 kernel: pci 0000:00:16.0: PCI bridge to [bus 0b] Jan 17 12:18:03.761859 kernel: pci 0000:00:16.0: bridge window [io 0x5000-0x5fff] Jan 17 12:18:03.761908 kernel: pci 0000:00:16.0: bridge window [mem 0xfd400000-0xfd4fffff] Jan 17 12:18:03.761959 kernel: pci 0000:00:16.0: bridge window [mem 0xc0200000-0xc03fffff 64bit pref] Jan 17 12:18:03.762008 kernel: pci 0000:00:16.1: PCI bridge to [bus 0c] Jan 17 12:18:03.762056 kernel: pci 0000:00:16.1: bridge window [io 0x9000-0x9fff] Jan 17 12:18:03.762114 kernel: pci 0000:00:16.1: bridge window [mem 0xfd000000-0xfd0fffff] Jan 17 12:18:03.762167 kernel: pci 0000:00:16.1: bridge window [mem 0xe7700000-0xe77fffff 64bit pref] Jan 17 12:18:03.762215 kernel: pci 0000:00:16.2: PCI bridge to [bus 0d] Jan 17 12:18:03.762273 kernel: pci 0000:00:16.2: bridge window [io 0xd000-0xdfff] Jan 17 12:18:03.762345 kernel: pci 0000:00:16.2: bridge window [mem 0xfcc00000-0xfccfffff] Jan 17 12:18:03.762396 kernel: pci 0000:00:16.2: bridge window [mem 0xe7300000-0xe73fffff 64bit pref] Jan 17 12:18:03.762444 kernel: pci 0000:00:16.3: PCI bridge to [bus 0e] Jan 17 12:18:03.762496 kernel: pci 0000:00:16.3: bridge window [mem 0xfc800000-0xfc8fffff] Jan 17 12:18:03.762544 kernel: pci 0000:00:16.3: bridge window [mem 0xe6f00000-0xe6ffffff 64bit pref] Jan 17 12:18:03.762592 kernel: pci 0000:00:16.4: PCI bridge to [bus 0f] Jan 17 12:18:03.762641 kernel: pci 0000:00:16.4: bridge window [mem 0xfc400000-0xfc4fffff] Jan 17 12:18:03.762690 kernel: pci 0000:00:16.4: bridge window [mem 0xe6b00000-0xe6bfffff 64bit pref] Jan 17 12:18:03.762737 kernel: pci 0000:00:16.5: PCI bridge to [bus 10] Jan 17 12:18:03.762785 kernel: pci 0000:00:16.5: bridge window [mem 0xfc000000-0xfc0fffff] Jan 17 12:18:03.762834 kernel: pci 0000:00:16.5: bridge window [mem 0xe6700000-0xe67fffff 64bit pref] Jan 17 12:18:03.762882 kernel: pci 0000:00:16.6: PCI bridge to [bus 11] Jan 17 12:18:03.762933 kernel: pci 0000:00:16.6: bridge window [mem 0xfbc00000-0xfbcfffff] Jan 17 12:18:03.762992 kernel: pci 0000:00:16.6: bridge window [mem 0xe6300000-0xe63fffff 64bit pref] Jan 17 12:18:03.763050 kernel: pci 0000:00:16.7: PCI bridge to [bus 12] Jan 17 12:18:03.763149 kernel: pci 0000:00:16.7: bridge window [mem 0xfb800000-0xfb8fffff] Jan 17 12:18:03.763205 kernel: pci 0000:00:16.7: bridge window [mem 0xe5f00000-0xe5ffffff 64bit pref] Jan 17 12:18:03.763254 kernel: pci 0000:00:17.0: PCI bridge to [bus 13] Jan 17 12:18:03.763302 kernel: pci 0000:00:17.0: bridge window [io 0x6000-0x6fff] Jan 17 12:18:03.763350 kernel: pci 0000:00:17.0: bridge window [mem 0xfd300000-0xfd3fffff] Jan 17 12:18:03.763398 kernel: pci 0000:00:17.0: bridge window [mem 0xe7a00000-0xe7afffff 64bit pref] Jan 17 12:18:03.763446 kernel: pci 0000:00:17.1: PCI bridge to [bus 14] Jan 17 12:18:03.763497 kernel: pci 0000:00:17.1: bridge window [io 0xa000-0xafff] Jan 17 12:18:03.763546 kernel: pci 0000:00:17.1: bridge window [mem 0xfcf00000-0xfcffffff] Jan 17 12:18:03.763594 kernel: pci 0000:00:17.1: bridge window [mem 0xe7600000-0xe76fffff 64bit pref] Jan 17 12:18:03.763643 kernel: pci 0000:00:17.2: PCI bridge to [bus 15] Jan 17 12:18:03.763692 kernel: pci 0000:00:17.2: bridge window [io 0xe000-0xefff] Jan 17 12:18:03.763741 kernel: pci 0000:00:17.2: bridge window [mem 0xfcb00000-0xfcbfffff] Jan 17 12:18:03.763789 kernel: pci 0000:00:17.2: bridge window [mem 0xe7200000-0xe72fffff 64bit pref] Jan 17 12:18:03.763849 kernel: pci 0000:00:17.3: PCI bridge to [bus 16] Jan 17 12:18:03.763898 kernel: pci 0000:00:17.3: bridge window [mem 0xfc700000-0xfc7fffff] Jan 17 12:18:03.763948 kernel: pci 0000:00:17.3: bridge window [mem 0xe6e00000-0xe6efffff 64bit pref] Jan 17 12:18:03.763997 kernel: pci 0000:00:17.4: PCI bridge to [bus 17] Jan 17 12:18:03.764047 kernel: pci 0000:00:17.4: bridge window [mem 0xfc300000-0xfc3fffff] Jan 17 12:18:03.764138 kernel: pci 0000:00:17.4: bridge window [mem 0xe6a00000-0xe6afffff 64bit pref] Jan 17 12:18:03.764191 kernel: pci 0000:00:17.5: PCI bridge to [bus 18] Jan 17 12:18:03.764240 kernel: pci 0000:00:17.5: bridge window [mem 0xfbf00000-0xfbffffff] Jan 17 12:18:03.764288 kernel: pci 0000:00:17.5: bridge window [mem 0xe6600000-0xe66fffff 64bit pref] Jan 17 12:18:03.764336 kernel: pci 0000:00:17.6: PCI bridge to [bus 19] Jan 17 12:18:03.764384 kernel: pci 0000:00:17.6: bridge window [mem 0xfbb00000-0xfbbfffff] Jan 17 12:18:03.764437 kernel: pci 0000:00:17.6: bridge window [mem 0xe6200000-0xe62fffff 64bit pref] Jan 17 12:18:03.764488 kernel: pci 0000:00:17.7: PCI bridge to [bus 1a] Jan 17 12:18:03.764535 kernel: pci 0000:00:17.7: bridge window [mem 0xfb700000-0xfb7fffff] Jan 17 12:18:03.764583 kernel: pci 0000:00:17.7: bridge window [mem 0xe5e00000-0xe5efffff 64bit pref] Jan 17 12:18:03.764630 kernel: pci 0000:00:18.0: PCI bridge to [bus 1b] Jan 17 12:18:03.764678 kernel: pci 0000:00:18.0: bridge window [io 0x7000-0x7fff] Jan 17 12:18:03.764726 kernel: pci 0000:00:18.0: bridge window [mem 0xfd200000-0xfd2fffff] Jan 17 12:18:03.764773 kernel: pci 0000:00:18.0: bridge window [mem 0xe7900000-0xe79fffff 64bit pref] Jan 17 12:18:03.764822 kernel: pci 0000:00:18.1: PCI bridge to [bus 1c] Jan 17 12:18:03.764871 kernel: pci 0000:00:18.1: bridge window [io 0xb000-0xbfff] Jan 17 12:18:03.764921 kernel: pci 0000:00:18.1: bridge window [mem 0xfce00000-0xfcefffff] Jan 17 12:18:03.764969 kernel: pci 0000:00:18.1: bridge window [mem 0xe7500000-0xe75fffff 64bit pref] Jan 17 12:18:03.765017 kernel: pci 0000:00:18.2: PCI bridge to [bus 1d] Jan 17 12:18:03.765064 kernel: pci 0000:00:18.2: bridge window [mem 0xfca00000-0xfcafffff] Jan 17 12:18:03.765176 kernel: pci 0000:00:18.2: bridge window [mem 0xe7100000-0xe71fffff 64bit pref] Jan 17 12:18:03.765225 kernel: pci 0000:00:18.3: PCI bridge to [bus 1e] Jan 17 12:18:03.765272 kernel: pci 0000:00:18.3: bridge window [mem 0xfc600000-0xfc6fffff] Jan 17 12:18:03.765320 kernel: pci 0000:00:18.3: bridge window [mem 0xe6d00000-0xe6dfffff 64bit pref] Jan 17 12:18:03.765368 kernel: pci 0000:00:18.4: PCI bridge to [bus 1f] Jan 17 12:18:03.765416 kernel: pci 0000:00:18.4: bridge window [mem 0xfc200000-0xfc2fffff] Jan 17 12:18:03.765473 kernel: pci 0000:00:18.4: bridge window [mem 0xe6900000-0xe69fffff 64bit pref] Jan 17 12:18:03.765526 kernel: pci 0000:00:18.5: PCI bridge to [bus 20] Jan 17 12:18:03.765574 kernel: pci 0000:00:18.5: bridge window [mem 0xfbe00000-0xfbefffff] Jan 17 12:18:03.765621 kernel: pci 0000:00:18.5: bridge window [mem 0xe6500000-0xe65fffff 64bit pref] Jan 17 12:18:03.765668 kernel: pci 0000:00:18.6: PCI bridge to [bus 21] Jan 17 12:18:03.765716 kernel: pci 0000:00:18.6: bridge window [mem 0xfba00000-0xfbafffff] Jan 17 12:18:03.765762 kernel: pci 0000:00:18.6: bridge window [mem 0xe6100000-0xe61fffff 64bit pref] Jan 17 12:18:03.765810 kernel: pci 0000:00:18.7: PCI bridge to [bus 22] Jan 17 12:18:03.765858 kernel: pci 0000:00:18.7: bridge window [mem 0xfb600000-0xfb6fffff] Jan 17 12:18:03.765908 kernel: pci 0000:00:18.7: bridge window [mem 0xe5d00000-0xe5dfffff 64bit pref] Jan 17 12:18:03.765956 kernel: pci_bus 0000:00: resource 4 [mem 0x000a0000-0x000bffff window] Jan 17 12:18:03.766000 kernel: pci_bus 0000:00: resource 5 [mem 0x000cc000-0x000dbfff window] Jan 17 12:18:03.766042 kernel: pci_bus 0000:00: resource 6 [mem 0xc0000000-0xfebfffff window] Jan 17 12:18:03.766084 kernel: pci_bus 0000:00: resource 7 [io 0x0000-0x0cf7 window] Jan 17 12:18:03.767254 kernel: pci_bus 0000:00: resource 8 [io 0x0d00-0xfeff window] Jan 17 12:18:03.767309 kernel: pci_bus 0000:02: resource 0 [io 0x2000-0x3fff] Jan 17 12:18:03.767354 kernel: pci_bus 0000:02: resource 1 [mem 0xfd600000-0xfdffffff] Jan 17 12:18:03.767406 kernel: pci_bus 0000:02: resource 2 [mem 0xe7b00000-0xe7ffffff 64bit pref] Jan 17 12:18:03.767450 kernel: pci_bus 0000:02: resource 4 [mem 0x000a0000-0x000bffff window] Jan 17 12:18:03.767493 kernel: pci_bus 0000:02: resource 5 [mem 0x000cc000-0x000dbfff window] Jan 17 12:18:03.767536 kernel: pci_bus 0000:02: resource 6 [mem 0xc0000000-0xfebfffff window] Jan 17 12:18:03.767579 kernel: pci_bus 0000:02: resource 7 [io 0x0000-0x0cf7 window] Jan 17 12:18:03.767622 kernel: pci_bus 0000:02: resource 8 [io 0x0d00-0xfeff window] Jan 17 12:18:03.767669 kernel: pci_bus 0000:03: resource 0 [io 0x4000-0x4fff] Jan 17 12:18:03.767716 kernel: pci_bus 0000:03: resource 1 [mem 0xfd500000-0xfd5fffff] Jan 17 12:18:03.767760 kernel: pci_bus 0000:03: resource 2 [mem 0xc0000000-0xc01fffff 64bit pref] Jan 17 12:18:03.767807 kernel: pci_bus 0000:04: resource 0 [io 0x8000-0x8fff] Jan 17 12:18:03.767851 kernel: pci_bus 0000:04: resource 1 [mem 0xfd100000-0xfd1fffff] Jan 17 12:18:03.767894 kernel: pci_bus 0000:04: resource 2 [mem 0xe7800000-0xe78fffff 64bit pref] Jan 17 12:18:03.767941 kernel: pci_bus 0000:05: resource 0 [io 0xc000-0xcfff] Jan 17 12:18:03.767985 kernel: pci_bus 0000:05: resource 1 [mem 0xfcd00000-0xfcdfffff] Jan 17 12:18:03.768031 kernel: pci_bus 0000:05: resource 2 [mem 0xe7400000-0xe74fffff 64bit pref] Jan 17 12:18:03.768077 kernel: pci_bus 0000:06: resource 1 [mem 0xfc900000-0xfc9fffff] Jan 17 12:18:03.768134 kernel: pci_bus 0000:06: resource 2 [mem 0xe7000000-0xe70fffff 64bit pref] Jan 17 12:18:03.768197 kernel: pci_bus 0000:07: resource 1 [mem 0xfc500000-0xfc5fffff] Jan 17 12:18:03.768240 kernel: pci_bus 0000:07: resource 2 [mem 0xe6c00000-0xe6cfffff 64bit pref] Jan 17 12:18:03.768290 kernel: pci_bus 0000:08: resource 1 [mem 0xfc100000-0xfc1fffff] Jan 17 12:18:03.768336 kernel: pci_bus 0000:08: resource 2 [mem 0xe6800000-0xe68fffff 64bit pref] Jan 17 12:18:03.768384 kernel: pci_bus 0000:09: resource 1 [mem 0xfbd00000-0xfbdfffff] Jan 17 12:18:03.768457 kernel: pci_bus 0000:09: resource 2 [mem 0xe6400000-0xe64fffff 64bit pref] Jan 17 12:18:03.768507 kernel: pci_bus 0000:0a: resource 1 [mem 0xfb900000-0xfb9fffff] Jan 17 12:18:03.768560 kernel: pci_bus 0000:0a: resource 2 [mem 0xe6000000-0xe60fffff 64bit pref] Jan 17 12:18:03.768610 kernel: pci_bus 0000:0b: resource 0 [io 0x5000-0x5fff] Jan 17 12:18:03.768657 kernel: pci_bus 0000:0b: resource 1 [mem 0xfd400000-0xfd4fffff] Jan 17 12:18:03.768702 kernel: pci_bus 0000:0b: resource 2 [mem 0xc0200000-0xc03fffff 64bit pref] Jan 17 12:18:03.768749 kernel: pci_bus 0000:0c: resource 0 [io 0x9000-0x9fff] Jan 17 12:18:03.768794 kernel: pci_bus 0000:0c: resource 1 [mem 0xfd000000-0xfd0fffff] Jan 17 12:18:03.768839 kernel: pci_bus 0000:0c: resource 2 [mem 0xe7700000-0xe77fffff 64bit pref] Jan 17 12:18:03.768887 kernel: pci_bus 0000:0d: resource 0 [io 0xd000-0xdfff] Jan 17 12:18:03.768935 kernel: pci_bus 0000:0d: resource 1 [mem 0xfcc00000-0xfccfffff] Jan 17 12:18:03.768982 kernel: pci_bus 0000:0d: resource 2 [mem 0xe7300000-0xe73fffff 64bit pref] Jan 17 12:18:03.769030 kernel: pci_bus 0000:0e: resource 1 [mem 0xfc800000-0xfc8fffff] Jan 17 12:18:03.769076 kernel: pci_bus 0000:0e: resource 2 [mem 0xe6f00000-0xe6ffffff 64bit pref] Jan 17 12:18:03.770118 kernel: pci_bus 0000:0f: resource 1 [mem 0xfc400000-0xfc4fffff] Jan 17 12:18:03.770172 kernel: pci_bus 0000:0f: resource 2 [mem 0xe6b00000-0xe6bfffff 64bit pref] Jan 17 12:18:03.770223 kernel: pci_bus 0000:10: resource 1 [mem 0xfc000000-0xfc0fffff] Jan 17 12:18:03.770272 kernel: pci_bus 0000:10: resource 2 [mem 0xe6700000-0xe67fffff 64bit pref] Jan 17 12:18:03.770323 kernel: pci_bus 0000:11: resource 1 [mem 0xfbc00000-0xfbcfffff] Jan 17 12:18:03.770368 kernel: pci_bus 0000:11: resource 2 [mem 0xe6300000-0xe63fffff 64bit pref] Jan 17 12:18:03.770454 kernel: pci_bus 0000:12: resource 1 [mem 0xfb800000-0xfb8fffff] Jan 17 12:18:03.770498 kernel: pci_bus 0000:12: resource 2 [mem 0xe5f00000-0xe5ffffff 64bit pref] Jan 17 12:18:03.770546 kernel: pci_bus 0000:13: resource 0 [io 0x6000-0x6fff] Jan 17 12:18:03.770593 kernel: pci_bus 0000:13: resource 1 [mem 0xfd300000-0xfd3fffff] Jan 17 12:18:03.770637 kernel: pci_bus 0000:13: resource 2 [mem 0xe7a00000-0xe7afffff 64bit pref] Jan 17 12:18:03.770685 kernel: pci_bus 0000:14: resource 0 [io 0xa000-0xafff] Jan 17 12:18:03.770730 kernel: pci_bus 0000:14: resource 1 [mem 0xfcf00000-0xfcffffff] Jan 17 12:18:03.770774 kernel: pci_bus 0000:14: resource 2 [mem 0xe7600000-0xe76fffff 64bit pref] Jan 17 12:18:03.770821 kernel: pci_bus 0000:15: resource 0 [io 0xe000-0xefff] Jan 17 12:18:03.770865 kernel: pci_bus 0000:15: resource 1 [mem 0xfcb00000-0xfcbfffff] Jan 17 12:18:03.770911 kernel: pci_bus 0000:15: resource 2 [mem 0xe7200000-0xe72fffff 64bit pref] Jan 17 12:18:03.770958 kernel: pci_bus 0000:16: resource 1 [mem 0xfc700000-0xfc7fffff] Jan 17 12:18:03.771003 kernel: pci_bus 0000:16: resource 2 [mem 0xe6e00000-0xe6efffff 64bit pref] Jan 17 12:18:03.771051 kernel: pci_bus 0000:17: resource 1 [mem 0xfc300000-0xfc3fffff] Jan 17 12:18:03.771096 kernel: pci_bus 0000:17: resource 2 [mem 0xe6a00000-0xe6afffff 64bit pref] Jan 17 12:18:03.771454 kernel: pci_bus 0000:18: resource 1 [mem 0xfbf00000-0xfbffffff] Jan 17 12:18:03.771506 kernel: pci_bus 0000:18: resource 2 [mem 0xe6600000-0xe66fffff 64bit pref] Jan 17 12:18:03.771560 kernel: pci_bus 0000:19: resource 1 [mem 0xfbb00000-0xfbbfffff] Jan 17 12:18:03.771605 kernel: pci_bus 0000:19: resource 2 [mem 0xe6200000-0xe62fffff 64bit pref] Jan 17 12:18:03.771653 kernel: pci_bus 0000:1a: resource 1 [mem 0xfb700000-0xfb7fffff] Jan 17 12:18:03.771698 kernel: pci_bus 0000:1a: resource 2 [mem 0xe5e00000-0xe5efffff 64bit pref] Jan 17 12:18:03.771750 kernel: pci_bus 0000:1b: resource 0 [io 0x7000-0x7fff] Jan 17 12:18:03.771798 kernel: pci_bus 0000:1b: resource 1 [mem 0xfd200000-0xfd2fffff] Jan 17 12:18:03.771842 kernel: pci_bus 0000:1b: resource 2 [mem 0xe7900000-0xe79fffff 64bit pref] Jan 17 12:18:03.771891 kernel: pci_bus 0000:1c: resource 0 [io 0xb000-0xbfff] Jan 17 12:18:03.771937 kernel: pci_bus 0000:1c: resource 1 [mem 0xfce00000-0xfcefffff] Jan 17 12:18:03.771982 kernel: pci_bus 0000:1c: resource 2 [mem 0xe7500000-0xe75fffff 64bit pref] Jan 17 12:18:03.772030 kernel: pci_bus 0000:1d: resource 1 [mem 0xfca00000-0xfcafffff] Jan 17 12:18:03.772078 kernel: pci_bus 0000:1d: resource 2 [mem 0xe7100000-0xe71fffff 64bit pref] Jan 17 12:18:03.772146 kernel: pci_bus 0000:1e: resource 1 [mem 0xfc600000-0xfc6fffff] Jan 17 12:18:03.772194 kernel: pci_bus 0000:1e: resource 2 [mem 0xe6d00000-0xe6dfffff 64bit pref] Jan 17 12:18:03.772242 kernel: pci_bus 0000:1f: resource 1 [mem 0xfc200000-0xfc2fffff] Jan 17 12:18:03.772287 kernel: pci_bus 0000:1f: resource 2 [mem 0xe6900000-0xe69fffff 64bit pref] Jan 17 12:18:03.772476 kernel: pci_bus 0000:20: resource 1 [mem 0xfbe00000-0xfbefffff] Jan 17 12:18:03.772526 kernel: pci_bus 0000:20: resource 2 [mem 0xe6500000-0xe65fffff 64bit pref] Jan 17 12:18:03.772582 kernel: pci_bus 0000:21: resource 1 [mem 0xfba00000-0xfbafffff] Jan 17 12:18:03.772627 kernel: pci_bus 0000:21: resource 2 [mem 0xe6100000-0xe61fffff 64bit pref] Jan 17 12:18:03.772676 kernel: pci_bus 0000:22: resource 1 [mem 0xfb600000-0xfb6fffff] Jan 17 12:18:03.772721 kernel: pci_bus 0000:22: resource 2 [mem 0xe5d00000-0xe5dfffff 64bit pref] Jan 17 12:18:03.772774 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers Jan 17 12:18:03.772784 kernel: PCI: CLS 32 bytes, default 64 Jan 17 12:18:03.772793 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Jan 17 12:18:03.772800 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x311fd3cd494, max_idle_ns: 440795223879 ns Jan 17 12:18:03.772806 kernel: clocksource: Switched to clocksource tsc Jan 17 12:18:03.772812 kernel: Initialise system trusted keyrings Jan 17 12:18:03.772818 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 Jan 17 12:18:03.772824 kernel: Key type asymmetric registered Jan 17 12:18:03.772830 kernel: Asymmetric key parser 'x509' registered Jan 17 12:18:03.772836 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Jan 17 12:18:03.772842 kernel: io scheduler mq-deadline registered Jan 17 12:18:03.772850 kernel: io scheduler kyber registered Jan 17 12:18:03.772856 kernel: io scheduler bfq registered Jan 17 12:18:03.772907 kernel: pcieport 0000:00:15.0: PME: Signaling with IRQ 24 Jan 17 12:18:03.772958 kernel: pcieport 0000:00:15.0: pciehp: Slot #160 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.773008 kernel: pcieport 0000:00:15.1: PME: Signaling with IRQ 25 Jan 17 12:18:03.773058 kernel: pcieport 0000:00:15.1: pciehp: Slot #161 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.773118 kernel: pcieport 0000:00:15.2: PME: Signaling with IRQ 26 Jan 17 12:18:03.773172 kernel: pcieport 0000:00:15.2: pciehp: Slot #162 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.773225 kernel: pcieport 0000:00:15.3: PME: Signaling with IRQ 27 Jan 17 12:18:03.773275 kernel: pcieport 0000:00:15.3: pciehp: Slot #163 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.773324 kernel: pcieport 0000:00:15.4: PME: Signaling with IRQ 28 Jan 17 12:18:03.773374 kernel: pcieport 0000:00:15.4: pciehp: Slot #164 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.773423 kernel: pcieport 0000:00:15.5: PME: Signaling with IRQ 29 Jan 17 12:18:03.773472 kernel: pcieport 0000:00:15.5: pciehp: Slot #165 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.773523 kernel: pcieport 0000:00:15.6: PME: Signaling with IRQ 30 Jan 17 12:18:03.773573 kernel: pcieport 0000:00:15.6: pciehp: Slot #166 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.773622 kernel: pcieport 0000:00:15.7: PME: Signaling with IRQ 31 Jan 17 12:18:03.773671 kernel: pcieport 0000:00:15.7: pciehp: Slot #167 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.773720 kernel: pcieport 0000:00:16.0: PME: Signaling with IRQ 32 Jan 17 12:18:03.773772 kernel: pcieport 0000:00:16.0: pciehp: Slot #192 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.773822 kernel: pcieport 0000:00:16.1: PME: Signaling with IRQ 33 Jan 17 12:18:03.773880 kernel: pcieport 0000:00:16.1: pciehp: Slot #193 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.773939 kernel: pcieport 0000:00:16.2: PME: Signaling with IRQ 34 Jan 17 12:18:03.773988 kernel: pcieport 0000:00:16.2: pciehp: Slot #194 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.774037 kernel: pcieport 0000:00:16.3: PME: Signaling with IRQ 35 Jan 17 12:18:03.774087 kernel: pcieport 0000:00:16.3: pciehp: Slot #195 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.774215 kernel: pcieport 0000:00:16.4: PME: Signaling with IRQ 36 Jan 17 12:18:03.774266 kernel: pcieport 0000:00:16.4: pciehp: Slot #196 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.774318 kernel: pcieport 0000:00:16.5: PME: Signaling with IRQ 37 Jan 17 12:18:03.774369 kernel: pcieport 0000:00:16.5: pciehp: Slot #197 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.774422 kernel: pcieport 0000:00:16.6: PME: Signaling with IRQ 38 Jan 17 12:18:03.774474 kernel: pcieport 0000:00:16.6: pciehp: Slot #198 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.774522 kernel: pcieport 0000:00:16.7: PME: Signaling with IRQ 39 Jan 17 12:18:03.774571 kernel: pcieport 0000:00:16.7: pciehp: Slot #199 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.774620 kernel: pcieport 0000:00:17.0: PME: Signaling with IRQ 40 Jan 17 12:18:03.774670 kernel: pcieport 0000:00:17.0: pciehp: Slot #224 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.774718 kernel: pcieport 0000:00:17.1: PME: Signaling with IRQ 41 Jan 17 12:18:03.774770 kernel: pcieport 0000:00:17.1: pciehp: Slot #225 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.774819 kernel: pcieport 0000:00:17.2: PME: Signaling with IRQ 42 Jan 17 12:18:03.774868 kernel: pcieport 0000:00:17.2: pciehp: Slot #226 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.774928 kernel: pcieport 0000:00:17.3: PME: Signaling with IRQ 43 Jan 17 12:18:03.774984 kernel: pcieport 0000:00:17.3: pciehp: Slot #227 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.776155 kernel: pcieport 0000:00:17.4: PME: Signaling with IRQ 44 Jan 17 12:18:03.776217 kernel: pcieport 0000:00:17.4: pciehp: Slot #228 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.776269 kernel: pcieport 0000:00:17.5: PME: Signaling with IRQ 45 Jan 17 12:18:03.776320 kernel: pcieport 0000:00:17.5: pciehp: Slot #229 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.776370 kernel: pcieport 0000:00:17.6: PME: Signaling with IRQ 46 Jan 17 12:18:03.776419 kernel: pcieport 0000:00:17.6: pciehp: Slot #230 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.776469 kernel: pcieport 0000:00:17.7: PME: Signaling with IRQ 47 Jan 17 12:18:03.776556 kernel: pcieport 0000:00:17.7: pciehp: Slot #231 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.776604 kernel: pcieport 0000:00:18.0: PME: Signaling with IRQ 48 Jan 17 12:18:03.776652 kernel: pcieport 0000:00:18.0: pciehp: Slot #256 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.776701 kernel: pcieport 0000:00:18.1: PME: Signaling with IRQ 49 Jan 17 12:18:03.776749 kernel: pcieport 0000:00:18.1: pciehp: Slot #257 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.776800 kernel: pcieport 0000:00:18.2: PME: Signaling with IRQ 50 Jan 17 12:18:03.776851 kernel: pcieport 0000:00:18.2: pciehp: Slot #258 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.776900 kernel: pcieport 0000:00:18.3: PME: Signaling with IRQ 51 Jan 17 12:18:03.776950 kernel: pcieport 0000:00:18.3: pciehp: Slot #259 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.777000 kernel: pcieport 0000:00:18.4: PME: Signaling with IRQ 52 Jan 17 12:18:03.777049 kernel: pcieport 0000:00:18.4: pciehp: Slot #260 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.777232 kernel: pcieport 0000:00:18.5: PME: Signaling with IRQ 53 Jan 17 12:18:03.777292 kernel: pcieport 0000:00:18.5: pciehp: Slot #261 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.777342 kernel: pcieport 0000:00:18.6: PME: Signaling with IRQ 54 Jan 17 12:18:03.777397 kernel: pcieport 0000:00:18.6: pciehp: Slot #262 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.777448 kernel: pcieport 0000:00:18.7: PME: Signaling with IRQ 55 Jan 17 12:18:03.777498 kernel: pcieport 0000:00:18.7: pciehp: Slot #263 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jan 17 12:18:03.777509 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Jan 17 12:18:03.777516 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Jan 17 12:18:03.777522 kernel: 00:05: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Jan 17 12:18:03.777530 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBC,PNP0f13:MOUS] at 0x60,0x64 irq 1,12 Jan 17 12:18:03.777536 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Jan 17 12:18:03.777542 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Jan 17 12:18:03.777593 kernel: rtc_cmos 00:01: registered as rtc0 Jan 17 12:18:03.777641 kernel: rtc_cmos 00:01: setting system clock to 2025-01-17T12:18:03 UTC (1737116283) Jan 17 12:18:03.777686 kernel: rtc_cmos 00:01: alarms up to one month, y3k, 114 bytes nvram Jan 17 12:18:03.777695 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input1 Jan 17 12:18:03.777701 kernel: intel_pstate: CPU model not supported Jan 17 12:18:03.777708 kernel: NET: Registered PF_INET6 protocol family Jan 17 12:18:03.777714 kernel: Segment Routing with IPv6 Jan 17 12:18:03.777720 kernel: In-situ OAM (IOAM) with IPv6 Jan 17 12:18:03.777726 kernel: NET: Registered PF_PACKET protocol family Jan 17 12:18:03.777732 kernel: Key type dns_resolver registered Jan 17 12:18:03.777740 kernel: IPI shorthand broadcast: enabled Jan 17 12:18:03.777746 kernel: sched_clock: Marking stable (882003491, 224063293)->(1160789631, -54722847) Jan 17 12:18:03.777752 kernel: registered taskstats version 1 Jan 17 12:18:03.777759 kernel: Loading compiled-in X.509 certificates Jan 17 12:18:03.777765 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.71-flatcar: 6baa290b0089ed5c4c5f7248306af816ac8c7f80' Jan 17 12:18:03.777771 kernel: Key type .fscrypt registered Jan 17 12:18:03.777777 kernel: Key type fscrypt-provisioning registered Jan 17 12:18:03.777783 kernel: ima: No TPM chip found, activating TPM-bypass! Jan 17 12:18:03.777790 kernel: ima: Allocated hash algorithm: sha1 Jan 17 12:18:03.777796 kernel: ima: No architecture policies found Jan 17 12:18:03.777803 kernel: clk: Disabling unused clocks Jan 17 12:18:03.777809 kernel: Freeing unused kernel image (initmem) memory: 42848K Jan 17 12:18:03.777815 kernel: Write protecting the kernel read-only data: 36864k Jan 17 12:18:03.777821 kernel: Freeing unused kernel image (rodata/data gap) memory: 1848K Jan 17 12:18:03.777827 kernel: Run /init as init process Jan 17 12:18:03.777833 kernel: with arguments: Jan 17 12:18:03.777839 kernel: /init Jan 17 12:18:03.777845 kernel: with environment: Jan 17 12:18:03.777852 kernel: HOME=/ Jan 17 12:18:03.777858 kernel: TERM=linux Jan 17 12:18:03.777864 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Jan 17 12:18:03.777872 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Jan 17 12:18:03.777879 systemd[1]: Detected virtualization vmware. Jan 17 12:18:03.777886 systemd[1]: Detected architecture x86-64. Jan 17 12:18:03.777892 systemd[1]: Running in initrd. Jan 17 12:18:03.777898 systemd[1]: No hostname configured, using default hostname. Jan 17 12:18:03.777906 systemd[1]: Hostname set to <localhost>. Jan 17 12:18:03.777912 systemd[1]: Initializing machine ID from random generator. Jan 17 12:18:03.777919 systemd[1]: Queued start job for default target initrd.target. Jan 17 12:18:03.777925 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 17 12:18:03.777931 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 17 12:18:03.777938 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Jan 17 12:18:03.777945 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jan 17 12:18:03.777951 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Jan 17 12:18:03.777959 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Jan 17 12:18:03.777966 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Jan 17 12:18:03.777973 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Jan 17 12:18:03.777980 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 17 12:18:03.777986 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jan 17 12:18:03.777992 systemd[1]: Reached target paths.target - Path Units. Jan 17 12:18:03.778000 systemd[1]: Reached target slices.target - Slice Units. Jan 17 12:18:03.778006 systemd[1]: Reached target swap.target - Swaps. Jan 17 12:18:03.778013 systemd[1]: Reached target timers.target - Timer Units. Jan 17 12:18:03.778019 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Jan 17 12:18:03.778025 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jan 17 12:18:03.778032 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Jan 17 12:18:03.778038 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Jan 17 12:18:03.778045 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jan 17 12:18:03.778051 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jan 17 12:18:03.778059 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jan 17 12:18:03.778065 systemd[1]: Reached target sockets.target - Socket Units. Jan 17 12:18:03.778071 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Jan 17 12:18:03.778078 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jan 17 12:18:03.778084 systemd[1]: Finished network-cleanup.service - Network Cleanup. Jan 17 12:18:03.778090 systemd[1]: Starting systemd-fsck-usr.service... Jan 17 12:18:03.778097 systemd[1]: Starting systemd-journald.service - Journal Service... Jan 17 12:18:03.779126 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jan 17 12:18:03.779134 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 17 12:18:03.779158 systemd-journald[216]: Collecting audit messages is disabled. Jan 17 12:18:03.779174 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Jan 17 12:18:03.779181 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jan 17 12:18:03.779187 systemd[1]: Finished systemd-fsck-usr.service. Jan 17 12:18:03.779196 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Jan 17 12:18:03.779203 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jan 17 12:18:03.779210 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Jan 17 12:18:03.779217 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 17 12:18:03.779224 kernel: Bridge firewalling registered Jan 17 12:18:03.779231 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jan 17 12:18:03.779237 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jan 17 12:18:03.779244 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jan 17 12:18:03.779250 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jan 17 12:18:03.779258 systemd-journald[216]: Journal started Jan 17 12:18:03.779273 systemd-journald[216]: Runtime Journal (/run/log/journal/4812979c72774891b284461cf9c463ae) is 4.8M, max 38.6M, 33.8M free. Jan 17 12:18:03.741714 systemd-modules-load[217]: Inserted module 'overlay' Jan 17 12:18:03.782211 systemd[1]: Started systemd-journald.service - Journal Service. Jan 17 12:18:03.767233 systemd-modules-load[217]: Inserted module 'br_netfilter' Jan 17 12:18:03.782030 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 17 12:18:03.784092 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jan 17 12:18:03.784515 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jan 17 12:18:03.788802 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 17 12:18:03.789640 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Jan 17 12:18:03.789918 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 17 12:18:03.791168 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jan 17 12:18:03.797566 dracut-cmdline[248]: dracut-dracut-053 Jan 17 12:18:03.800548 dracut-cmdline[248]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=vmware flatcar.autologin verity.usrhash=bf1e0d81a0170850ab02d370c1a7c7a3f5983c980b3730f748240a3bda2dbb2e Jan 17 12:18:03.815513 systemd-resolved[250]: Positive Trust Anchors: Jan 17 12:18:03.815524 systemd-resolved[250]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jan 17 12:18:03.815545 systemd-resolved[250]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jan 17 12:18:03.818019 systemd-resolved[250]: Defaulting to hostname 'linux'. Jan 17 12:18:03.818582 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jan 17 12:18:03.818713 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jan 17 12:18:03.845112 kernel: SCSI subsystem initialized Jan 17 12:18:03.850109 kernel: Loading iSCSI transport class v2.0-870. Jan 17 12:18:03.857111 kernel: iscsi: registered transport (tcp) Jan 17 12:18:03.869421 kernel: iscsi: registered transport (qla4xxx) Jan 17 12:18:03.869441 kernel: QLogic iSCSI HBA Driver Jan 17 12:18:03.888756 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Jan 17 12:18:03.892181 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Jan 17 12:18:03.906889 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Jan 17 12:18:03.906913 kernel: device-mapper: uevent: version 1.0.3 Jan 17 12:18:03.906922 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Jan 17 12:18:03.937140 kernel: raid6: avx2x4 gen() 52549 MB/s Jan 17 12:18:03.954142 kernel: raid6: avx2x2 gen() 53087 MB/s Jan 17 12:18:03.971415 kernel: raid6: avx2x1 gen() 44554 MB/s Jan 17 12:18:03.971442 kernel: raid6: using algorithm avx2x2 gen() 53087 MB/s Jan 17 12:18:03.989366 kernel: raid6: .... xor() 31005 MB/s, rmw enabled Jan 17 12:18:03.989394 kernel: raid6: using avx2x2 recovery algorithm Jan 17 12:18:04.003117 kernel: xor: automatically using best checksumming function avx Jan 17 12:18:04.099119 kernel: Btrfs loaded, zoned=no, fsverity=no Jan 17 12:18:04.104515 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Jan 17 12:18:04.109204 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 17 12:18:04.116393 systemd-udevd[433]: Using default interface naming scheme 'v255'. Jan 17 12:18:04.118837 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 17 12:18:04.124189 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Jan 17 12:18:04.130752 dracut-pre-trigger[436]: rd.md=0: removing MD RAID activation Jan 17 12:18:04.145339 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Jan 17 12:18:04.150214 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jan 17 12:18:04.220477 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jan 17 12:18:04.226201 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Jan 17 12:18:04.234791 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Jan 17 12:18:04.235748 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Jan 17 12:18:04.236047 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 17 12:18:04.236301 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jan 17 12:18:04.240192 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Jan 17 12:18:04.247523 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Jan 17 12:18:04.291158 kernel: VMware PVSCSI driver - version 1.0.7.0-k Jan 17 12:18:04.292127 kernel: vmw_pvscsi: using 64bit dma Jan 17 12:18:04.296169 kernel: vmw_pvscsi: max_id: 16 Jan 17 12:18:04.296189 kernel: vmw_pvscsi: setting ring_pages to 8 Jan 17 12:18:04.303253 kernel: VMware vmxnet3 virtual NIC driver - version 1.7.0.0-k-NAPI Jan 17 12:18:04.303271 kernel: vmw_pvscsi: enabling reqCallThreshold Jan 17 12:18:04.303279 kernel: vmw_pvscsi: driver-based request coalescing enabled Jan 17 12:18:04.303287 kernel: vmw_pvscsi: using MSI-X Jan 17 12:18:04.304610 kernel: vmxnet3 0000:0b:00.0: # of Tx queues : 2, # of Rx queues : 2 Jan 17 12:18:04.316042 kernel: scsi host0: VMware PVSCSI storage adapter rev 2, req/cmp/msg rings: 8/8/1 pages, cmd_per_lun=254 Jan 17 12:18:04.316068 kernel: cryptd: max_cpu_qlen set to 1000 Jan 17 12:18:04.316077 kernel: libata version 3.00 loaded. Jan 17 12:18:04.316087 kernel: vmxnet3 0000:0b:00.0 eth0: NIC Link is Up 10000 Mbps Jan 17 12:18:04.316192 kernel: ata_piix 0000:00:07.1: version 2.13 Jan 17 12:18:04.327980 kernel: vmw_pvscsi 0000:03:00.0: VMware PVSCSI rev 2 host #0 Jan 17 12:18:04.328065 kernel: scsi 0:0:0:0: Direct-Access VMware Virtual disk 2.0 PQ: 0 ANSI: 6 Jan 17 12:18:04.328160 kernel: scsi host1: ata_piix Jan 17 12:18:04.328221 kernel: scsi host2: ata_piix Jan 17 12:18:04.328281 kernel: ata1: PATA max UDMA/33 cmd 0x1f0 ctl 0x3f6 bmdma 0x1060 irq 14 Jan 17 12:18:04.328289 kernel: vmxnet3 0000:0b:00.0 ens192: renamed from eth0 Jan 17 12:18:04.328355 kernel: ata2: PATA max UDMA/33 cmd 0x170 ctl 0x376 bmdma 0x1068 irq 15 Jan 17 12:18:04.328364 kernel: AVX2 version of gcm_enc/dec engaged. Jan 17 12:18:04.328371 kernel: AES CTR mode by8 optimization enabled Jan 17 12:18:04.330557 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jan 17 12:18:04.330781 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 17 12:18:04.330940 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jan 17 12:18:04.331029 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 17 12:18:04.331093 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 17 12:18:04.331200 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Jan 17 12:18:04.337433 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 17 12:18:04.348443 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 17 12:18:04.349271 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jan 17 12:18:04.361492 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 17 12:18:04.497120 kernel: ata2.00: ATAPI: VMware Virtual IDE CDROM Drive, 00000001, max UDMA/33 Jan 17 12:18:04.504236 kernel: scsi 2:0:0:0: CD-ROM NECVMWar VMware IDE CDR10 1.00 PQ: 0 ANSI: 5 Jan 17 12:18:04.516412 kernel: sd 0:0:0:0: [sda] 17805312 512-byte logical blocks: (9.12 GB/8.49 GiB) Jan 17 12:18:04.522323 kernel: sd 0:0:0:0: [sda] Write Protect is off Jan 17 12:18:04.522391 kernel: sd 0:0:0:0: [sda] Mode Sense: 31 00 00 00 Jan 17 12:18:04.522452 kernel: sd 0:0:0:0: [sda] Cache data unavailable Jan 17 12:18:04.522510 kernel: sd 0:0:0:0: [sda] Assuming drive cache: write through Jan 17 12:18:04.522567 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 1x/1x writer dvd-ram cd/rw xa/form2 cdda tray Jan 17 12:18:04.532814 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Jan 17 12:18:04.532824 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jan 17 12:18:04.532837 kernel: sd 0:0:0:0: [sda] Attached SCSI disk Jan 17 12:18:04.532902 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Jan 17 12:18:04.692609 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/sda6 scanned by (udev-worker) (483) Jan 17 12:18:04.700799 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Virtual_disk OEM. Jan 17 12:18:04.704027 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Virtual_disk EFI-SYSTEM. Jan 17 12:18:04.707186 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Virtual_disk ROOT. Jan 17 12:18:04.713114 kernel: BTRFS: device fsid e459b8ee-f1f7-4c3d-a087-3f1955f52c85 devid 1 transid 36 /dev/sda3 scanned by (udev-worker) (481) Jan 17 12:18:04.718819 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Virtual_disk USR-A. Jan 17 12:18:04.719118 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Virtual_disk USR-A. Jan 17 12:18:04.727281 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Jan 17 12:18:04.757122 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jan 17 12:18:04.762113 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jan 17 12:18:05.770155 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jan 17 12:18:05.771074 disk-uuid[588]: The operation has completed successfully. Jan 17 12:18:05.806248 systemd[1]: disk-uuid.service: Deactivated successfully. Jan 17 12:18:05.806307 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Jan 17 12:18:05.811367 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Jan 17 12:18:05.812997 sh[605]: Success Jan 17 12:18:05.821127 kernel: device-mapper: verity: sha256 using implementation "sha256-avx2" Jan 17 12:18:05.866185 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Jan 17 12:18:05.873530 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Jan 17 12:18:05.873995 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Jan 17 12:18:05.891974 kernel: BTRFS info (device dm-0): first mount of filesystem e459b8ee-f1f7-4c3d-a087-3f1955f52c85 Jan 17 12:18:05.891995 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Jan 17 12:18:05.892004 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Jan 17 12:18:05.892011 kernel: BTRFS info (device dm-0): disabling log replay at mount time Jan 17 12:18:05.892018 kernel: BTRFS info (device dm-0): using free space tree Jan 17 12:18:05.898113 kernel: BTRFS info (device dm-0): enabling ssd optimizations Jan 17 12:18:05.898888 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Jan 17 12:18:05.906187 systemd[1]: Starting afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments... Jan 17 12:18:05.907174 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Jan 17 12:18:05.923506 kernel: BTRFS info (device sda6): first mount of filesystem a70a40d6-5ab2-4665-81b1-b8e9f58c5ff8 Jan 17 12:18:05.923534 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Jan 17 12:18:05.923542 kernel: BTRFS info (device sda6): using free space tree Jan 17 12:18:05.928181 kernel: BTRFS info (device sda6): enabling ssd optimizations Jan 17 12:18:05.934701 systemd[1]: mnt-oem.mount: Deactivated successfully. Jan 17 12:18:05.936426 kernel: BTRFS info (device sda6): last unmount of filesystem a70a40d6-5ab2-4665-81b1-b8e9f58c5ff8 Jan 17 12:18:05.942847 systemd[1]: Finished ignition-setup.service - Ignition (setup). Jan 17 12:18:05.948175 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Jan 17 12:18:05.970317 systemd[1]: Finished afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments. Jan 17 12:18:05.979320 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Jan 17 12:18:06.017350 ignition[664]: Ignition 2.19.0 Jan 17 12:18:06.017358 ignition[664]: Stage: fetch-offline Jan 17 12:18:06.017383 ignition[664]: no configs at "/usr/lib/ignition/base.d" Jan 17 12:18:06.017390 ignition[664]: no config dir at "/usr/lib/ignition/base.platform.d/vmware" Jan 17 12:18:06.017443 ignition[664]: parsed url from cmdline: "" Jan 17 12:18:06.017445 ignition[664]: no config URL provided Jan 17 12:18:06.017448 ignition[664]: reading system config file "/usr/lib/ignition/user.ign" Jan 17 12:18:06.017453 ignition[664]: no config at "/usr/lib/ignition/user.ign" Jan 17 12:18:06.017833 ignition[664]: config successfully fetched Jan 17 12:18:06.017850 ignition[664]: parsing config with SHA512: 2b2b8494d6bfd05329bb7d196c73328bab11e189c274c6ed5a8b4488d854439289a668dac01d59a5ee639bf6f8486a3245506eb1d49630a99d3fdd97bea97388 Jan 17 12:18:06.020578 unknown[664]: fetched base config from "system" Jan 17 12:18:06.020584 unknown[664]: fetched user config from "vmware" Jan 17 12:18:06.022241 ignition[664]: fetch-offline: fetch-offline passed Jan 17 12:18:06.022296 ignition[664]: Ignition finished successfully Jan 17 12:18:06.023023 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Jan 17 12:18:06.045452 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jan 17 12:18:06.050332 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jan 17 12:18:06.061287 systemd-networkd[800]: lo: Link UP Jan 17 12:18:06.061292 systemd-networkd[800]: lo: Gained carrier Jan 17 12:18:06.062133 systemd-networkd[800]: Enumeration completed Jan 17 12:18:06.062262 systemd[1]: Started systemd-networkd.service - Network Configuration. Jan 17 12:18:06.062383 systemd[1]: Reached target network.target - Network. Jan 17 12:18:06.062406 systemd-networkd[800]: ens192: Configuring with /etc/systemd/network/10-dracut-cmdline-99.network. Jan 17 12:18:06.062461 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Jan 17 12:18:06.065947 kernel: vmxnet3 0000:0b:00.0 ens192: intr type 3, mode 0, 3 vectors allocated Jan 17 12:18:06.066059 kernel: vmxnet3 0000:0b:00.0 ens192: NIC Link is Up 10000 Mbps Jan 17 12:18:06.065819 systemd-networkd[800]: ens192: Link UP Jan 17 12:18:06.065821 systemd-networkd[800]: ens192: Gained carrier Jan 17 12:18:06.072147 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Jan 17 12:18:06.078749 ignition[802]: Ignition 2.19.0 Jan 17 12:18:06.078754 ignition[802]: Stage: kargs Jan 17 12:18:06.078859 ignition[802]: no configs at "/usr/lib/ignition/base.d" Jan 17 12:18:06.078865 ignition[802]: no config dir at "/usr/lib/ignition/base.platform.d/vmware" Jan 17 12:18:06.079373 ignition[802]: kargs: kargs passed Jan 17 12:18:06.079400 ignition[802]: Ignition finished successfully Jan 17 12:18:06.080612 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Jan 17 12:18:06.086313 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Jan 17 12:18:06.092991 ignition[809]: Ignition 2.19.0 Jan 17 12:18:06.092999 ignition[809]: Stage: disks Jan 17 12:18:06.093093 ignition[809]: no configs at "/usr/lib/ignition/base.d" Jan 17 12:18:06.093107 ignition[809]: no config dir at "/usr/lib/ignition/base.platform.d/vmware" Jan 17 12:18:06.093590 ignition[809]: disks: disks passed Jan 17 12:18:06.093616 ignition[809]: Ignition finished successfully Jan 17 12:18:06.094207 systemd[1]: Finished ignition-disks.service - Ignition (disks). Jan 17 12:18:06.094583 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Jan 17 12:18:06.094704 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Jan 17 12:18:06.094886 systemd[1]: Reached target local-fs.target - Local File Systems. Jan 17 12:18:06.095065 systemd[1]: Reached target sysinit.target - System Initialization. Jan 17 12:18:06.095266 systemd[1]: Reached target basic.target - Basic System. Jan 17 12:18:06.098183 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Jan 17 12:18:06.140989 systemd-fsck[817]: ROOT: clean, 14/1628000 files, 120691/1617920 blocks Jan 17 12:18:06.141978 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Jan 17 12:18:06.148117 systemd[1]: Mounting sysroot.mount - /sysroot... Jan 17 12:18:06.208810 systemd[1]: Mounted sysroot.mount - /sysroot. Jan 17 12:18:06.209111 kernel: EXT4-fs (sda9): mounted filesystem 0ba4fe0e-76d7-406f-b570-4642d86198f6 r/w with ordered data mode. Quota mode: none. Jan 17 12:18:06.209324 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Jan 17 12:18:06.214278 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jan 17 12:18:06.215147 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Jan 17 12:18:06.215936 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Jan 17 12:18:06.215961 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Jan 17 12:18:06.215974 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Jan 17 12:18:06.219157 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Jan 17 12:18:06.219671 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Jan 17 12:18:06.223108 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/sda6 scanned by mount (825) Jan 17 12:18:06.225280 kernel: BTRFS info (device sda6): first mount of filesystem a70a40d6-5ab2-4665-81b1-b8e9f58c5ff8 Jan 17 12:18:06.225298 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Jan 17 12:18:06.225309 kernel: BTRFS info (device sda6): using free space tree Jan 17 12:18:06.230111 kernel: BTRFS info (device sda6): enabling ssd optimizations Jan 17 12:18:06.230335 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jan 17 12:18:06.250387 initrd-setup-root[849]: cut: /sysroot/etc/passwd: No such file or directory Jan 17 12:18:06.252699 initrd-setup-root[856]: cut: /sysroot/etc/group: No such file or directory Jan 17 12:18:06.254802 initrd-setup-root[863]: cut: /sysroot/etc/shadow: No such file or directory Jan 17 12:18:06.257023 initrd-setup-root[870]: cut: /sysroot/etc/gshadow: No such file or directory Jan 17 12:18:06.306431 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Jan 17 12:18:06.311246 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Jan 17 12:18:06.313587 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Jan 17 12:18:06.318178 kernel: BTRFS info (device sda6): last unmount of filesystem a70a40d6-5ab2-4665-81b1-b8e9f58c5ff8 Jan 17 12:18:06.345767 ignition[938]: INFO : Ignition 2.19.0 Jan 17 12:18:06.345767 ignition[938]: INFO : Stage: mount Jan 17 12:18:06.345767 ignition[938]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 17 12:18:06.345767 ignition[938]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/vmware" Jan 17 12:18:06.345767 ignition[938]: INFO : mount: mount passed Jan 17 12:18:06.345767 ignition[938]: INFO : Ignition finished successfully Jan 17 12:18:06.346545 systemd[1]: Finished ignition-mount.service - Ignition (mount). Jan 17 12:18:06.350312 systemd[1]: Starting ignition-files.service - Ignition (files)... Jan 17 12:18:06.350565 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Jan 17 12:18:06.887776 systemd[1]: sysroot-oem.mount: Deactivated successfully. Jan 17 12:18:06.894324 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jan 17 12:18:06.902116 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/sda6 scanned by mount (951) Jan 17 12:18:06.904965 kernel: BTRFS info (device sda6): first mount of filesystem a70a40d6-5ab2-4665-81b1-b8e9f58c5ff8 Jan 17 12:18:06.904985 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Jan 17 12:18:06.904996 kernel: BTRFS info (device sda6): using free space tree Jan 17 12:18:06.909111 kernel: BTRFS info (device sda6): enabling ssd optimizations Jan 17 12:18:06.909321 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jan 17 12:18:06.927986 ignition[968]: INFO : Ignition 2.19.0 Jan 17 12:18:06.927986 ignition[968]: INFO : Stage: files Jan 17 12:18:06.927986 ignition[968]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 17 12:18:06.927986 ignition[968]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/vmware" Jan 17 12:18:06.927986 ignition[968]: DEBUG : files: compiled without relabeling support, skipping Jan 17 12:18:06.928837 ignition[968]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Jan 17 12:18:06.928980 ignition[968]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Jan 17 12:18:06.931094 ignition[968]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Jan 17 12:18:06.931318 ignition[968]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Jan 17 12:18:06.931593 unknown[968]: wrote ssh authorized keys file for user: core Jan 17 12:18:06.931779 ignition[968]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Jan 17 12:18:06.933013 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Jan 17 12:18:06.933355 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-amd64.tar.gz: attempt #1 Jan 17 12:18:06.978452 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Jan 17 12:18:07.082782 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Jan 17 12:18:07.083555 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Jan 17 12:18:07.083555 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Jan 17 12:18:07.083555 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Jan 17 12:18:07.083555 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Jan 17 12:18:07.083555 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Jan 17 12:18:07.083555 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Jan 17 12:18:07.083555 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Jan 17 12:18:07.083555 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Jan 17 12:18:07.083555 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Jan 17 12:18:07.085157 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Jan 17 12:18:07.085157 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.0-x86-64.raw" Jan 17 12:18:07.085157 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.0-x86-64.raw" Jan 17 12:18:07.085157 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.0-x86-64.raw" Jan 17 12:18:07.085157 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.31.0-x86-64.raw: attempt #1 Jan 17 12:18:07.599089 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Jan 17 12:18:07.782636 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.0-x86-64.raw" Jan 17 12:18:07.782636 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/etc/systemd/network/00-vmware.network" Jan 17 12:18:07.783134 ignition[968]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/etc/systemd/network/00-vmware.network" Jan 17 12:18:07.783134 ignition[968]: INFO : files: op(c): [started] processing unit "prepare-helm.service" Jan 17 12:18:07.786931 ignition[968]: INFO : files: op(c): op(d): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jan 17 12:18:07.787119 ignition[968]: INFO : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jan 17 12:18:07.787119 ignition[968]: INFO : files: op(c): [finished] processing unit "prepare-helm.service" Jan 17 12:18:07.787119 ignition[968]: INFO : files: op(e): [started] processing unit "coreos-metadata.service" Jan 17 12:18:07.787119 ignition[968]: INFO : files: op(e): op(f): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Jan 17 12:18:07.787119 ignition[968]: INFO : files: op(e): op(f): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Jan 17 12:18:07.787119 ignition[968]: INFO : files: op(e): [finished] processing unit "coreos-metadata.service" Jan 17 12:18:07.787119 ignition[968]: INFO : files: op(10): [started] setting preset to disabled for "coreos-metadata.service" Jan 17 12:18:08.130490 systemd-networkd[800]: ens192: Gained IPv6LL Jan 17 12:18:08.196996 ignition[968]: INFO : files: op(10): op(11): [started] removing enablement symlink(s) for "coreos-metadata.service" Jan 17 12:18:08.201122 ignition[968]: INFO : files: op(10): op(11): [finished] removing enablement symlink(s) for "coreos-metadata.service" Jan 17 12:18:08.201122 ignition[968]: INFO : files: op(10): [finished] setting preset to disabled for "coreos-metadata.service" Jan 17 12:18:08.201122 ignition[968]: INFO : files: op(12): [started] setting preset to enabled for "prepare-helm.service" Jan 17 12:18:08.201122 ignition[968]: INFO : files: op(12): [finished] setting preset to enabled for "prepare-helm.service" Jan 17 12:18:08.201905 ignition[968]: INFO : files: createResultFile: createFiles: op(13): [started] writing file "/sysroot/etc/.ignition-result.json" Jan 17 12:18:08.201905 ignition[968]: INFO : files: createResultFile: createFiles: op(13): [finished] writing file "/sysroot/etc/.ignition-result.json" Jan 17 12:18:08.201905 ignition[968]: INFO : files: files passed Jan 17 12:18:08.201905 ignition[968]: INFO : Ignition finished successfully Jan 17 12:18:08.202120 systemd[1]: Finished ignition-files.service - Ignition (files). Jan 17 12:18:08.212325 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Jan 17 12:18:08.213585 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Jan 17 12:18:08.215033 systemd[1]: ignition-quench.service: Deactivated successfully. Jan 17 12:18:08.215084 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Jan 17 12:18:08.220512 initrd-setup-root-after-ignition[998]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jan 17 12:18:08.220512 initrd-setup-root-after-ignition[998]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Jan 17 12:18:08.220933 initrd-setup-root-after-ignition[1002]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jan 17 12:18:08.221920 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Jan 17 12:18:08.222399 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Jan 17 12:18:08.226286 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Jan 17 12:18:08.237566 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Jan 17 12:18:08.237615 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Jan 17 12:18:08.237914 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Jan 17 12:18:08.238018 systemd[1]: Reached target initrd.target - Initrd Default Target. Jan 17 12:18:08.238146 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Jan 17 12:18:08.240177 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Jan 17 12:18:08.247182 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jan 17 12:18:08.251205 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Jan 17 12:18:08.256079 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Jan 17 12:18:08.256358 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 17 12:18:08.256517 systemd[1]: Stopped target timers.target - Timer Units. Jan 17 12:18:08.256651 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Jan 17 12:18:08.256717 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jan 17 12:18:08.256931 systemd[1]: Stopped target initrd.target - Initrd Default Target. Jan 17 12:18:08.257146 systemd[1]: Stopped target basic.target - Basic System. Jan 17 12:18:08.257311 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Jan 17 12:18:08.257504 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Jan 17 12:18:08.257701 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Jan 17 12:18:08.257888 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Jan 17 12:18:08.258059 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Jan 17 12:18:08.258463 systemd[1]: Stopped target sysinit.target - System Initialization. Jan 17 12:18:08.258637 systemd[1]: Stopped target local-fs.target - Local File Systems. Jan 17 12:18:08.258804 systemd[1]: Stopped target swap.target - Swaps. Jan 17 12:18:08.258961 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Jan 17 12:18:08.259018 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Jan 17 12:18:08.259306 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Jan 17 12:18:08.259540 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 17 12:18:08.259685 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Jan 17 12:18:08.259726 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 17 12:18:08.259870 systemd[1]: dracut-initqueue.service: Deactivated successfully. Jan 17 12:18:08.259927 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Jan 17 12:18:08.260177 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Jan 17 12:18:08.260237 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Jan 17 12:18:08.260500 systemd[1]: Stopped target paths.target - Path Units. Jan 17 12:18:08.260623 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Jan 17 12:18:08.263120 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 17 12:18:08.263283 systemd[1]: Stopped target slices.target - Slice Units. Jan 17 12:18:08.263497 systemd[1]: Stopped target sockets.target - Socket Units. Jan 17 12:18:08.263669 systemd[1]: iscsid.socket: Deactivated successfully. Jan 17 12:18:08.263731 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Jan 17 12:18:08.263918 systemd[1]: iscsiuio.socket: Deactivated successfully. Jan 17 12:18:08.263961 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jan 17 12:18:08.264206 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Jan 17 12:18:08.264263 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Jan 17 12:18:08.264496 systemd[1]: ignition-files.service: Deactivated successfully. Jan 17 12:18:08.264548 systemd[1]: Stopped ignition-files.service - Ignition (files). Jan 17 12:18:08.269187 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Jan 17 12:18:08.271218 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Jan 17 12:18:08.271328 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Jan 17 12:18:08.271416 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Jan 17 12:18:08.271667 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Jan 17 12:18:08.271742 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Jan 17 12:18:08.274044 systemd[1]: initrd-cleanup.service: Deactivated successfully. Jan 17 12:18:08.274131 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Jan 17 12:18:08.278107 ignition[1022]: INFO : Ignition 2.19.0 Jan 17 12:18:08.278107 ignition[1022]: INFO : Stage: umount Jan 17 12:18:08.278107 ignition[1022]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 17 12:18:08.278107 ignition[1022]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/vmware" Jan 17 12:18:08.279362 ignition[1022]: INFO : umount: umount passed Jan 17 12:18:08.279362 ignition[1022]: INFO : Ignition finished successfully Jan 17 12:18:08.279997 systemd[1]: ignition-mount.service: Deactivated successfully. Jan 17 12:18:08.280198 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Jan 17 12:18:08.280664 systemd[1]: Stopped target network.target - Network. Jan 17 12:18:08.280880 systemd[1]: ignition-disks.service: Deactivated successfully. Jan 17 12:18:08.281029 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Jan 17 12:18:08.281292 systemd[1]: ignition-kargs.service: Deactivated successfully. Jan 17 12:18:08.281403 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Jan 17 12:18:08.281612 systemd[1]: ignition-setup.service: Deactivated successfully. Jan 17 12:18:08.281652 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Jan 17 12:18:08.281864 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Jan 17 12:18:08.281884 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Jan 17 12:18:08.282207 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Jan 17 12:18:08.282592 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Jan 17 12:18:08.283513 systemd[1]: sysroot-boot.mount: Deactivated successfully. Jan 17 12:18:08.288212 systemd[1]: systemd-resolved.service: Deactivated successfully. Jan 17 12:18:08.288272 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Jan 17 12:18:08.288830 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Jan 17 12:18:08.288862 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 17 12:18:08.290052 systemd[1]: systemd-networkd.service: Deactivated successfully. Jan 17 12:18:08.290118 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Jan 17 12:18:08.290419 systemd[1]: systemd-networkd.socket: Deactivated successfully. Jan 17 12:18:08.290456 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Jan 17 12:18:08.293196 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Jan 17 12:18:08.293284 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Jan 17 12:18:08.293309 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jan 17 12:18:08.293422 systemd[1]: afterburn-network-kargs.service: Deactivated successfully. Jan 17 12:18:08.293442 systemd[1]: Stopped afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments. Jan 17 12:18:08.293567 systemd[1]: systemd-sysctl.service: Deactivated successfully. Jan 17 12:18:08.293602 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Jan 17 12:18:08.294531 systemd[1]: systemd-modules-load.service: Deactivated successfully. Jan 17 12:18:08.294554 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Jan 17 12:18:08.295794 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 17 12:18:08.300258 systemd[1]: network-cleanup.service: Deactivated successfully. Jan 17 12:18:08.300321 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Jan 17 12:18:08.300675 systemd[1]: systemd-udevd.service: Deactivated successfully. Jan 17 12:18:08.300741 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 17 12:18:08.301281 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Jan 17 12:18:08.301308 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Jan 17 12:18:08.301511 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Jan 17 12:18:08.301528 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Jan 17 12:18:08.301675 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Jan 17 12:18:08.301696 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Jan 17 12:18:08.301952 systemd[1]: dracut-cmdline.service: Deactivated successfully. Jan 17 12:18:08.301973 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Jan 17 12:18:08.302244 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jan 17 12:18:08.302264 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 17 12:18:08.306390 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Jan 17 12:18:08.306498 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Jan 17 12:18:08.306525 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 17 12:18:08.306648 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Jan 17 12:18:08.306669 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jan 17 12:18:08.306790 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Jan 17 12:18:08.306811 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Jan 17 12:18:08.306930 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 17 12:18:08.306949 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 17 12:18:08.309117 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Jan 17 12:18:08.309184 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Jan 17 12:18:08.386151 systemd[1]: sysroot-boot.service: Deactivated successfully. Jan 17 12:18:08.386232 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Jan 17 12:18:08.386718 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Jan 17 12:18:08.386888 systemd[1]: initrd-setup-root.service: Deactivated successfully. Jan 17 12:18:08.386926 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Jan 17 12:18:08.389204 systemd[1]: Starting initrd-switch-root.service - Switch Root... Jan 17 12:18:08.402132 systemd[1]: Switching root. Jan 17 12:18:08.430620 systemd-journald[216]: Journal stopped Jan 17 12:18:09.448681 systemd-journald[216]: Received SIGTERM from PID 1 (systemd). Jan 17 12:18:09.448704 kernel: SELinux: policy capability network_peer_controls=1 Jan 17 12:18:09.448712 kernel: SELinux: policy capability open_perms=1 Jan 17 12:18:09.448718 kernel: SELinux: policy capability extended_socket_class=1 Jan 17 12:18:09.448723 kernel: SELinux: policy capability always_check_network=0 Jan 17 12:18:09.448729 kernel: SELinux: policy capability cgroup_seclabel=1 Jan 17 12:18:09.448736 kernel: SELinux: policy capability nnp_nosuid_transition=1 Jan 17 12:18:09.448742 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Jan 17 12:18:09.448748 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Jan 17 12:18:09.448754 kernel: audit: type=1403 audit(1737116288.995:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Jan 17 12:18:09.448760 systemd[1]: Successfully loaded SELinux policy in 34.682ms. Jan 17 12:18:09.448767 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 6.773ms. Jan 17 12:18:09.448774 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Jan 17 12:18:09.448782 systemd[1]: Detected virtualization vmware. Jan 17 12:18:09.448789 systemd[1]: Detected architecture x86-64. Jan 17 12:18:09.448796 systemd[1]: Detected first boot. Jan 17 12:18:09.448803 systemd[1]: Initializing machine ID from random generator. Jan 17 12:18:09.448811 zram_generator::config[1064]: No configuration found. Jan 17 12:18:09.448819 systemd[1]: Populated /etc with preset unit settings. Jan 17 12:18:09.448826 systemd[1]: /etc/systemd/system/coreos-metadata.service:11: Ignoring unknown escape sequences: "echo "COREOS_CUSTOM_PRIVATE_IPV4=$(ip addr show ens192 | grep "inet 10." | grep -Po "inet \K[\d.]+") Jan 17 12:18:09.448833 systemd[1]: COREOS_CUSTOM_PUBLIC_IPV4=$(ip addr show ens192 | grep -v "inet 10." | grep -Po "inet \K[\d.]+")" > ${OUTPUT}" Jan 17 12:18:09.448840 systemd[1]: initrd-switch-root.service: Deactivated successfully. Jan 17 12:18:09.448846 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Jan 17 12:18:09.448853 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Jan 17 12:18:09.448861 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Jan 17 12:18:09.448868 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Jan 17 12:18:09.448875 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Jan 17 12:18:09.448881 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Jan 17 12:18:09.448888 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Jan 17 12:18:09.448895 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Jan 17 12:18:09.448902 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Jan 17 12:18:09.448910 systemd[1]: Created slice user.slice - User and Session Slice. Jan 17 12:18:09.448917 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 17 12:18:09.448924 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 17 12:18:09.448930 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Jan 17 12:18:09.448937 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Jan 17 12:18:09.448944 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Jan 17 12:18:09.448951 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jan 17 12:18:09.448958 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Jan 17 12:18:09.448966 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 17 12:18:09.448973 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Jan 17 12:18:09.448981 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Jan 17 12:18:09.448988 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Jan 17 12:18:09.448995 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Jan 17 12:18:09.449002 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 17 12:18:09.449009 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jan 17 12:18:09.449016 systemd[1]: Reached target slices.target - Slice Units. Jan 17 12:18:09.449025 systemd[1]: Reached target swap.target - Swaps. Jan 17 12:18:09.449032 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Jan 17 12:18:09.449038 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Jan 17 12:18:09.449046 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jan 17 12:18:09.449053 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jan 17 12:18:09.449061 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jan 17 12:18:09.449068 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Jan 17 12:18:09.449075 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Jan 17 12:18:09.449082 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Jan 17 12:18:09.449090 systemd[1]: Mounting media.mount - External Media Directory... Jan 17 12:18:09.449097 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 17 12:18:09.450602 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Jan 17 12:18:09.450612 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Jan 17 12:18:09.450622 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Jan 17 12:18:09.450630 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Jan 17 12:18:09.450637 systemd[1]: Reached target machines.target - Containers. Jan 17 12:18:09.450643 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Jan 17 12:18:09.450650 systemd[1]: Starting ignition-delete-config.service - Ignition (delete config)... Jan 17 12:18:09.450657 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jan 17 12:18:09.450664 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Jan 17 12:18:09.450670 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 17 12:18:09.450678 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jan 17 12:18:09.450685 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 17 12:18:09.450692 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Jan 17 12:18:09.450699 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 17 12:18:09.450705 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Jan 17 12:18:09.450712 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Jan 17 12:18:09.450719 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Jan 17 12:18:09.450726 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Jan 17 12:18:09.450733 systemd[1]: Stopped systemd-fsck-usr.service. Jan 17 12:18:09.450741 systemd[1]: Starting systemd-journald.service - Journal Service... Jan 17 12:18:09.450748 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jan 17 12:18:09.450755 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Jan 17 12:18:09.450761 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Jan 17 12:18:09.450768 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jan 17 12:18:09.450775 systemd[1]: verity-setup.service: Deactivated successfully. Jan 17 12:18:09.450782 systemd[1]: Stopped verity-setup.service. Jan 17 12:18:09.450789 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 17 12:18:09.450808 systemd-journald[1154]: Collecting audit messages is disabled. Jan 17 12:18:09.450823 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Jan 17 12:18:09.450846 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Jan 17 12:18:09.450853 systemd[1]: Mounted media.mount - External Media Directory. Jan 17 12:18:09.450861 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Jan 17 12:18:09.450869 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Jan 17 12:18:09.450876 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Jan 17 12:18:09.450883 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jan 17 12:18:09.450890 systemd[1]: modprobe@configfs.service: Deactivated successfully. Jan 17 12:18:09.450897 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Jan 17 12:18:09.450903 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 17 12:18:09.450910 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 17 12:18:09.450917 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 17 12:18:09.450925 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 17 12:18:09.450933 systemd-journald[1154]: Journal started Jan 17 12:18:09.450947 systemd-journald[1154]: Runtime Journal (/run/log/journal/e217d58a8d7344ad96ea552369ca3ed6) is 4.8M, max 38.6M, 33.8M free. Jan 17 12:18:09.291843 systemd[1]: Queued start job for default target multi-user.target. Jan 17 12:18:09.452387 systemd[1]: Started systemd-journald.service - Journal Service. Jan 17 12:18:09.307400 systemd[1]: Unnecessary job was removed for dev-sda6.device - /dev/sda6. Jan 17 12:18:09.307623 systemd[1]: systemd-journald.service: Deactivated successfully. Jan 17 12:18:09.452729 jq[1131]: true Jan 17 12:18:09.455604 kernel: fuse: init (API version 7.39) Jan 17 12:18:09.459280 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jan 17 12:18:09.459537 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Jan 17 12:18:09.460272 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Jan 17 12:18:09.461186 jq[1166]: true Jan 17 12:18:09.461862 systemd[1]: modprobe@fuse.service: Deactivated successfully. Jan 17 12:18:09.461939 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Jan 17 12:18:09.472385 kernel: loop: module loaded Jan 17 12:18:09.472357 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 17 12:18:09.472479 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 17 12:18:09.473611 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Jan 17 12:18:09.479566 systemd[1]: Reached target network-pre.target - Preparation for Network. Jan 17 12:18:09.493477 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Jan 17 12:18:09.498169 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Jan 17 12:18:09.498302 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Jan 17 12:18:09.498322 systemd[1]: Reached target local-fs.target - Local File Systems. Jan 17 12:18:09.498980 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Jan 17 12:18:09.506187 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Jan 17 12:18:09.508776 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Jan 17 12:18:09.509175 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 17 12:18:09.512137 kernel: ACPI: bus type drm_connector registered Jan 17 12:18:09.512417 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Jan 17 12:18:09.515810 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Jan 17 12:18:09.515988 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jan 17 12:18:09.522189 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Jan 17 12:18:09.522334 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jan 17 12:18:09.524185 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jan 17 12:18:09.526192 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Jan 17 12:18:09.529203 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Jan 17 12:18:09.530237 systemd[1]: modprobe@drm.service: Deactivated successfully. Jan 17 12:18:09.530331 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jan 17 12:18:09.530588 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Jan 17 12:18:09.530729 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Jan 17 12:18:09.531426 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Jan 17 12:18:09.548223 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Jan 17 12:18:09.548412 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Jan 17 12:18:09.552966 systemd-journald[1154]: Time spent on flushing to /var/log/journal/e217d58a8d7344ad96ea552369ca3ed6 is 75.569ms for 1837 entries. Jan 17 12:18:09.552966 systemd-journald[1154]: System Journal (/var/log/journal/e217d58a8d7344ad96ea552369ca3ed6) is 8.0M, max 584.8M, 576.8M free. Jan 17 12:18:09.650655 systemd-journald[1154]: Received client request to flush runtime journal. Jan 17 12:18:09.650687 kernel: loop0: detected capacity change from 0 to 2976 Jan 17 12:18:09.650698 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Jan 17 12:18:09.553655 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Jan 17 12:18:09.617279 ignition[1174]: Ignition 2.19.0 Jan 17 12:18:09.611570 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jan 17 12:18:09.617484 ignition[1174]: deleting config from guestinfo properties Jan 17 12:18:09.614640 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Jan 17 12:18:09.627383 ignition[1174]: Successfully deleted config Jan 17 12:18:09.615302 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Jan 17 12:18:09.615588 systemd-tmpfiles[1202]: ACLs are not supported, ignoring. Jan 17 12:18:09.615688 systemd-tmpfiles[1202]: ACLs are not supported, ignoring. Jan 17 12:18:09.622187 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jan 17 12:18:09.628829 systemd[1]: Starting systemd-sysusers.service - Create System Users... Jan 17 12:18:09.629097 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jan 17 12:18:09.632281 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Jan 17 12:18:09.647216 systemd[1]: Finished ignition-delete-config.service - Ignition (delete config). Jan 17 12:18:09.652312 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Jan 17 12:18:09.654980 udevadm[1226]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Jan 17 12:18:09.658112 kernel: loop1: detected capacity change from 0 to 205544 Jan 17 12:18:09.679243 systemd[1]: Finished systemd-sysusers.service - Create System Users. Jan 17 12:18:09.685234 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jan 17 12:18:09.696113 kernel: loop2: detected capacity change from 0 to 142488 Jan 17 12:18:09.701151 systemd-tmpfiles[1234]: ACLs are not supported, ignoring. Jan 17 12:18:09.701162 systemd-tmpfiles[1234]: ACLs are not supported, ignoring. Jan 17 12:18:09.706239 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 17 12:18:09.743141 kernel: loop3: detected capacity change from 0 to 140768 Jan 17 12:18:09.790118 kernel: loop4: detected capacity change from 0 to 2976 Jan 17 12:18:09.813157 kernel: loop5: detected capacity change from 0 to 205544 Jan 17 12:18:09.829125 kernel: loop6: detected capacity change from 0 to 142488 Jan 17 12:18:09.863167 kernel: loop7: detected capacity change from 0 to 140768 Jan 17 12:18:09.915765 (sd-merge)[1239]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-vmware'. Jan 17 12:18:09.916111 (sd-merge)[1239]: Merged extensions into '/usr'. Jan 17 12:18:09.919854 systemd[1]: Reloading requested from client PID 1201 ('systemd-sysext') (unit systemd-sysext.service)... Jan 17 12:18:09.919861 systemd[1]: Reloading... Jan 17 12:18:09.968114 zram_generator::config[1262]: No configuration found. Jan 17 12:18:10.069819 systemd[1]: /etc/systemd/system/coreos-metadata.service:11: Ignoring unknown escape sequences: "echo "COREOS_CUSTOM_PRIVATE_IPV4=$(ip addr show ens192 | grep "inet 10." | grep -Po "inet \K[\d.]+") Jan 17 12:18:10.085673 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jan 17 12:18:10.113518 systemd[1]: Reloading finished in 193 ms. Jan 17 12:18:10.121785 ldconfig[1196]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Jan 17 12:18:10.132736 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Jan 17 12:18:10.133029 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Jan 17 12:18:10.133329 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Jan 17 12:18:10.140333 systemd[1]: Starting ensure-sysext.service... Jan 17 12:18:10.141220 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jan 17 12:18:10.143186 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 17 12:18:10.149081 systemd[1]: Reloading requested from client PID 1322 ('systemctl') (unit ensure-sysext.service)... Jan 17 12:18:10.149094 systemd[1]: Reloading... Jan 17 12:18:10.160721 systemd-tmpfiles[1323]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Jan 17 12:18:10.160930 systemd-tmpfiles[1323]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Jan 17 12:18:10.161638 systemd-tmpfiles[1323]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Jan 17 12:18:10.161850 systemd-tmpfiles[1323]: ACLs are not supported, ignoring. Jan 17 12:18:10.161927 systemd-tmpfiles[1323]: ACLs are not supported, ignoring. Jan 17 12:18:10.164083 systemd-tmpfiles[1323]: Detected autofs mount point /boot during canonicalization of boot. Jan 17 12:18:10.164150 systemd-tmpfiles[1323]: Skipping /boot Jan 17 12:18:10.169256 systemd-tmpfiles[1323]: Detected autofs mount point /boot during canonicalization of boot. Jan 17 12:18:10.169308 systemd-tmpfiles[1323]: Skipping /boot Jan 17 12:18:10.171164 systemd-udevd[1324]: Using default interface naming scheme 'v255'. Jan 17 12:18:10.186118 zram_generator::config[1348]: No configuration found. Jan 17 12:18:10.269261 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input3 Jan 17 12:18:10.278113 kernel: ACPI: button: Power Button [PWRF] Jan 17 12:18:10.299118 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 36 scanned by (udev-worker) (1363) Jan 17 12:18:10.310768 systemd[1]: /etc/systemd/system/coreos-metadata.service:11: Ignoring unknown escape sequences: "echo "COREOS_CUSTOM_PRIVATE_IPV4=$(ip addr show ens192 | grep "inet 10." | grep -Po "inet \K[\d.]+") Jan 17 12:18:10.332483 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jan 17 12:18:10.356116 kernel: piix4_smbus 0000:00:07.3: SMBus Host Controller not enabled! Jan 17 12:18:10.363132 kernel: vmw_vmci 0000:00:07.7: Using capabilities 0xc Jan 17 12:18:10.366614 kernel: Guest personality initialized and is active Jan 17 12:18:10.366642 kernel: VMCI host device registered (name=vmci, major=10, minor=125) Jan 17 12:18:10.366652 kernel: Initialized host personality Jan 17 12:18:10.380026 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Jan 17 12:18:10.380216 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Virtual_disk OEM. Jan 17 12:18:10.380847 systemd[1]: Reloading finished in 231 ms. Jan 17 12:18:10.381127 kernel: input: ImPS/2 Generic Wheel Mouse as /devices/platform/i8042/serio1/input/input4 Jan 17 12:18:10.390228 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 17 12:18:10.395338 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 17 12:18:10.410893 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 17 12:18:10.415116 kernel: mousedev: PS/2 mouse device common for all mice Jan 17 12:18:10.415258 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Jan 17 12:18:10.417634 (udev-worker)[1364]: id: Truncating stdout of 'dmi_memory_id' up to 16384 byte. Jan 17 12:18:10.418891 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Jan 17 12:18:10.420675 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 17 12:18:10.422816 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 17 12:18:10.425278 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 17 12:18:10.425445 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 17 12:18:10.426593 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Jan 17 12:18:10.429651 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Jan 17 12:18:10.431947 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jan 17 12:18:10.435297 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jan 17 12:18:10.436338 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Jan 17 12:18:10.436603 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 17 12:18:10.437613 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 17 12:18:10.438148 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 17 12:18:10.438472 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 17 12:18:10.438747 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 17 12:18:10.443244 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 17 12:18:10.446454 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 17 12:18:10.452338 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jan 17 12:18:10.453118 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 17 12:18:10.453277 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 17 12:18:10.453413 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jan 17 12:18:10.454021 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 17 12:18:10.454118 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 17 12:18:10.456752 systemd[1]: Finished ensure-sysext.service. Jan 17 12:18:10.457433 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Jan 17 12:18:10.466326 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Jan 17 12:18:10.470148 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Jan 17 12:18:10.472932 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 17 12:18:10.473373 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Jan 17 12:18:10.473625 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 17 12:18:10.473704 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 17 12:18:10.474203 systemd[1]: modprobe@drm.service: Deactivated successfully. Jan 17 12:18:10.474279 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jan 17 12:18:10.477322 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 17 12:18:10.477408 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 17 12:18:10.489268 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Jan 17 12:18:10.489397 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jan 17 12:18:10.489434 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jan 17 12:18:10.491130 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Jan 17 12:18:10.491417 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Jan 17 12:18:10.496884 systemd[1]: Starting systemd-update-done.service - Update is Completed... Jan 17 12:18:10.497136 augenrules[1484]: No rules Jan 17 12:18:10.497442 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Jan 17 12:18:10.505929 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Jan 17 12:18:10.506328 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Jan 17 12:18:10.511543 systemd[1]: Started systemd-userdbd.service - User Database Manager. Jan 17 12:18:10.512117 lvm[1479]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Jan 17 12:18:10.517367 systemd[1]: Finished systemd-update-done.service - Update is Completed. Jan 17 12:18:10.534299 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Jan 17 12:18:10.534478 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jan 17 12:18:10.538270 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Jan 17 12:18:10.551607 lvm[1502]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Jan 17 12:18:10.573153 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Jan 17 12:18:10.575673 systemd-networkd[1453]: lo: Link UP Jan 17 12:18:10.575679 systemd-networkd[1453]: lo: Gained carrier Jan 17 12:18:10.576391 systemd-networkd[1453]: Enumeration completed Jan 17 12:18:10.576455 systemd-timesyncd[1469]: No network connectivity, watching for changes. Jan 17 12:18:10.577362 systemd-networkd[1453]: ens192: Configuring with /etc/systemd/network/00-vmware.network. Jan 17 12:18:10.578189 kernel: vmxnet3 0000:0b:00.0 ens192: intr type 3, mode 0, 3 vectors allocated Jan 17 12:18:10.578303 kernel: vmxnet3 0000:0b:00.0 ens192: NIC Link is Up 10000 Mbps Jan 17 12:18:10.579872 systemd-networkd[1453]: ens192: Link UP Jan 17 12:18:10.579966 systemd-networkd[1453]: ens192: Gained carrier Jan 17 12:18:10.580041 systemd[1]: Started systemd-networkd.service - Network Configuration. Jan 17 12:18:10.580454 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 17 12:18:10.581641 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Jan 17 12:18:10.581918 systemd[1]: Reached target time-set.target - System Time Set. Jan 17 12:18:10.582848 systemd-timesyncd[1469]: Network configuration changed, trying to establish connection. Jan 17 12:18:10.585236 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Jan 17 12:18:10.589433 systemd-resolved[1454]: Positive Trust Anchors: Jan 17 12:18:10.589442 systemd-resolved[1454]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jan 17 12:18:10.589464 systemd-resolved[1454]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jan 17 12:18:10.592121 systemd-resolved[1454]: Defaulting to hostname 'linux'. Jan 17 12:18:10.593044 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jan 17 12:18:10.593209 systemd[1]: Reached target network.target - Network. Jan 17 12:18:10.593296 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jan 17 12:18:10.593403 systemd[1]: Reached target sysinit.target - System Initialization. Jan 17 12:18:10.593572 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Jan 17 12:18:10.593698 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Jan 17 12:18:10.593880 systemd[1]: Started logrotate.timer - Daily rotation of log files. Jan 17 12:18:10.594021 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Jan 17 12:18:10.594135 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Jan 17 12:18:10.594237 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Jan 17 12:18:10.594255 systemd[1]: Reached target paths.target - Path Units. Jan 17 12:18:10.594338 systemd[1]: Reached target timers.target - Timer Units. Jan 17 12:18:10.594861 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Jan 17 12:18:10.595857 systemd[1]: Starting docker.socket - Docker Socket for the API... Jan 17 12:18:10.599071 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Jan 17 12:18:10.599463 systemd[1]: Listening on docker.socket - Docker Socket for the API. Jan 17 12:18:10.599596 systemd[1]: Reached target sockets.target - Socket Units. Jan 17 12:18:10.599679 systemd[1]: Reached target basic.target - Basic System. Jan 17 12:18:10.599784 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Jan 17 12:18:10.599800 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Jan 17 12:18:10.600501 systemd[1]: Starting containerd.service - containerd container runtime... Jan 17 12:18:10.603176 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Jan 17 12:18:10.604415 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Jan 17 12:18:10.607126 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Jan 17 12:18:10.607227 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Jan 17 12:18:10.608253 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Jan 17 12:18:10.612047 jq[1514]: false Jan 17 12:18:10.613205 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Jan 17 12:18:10.615964 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Jan 17 12:18:10.619254 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Jan 17 12:18:10.619754 dbus-daemon[1513]: [system] SELinux support is enabled Jan 17 12:18:10.624186 systemd[1]: Starting systemd-logind.service - User Login Management... Jan 17 12:18:10.624481 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Jan 17 12:18:10.626066 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Jan 17 12:18:10.627219 systemd[1]: Starting update-engine.service - Update Engine... Jan 17 12:18:10.630499 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Jan 17 12:18:10.630992 extend-filesystems[1515]: Found loop4 Jan 17 12:18:10.633321 extend-filesystems[1515]: Found loop5 Jan 17 12:18:10.633321 extend-filesystems[1515]: Found loop6 Jan 17 12:18:10.633321 extend-filesystems[1515]: Found loop7 Jan 17 12:18:10.633321 extend-filesystems[1515]: Found sda Jan 17 12:18:10.633321 extend-filesystems[1515]: Found sda1 Jan 17 12:18:10.633321 extend-filesystems[1515]: Found sda2 Jan 17 12:18:10.633321 extend-filesystems[1515]: Found sda3 Jan 17 12:18:10.633321 extend-filesystems[1515]: Found usr Jan 17 12:18:10.633321 extend-filesystems[1515]: Found sda4 Jan 17 12:18:10.633321 extend-filesystems[1515]: Found sda6 Jan 17 12:18:10.633321 extend-filesystems[1515]: Found sda7 Jan 17 12:18:10.633321 extend-filesystems[1515]: Found sda9 Jan 17 12:18:10.633321 extend-filesystems[1515]: Checking size of /dev/sda9 Jan 17 12:18:10.634947 systemd[1]: Starting vgauthd.service - VGAuth Service for open-vm-tools... Jan 17 12:18:10.635304 systemd[1]: Started dbus.service - D-Bus System Message Bus. Jan 17 12:18:10.645310 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Jan 17 12:18:10.645430 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Jan 17 12:18:10.648713 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Jan 17 12:18:10.648804 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Jan 17 12:18:10.650493 update_engine[1525]: I20250117 12:18:10.650436 1525 main.cc:92] Flatcar Update Engine starting Jan 17 12:18:10.652602 update_engine[1525]: I20250117 12:18:10.652585 1525 update_check_scheduler.cc:74] Next update check in 6m55s Jan 17 12:18:10.654552 extend-filesystems[1515]: Old size kept for /dev/sda9 Jan 17 12:18:10.654685 extend-filesystems[1515]: Found sr0 Jan 17 12:18:10.657116 jq[1526]: true Jan 17 12:18:10.657421 systemd[1]: extend-filesystems.service: Deactivated successfully. Jan 17 12:18:10.657534 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Jan 17 12:18:10.657806 systemd[1]: motdgen.service: Deactivated successfully. Jan 17 12:18:10.657892 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Jan 17 12:18:10.660130 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Jan 17 12:18:10.660160 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Jan 17 12:18:10.660739 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Jan 17 12:18:10.660753 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Jan 17 12:18:10.661010 systemd[1]: Started update-engine.service - Update Engine. Jan 17 12:18:10.664178 systemd[1]: Started locksmithd.service - Cluster reboot manager. Jan 17 12:18:10.666913 (ntainerd)[1544]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Jan 17 12:18:10.673265 systemd[1]: Started vgauthd.service - VGAuth Service for open-vm-tools. Jan 17 12:18:10.676857 systemd[1]: Starting vmtoolsd.service - Service for virtual machines hosted on VMware... Jan 17 12:18:10.681125 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 36 scanned by (udev-worker) (1360) Jan 17 12:18:10.681164 jq[1550]: true Jan 17 12:18:10.690567 systemd-logind[1522]: Watching system buttons on /dev/input/event1 (Power Button) Jan 17 12:18:10.690717 systemd-logind[1522]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Jan 17 12:18:10.691066 systemd-logind[1522]: New seat seat0. Jan 17 12:18:10.700351 systemd[1]: Started systemd-logind.service - User Login Management. Jan 17 12:18:10.703930 tar[1536]: linux-amd64/helm Jan 17 12:18:10.728195 systemd[1]: Started vmtoolsd.service - Service for virtual machines hosted on VMware. Jan 17 12:18:10.740088 unknown[1553]: Pref_Init: Using '/etc/vmware-tools/vgauth.conf' as preferences filepath Jan 17 12:18:10.748162 unknown[1553]: Core dump limit set to -1 Jan 17 12:18:10.763414 bash[1575]: Updated "/home/core/.ssh/authorized_keys" Jan 17 12:18:10.763607 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Jan 17 12:18:10.766174 kernel: NET: Registered PF_VSOCK protocol family Jan 17 12:18:10.764458 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Jan 17 12:19:23.901976 systemd-timesyncd[1469]: Contacted time server 50.205.57.38:123 (3.flatcar.pool.ntp.org). Jan 17 12:19:23.902088 systemd-timesyncd[1469]: Initial clock synchronization to Fri 2025-01-17 12:19:23.901600 UTC. Jan 17 12:19:23.904700 systemd-resolved[1454]: Clock change detected. Flushing caches. Jan 17 12:19:23.941130 locksmithd[1548]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Jan 17 12:19:23.952159 sshd_keygen[1554]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Jan 17 12:19:23.971974 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Jan 17 12:19:23.977858 systemd[1]: Starting issuegen.service - Generate /run/issue... Jan 17 12:19:23.981849 systemd[1]: issuegen.service: Deactivated successfully. Jan 17 12:19:23.981981 systemd[1]: Finished issuegen.service - Generate /run/issue. Jan 17 12:19:23.983919 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Jan 17 12:19:24.006610 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Jan 17 12:19:24.010904 systemd[1]: Started getty@tty1.service - Getty on tty1. Jan 17 12:19:24.011612 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Jan 17 12:19:24.011801 systemd[1]: Reached target getty.target - Login Prompts. Jan 17 12:19:24.061590 containerd[1544]: time="2025-01-17T12:19:24.061532912Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Jan 17 12:19:24.083776 containerd[1544]: time="2025-01-17T12:19:24.083720802Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Jan 17 12:19:24.085721 containerd[1544]: time="2025-01-17T12:19:24.085506125Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.71-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Jan 17 12:19:24.085721 containerd[1544]: time="2025-01-17T12:19:24.085522435Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Jan 17 12:19:24.085721 containerd[1544]: time="2025-01-17T12:19:24.085532562Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Jan 17 12:19:24.085721 containerd[1544]: time="2025-01-17T12:19:24.085619922Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Jan 17 12:19:24.085721 containerd[1544]: time="2025-01-17T12:19:24.085630071Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Jan 17 12:19:24.085721 containerd[1544]: time="2025-01-17T12:19:24.085662628Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Jan 17 12:19:24.085721 containerd[1544]: time="2025-01-17T12:19:24.085679928Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Jan 17 12:19:24.086293 containerd[1544]: time="2025-01-17T12:19:24.086100595Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Jan 17 12:19:24.086293 containerd[1544]: time="2025-01-17T12:19:24.086112411Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Jan 17 12:19:24.086293 containerd[1544]: time="2025-01-17T12:19:24.086120770Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Jan 17 12:19:24.086293 containerd[1544]: time="2025-01-17T12:19:24.086126584Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Jan 17 12:19:24.086293 containerd[1544]: time="2025-01-17T12:19:24.086169895Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Jan 17 12:19:24.086293 containerd[1544]: time="2025-01-17T12:19:24.086277427Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Jan 17 12:19:24.086437 containerd[1544]: time="2025-01-17T12:19:24.086427806Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Jan 17 12:19:24.086468 containerd[1544]: time="2025-01-17T12:19:24.086462218Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Jan 17 12:19:24.086538 containerd[1544]: time="2025-01-17T12:19:24.086528934Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Jan 17 12:19:24.086601 containerd[1544]: time="2025-01-17T12:19:24.086584029Z" level=info msg="metadata content store policy set" policy=shared Jan 17 12:19:24.088147 containerd[1544]: time="2025-01-17T12:19:24.087974804Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Jan 17 12:19:24.088305 containerd[1544]: time="2025-01-17T12:19:24.088186098Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Jan 17 12:19:24.088305 containerd[1544]: time="2025-01-17T12:19:24.088200776Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Jan 17 12:19:24.088305 containerd[1544]: time="2025-01-17T12:19:24.088209761Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Jan 17 12:19:24.088305 containerd[1544]: time="2025-01-17T12:19:24.088217928Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Jan 17 12:19:24.088305 containerd[1544]: time="2025-01-17T12:19:24.088282338Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Jan 17 12:19:24.088510 containerd[1544]: time="2025-01-17T12:19:24.088501003Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Jan 17 12:19:24.088685 containerd[1544]: time="2025-01-17T12:19:24.088590790Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Jan 17 12:19:24.088685 containerd[1544]: time="2025-01-17T12:19:24.088602000Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Jan 17 12:19:24.088685 containerd[1544]: time="2025-01-17T12:19:24.088609744Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Jan 17 12:19:24.088685 containerd[1544]: time="2025-01-17T12:19:24.088617208Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Jan 17 12:19:24.088685 containerd[1544]: time="2025-01-17T12:19:24.088624344Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Jan 17 12:19:24.088685 containerd[1544]: time="2025-01-17T12:19:24.088631738Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Jan 17 12:19:24.088685 containerd[1544]: time="2025-01-17T12:19:24.088639492Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Jan 17 12:19:24.088685 containerd[1544]: time="2025-01-17T12:19:24.088647040Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Jan 17 12:19:24.088685 containerd[1544]: time="2025-01-17T12:19:24.088659803Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Jan 17 12:19:24.089048 containerd[1544]: time="2025-01-17T12:19:24.088668145Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Jan 17 12:19:24.089048 containerd[1544]: time="2025-01-17T12:19:24.088823730Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Jan 17 12:19:24.089048 containerd[1544]: time="2025-01-17T12:19:24.088837809Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Jan 17 12:19:24.089048 containerd[1544]: time="2025-01-17T12:19:24.088845741Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Jan 17 12:19:24.089048 containerd[1544]: time="2025-01-17T12:19:24.088852551Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Jan 17 12:19:24.089048 containerd[1544]: time="2025-01-17T12:19:24.088861997Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Jan 17 12:19:24.089048 containerd[1544]: time="2025-01-17T12:19:24.088872333Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Jan 17 12:19:24.089048 containerd[1544]: time="2025-01-17T12:19:24.088883124Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Jan 17 12:19:24.089048 containerd[1544]: time="2025-01-17T12:19:24.088915780Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Jan 17 12:19:24.089048 containerd[1544]: time="2025-01-17T12:19:24.088943574Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Jan 17 12:19:24.089048 containerd[1544]: time="2025-01-17T12:19:24.088951317Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Jan 17 12:19:24.089048 containerd[1544]: time="2025-01-17T12:19:24.088960533Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Jan 17 12:19:24.089048 containerd[1544]: time="2025-01-17T12:19:24.088966953Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Jan 17 12:19:24.089048 containerd[1544]: time="2025-01-17T12:19:24.088973406Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Jan 17 12:19:24.089234 containerd[1544]: time="2025-01-17T12:19:24.088980316Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Jan 17 12:19:24.089234 containerd[1544]: time="2025-01-17T12:19:24.089003725Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Jan 17 12:19:24.089234 containerd[1544]: time="2025-01-17T12:19:24.089014732Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Jan 17 12:19:24.089234 containerd[1544]: time="2025-01-17T12:19:24.089021140Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Jan 17 12:19:24.089234 containerd[1544]: time="2025-01-17T12:19:24.089026683Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Jan 17 12:19:24.089696 containerd[1544]: time="2025-01-17T12:19:24.089415524Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Jan 17 12:19:24.089696 containerd[1544]: time="2025-01-17T12:19:24.089431582Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Jan 17 12:19:24.089696 containerd[1544]: time="2025-01-17T12:19:24.089438482Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Jan 17 12:19:24.089696 containerd[1544]: time="2025-01-17T12:19:24.089445283Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Jan 17 12:19:24.089696 containerd[1544]: time="2025-01-17T12:19:24.089450780Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Jan 17 12:19:24.089696 containerd[1544]: time="2025-01-17T12:19:24.089457335Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Jan 17 12:19:24.089696 containerd[1544]: time="2025-01-17T12:19:24.089497867Z" level=info msg="NRI interface is disabled by configuration." Jan 17 12:19:24.089696 containerd[1544]: time="2025-01-17T12:19:24.089505808Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Jan 17 12:19:24.089828 containerd[1544]: time="2025-01-17T12:19:24.089653415Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Jan 17 12:19:24.089927 containerd[1544]: time="2025-01-17T12:19:24.089919277Z" level=info msg="Connect containerd service" Jan 17 12:19:24.090199 containerd[1544]: time="2025-01-17T12:19:24.089985787Z" level=info msg="using legacy CRI server" Jan 17 12:19:24.090199 containerd[1544]: time="2025-01-17T12:19:24.089993582Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Jan 17 12:19:24.090199 containerd[1544]: time="2025-01-17T12:19:24.090048035Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Jan 17 12:19:24.090928 containerd[1544]: time="2025-01-17T12:19:24.090849270Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Jan 17 12:19:24.090988 containerd[1544]: time="2025-01-17T12:19:24.090969014Z" level=info msg="Start subscribing containerd event" Jan 17 12:19:24.091147 containerd[1544]: time="2025-01-17T12:19:24.091139143Z" level=info msg="Start recovering state" Jan 17 12:19:24.091208 containerd[1544]: time="2025-01-17T12:19:24.091200497Z" level=info msg="Start event monitor" Jan 17 12:19:24.091349 containerd[1544]: time="2025-01-17T12:19:24.091341592Z" level=info msg="Start snapshots syncer" Jan 17 12:19:24.091691 containerd[1544]: time="2025-01-17T12:19:24.091374894Z" level=info msg="Start cni network conf syncer for default" Jan 17 12:19:24.091691 containerd[1544]: time="2025-01-17T12:19:24.091382243Z" level=info msg="Start streaming server" Jan 17 12:19:24.091691 containerd[1544]: time="2025-01-17T12:19:24.091307378Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Jan 17 12:19:24.091691 containerd[1544]: time="2025-01-17T12:19:24.091445094Z" level=info msg=serving... address=/run/containerd/containerd.sock Jan 17 12:19:24.092106 containerd[1544]: time="2025-01-17T12:19:24.092096761Z" level=info msg="containerd successfully booted in 0.031141s" Jan 17 12:19:24.092145 systemd[1]: Started containerd.service - containerd container runtime. Jan 17 12:19:24.235076 tar[1536]: linux-amd64/LICENSE Jan 17 12:19:24.235435 tar[1536]: linux-amd64/README.md Jan 17 12:19:24.248388 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Jan 17 12:19:25.734981 systemd-networkd[1453]: ens192: Gained IPv6LL Jan 17 12:19:25.736141 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Jan 17 12:19:25.737036 systemd[1]: Reached target network-online.target - Network is Online. Jan 17 12:19:25.746951 systemd[1]: Starting coreos-metadata.service - VMware metadata agent... Jan 17 12:19:25.748502 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 17 12:19:25.749837 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Jan 17 12:19:25.777017 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Jan 17 12:19:25.778509 systemd[1]: coreos-metadata.service: Deactivated successfully. Jan 17 12:19:25.778693 systemd[1]: Finished coreos-metadata.service - VMware metadata agent. Jan 17 12:19:25.779280 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Jan 17 12:19:26.503571 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 17 12:19:26.504171 systemd[1]: Reached target multi-user.target - Multi-User System. Jan 17 12:19:26.506499 systemd[1]: Startup finished in 961ms (kernel) + 5.369s (initrd) + 4.419s (userspace) = 10.750s. Jan 17 12:19:26.511466 (kubelet)[1692]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 17 12:19:26.534954 login[1604]: pam_unix(login:session): session opened for user core(uid=500) by LOGIN(uid=0) Jan 17 12:19:26.537803 login[1605]: pam_unix(login:session): session opened for user core(uid=500) by LOGIN(uid=0) Jan 17 12:19:26.543370 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Jan 17 12:19:26.551882 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Jan 17 12:19:26.554583 systemd-logind[1522]: New session 2 of user core. Jan 17 12:19:26.558964 systemd-logind[1522]: New session 1 of user core. Jan 17 12:19:26.563374 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Jan 17 12:19:26.569112 systemd[1]: Starting user@500.service - User Manager for UID 500... Jan 17 12:19:26.571591 (systemd)[1699]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Jan 17 12:19:26.679824 systemd[1699]: Queued start job for default target default.target. Jan 17 12:19:26.684577 systemd[1699]: Created slice app.slice - User Application Slice. Jan 17 12:19:26.684601 systemd[1699]: Reached target paths.target - Paths. Jan 17 12:19:26.684615 systemd[1699]: Reached target timers.target - Timers. Jan 17 12:19:26.685451 systemd[1699]: Starting dbus.socket - D-Bus User Message Bus Socket... Jan 17 12:19:26.693639 systemd[1699]: Listening on dbus.socket - D-Bus User Message Bus Socket. Jan 17 12:19:26.693691 systemd[1699]: Reached target sockets.target - Sockets. Jan 17 12:19:26.694124 systemd[1699]: Reached target basic.target - Basic System. Jan 17 12:19:26.694162 systemd[1699]: Reached target default.target - Main User Target. Jan 17 12:19:26.694188 systemd[1699]: Startup finished in 118ms. Jan 17 12:19:26.694195 systemd[1]: Started user@500.service - User Manager for UID 500. Jan 17 12:19:26.702776 systemd[1]: Started session-1.scope - Session 1 of User core. Jan 17 12:19:26.703560 systemd[1]: Started session-2.scope - Session 2 of User core. Jan 17 12:19:27.079273 kubelet[1692]: E0117 12:19:27.079235 1692 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 17 12:19:27.080727 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 17 12:19:27.080851 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 17 12:19:37.331091 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Jan 17 12:19:37.341878 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 17 12:19:37.446081 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 17 12:19:37.449343 (kubelet)[1742]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 17 12:19:37.494755 kubelet[1742]: E0117 12:19:37.494702 1742 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 17 12:19:37.496997 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 17 12:19:37.497080 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 17 12:19:47.747405 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Jan 17 12:19:47.756809 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 17 12:19:47.818566 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 17 12:19:47.821566 (kubelet)[1757]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 17 12:19:47.844654 kubelet[1757]: E0117 12:19:47.844621 1757 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 17 12:19:47.845806 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 17 12:19:47.845890 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 17 12:19:58.096145 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Jan 17 12:19:58.107020 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 17 12:19:58.338500 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 17 12:19:58.341118 (kubelet)[1772]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 17 12:19:58.361048 kubelet[1772]: E0117 12:19:58.360950 1772 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 17 12:19:58.362572 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 17 12:19:58.362664 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 17 12:20:03.977574 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Jan 17 12:20:03.978450 systemd[1]: Started sshd@0-139.178.70.100:22-147.75.109.163:55024.service - OpenSSH per-connection server daemon (147.75.109.163:55024). Jan 17 12:20:04.009949 sshd[1780]: Accepted publickey for core from 147.75.109.163 port 55024 ssh2: RSA SHA256:d86Zfld7pfipwDMCy9Zh9gJz3C7zt8CsQJU6anwQyxc Jan 17 12:20:04.010995 sshd[1780]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 12:20:04.014362 systemd-logind[1522]: New session 3 of user core. Jan 17 12:20:04.020851 systemd[1]: Started session-3.scope - Session 3 of User core. Jan 17 12:20:04.075819 systemd[1]: Started sshd@1-139.178.70.100:22-147.75.109.163:55026.service - OpenSSH per-connection server daemon (147.75.109.163:55026). Jan 17 12:20:04.099991 sshd[1785]: Accepted publickey for core from 147.75.109.163 port 55026 ssh2: RSA SHA256:d86Zfld7pfipwDMCy9Zh9gJz3C7zt8CsQJU6anwQyxc Jan 17 12:20:04.100937 sshd[1785]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 12:20:04.105071 systemd-logind[1522]: New session 4 of user core. Jan 17 12:20:04.107801 systemd[1]: Started session-4.scope - Session 4 of User core. Jan 17 12:20:04.161514 sshd[1785]: pam_unix(sshd:session): session closed for user core Jan 17 12:20:04.168514 systemd[1]: sshd@1-139.178.70.100:22-147.75.109.163:55026.service: Deactivated successfully. Jan 17 12:20:04.169834 systemd[1]: session-4.scope: Deactivated successfully. Jan 17 12:20:04.170810 systemd-logind[1522]: Session 4 logged out. Waiting for processes to exit. Jan 17 12:20:04.171850 systemd[1]: Started sshd@2-139.178.70.100:22-147.75.109.163:55036.service - OpenSSH per-connection server daemon (147.75.109.163:55036). Jan 17 12:20:04.172444 systemd-logind[1522]: Removed session 4. Jan 17 12:20:04.199766 sshd[1792]: Accepted publickey for core from 147.75.109.163 port 55036 ssh2: RSA SHA256:d86Zfld7pfipwDMCy9Zh9gJz3C7zt8CsQJU6anwQyxc Jan 17 12:20:04.200515 sshd[1792]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 12:20:04.203727 systemd-logind[1522]: New session 5 of user core. Jan 17 12:20:04.210758 systemd[1]: Started session-5.scope - Session 5 of User core. Jan 17 12:20:04.257034 sshd[1792]: pam_unix(sshd:session): session closed for user core Jan 17 12:20:04.265962 systemd[1]: sshd@2-139.178.70.100:22-147.75.109.163:55036.service: Deactivated successfully. Jan 17 12:20:04.266662 systemd[1]: session-5.scope: Deactivated successfully. Jan 17 12:20:04.267408 systemd-logind[1522]: Session 5 logged out. Waiting for processes to exit. Jan 17 12:20:04.268055 systemd[1]: Started sshd@3-139.178.70.100:22-147.75.109.163:55052.service - OpenSSH per-connection server daemon (147.75.109.163:55052). Jan 17 12:20:04.269813 systemd-logind[1522]: Removed session 5. Jan 17 12:20:04.292415 sshd[1799]: Accepted publickey for core from 147.75.109.163 port 55052 ssh2: RSA SHA256:d86Zfld7pfipwDMCy9Zh9gJz3C7zt8CsQJU6anwQyxc Jan 17 12:20:04.293140 sshd[1799]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 12:20:04.295721 systemd-logind[1522]: New session 6 of user core. Jan 17 12:20:04.301744 systemd[1]: Started session-6.scope - Session 6 of User core. Jan 17 12:20:04.350523 sshd[1799]: pam_unix(sshd:session): session closed for user core Jan 17 12:20:04.361140 systemd[1]: sshd@3-139.178.70.100:22-147.75.109.163:55052.service: Deactivated successfully. Jan 17 12:20:04.361945 systemd[1]: session-6.scope: Deactivated successfully. Jan 17 12:20:04.362710 systemd-logind[1522]: Session 6 logged out. Waiting for processes to exit. Jan 17 12:20:04.363452 systemd[1]: Started sshd@4-139.178.70.100:22-147.75.109.163:55054.service - OpenSSH per-connection server daemon (147.75.109.163:55054). Jan 17 12:20:04.364884 systemd-logind[1522]: Removed session 6. Jan 17 12:20:04.390157 sshd[1806]: Accepted publickey for core from 147.75.109.163 port 55054 ssh2: RSA SHA256:d86Zfld7pfipwDMCy9Zh9gJz3C7zt8CsQJU6anwQyxc Jan 17 12:20:04.391114 sshd[1806]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 12:20:04.395545 systemd-logind[1522]: New session 7 of user core. Jan 17 12:20:04.400821 systemd[1]: Started session-7.scope - Session 7 of User core. Jan 17 12:20:04.456220 sudo[1809]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Jan 17 12:20:04.456627 sudo[1809]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 17 12:20:04.468443 sudo[1809]: pam_unix(sudo:session): session closed for user root Jan 17 12:20:04.469469 sshd[1806]: pam_unix(sshd:session): session closed for user core Jan 17 12:20:04.478378 systemd[1]: sshd@4-139.178.70.100:22-147.75.109.163:55054.service: Deactivated successfully. Jan 17 12:20:04.479157 systemd[1]: session-7.scope: Deactivated successfully. Jan 17 12:20:04.480104 systemd-logind[1522]: Session 7 logged out. Waiting for processes to exit. Jan 17 12:20:04.484891 systemd[1]: Started sshd@5-139.178.70.100:22-147.75.109.163:55058.service - OpenSSH per-connection server daemon (147.75.109.163:55058). Jan 17 12:20:04.486101 systemd-logind[1522]: Removed session 7. Jan 17 12:20:04.508840 sshd[1814]: Accepted publickey for core from 147.75.109.163 port 55058 ssh2: RSA SHA256:d86Zfld7pfipwDMCy9Zh9gJz3C7zt8CsQJU6anwQyxc Jan 17 12:20:04.509701 sshd[1814]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 12:20:04.514205 systemd-logind[1522]: New session 8 of user core. Jan 17 12:20:04.520796 systemd[1]: Started session-8.scope - Session 8 of User core. Jan 17 12:20:04.569847 sudo[1818]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Jan 17 12:20:04.570292 sudo[1818]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 17 12:20:04.572579 sudo[1818]: pam_unix(sudo:session): session closed for user root Jan 17 12:20:04.575920 sudo[1817]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Jan 17 12:20:04.576102 sudo[1817]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 17 12:20:04.583826 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Jan 17 12:20:04.585686 auditctl[1821]: No rules Jan 17 12:20:04.585924 systemd[1]: audit-rules.service: Deactivated successfully. Jan 17 12:20:04.586070 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Jan 17 12:20:04.587258 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Jan 17 12:20:04.603299 augenrules[1839]: No rules Jan 17 12:20:04.604175 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Jan 17 12:20:04.605055 sudo[1817]: pam_unix(sudo:session): session closed for user root Jan 17 12:20:04.605974 sshd[1814]: pam_unix(sshd:session): session closed for user core Jan 17 12:20:04.610184 systemd[1]: sshd@5-139.178.70.100:22-147.75.109.163:55058.service: Deactivated successfully. Jan 17 12:20:04.611298 systemd[1]: session-8.scope: Deactivated successfully. Jan 17 12:20:04.612102 systemd-logind[1522]: Session 8 logged out. Waiting for processes to exit. Jan 17 12:20:04.616871 systemd[1]: Started sshd@6-139.178.70.100:22-147.75.109.163:55068.service - OpenSSH per-connection server daemon (147.75.109.163:55068). Jan 17 12:20:04.617783 systemd-logind[1522]: Removed session 8. Jan 17 12:20:04.639187 sshd[1847]: Accepted publickey for core from 147.75.109.163 port 55068 ssh2: RSA SHA256:d86Zfld7pfipwDMCy9Zh9gJz3C7zt8CsQJU6anwQyxc Jan 17 12:20:04.639907 sshd[1847]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 12:20:04.642157 systemd-logind[1522]: New session 9 of user core. Jan 17 12:20:04.661843 systemd[1]: Started session-9.scope - Session 9 of User core. Jan 17 12:20:04.712241 sudo[1850]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Jan 17 12:20:04.712480 sudo[1850]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 17 12:20:05.004920 (dockerd)[1866]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Jan 17 12:20:05.005434 systemd[1]: Starting docker.service - Docker Application Container Engine... Jan 17 12:20:05.262786 dockerd[1866]: time="2025-01-17T12:20:05.262694944Z" level=info msg="Starting up" Jan 17 12:20:05.410549 systemd[1]: var-lib-docker-check\x2doverlayfs\x2dsupport2067424599-merged.mount: Deactivated successfully. Jan 17 12:20:05.530539 dockerd[1866]: time="2025-01-17T12:20:05.530245142Z" level=info msg="Loading containers: start." Jan 17 12:20:05.739699 kernel: Initializing XFRM netlink socket Jan 17 12:20:05.898645 systemd-networkd[1453]: docker0: Link UP Jan 17 12:20:05.908694 dockerd[1866]: time="2025-01-17T12:20:05.908380251Z" level=info msg="Loading containers: done." Jan 17 12:20:05.918005 dockerd[1866]: time="2025-01-17T12:20:05.917935259Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Jan 17 12:20:05.918361 dockerd[1866]: time="2025-01-17T12:20:05.918125634Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Jan 17 12:20:05.918361 dockerd[1866]: time="2025-01-17T12:20:05.918212537Z" level=info msg="Daemon has completed initialization" Jan 17 12:20:05.936281 dockerd[1866]: time="2025-01-17T12:20:05.935942489Z" level=info msg="API listen on /run/docker.sock" Jan 17 12:20:05.936075 systemd[1]: Started docker.service - Docker Application Container Engine. Jan 17 12:20:06.910939 containerd[1544]: time="2025-01-17T12:20:06.910911072Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.5\"" Jan 17 12:20:07.678953 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount544283267.mount: Deactivated successfully. Jan 17 12:20:08.587360 containerd[1544]: time="2025-01-17T12:20:08.587291845Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.31.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:20:08.587943 containerd[1544]: time="2025-01-17T12:20:08.587915867Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.31.5: active requests=0, bytes read=27976721" Jan 17 12:20:08.589280 containerd[1544]: time="2025-01-17T12:20:08.589251536Z" level=info msg="ImageCreate event name:\"sha256:2212e74642e45d72a36f297bea139f607ce4ccc4792966a8e9c4d30e04a4a6fb\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:20:08.590739 containerd[1544]: time="2025-01-17T12:20:08.590710128Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:fc4b366c0036b90d147f3b58244cf7d5f1f42b0db539f0fe83a8fc6e25a434ab\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:20:08.591431 containerd[1544]: time="2025-01-17T12:20:08.591329620Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.31.5\" with image id \"sha256:2212e74642e45d72a36f297bea139f607ce4ccc4792966a8e9c4d30e04a4a6fb\", repo tag \"registry.k8s.io/kube-apiserver:v1.31.5\", repo digest \"registry.k8s.io/kube-apiserver@sha256:fc4b366c0036b90d147f3b58244cf7d5f1f42b0db539f0fe83a8fc6e25a434ab\", size \"27973521\" in 1.680394322s" Jan 17 12:20:08.591431 containerd[1544]: time="2025-01-17T12:20:08.591349156Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.5\" returns image reference \"sha256:2212e74642e45d72a36f297bea139f607ce4ccc4792966a8e9c4d30e04a4a6fb\"" Jan 17 12:20:08.592609 containerd[1544]: time="2025-01-17T12:20:08.592557054Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.5\"" Jan 17 12:20:08.612868 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 4. Jan 17 12:20:08.618806 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 17 12:20:08.716937 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 17 12:20:08.719449 (kubelet)[2065]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 17 12:20:08.741720 kubelet[2065]: E0117 12:20:08.741689 2065 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 17 12:20:08.742659 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 17 12:20:08.742751 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 17 12:20:08.758723 update_engine[1525]: I20250117 12:20:08.758692 1525 update_attempter.cc:509] Updating boot flags... Jan 17 12:20:08.787714 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 36 scanned by (udev-worker) (2078) Jan 17 12:20:08.877700 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 36 scanned by (udev-worker) (2081) Jan 17 12:20:10.077570 containerd[1544]: time="2025-01-17T12:20:10.077535513Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.31.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:20:10.083013 containerd[1544]: time="2025-01-17T12:20:10.082984192Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.31.5: active requests=0, bytes read=24701143" Jan 17 12:20:10.093139 containerd[1544]: time="2025-01-17T12:20:10.093110642Z" level=info msg="ImageCreate event name:\"sha256:d7fccb640e0edce9c47bd71f2b2ce328b824bea199bfe5838dda3fe2af6372f2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:20:10.111789 containerd[1544]: time="2025-01-17T12:20:10.111751344Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:848cf42bf6c3c5ccac232b76c901c309edb3ebeac4d856885af0fc718798207e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:20:10.112533 containerd[1544]: time="2025-01-17T12:20:10.112514249Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.31.5\" with image id \"sha256:d7fccb640e0edce9c47bd71f2b2ce328b824bea199bfe5838dda3fe2af6372f2\", repo tag \"registry.k8s.io/kube-controller-manager:v1.31.5\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:848cf42bf6c3c5ccac232b76c901c309edb3ebeac4d856885af0fc718798207e\", size \"26147725\" in 1.519941256s" Jan 17 12:20:10.112588 containerd[1544]: time="2025-01-17T12:20:10.112532784Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.5\" returns image reference \"sha256:d7fccb640e0edce9c47bd71f2b2ce328b824bea199bfe5838dda3fe2af6372f2\"" Jan 17 12:20:10.112846 containerd[1544]: time="2025-01-17T12:20:10.112824609Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.5\"" Jan 17 12:20:11.338221 containerd[1544]: time="2025-01-17T12:20:11.337919141Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.31.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:20:11.338503 containerd[1544]: time="2025-01-17T12:20:11.338459903Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.31.5: active requests=0, bytes read=18652053" Jan 17 12:20:11.338760 containerd[1544]: time="2025-01-17T12:20:11.338741731Z" level=info msg="ImageCreate event name:\"sha256:4b2fb209f5d1efc0fc980c5acda28886e4eb6ab4820173976bdd441cbd2ee09a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:20:11.340553 containerd[1544]: time="2025-01-17T12:20:11.340527205Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:0e01fd956ba32a7fa08f6b6da24e8c49015905c8e2cf752978d495e44cd4a8a9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:20:11.341166 containerd[1544]: time="2025-01-17T12:20:11.341145612Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.31.5\" with image id \"sha256:4b2fb209f5d1efc0fc980c5acda28886e4eb6ab4820173976bdd441cbd2ee09a\", repo tag \"registry.k8s.io/kube-scheduler:v1.31.5\", repo digest \"registry.k8s.io/kube-scheduler@sha256:0e01fd956ba32a7fa08f6b6da24e8c49015905c8e2cf752978d495e44cd4a8a9\", size \"20098653\" in 1.227662692s" Jan 17 12:20:11.341196 containerd[1544]: time="2025-01-17T12:20:11.341169756Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.5\" returns image reference \"sha256:4b2fb209f5d1efc0fc980c5acda28886e4eb6ab4820173976bdd441cbd2ee09a\"" Jan 17 12:20:11.341741 containerd[1544]: time="2025-01-17T12:20:11.341647099Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.5\"" Jan 17 12:20:12.434913 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1621282553.mount: Deactivated successfully. Jan 17 12:20:12.793595 containerd[1544]: time="2025-01-17T12:20:12.793356729Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.31.5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:20:12.798139 containerd[1544]: time="2025-01-17T12:20:12.797997056Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.31.5: active requests=0, bytes read=30231128" Jan 17 12:20:12.803377 containerd[1544]: time="2025-01-17T12:20:12.803354846Z" level=info msg="ImageCreate event name:\"sha256:34018aef09a62f8b40bdd1d2e1bf6c48f359cab492d51059a09e20745ab02ce2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:20:12.813176 containerd[1544]: time="2025-01-17T12:20:12.813131271Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:c00685cc45c1fb539c5bbd8d24d2577f96e9399efac1670f688f654b30f8c64c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:20:12.813830 containerd[1544]: time="2025-01-17T12:20:12.813506788Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.31.5\" with image id \"sha256:34018aef09a62f8b40bdd1d2e1bf6c48f359cab492d51059a09e20745ab02ce2\", repo tag \"registry.k8s.io/kube-proxy:v1.31.5\", repo digest \"registry.k8s.io/kube-proxy@sha256:c00685cc45c1fb539c5bbd8d24d2577f96e9399efac1670f688f654b30f8c64c\", size \"30230147\" in 1.471816169s" Jan 17 12:20:12.813830 containerd[1544]: time="2025-01-17T12:20:12.813529385Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.5\" returns image reference \"sha256:34018aef09a62f8b40bdd1d2e1bf6c48f359cab492d51059a09e20745ab02ce2\"" Jan 17 12:20:12.814028 containerd[1544]: time="2025-01-17T12:20:12.814006870Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\"" Jan 17 12:20:13.442899 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2511496782.mount: Deactivated successfully. Jan 17 12:20:14.200985 containerd[1544]: time="2025-01-17T12:20:14.200938220Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:20:14.201942 containerd[1544]: time="2025-01-17T12:20:14.201906849Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.1: active requests=0, bytes read=18185761" Jan 17 12:20:14.202468 containerd[1544]: time="2025-01-17T12:20:14.202437690Z" level=info msg="ImageCreate event name:\"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:20:14.204639 containerd[1544]: time="2025-01-17T12:20:14.204611472Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:20:14.206149 containerd[1544]: time="2025-01-17T12:20:14.205616770Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.1\" with image id \"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\", size \"18182961\" in 1.391584462s" Jan 17 12:20:14.206149 containerd[1544]: time="2025-01-17T12:20:14.205640328Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\" returns image reference \"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\"" Jan 17 12:20:14.206149 containerd[1544]: time="2025-01-17T12:20:14.206077409Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Jan 17 12:20:14.694132 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4206722492.mount: Deactivated successfully. Jan 17 12:20:14.696626 containerd[1544]: time="2025-01-17T12:20:14.696572312Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:20:14.697294 containerd[1544]: time="2025-01-17T12:20:14.697215809Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Jan 17 12:20:14.697294 containerd[1544]: time="2025-01-17T12:20:14.697264383Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:20:14.699223 containerd[1544]: time="2025-01-17T12:20:14.699196599Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:20:14.700232 containerd[1544]: time="2025-01-17T12:20:14.700057956Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 493.963236ms" Jan 17 12:20:14.700232 containerd[1544]: time="2025-01-17T12:20:14.700107295Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Jan 17 12:20:14.700967 containerd[1544]: time="2025-01-17T12:20:14.700714605Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\"" Jan 17 12:20:15.186522 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2885415681.mount: Deactivated successfully. Jan 17 12:20:18.164077 containerd[1544]: time="2025-01-17T12:20:18.164033154Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.15-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:20:18.165080 containerd[1544]: time="2025-01-17T12:20:18.164939855Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.15-0: active requests=0, bytes read=56779973" Jan 17 12:20:18.165467 containerd[1544]: time="2025-01-17T12:20:18.165450322Z" level=info msg="ImageCreate event name:\"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:20:18.167170 containerd[1544]: time="2025-01-17T12:20:18.167154612Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:20:18.167950 containerd[1544]: time="2025-01-17T12:20:18.167934806Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.15-0\" with image id \"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\", repo tag \"registry.k8s.io/etcd:3.5.15-0\", repo digest \"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\", size \"56909194\" in 3.467196551s" Jan 17 12:20:18.168073 containerd[1544]: time="2025-01-17T12:20:18.167995442Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\" returns image reference \"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\"" Jan 17 12:20:18.783993 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 5. Jan 17 12:20:18.790866 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 17 12:20:19.319774 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 17 12:20:19.322620 (kubelet)[2231]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 17 12:20:19.357832 kubelet[2231]: E0117 12:20:19.357798 2231 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 17 12:20:19.360064 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 17 12:20:19.360167 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 17 12:20:20.702526 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 17 12:20:20.711950 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 17 12:20:20.729411 systemd[1]: Reloading requested from client PID 2246 ('systemctl') (unit session-9.scope)... Jan 17 12:20:20.729420 systemd[1]: Reloading... Jan 17 12:20:20.795701 zram_generator::config[2284]: No configuration found. Jan 17 12:20:20.855910 systemd[1]: /etc/systemd/system/coreos-metadata.service:11: Ignoring unknown escape sequences: "echo "COREOS_CUSTOM_PRIVATE_IPV4=$(ip addr show ens192 | grep "inet 10." | grep -Po "inet \K[\d.]+") Jan 17 12:20:20.872444 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jan 17 12:20:20.915596 systemd[1]: Reloading finished in 185 ms. Jan 17 12:20:20.963130 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Jan 17 12:20:20.963200 systemd[1]: kubelet.service: Failed with result 'signal'. Jan 17 12:20:20.963419 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 17 12:20:20.967937 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 17 12:20:21.262061 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 17 12:20:21.267849 (kubelet)[2351]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jan 17 12:20:21.290608 kubelet[2351]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 17 12:20:21.290608 kubelet[2351]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Jan 17 12:20:21.290608 kubelet[2351]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 17 12:20:21.297312 kubelet[2351]: I0117 12:20:21.297286 2351 server.go:206] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 17 12:20:21.543151 kubelet[2351]: I0117 12:20:21.542849 2351 server.go:486] "Kubelet version" kubeletVersion="v1.31.0" Jan 17 12:20:21.543151 kubelet[2351]: I0117 12:20:21.542880 2351 server.go:488] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 17 12:20:21.543151 kubelet[2351]: I0117 12:20:21.543055 2351 server.go:929] "Client rotation is on, will bootstrap in background" Jan 17 12:20:21.647772 kubelet[2351]: I0117 12:20:21.647310 2351 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jan 17 12:20:21.648688 kubelet[2351]: E0117 12:20:21.648213 2351 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://139.178.70.100:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 139.178.70.100:6443: connect: connection refused" logger="UnhandledError" Jan 17 12:20:21.653755 kubelet[2351]: E0117 12:20:21.653710 2351 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Jan 17 12:20:21.653755 kubelet[2351]: I0117 12:20:21.653735 2351 server.go:1403] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Jan 17 12:20:21.657863 kubelet[2351]: I0117 12:20:21.657845 2351 server.go:744] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jan 17 12:20:21.658929 kubelet[2351]: I0117 12:20:21.658913 2351 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Jan 17 12:20:21.659055 kubelet[2351]: I0117 12:20:21.659027 2351 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 17 12:20:21.659184 kubelet[2351]: I0117 12:20:21.659054 2351 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jan 17 12:20:21.659258 kubelet[2351]: I0117 12:20:21.659193 2351 topology_manager.go:138] "Creating topology manager with none policy" Jan 17 12:20:21.659258 kubelet[2351]: I0117 12:20:21.659199 2351 container_manager_linux.go:300] "Creating device plugin manager" Jan 17 12:20:21.659299 kubelet[2351]: I0117 12:20:21.659283 2351 state_mem.go:36] "Initialized new in-memory state store" Jan 17 12:20:21.661041 kubelet[2351]: I0117 12:20:21.660889 2351 kubelet.go:408] "Attempting to sync node with API server" Jan 17 12:20:21.661041 kubelet[2351]: I0117 12:20:21.660904 2351 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 17 12:20:21.661041 kubelet[2351]: I0117 12:20:21.660933 2351 kubelet.go:314] "Adding apiserver pod source" Jan 17 12:20:21.661041 kubelet[2351]: I0117 12:20:21.660947 2351 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 17 12:20:21.665250 kubelet[2351]: W0117 12:20:21.664888 2351 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://139.178.70.100:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 139.178.70.100:6443: connect: connection refused Jan 17 12:20:21.665250 kubelet[2351]: E0117 12:20:21.664929 2351 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://139.178.70.100:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 139.178.70.100:6443: connect: connection refused" logger="UnhandledError" Jan 17 12:20:21.666240 kubelet[2351]: W0117 12:20:21.666218 2351 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://139.178.70.100:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 139.178.70.100:6443: connect: connection refused Jan 17 12:20:21.666309 kubelet[2351]: E0117 12:20:21.666300 2351 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://139.178.70.100:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 139.178.70.100:6443: connect: connection refused" logger="UnhandledError" Jan 17 12:20:21.666403 kubelet[2351]: I0117 12:20:21.666395 2351 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Jan 17 12:20:21.668266 kubelet[2351]: I0117 12:20:21.668256 2351 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jan 17 12:20:21.668898 kubelet[2351]: W0117 12:20:21.668889 2351 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Jan 17 12:20:21.671101 kubelet[2351]: I0117 12:20:21.671082 2351 server.go:1269] "Started kubelet" Jan 17 12:20:21.674690 kubelet[2351]: I0117 12:20:21.674118 2351 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Jan 17 12:20:21.675487 kubelet[2351]: I0117 12:20:21.675478 2351 server.go:460] "Adding debug handlers to kubelet server" Jan 17 12:20:21.676223 kubelet[2351]: I0117 12:20:21.676187 2351 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 17 12:20:21.676439 kubelet[2351]: I0117 12:20:21.676424 2351 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 17 12:20:21.676903 kubelet[2351]: I0117 12:20:21.676896 2351 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 17 12:20:21.680840 kubelet[2351]: I0117 12:20:21.680827 2351 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Jan 17 12:20:21.681597 kubelet[2351]: E0117 12:20:21.677942 2351 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://139.178.70.100:6443/api/v1/namespaces/default/events\": dial tcp 139.178.70.100:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.181b7a2e8f335bc5 default 0 0001-01-01 00:00:00 +0000 UTC <nil> <nil> map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-01-17 12:20:21.671050181 +0000 UTC m=+0.401426802,LastTimestamp:2025-01-17 12:20:21.671050181 +0000 UTC m=+0.401426802,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Jan 17 12:20:21.682148 kubelet[2351]: E0117 12:20:21.681801 2351 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"localhost\" not found" Jan 17 12:20:21.682148 kubelet[2351]: I0117 12:20:21.681833 2351 volume_manager.go:289] "Starting Kubelet Volume Manager" Jan 17 12:20:21.682148 kubelet[2351]: I0117 12:20:21.681949 2351 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Jan 17 12:20:21.683885 kubelet[2351]: I0117 12:20:21.683870 2351 reconciler.go:26] "Reconciler: start to sync state" Jan 17 12:20:21.684323 kubelet[2351]: E0117 12:20:21.684295 2351 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://139.178.70.100:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 139.178.70.100:6443: connect: connection refused" interval="200ms" Jan 17 12:20:21.685053 kubelet[2351]: W0117 12:20:21.685031 2351 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://139.178.70.100:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 139.178.70.100:6443: connect: connection refused Jan 17 12:20:21.685089 kubelet[2351]: E0117 12:20:21.685069 2351 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://139.178.70.100:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 139.178.70.100:6443: connect: connection refused" logger="UnhandledError" Jan 17 12:20:21.685579 kubelet[2351]: I0117 12:20:21.685561 2351 factory.go:221] Registration of the containerd container factory successfully Jan 17 12:20:21.685579 kubelet[2351]: I0117 12:20:21.685574 2351 factory.go:221] Registration of the systemd container factory successfully Jan 17 12:20:21.686192 kubelet[2351]: I0117 12:20:21.686174 2351 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jan 17 12:20:21.691373 kubelet[2351]: I0117 12:20:21.691346 2351 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jan 17 12:20:21.692351 kubelet[2351]: I0117 12:20:21.692150 2351 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jan 17 12:20:21.692351 kubelet[2351]: I0117 12:20:21.692170 2351 status_manager.go:217] "Starting to sync pod status with apiserver" Jan 17 12:20:21.692351 kubelet[2351]: I0117 12:20:21.692182 2351 kubelet.go:2321] "Starting kubelet main sync loop" Jan 17 12:20:21.692351 kubelet[2351]: E0117 12:20:21.692204 2351 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 17 12:20:21.696343 kubelet[2351]: W0117 12:20:21.696315 2351 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://139.178.70.100:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 139.178.70.100:6443: connect: connection refused Jan 17 12:20:21.696586 kubelet[2351]: E0117 12:20:21.696573 2351 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://139.178.70.100:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 139.178.70.100:6443: connect: connection refused" logger="UnhandledError" Jan 17 12:20:21.697888 kubelet[2351]: E0117 12:20:21.697021 2351 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jan 17 12:20:21.708358 kubelet[2351]: I0117 12:20:21.708339 2351 cpu_manager.go:214] "Starting CPU manager" policy="none" Jan 17 12:20:21.708358 kubelet[2351]: I0117 12:20:21.708350 2351 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Jan 17 12:20:21.708358 kubelet[2351]: I0117 12:20:21.708363 2351 state_mem.go:36] "Initialized new in-memory state store" Jan 17 12:20:21.714128 kubelet[2351]: I0117 12:20:21.714108 2351 policy_none.go:49] "None policy: Start" Jan 17 12:20:21.714471 kubelet[2351]: I0117 12:20:21.714455 2351 memory_manager.go:170] "Starting memorymanager" policy="None" Jan 17 12:20:21.714471 kubelet[2351]: I0117 12:20:21.714470 2351 state_mem.go:35] "Initializing new in-memory state store" Jan 17 12:20:21.729527 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Jan 17 12:20:21.741211 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Jan 17 12:20:21.743758 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Jan 17 12:20:21.757246 kubelet[2351]: I0117 12:20:21.757230 2351 manager.go:510] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jan 17 12:20:21.757551 kubelet[2351]: I0117 12:20:21.757355 2351 eviction_manager.go:189] "Eviction manager: starting control loop" Jan 17 12:20:21.757551 kubelet[2351]: I0117 12:20:21.757365 2351 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 17 12:20:21.757551 kubelet[2351]: I0117 12:20:21.757510 2351 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 17 12:20:21.758904 kubelet[2351]: E0117 12:20:21.758886 2351 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Jan 17 12:20:21.799713 systemd[1]: Created slice kubepods-burstable-pod6c53cb858e6b48feb804002d9332726b.slice - libcontainer container kubepods-burstable-pod6c53cb858e6b48feb804002d9332726b.slice. Jan 17 12:20:21.813235 systemd[1]: Created slice kubepods-burstable-podfa5289f3c0ba7f1736282e713231ffc5.slice - libcontainer container kubepods-burstable-podfa5289f3c0ba7f1736282e713231ffc5.slice. Jan 17 12:20:21.822810 systemd[1]: Created slice kubepods-burstable-podc988230cd0d49eebfaffbefbe8c74a10.slice - libcontainer container kubepods-burstable-podc988230cd0d49eebfaffbefbe8c74a10.slice. Jan 17 12:20:21.858926 kubelet[2351]: I0117 12:20:21.858875 2351 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Jan 17 12:20:21.859125 kubelet[2351]: E0117 12:20:21.859110 2351 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://139.178.70.100:6443/api/v1/nodes\": dial tcp 139.178.70.100:6443: connect: connection refused" node="localhost" Jan 17 12:20:21.884710 kubelet[2351]: I0117 12:20:21.884491 2351 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/6c53cb858e6b48feb804002d9332726b-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"6c53cb858e6b48feb804002d9332726b\") " pod="kube-system/kube-apiserver-localhost" Jan 17 12:20:21.884710 kubelet[2351]: I0117 12:20:21.884516 2351 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/6c53cb858e6b48feb804002d9332726b-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"6c53cb858e6b48feb804002d9332726b\") " pod="kube-system/kube-apiserver-localhost" Jan 17 12:20:21.884710 kubelet[2351]: I0117 12:20:21.884526 2351 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/fa5289f3c0ba7f1736282e713231ffc5-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"fa5289f3c0ba7f1736282e713231ffc5\") " pod="kube-system/kube-controller-manager-localhost" Jan 17 12:20:21.884710 kubelet[2351]: I0117 12:20:21.884536 2351 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/fa5289f3c0ba7f1736282e713231ffc5-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"fa5289f3c0ba7f1736282e713231ffc5\") " pod="kube-system/kube-controller-manager-localhost" Jan 17 12:20:21.884710 kubelet[2351]: I0117 12:20:21.884544 2351 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/fa5289f3c0ba7f1736282e713231ffc5-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"fa5289f3c0ba7f1736282e713231ffc5\") " pod="kube-system/kube-controller-manager-localhost" Jan 17 12:20:21.884861 kubelet[2351]: I0117 12:20:21.884553 2351 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/c988230cd0d49eebfaffbefbe8c74a10-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"c988230cd0d49eebfaffbefbe8c74a10\") " pod="kube-system/kube-scheduler-localhost" Jan 17 12:20:21.884861 kubelet[2351]: I0117 12:20:21.884560 2351 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/6c53cb858e6b48feb804002d9332726b-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"6c53cb858e6b48feb804002d9332726b\") " pod="kube-system/kube-apiserver-localhost" Jan 17 12:20:21.884861 kubelet[2351]: I0117 12:20:21.884568 2351 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/fa5289f3c0ba7f1736282e713231ffc5-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"fa5289f3c0ba7f1736282e713231ffc5\") " pod="kube-system/kube-controller-manager-localhost" Jan 17 12:20:21.884861 kubelet[2351]: I0117 12:20:21.884575 2351 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/fa5289f3c0ba7f1736282e713231ffc5-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"fa5289f3c0ba7f1736282e713231ffc5\") " pod="kube-system/kube-controller-manager-localhost" Jan 17 12:20:21.884861 kubelet[2351]: E0117 12:20:21.884787 2351 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://139.178.70.100:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 139.178.70.100:6443: connect: connection refused" interval="400ms" Jan 17 12:20:22.060486 kubelet[2351]: I0117 12:20:22.060415 2351 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Jan 17 12:20:22.060908 kubelet[2351]: E0117 12:20:22.060664 2351 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://139.178.70.100:6443/api/v1/nodes\": dial tcp 139.178.70.100:6443: connect: connection refused" node="localhost" Jan 17 12:20:22.112695 containerd[1544]: time="2025-01-17T12:20:22.112651324Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:6c53cb858e6b48feb804002d9332726b,Namespace:kube-system,Attempt:0,}" Jan 17 12:20:22.126354 containerd[1544]: time="2025-01-17T12:20:22.126292298Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:fa5289f3c0ba7f1736282e713231ffc5,Namespace:kube-system,Attempt:0,}" Jan 17 12:20:22.126534 containerd[1544]: time="2025-01-17T12:20:22.126292312Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:c988230cd0d49eebfaffbefbe8c74a10,Namespace:kube-system,Attempt:0,}" Jan 17 12:20:22.285499 kubelet[2351]: E0117 12:20:22.285465 2351 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://139.178.70.100:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 139.178.70.100:6443: connect: connection refused" interval="800ms" Jan 17 12:20:22.462074 kubelet[2351]: I0117 12:20:22.461785 2351 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Jan 17 12:20:22.462345 kubelet[2351]: E0117 12:20:22.462026 2351 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://139.178.70.100:6443/api/v1/nodes\": dial tcp 139.178.70.100:6443: connect: connection refused" node="localhost" Jan 17 12:20:22.620385 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2546826895.mount: Deactivated successfully. Jan 17 12:20:22.624692 containerd[1544]: time="2025-01-17T12:20:22.624089913Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 17 12:20:22.626104 containerd[1544]: time="2025-01-17T12:20:22.626082707Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312056" Jan 17 12:20:22.626779 containerd[1544]: time="2025-01-17T12:20:22.626766060Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 17 12:20:22.628648 containerd[1544]: time="2025-01-17T12:20:22.628617752Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 17 12:20:22.628730 containerd[1544]: time="2025-01-17T12:20:22.628715306Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Jan 17 12:20:22.630522 containerd[1544]: time="2025-01-17T12:20:22.630507592Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 17 12:20:22.631728 containerd[1544]: time="2025-01-17T12:20:22.630584687Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Jan 17 12:20:22.631855 containerd[1544]: time="2025-01-17T12:20:22.631838663Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 505.495606ms" Jan 17 12:20:22.633166 containerd[1544]: time="2025-01-17T12:20:22.633150807Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 17 12:20:22.633436 containerd[1544]: time="2025-01-17T12:20:22.633424594Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 520.688607ms" Jan 17 12:20:22.634813 containerd[1544]: time="2025-01-17T12:20:22.634797392Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 508.322643ms" Jan 17 12:20:22.683997 kubelet[2351]: W0117 12:20:22.683886 2351 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://139.178.70.100:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 139.178.70.100:6443: connect: connection refused Jan 17 12:20:22.685481 kubelet[2351]: E0117 12:20:22.685464 2351 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://139.178.70.100:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 139.178.70.100:6443: connect: connection refused" logger="UnhandledError" Jan 17 12:20:22.790698 containerd[1544]: time="2025-01-17T12:20:22.790244589Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 17 12:20:22.790698 containerd[1544]: time="2025-01-17T12:20:22.790289222Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 17 12:20:22.790698 containerd[1544]: time="2025-01-17T12:20:22.790319391Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 12:20:22.790871 containerd[1544]: time="2025-01-17T12:20:22.790668914Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 12:20:22.796006 containerd[1544]: time="2025-01-17T12:20:22.795865582Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 17 12:20:22.796006 containerd[1544]: time="2025-01-17T12:20:22.795892162Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 17 12:20:22.796006 containerd[1544]: time="2025-01-17T12:20:22.795899242Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 12:20:22.796006 containerd[1544]: time="2025-01-17T12:20:22.795947339Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 12:20:22.800240 containerd[1544]: time="2025-01-17T12:20:22.800207256Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 17 12:20:22.800691 containerd[1544]: time="2025-01-17T12:20:22.800664040Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 17 12:20:22.800752 containerd[1544]: time="2025-01-17T12:20:22.800740183Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 12:20:22.800873 containerd[1544]: time="2025-01-17T12:20:22.800846270Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 12:20:22.815808 systemd[1]: Started cri-containerd-c4fd72784ab4aeaa057df83b3ef28e177a5356f08de9dd8fd046ed1282821721.scope - libcontainer container c4fd72784ab4aeaa057df83b3ef28e177a5356f08de9dd8fd046ed1282821721. Jan 17 12:20:22.818187 systemd[1]: Started cri-containerd-7d1b3bc176a7018dc0f70320b49d11976f1d04703499690312d7ea261a68b493.scope - libcontainer container 7d1b3bc176a7018dc0f70320b49d11976f1d04703499690312d7ea261a68b493. Jan 17 12:20:22.822014 systemd[1]: Started cri-containerd-72a4f26efc1b9e15f6a17fd90c192878b636bc14a7593eb899776966c31fdd59.scope - libcontainer container 72a4f26efc1b9e15f6a17fd90c192878b636bc14a7593eb899776966c31fdd59. Jan 17 12:20:22.872777 containerd[1544]: time="2025-01-17T12:20:22.872564055Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:6c53cb858e6b48feb804002d9332726b,Namespace:kube-system,Attempt:0,} returns sandbox id \"7d1b3bc176a7018dc0f70320b49d11976f1d04703499690312d7ea261a68b493\"" Jan 17 12:20:22.885000 containerd[1544]: time="2025-01-17T12:20:22.884968151Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:fa5289f3c0ba7f1736282e713231ffc5,Namespace:kube-system,Attempt:0,} returns sandbox id \"c4fd72784ab4aeaa057df83b3ef28e177a5356f08de9dd8fd046ed1282821721\"" Jan 17 12:20:22.885378 containerd[1544]: time="2025-01-17T12:20:22.885292938Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:c988230cd0d49eebfaffbefbe8c74a10,Namespace:kube-system,Attempt:0,} returns sandbox id \"72a4f26efc1b9e15f6a17fd90c192878b636bc14a7593eb899776966c31fdd59\"" Jan 17 12:20:22.889278 containerd[1544]: time="2025-01-17T12:20:22.889257746Z" level=info msg="CreateContainer within sandbox \"c4fd72784ab4aeaa057df83b3ef28e177a5356f08de9dd8fd046ed1282821721\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Jan 17 12:20:22.889367 containerd[1544]: time="2025-01-17T12:20:22.889347544Z" level=info msg="CreateContainer within sandbox \"7d1b3bc176a7018dc0f70320b49d11976f1d04703499690312d7ea261a68b493\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Jan 17 12:20:22.889441 containerd[1544]: time="2025-01-17T12:20:22.889424980Z" level=info msg="CreateContainer within sandbox \"72a4f26efc1b9e15f6a17fd90c192878b636bc14a7593eb899776966c31fdd59\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Jan 17 12:20:23.034389 kubelet[2351]: W0117 12:20:23.034343 2351 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://139.178.70.100:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 139.178.70.100:6443: connect: connection refused Jan 17 12:20:23.034389 kubelet[2351]: E0117 12:20:23.034388 2351 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://139.178.70.100:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 139.178.70.100:6443: connect: connection refused" logger="UnhandledError" Jan 17 12:20:23.085116 kubelet[2351]: W0117 12:20:23.085076 2351 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://139.178.70.100:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 139.178.70.100:6443: connect: connection refused Jan 17 12:20:23.085185 kubelet[2351]: E0117 12:20:23.085122 2351 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://139.178.70.100:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 139.178.70.100:6443: connect: connection refused" logger="UnhandledError" Jan 17 12:20:23.086272 kubelet[2351]: E0117 12:20:23.086249 2351 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://139.178.70.100:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 139.178.70.100:6443: connect: connection refused" interval="1.6s" Jan 17 12:20:23.141959 kubelet[2351]: W0117 12:20:23.141882 2351 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://139.178.70.100:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 139.178.70.100:6443: connect: connection refused Jan 17 12:20:23.141959 kubelet[2351]: E0117 12:20:23.141939 2351 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://139.178.70.100:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 139.178.70.100:6443: connect: connection refused" logger="UnhandledError" Jan 17 12:20:23.263772 kubelet[2351]: I0117 12:20:23.263512 2351 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Jan 17 12:20:23.263772 kubelet[2351]: E0117 12:20:23.263755 2351 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://139.178.70.100:6443/api/v1/nodes\": dial tcp 139.178.70.100:6443: connect: connection refused" node="localhost" Jan 17 12:20:23.273571 containerd[1544]: time="2025-01-17T12:20:23.273471851Z" level=info msg="CreateContainer within sandbox \"72a4f26efc1b9e15f6a17fd90c192878b636bc14a7593eb899776966c31fdd59\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"d8e0b13a9afdc5e4d3c3b3554a96ed845d9f3dcbc73178e6ea796272040c545d\"" Jan 17 12:20:23.276490 containerd[1544]: time="2025-01-17T12:20:23.274025668Z" level=info msg="StartContainer for \"d8e0b13a9afdc5e4d3c3b3554a96ed845d9f3dcbc73178e6ea796272040c545d\"" Jan 17 12:20:23.278449 containerd[1544]: time="2025-01-17T12:20:23.278361451Z" level=info msg="CreateContainer within sandbox \"c4fd72784ab4aeaa057df83b3ef28e177a5356f08de9dd8fd046ed1282821721\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"0db50dd0a92a2e12f3f6e6a17525a339cf95e86b2eeb5bcdac77512a8441679f\"" Jan 17 12:20:23.278896 containerd[1544]: time="2025-01-17T12:20:23.278825310Z" level=info msg="StartContainer for \"0db50dd0a92a2e12f3f6e6a17525a339cf95e86b2eeb5bcdac77512a8441679f\"" Jan 17 12:20:23.282461 containerd[1544]: time="2025-01-17T12:20:23.282410938Z" level=info msg="CreateContainer within sandbox \"7d1b3bc176a7018dc0f70320b49d11976f1d04703499690312d7ea261a68b493\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"07ef263ca0edf1b82ab0d8ccc95e7250b3c26977e0b3952521544ad564b27977\"" Jan 17 12:20:23.282735 containerd[1544]: time="2025-01-17T12:20:23.282667735Z" level=info msg="StartContainer for \"07ef263ca0edf1b82ab0d8ccc95e7250b3c26977e0b3952521544ad564b27977\"" Jan 17 12:20:23.301800 systemd[1]: Started cri-containerd-0db50dd0a92a2e12f3f6e6a17525a339cf95e86b2eeb5bcdac77512a8441679f.scope - libcontainer container 0db50dd0a92a2e12f3f6e6a17525a339cf95e86b2eeb5bcdac77512a8441679f. Jan 17 12:20:23.305461 systemd[1]: Started cri-containerd-07ef263ca0edf1b82ab0d8ccc95e7250b3c26977e0b3952521544ad564b27977.scope - libcontainer container 07ef263ca0edf1b82ab0d8ccc95e7250b3c26977e0b3952521544ad564b27977. Jan 17 12:20:23.306186 systemd[1]: Started cri-containerd-d8e0b13a9afdc5e4d3c3b3554a96ed845d9f3dcbc73178e6ea796272040c545d.scope - libcontainer container d8e0b13a9afdc5e4d3c3b3554a96ed845d9f3dcbc73178e6ea796272040c545d. Jan 17 12:20:23.341940 containerd[1544]: time="2025-01-17T12:20:23.341887094Z" level=info msg="StartContainer for \"0db50dd0a92a2e12f3f6e6a17525a339cf95e86b2eeb5bcdac77512a8441679f\" returns successfully" Jan 17 12:20:23.359772 containerd[1544]: time="2025-01-17T12:20:23.359502153Z" level=info msg="StartContainer for \"d8e0b13a9afdc5e4d3c3b3554a96ed845d9f3dcbc73178e6ea796272040c545d\" returns successfully" Jan 17 12:20:23.359772 containerd[1544]: time="2025-01-17T12:20:23.359557721Z" level=info msg="StartContainer for \"07ef263ca0edf1b82ab0d8ccc95e7250b3c26977e0b3952521544ad564b27977\" returns successfully" Jan 17 12:20:23.655274 kubelet[2351]: E0117 12:20:23.655013 2351 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://139.178.70.100:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 139.178.70.100:6443: connect: connection refused" logger="UnhandledError" Jan 17 12:20:24.866695 kubelet[2351]: I0117 12:20:24.865424 2351 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Jan 17 12:20:24.984405 kubelet[2351]: E0117 12:20:24.984372 2351 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Jan 17 12:20:25.061225 kubelet[2351]: I0117 12:20:25.061199 2351 kubelet_node_status.go:75] "Successfully registered node" node="localhost" Jan 17 12:20:25.061225 kubelet[2351]: E0117 12:20:25.061224 2351 kubelet_node_status.go:535] "Error updating node status, will retry" err="error getting node \"localhost\": node \"localhost\" not found" Jan 17 12:20:25.668982 kubelet[2351]: I0117 12:20:25.668910 2351 apiserver.go:52] "Watching apiserver" Jan 17 12:20:25.683068 kubelet[2351]: I0117 12:20:25.683055 2351 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Jan 17 12:20:26.943480 systemd[1]: Reloading requested from client PID 2618 ('systemctl') (unit session-9.scope)... Jan 17 12:20:26.943491 systemd[1]: Reloading... Jan 17 12:20:27.006707 zram_generator::config[2656]: No configuration found. Jan 17 12:20:27.073705 systemd[1]: /etc/systemd/system/coreos-metadata.service:11: Ignoring unknown escape sequences: "echo "COREOS_CUSTOM_PRIVATE_IPV4=$(ip addr show ens192 | grep "inet 10." | grep -Po "inet \K[\d.]+") Jan 17 12:20:27.089485 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jan 17 12:20:27.142194 systemd[1]: Reloading finished in 198 ms. Jan 17 12:20:27.167314 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Jan 17 12:20:27.180473 systemd[1]: kubelet.service: Deactivated successfully. Jan 17 12:20:27.180624 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 17 12:20:27.184994 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 17 12:20:27.503488 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 17 12:20:27.512955 (kubelet)[2723]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jan 17 12:20:27.543039 kubelet[2723]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 17 12:20:27.543039 kubelet[2723]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Jan 17 12:20:27.543039 kubelet[2723]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 17 12:20:27.544606 kubelet[2723]: I0117 12:20:27.544420 2723 server.go:206] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 17 12:20:27.548698 kubelet[2723]: I0117 12:20:27.548151 2723 server.go:486] "Kubelet version" kubeletVersion="v1.31.0" Jan 17 12:20:27.548698 kubelet[2723]: I0117 12:20:27.548163 2723 server.go:488] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 17 12:20:27.548698 kubelet[2723]: I0117 12:20:27.548286 2723 server.go:929] "Client rotation is on, will bootstrap in background" Jan 17 12:20:27.549025 kubelet[2723]: I0117 12:20:27.549012 2723 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Jan 17 12:20:27.551206 kubelet[2723]: I0117 12:20:27.551132 2723 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jan 17 12:20:27.556289 kubelet[2723]: E0117 12:20:27.556272 2723 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Jan 17 12:20:27.556289 kubelet[2723]: I0117 12:20:27.556289 2723 server.go:1403] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Jan 17 12:20:27.558683 kubelet[2723]: I0117 12:20:27.558365 2723 server.go:744] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jan 17 12:20:27.558683 kubelet[2723]: I0117 12:20:27.558426 2723 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Jan 17 12:20:27.558683 kubelet[2723]: I0117 12:20:27.558481 2723 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 17 12:20:27.558683 kubelet[2723]: I0117 12:20:27.558492 2723 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jan 17 12:20:27.558801 kubelet[2723]: I0117 12:20:27.558633 2723 topology_manager.go:138] "Creating topology manager with none policy" Jan 17 12:20:27.558801 kubelet[2723]: I0117 12:20:27.558639 2723 container_manager_linux.go:300] "Creating device plugin manager" Jan 17 12:20:27.558801 kubelet[2723]: I0117 12:20:27.558656 2723 state_mem.go:36] "Initialized new in-memory state store" Jan 17 12:20:27.558801 kubelet[2723]: I0117 12:20:27.558750 2723 kubelet.go:408] "Attempting to sync node with API server" Jan 17 12:20:27.558801 kubelet[2723]: I0117 12:20:27.558758 2723 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 17 12:20:27.558801 kubelet[2723]: I0117 12:20:27.558774 2723 kubelet.go:314] "Adding apiserver pod source" Jan 17 12:20:27.558801 kubelet[2723]: I0117 12:20:27.558783 2723 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 17 12:20:27.560391 kubelet[2723]: I0117 12:20:27.560373 2723 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Jan 17 12:20:27.560698 kubelet[2723]: I0117 12:20:27.560688 2723 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jan 17 12:20:27.560997 kubelet[2723]: I0117 12:20:27.560981 2723 server.go:1269] "Started kubelet" Jan 17 12:20:27.570390 kubelet[2723]: I0117 12:20:27.570376 2723 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 17 12:20:27.571064 kubelet[2723]: E0117 12:20:27.571053 2723 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jan 17 12:20:27.573171 kubelet[2723]: I0117 12:20:27.572619 2723 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Jan 17 12:20:27.573269 kubelet[2723]: I0117 12:20:27.573262 2723 server.go:460] "Adding debug handlers to kubelet server" Jan 17 12:20:27.573934 kubelet[2723]: I0117 12:20:27.573911 2723 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 17 12:20:27.574066 kubelet[2723]: I0117 12:20:27.574059 2723 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 17 12:20:27.574395 kubelet[2723]: I0117 12:20:27.574387 2723 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Jan 17 12:20:27.575377 kubelet[2723]: I0117 12:20:27.575369 2723 volume_manager.go:289] "Starting Kubelet Volume Manager" Jan 17 12:20:27.575661 kubelet[2723]: I0117 12:20:27.575654 2723 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Jan 17 12:20:27.575775 kubelet[2723]: I0117 12:20:27.575770 2723 reconciler.go:26] "Reconciler: start to sync state" Jan 17 12:20:27.576168 kubelet[2723]: I0117 12:20:27.576160 2723 factory.go:221] Registration of the systemd container factory successfully Jan 17 12:20:27.576261 kubelet[2723]: I0117 12:20:27.576251 2723 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jan 17 12:20:27.582689 kubelet[2723]: I0117 12:20:27.581935 2723 factory.go:221] Registration of the containerd container factory successfully Jan 17 12:20:27.582689 kubelet[2723]: I0117 12:20:27.581942 2723 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jan 17 12:20:27.583410 kubelet[2723]: I0117 12:20:27.583396 2723 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jan 17 12:20:27.583440 kubelet[2723]: I0117 12:20:27.583413 2723 status_manager.go:217] "Starting to sync pod status with apiserver" Jan 17 12:20:27.583440 kubelet[2723]: I0117 12:20:27.583425 2723 kubelet.go:2321] "Starting kubelet main sync loop" Jan 17 12:20:27.583478 kubelet[2723]: E0117 12:20:27.583447 2723 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 17 12:20:27.618922 kubelet[2723]: I0117 12:20:27.618903 2723 cpu_manager.go:214] "Starting CPU manager" policy="none" Jan 17 12:20:27.619062 kubelet[2723]: I0117 12:20:27.619055 2723 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Jan 17 12:20:27.619115 kubelet[2723]: I0117 12:20:27.619109 2723 state_mem.go:36] "Initialized new in-memory state store" Jan 17 12:20:27.619275 kubelet[2723]: I0117 12:20:27.619268 2723 state_mem.go:88] "Updated default CPUSet" cpuSet="" Jan 17 12:20:27.619323 kubelet[2723]: I0117 12:20:27.619310 2723 state_mem.go:96] "Updated CPUSet assignments" assignments={} Jan 17 12:20:27.619363 kubelet[2723]: I0117 12:20:27.619360 2723 policy_none.go:49] "None policy: Start" Jan 17 12:20:27.619751 kubelet[2723]: I0117 12:20:27.619743 2723 memory_manager.go:170] "Starting memorymanager" policy="None" Jan 17 12:20:27.619810 kubelet[2723]: I0117 12:20:27.619804 2723 state_mem.go:35] "Initializing new in-memory state store" Jan 17 12:20:27.619947 kubelet[2723]: I0117 12:20:27.619936 2723 state_mem.go:75] "Updated machine memory state" Jan 17 12:20:27.622387 kubelet[2723]: I0117 12:20:27.622379 2723 manager.go:510] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jan 17 12:20:27.623436 kubelet[2723]: I0117 12:20:27.623428 2723 eviction_manager.go:189] "Eviction manager: starting control loop" Jan 17 12:20:27.623911 kubelet[2723]: I0117 12:20:27.623892 2723 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 17 12:20:27.624071 kubelet[2723]: I0117 12:20:27.624054 2723 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 17 12:20:27.690243 kubelet[2723]: E0117 12:20:27.690220 2723 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Jan 17 12:20:27.690852 kubelet[2723]: E0117 12:20:27.690829 2723 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost" Jan 17 12:20:27.727047 kubelet[2723]: I0117 12:20:27.727030 2723 kubelet_node_status.go:72] "Attempting to register node" node="localhost" Jan 17 12:20:27.732189 kubelet[2723]: I0117 12:20:27.732171 2723 kubelet_node_status.go:111] "Node was previously registered" node="localhost" Jan 17 12:20:27.732357 kubelet[2723]: I0117 12:20:27.732338 2723 kubelet_node_status.go:75] "Successfully registered node" node="localhost" Jan 17 12:20:27.776222 kubelet[2723]: I0117 12:20:27.776148 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/6c53cb858e6b48feb804002d9332726b-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"6c53cb858e6b48feb804002d9332726b\") " pod="kube-system/kube-apiserver-localhost" Jan 17 12:20:27.776652 kubelet[2723]: I0117 12:20:27.776347 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/6c53cb858e6b48feb804002d9332726b-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"6c53cb858e6b48feb804002d9332726b\") " pod="kube-system/kube-apiserver-localhost" Jan 17 12:20:27.776652 kubelet[2723]: I0117 12:20:27.776367 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/fa5289f3c0ba7f1736282e713231ffc5-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"fa5289f3c0ba7f1736282e713231ffc5\") " pod="kube-system/kube-controller-manager-localhost" Jan 17 12:20:27.776652 kubelet[2723]: I0117 12:20:27.776381 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/fa5289f3c0ba7f1736282e713231ffc5-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"fa5289f3c0ba7f1736282e713231ffc5\") " pod="kube-system/kube-controller-manager-localhost" Jan 17 12:20:27.776652 kubelet[2723]: I0117 12:20:27.776395 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/fa5289f3c0ba7f1736282e713231ffc5-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"fa5289f3c0ba7f1736282e713231ffc5\") " pod="kube-system/kube-controller-manager-localhost" Jan 17 12:20:27.776652 kubelet[2723]: I0117 12:20:27.776407 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/fa5289f3c0ba7f1736282e713231ffc5-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"fa5289f3c0ba7f1736282e713231ffc5\") " pod="kube-system/kube-controller-manager-localhost" Jan 17 12:20:27.776811 kubelet[2723]: I0117 12:20:27.776418 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/fa5289f3c0ba7f1736282e713231ffc5-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"fa5289f3c0ba7f1736282e713231ffc5\") " pod="kube-system/kube-controller-manager-localhost" Jan 17 12:20:27.776811 kubelet[2723]: I0117 12:20:27.776605 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/c988230cd0d49eebfaffbefbe8c74a10-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"c988230cd0d49eebfaffbefbe8c74a10\") " pod="kube-system/kube-scheduler-localhost" Jan 17 12:20:27.776811 kubelet[2723]: I0117 12:20:27.776618 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/6c53cb858e6b48feb804002d9332726b-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"6c53cb858e6b48feb804002d9332726b\") " pod="kube-system/kube-apiserver-localhost" Jan 17 12:20:28.559909 kubelet[2723]: I0117 12:20:28.559887 2723 apiserver.go:52] "Watching apiserver" Jan 17 12:20:28.576778 kubelet[2723]: I0117 12:20:28.576744 2723 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Jan 17 12:20:28.641649 kubelet[2723]: I0117 12:20:28.641536 2723 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.641524978 podStartE2EDuration="1.641524978s" podCreationTimestamp="2025-01-17 12:20:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-17 12:20:28.63095262 +0000 UTC m=+1.114425204" watchObservedRunningTime="2025-01-17 12:20:28.641524978 +0000 UTC m=+1.124997562" Jan 17 12:20:28.641877 kubelet[2723]: I0117 12:20:28.641776 2723 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=2.641769686 podStartE2EDuration="2.641769686s" podCreationTimestamp="2025-01-17 12:20:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-17 12:20:28.641463807 +0000 UTC m=+1.124936400" watchObservedRunningTime="2025-01-17 12:20:28.641769686 +0000 UTC m=+1.125242272" Jan 17 12:20:28.666998 kubelet[2723]: I0117 12:20:28.666735 2723 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=2.666722049 podStartE2EDuration="2.666722049s" podCreationTimestamp="2025-01-17 12:20:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-17 12:20:28.66655642 +0000 UTC m=+1.150029007" watchObservedRunningTime="2025-01-17 12:20:28.666722049 +0000 UTC m=+1.150194630" Jan 17 12:20:31.925560 sudo[1850]: pam_unix(sudo:session): session closed for user root Jan 17 12:20:31.927240 sshd[1847]: pam_unix(sshd:session): session closed for user core Jan 17 12:20:31.928829 systemd[1]: sshd@6-139.178.70.100:22-147.75.109.163:55068.service: Deactivated successfully. Jan 17 12:20:31.929905 systemd[1]: session-9.scope: Deactivated successfully. Jan 17 12:20:31.930014 systemd[1]: session-9.scope: Consumed 3.109s CPU time, 142.1M memory peak, 0B memory swap peak. Jan 17 12:20:31.930750 systemd-logind[1522]: Session 9 logged out. Waiting for processes to exit. Jan 17 12:20:31.931332 systemd-logind[1522]: Removed session 9. Jan 17 12:20:32.523757 kubelet[2723]: I0117 12:20:32.523658 2723 kuberuntime_manager.go:1633] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Jan 17 12:20:32.524010 containerd[1544]: time="2025-01-17T12:20:32.523944133Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Jan 17 12:20:32.524136 kubelet[2723]: I0117 12:20:32.524048 2723 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Jan 17 12:20:33.115904 systemd[1]: Created slice kubepods-besteffort-pod748f4fc1_6b19_4928_bff2_5fd3a7b6e50e.slice - libcontainer container kubepods-besteffort-pod748f4fc1_6b19_4928_bff2_5fd3a7b6e50e.slice. Jan 17 12:20:33.208653 kubelet[2723]: I0117 12:20:33.208549 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdl4g\" (UniqueName: \"kubernetes.io/projected/748f4fc1-6b19-4928-bff2-5fd3a7b6e50e-kube-api-access-fdl4g\") pod \"kube-proxy-7jz56\" (UID: \"748f4fc1-6b19-4928-bff2-5fd3a7b6e50e\") " pod="kube-system/kube-proxy-7jz56" Jan 17 12:20:33.208653 kubelet[2723]: I0117 12:20:33.208579 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/748f4fc1-6b19-4928-bff2-5fd3a7b6e50e-kube-proxy\") pod \"kube-proxy-7jz56\" (UID: \"748f4fc1-6b19-4928-bff2-5fd3a7b6e50e\") " pod="kube-system/kube-proxy-7jz56" Jan 17 12:20:33.208653 kubelet[2723]: I0117 12:20:33.208592 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/748f4fc1-6b19-4928-bff2-5fd3a7b6e50e-xtables-lock\") pod \"kube-proxy-7jz56\" (UID: \"748f4fc1-6b19-4928-bff2-5fd3a7b6e50e\") " pod="kube-system/kube-proxy-7jz56" Jan 17 12:20:33.208653 kubelet[2723]: I0117 12:20:33.208600 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/748f4fc1-6b19-4928-bff2-5fd3a7b6e50e-lib-modules\") pod \"kube-proxy-7jz56\" (UID: \"748f4fc1-6b19-4928-bff2-5fd3a7b6e50e\") " pod="kube-system/kube-proxy-7jz56" Jan 17 12:20:33.422308 containerd[1544]: time="2025-01-17T12:20:33.422103009Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-7jz56,Uid:748f4fc1-6b19-4928-bff2-5fd3a7b6e50e,Namespace:kube-system,Attempt:0,}" Jan 17 12:20:33.444090 systemd[1]: Created slice kubepods-besteffort-podef1d5a58_97f0_4379_b0e0_cae489254453.slice - libcontainer container kubepods-besteffort-podef1d5a58_97f0_4379_b0e0_cae489254453.slice. Jan 17 12:20:33.457648 containerd[1544]: time="2025-01-17T12:20:33.457495225Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 17 12:20:33.457648 containerd[1544]: time="2025-01-17T12:20:33.457529351Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 17 12:20:33.457648 containerd[1544]: time="2025-01-17T12:20:33.457537399Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 12:20:33.457648 containerd[1544]: time="2025-01-17T12:20:33.457583323Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 12:20:33.471776 systemd[1]: Started cri-containerd-5dec8a8412931ceca2bd60fc464051fb4cf44577e5a9a956f0f9464ce4332585.scope - libcontainer container 5dec8a8412931ceca2bd60fc464051fb4cf44577e5a9a956f0f9464ce4332585. Jan 17 12:20:33.485972 containerd[1544]: time="2025-01-17T12:20:33.485947441Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-7jz56,Uid:748f4fc1-6b19-4928-bff2-5fd3a7b6e50e,Namespace:kube-system,Attempt:0,} returns sandbox id \"5dec8a8412931ceca2bd60fc464051fb4cf44577e5a9a956f0f9464ce4332585\"" Jan 17 12:20:33.488483 containerd[1544]: time="2025-01-17T12:20:33.488455617Z" level=info msg="CreateContainer within sandbox \"5dec8a8412931ceca2bd60fc464051fb4cf44577e5a9a956f0f9464ce4332585\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Jan 17 12:20:33.496202 containerd[1544]: time="2025-01-17T12:20:33.496038365Z" level=info msg="CreateContainer within sandbox \"5dec8a8412931ceca2bd60fc464051fb4cf44577e5a9a956f0f9464ce4332585\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"996d103a3afe46465b1f7cddae2a71719dc11be0b7753fbdd5b51f61a2b933e8\"" Jan 17 12:20:33.497681 containerd[1544]: time="2025-01-17T12:20:33.497645720Z" level=info msg="StartContainer for \"996d103a3afe46465b1f7cddae2a71719dc11be0b7753fbdd5b51f61a2b933e8\"" Jan 17 12:20:33.518822 systemd[1]: Started cri-containerd-996d103a3afe46465b1f7cddae2a71719dc11be0b7753fbdd5b51f61a2b933e8.scope - libcontainer container 996d103a3afe46465b1f7cddae2a71719dc11be0b7753fbdd5b51f61a2b933e8. Jan 17 12:20:33.542285 containerd[1544]: time="2025-01-17T12:20:33.541899218Z" level=info msg="StartContainer for \"996d103a3afe46465b1f7cddae2a71719dc11be0b7753fbdd5b51f61a2b933e8\" returns successfully" Jan 17 12:20:33.611959 kubelet[2723]: I0117 12:20:33.611935 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/ef1d5a58-97f0-4379-b0e0-cae489254453-var-lib-calico\") pod \"tigera-operator-76c4976dd7-2sjgp\" (UID: \"ef1d5a58-97f0-4379-b0e0-cae489254453\") " pod="tigera-operator/tigera-operator-76c4976dd7-2sjgp" Jan 17 12:20:33.612345 kubelet[2723]: I0117 12:20:33.612302 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hx5th\" (UniqueName: \"kubernetes.io/projected/ef1d5a58-97f0-4379-b0e0-cae489254453-kube-api-access-hx5th\") pod \"tigera-operator-76c4976dd7-2sjgp\" (UID: \"ef1d5a58-97f0-4379-b0e0-cae489254453\") " pod="tigera-operator/tigera-operator-76c4976dd7-2sjgp" Jan 17 12:20:33.630533 kubelet[2723]: I0117 12:20:33.630490 2723 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-7jz56" podStartSLOduration=0.630475435 podStartE2EDuration="630.475435ms" podCreationTimestamp="2025-01-17 12:20:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-17 12:20:33.630454901 +0000 UTC m=+6.113927487" watchObservedRunningTime="2025-01-17 12:20:33.630475435 +0000 UTC m=+6.113948022" Jan 17 12:20:33.748194 containerd[1544]: time="2025-01-17T12:20:33.748021145Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-76c4976dd7-2sjgp,Uid:ef1d5a58-97f0-4379-b0e0-cae489254453,Namespace:tigera-operator,Attempt:0,}" Jan 17 12:20:33.767183 containerd[1544]: time="2025-01-17T12:20:33.767019002Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 17 12:20:33.767183 containerd[1544]: time="2025-01-17T12:20:33.767081791Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 17 12:20:33.767428 containerd[1544]: time="2025-01-17T12:20:33.767171344Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 12:20:33.767479 containerd[1544]: time="2025-01-17T12:20:33.767410081Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 12:20:33.784797 systemd[1]: Started cri-containerd-87ed2c7c805af5a41e1c2ff0eee96141e4725d05056c8509ad78daaffc83bf0d.scope - libcontainer container 87ed2c7c805af5a41e1c2ff0eee96141e4725d05056c8509ad78daaffc83bf0d. Jan 17 12:20:33.812189 containerd[1544]: time="2025-01-17T12:20:33.812155541Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-76c4976dd7-2sjgp,Uid:ef1d5a58-97f0-4379-b0e0-cae489254453,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"87ed2c7c805af5a41e1c2ff0eee96141e4725d05056c8509ad78daaffc83bf0d\"" Jan 17 12:20:33.813078 containerd[1544]: time="2025-01-17T12:20:33.813063838Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.36.2\"" Jan 17 12:20:34.328330 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3223442001.mount: Deactivated successfully. Jan 17 12:20:37.965563 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3603106044.mount: Deactivated successfully. Jan 17 12:20:38.352627 containerd[1544]: time="2025-01-17T12:20:38.352558211Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.36.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:20:38.355310 containerd[1544]: time="2025-01-17T12:20:38.355273218Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.36.2: active requests=0, bytes read=21764301" Jan 17 12:20:38.360771 containerd[1544]: time="2025-01-17T12:20:38.358880122Z" level=info msg="ImageCreate event name:\"sha256:3045aa4a360d468ed15090f280e94c54bf4678269a6e863a9ebcf5b31534a346\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:20:38.372152 containerd[1544]: time="2025-01-17T12:20:38.372102459Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:fc9ea45f2475fd99db1b36d2ff180a50017b1a5ea0e82a171c6b439b3a620764\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:20:38.372909 containerd[1544]: time="2025-01-17T12:20:38.372501955Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.36.2\" with image id \"sha256:3045aa4a360d468ed15090f280e94c54bf4678269a6e863a9ebcf5b31534a346\", repo tag \"quay.io/tigera/operator:v1.36.2\", repo digest \"quay.io/tigera/operator@sha256:fc9ea45f2475fd99db1b36d2ff180a50017b1a5ea0e82a171c6b439b3a620764\", size \"21758492\" in 4.559421225s" Jan 17 12:20:38.372909 containerd[1544]: time="2025-01-17T12:20:38.372522494Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.36.2\" returns image reference \"sha256:3045aa4a360d468ed15090f280e94c54bf4678269a6e863a9ebcf5b31534a346\"" Jan 17 12:20:38.477184 containerd[1544]: time="2025-01-17T12:20:38.477145225Z" level=info msg="CreateContainer within sandbox \"87ed2c7c805af5a41e1c2ff0eee96141e4725d05056c8509ad78daaffc83bf0d\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Jan 17 12:20:38.514741 containerd[1544]: time="2025-01-17T12:20:38.514715299Z" level=info msg="CreateContainer within sandbox \"87ed2c7c805af5a41e1c2ff0eee96141e4725d05056c8509ad78daaffc83bf0d\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"eeac3f3a92d898c8bdad4bf192788c0c9079ff7c8aa254096943ef192381c211\"" Jan 17 12:20:38.529298 containerd[1544]: time="2025-01-17T12:20:38.529270233Z" level=info msg="StartContainer for \"eeac3f3a92d898c8bdad4bf192788c0c9079ff7c8aa254096943ef192381c211\"" Jan 17 12:20:38.545789 systemd[1]: Started cri-containerd-eeac3f3a92d898c8bdad4bf192788c0c9079ff7c8aa254096943ef192381c211.scope - libcontainer container eeac3f3a92d898c8bdad4bf192788c0c9079ff7c8aa254096943ef192381c211. Jan 17 12:20:38.567074 containerd[1544]: time="2025-01-17T12:20:38.567039695Z" level=info msg="StartContainer for \"eeac3f3a92d898c8bdad4bf192788c0c9079ff7c8aa254096943ef192381c211\" returns successfully" Jan 17 12:20:39.585508 kubelet[2723]: I0117 12:20:39.585426 2723 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-76c4976dd7-2sjgp" podStartSLOduration=1.974229161 podStartE2EDuration="6.585413074s" podCreationTimestamp="2025-01-17 12:20:33 +0000 UTC" firstStartedPulling="2025-01-17 12:20:33.812811799 +0000 UTC m=+6.296284382" lastFinishedPulling="2025-01-17 12:20:38.423995711 +0000 UTC m=+10.907468295" observedRunningTime="2025-01-17 12:20:38.640463739 +0000 UTC m=+11.123936327" watchObservedRunningTime="2025-01-17 12:20:39.585413074 +0000 UTC m=+12.068885662" Jan 17 12:20:41.426209 systemd[1]: Created slice kubepods-besteffort-podae5d27f6_477c_469b_a2ed_6749e526d2dd.slice - libcontainer container kubepods-besteffort-podae5d27f6_477c_469b_a2ed_6749e526d2dd.slice. Jan 17 12:20:41.472707 kubelet[2723]: I0117 12:20:41.471909 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ae5d27f6-477c-469b-a2ed-6749e526d2dd-tigera-ca-bundle\") pod \"calico-typha-6b6f478798-z2dzj\" (UID: \"ae5d27f6-477c-469b-a2ed-6749e526d2dd\") " pod="calico-system/calico-typha-6b6f478798-z2dzj" Jan 17 12:20:41.472707 kubelet[2723]: I0117 12:20:41.471942 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/4fe763aa-7473-4bca-bce2-d68bbf9c2ef4-cni-log-dir\") pod \"calico-node-d5fqz\" (UID: \"4fe763aa-7473-4bca-bce2-d68bbf9c2ef4\") " pod="calico-system/calico-node-d5fqz" Jan 17 12:20:41.472707 kubelet[2723]: I0117 12:20:41.471959 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/4fe763aa-7473-4bca-bce2-d68bbf9c2ef4-flexvol-driver-host\") pod \"calico-node-d5fqz\" (UID: \"4fe763aa-7473-4bca-bce2-d68bbf9c2ef4\") " pod="calico-system/calico-node-d5fqz" Jan 17 12:20:41.472707 kubelet[2723]: I0117 12:20:41.471969 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4fe763aa-7473-4bca-bce2-d68bbf9c2ef4-tigera-ca-bundle\") pod \"calico-node-d5fqz\" (UID: \"4fe763aa-7473-4bca-bce2-d68bbf9c2ef4\") " pod="calico-system/calico-node-d5fqz" Jan 17 12:20:41.472707 kubelet[2723]: I0117 12:20:41.471978 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/4fe763aa-7473-4bca-bce2-d68bbf9c2ef4-node-certs\") pod \"calico-node-d5fqz\" (UID: \"4fe763aa-7473-4bca-bce2-d68bbf9c2ef4\") " pod="calico-system/calico-node-d5fqz" Jan 17 12:20:41.473043 kubelet[2723]: I0117 12:20:41.471987 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/4fe763aa-7473-4bca-bce2-d68bbf9c2ef4-var-run-calico\") pod \"calico-node-d5fqz\" (UID: \"4fe763aa-7473-4bca-bce2-d68bbf9c2ef4\") " pod="calico-system/calico-node-d5fqz" Jan 17 12:20:41.473043 kubelet[2723]: I0117 12:20:41.471995 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/4fe763aa-7473-4bca-bce2-d68bbf9c2ef4-cni-bin-dir\") pod \"calico-node-d5fqz\" (UID: \"4fe763aa-7473-4bca-bce2-d68bbf9c2ef4\") " pod="calico-system/calico-node-d5fqz" Jan 17 12:20:41.473043 kubelet[2723]: I0117 12:20:41.472007 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/ae5d27f6-477c-469b-a2ed-6749e526d2dd-typha-certs\") pod \"calico-typha-6b6f478798-z2dzj\" (UID: \"ae5d27f6-477c-469b-a2ed-6749e526d2dd\") " pod="calico-system/calico-typha-6b6f478798-z2dzj" Jan 17 12:20:41.473043 kubelet[2723]: I0117 12:20:41.472039 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dz8fk\" (UniqueName: \"kubernetes.io/projected/4fe763aa-7473-4bca-bce2-d68bbf9c2ef4-kube-api-access-dz8fk\") pod \"calico-node-d5fqz\" (UID: \"4fe763aa-7473-4bca-bce2-d68bbf9c2ef4\") " pod="calico-system/calico-node-d5fqz" Jan 17 12:20:41.473043 kubelet[2723]: I0117 12:20:41.472054 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4fe763aa-7473-4bca-bce2-d68bbf9c2ef4-lib-modules\") pod \"calico-node-d5fqz\" (UID: \"4fe763aa-7473-4bca-bce2-d68bbf9c2ef4\") " pod="calico-system/calico-node-d5fqz" Jan 17 12:20:41.474886 kubelet[2723]: I0117 12:20:41.472065 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/4fe763aa-7473-4bca-bce2-d68bbf9c2ef4-policysync\") pod \"calico-node-d5fqz\" (UID: \"4fe763aa-7473-4bca-bce2-d68bbf9c2ef4\") " pod="calico-system/calico-node-d5fqz" Jan 17 12:20:41.474886 kubelet[2723]: I0117 12:20:41.472076 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/4fe763aa-7473-4bca-bce2-d68bbf9c2ef4-var-lib-calico\") pod \"calico-node-d5fqz\" (UID: \"4fe763aa-7473-4bca-bce2-d68bbf9c2ef4\") " pod="calico-system/calico-node-d5fqz" Jan 17 12:20:41.474886 kubelet[2723]: I0117 12:20:41.472094 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/4fe763aa-7473-4bca-bce2-d68bbf9c2ef4-xtables-lock\") pod \"calico-node-d5fqz\" (UID: \"4fe763aa-7473-4bca-bce2-d68bbf9c2ef4\") " pod="calico-system/calico-node-d5fqz" Jan 17 12:20:41.474886 kubelet[2723]: I0117 12:20:41.472111 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fn7cm\" (UniqueName: \"kubernetes.io/projected/ae5d27f6-477c-469b-a2ed-6749e526d2dd-kube-api-access-fn7cm\") pod \"calico-typha-6b6f478798-z2dzj\" (UID: \"ae5d27f6-477c-469b-a2ed-6749e526d2dd\") " pod="calico-system/calico-typha-6b6f478798-z2dzj" Jan 17 12:20:41.474886 kubelet[2723]: I0117 12:20:41.472125 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/4fe763aa-7473-4bca-bce2-d68bbf9c2ef4-cni-net-dir\") pod \"calico-node-d5fqz\" (UID: \"4fe763aa-7473-4bca-bce2-d68bbf9c2ef4\") " pod="calico-system/calico-node-d5fqz" Jan 17 12:20:41.474106 systemd[1]: Created slice kubepods-besteffort-pod4fe763aa_7473_4bca_bce2_d68bbf9c2ef4.slice - libcontainer container kubepods-besteffort-pod4fe763aa_7473_4bca_bce2_d68bbf9c2ef4.slice. Jan 17 12:20:41.585755 kubelet[2723]: E0117 12:20:41.584951 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.585755 kubelet[2723]: W0117 12:20:41.584980 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.585755 kubelet[2723]: E0117 12:20:41.585007 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.586220 kubelet[2723]: E0117 12:20:41.586204 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.586220 kubelet[2723]: W0117 12:20:41.586215 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.587234 kubelet[2723]: E0117 12:20:41.587081 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.587234 kubelet[2723]: W0117 12:20:41.587089 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.587845 kubelet[2723]: E0117 12:20:41.587559 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.587845 kubelet[2723]: W0117 12:20:41.587570 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.587845 kubelet[2723]: E0117 12:20:41.587582 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.588235 kubelet[2723]: E0117 12:20:41.588195 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.588235 kubelet[2723]: W0117 12:20:41.588203 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.588235 kubelet[2723]: E0117 12:20:41.588212 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.588372 kubelet[2723]: E0117 12:20:41.588347 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.588372 kubelet[2723]: W0117 12:20:41.588355 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.588372 kubelet[2723]: E0117 12:20:41.588363 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.590139 kubelet[2723]: E0117 12:20:41.590048 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.590139 kubelet[2723]: E0117 12:20:41.590069 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.594814 kubelet[2723]: E0117 12:20:41.594759 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.594814 kubelet[2723]: W0117 12:20:41.594787 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.594814 kubelet[2723]: E0117 12:20:41.594807 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.597160 kubelet[2723]: E0117 12:20:41.597140 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.597160 kubelet[2723]: W0117 12:20:41.597155 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.597213 kubelet[2723]: E0117 12:20:41.597169 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.597328 kubelet[2723]: E0117 12:20:41.597310 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.597328 kubelet[2723]: W0117 12:20:41.597323 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.597328 kubelet[2723]: E0117 12:20:41.597329 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.597502 kubelet[2723]: E0117 12:20:41.597490 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.597502 kubelet[2723]: W0117 12:20:41.597498 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.597548 kubelet[2723]: E0117 12:20:41.597503 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.607745 kubelet[2723]: E0117 12:20:41.607701 2723 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-zj6v2" podUID="d5c6d7d3-a76d-4ec7-b919-828df6f31125" Jan 17 12:20:41.620904 kubelet[2723]: E0117 12:20:41.620885 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.620904 kubelet[2723]: W0117 12:20:41.620898 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.621052 kubelet[2723]: E0117 12:20:41.620915 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.626890 kubelet[2723]: E0117 12:20:41.626768 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.626890 kubelet[2723]: W0117 12:20:41.626784 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.626890 kubelet[2723]: E0117 12:20:41.626798 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.672793 kubelet[2723]: E0117 12:20:41.672768 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.672793 kubelet[2723]: W0117 12:20:41.672785 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.672911 kubelet[2723]: E0117 12:20:41.672804 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.672977 kubelet[2723]: E0117 12:20:41.672966 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.672977 kubelet[2723]: W0117 12:20:41.672975 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.673025 kubelet[2723]: E0117 12:20:41.672982 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.673124 kubelet[2723]: E0117 12:20:41.673113 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.673124 kubelet[2723]: W0117 12:20:41.673120 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.673169 kubelet[2723]: E0117 12:20:41.673143 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.673270 kubelet[2723]: E0117 12:20:41.673259 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.673270 kubelet[2723]: W0117 12:20:41.673267 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.673317 kubelet[2723]: E0117 12:20:41.673272 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.673414 kubelet[2723]: E0117 12:20:41.673400 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.673414 kubelet[2723]: W0117 12:20:41.673409 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.673477 kubelet[2723]: E0117 12:20:41.673416 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.673573 kubelet[2723]: E0117 12:20:41.673562 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.673573 kubelet[2723]: W0117 12:20:41.673569 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.673573 kubelet[2723]: E0117 12:20:41.673574 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.673711 kubelet[2723]: E0117 12:20:41.673699 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.673711 kubelet[2723]: W0117 12:20:41.673707 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.673762 kubelet[2723]: E0117 12:20:41.673713 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.673850 kubelet[2723]: E0117 12:20:41.673838 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.673850 kubelet[2723]: W0117 12:20:41.673846 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.673850 kubelet[2723]: E0117 12:20:41.673851 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.673985 kubelet[2723]: E0117 12:20:41.673969 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.673985 kubelet[2723]: W0117 12:20:41.673981 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.673985 kubelet[2723]: E0117 12:20:41.673986 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.674133 kubelet[2723]: E0117 12:20:41.674120 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.674133 kubelet[2723]: W0117 12:20:41.674128 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.674133 kubelet[2723]: E0117 12:20:41.674133 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.674259 kubelet[2723]: E0117 12:20:41.674248 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.674284 kubelet[2723]: W0117 12:20:41.674261 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.674284 kubelet[2723]: E0117 12:20:41.674268 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.674403 kubelet[2723]: E0117 12:20:41.674392 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.674403 kubelet[2723]: W0117 12:20:41.674400 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.674448 kubelet[2723]: E0117 12:20:41.674405 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.674543 kubelet[2723]: E0117 12:20:41.674530 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.674543 kubelet[2723]: W0117 12:20:41.674543 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.674598 kubelet[2723]: E0117 12:20:41.674549 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.674713 kubelet[2723]: E0117 12:20:41.674682 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.674713 kubelet[2723]: W0117 12:20:41.674698 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.674713 kubelet[2723]: E0117 12:20:41.674707 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.674841 kubelet[2723]: E0117 12:20:41.674829 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.674841 kubelet[2723]: W0117 12:20:41.674837 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.674841 kubelet[2723]: E0117 12:20:41.674842 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.674990 kubelet[2723]: E0117 12:20:41.674978 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.674990 kubelet[2723]: W0117 12:20:41.674986 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.674990 kubelet[2723]: E0117 12:20:41.674991 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.675149 kubelet[2723]: E0117 12:20:41.675137 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.675149 kubelet[2723]: W0117 12:20:41.675144 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.675149 kubelet[2723]: E0117 12:20:41.675149 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.675294 kubelet[2723]: E0117 12:20:41.675283 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.675294 kubelet[2723]: W0117 12:20:41.675290 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.675294 kubelet[2723]: E0117 12:20:41.675295 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.675423 kubelet[2723]: E0117 12:20:41.675411 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.675423 kubelet[2723]: W0117 12:20:41.675418 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.675423 kubelet[2723]: E0117 12:20:41.675423 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.675561 kubelet[2723]: E0117 12:20:41.675551 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.675561 kubelet[2723]: W0117 12:20:41.675558 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.675605 kubelet[2723]: E0117 12:20:41.675563 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.675721 kubelet[2723]: E0117 12:20:41.675708 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.675721 kubelet[2723]: W0117 12:20:41.675716 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.675721 kubelet[2723]: E0117 12:20:41.675720 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.675874 kubelet[2723]: E0117 12:20:41.675860 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.675874 kubelet[2723]: W0117 12:20:41.675868 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.675874 kubelet[2723]: E0117 12:20:41.675874 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.676019 kubelet[2723]: E0117 12:20:41.676007 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.676019 kubelet[2723]: W0117 12:20:41.676015 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.676068 kubelet[2723]: E0117 12:20:41.676021 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.676162 kubelet[2723]: E0117 12:20:41.676151 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.676162 kubelet[2723]: W0117 12:20:41.676158 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.676162 kubelet[2723]: E0117 12:20:41.676163 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.676925 kubelet[2723]: E0117 12:20:41.676283 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.676925 kubelet[2723]: W0117 12:20:41.676290 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.676925 kubelet[2723]: E0117 12:20:41.676303 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.677914 kubelet[2723]: E0117 12:20:41.677866 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.678057 kubelet[2723]: W0117 12:20:41.677949 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.678057 kubelet[2723]: E0117 12:20:41.677964 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.678310 kubelet[2723]: E0117 12:20:41.678250 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.678310 kubelet[2723]: W0117 12:20:41.678257 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.678310 kubelet[2723]: E0117 12:20:41.678262 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.678477 kubelet[2723]: E0117 12:20:41.678471 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.678611 kubelet[2723]: W0117 12:20:41.678580 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.678611 kubelet[2723]: E0117 12:20:41.678597 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.679106 kubelet[2723]: E0117 12:20:41.679085 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.679374 kubelet[2723]: W0117 12:20:41.679232 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.679374 kubelet[2723]: E0117 12:20:41.679243 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.679469 kubelet[2723]: E0117 12:20:41.679461 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.679561 kubelet[2723]: W0117 12:20:41.679502 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.679561 kubelet[2723]: E0117 12:20:41.679513 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.679702 kubelet[2723]: E0117 12:20:41.679652 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.679702 kubelet[2723]: W0117 12:20:41.679658 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.679702 kubelet[2723]: E0117 12:20:41.679666 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.679949 kubelet[2723]: E0117 12:20:41.679872 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.679949 kubelet[2723]: W0117 12:20:41.679878 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.679949 kubelet[2723]: E0117 12:20:41.679886 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.680142 kubelet[2723]: E0117 12:20:41.680081 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.680142 kubelet[2723]: W0117 12:20:41.680088 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.680142 kubelet[2723]: E0117 12:20:41.680096 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.680273 kubelet[2723]: E0117 12:20:41.680239 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.680273 kubelet[2723]: W0117 12:20:41.680246 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.680273 kubelet[2723]: E0117 12:20:41.680251 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.680513 kubelet[2723]: E0117 12:20:41.680431 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.680513 kubelet[2723]: W0117 12:20:41.680437 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.680513 kubelet[2723]: E0117 12:20:41.680442 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.680663 kubelet[2723]: E0117 12:20:41.680657 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.680730 kubelet[2723]: W0117 12:20:41.680692 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.680730 kubelet[2723]: E0117 12:20:41.680700 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.680860 kubelet[2723]: I0117 12:20:41.680785 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d5c6d7d3-a76d-4ec7-b919-828df6f31125-kubelet-dir\") pod \"csi-node-driver-zj6v2\" (UID: \"d5c6d7d3-a76d-4ec7-b919-828df6f31125\") " pod="calico-system/csi-node-driver-zj6v2" Jan 17 12:20:41.680924 kubelet[2723]: E0117 12:20:41.680918 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.681009 kubelet[2723]: W0117 12:20:41.680951 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.681009 kubelet[2723]: E0117 12:20:41.680972 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.681009 kubelet[2723]: I0117 12:20:41.680983 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d5c6d7d3-a76d-4ec7-b919-828df6f31125-registration-dir\") pod \"csi-node-driver-zj6v2\" (UID: \"d5c6d7d3-a76d-4ec7-b919-828df6f31125\") " pod="calico-system/csi-node-driver-zj6v2" Jan 17 12:20:41.681263 kubelet[2723]: E0117 12:20:41.681192 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.681263 kubelet[2723]: W0117 12:20:41.681199 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.681263 kubelet[2723]: E0117 12:20:41.681206 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.681263 kubelet[2723]: I0117 12:20:41.681215 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d5c6d7d3-a76d-4ec7-b919-828df6f31125-socket-dir\") pod \"csi-node-driver-zj6v2\" (UID: \"d5c6d7d3-a76d-4ec7-b919-828df6f31125\") " pod="calico-system/csi-node-driver-zj6v2" Jan 17 12:20:41.681467 kubelet[2723]: E0117 12:20:41.681392 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.681467 kubelet[2723]: W0117 12:20:41.681398 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.681467 kubelet[2723]: E0117 12:20:41.681408 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.681467 kubelet[2723]: I0117 12:20:41.681417 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86qpv\" (UniqueName: \"kubernetes.io/projected/d5c6d7d3-a76d-4ec7-b919-828df6f31125-kube-api-access-86qpv\") pod \"csi-node-driver-zj6v2\" (UID: \"d5c6d7d3-a76d-4ec7-b919-828df6f31125\") " pod="calico-system/csi-node-driver-zj6v2" Jan 17 12:20:41.681743 kubelet[2723]: E0117 12:20:41.681630 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.681743 kubelet[2723]: W0117 12:20:41.681637 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.681743 kubelet[2723]: E0117 12:20:41.681645 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.681743 kubelet[2723]: I0117 12:20:41.681654 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/d5c6d7d3-a76d-4ec7-b919-828df6f31125-varrun\") pod \"csi-node-driver-zj6v2\" (UID: \"d5c6d7d3-a76d-4ec7-b919-828df6f31125\") " pod="calico-system/csi-node-driver-zj6v2" Jan 17 12:20:41.681942 kubelet[2723]: E0117 12:20:41.681802 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.681942 kubelet[2723]: W0117 12:20:41.681807 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.681942 kubelet[2723]: E0117 12:20:41.681920 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.682060 kubelet[2723]: E0117 12:20:41.682021 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.682060 kubelet[2723]: W0117 12:20:41.682028 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.682060 kubelet[2723]: E0117 12:20:41.682041 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.682197 kubelet[2723]: E0117 12:20:41.682191 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.682288 kubelet[2723]: W0117 12:20:41.682232 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.682288 kubelet[2723]: E0117 12:20:41.682247 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.682419 kubelet[2723]: E0117 12:20:41.682370 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.682419 kubelet[2723]: W0117 12:20:41.682375 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.682419 kubelet[2723]: E0117 12:20:41.682382 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.682544 kubelet[2723]: E0117 12:20:41.682504 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.682544 kubelet[2723]: W0117 12:20:41.682509 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.682544 kubelet[2723]: E0117 12:20:41.682519 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.682668 kubelet[2723]: E0117 12:20:41.682663 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.682753 kubelet[2723]: W0117 12:20:41.682715 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.682753 kubelet[2723]: E0117 12:20:41.682720 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.682987 kubelet[2723]: E0117 12:20:41.682929 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.682987 kubelet[2723]: W0117 12:20:41.682935 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.682987 kubelet[2723]: E0117 12:20:41.682940 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.683097 kubelet[2723]: E0117 12:20:41.683041 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.683097 kubelet[2723]: W0117 12:20:41.683046 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.683097 kubelet[2723]: E0117 12:20:41.683052 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.683340 kubelet[2723]: E0117 12:20:41.683279 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.683340 kubelet[2723]: W0117 12:20:41.683285 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.683340 kubelet[2723]: E0117 12:20:41.683290 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.683475 kubelet[2723]: E0117 12:20:41.683395 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.683475 kubelet[2723]: W0117 12:20:41.683399 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.683475 kubelet[2723]: E0117 12:20:41.683404 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.778275 containerd[1544]: time="2025-01-17T12:20:41.778246236Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-6b6f478798-z2dzj,Uid:ae5d27f6-477c-469b-a2ed-6749e526d2dd,Namespace:calico-system,Attempt:0,}" Jan 17 12:20:41.780452 containerd[1544]: time="2025-01-17T12:20:41.780400197Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-d5fqz,Uid:4fe763aa-7473-4bca-bce2-d68bbf9c2ef4,Namespace:calico-system,Attempt:0,}" Jan 17 12:20:41.782216 kubelet[2723]: E0117 12:20:41.782199 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.782216 kubelet[2723]: W0117 12:20:41.782210 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.782350 kubelet[2723]: E0117 12:20:41.782225 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.782378 kubelet[2723]: E0117 12:20:41.782365 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.782378 kubelet[2723]: W0117 12:20:41.782370 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.782419 kubelet[2723]: E0117 12:20:41.782381 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.782613 kubelet[2723]: E0117 12:20:41.782477 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.782613 kubelet[2723]: W0117 12:20:41.782483 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.782613 kubelet[2723]: E0117 12:20:41.782488 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.791117 kubelet[2723]: E0117 12:20:41.782667 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.791117 kubelet[2723]: W0117 12:20:41.782683 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.791117 kubelet[2723]: E0117 12:20:41.782691 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.791117 kubelet[2723]: E0117 12:20:41.782788 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.791117 kubelet[2723]: W0117 12:20:41.782816 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.791117 kubelet[2723]: E0117 12:20:41.782826 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.791117 kubelet[2723]: E0117 12:20:41.782928 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.791117 kubelet[2723]: W0117 12:20:41.782932 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.791117 kubelet[2723]: E0117 12:20:41.782936 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.791117 kubelet[2723]: E0117 12:20:41.783011 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.791309 kubelet[2723]: W0117 12:20:41.783015 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.791309 kubelet[2723]: E0117 12:20:41.783019 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.791309 kubelet[2723]: E0117 12:20:41.783092 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.791309 kubelet[2723]: W0117 12:20:41.783096 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.791309 kubelet[2723]: E0117 12:20:41.783107 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.791309 kubelet[2723]: E0117 12:20:41.783223 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.791309 kubelet[2723]: W0117 12:20:41.783227 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.791309 kubelet[2723]: E0117 12:20:41.783234 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.791309 kubelet[2723]: E0117 12:20:41.783458 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.791309 kubelet[2723]: W0117 12:20:41.783464 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.791469 kubelet[2723]: E0117 12:20:41.783475 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.791469 kubelet[2723]: E0117 12:20:41.784626 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.791469 kubelet[2723]: W0117 12:20:41.784634 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.791469 kubelet[2723]: E0117 12:20:41.784647 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.791469 kubelet[2723]: E0117 12:20:41.784780 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.791469 kubelet[2723]: W0117 12:20:41.784795 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.791469 kubelet[2723]: E0117 12:20:41.784810 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.791469 kubelet[2723]: E0117 12:20:41.785009 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.791469 kubelet[2723]: W0117 12:20:41.785019 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.791469 kubelet[2723]: E0117 12:20:41.785037 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.791627 kubelet[2723]: E0117 12:20:41.785132 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.791627 kubelet[2723]: W0117 12:20:41.785137 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.791627 kubelet[2723]: E0117 12:20:41.785265 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.791627 kubelet[2723]: E0117 12:20:41.785284 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.791627 kubelet[2723]: W0117 12:20:41.785289 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.791627 kubelet[2723]: E0117 12:20:41.785306 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.791627 kubelet[2723]: E0117 12:20:41.785421 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.791627 kubelet[2723]: W0117 12:20:41.785428 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.791627 kubelet[2723]: E0117 12:20:41.785438 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.791627 kubelet[2723]: E0117 12:20:41.785577 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.791797 kubelet[2723]: W0117 12:20:41.785581 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.791797 kubelet[2723]: E0117 12:20:41.785589 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.791797 kubelet[2723]: E0117 12:20:41.785722 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.791797 kubelet[2723]: W0117 12:20:41.785728 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.791797 kubelet[2723]: E0117 12:20:41.785733 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.791797 kubelet[2723]: E0117 12:20:41.786034 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.791797 kubelet[2723]: W0117 12:20:41.786049 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.791797 kubelet[2723]: E0117 12:20:41.786056 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.791797 kubelet[2723]: E0117 12:20:41.786150 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.791797 kubelet[2723]: W0117 12:20:41.786156 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.791949 kubelet[2723]: E0117 12:20:41.786161 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.791949 kubelet[2723]: E0117 12:20:41.786245 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.791949 kubelet[2723]: W0117 12:20:41.786249 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.791949 kubelet[2723]: E0117 12:20:41.786253 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.791949 kubelet[2723]: E0117 12:20:41.786375 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.791949 kubelet[2723]: W0117 12:20:41.786379 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.791949 kubelet[2723]: E0117 12:20:41.786384 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.791949 kubelet[2723]: E0117 12:20:41.786470 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.791949 kubelet[2723]: W0117 12:20:41.786474 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.791949 kubelet[2723]: E0117 12:20:41.786478 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.794764 kubelet[2723]: E0117 12:20:41.786562 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.794764 kubelet[2723]: W0117 12:20:41.786566 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.794764 kubelet[2723]: E0117 12:20:41.786571 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.794764 kubelet[2723]: E0117 12:20:41.788040 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.794764 kubelet[2723]: W0117 12:20:41.788049 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.794764 kubelet[2723]: E0117 12:20:41.788059 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.795324 kubelet[2723]: E0117 12:20:41.795275 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:41.795324 kubelet[2723]: W0117 12:20:41.795284 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:41.795324 kubelet[2723]: E0117 12:20:41.795297 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:41.850521 containerd[1544]: time="2025-01-17T12:20:41.850378628Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 17 12:20:41.850720 containerd[1544]: time="2025-01-17T12:20:41.850572844Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 17 12:20:41.850900 containerd[1544]: time="2025-01-17T12:20:41.850779015Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 12:20:41.851252 containerd[1544]: time="2025-01-17T12:20:41.851214375Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 12:20:41.856022 containerd[1544]: time="2025-01-17T12:20:41.855512056Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 17 12:20:41.856022 containerd[1544]: time="2025-01-17T12:20:41.855557438Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 17 12:20:41.856022 containerd[1544]: time="2025-01-17T12:20:41.855567440Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 12:20:41.856022 containerd[1544]: time="2025-01-17T12:20:41.855613222Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 12:20:41.879825 systemd[1]: Started cri-containerd-9c5e27d62dd2e2f1a1c170ca46592bc5b6138340c94cbce7832614441af23571.scope - libcontainer container 9c5e27d62dd2e2f1a1c170ca46592bc5b6138340c94cbce7832614441af23571. Jan 17 12:20:41.883955 systemd[1]: Started cri-containerd-31098533b87e8c32bd5c7953206fce8bc0493676630ea682290d7f0996cd132c.scope - libcontainer container 31098533b87e8c32bd5c7953206fce8bc0493676630ea682290d7f0996cd132c. Jan 17 12:20:41.905661 containerd[1544]: time="2025-01-17T12:20:41.905484191Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-d5fqz,Uid:4fe763aa-7473-4bca-bce2-d68bbf9c2ef4,Namespace:calico-system,Attempt:0,} returns sandbox id \"31098533b87e8c32bd5c7953206fce8bc0493676630ea682290d7f0996cd132c\"" Jan 17 12:20:41.927208 containerd[1544]: time="2025-01-17T12:20:41.927113907Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-6b6f478798-z2dzj,Uid:ae5d27f6-477c-469b-a2ed-6749e526d2dd,Namespace:calico-system,Attempt:0,} returns sandbox id \"9c5e27d62dd2e2f1a1c170ca46592bc5b6138340c94cbce7832614441af23571\"" Jan 17 12:20:41.960292 containerd[1544]: time="2025-01-17T12:20:41.960129881Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.29.1\"" Jan 17 12:20:43.344535 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4037878396.mount: Deactivated successfully. Jan 17 12:20:43.614185 kubelet[2723]: E0117 12:20:43.614094 2723 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-zj6v2" podUID="d5c6d7d3-a76d-4ec7-b919-828df6f31125" Jan 17 12:20:43.957507 containerd[1544]: time="2025-01-17T12:20:43.957014541Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:20:43.957928 containerd[1544]: time="2025-01-17T12:20:43.957899413Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.29.1: active requests=0, bytes read=31343363" Jan 17 12:20:43.958527 containerd[1544]: time="2025-01-17T12:20:43.958500522Z" level=info msg="ImageCreate event name:\"sha256:4cb3738506f5a9c530033d1e24fd6b9ec618518a2ec8b012ded33572be06ab44\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:20:43.959987 containerd[1544]: time="2025-01-17T12:20:43.959947718Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:768a194e1115c73bcbf35edb7afd18a63e16e08d940c79993565b6a3cca2da7c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:20:43.961689 containerd[1544]: time="2025-01-17T12:20:43.960177077Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.29.1\" with image id \"sha256:4cb3738506f5a9c530033d1e24fd6b9ec618518a2ec8b012ded33572be06ab44\", repo tag \"ghcr.io/flatcar/calico/typha:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:768a194e1115c73bcbf35edb7afd18a63e16e08d940c79993565b6a3cca2da7c\", size \"31343217\" in 2.000020543s" Jan 17 12:20:43.961689 containerd[1544]: time="2025-01-17T12:20:43.960196039Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.29.1\" returns image reference \"sha256:4cb3738506f5a9c530033d1e24fd6b9ec618518a2ec8b012ded33572be06ab44\"" Jan 17 12:20:43.966044 containerd[1544]: time="2025-01-17T12:20:43.966020395Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\"" Jan 17 12:20:43.981306 containerd[1544]: time="2025-01-17T12:20:43.981281181Z" level=info msg="CreateContainer within sandbox \"9c5e27d62dd2e2f1a1c170ca46592bc5b6138340c94cbce7832614441af23571\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Jan 17 12:20:44.003524 containerd[1544]: time="2025-01-17T12:20:44.003501460Z" level=info msg="CreateContainer within sandbox \"9c5e27d62dd2e2f1a1c170ca46592bc5b6138340c94cbce7832614441af23571\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"d23a28e67d081070fcb7adee96df81c99e1914e262e79c3ec0a826fef091abd9\"" Jan 17 12:20:44.004012 containerd[1544]: time="2025-01-17T12:20:44.003998724Z" level=info msg="StartContainer for \"d23a28e67d081070fcb7adee96df81c99e1914e262e79c3ec0a826fef091abd9\"" Jan 17 12:20:44.026941 systemd[1]: Started cri-containerd-d23a28e67d081070fcb7adee96df81c99e1914e262e79c3ec0a826fef091abd9.scope - libcontainer container d23a28e67d081070fcb7adee96df81c99e1914e262e79c3ec0a826fef091abd9. Jan 17 12:20:44.064294 containerd[1544]: time="2025-01-17T12:20:44.064271742Z" level=info msg="StartContainer for \"d23a28e67d081070fcb7adee96df81c99e1914e262e79c3ec0a826fef091abd9\" returns successfully" Jan 17 12:20:44.646772 kubelet[2723]: E0117 12:20:44.646535 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:44.646772 kubelet[2723]: W0117 12:20:44.646551 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:44.646772 kubelet[2723]: E0117 12:20:44.646564 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:44.646772 kubelet[2723]: E0117 12:20:44.646666 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:44.646772 kubelet[2723]: W0117 12:20:44.646686 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:44.646772 kubelet[2723]: E0117 12:20:44.646692 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:44.646772 kubelet[2723]: E0117 12:20:44.646771 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:44.646772 kubelet[2723]: W0117 12:20:44.646776 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:44.646772 kubelet[2723]: E0117 12:20:44.646781 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:44.647448 kubelet[2723]: E0117 12:20:44.646878 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:44.647448 kubelet[2723]: W0117 12:20:44.646882 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:44.647448 kubelet[2723]: E0117 12:20:44.646887 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:44.647448 kubelet[2723]: E0117 12:20:44.646967 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:44.647448 kubelet[2723]: W0117 12:20:44.646971 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:44.647448 kubelet[2723]: E0117 12:20:44.646976 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:44.647715 kubelet[2723]: E0117 12:20:44.647704 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:44.647715 kubelet[2723]: W0117 12:20:44.647714 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:44.648087 kubelet[2723]: E0117 12:20:44.647720 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:44.648087 kubelet[2723]: E0117 12:20:44.647803 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:44.648087 kubelet[2723]: W0117 12:20:44.647807 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:44.648087 kubelet[2723]: E0117 12:20:44.647811 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:44.648087 kubelet[2723]: E0117 12:20:44.647916 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:44.648087 kubelet[2723]: W0117 12:20:44.647920 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:44.648087 kubelet[2723]: E0117 12:20:44.647925 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:44.648235 kubelet[2723]: E0117 12:20:44.648195 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:44.648235 kubelet[2723]: W0117 12:20:44.648200 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:44.648235 kubelet[2723]: E0117 12:20:44.648205 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:44.648419 kubelet[2723]: E0117 12:20:44.648405 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:44.648419 kubelet[2723]: W0117 12:20:44.648411 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:44.648419 kubelet[2723]: E0117 12:20:44.648419 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:44.648572 kubelet[2723]: E0117 12:20:44.648523 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:44.648572 kubelet[2723]: W0117 12:20:44.648531 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:44.648572 kubelet[2723]: E0117 12:20:44.648536 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:44.648630 kubelet[2723]: E0117 12:20:44.648622 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:44.648630 kubelet[2723]: W0117 12:20:44.648626 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:44.648663 kubelet[2723]: E0117 12:20:44.648630 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:44.648875 kubelet[2723]: E0117 12:20:44.648736 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:44.648875 kubelet[2723]: W0117 12:20:44.648743 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:44.648875 kubelet[2723]: E0117 12:20:44.648747 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:44.648875 kubelet[2723]: E0117 12:20:44.648841 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:44.648875 kubelet[2723]: W0117 12:20:44.648845 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:44.648875 kubelet[2723]: E0117 12:20:44.648850 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:44.649101 kubelet[2723]: E0117 12:20:44.648934 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:44.649101 kubelet[2723]: W0117 12:20:44.648938 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:44.649101 kubelet[2723]: E0117 12:20:44.648942 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:44.668338 kubelet[2723]: I0117 12:20:44.667635 2723 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-6b6f478798-z2dzj" podStartSLOduration=1.661493812 podStartE2EDuration="3.667624424s" podCreationTimestamp="2025-01-17 12:20:41 +0000 UTC" firstStartedPulling="2025-01-17 12:20:41.959735949 +0000 UTC m=+14.443208531" lastFinishedPulling="2025-01-17 12:20:43.965866559 +0000 UTC m=+16.449339143" observedRunningTime="2025-01-17 12:20:44.666240464 +0000 UTC m=+17.149713056" watchObservedRunningTime="2025-01-17 12:20:44.667624424 +0000 UTC m=+17.151097011" Jan 17 12:20:44.707311 kubelet[2723]: E0117 12:20:44.707027 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:44.707311 kubelet[2723]: W0117 12:20:44.707041 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:44.707311 kubelet[2723]: E0117 12:20:44.707279 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:44.707508 kubelet[2723]: E0117 12:20:44.707435 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:44.707508 kubelet[2723]: W0117 12:20:44.707440 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:44.707508 kubelet[2723]: E0117 12:20:44.707449 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:44.707597 kubelet[2723]: E0117 12:20:44.707549 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:44.707597 kubelet[2723]: W0117 12:20:44.707554 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:44.707597 kubelet[2723]: E0117 12:20:44.707561 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:44.707700 kubelet[2723]: E0117 12:20:44.707691 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:44.707700 kubelet[2723]: W0117 12:20:44.707697 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:44.707760 kubelet[2723]: E0117 12:20:44.707704 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:44.707801 kubelet[2723]: E0117 12:20:44.707796 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:44.707801 kubelet[2723]: W0117 12:20:44.707800 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:44.707862 kubelet[2723]: E0117 12:20:44.707816 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:44.707906 kubelet[2723]: E0117 12:20:44.707901 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:44.707906 kubelet[2723]: W0117 12:20:44.707906 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:44.707961 kubelet[2723]: E0117 12:20:44.707914 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:44.708029 kubelet[2723]: E0117 12:20:44.708016 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:44.708029 kubelet[2723]: W0117 12:20:44.708024 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:44.708029 kubelet[2723]: E0117 12:20:44.708031 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:44.708367 kubelet[2723]: E0117 12:20:44.708298 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:44.708367 kubelet[2723]: W0117 12:20:44.708308 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:44.708367 kubelet[2723]: E0117 12:20:44.708322 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:44.708523 kubelet[2723]: E0117 12:20:44.708423 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:44.708523 kubelet[2723]: W0117 12:20:44.708428 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:44.708523 kubelet[2723]: E0117 12:20:44.708434 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:44.708634 kubelet[2723]: E0117 12:20:44.708605 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:44.708634 kubelet[2723]: W0117 12:20:44.708611 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:44.708634 kubelet[2723]: E0117 12:20:44.708625 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:44.708890 kubelet[2723]: E0117 12:20:44.708839 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:44.708890 kubelet[2723]: W0117 12:20:44.708844 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:44.708890 kubelet[2723]: E0117 12:20:44.708859 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:44.708997 kubelet[2723]: E0117 12:20:44.708947 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:44.708997 kubelet[2723]: W0117 12:20:44.708951 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:44.709098 kubelet[2723]: E0117 12:20:44.709044 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:44.709241 kubelet[2723]: E0117 12:20:44.709202 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:44.709241 kubelet[2723]: W0117 12:20:44.709208 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:44.709241 kubelet[2723]: E0117 12:20:44.709217 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:44.709451 kubelet[2723]: E0117 12:20:44.709300 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:44.709451 kubelet[2723]: W0117 12:20:44.709305 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:44.709451 kubelet[2723]: E0117 12:20:44.709310 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:44.709451 kubelet[2723]: E0117 12:20:44.709386 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:44.709451 kubelet[2723]: W0117 12:20:44.709390 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:44.709451 kubelet[2723]: E0117 12:20:44.709397 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:44.709551 kubelet[2723]: E0117 12:20:44.709483 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:44.709551 kubelet[2723]: W0117 12:20:44.709488 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:44.709551 kubelet[2723]: E0117 12:20:44.709492 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:44.709696 kubelet[2723]: E0117 12:20:44.709687 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:44.709696 kubelet[2723]: W0117 12:20:44.709693 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:44.709765 kubelet[2723]: E0117 12:20:44.709700 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:44.709790 kubelet[2723]: E0117 12:20:44.709786 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:44.709807 kubelet[2723]: W0117 12:20:44.709790 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:44.709807 kubelet[2723]: E0117 12:20:44.709794 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:45.585076 kubelet[2723]: E0117 12:20:45.584870 2723 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-zj6v2" podUID="d5c6d7d3-a76d-4ec7-b919-828df6f31125" Jan 17 12:20:45.645783 kubelet[2723]: I0117 12:20:45.645760 2723 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 17 12:20:45.654297 kubelet[2723]: E0117 12:20:45.654282 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:45.654297 kubelet[2723]: W0117 12:20:45.654293 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:45.654505 kubelet[2723]: E0117 12:20:45.654308 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:45.654505 kubelet[2723]: E0117 12:20:45.654427 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:45.654505 kubelet[2723]: W0117 12:20:45.654432 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:45.654505 kubelet[2723]: E0117 12:20:45.654437 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:45.654593 kubelet[2723]: E0117 12:20:45.654530 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:45.654593 kubelet[2723]: W0117 12:20:45.654535 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:45.654593 kubelet[2723]: E0117 12:20:45.654540 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:45.654653 kubelet[2723]: E0117 12:20:45.654638 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:45.654653 kubelet[2723]: W0117 12:20:45.654652 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:45.654716 kubelet[2723]: E0117 12:20:45.654657 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:45.654788 kubelet[2723]: E0117 12:20:45.654777 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:45.654788 kubelet[2723]: W0117 12:20:45.654785 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:45.654829 kubelet[2723]: E0117 12:20:45.654790 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:45.654893 kubelet[2723]: E0117 12:20:45.654885 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:45.654915 kubelet[2723]: W0117 12:20:45.654893 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:45.654915 kubelet[2723]: E0117 12:20:45.654899 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:45.655010 kubelet[2723]: E0117 12:20:45.655000 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:45.655010 kubelet[2723]: W0117 12:20:45.655008 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:45.655050 kubelet[2723]: E0117 12:20:45.655012 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:45.655117 kubelet[2723]: E0117 12:20:45.655108 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:45.655117 kubelet[2723]: W0117 12:20:45.655116 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:45.655158 kubelet[2723]: E0117 12:20:45.655122 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:45.655235 kubelet[2723]: E0117 12:20:45.655226 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:45.655235 kubelet[2723]: W0117 12:20:45.655233 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:45.655274 kubelet[2723]: E0117 12:20:45.655238 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:45.655338 kubelet[2723]: E0117 12:20:45.655329 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:45.655338 kubelet[2723]: W0117 12:20:45.655336 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:45.659968 kubelet[2723]: E0117 12:20:45.655341 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:45.659968 kubelet[2723]: E0117 12:20:45.655432 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:45.659968 kubelet[2723]: W0117 12:20:45.655436 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:45.659968 kubelet[2723]: E0117 12:20:45.655441 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:45.659968 kubelet[2723]: E0117 12:20:45.655555 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:45.659968 kubelet[2723]: W0117 12:20:45.655560 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:45.659968 kubelet[2723]: E0117 12:20:45.655565 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:45.659968 kubelet[2723]: E0117 12:20:45.655662 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:45.659968 kubelet[2723]: W0117 12:20:45.655667 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:45.659968 kubelet[2723]: E0117 12:20:45.655682 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:45.660162 kubelet[2723]: E0117 12:20:45.655781 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:45.660162 kubelet[2723]: W0117 12:20:45.655785 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:45.660162 kubelet[2723]: E0117 12:20:45.655791 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:45.660162 kubelet[2723]: E0117 12:20:45.655881 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:45.660162 kubelet[2723]: W0117 12:20:45.655885 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:45.660162 kubelet[2723]: E0117 12:20:45.655889 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:45.713320 kubelet[2723]: E0117 12:20:45.713300 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:45.713320 kubelet[2723]: W0117 12:20:45.713314 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:45.713320 kubelet[2723]: E0117 12:20:45.713327 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:45.713500 kubelet[2723]: E0117 12:20:45.713490 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:45.713500 kubelet[2723]: W0117 12:20:45.713499 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:45.713540 kubelet[2723]: E0117 12:20:45.713511 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:45.713629 kubelet[2723]: E0117 12:20:45.713621 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:45.713629 kubelet[2723]: W0117 12:20:45.713628 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:45.713669 kubelet[2723]: E0117 12:20:45.713643 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:45.713758 kubelet[2723]: E0117 12:20:45.713749 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:45.713784 kubelet[2723]: W0117 12:20:45.713756 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:45.713784 kubelet[2723]: E0117 12:20:45.713773 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:45.713899 kubelet[2723]: E0117 12:20:45.713888 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:45.713899 kubelet[2723]: W0117 12:20:45.713896 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:45.713989 kubelet[2723]: E0117 12:20:45.713980 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:45.714105 kubelet[2723]: E0117 12:20:45.714095 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:45.714105 kubelet[2723]: W0117 12:20:45.714102 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:45.714144 kubelet[2723]: E0117 12:20:45.714109 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:45.714219 kubelet[2723]: E0117 12:20:45.714205 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:45.714241 kubelet[2723]: W0117 12:20:45.714229 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:45.714262 kubelet[2723]: E0117 12:20:45.714239 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:45.714366 kubelet[2723]: E0117 12:20:45.714356 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:45.714366 kubelet[2723]: W0117 12:20:45.714363 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:45.714456 kubelet[2723]: E0117 12:20:45.714419 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:45.714483 kubelet[2723]: E0117 12:20:45.714477 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:45.714483 kubelet[2723]: W0117 12:20:45.714481 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:45.714519 kubelet[2723]: E0117 12:20:45.714491 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:45.714614 kubelet[2723]: E0117 12:20:45.714603 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:45.714639 kubelet[2723]: W0117 12:20:45.714619 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:45.714639 kubelet[2723]: E0117 12:20:45.714629 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:45.714752 kubelet[2723]: E0117 12:20:45.714743 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:45.714777 kubelet[2723]: W0117 12:20:45.714752 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:45.714799 kubelet[2723]: E0117 12:20:45.714776 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:45.714886 kubelet[2723]: E0117 12:20:45.714876 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:45.714886 kubelet[2723]: W0117 12:20:45.714883 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:45.714928 kubelet[2723]: E0117 12:20:45.714889 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:45.715016 kubelet[2723]: E0117 12:20:45.715007 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:45.715016 kubelet[2723]: W0117 12:20:45.715014 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:45.715059 kubelet[2723]: E0117 12:20:45.715023 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:45.715510 kubelet[2723]: E0117 12:20:45.715206 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:45.715510 kubelet[2723]: W0117 12:20:45.715212 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:45.715510 kubelet[2723]: E0117 12:20:45.715223 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:45.715510 kubelet[2723]: E0117 12:20:45.715323 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:45.715510 kubelet[2723]: W0117 12:20:45.715328 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:45.715510 kubelet[2723]: E0117 12:20:45.715338 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:45.715510 kubelet[2723]: E0117 12:20:45.715445 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:45.715510 kubelet[2723]: W0117 12:20:45.715449 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:45.715510 kubelet[2723]: E0117 12:20:45.715461 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:45.715732 kubelet[2723]: E0117 12:20:45.715721 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:45.715732 kubelet[2723]: W0117 12:20:45.715730 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:45.715771 kubelet[2723]: E0117 12:20:45.715739 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:45.715839 kubelet[2723]: E0117 12:20:45.715828 2723 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 17 12:20:45.715839 kubelet[2723]: W0117 12:20:45.715837 2723 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 17 12:20:45.715879 kubelet[2723]: E0117 12:20:45.715842 2723 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 17 12:20:45.762753 containerd[1544]: time="2025-01-17T12:20:45.762724711Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:20:45.763254 containerd[1544]: time="2025-01-17T12:20:45.763224878Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1: active requests=0, bytes read=5362121" Jan 17 12:20:45.763836 containerd[1544]: time="2025-01-17T12:20:45.763522837Z" level=info msg="ImageCreate event name:\"sha256:2b7452b763ec8833ca0386ada5fd066e552a9b3b02b8538a5e34cc3d6d3840a6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:20:45.764502 containerd[1544]: time="2025-01-17T12:20:45.764481287Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:a63f8b4ff531912d12d143664eb263fdbc6cd7b3ff4aa777dfb6e318a090462c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:20:45.765126 containerd[1544]: time="2025-01-17T12:20:45.764875476Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\" with image id \"sha256:2b7452b763ec8833ca0386ada5fd066e552a9b3b02b8538a5e34cc3d6d3840a6\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:a63f8b4ff531912d12d143664eb263fdbc6cd7b3ff4aa777dfb6e318a090462c\", size \"6855165\" in 1.798832223s" Jan 17 12:20:45.765126 containerd[1544]: time="2025-01-17T12:20:45.764895897Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\" returns image reference \"sha256:2b7452b763ec8833ca0386ada5fd066e552a9b3b02b8538a5e34cc3d6d3840a6\"" Jan 17 12:20:45.766755 containerd[1544]: time="2025-01-17T12:20:45.766726784Z" level=info msg="CreateContainer within sandbox \"31098533b87e8c32bd5c7953206fce8bc0493676630ea682290d7f0996cd132c\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Jan 17 12:20:45.772954 containerd[1544]: time="2025-01-17T12:20:45.772932105Z" level=info msg="CreateContainer within sandbox \"31098533b87e8c32bd5c7953206fce8bc0493676630ea682290d7f0996cd132c\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"9251edc4e9998312e8e0470f1734b575f7c79bbd8039ddfb1e4497b2f5b702a6\"" Jan 17 12:20:45.775014 containerd[1544]: time="2025-01-17T12:20:45.774013390Z" level=info msg="StartContainer for \"9251edc4e9998312e8e0470f1734b575f7c79bbd8039ddfb1e4497b2f5b702a6\"" Jan 17 12:20:45.799761 systemd[1]: Started cri-containerd-9251edc4e9998312e8e0470f1734b575f7c79bbd8039ddfb1e4497b2f5b702a6.scope - libcontainer container 9251edc4e9998312e8e0470f1734b575f7c79bbd8039ddfb1e4497b2f5b702a6. Jan 17 12:20:45.818570 containerd[1544]: time="2025-01-17T12:20:45.818466921Z" level=info msg="StartContainer for \"9251edc4e9998312e8e0470f1734b575f7c79bbd8039ddfb1e4497b2f5b702a6\" returns successfully" Jan 17 12:20:45.831140 systemd[1]: cri-containerd-9251edc4e9998312e8e0470f1734b575f7c79bbd8039ddfb1e4497b2f5b702a6.scope: Deactivated successfully. Jan 17 12:20:45.901745 containerd[1544]: time="2025-01-17T12:20:45.882330127Z" level=info msg="shim disconnected" id=9251edc4e9998312e8e0470f1734b575f7c79bbd8039ddfb1e4497b2f5b702a6 namespace=k8s.io Jan 17 12:20:45.902636 containerd[1544]: time="2025-01-17T12:20:45.901731425Z" level=warning msg="cleaning up after shim disconnected" id=9251edc4e9998312e8e0470f1734b575f7c79bbd8039ddfb1e4497b2f5b702a6 namespace=k8s.io Jan 17 12:20:45.902636 containerd[1544]: time="2025-01-17T12:20:45.901814870Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 17 12:20:45.973931 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-9251edc4e9998312e8e0470f1734b575f7c79bbd8039ddfb1e4497b2f5b702a6-rootfs.mount: Deactivated successfully. Jan 17 12:20:46.657991 containerd[1544]: time="2025-01-17T12:20:46.657934630Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.1\"" Jan 17 12:20:47.584207 kubelet[2723]: E0117 12:20:47.583982 2723 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-zj6v2" podUID="d5c6d7d3-a76d-4ec7-b919-828df6f31125" Jan 17 12:20:49.584529 kubelet[2723]: E0117 12:20:49.583812 2723 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-zj6v2" podUID="d5c6d7d3-a76d-4ec7-b919-828df6f31125" Jan 17 12:20:49.641764 containerd[1544]: time="2025-01-17T12:20:49.641712574Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:20:49.644602 containerd[1544]: time="2025-01-17T12:20:49.642303781Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.29.1: active requests=0, bytes read=96154154" Jan 17 12:20:49.644602 containerd[1544]: time="2025-01-17T12:20:49.642687946Z" level=info msg="ImageCreate event name:\"sha256:7dd6ea186aba0d7a1791a79d426fe854527ca95192b26bbd19e8baf8373f7d0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:20:49.648541 containerd[1544]: time="2025-01-17T12:20:49.647895741Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:21e759d51c90dfb34fc1397dc180dd3a3fb564c2b0580d2f61ffe108f2a3c94b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:20:49.648541 containerd[1544]: time="2025-01-17T12:20:49.648464817Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.29.1\" with image id \"sha256:7dd6ea186aba0d7a1791a79d426fe854527ca95192b26bbd19e8baf8373f7d0e\", repo tag \"ghcr.io/flatcar/calico/cni:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:21e759d51c90dfb34fc1397dc180dd3a3fb564c2b0580d2f61ffe108f2a3c94b\", size \"97647238\" in 2.990347164s" Jan 17 12:20:49.648541 containerd[1544]: time="2025-01-17T12:20:49.648486058Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.1\" returns image reference \"sha256:7dd6ea186aba0d7a1791a79d426fe854527ca95192b26bbd19e8baf8373f7d0e\"" Jan 17 12:20:49.650405 containerd[1544]: time="2025-01-17T12:20:49.650375869Z" level=info msg="CreateContainer within sandbox \"31098533b87e8c32bd5c7953206fce8bc0493676630ea682290d7f0996cd132c\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Jan 17 12:20:49.658893 containerd[1544]: time="2025-01-17T12:20:49.658814663Z" level=info msg="CreateContainer within sandbox \"31098533b87e8c32bd5c7953206fce8bc0493676630ea682290d7f0996cd132c\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"98b0c36065208c15f7f1c5b955ac71cb49aed89f19332ea5e1ab96a0d00c8efa\"" Jan 17 12:20:49.659854 containerd[1544]: time="2025-01-17T12:20:49.659496606Z" level=info msg="StartContainer for \"98b0c36065208c15f7f1c5b955ac71cb49aed89f19332ea5e1ab96a0d00c8efa\"" Jan 17 12:20:49.714994 systemd[1]: Started cri-containerd-98b0c36065208c15f7f1c5b955ac71cb49aed89f19332ea5e1ab96a0d00c8efa.scope - libcontainer container 98b0c36065208c15f7f1c5b955ac71cb49aed89f19332ea5e1ab96a0d00c8efa. Jan 17 12:20:49.736251 containerd[1544]: time="2025-01-17T12:20:49.736218655Z" level=info msg="StartContainer for \"98b0c36065208c15f7f1c5b955ac71cb49aed89f19332ea5e1ab96a0d00c8efa\" returns successfully" Jan 17 12:20:51.140799 systemd[1]: cri-containerd-98b0c36065208c15f7f1c5b955ac71cb49aed89f19332ea5e1ab96a0d00c8efa.scope: Deactivated successfully. Jan 17 12:20:51.163391 kubelet[2723]: I0117 12:20:51.163016 2723 kubelet_node_status.go:488] "Fast updating node status as it just became ready" Jan 17 12:20:51.168409 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-98b0c36065208c15f7f1c5b955ac71cb49aed89f19332ea5e1ab96a0d00c8efa-rootfs.mount: Deactivated successfully. Jan 17 12:20:51.170213 containerd[1544]: time="2025-01-17T12:20:51.170165770Z" level=info msg="shim disconnected" id=98b0c36065208c15f7f1c5b955ac71cb49aed89f19332ea5e1ab96a0d00c8efa namespace=k8s.io Jan 17 12:20:51.171424 containerd[1544]: time="2025-01-17T12:20:51.170464857Z" level=warning msg="cleaning up after shim disconnected" id=98b0c36065208c15f7f1c5b955ac71cb49aed89f19332ea5e1ab96a0d00c8efa namespace=k8s.io Jan 17 12:20:51.171424 containerd[1544]: time="2025-01-17T12:20:51.170478063Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 17 12:20:51.214076 kubelet[2723]: W0117 12:20:51.213166 2723 reflector.go:561] object-"calico-apiserver"/"calico-apiserver-certs": failed to list *v1.Secret: secrets "calico-apiserver-certs" is forbidden: User "system:node:localhost" cannot list resource "secrets" in API group "" in the namespace "calico-apiserver": no relationship found between node 'localhost' and this object Jan 17 12:20:51.213989 systemd[1]: Created slice kubepods-burstable-pod7756b0ac_5d50_42a5_ba21_9f884a8f9486.slice - libcontainer container kubepods-burstable-pod7756b0ac_5d50_42a5_ba21_9f884a8f9486.slice. Jan 17 12:20:51.217523 kubelet[2723]: E0117 12:20:51.214994 2723 reflector.go:158] "Unhandled Error" err="object-\"calico-apiserver\"/\"calico-apiserver-certs\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"calico-apiserver-certs\" is forbidden: User \"system:node:localhost\" cannot list resource \"secrets\" in API group \"\" in the namespace \"calico-apiserver\": no relationship found between node 'localhost' and this object" logger="UnhandledError" Jan 17 12:20:51.217523 kubelet[2723]: W0117 12:20:51.214683 2723 reflector.go:561] object-"calico-apiserver"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:localhost" cannot list resource "configmaps" in API group "" in the namespace "calico-apiserver": no relationship found between node 'localhost' and this object Jan 17 12:20:51.217523 kubelet[2723]: E0117 12:20:51.215015 2723 reflector.go:158] "Unhandled Error" err="object-\"calico-apiserver\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:localhost\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"calico-apiserver\": no relationship found between node 'localhost' and this object" logger="UnhandledError" Jan 17 12:20:51.218244 systemd[1]: Created slice kubepods-burstable-pod95fe1c4f_c274_4a0d_aa79_7b5cb4353bdf.slice - libcontainer container kubepods-burstable-pod95fe1c4f_c274_4a0d_aa79_7b5cb4353bdf.slice. Jan 17 12:20:51.221947 systemd[1]: Created slice kubepods-besteffort-pod9c19e775_a72c_40f9_9f63_2947d850da10.slice - libcontainer container kubepods-besteffort-pod9c19e775_a72c_40f9_9f63_2947d850da10.slice. Jan 17 12:20:51.226492 systemd[1]: Created slice kubepods-besteffort-pod0c7fe481_2df4_4da8_b215_b904f488836f.slice - libcontainer container kubepods-besteffort-pod0c7fe481_2df4_4da8_b215_b904f488836f.slice. Jan 17 12:20:51.231202 systemd[1]: Created slice kubepods-besteffort-podcb76a1e6_a555_4ec5_bb3d_c19efe7e0c24.slice - libcontainer container kubepods-besteffort-podcb76a1e6_a555_4ec5_bb3d_c19efe7e0c24.slice. Jan 17 12:20:51.248489 kubelet[2723]: I0117 12:20:51.248388 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4mg4\" (UniqueName: \"kubernetes.io/projected/95fe1c4f-c274-4a0d-aa79-7b5cb4353bdf-kube-api-access-k4mg4\") pod \"coredns-6f6b679f8f-bq2mf\" (UID: \"95fe1c4f-c274-4a0d-aa79-7b5cb4353bdf\") " pod="kube-system/coredns-6f6b679f8f-bq2mf" Jan 17 12:20:51.248489 kubelet[2723]: I0117 12:20:51.248425 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7756b0ac-5d50-42a5-ba21-9f884a8f9486-config-volume\") pod \"coredns-6f6b679f8f-6f26n\" (UID: \"7756b0ac-5d50-42a5-ba21-9f884a8f9486\") " pod="kube-system/coredns-6f6b679f8f-6f26n" Jan 17 12:20:51.248489 kubelet[2723]: I0117 12:20:51.248439 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nknx\" (UniqueName: \"kubernetes.io/projected/7756b0ac-5d50-42a5-ba21-9f884a8f9486-kube-api-access-8nknx\") pod \"coredns-6f6b679f8f-6f26n\" (UID: \"7756b0ac-5d50-42a5-ba21-9f884a8f9486\") " pod="kube-system/coredns-6f6b679f8f-6f26n" Jan 17 12:20:51.248489 kubelet[2723]: I0117 12:20:51.248451 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/95fe1c4f-c274-4a0d-aa79-7b5cb4353bdf-config-volume\") pod \"coredns-6f6b679f8f-bq2mf\" (UID: \"95fe1c4f-c274-4a0d-aa79-7b5cb4353bdf\") " pod="kube-system/coredns-6f6b679f8f-bq2mf" Jan 17 12:20:51.349659 kubelet[2723]: I0117 12:20:51.348825 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/9c19e775-a72c-40f9-9f63-2947d850da10-calico-apiserver-certs\") pod \"calico-apiserver-6b697d6776-n97bk\" (UID: \"9c19e775-a72c-40f9-9f63-2947d850da10\") " pod="calico-apiserver/calico-apiserver-6b697d6776-n97bk" Jan 17 12:20:51.349659 kubelet[2723]: I0117 12:20:51.348863 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mr2hb\" (UniqueName: \"kubernetes.io/projected/0c7fe481-2df4-4da8-b215-b904f488836f-kube-api-access-mr2hb\") pod \"calico-kube-controllers-784b5c8b57-wvh2g\" (UID: \"0c7fe481-2df4-4da8-b215-b904f488836f\") " pod="calico-system/calico-kube-controllers-784b5c8b57-wvh2g" Jan 17 12:20:51.349659 kubelet[2723]: I0117 12:20:51.348879 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbjj7\" (UniqueName: \"kubernetes.io/projected/9c19e775-a72c-40f9-9f63-2947d850da10-kube-api-access-zbjj7\") pod \"calico-apiserver-6b697d6776-n97bk\" (UID: \"9c19e775-a72c-40f9-9f63-2947d850da10\") " pod="calico-apiserver/calico-apiserver-6b697d6776-n97bk" Jan 17 12:20:51.349659 kubelet[2723]: I0117 12:20:51.348894 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzqzn\" (UniqueName: \"kubernetes.io/projected/cb76a1e6-a555-4ec5-bb3d-c19efe7e0c24-kube-api-access-mzqzn\") pod \"calico-apiserver-6b697d6776-wzkmk\" (UID: \"cb76a1e6-a555-4ec5-bb3d-c19efe7e0c24\") " pod="calico-apiserver/calico-apiserver-6b697d6776-wzkmk" Jan 17 12:20:51.349659 kubelet[2723]: I0117 12:20:51.348909 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/cb76a1e6-a555-4ec5-bb3d-c19efe7e0c24-calico-apiserver-certs\") pod \"calico-apiserver-6b697d6776-wzkmk\" (UID: \"cb76a1e6-a555-4ec5-bb3d-c19efe7e0c24\") " pod="calico-apiserver/calico-apiserver-6b697d6776-wzkmk" Jan 17 12:20:51.349857 kubelet[2723]: I0117 12:20:51.348929 2723 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0c7fe481-2df4-4da8-b215-b904f488836f-tigera-ca-bundle\") pod \"calico-kube-controllers-784b5c8b57-wvh2g\" (UID: \"0c7fe481-2df4-4da8-b215-b904f488836f\") " pod="calico-system/calico-kube-controllers-784b5c8b57-wvh2g" Jan 17 12:20:51.525315 containerd[1544]: time="2025-01-17T12:20:51.524855425Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-bq2mf,Uid:95fe1c4f-c274-4a0d-aa79-7b5cb4353bdf,Namespace:kube-system,Attempt:0,}" Jan 17 12:20:51.525417 containerd[1544]: time="2025-01-17T12:20:51.524871522Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-6f26n,Uid:7756b0ac-5d50-42a5-ba21-9f884a8f9486,Namespace:kube-system,Attempt:0,}" Jan 17 12:20:51.531241 containerd[1544]: time="2025-01-17T12:20:51.531211633Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-784b5c8b57-wvh2g,Uid:0c7fe481-2df4-4da8-b215-b904f488836f,Namespace:calico-system,Attempt:0,}" Jan 17 12:20:51.594400 systemd[1]: Created slice kubepods-besteffort-podd5c6d7d3_a76d_4ec7_b919_828df6f31125.slice - libcontainer container kubepods-besteffort-podd5c6d7d3_a76d_4ec7_b919_828df6f31125.slice. Jan 17 12:20:51.596195 containerd[1544]: time="2025-01-17T12:20:51.596176557Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-zj6v2,Uid:d5c6d7d3-a76d-4ec7-b919-828df6f31125,Namespace:calico-system,Attempt:0,}" Jan 17 12:20:51.690760 containerd[1544]: time="2025-01-17T12:20:51.690699535Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.1\"" Jan 17 12:20:51.764269 containerd[1544]: time="2025-01-17T12:20:51.764188354Z" level=error msg="Failed to destroy network for sandbox \"d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 17 12:20:51.764516 containerd[1544]: time="2025-01-17T12:20:51.764446683Z" level=error msg="Failed to destroy network for sandbox \"94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 17 12:20:51.767351 containerd[1544]: time="2025-01-17T12:20:51.767278534Z" level=error msg="encountered an error cleaning up failed sandbox \"94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 17 12:20:51.767482 containerd[1544]: time="2025-01-17T12:20:51.767338462Z" level=error msg="Failed to destroy network for sandbox \"711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 17 12:20:51.767656 containerd[1544]: time="2025-01-17T12:20:51.767642557Z" level=error msg="encountered an error cleaning up failed sandbox \"711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 17 12:20:51.767771 containerd[1544]: time="2025-01-17T12:20:51.767715534Z" level=error msg="Failed to destroy network for sandbox \"81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 17 12:20:51.767934 containerd[1544]: time="2025-01-17T12:20:51.767921975Z" level=error msg="encountered an error cleaning up failed sandbox \"81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 17 12:20:51.768062 containerd[1544]: time="2025-01-17T12:20:51.767279604Z" level=error msg="encountered an error cleaning up failed sandbox \"d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 17 12:20:51.775903 containerd[1544]: time="2025-01-17T12:20:51.775752090Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-784b5c8b57-wvh2g,Uid:0c7fe481-2df4-4da8-b215-b904f488836f,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 17 12:20:51.776004 containerd[1544]: time="2025-01-17T12:20:51.775803104Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-zj6v2,Uid:d5c6d7d3-a76d-4ec7-b919-828df6f31125,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 17 12:20:51.779439 containerd[1544]: time="2025-01-17T12:20:51.775814908Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-bq2mf,Uid:95fe1c4f-c274-4a0d-aa79-7b5cb4353bdf,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 17 12:20:51.779439 containerd[1544]: time="2025-01-17T12:20:51.775822553Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-6f26n,Uid:7756b0ac-5d50-42a5-ba21-9f884a8f9486,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 17 12:20:51.780084 kubelet[2723]: E0117 12:20:51.779774 2723 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 17 12:20:51.780084 kubelet[2723]: E0117 12:20:51.779822 2723 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-6f26n" Jan 17 12:20:51.780084 kubelet[2723]: E0117 12:20:51.779836 2723 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-6f26n" Jan 17 12:20:51.780428 kubelet[2723]: E0117 12:20:51.779861 2723 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-6f6b679f8f-6f26n_kube-system(7756b0ac-5d50-42a5-ba21-9f884a8f9486)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-6f6b679f8f-6f26n_kube-system(7756b0ac-5d50-42a5-ba21-9f884a8f9486)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-6f6b679f8f-6f26n" podUID="7756b0ac-5d50-42a5-ba21-9f884a8f9486" Jan 17 12:20:51.780428 kubelet[2723]: E0117 12:20:51.779957 2723 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 17 12:20:51.780428 kubelet[2723]: E0117 12:20:51.779979 2723 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 17 12:20:51.780428 kubelet[2723]: E0117 12:20:51.779982 2723 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-zj6v2" Jan 17 12:20:51.780526 kubelet[2723]: E0117 12:20:51.779992 2723 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-bq2mf" Jan 17 12:20:51.780526 kubelet[2723]: E0117 12:20:51.780001 2723 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-bq2mf" Jan 17 12:20:51.780526 kubelet[2723]: E0117 12:20:51.780002 2723 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 17 12:20:51.780526 kubelet[2723]: E0117 12:20:51.780012 2723 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-784b5c8b57-wvh2g" Jan 17 12:20:51.780613 kubelet[2723]: E0117 12:20:51.780014 2723 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-6f6b679f8f-bq2mf_kube-system(95fe1c4f-c274-4a0d-aa79-7b5cb4353bdf)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-6f6b679f8f-bq2mf_kube-system(95fe1c4f-c274-4a0d-aa79-7b5cb4353bdf)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-6f6b679f8f-bq2mf" podUID="95fe1c4f-c274-4a0d-aa79-7b5cb4353bdf" Jan 17 12:20:51.780613 kubelet[2723]: E0117 12:20:51.780027 2723 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-784b5c8b57-wvh2g" Jan 17 12:20:51.780613 kubelet[2723]: E0117 12:20:51.780047 2723 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-784b5c8b57-wvh2g_calico-system(0c7fe481-2df4-4da8-b215-b904f488836f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-784b5c8b57-wvh2g_calico-system(0c7fe481-2df4-4da8-b215-b904f488836f)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-784b5c8b57-wvh2g" podUID="0c7fe481-2df4-4da8-b215-b904f488836f" Jan 17 12:20:51.780715 kubelet[2723]: E0117 12:20:51.779993 2723 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-zj6v2" Jan 17 12:20:51.780715 kubelet[2723]: E0117 12:20:51.780089 2723 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-zj6v2_calico-system(d5c6d7d3-a76d-4ec7-b919-828df6f31125)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-zj6v2_calico-system(d5c6d7d3-a76d-4ec7-b919-828df6f31125)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-zj6v2" podUID="d5c6d7d3-a76d-4ec7-b919-828df6f31125" Jan 17 12:20:52.167479 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef-shm.mount: Deactivated successfully. Jan 17 12:20:52.167534 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1-shm.mount: Deactivated successfully. Jan 17 12:20:52.449774 kubelet[2723]: E0117 12:20:52.449639 2723 secret.go:188] Couldn't get secret calico-apiserver/calico-apiserver-certs: failed to sync secret cache: timed out waiting for the condition Jan 17 12:20:52.449774 kubelet[2723]: E0117 12:20:52.449722 2723 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cb76a1e6-a555-4ec5-bb3d-c19efe7e0c24-calico-apiserver-certs podName:cb76a1e6-a555-4ec5-bb3d-c19efe7e0c24 nodeName:}" failed. No retries permitted until 2025-01-17 12:20:52.94970674 +0000 UTC m=+25.433179338 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "calico-apiserver-certs" (UniqueName: "kubernetes.io/secret/cb76a1e6-a555-4ec5-bb3d-c19efe7e0c24-calico-apiserver-certs") pod "calico-apiserver-6b697d6776-wzkmk" (UID: "cb76a1e6-a555-4ec5-bb3d-c19efe7e0c24") : failed to sync secret cache: timed out waiting for the condition Jan 17 12:20:52.450709 kubelet[2723]: E0117 12:20:52.450694 2723 secret.go:188] Couldn't get secret calico-apiserver/calico-apiserver-certs: failed to sync secret cache: timed out waiting for the condition Jan 17 12:20:52.450745 kubelet[2723]: E0117 12:20:52.450728 2723 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9c19e775-a72c-40f9-9f63-2947d850da10-calico-apiserver-certs podName:9c19e775-a72c-40f9-9f63-2947d850da10 nodeName:}" failed. No retries permitted until 2025-01-17 12:20:52.950720259 +0000 UTC m=+25.434192842 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "calico-apiserver-certs" (UniqueName: "kubernetes.io/secret/9c19e775-a72c-40f9-9f63-2947d850da10-calico-apiserver-certs") pod "calico-apiserver-6b697d6776-n97bk" (UID: "9c19e775-a72c-40f9-9f63-2947d850da10") : failed to sync secret cache: timed out waiting for the condition Jan 17 12:20:52.678839 kubelet[2723]: I0117 12:20:52.678787 2723 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" Jan 17 12:20:52.680289 kubelet[2723]: I0117 12:20:52.680277 2723 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" Jan 17 12:20:52.693983 kubelet[2723]: I0117 12:20:52.693587 2723 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" Jan 17 12:20:52.694657 kubelet[2723]: I0117 12:20:52.694588 2723 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" Jan 17 12:20:52.713229 containerd[1544]: time="2025-01-17T12:20:52.713179622Z" level=info msg="StopPodSandbox for \"d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1\"" Jan 17 12:20:52.714499 containerd[1544]: time="2025-01-17T12:20:52.713704261Z" level=info msg="StopPodSandbox for \"711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21\"" Jan 17 12:20:52.714499 containerd[1544]: time="2025-01-17T12:20:52.714241957Z" level=info msg="Ensure that sandbox d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1 in task-service has been cleanup successfully" Jan 17 12:20:52.715519 containerd[1544]: time="2025-01-17T12:20:52.715508915Z" level=info msg="StopPodSandbox for \"81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef\"" Jan 17 12:20:52.715689 containerd[1544]: time="2025-01-17T12:20:52.715664484Z" level=info msg="Ensure that sandbox 81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef in task-service has been cleanup successfully" Jan 17 12:20:52.716712 containerd[1544]: time="2025-01-17T12:20:52.716696764Z" level=info msg="Ensure that sandbox 711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21 in task-service has been cleanup successfully" Jan 17 12:20:52.717975 containerd[1544]: time="2025-01-17T12:20:52.717962890Z" level=info msg="StopPodSandbox for \"94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe\"" Jan 17 12:20:52.718158 containerd[1544]: time="2025-01-17T12:20:52.718147441Z" level=info msg="Ensure that sandbox 94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe in task-service has been cleanup successfully" Jan 17 12:20:52.745582 containerd[1544]: time="2025-01-17T12:20:52.745549335Z" level=error msg="StopPodSandbox for \"81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef\" failed" error="failed to destroy network for sandbox \"81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 17 12:20:52.745888 kubelet[2723]: E0117 12:20:52.745862 2723 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" Jan 17 12:20:52.746000 kubelet[2723]: E0117 12:20:52.745968 2723 kuberuntime_manager.go:1477] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef"} Jan 17 12:20:52.746096 containerd[1544]: time="2025-01-17T12:20:52.746082224Z" level=error msg="StopPodSandbox for \"711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21\" failed" error="failed to destroy network for sandbox \"711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 17 12:20:52.746161 kubelet[2723]: E0117 12:20:52.746146 2723 kuberuntime_manager.go:1077] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"95fe1c4f-c274-4a0d-aa79-7b5cb4353bdf\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Jan 17 12:20:52.746304 kubelet[2723]: E0117 12:20:52.746241 2723 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"95fe1c4f-c274-4a0d-aa79-7b5cb4353bdf\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-6f6b679f8f-bq2mf" podUID="95fe1c4f-c274-4a0d-aa79-7b5cb4353bdf" Jan 17 12:20:52.746650 kubelet[2723]: E0117 12:20:52.746595 2723 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" Jan 17 12:20:52.746650 kubelet[2723]: E0117 12:20:52.746611 2723 kuberuntime_manager.go:1477] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21"} Jan 17 12:20:52.746650 kubelet[2723]: E0117 12:20:52.746625 2723 kuberuntime_manager.go:1077] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"d5c6d7d3-a76d-4ec7-b919-828df6f31125\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Jan 17 12:20:52.746650 kubelet[2723]: E0117 12:20:52.746637 2723 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"d5c6d7d3-a76d-4ec7-b919-828df6f31125\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-zj6v2" podUID="d5c6d7d3-a76d-4ec7-b919-828df6f31125" Jan 17 12:20:52.748877 containerd[1544]: time="2025-01-17T12:20:52.748784502Z" level=error msg="StopPodSandbox for \"d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1\" failed" error="failed to destroy network for sandbox \"d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 17 12:20:52.748998 kubelet[2723]: E0117 12:20:52.748937 2723 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" Jan 17 12:20:52.748998 kubelet[2723]: E0117 12:20:52.748956 2723 kuberuntime_manager.go:1477] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1"} Jan 17 12:20:52.748998 kubelet[2723]: E0117 12:20:52.748972 2723 kuberuntime_manager.go:1077] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"7756b0ac-5d50-42a5-ba21-9f884a8f9486\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Jan 17 12:20:52.748998 kubelet[2723]: E0117 12:20:52.748983 2723 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"7756b0ac-5d50-42a5-ba21-9f884a8f9486\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-6f6b679f8f-6f26n" podUID="7756b0ac-5d50-42a5-ba21-9f884a8f9486" Jan 17 12:20:52.754128 containerd[1544]: time="2025-01-17T12:20:52.754092157Z" level=error msg="StopPodSandbox for \"94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe\" failed" error="failed to destroy network for sandbox \"94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 17 12:20:52.754597 kubelet[2723]: E0117 12:20:52.754464 2723 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" Jan 17 12:20:52.754597 kubelet[2723]: E0117 12:20:52.754509 2723 kuberuntime_manager.go:1477] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe"} Jan 17 12:20:52.754597 kubelet[2723]: E0117 12:20:52.754529 2723 kuberuntime_manager.go:1077] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"0c7fe481-2df4-4da8-b215-b904f488836f\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Jan 17 12:20:52.754597 kubelet[2723]: E0117 12:20:52.754543 2723 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"0c7fe481-2df4-4da8-b215-b904f488836f\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-784b5c8b57-wvh2g" podUID="0c7fe481-2df4-4da8-b215-b904f488836f" Jan 17 12:20:53.028232 containerd[1544]: time="2025-01-17T12:20:53.027915272Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6b697d6776-n97bk,Uid:9c19e775-a72c-40f9-9f63-2947d850da10,Namespace:calico-apiserver,Attempt:0,}" Jan 17 12:20:53.038139 containerd[1544]: time="2025-01-17T12:20:53.038119681Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6b697d6776-wzkmk,Uid:cb76a1e6-a555-4ec5-bb3d-c19efe7e0c24,Namespace:calico-apiserver,Attempt:0,}" Jan 17 12:20:53.104459 containerd[1544]: time="2025-01-17T12:20:53.104431165Z" level=error msg="Failed to destroy network for sandbox \"fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 17 12:20:53.104750 containerd[1544]: time="2025-01-17T12:20:53.104734121Z" level=error msg="encountered an error cleaning up failed sandbox \"fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 17 12:20:53.104830 containerd[1544]: time="2025-01-17T12:20:53.104815556Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6b697d6776-n97bk,Uid:9c19e775-a72c-40f9-9f63-2947d850da10,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 17 12:20:53.105035 kubelet[2723]: E0117 12:20:53.105016 2723 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 17 12:20:53.105195 kubelet[2723]: E0117 12:20:53.105182 2723 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-6b697d6776-n97bk" Jan 17 12:20:53.105258 kubelet[2723]: E0117 12:20:53.105248 2723 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-6b697d6776-n97bk" Jan 17 12:20:53.105865 kubelet[2723]: E0117 12:20:53.105849 2723 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-6b697d6776-n97bk_calico-apiserver(9c19e775-a72c-40f9-9f63-2947d850da10)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-6b697d6776-n97bk_calico-apiserver(9c19e775-a72c-40f9-9f63-2947d850da10)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-6b697d6776-n97bk" podUID="9c19e775-a72c-40f9-9f63-2947d850da10" Jan 17 12:20:53.108784 containerd[1544]: time="2025-01-17T12:20:53.108759557Z" level=error msg="Failed to destroy network for sandbox \"977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 17 12:20:53.109058 containerd[1544]: time="2025-01-17T12:20:53.109043688Z" level=error msg="encountered an error cleaning up failed sandbox \"977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 17 12:20:53.109691 containerd[1544]: time="2025-01-17T12:20:53.109121137Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6b697d6776-wzkmk,Uid:cb76a1e6-a555-4ec5-bb3d-c19efe7e0c24,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 17 12:20:53.109747 kubelet[2723]: E0117 12:20:53.109214 2723 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 17 12:20:53.109747 kubelet[2723]: E0117 12:20:53.109245 2723 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-6b697d6776-wzkmk" Jan 17 12:20:53.109747 kubelet[2723]: E0117 12:20:53.109257 2723 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-6b697d6776-wzkmk" Jan 17 12:20:53.109825 kubelet[2723]: E0117 12:20:53.109281 2723 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-6b697d6776-wzkmk_calico-apiserver(cb76a1e6-a555-4ec5-bb3d-c19efe7e0c24)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-6b697d6776-wzkmk_calico-apiserver(cb76a1e6-a555-4ec5-bb3d-c19efe7e0c24)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-6b697d6776-wzkmk" podUID="cb76a1e6-a555-4ec5-bb3d-c19efe7e0c24" Jan 17 12:20:53.167814 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4-shm.mount: Deactivated successfully. Jan 17 12:20:53.697115 containerd[1544]: time="2025-01-17T12:20:53.696868868Z" level=info msg="StopPodSandbox for \"977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598\"" Jan 17 12:20:53.697115 containerd[1544]: time="2025-01-17T12:20:53.697063350Z" level=info msg="Ensure that sandbox 977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598 in task-service has been cleanup successfully" Jan 17 12:20:53.697425 kubelet[2723]: I0117 12:20:53.696470 2723 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" Jan 17 12:20:53.698214 kubelet[2723]: I0117 12:20:53.697854 2723 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" Jan 17 12:20:53.698259 containerd[1544]: time="2025-01-17T12:20:53.698107995Z" level=info msg="StopPodSandbox for \"fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4\"" Jan 17 12:20:53.698369 containerd[1544]: time="2025-01-17T12:20:53.698206053Z" level=info msg="Ensure that sandbox fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4 in task-service has been cleanup successfully" Jan 17 12:20:53.717738 containerd[1544]: time="2025-01-17T12:20:53.717706778Z" level=error msg="StopPodSandbox for \"977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598\" failed" error="failed to destroy network for sandbox \"977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 17 12:20:53.717995 kubelet[2723]: E0117 12:20:53.717839 2723 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" Jan 17 12:20:53.717995 kubelet[2723]: E0117 12:20:53.717869 2723 kuberuntime_manager.go:1477] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598"} Jan 17 12:20:53.717995 kubelet[2723]: E0117 12:20:53.717892 2723 kuberuntime_manager.go:1077] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"cb76a1e6-a555-4ec5-bb3d-c19efe7e0c24\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Jan 17 12:20:53.717995 kubelet[2723]: E0117 12:20:53.717905 2723 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"cb76a1e6-a555-4ec5-bb3d-c19efe7e0c24\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-6b697d6776-wzkmk" podUID="cb76a1e6-a555-4ec5-bb3d-c19efe7e0c24" Jan 17 12:20:53.736474 containerd[1544]: time="2025-01-17T12:20:53.736446990Z" level=error msg="StopPodSandbox for \"fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4\" failed" error="failed to destroy network for sandbox \"fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 17 12:20:53.736739 kubelet[2723]: E0117 12:20:53.736710 2723 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" Jan 17 12:20:53.736787 kubelet[2723]: E0117 12:20:53.736750 2723 kuberuntime_manager.go:1477] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4"} Jan 17 12:20:53.736833 kubelet[2723]: E0117 12:20:53.736781 2723 kuberuntime_manager.go:1077] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"9c19e775-a72c-40f9-9f63-2947d850da10\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Jan 17 12:20:53.736833 kubelet[2723]: E0117 12:20:53.736803 2723 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"9c19e775-a72c-40f9-9f63-2947d850da10\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-6b697d6776-n97bk" podUID="9c19e775-a72c-40f9-9f63-2947d850da10" Jan 17 12:20:56.033208 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount907523925.mount: Deactivated successfully. Jan 17 12:20:56.470378 containerd[1544]: time="2025-01-17T12:20:56.470336217Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.29.1: active requests=0, bytes read=142742010" Jan 17 12:20:56.476021 containerd[1544]: time="2025-01-17T12:20:56.470390770Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:20:56.481614 containerd[1544]: time="2025-01-17T12:20:56.481583495Z" level=info msg="ImageCreate event name:\"sha256:feb26d4585d68e875d9bd9bd6c27ea9f2d5c9ed9ef70f8b8cb0ebb0559a1d664\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:20:56.502698 containerd[1544]: time="2025-01-17T12:20:56.502661880Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:99c3917516efe1f807a0cfdf2d14b628b7c5cc6bd8a9ee5a253154f31756bea1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:20:56.508944 containerd[1544]: time="2025-01-17T12:20:56.506114959Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.29.1\" with image id \"sha256:feb26d4585d68e875d9bd9bd6c27ea9f2d5c9ed9ef70f8b8cb0ebb0559a1d664\", repo tag \"ghcr.io/flatcar/calico/node:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/node@sha256:99c3917516efe1f807a0cfdf2d14b628b7c5cc6bd8a9ee5a253154f31756bea1\", size \"142741872\" in 4.812818411s" Jan 17 12:20:56.508944 containerd[1544]: time="2025-01-17T12:20:56.506137283Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.1\" returns image reference \"sha256:feb26d4585d68e875d9bd9bd6c27ea9f2d5c9ed9ef70f8b8cb0ebb0559a1d664\"" Jan 17 12:20:56.784670 containerd[1544]: time="2025-01-17T12:20:56.784476417Z" level=info msg="CreateContainer within sandbox \"31098533b87e8c32bd5c7953206fce8bc0493676630ea682290d7f0996cd132c\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Jan 17 12:20:56.899100 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount598338799.mount: Deactivated successfully. Jan 17 12:20:56.905593 containerd[1544]: time="2025-01-17T12:20:56.905543995Z" level=info msg="CreateContainer within sandbox \"31098533b87e8c32bd5c7953206fce8bc0493676630ea682290d7f0996cd132c\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"bdd3e8c0d8dc609233a7628780604ac329962b996666ed0d28c27db3140b9f1c\"" Jan 17 12:20:56.908443 containerd[1544]: time="2025-01-17T12:20:56.908423874Z" level=info msg="StartContainer for \"bdd3e8c0d8dc609233a7628780604ac329962b996666ed0d28c27db3140b9f1c\"" Jan 17 12:20:57.103437 systemd[1]: run-containerd-runc-k8s.io-bdd3e8c0d8dc609233a7628780604ac329962b996666ed0d28c27db3140b9f1c-runc.2xE0dB.mount: Deactivated successfully. Jan 17 12:20:57.112758 systemd[1]: Started cri-containerd-bdd3e8c0d8dc609233a7628780604ac329962b996666ed0d28c27db3140b9f1c.scope - libcontainer container bdd3e8c0d8dc609233a7628780604ac329962b996666ed0d28c27db3140b9f1c. Jan 17 12:20:57.138155 containerd[1544]: time="2025-01-17T12:20:57.138087645Z" level=info msg="StartContainer for \"bdd3e8c0d8dc609233a7628780604ac329962b996666ed0d28c27db3140b9f1c\" returns successfully" Jan 17 12:20:57.542721 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Jan 17 12:20:57.544755 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld <Jason@zx2c4.com>. All Rights Reserved. Jan 17 12:20:58.092629 kubelet[2723]: I0117 12:20:58.078077 2723 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-d5fqz" podStartSLOduration=2.340183617 podStartE2EDuration="16.949479127s" podCreationTimestamp="2025-01-17 12:20:41 +0000 UTC" firstStartedPulling="2025-01-17 12:20:41.960583353 +0000 UTC m=+14.444055938" lastFinishedPulling="2025-01-17 12:20:56.569878864 +0000 UTC m=+29.053351448" observedRunningTime="2025-01-17 12:20:57.84048207 +0000 UTC m=+30.323954664" watchObservedRunningTime="2025-01-17 12:20:57.949479127 +0000 UTC m=+30.432951721" Jan 17 12:21:03.585543 containerd[1544]: time="2025-01-17T12:21:03.585502687Z" level=info msg="StopPodSandbox for \"711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21\"" Jan 17 12:21:03.886696 containerd[1544]: 2025-01-17 12:21:03.646 [INFO][4143] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" Jan 17 12:21:03.886696 containerd[1544]: 2025-01-17 12:21:03.647 [INFO][4143] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" iface="eth0" netns="/var/run/netns/cni-409ddbcb-7173-8c43-6452-843b23522320" Jan 17 12:21:03.886696 containerd[1544]: 2025-01-17 12:21:03.647 [INFO][4143] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" iface="eth0" netns="/var/run/netns/cni-409ddbcb-7173-8c43-6452-843b23522320" Jan 17 12:21:03.886696 containerd[1544]: 2025-01-17 12:21:03.651 [INFO][4143] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" iface="eth0" netns="/var/run/netns/cni-409ddbcb-7173-8c43-6452-843b23522320" Jan 17 12:21:03.886696 containerd[1544]: 2025-01-17 12:21:03.651 [INFO][4143] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" Jan 17 12:21:03.886696 containerd[1544]: 2025-01-17 12:21:03.651 [INFO][4143] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" Jan 17 12:21:03.886696 containerd[1544]: 2025-01-17 12:21:03.861 [INFO][4149] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" HandleID="k8s-pod-network.711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" Workload="localhost-k8s-csi--node--driver--zj6v2-eth0" Jan 17 12:21:03.886696 containerd[1544]: 2025-01-17 12:21:03.866 [INFO][4149] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jan 17 12:21:03.886696 containerd[1544]: 2025-01-17 12:21:03.867 [INFO][4149] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jan 17 12:21:03.886696 containerd[1544]: 2025-01-17 12:21:03.881 [WARNING][4149] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" HandleID="k8s-pod-network.711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" Workload="localhost-k8s-csi--node--driver--zj6v2-eth0" Jan 17 12:21:03.886696 containerd[1544]: 2025-01-17 12:21:03.881 [INFO][4149] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" HandleID="k8s-pod-network.711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" Workload="localhost-k8s-csi--node--driver--zj6v2-eth0" Jan 17 12:21:03.886696 containerd[1544]: 2025-01-17 12:21:03.882 [INFO][4149] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jan 17 12:21:03.886696 containerd[1544]: 2025-01-17 12:21:03.883 [INFO][4143] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" Jan 17 12:21:03.886546 systemd[1]: run-netns-cni\x2d409ddbcb\x2d7173\x2d8c43\x2d6452\x2d843b23522320.mount: Deactivated successfully. Jan 17 12:21:03.890856 containerd[1544]: time="2025-01-17T12:21:03.890802649Z" level=info msg="TearDown network for sandbox \"711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21\" successfully" Jan 17 12:21:03.890856 containerd[1544]: time="2025-01-17T12:21:03.890828117Z" level=info msg="StopPodSandbox for \"711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21\" returns successfully" Jan 17 12:21:03.891528 containerd[1544]: time="2025-01-17T12:21:03.891445439Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-zj6v2,Uid:d5c6d7d3-a76d-4ec7-b919-828df6f31125,Namespace:calico-system,Attempt:1,}" Jan 17 12:21:03.988033 systemd-networkd[1453]: calic5bb7ed3b64: Link UP Jan 17 12:21:03.988133 systemd-networkd[1453]: calic5bb7ed3b64: Gained carrier Jan 17 12:21:04.000975 containerd[1544]: 2025-01-17 12:21:03.916 [INFO][4158] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Jan 17 12:21:04.000975 containerd[1544]: 2025-01-17 12:21:03.926 [INFO][4158] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-csi--node--driver--zj6v2-eth0 csi-node-driver- calico-system d5c6d7d3-a76d-4ec7-b919-828df6f31125 755 0 2025-01-17 12:20:41 +0000 UTC <nil> <nil> map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:56747c9949 k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s localhost csi-node-driver-zj6v2 eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] calic5bb7ed3b64 [] []}} ContainerID="775d8d469d2ab031a2239f0071ad078481fb2c0911d17de24c2d4477a636c3c0" Namespace="calico-system" Pod="csi-node-driver-zj6v2" WorkloadEndpoint="localhost-k8s-csi--node--driver--zj6v2-" Jan 17 12:21:04.000975 containerd[1544]: 2025-01-17 12:21:03.926 [INFO][4158] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="775d8d469d2ab031a2239f0071ad078481fb2c0911d17de24c2d4477a636c3c0" Namespace="calico-system" Pod="csi-node-driver-zj6v2" WorkloadEndpoint="localhost-k8s-csi--node--driver--zj6v2-eth0" Jan 17 12:21:04.000975 containerd[1544]: 2025-01-17 12:21:03.952 [INFO][4169] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="775d8d469d2ab031a2239f0071ad078481fb2c0911d17de24c2d4477a636c3c0" HandleID="k8s-pod-network.775d8d469d2ab031a2239f0071ad078481fb2c0911d17de24c2d4477a636c3c0" Workload="localhost-k8s-csi--node--driver--zj6v2-eth0" Jan 17 12:21:04.000975 containerd[1544]: 2025-01-17 12:21:03.959 [INFO][4169] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="775d8d469d2ab031a2239f0071ad078481fb2c0911d17de24c2d4477a636c3c0" HandleID="k8s-pod-network.775d8d469d2ab031a2239f0071ad078481fb2c0911d17de24c2d4477a636c3c0" Workload="localhost-k8s-csi--node--driver--zj6v2-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000291f00), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"csi-node-driver-zj6v2", "timestamp":"2025-01-17 12:21:03.952052965 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 17 12:21:04.000975 containerd[1544]: 2025-01-17 12:21:03.959 [INFO][4169] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jan 17 12:21:04.000975 containerd[1544]: 2025-01-17 12:21:03.959 [INFO][4169] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jan 17 12:21:04.000975 containerd[1544]: 2025-01-17 12:21:03.959 [INFO][4169] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 17 12:21:04.000975 containerd[1544]: 2025-01-17 12:21:03.961 [INFO][4169] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.775d8d469d2ab031a2239f0071ad078481fb2c0911d17de24c2d4477a636c3c0" host="localhost" Jan 17 12:21:04.000975 containerd[1544]: 2025-01-17 12:21:03.967 [INFO][4169] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" Jan 17 12:21:04.000975 containerd[1544]: 2025-01-17 12:21:03.969 [INFO][4169] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Jan 17 12:21:04.000975 containerd[1544]: 2025-01-17 12:21:03.970 [INFO][4169] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 17 12:21:04.000975 containerd[1544]: 2025-01-17 12:21:03.971 [INFO][4169] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 17 12:21:04.000975 containerd[1544]: 2025-01-17 12:21:03.971 [INFO][4169] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.775d8d469d2ab031a2239f0071ad078481fb2c0911d17de24c2d4477a636c3c0" host="localhost" Jan 17 12:21:04.000975 containerd[1544]: 2025-01-17 12:21:03.972 [INFO][4169] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.775d8d469d2ab031a2239f0071ad078481fb2c0911d17de24c2d4477a636c3c0 Jan 17 12:21:04.000975 containerd[1544]: 2025-01-17 12:21:03.974 [INFO][4169] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.775d8d469d2ab031a2239f0071ad078481fb2c0911d17de24c2d4477a636c3c0" host="localhost" Jan 17 12:21:04.000975 containerd[1544]: 2025-01-17 12:21:03.977 [INFO][4169] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.129/26] block=192.168.88.128/26 handle="k8s-pod-network.775d8d469d2ab031a2239f0071ad078481fb2c0911d17de24c2d4477a636c3c0" host="localhost" Jan 17 12:21:04.000975 containerd[1544]: 2025-01-17 12:21:03.977 [INFO][4169] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.129/26] handle="k8s-pod-network.775d8d469d2ab031a2239f0071ad078481fb2c0911d17de24c2d4477a636c3c0" host="localhost" Jan 17 12:21:04.000975 containerd[1544]: 2025-01-17 12:21:03.977 [INFO][4169] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jan 17 12:21:04.000975 containerd[1544]: 2025-01-17 12:21:03.977 [INFO][4169] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.129/26] IPv6=[] ContainerID="775d8d469d2ab031a2239f0071ad078481fb2c0911d17de24c2d4477a636c3c0" HandleID="k8s-pod-network.775d8d469d2ab031a2239f0071ad078481fb2c0911d17de24c2d4477a636c3c0" Workload="localhost-k8s-csi--node--driver--zj6v2-eth0" Jan 17 12:21:04.001562 containerd[1544]: 2025-01-17 12:21:03.979 [INFO][4158] cni-plugin/k8s.go 386: Populated endpoint ContainerID="775d8d469d2ab031a2239f0071ad078481fb2c0911d17de24c2d4477a636c3c0" Namespace="calico-system" Pod="csi-node-driver-zj6v2" WorkloadEndpoint="localhost-k8s-csi--node--driver--zj6v2-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--zj6v2-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"d5c6d7d3-a76d-4ec7-b919-828df6f31125", ResourceVersion:"755", Generation:0, CreationTimestamp:time.Date(2025, time.January, 17, 12, 20, 41, 0, time.Local), DeletionTimestamp:<nil>, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"56747c9949", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"csi-node-driver-zj6v2", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calic5bb7ed3b64", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 17 12:21:04.001562 containerd[1544]: 2025-01-17 12:21:03.979 [INFO][4158] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.129/32] ContainerID="775d8d469d2ab031a2239f0071ad078481fb2c0911d17de24c2d4477a636c3c0" Namespace="calico-system" Pod="csi-node-driver-zj6v2" WorkloadEndpoint="localhost-k8s-csi--node--driver--zj6v2-eth0" Jan 17 12:21:04.001562 containerd[1544]: 2025-01-17 12:21:03.979 [INFO][4158] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calic5bb7ed3b64 ContainerID="775d8d469d2ab031a2239f0071ad078481fb2c0911d17de24c2d4477a636c3c0" Namespace="calico-system" Pod="csi-node-driver-zj6v2" WorkloadEndpoint="localhost-k8s-csi--node--driver--zj6v2-eth0" Jan 17 12:21:04.001562 containerd[1544]: 2025-01-17 12:21:03.988 [INFO][4158] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="775d8d469d2ab031a2239f0071ad078481fb2c0911d17de24c2d4477a636c3c0" Namespace="calico-system" Pod="csi-node-driver-zj6v2" WorkloadEndpoint="localhost-k8s-csi--node--driver--zj6v2-eth0" Jan 17 12:21:04.001562 containerd[1544]: 2025-01-17 12:21:03.989 [INFO][4158] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="775d8d469d2ab031a2239f0071ad078481fb2c0911d17de24c2d4477a636c3c0" Namespace="calico-system" Pod="csi-node-driver-zj6v2" WorkloadEndpoint="localhost-k8s-csi--node--driver--zj6v2-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--zj6v2-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"d5c6d7d3-a76d-4ec7-b919-828df6f31125", ResourceVersion:"755", Generation:0, CreationTimestamp:time.Date(2025, time.January, 17, 12, 20, 41, 0, time.Local), DeletionTimestamp:<nil>, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"56747c9949", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"775d8d469d2ab031a2239f0071ad078481fb2c0911d17de24c2d4477a636c3c0", Pod:"csi-node-driver-zj6v2", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calic5bb7ed3b64", MAC:"c2:f3:6a:14:ed:81", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 17 12:21:04.001562 containerd[1544]: 2025-01-17 12:21:03.998 [INFO][4158] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="775d8d469d2ab031a2239f0071ad078481fb2c0911d17de24c2d4477a636c3c0" Namespace="calico-system" Pod="csi-node-driver-zj6v2" WorkloadEndpoint="localhost-k8s-csi--node--driver--zj6v2-eth0" Jan 17 12:21:04.023823 containerd[1544]: time="2025-01-17T12:21:04.023748930Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 17 12:21:04.023823 containerd[1544]: time="2025-01-17T12:21:04.023782296Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 17 12:21:04.023823 containerd[1544]: time="2025-01-17T12:21:04.023789811Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 12:21:04.025864 containerd[1544]: time="2025-01-17T12:21:04.023845075Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 12:21:04.045815 systemd[1]: Started cri-containerd-775d8d469d2ab031a2239f0071ad078481fb2c0911d17de24c2d4477a636c3c0.scope - libcontainer container 775d8d469d2ab031a2239f0071ad078481fb2c0911d17de24c2d4477a636c3c0. Jan 17 12:21:04.053802 systemd-resolved[1454]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 17 12:21:04.065648 containerd[1544]: time="2025-01-17T12:21:04.065342747Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-zj6v2,Uid:d5c6d7d3-a76d-4ec7-b919-828df6f31125,Namespace:calico-system,Attempt:1,} returns sandbox id \"775d8d469d2ab031a2239f0071ad078481fb2c0911d17de24c2d4477a636c3c0\"" Jan 17 12:21:04.069501 containerd[1544]: time="2025-01-17T12:21:04.069433641Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.1\"" Jan 17 12:21:04.585366 containerd[1544]: time="2025-01-17T12:21:04.584658308Z" level=info msg="StopPodSandbox for \"977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598\"" Jan 17 12:21:04.585366 containerd[1544]: time="2025-01-17T12:21:04.584782009Z" level=info msg="StopPodSandbox for \"94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe\"" Jan 17 12:21:04.654768 containerd[1544]: 2025-01-17 12:21:04.621 [INFO][4273] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" Jan 17 12:21:04.654768 containerd[1544]: 2025-01-17 12:21:04.622 [INFO][4273] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" iface="eth0" netns="/var/run/netns/cni-6556479f-407d-068c-bacc-283e0b123fdf" Jan 17 12:21:04.654768 containerd[1544]: 2025-01-17 12:21:04.622 [INFO][4273] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" iface="eth0" netns="/var/run/netns/cni-6556479f-407d-068c-bacc-283e0b123fdf" Jan 17 12:21:04.654768 containerd[1544]: 2025-01-17 12:21:04.622 [INFO][4273] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" iface="eth0" netns="/var/run/netns/cni-6556479f-407d-068c-bacc-283e0b123fdf" Jan 17 12:21:04.654768 containerd[1544]: 2025-01-17 12:21:04.622 [INFO][4273] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" Jan 17 12:21:04.654768 containerd[1544]: 2025-01-17 12:21:04.622 [INFO][4273] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" Jan 17 12:21:04.654768 containerd[1544]: 2025-01-17 12:21:04.645 [INFO][4284] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" HandleID="k8s-pod-network.977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" Workload="localhost-k8s-calico--apiserver--6b697d6776--wzkmk-eth0" Jan 17 12:21:04.654768 containerd[1544]: 2025-01-17 12:21:04.646 [INFO][4284] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jan 17 12:21:04.654768 containerd[1544]: 2025-01-17 12:21:04.646 [INFO][4284] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jan 17 12:21:04.654768 containerd[1544]: 2025-01-17 12:21:04.650 [WARNING][4284] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" HandleID="k8s-pod-network.977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" Workload="localhost-k8s-calico--apiserver--6b697d6776--wzkmk-eth0" Jan 17 12:21:04.654768 containerd[1544]: 2025-01-17 12:21:04.650 [INFO][4284] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" HandleID="k8s-pod-network.977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" Workload="localhost-k8s-calico--apiserver--6b697d6776--wzkmk-eth0" Jan 17 12:21:04.654768 containerd[1544]: 2025-01-17 12:21:04.651 [INFO][4284] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jan 17 12:21:04.654768 containerd[1544]: 2025-01-17 12:21:04.653 [INFO][4273] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" Jan 17 12:21:04.655916 containerd[1544]: time="2025-01-17T12:21:04.655836476Z" level=info msg="TearDown network for sandbox \"977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598\" successfully" Jan 17 12:21:04.655916 containerd[1544]: time="2025-01-17T12:21:04.655857125Z" level=info msg="StopPodSandbox for \"977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598\" returns successfully" Jan 17 12:21:04.656353 containerd[1544]: time="2025-01-17T12:21:04.656335030Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6b697d6776-wzkmk,Uid:cb76a1e6-a555-4ec5-bb3d-c19efe7e0c24,Namespace:calico-apiserver,Attempt:1,}" Jan 17 12:21:04.658691 containerd[1544]: 2025-01-17 12:21:04.628 [INFO][4269] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" Jan 17 12:21:04.658691 containerd[1544]: 2025-01-17 12:21:04.629 [INFO][4269] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" iface="eth0" netns="/var/run/netns/cni-09475e86-9098-13f4-c2c0-7d868c96848c" Jan 17 12:21:04.658691 containerd[1544]: 2025-01-17 12:21:04.629 [INFO][4269] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" iface="eth0" netns="/var/run/netns/cni-09475e86-9098-13f4-c2c0-7d868c96848c" Jan 17 12:21:04.658691 containerd[1544]: 2025-01-17 12:21:04.629 [INFO][4269] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" iface="eth0" netns="/var/run/netns/cni-09475e86-9098-13f4-c2c0-7d868c96848c" Jan 17 12:21:04.658691 containerd[1544]: 2025-01-17 12:21:04.629 [INFO][4269] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" Jan 17 12:21:04.658691 containerd[1544]: 2025-01-17 12:21:04.630 [INFO][4269] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" Jan 17 12:21:04.658691 containerd[1544]: 2025-01-17 12:21:04.647 [INFO][4288] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" HandleID="k8s-pod-network.94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" Workload="localhost-k8s-calico--kube--controllers--784b5c8b57--wvh2g-eth0" Jan 17 12:21:04.658691 containerd[1544]: 2025-01-17 12:21:04.647 [INFO][4288] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jan 17 12:21:04.658691 containerd[1544]: 2025-01-17 12:21:04.651 [INFO][4288] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jan 17 12:21:04.658691 containerd[1544]: 2025-01-17 12:21:04.654 [WARNING][4288] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" HandleID="k8s-pod-network.94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" Workload="localhost-k8s-calico--kube--controllers--784b5c8b57--wvh2g-eth0" Jan 17 12:21:04.658691 containerd[1544]: 2025-01-17 12:21:04.654 [INFO][4288] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" HandleID="k8s-pod-network.94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" Workload="localhost-k8s-calico--kube--controllers--784b5c8b57--wvh2g-eth0" Jan 17 12:21:04.658691 containerd[1544]: 2025-01-17 12:21:04.655 [INFO][4288] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jan 17 12:21:04.658691 containerd[1544]: 2025-01-17 12:21:04.657 [INFO][4269] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" Jan 17 12:21:04.658691 containerd[1544]: time="2025-01-17T12:21:04.658629177Z" level=info msg="TearDown network for sandbox \"94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe\" successfully" Jan 17 12:21:04.658691 containerd[1544]: time="2025-01-17T12:21:04.658643960Z" level=info msg="StopPodSandbox for \"94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe\" returns successfully" Jan 17 12:21:04.659307 containerd[1544]: time="2025-01-17T12:21:04.659163516Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-784b5c8b57-wvh2g,Uid:0c7fe481-2df4-4da8-b215-b904f488836f,Namespace:calico-system,Attempt:1,}" Jan 17 12:21:04.743082 systemd-networkd[1453]: cali4a2f335bdbd: Link UP Jan 17 12:21:04.743561 systemd-networkd[1453]: cali4a2f335bdbd: Gained carrier Jan 17 12:21:04.753667 containerd[1544]: 2025-01-17 12:21:04.686 [INFO][4297] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Jan 17 12:21:04.753667 containerd[1544]: 2025-01-17 12:21:04.694 [INFO][4297] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--6b697d6776--wzkmk-eth0 calico-apiserver-6b697d6776- calico-apiserver cb76a1e6-a555-4ec5-bb3d-c19efe7e0c24 764 0 2025-01-17 12:20:41 +0000 UTC <nil> <nil> map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:6b697d6776 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-6b697d6776-wzkmk eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali4a2f335bdbd [] []}} ContainerID="c42d613ff2110abb6ca0e2360ab25abb8b1b10e7bcc7fdb0277c7f14f9b056bd" Namespace="calico-apiserver" Pod="calico-apiserver-6b697d6776-wzkmk" WorkloadEndpoint="localhost-k8s-calico--apiserver--6b697d6776--wzkmk-" Jan 17 12:21:04.753667 containerd[1544]: 2025-01-17 12:21:04.694 [INFO][4297] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="c42d613ff2110abb6ca0e2360ab25abb8b1b10e7bcc7fdb0277c7f14f9b056bd" Namespace="calico-apiserver" Pod="calico-apiserver-6b697d6776-wzkmk" WorkloadEndpoint="localhost-k8s-calico--apiserver--6b697d6776--wzkmk-eth0" Jan 17 12:21:04.753667 containerd[1544]: 2025-01-17 12:21:04.717 [INFO][4321] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="c42d613ff2110abb6ca0e2360ab25abb8b1b10e7bcc7fdb0277c7f14f9b056bd" HandleID="k8s-pod-network.c42d613ff2110abb6ca0e2360ab25abb8b1b10e7bcc7fdb0277c7f14f9b056bd" Workload="localhost-k8s-calico--apiserver--6b697d6776--wzkmk-eth0" Jan 17 12:21:04.753667 containerd[1544]: 2025-01-17 12:21:04.722 [INFO][4321] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="c42d613ff2110abb6ca0e2360ab25abb8b1b10e7bcc7fdb0277c7f14f9b056bd" HandleID="k8s-pod-network.c42d613ff2110abb6ca0e2360ab25abb8b1b10e7bcc7fdb0277c7f14f9b056bd" Workload="localhost-k8s-calico--apiserver--6b697d6776--wzkmk-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00011dc30), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-6b697d6776-wzkmk", "timestamp":"2025-01-17 12:21:04.716989136 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 17 12:21:04.753667 containerd[1544]: 2025-01-17 12:21:04.722 [INFO][4321] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jan 17 12:21:04.753667 containerd[1544]: 2025-01-17 12:21:04.722 [INFO][4321] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jan 17 12:21:04.753667 containerd[1544]: 2025-01-17 12:21:04.723 [INFO][4321] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 17 12:21:04.753667 containerd[1544]: 2025-01-17 12:21:04.724 [INFO][4321] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.c42d613ff2110abb6ca0e2360ab25abb8b1b10e7bcc7fdb0277c7f14f9b056bd" host="localhost" Jan 17 12:21:04.753667 containerd[1544]: 2025-01-17 12:21:04.727 [INFO][4321] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" Jan 17 12:21:04.753667 containerd[1544]: 2025-01-17 12:21:04.730 [INFO][4321] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Jan 17 12:21:04.753667 containerd[1544]: 2025-01-17 12:21:04.731 [INFO][4321] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 17 12:21:04.753667 containerd[1544]: 2025-01-17 12:21:04.732 [INFO][4321] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 17 12:21:04.753667 containerd[1544]: 2025-01-17 12:21:04.732 [INFO][4321] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.c42d613ff2110abb6ca0e2360ab25abb8b1b10e7bcc7fdb0277c7f14f9b056bd" host="localhost" Jan 17 12:21:04.753667 containerd[1544]: 2025-01-17 12:21:04.733 [INFO][4321] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.c42d613ff2110abb6ca0e2360ab25abb8b1b10e7bcc7fdb0277c7f14f9b056bd Jan 17 12:21:04.753667 containerd[1544]: 2025-01-17 12:21:04.735 [INFO][4321] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.c42d613ff2110abb6ca0e2360ab25abb8b1b10e7bcc7fdb0277c7f14f9b056bd" host="localhost" Jan 17 12:21:04.753667 containerd[1544]: 2025-01-17 12:21:04.738 [INFO][4321] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.130/26] block=192.168.88.128/26 handle="k8s-pod-network.c42d613ff2110abb6ca0e2360ab25abb8b1b10e7bcc7fdb0277c7f14f9b056bd" host="localhost" Jan 17 12:21:04.753667 containerd[1544]: 2025-01-17 12:21:04.738 [INFO][4321] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.130/26] handle="k8s-pod-network.c42d613ff2110abb6ca0e2360ab25abb8b1b10e7bcc7fdb0277c7f14f9b056bd" host="localhost" Jan 17 12:21:04.753667 containerd[1544]: 2025-01-17 12:21:04.738 [INFO][4321] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jan 17 12:21:04.753667 containerd[1544]: 2025-01-17 12:21:04.738 [INFO][4321] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.130/26] IPv6=[] ContainerID="c42d613ff2110abb6ca0e2360ab25abb8b1b10e7bcc7fdb0277c7f14f9b056bd" HandleID="k8s-pod-network.c42d613ff2110abb6ca0e2360ab25abb8b1b10e7bcc7fdb0277c7f14f9b056bd" Workload="localhost-k8s-calico--apiserver--6b697d6776--wzkmk-eth0" Jan 17 12:21:04.754251 containerd[1544]: 2025-01-17 12:21:04.740 [INFO][4297] cni-plugin/k8s.go 386: Populated endpoint ContainerID="c42d613ff2110abb6ca0e2360ab25abb8b1b10e7bcc7fdb0277c7f14f9b056bd" Namespace="calico-apiserver" Pod="calico-apiserver-6b697d6776-wzkmk" WorkloadEndpoint="localhost-k8s-calico--apiserver--6b697d6776--wzkmk-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--6b697d6776--wzkmk-eth0", GenerateName:"calico-apiserver-6b697d6776-", Namespace:"calico-apiserver", SelfLink:"", UID:"cb76a1e6-a555-4ec5-bb3d-c19efe7e0c24", ResourceVersion:"764", Generation:0, CreationTimestamp:time.Date(2025, time.January, 17, 12, 20, 41, 0, time.Local), DeletionTimestamp:<nil>, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6b697d6776", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-6b697d6776-wzkmk", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali4a2f335bdbd", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 17 12:21:04.754251 containerd[1544]: 2025-01-17 12:21:04.740 [INFO][4297] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.130/32] ContainerID="c42d613ff2110abb6ca0e2360ab25abb8b1b10e7bcc7fdb0277c7f14f9b056bd" Namespace="calico-apiserver" Pod="calico-apiserver-6b697d6776-wzkmk" WorkloadEndpoint="localhost-k8s-calico--apiserver--6b697d6776--wzkmk-eth0" Jan 17 12:21:04.754251 containerd[1544]: 2025-01-17 12:21:04.740 [INFO][4297] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali4a2f335bdbd ContainerID="c42d613ff2110abb6ca0e2360ab25abb8b1b10e7bcc7fdb0277c7f14f9b056bd" Namespace="calico-apiserver" Pod="calico-apiserver-6b697d6776-wzkmk" WorkloadEndpoint="localhost-k8s-calico--apiserver--6b697d6776--wzkmk-eth0" Jan 17 12:21:04.754251 containerd[1544]: 2025-01-17 12:21:04.742 [INFO][4297] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="c42d613ff2110abb6ca0e2360ab25abb8b1b10e7bcc7fdb0277c7f14f9b056bd" Namespace="calico-apiserver" Pod="calico-apiserver-6b697d6776-wzkmk" WorkloadEndpoint="localhost-k8s-calico--apiserver--6b697d6776--wzkmk-eth0" Jan 17 12:21:04.754251 containerd[1544]: 2025-01-17 12:21:04.743 [INFO][4297] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="c42d613ff2110abb6ca0e2360ab25abb8b1b10e7bcc7fdb0277c7f14f9b056bd" Namespace="calico-apiserver" Pod="calico-apiserver-6b697d6776-wzkmk" WorkloadEndpoint="localhost-k8s-calico--apiserver--6b697d6776--wzkmk-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--6b697d6776--wzkmk-eth0", GenerateName:"calico-apiserver-6b697d6776-", Namespace:"calico-apiserver", SelfLink:"", UID:"cb76a1e6-a555-4ec5-bb3d-c19efe7e0c24", ResourceVersion:"764", Generation:0, CreationTimestamp:time.Date(2025, time.January, 17, 12, 20, 41, 0, time.Local), DeletionTimestamp:<nil>, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6b697d6776", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"c42d613ff2110abb6ca0e2360ab25abb8b1b10e7bcc7fdb0277c7f14f9b056bd", Pod:"calico-apiserver-6b697d6776-wzkmk", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali4a2f335bdbd", MAC:"8a:92:70:fa:41:17", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 17 12:21:04.754251 containerd[1544]: 2025-01-17 12:21:04.752 [INFO][4297] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="c42d613ff2110abb6ca0e2360ab25abb8b1b10e7bcc7fdb0277c7f14f9b056bd" Namespace="calico-apiserver" Pod="calico-apiserver-6b697d6776-wzkmk" WorkloadEndpoint="localhost-k8s-calico--apiserver--6b697d6776--wzkmk-eth0" Jan 17 12:21:04.765304 containerd[1544]: time="2025-01-17T12:21:04.765086245Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 17 12:21:04.765304 containerd[1544]: time="2025-01-17T12:21:04.765124230Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 17 12:21:04.765304 containerd[1544]: time="2025-01-17T12:21:04.765142426Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 12:21:04.765304 containerd[1544]: time="2025-01-17T12:21:04.765225531Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 12:21:04.778928 systemd[1]: Started cri-containerd-c42d613ff2110abb6ca0e2360ab25abb8b1b10e7bcc7fdb0277c7f14f9b056bd.scope - libcontainer container c42d613ff2110abb6ca0e2360ab25abb8b1b10e7bcc7fdb0277c7f14f9b056bd. Jan 17 12:21:04.787784 systemd-resolved[1454]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 17 12:21:04.808390 containerd[1544]: time="2025-01-17T12:21:04.808361999Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6b697d6776-wzkmk,Uid:cb76a1e6-a555-4ec5-bb3d-c19efe7e0c24,Namespace:calico-apiserver,Attempt:1,} returns sandbox id \"c42d613ff2110abb6ca0e2360ab25abb8b1b10e7bcc7fdb0277c7f14f9b056bd\"" Jan 17 12:21:04.840769 systemd-networkd[1453]: cali1dda6d62550: Link UP Jan 17 12:21:04.841281 systemd-networkd[1453]: cali1dda6d62550: Gained carrier Jan 17 12:21:04.851418 containerd[1544]: 2025-01-17 12:21:04.689 [INFO][4308] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Jan 17 12:21:04.851418 containerd[1544]: 2025-01-17 12:21:04.696 [INFO][4308] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--kube--controllers--784b5c8b57--wvh2g-eth0 calico-kube-controllers-784b5c8b57- calico-system 0c7fe481-2df4-4da8-b215-b904f488836f 765 0 2025-01-17 12:20:41 +0000 UTC <nil> <nil> map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:784b5c8b57 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s localhost calico-kube-controllers-784b5c8b57-wvh2g eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali1dda6d62550 [] []}} ContainerID="99dd9e4ee20f35388085cfc493ed39315999eee19d59ba8391cf9393e4a2f823" Namespace="calico-system" Pod="calico-kube-controllers-784b5c8b57-wvh2g" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--784b5c8b57--wvh2g-" Jan 17 12:21:04.851418 containerd[1544]: 2025-01-17 12:21:04.697 [INFO][4308] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="99dd9e4ee20f35388085cfc493ed39315999eee19d59ba8391cf9393e4a2f823" Namespace="calico-system" Pod="calico-kube-controllers-784b5c8b57-wvh2g" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--784b5c8b57--wvh2g-eth0" Jan 17 12:21:04.851418 containerd[1544]: 2025-01-17 12:21:04.725 [INFO][4325] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="99dd9e4ee20f35388085cfc493ed39315999eee19d59ba8391cf9393e4a2f823" HandleID="k8s-pod-network.99dd9e4ee20f35388085cfc493ed39315999eee19d59ba8391cf9393e4a2f823" Workload="localhost-k8s-calico--kube--controllers--784b5c8b57--wvh2g-eth0" Jan 17 12:21:04.851418 containerd[1544]: 2025-01-17 12:21:04.731 [INFO][4325] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="99dd9e4ee20f35388085cfc493ed39315999eee19d59ba8391cf9393e4a2f823" HandleID="k8s-pod-network.99dd9e4ee20f35388085cfc493ed39315999eee19d59ba8391cf9393e4a2f823" Workload="localhost-k8s-calico--kube--controllers--784b5c8b57--wvh2g-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002ed6e0), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"calico-kube-controllers-784b5c8b57-wvh2g", "timestamp":"2025-01-17 12:21:04.725896327 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 17 12:21:04.851418 containerd[1544]: 2025-01-17 12:21:04.731 [INFO][4325] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jan 17 12:21:04.851418 containerd[1544]: 2025-01-17 12:21:04.739 [INFO][4325] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jan 17 12:21:04.851418 containerd[1544]: 2025-01-17 12:21:04.739 [INFO][4325] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 17 12:21:04.851418 containerd[1544]: 2025-01-17 12:21:04.824 [INFO][4325] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.99dd9e4ee20f35388085cfc493ed39315999eee19d59ba8391cf9393e4a2f823" host="localhost" Jan 17 12:21:04.851418 containerd[1544]: 2025-01-17 12:21:04.826 [INFO][4325] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" Jan 17 12:21:04.851418 containerd[1544]: 2025-01-17 12:21:04.830 [INFO][4325] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Jan 17 12:21:04.851418 containerd[1544]: 2025-01-17 12:21:04.830 [INFO][4325] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 17 12:21:04.851418 containerd[1544]: 2025-01-17 12:21:04.832 [INFO][4325] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 17 12:21:04.851418 containerd[1544]: 2025-01-17 12:21:04.832 [INFO][4325] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.99dd9e4ee20f35388085cfc493ed39315999eee19d59ba8391cf9393e4a2f823" host="localhost" Jan 17 12:21:04.851418 containerd[1544]: 2025-01-17 12:21:04.832 [INFO][4325] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.99dd9e4ee20f35388085cfc493ed39315999eee19d59ba8391cf9393e4a2f823 Jan 17 12:21:04.851418 containerd[1544]: 2025-01-17 12:21:04.834 [INFO][4325] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.99dd9e4ee20f35388085cfc493ed39315999eee19d59ba8391cf9393e4a2f823" host="localhost" Jan 17 12:21:04.851418 containerd[1544]: 2025-01-17 12:21:04.837 [INFO][4325] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.131/26] block=192.168.88.128/26 handle="k8s-pod-network.99dd9e4ee20f35388085cfc493ed39315999eee19d59ba8391cf9393e4a2f823" host="localhost" Jan 17 12:21:04.851418 containerd[1544]: 2025-01-17 12:21:04.837 [INFO][4325] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.131/26] handle="k8s-pod-network.99dd9e4ee20f35388085cfc493ed39315999eee19d59ba8391cf9393e4a2f823" host="localhost" Jan 17 12:21:04.851418 containerd[1544]: 2025-01-17 12:21:04.837 [INFO][4325] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jan 17 12:21:04.851418 containerd[1544]: 2025-01-17 12:21:04.837 [INFO][4325] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.131/26] IPv6=[] ContainerID="99dd9e4ee20f35388085cfc493ed39315999eee19d59ba8391cf9393e4a2f823" HandleID="k8s-pod-network.99dd9e4ee20f35388085cfc493ed39315999eee19d59ba8391cf9393e4a2f823" Workload="localhost-k8s-calico--kube--controllers--784b5c8b57--wvh2g-eth0" Jan 17 12:21:04.851972 containerd[1544]: 2025-01-17 12:21:04.839 [INFO][4308] cni-plugin/k8s.go 386: Populated endpoint ContainerID="99dd9e4ee20f35388085cfc493ed39315999eee19d59ba8391cf9393e4a2f823" Namespace="calico-system" Pod="calico-kube-controllers-784b5c8b57-wvh2g" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--784b5c8b57--wvh2g-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--784b5c8b57--wvh2g-eth0", GenerateName:"calico-kube-controllers-784b5c8b57-", Namespace:"calico-system", SelfLink:"", UID:"0c7fe481-2df4-4da8-b215-b904f488836f", ResourceVersion:"765", Generation:0, CreationTimestamp:time.Date(2025, time.January, 17, 12, 20, 41, 0, time.Local), DeletionTimestamp:<nil>, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"784b5c8b57", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-kube-controllers-784b5c8b57-wvh2g", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali1dda6d62550", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 17 12:21:04.851972 containerd[1544]: 2025-01-17 12:21:04.839 [INFO][4308] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.131/32] ContainerID="99dd9e4ee20f35388085cfc493ed39315999eee19d59ba8391cf9393e4a2f823" Namespace="calico-system" Pod="calico-kube-controllers-784b5c8b57-wvh2g" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--784b5c8b57--wvh2g-eth0" Jan 17 12:21:04.851972 containerd[1544]: 2025-01-17 12:21:04.839 [INFO][4308] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali1dda6d62550 ContainerID="99dd9e4ee20f35388085cfc493ed39315999eee19d59ba8391cf9393e4a2f823" Namespace="calico-system" Pod="calico-kube-controllers-784b5c8b57-wvh2g" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--784b5c8b57--wvh2g-eth0" Jan 17 12:21:04.851972 containerd[1544]: 2025-01-17 12:21:04.842 [INFO][4308] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="99dd9e4ee20f35388085cfc493ed39315999eee19d59ba8391cf9393e4a2f823" Namespace="calico-system" Pod="calico-kube-controllers-784b5c8b57-wvh2g" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--784b5c8b57--wvh2g-eth0" Jan 17 12:21:04.851972 containerd[1544]: 2025-01-17 12:21:04.843 [INFO][4308] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="99dd9e4ee20f35388085cfc493ed39315999eee19d59ba8391cf9393e4a2f823" Namespace="calico-system" Pod="calico-kube-controllers-784b5c8b57-wvh2g" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--784b5c8b57--wvh2g-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--784b5c8b57--wvh2g-eth0", GenerateName:"calico-kube-controllers-784b5c8b57-", Namespace:"calico-system", SelfLink:"", UID:"0c7fe481-2df4-4da8-b215-b904f488836f", ResourceVersion:"765", Generation:0, CreationTimestamp:time.Date(2025, time.January, 17, 12, 20, 41, 0, time.Local), DeletionTimestamp:<nil>, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"784b5c8b57", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"99dd9e4ee20f35388085cfc493ed39315999eee19d59ba8391cf9393e4a2f823", Pod:"calico-kube-controllers-784b5c8b57-wvh2g", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali1dda6d62550", MAC:"16:49:b5:d3:fb:f4", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 17 12:21:04.851972 containerd[1544]: 2025-01-17 12:21:04.850 [INFO][4308] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="99dd9e4ee20f35388085cfc493ed39315999eee19d59ba8391cf9393e4a2f823" Namespace="calico-system" Pod="calico-kube-controllers-784b5c8b57-wvh2g" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--784b5c8b57--wvh2g-eth0" Jan 17 12:21:04.864613 containerd[1544]: time="2025-01-17T12:21:04.864456129Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 17 12:21:04.864613 containerd[1544]: time="2025-01-17T12:21:04.864539326Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 17 12:21:04.864613 containerd[1544]: time="2025-01-17T12:21:04.864554484Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 12:21:04.866812 containerd[1544]: time="2025-01-17T12:21:04.864944619Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 12:21:04.877781 systemd[1]: Started cri-containerd-99dd9e4ee20f35388085cfc493ed39315999eee19d59ba8391cf9393e4a2f823.scope - libcontainer container 99dd9e4ee20f35388085cfc493ed39315999eee19d59ba8391cf9393e4a2f823. Jan 17 12:21:04.886067 systemd-resolved[1454]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 17 12:21:04.889201 systemd[1]: run-containerd-runc-k8s.io-775d8d469d2ab031a2239f0071ad078481fb2c0911d17de24c2d4477a636c3c0-runc.GsCiHT.mount: Deactivated successfully. Jan 17 12:21:04.889405 systemd[1]: run-netns-cni\x2d6556479f\x2d407d\x2d068c\x2dbacc\x2d283e0b123fdf.mount: Deactivated successfully. Jan 17 12:21:04.889493 systemd[1]: run-netns-cni\x2d09475e86\x2d9098\x2d13f4\x2dc2c0\x2d7d868c96848c.mount: Deactivated successfully. Jan 17 12:21:04.911930 containerd[1544]: time="2025-01-17T12:21:04.911903440Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-784b5c8b57-wvh2g,Uid:0c7fe481-2df4-4da8-b215-b904f488836f,Namespace:calico-system,Attempt:1,} returns sandbox id \"99dd9e4ee20f35388085cfc493ed39315999eee19d59ba8391cf9393e4a2f823\"" Jan 17 12:21:05.586852 containerd[1544]: time="2025-01-17T12:21:05.585461731Z" level=info msg="StopPodSandbox for \"fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4\"" Jan 17 12:21:05.588459 containerd[1544]: time="2025-01-17T12:21:05.588424650Z" level=info msg="StopPodSandbox for \"81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef\"" Jan 17 12:21:05.647267 containerd[1544]: 2025-01-17 12:21:05.624 [INFO][4487] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" Jan 17 12:21:05.647267 containerd[1544]: 2025-01-17 12:21:05.625 [INFO][4487] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" iface="eth0" netns="/var/run/netns/cni-15e0fb59-d9ce-af77-9e9f-e6e4f9922308" Jan 17 12:21:05.647267 containerd[1544]: 2025-01-17 12:21:05.625 [INFO][4487] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" iface="eth0" netns="/var/run/netns/cni-15e0fb59-d9ce-af77-9e9f-e6e4f9922308" Jan 17 12:21:05.647267 containerd[1544]: 2025-01-17 12:21:05.625 [INFO][4487] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" iface="eth0" netns="/var/run/netns/cni-15e0fb59-d9ce-af77-9e9f-e6e4f9922308" Jan 17 12:21:05.647267 containerd[1544]: 2025-01-17 12:21:05.625 [INFO][4487] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" Jan 17 12:21:05.647267 containerd[1544]: 2025-01-17 12:21:05.625 [INFO][4487] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" Jan 17 12:21:05.647267 containerd[1544]: 2025-01-17 12:21:05.640 [INFO][4500] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" HandleID="k8s-pod-network.fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" Workload="localhost-k8s-calico--apiserver--6b697d6776--n97bk-eth0" Jan 17 12:21:05.647267 containerd[1544]: 2025-01-17 12:21:05.640 [INFO][4500] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jan 17 12:21:05.647267 containerd[1544]: 2025-01-17 12:21:05.640 [INFO][4500] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jan 17 12:21:05.647267 containerd[1544]: 2025-01-17 12:21:05.644 [WARNING][4500] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" HandleID="k8s-pod-network.fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" Workload="localhost-k8s-calico--apiserver--6b697d6776--n97bk-eth0" Jan 17 12:21:05.647267 containerd[1544]: 2025-01-17 12:21:05.644 [INFO][4500] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" HandleID="k8s-pod-network.fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" Workload="localhost-k8s-calico--apiserver--6b697d6776--n97bk-eth0" Jan 17 12:21:05.647267 containerd[1544]: 2025-01-17 12:21:05.645 [INFO][4500] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jan 17 12:21:05.647267 containerd[1544]: 2025-01-17 12:21:05.646 [INFO][4487] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" Jan 17 12:21:05.648029 containerd[1544]: time="2025-01-17T12:21:05.647756502Z" level=info msg="TearDown network for sandbox \"fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4\" successfully" Jan 17 12:21:05.648029 containerd[1544]: time="2025-01-17T12:21:05.647782995Z" level=info msg="StopPodSandbox for \"fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4\" returns successfully" Jan 17 12:21:05.649103 systemd[1]: run-netns-cni\x2d15e0fb59\x2dd9ce\x2daf77\x2d9e9f\x2de6e4f9922308.mount: Deactivated successfully. Jan 17 12:21:05.649311 containerd[1544]: time="2025-01-17T12:21:05.649297348Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6b697d6776-n97bk,Uid:9c19e775-a72c-40f9-9f63-2947d850da10,Namespace:calico-apiserver,Attempt:1,}" Jan 17 12:21:05.660958 containerd[1544]: 2025-01-17 12:21:05.621 [INFO][4488] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" Jan 17 12:21:05.660958 containerd[1544]: 2025-01-17 12:21:05.622 [INFO][4488] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" iface="eth0" netns="/var/run/netns/cni-3c9ca590-a110-ec11-796d-b04a6e16fac3" Jan 17 12:21:05.660958 containerd[1544]: 2025-01-17 12:21:05.622 [INFO][4488] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" iface="eth0" netns="/var/run/netns/cni-3c9ca590-a110-ec11-796d-b04a6e16fac3" Jan 17 12:21:05.660958 containerd[1544]: 2025-01-17 12:21:05.623 [INFO][4488] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" iface="eth0" netns="/var/run/netns/cni-3c9ca590-a110-ec11-796d-b04a6e16fac3" Jan 17 12:21:05.660958 containerd[1544]: 2025-01-17 12:21:05.623 [INFO][4488] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" Jan 17 12:21:05.660958 containerd[1544]: 2025-01-17 12:21:05.623 [INFO][4488] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" Jan 17 12:21:05.660958 containerd[1544]: 2025-01-17 12:21:05.653 [INFO][4499] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" HandleID="k8s-pod-network.81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" Workload="localhost-k8s-coredns--6f6b679f8f--bq2mf-eth0" Jan 17 12:21:05.660958 containerd[1544]: 2025-01-17 12:21:05.653 [INFO][4499] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jan 17 12:21:05.660958 containerd[1544]: 2025-01-17 12:21:05.653 [INFO][4499] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jan 17 12:21:05.660958 containerd[1544]: 2025-01-17 12:21:05.657 [WARNING][4499] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" HandleID="k8s-pod-network.81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" Workload="localhost-k8s-coredns--6f6b679f8f--bq2mf-eth0" Jan 17 12:21:05.660958 containerd[1544]: 2025-01-17 12:21:05.657 [INFO][4499] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" HandleID="k8s-pod-network.81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" Workload="localhost-k8s-coredns--6f6b679f8f--bq2mf-eth0" Jan 17 12:21:05.660958 containerd[1544]: 2025-01-17 12:21:05.658 [INFO][4499] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jan 17 12:21:05.660958 containerd[1544]: 2025-01-17 12:21:05.659 [INFO][4488] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" Jan 17 12:21:05.662578 containerd[1544]: time="2025-01-17T12:21:05.661364258Z" level=info msg="TearDown network for sandbox \"81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef\" successfully" Jan 17 12:21:05.662578 containerd[1544]: time="2025-01-17T12:21:05.661380794Z" level=info msg="StopPodSandbox for \"81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef\" returns successfully" Jan 17 12:21:05.663589 systemd[1]: run-netns-cni\x2d3c9ca590\x2da110\x2dec11\x2d796d\x2db04a6e16fac3.mount: Deactivated successfully. Jan 17 12:21:05.663800 containerd[1544]: time="2025-01-17T12:21:05.663659494Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-bq2mf,Uid:95fe1c4f-c274-4a0d-aa79-7b5cb4353bdf,Namespace:kube-system,Attempt:1,}" Jan 17 12:21:05.766827 systemd-networkd[1453]: calic5bb7ed3b64: Gained IPv6LL Jan 17 12:21:05.785723 systemd-networkd[1453]: cali1b1536c4fdb: Link UP Jan 17 12:21:05.785845 systemd-networkd[1453]: cali1b1536c4fdb: Gained carrier Jan 17 12:21:05.796046 containerd[1544]: 2025-01-17 12:21:05.705 [INFO][4512] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Jan 17 12:21:05.796046 containerd[1544]: 2025-01-17 12:21:05.715 [INFO][4512] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--6b697d6776--n97bk-eth0 calico-apiserver-6b697d6776- calico-apiserver 9c19e775-a72c-40f9-9f63-2947d850da10 778 0 2025-01-17 12:20:41 +0000 UTC <nil> <nil> map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:6b697d6776 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-6b697d6776-n97bk eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali1b1536c4fdb [] []}} ContainerID="899c759b7380f5fb3825df86319a1ff1551827e14d94f8418cf3528733e21bf4" Namespace="calico-apiserver" Pod="calico-apiserver-6b697d6776-n97bk" WorkloadEndpoint="localhost-k8s-calico--apiserver--6b697d6776--n97bk-" Jan 17 12:21:05.796046 containerd[1544]: 2025-01-17 12:21:05.715 [INFO][4512] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="899c759b7380f5fb3825df86319a1ff1551827e14d94f8418cf3528733e21bf4" Namespace="calico-apiserver" Pod="calico-apiserver-6b697d6776-n97bk" WorkloadEndpoint="localhost-k8s-calico--apiserver--6b697d6776--n97bk-eth0" Jan 17 12:21:05.796046 containerd[1544]: 2025-01-17 12:21:05.748 [INFO][4535] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="899c759b7380f5fb3825df86319a1ff1551827e14d94f8418cf3528733e21bf4" HandleID="k8s-pod-network.899c759b7380f5fb3825df86319a1ff1551827e14d94f8418cf3528733e21bf4" Workload="localhost-k8s-calico--apiserver--6b697d6776--n97bk-eth0" Jan 17 12:21:05.796046 containerd[1544]: 2025-01-17 12:21:05.760 [INFO][4535] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="899c759b7380f5fb3825df86319a1ff1551827e14d94f8418cf3528733e21bf4" HandleID="k8s-pod-network.899c759b7380f5fb3825df86319a1ff1551827e14d94f8418cf3528733e21bf4" Workload="localhost-k8s-calico--apiserver--6b697d6776--n97bk-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00042d820), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-6b697d6776-n97bk", "timestamp":"2025-01-17 12:21:05.748632805 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 17 12:21:05.796046 containerd[1544]: 2025-01-17 12:21:05.760 [INFO][4535] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jan 17 12:21:05.796046 containerd[1544]: 2025-01-17 12:21:05.760 [INFO][4535] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jan 17 12:21:05.796046 containerd[1544]: 2025-01-17 12:21:05.760 [INFO][4535] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 17 12:21:05.796046 containerd[1544]: 2025-01-17 12:21:05.761 [INFO][4535] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.899c759b7380f5fb3825df86319a1ff1551827e14d94f8418cf3528733e21bf4" host="localhost" Jan 17 12:21:05.796046 containerd[1544]: 2025-01-17 12:21:05.763 [INFO][4535] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" Jan 17 12:21:05.796046 containerd[1544]: 2025-01-17 12:21:05.768 [INFO][4535] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Jan 17 12:21:05.796046 containerd[1544]: 2025-01-17 12:21:05.769 [INFO][4535] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 17 12:21:05.796046 containerd[1544]: 2025-01-17 12:21:05.771 [INFO][4535] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 17 12:21:05.796046 containerd[1544]: 2025-01-17 12:21:05.771 [INFO][4535] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.899c759b7380f5fb3825df86319a1ff1551827e14d94f8418cf3528733e21bf4" host="localhost" Jan 17 12:21:05.796046 containerd[1544]: 2025-01-17 12:21:05.772 [INFO][4535] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.899c759b7380f5fb3825df86319a1ff1551827e14d94f8418cf3528733e21bf4 Jan 17 12:21:05.796046 containerd[1544]: 2025-01-17 12:21:05.776 [INFO][4535] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.899c759b7380f5fb3825df86319a1ff1551827e14d94f8418cf3528733e21bf4" host="localhost" Jan 17 12:21:05.796046 containerd[1544]: 2025-01-17 12:21:05.781 [INFO][4535] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.132/26] block=192.168.88.128/26 handle="k8s-pod-network.899c759b7380f5fb3825df86319a1ff1551827e14d94f8418cf3528733e21bf4" host="localhost" Jan 17 12:21:05.796046 containerd[1544]: 2025-01-17 12:21:05.781 [INFO][4535] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.132/26] handle="k8s-pod-network.899c759b7380f5fb3825df86319a1ff1551827e14d94f8418cf3528733e21bf4" host="localhost" Jan 17 12:21:05.796046 containerd[1544]: 2025-01-17 12:21:05.781 [INFO][4535] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jan 17 12:21:05.796046 containerd[1544]: 2025-01-17 12:21:05.781 [INFO][4535] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.132/26] IPv6=[] ContainerID="899c759b7380f5fb3825df86319a1ff1551827e14d94f8418cf3528733e21bf4" HandleID="k8s-pod-network.899c759b7380f5fb3825df86319a1ff1551827e14d94f8418cf3528733e21bf4" Workload="localhost-k8s-calico--apiserver--6b697d6776--n97bk-eth0" Jan 17 12:21:05.800557 containerd[1544]: 2025-01-17 12:21:05.782 [INFO][4512] cni-plugin/k8s.go 386: Populated endpoint ContainerID="899c759b7380f5fb3825df86319a1ff1551827e14d94f8418cf3528733e21bf4" Namespace="calico-apiserver" Pod="calico-apiserver-6b697d6776-n97bk" WorkloadEndpoint="localhost-k8s-calico--apiserver--6b697d6776--n97bk-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--6b697d6776--n97bk-eth0", GenerateName:"calico-apiserver-6b697d6776-", Namespace:"calico-apiserver", SelfLink:"", UID:"9c19e775-a72c-40f9-9f63-2947d850da10", ResourceVersion:"778", Generation:0, CreationTimestamp:time.Date(2025, time.January, 17, 12, 20, 41, 0, time.Local), DeletionTimestamp:<nil>, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6b697d6776", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-6b697d6776-n97bk", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali1b1536c4fdb", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 17 12:21:05.800557 containerd[1544]: 2025-01-17 12:21:05.783 [INFO][4512] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.132/32] ContainerID="899c759b7380f5fb3825df86319a1ff1551827e14d94f8418cf3528733e21bf4" Namespace="calico-apiserver" Pod="calico-apiserver-6b697d6776-n97bk" WorkloadEndpoint="localhost-k8s-calico--apiserver--6b697d6776--n97bk-eth0" Jan 17 12:21:05.800557 containerd[1544]: 2025-01-17 12:21:05.783 [INFO][4512] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali1b1536c4fdb ContainerID="899c759b7380f5fb3825df86319a1ff1551827e14d94f8418cf3528733e21bf4" Namespace="calico-apiserver" Pod="calico-apiserver-6b697d6776-n97bk" WorkloadEndpoint="localhost-k8s-calico--apiserver--6b697d6776--n97bk-eth0" Jan 17 12:21:05.800557 containerd[1544]: 2025-01-17 12:21:05.786 [INFO][4512] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="899c759b7380f5fb3825df86319a1ff1551827e14d94f8418cf3528733e21bf4" Namespace="calico-apiserver" Pod="calico-apiserver-6b697d6776-n97bk" WorkloadEndpoint="localhost-k8s-calico--apiserver--6b697d6776--n97bk-eth0" Jan 17 12:21:05.800557 containerd[1544]: 2025-01-17 12:21:05.786 [INFO][4512] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="899c759b7380f5fb3825df86319a1ff1551827e14d94f8418cf3528733e21bf4" Namespace="calico-apiserver" Pod="calico-apiserver-6b697d6776-n97bk" WorkloadEndpoint="localhost-k8s-calico--apiserver--6b697d6776--n97bk-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--6b697d6776--n97bk-eth0", GenerateName:"calico-apiserver-6b697d6776-", Namespace:"calico-apiserver", SelfLink:"", UID:"9c19e775-a72c-40f9-9f63-2947d850da10", ResourceVersion:"778", Generation:0, CreationTimestamp:time.Date(2025, time.January, 17, 12, 20, 41, 0, time.Local), DeletionTimestamp:<nil>, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6b697d6776", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"899c759b7380f5fb3825df86319a1ff1551827e14d94f8418cf3528733e21bf4", Pod:"calico-apiserver-6b697d6776-n97bk", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali1b1536c4fdb", MAC:"ca:6f:47:88:29:8a", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 17 12:21:05.800557 containerd[1544]: 2025-01-17 12:21:05.793 [INFO][4512] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="899c759b7380f5fb3825df86319a1ff1551827e14d94f8418cf3528733e21bf4" Namespace="calico-apiserver" Pod="calico-apiserver-6b697d6776-n97bk" WorkloadEndpoint="localhost-k8s-calico--apiserver--6b697d6776--n97bk-eth0" Jan 17 12:21:05.818042 containerd[1544]: time="2025-01-17T12:21:05.817730645Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 17 12:21:05.818042 containerd[1544]: time="2025-01-17T12:21:05.817777183Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 17 12:21:05.818042 containerd[1544]: time="2025-01-17T12:21:05.817792950Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 12:21:05.818042 containerd[1544]: time="2025-01-17T12:21:05.817850012Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 12:21:05.836835 systemd[1]: Started cri-containerd-899c759b7380f5fb3825df86319a1ff1551827e14d94f8418cf3528733e21bf4.scope - libcontainer container 899c759b7380f5fb3825df86319a1ff1551827e14d94f8418cf3528733e21bf4. Jan 17 12:21:05.849544 systemd-resolved[1454]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 17 12:21:05.894389 systemd-networkd[1453]: cali2818078aafc: Link UP Jan 17 12:21:05.897778 systemd-networkd[1453]: cali2818078aafc: Gained carrier Jan 17 12:21:05.900705 containerd[1544]: time="2025-01-17T12:21:05.899131336Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-6b697d6776-n97bk,Uid:9c19e775-a72c-40f9-9f63-2947d850da10,Namespace:calico-apiserver,Attempt:1,} returns sandbox id \"899c759b7380f5fb3825df86319a1ff1551827e14d94f8418cf3528733e21bf4\"" Jan 17 12:21:05.910262 containerd[1544]: 2025-01-17 12:21:05.719 [INFO][4522] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Jan 17 12:21:05.910262 containerd[1544]: 2025-01-17 12:21:05.734 [INFO][4522] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--6f6b679f8f--bq2mf-eth0 coredns-6f6b679f8f- kube-system 95fe1c4f-c274-4a0d-aa79-7b5cb4353bdf 777 0 2025-01-17 12:20:33 +0000 UTC <nil> <nil> map[k8s-app:kube-dns pod-template-hash:6f6b679f8f projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-6f6b679f8f-bq2mf eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali2818078aafc [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="2979d448e752ef269fd63a5774b653a4d50161125cfc8a6dbc8abb731174203a" Namespace="kube-system" Pod="coredns-6f6b679f8f-bq2mf" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--bq2mf-" Jan 17 12:21:05.910262 containerd[1544]: 2025-01-17 12:21:05.734 [INFO][4522] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="2979d448e752ef269fd63a5774b653a4d50161125cfc8a6dbc8abb731174203a" Namespace="kube-system" Pod="coredns-6f6b679f8f-bq2mf" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--bq2mf-eth0" Jan 17 12:21:05.910262 containerd[1544]: 2025-01-17 12:21:05.773 [INFO][4539] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="2979d448e752ef269fd63a5774b653a4d50161125cfc8a6dbc8abb731174203a" HandleID="k8s-pod-network.2979d448e752ef269fd63a5774b653a4d50161125cfc8a6dbc8abb731174203a" Workload="localhost-k8s-coredns--6f6b679f8f--bq2mf-eth0" Jan 17 12:21:05.910262 containerd[1544]: 2025-01-17 12:21:05.858 [INFO][4539] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="2979d448e752ef269fd63a5774b653a4d50161125cfc8a6dbc8abb731174203a" HandleID="k8s-pod-network.2979d448e752ef269fd63a5774b653a4d50161125cfc8a6dbc8abb731174203a" Workload="localhost-k8s-coredns--6f6b679f8f--bq2mf-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0002ed6d0), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-6f6b679f8f-bq2mf", "timestamp":"2025-01-17 12:21:05.773603466 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 17 12:21:05.910262 containerd[1544]: 2025-01-17 12:21:05.858 [INFO][4539] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jan 17 12:21:05.910262 containerd[1544]: 2025-01-17 12:21:05.858 [INFO][4539] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jan 17 12:21:05.910262 containerd[1544]: 2025-01-17 12:21:05.858 [INFO][4539] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 17 12:21:05.910262 containerd[1544]: 2025-01-17 12:21:05.862 [INFO][4539] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.2979d448e752ef269fd63a5774b653a4d50161125cfc8a6dbc8abb731174203a" host="localhost" Jan 17 12:21:05.910262 containerd[1544]: 2025-01-17 12:21:05.865 [INFO][4539] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" Jan 17 12:21:05.910262 containerd[1544]: 2025-01-17 12:21:05.869 [INFO][4539] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Jan 17 12:21:05.910262 containerd[1544]: 2025-01-17 12:21:05.870 [INFO][4539] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 17 12:21:05.910262 containerd[1544]: 2025-01-17 12:21:05.873 [INFO][4539] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 17 12:21:05.910262 containerd[1544]: 2025-01-17 12:21:05.873 [INFO][4539] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.2979d448e752ef269fd63a5774b653a4d50161125cfc8a6dbc8abb731174203a" host="localhost" Jan 17 12:21:05.910262 containerd[1544]: 2025-01-17 12:21:05.875 [INFO][4539] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.2979d448e752ef269fd63a5774b653a4d50161125cfc8a6dbc8abb731174203a Jan 17 12:21:05.910262 containerd[1544]: 2025-01-17 12:21:05.880 [INFO][4539] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.2979d448e752ef269fd63a5774b653a4d50161125cfc8a6dbc8abb731174203a" host="localhost" Jan 17 12:21:05.910262 containerd[1544]: 2025-01-17 12:21:05.885 [INFO][4539] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.133/26] block=192.168.88.128/26 handle="k8s-pod-network.2979d448e752ef269fd63a5774b653a4d50161125cfc8a6dbc8abb731174203a" host="localhost" Jan 17 12:21:05.910262 containerd[1544]: 2025-01-17 12:21:05.885 [INFO][4539] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.133/26] handle="k8s-pod-network.2979d448e752ef269fd63a5774b653a4d50161125cfc8a6dbc8abb731174203a" host="localhost" Jan 17 12:21:05.910262 containerd[1544]: 2025-01-17 12:21:05.885 [INFO][4539] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jan 17 12:21:05.910262 containerd[1544]: 2025-01-17 12:21:05.885 [INFO][4539] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.133/26] IPv6=[] ContainerID="2979d448e752ef269fd63a5774b653a4d50161125cfc8a6dbc8abb731174203a" HandleID="k8s-pod-network.2979d448e752ef269fd63a5774b653a4d50161125cfc8a6dbc8abb731174203a" Workload="localhost-k8s-coredns--6f6b679f8f--bq2mf-eth0" Jan 17 12:21:05.913627 containerd[1544]: 2025-01-17 12:21:05.890 [INFO][4522] cni-plugin/k8s.go 386: Populated endpoint ContainerID="2979d448e752ef269fd63a5774b653a4d50161125cfc8a6dbc8abb731174203a" Namespace="kube-system" Pod="coredns-6f6b679f8f-bq2mf" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--bq2mf-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--6f6b679f8f--bq2mf-eth0", GenerateName:"coredns-6f6b679f8f-", Namespace:"kube-system", SelfLink:"", UID:"95fe1c4f-c274-4a0d-aa79-7b5cb4353bdf", ResourceVersion:"777", Generation:0, CreationTimestamp:time.Date(2025, time.January, 17, 12, 20, 33, 0, time.Local), DeletionTimestamp:<nil>, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"6f6b679f8f", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-6f6b679f8f-bq2mf", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali2818078aafc", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 17 12:21:05.913627 containerd[1544]: 2025-01-17 12:21:05.890 [INFO][4522] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.133/32] ContainerID="2979d448e752ef269fd63a5774b653a4d50161125cfc8a6dbc8abb731174203a" Namespace="kube-system" Pod="coredns-6f6b679f8f-bq2mf" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--bq2mf-eth0" Jan 17 12:21:05.913627 containerd[1544]: 2025-01-17 12:21:05.890 [INFO][4522] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali2818078aafc ContainerID="2979d448e752ef269fd63a5774b653a4d50161125cfc8a6dbc8abb731174203a" Namespace="kube-system" Pod="coredns-6f6b679f8f-bq2mf" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--bq2mf-eth0" Jan 17 12:21:05.913627 containerd[1544]: 2025-01-17 12:21:05.893 [INFO][4522] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="2979d448e752ef269fd63a5774b653a4d50161125cfc8a6dbc8abb731174203a" Namespace="kube-system" Pod="coredns-6f6b679f8f-bq2mf" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--bq2mf-eth0" Jan 17 12:21:05.913627 containerd[1544]: 2025-01-17 12:21:05.893 [INFO][4522] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="2979d448e752ef269fd63a5774b653a4d50161125cfc8a6dbc8abb731174203a" Namespace="kube-system" Pod="coredns-6f6b679f8f-bq2mf" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--bq2mf-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--6f6b679f8f--bq2mf-eth0", GenerateName:"coredns-6f6b679f8f-", Namespace:"kube-system", SelfLink:"", UID:"95fe1c4f-c274-4a0d-aa79-7b5cb4353bdf", ResourceVersion:"777", Generation:0, CreationTimestamp:time.Date(2025, time.January, 17, 12, 20, 33, 0, time.Local), DeletionTimestamp:<nil>, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"6f6b679f8f", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"2979d448e752ef269fd63a5774b653a4d50161125cfc8a6dbc8abb731174203a", Pod:"coredns-6f6b679f8f-bq2mf", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali2818078aafc", MAC:"2a:b9:d0:8e:b2:e9", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 17 12:21:05.913627 containerd[1544]: 2025-01-17 12:21:05.906 [INFO][4522] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="2979d448e752ef269fd63a5774b653a4d50161125cfc8a6dbc8abb731174203a" Namespace="kube-system" Pod="coredns-6f6b679f8f-bq2mf" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--bq2mf-eth0" Jan 17 12:21:05.928273 containerd[1544]: time="2025-01-17T12:21:05.928132490Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 17 12:21:05.928273 containerd[1544]: time="2025-01-17T12:21:05.928173946Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 17 12:21:05.928273 containerd[1544]: time="2025-01-17T12:21:05.928191329Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 12:21:05.928273 containerd[1544]: time="2025-01-17T12:21:05.928244368Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 12:21:05.945479 containerd[1544]: time="2025-01-17T12:21:05.945035591Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:21:05.945479 containerd[1544]: time="2025-01-17T12:21:05.945302697Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.29.1: active requests=0, bytes read=7902632" Jan 17 12:21:05.945479 containerd[1544]: time="2025-01-17T12:21:05.945457407Z" level=info msg="ImageCreate event name:\"sha256:bda8c42e04758c4f061339e213f50ccdc7502c4176fbf631aa12357e62b63540\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:21:05.946658 containerd[1544]: time="2025-01-17T12:21:05.946637336Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:eaa7e01fb16b603c155a67b81f16992281db7f831684c7b2081d3434587a7ff3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:21:05.946899 systemd[1]: Started cri-containerd-2979d448e752ef269fd63a5774b653a4d50161125cfc8a6dbc8abb731174203a.scope - libcontainer container 2979d448e752ef269fd63a5774b653a4d50161125cfc8a6dbc8abb731174203a. Jan 17 12:21:05.949181 containerd[1544]: time="2025-01-17T12:21:05.948823802Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.29.1\" with image id \"sha256:bda8c42e04758c4f061339e213f50ccdc7502c4176fbf631aa12357e62b63540\", repo tag \"ghcr.io/flatcar/calico/csi:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:eaa7e01fb16b603c155a67b81f16992281db7f831684c7b2081d3434587a7ff3\", size \"9395716\" in 1.87937061s" Jan 17 12:21:05.949181 containerd[1544]: time="2025-01-17T12:21:05.948842777Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.1\" returns image reference \"sha256:bda8c42e04758c4f061339e213f50ccdc7502c4176fbf631aa12357e62b63540\"" Jan 17 12:21:05.951588 containerd[1544]: time="2025-01-17T12:21:05.951298667Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.1\"" Jan 17 12:21:05.951990 containerd[1544]: time="2025-01-17T12:21:05.951973402Z" level=info msg="CreateContainer within sandbox \"775d8d469d2ab031a2239f0071ad078481fb2c0911d17de24c2d4477a636c3c0\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Jan 17 12:21:05.960775 systemd-resolved[1454]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 17 12:21:05.962379 containerd[1544]: time="2025-01-17T12:21:05.962290214Z" level=info msg="CreateContainer within sandbox \"775d8d469d2ab031a2239f0071ad078481fb2c0911d17de24c2d4477a636c3c0\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"84582fa60e720abac57db795d8c76d23502f485ec24f2870df2c25aeb3d0ee38\"" Jan 17 12:21:05.964256 containerd[1544]: time="2025-01-17T12:21:05.962924027Z" level=info msg="StartContainer for \"84582fa60e720abac57db795d8c76d23502f485ec24f2870df2c25aeb3d0ee38\"" Jan 17 12:21:05.986775 systemd[1]: Started cri-containerd-84582fa60e720abac57db795d8c76d23502f485ec24f2870df2c25aeb3d0ee38.scope - libcontainer container 84582fa60e720abac57db795d8c76d23502f485ec24f2870df2c25aeb3d0ee38. Jan 17 12:21:05.996598 containerd[1544]: time="2025-01-17T12:21:05.996572480Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-bq2mf,Uid:95fe1c4f-c274-4a0d-aa79-7b5cb4353bdf,Namespace:kube-system,Attempt:1,} returns sandbox id \"2979d448e752ef269fd63a5774b653a4d50161125cfc8a6dbc8abb731174203a\"" Jan 17 12:21:06.000205 containerd[1544]: time="2025-01-17T12:21:06.000187073Z" level=info msg="CreateContainer within sandbox \"2979d448e752ef269fd63a5774b653a4d50161125cfc8a6dbc8abb731174203a\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jan 17 12:21:06.014009 containerd[1544]: time="2025-01-17T12:21:06.013984457Z" level=info msg="StartContainer for \"84582fa60e720abac57db795d8c76d23502f485ec24f2870df2c25aeb3d0ee38\" returns successfully" Jan 17 12:21:06.015749 containerd[1544]: time="2025-01-17T12:21:06.015729089Z" level=info msg="CreateContainer within sandbox \"2979d448e752ef269fd63a5774b653a4d50161125cfc8a6dbc8abb731174203a\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"d0cf094d8ea31d1b0e405b47397399f2376a75f8ee580953019677bc46ed7a5b\"" Jan 17 12:21:06.016412 containerd[1544]: time="2025-01-17T12:21:06.016044316Z" level=info msg="StartContainer for \"d0cf094d8ea31d1b0e405b47397399f2376a75f8ee580953019677bc46ed7a5b\"" Jan 17 12:21:06.034909 systemd[1]: Started cri-containerd-d0cf094d8ea31d1b0e405b47397399f2376a75f8ee580953019677bc46ed7a5b.scope - libcontainer container d0cf094d8ea31d1b0e405b47397399f2376a75f8ee580953019677bc46ed7a5b. Jan 17 12:21:06.067043 containerd[1544]: time="2025-01-17T12:21:06.066992481Z" level=info msg="StartContainer for \"d0cf094d8ea31d1b0e405b47397399f2376a75f8ee580953019677bc46ed7a5b\" returns successfully" Jan 17 12:21:06.342806 systemd-networkd[1453]: cali4a2f335bdbd: Gained IPv6LL Jan 17 12:21:06.343075 systemd-networkd[1453]: cali1dda6d62550: Gained IPv6LL Jan 17 12:21:06.586639 containerd[1544]: time="2025-01-17T12:21:06.586462968Z" level=info msg="StopPodSandbox for \"d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1\"" Jan 17 12:21:06.705141 containerd[1544]: 2025-01-17 12:21:06.653 [INFO][4754] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" Jan 17 12:21:06.705141 containerd[1544]: 2025-01-17 12:21:06.653 [INFO][4754] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" iface="eth0" netns="/var/run/netns/cni-e8f0bf85-80f8-5296-3358-2c45e669e9f6" Jan 17 12:21:06.705141 containerd[1544]: 2025-01-17 12:21:06.654 [INFO][4754] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" iface="eth0" netns="/var/run/netns/cni-e8f0bf85-80f8-5296-3358-2c45e669e9f6" Jan 17 12:21:06.705141 containerd[1544]: 2025-01-17 12:21:06.654 [INFO][4754] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" iface="eth0" netns="/var/run/netns/cni-e8f0bf85-80f8-5296-3358-2c45e669e9f6" Jan 17 12:21:06.705141 containerd[1544]: 2025-01-17 12:21:06.654 [INFO][4754] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" Jan 17 12:21:06.705141 containerd[1544]: 2025-01-17 12:21:06.654 [INFO][4754] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" Jan 17 12:21:06.705141 containerd[1544]: 2025-01-17 12:21:06.668 [INFO][4770] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" HandleID="k8s-pod-network.d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" Workload="localhost-k8s-coredns--6f6b679f8f--6f26n-eth0" Jan 17 12:21:06.705141 containerd[1544]: 2025-01-17 12:21:06.668 [INFO][4770] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jan 17 12:21:06.705141 containerd[1544]: 2025-01-17 12:21:06.668 [INFO][4770] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jan 17 12:21:06.705141 containerd[1544]: 2025-01-17 12:21:06.702 [WARNING][4770] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" HandleID="k8s-pod-network.d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" Workload="localhost-k8s-coredns--6f6b679f8f--6f26n-eth0" Jan 17 12:21:06.705141 containerd[1544]: 2025-01-17 12:21:06.702 [INFO][4770] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" HandleID="k8s-pod-network.d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" Workload="localhost-k8s-coredns--6f6b679f8f--6f26n-eth0" Jan 17 12:21:06.705141 containerd[1544]: 2025-01-17 12:21:06.703 [INFO][4770] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jan 17 12:21:06.705141 containerd[1544]: 2025-01-17 12:21:06.704 [INFO][4754] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" Jan 17 12:21:06.706075 containerd[1544]: time="2025-01-17T12:21:06.705213138Z" level=info msg="TearDown network for sandbox \"d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1\" successfully" Jan 17 12:21:06.706075 containerd[1544]: time="2025-01-17T12:21:06.705229181Z" level=info msg="StopPodSandbox for \"d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1\" returns successfully" Jan 17 12:21:06.706075 containerd[1544]: time="2025-01-17T12:21:06.705716480Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-6f26n,Uid:7756b0ac-5d50-42a5-ba21-9f884a8f9486,Namespace:kube-system,Attempt:1,}" Jan 17 12:21:06.854975 systemd-networkd[1453]: cali1b1536c4fdb: Gained IPv6LL Jan 17 12:21:06.888413 systemd[1]: run-netns-cni\x2de8f0bf85\x2d80f8\x2d5296\x2d3358\x2d2c45e669e9f6.mount: Deactivated successfully. Jan 17 12:21:07.127690 systemd-networkd[1453]: cali5874a32a051: Link UP Jan 17 12:21:07.127805 systemd-networkd[1453]: cali5874a32a051: Gained carrier Jan 17 12:21:07.136372 kubelet[2723]: I0117 12:21:07.136194 2723 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-6f6b679f8f-bq2mf" podStartSLOduration=34.136178211 podStartE2EDuration="34.136178211s" podCreationTimestamp="2025-01-17 12:20:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-17 12:21:06.781907618 +0000 UTC m=+39.265380209" watchObservedRunningTime="2025-01-17 12:21:07.136178211 +0000 UTC m=+39.619650803" Jan 17 12:21:07.140287 containerd[1544]: 2025-01-17 12:21:07.075 [INFO][4777] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Jan 17 12:21:07.140287 containerd[1544]: 2025-01-17 12:21:07.081 [INFO][4777] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--6f6b679f8f--6f26n-eth0 coredns-6f6b679f8f- kube-system 7756b0ac-5d50-42a5-ba21-9f884a8f9486 798 0 2025-01-17 12:20:33 +0000 UTC <nil> <nil> map[k8s-app:kube-dns pod-template-hash:6f6b679f8f projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-6f6b679f8f-6f26n eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali5874a32a051 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="48c13b9221db34d5ac7a06035688d231254dc30dbe87dd08127ed8b235b5a728" Namespace="kube-system" Pod="coredns-6f6b679f8f-6f26n" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--6f26n-" Jan 17 12:21:07.140287 containerd[1544]: 2025-01-17 12:21:07.081 [INFO][4777] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="48c13b9221db34d5ac7a06035688d231254dc30dbe87dd08127ed8b235b5a728" Namespace="kube-system" Pod="coredns-6f6b679f8f-6f26n" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--6f26n-eth0" Jan 17 12:21:07.140287 containerd[1544]: 2025-01-17 12:21:07.101 [INFO][4788] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="48c13b9221db34d5ac7a06035688d231254dc30dbe87dd08127ed8b235b5a728" HandleID="k8s-pod-network.48c13b9221db34d5ac7a06035688d231254dc30dbe87dd08127ed8b235b5a728" Workload="localhost-k8s-coredns--6f6b679f8f--6f26n-eth0" Jan 17 12:21:07.140287 containerd[1544]: 2025-01-17 12:21:07.106 [INFO][4788] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="48c13b9221db34d5ac7a06035688d231254dc30dbe87dd08127ed8b235b5a728" HandleID="k8s-pod-network.48c13b9221db34d5ac7a06035688d231254dc30dbe87dd08127ed8b235b5a728" Workload="localhost-k8s-coredns--6f6b679f8f--6f26n-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000290b70), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-6f6b679f8f-6f26n", "timestamp":"2025-01-17 12:21:07.101001313 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 17 12:21:07.140287 containerd[1544]: 2025-01-17 12:21:07.106 [INFO][4788] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jan 17 12:21:07.140287 containerd[1544]: 2025-01-17 12:21:07.106 [INFO][4788] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jan 17 12:21:07.140287 containerd[1544]: 2025-01-17 12:21:07.106 [INFO][4788] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jan 17 12:21:07.140287 containerd[1544]: 2025-01-17 12:21:07.107 [INFO][4788] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.48c13b9221db34d5ac7a06035688d231254dc30dbe87dd08127ed8b235b5a728" host="localhost" Jan 17 12:21:07.140287 containerd[1544]: 2025-01-17 12:21:07.109 [INFO][4788] ipam/ipam.go 372: Looking up existing affinities for host host="localhost" Jan 17 12:21:07.140287 containerd[1544]: 2025-01-17 12:21:07.111 [INFO][4788] ipam/ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Jan 17 12:21:07.140287 containerd[1544]: 2025-01-17 12:21:07.112 [INFO][4788] ipam/ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jan 17 12:21:07.140287 containerd[1544]: 2025-01-17 12:21:07.113 [INFO][4788] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jan 17 12:21:07.140287 containerd[1544]: 2025-01-17 12:21:07.113 [INFO][4788] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.48c13b9221db34d5ac7a06035688d231254dc30dbe87dd08127ed8b235b5a728" host="localhost" Jan 17 12:21:07.140287 containerd[1544]: 2025-01-17 12:21:07.114 [INFO][4788] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.48c13b9221db34d5ac7a06035688d231254dc30dbe87dd08127ed8b235b5a728 Jan 17 12:21:07.140287 containerd[1544]: 2025-01-17 12:21:07.117 [INFO][4788] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.48c13b9221db34d5ac7a06035688d231254dc30dbe87dd08127ed8b235b5a728" host="localhost" Jan 17 12:21:07.140287 containerd[1544]: 2025-01-17 12:21:07.123 [INFO][4788] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.88.134/26] block=192.168.88.128/26 handle="k8s-pod-network.48c13b9221db34d5ac7a06035688d231254dc30dbe87dd08127ed8b235b5a728" host="localhost" Jan 17 12:21:07.140287 containerd[1544]: 2025-01-17 12:21:07.123 [INFO][4788] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.134/26] handle="k8s-pod-network.48c13b9221db34d5ac7a06035688d231254dc30dbe87dd08127ed8b235b5a728" host="localhost" Jan 17 12:21:07.140287 containerd[1544]: 2025-01-17 12:21:07.123 [INFO][4788] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jan 17 12:21:07.140287 containerd[1544]: 2025-01-17 12:21:07.123 [INFO][4788] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.88.134/26] IPv6=[] ContainerID="48c13b9221db34d5ac7a06035688d231254dc30dbe87dd08127ed8b235b5a728" HandleID="k8s-pod-network.48c13b9221db34d5ac7a06035688d231254dc30dbe87dd08127ed8b235b5a728" Workload="localhost-k8s-coredns--6f6b679f8f--6f26n-eth0" Jan 17 12:21:07.142186 containerd[1544]: 2025-01-17 12:21:07.124 [INFO][4777] cni-plugin/k8s.go 386: Populated endpoint ContainerID="48c13b9221db34d5ac7a06035688d231254dc30dbe87dd08127ed8b235b5a728" Namespace="kube-system" Pod="coredns-6f6b679f8f-6f26n" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--6f26n-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--6f6b679f8f--6f26n-eth0", GenerateName:"coredns-6f6b679f8f-", Namespace:"kube-system", SelfLink:"", UID:"7756b0ac-5d50-42a5-ba21-9f884a8f9486", ResourceVersion:"798", Generation:0, CreationTimestamp:time.Date(2025, time.January, 17, 12, 20, 33, 0, time.Local), DeletionTimestamp:<nil>, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"6f6b679f8f", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-6f6b679f8f-6f26n", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali5874a32a051", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 17 12:21:07.142186 containerd[1544]: 2025-01-17 12:21:07.124 [INFO][4777] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.88.134/32] ContainerID="48c13b9221db34d5ac7a06035688d231254dc30dbe87dd08127ed8b235b5a728" Namespace="kube-system" Pod="coredns-6f6b679f8f-6f26n" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--6f26n-eth0" Jan 17 12:21:07.142186 containerd[1544]: 2025-01-17 12:21:07.124 [INFO][4777] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali5874a32a051 ContainerID="48c13b9221db34d5ac7a06035688d231254dc30dbe87dd08127ed8b235b5a728" Namespace="kube-system" Pod="coredns-6f6b679f8f-6f26n" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--6f26n-eth0" Jan 17 12:21:07.142186 containerd[1544]: 2025-01-17 12:21:07.127 [INFO][4777] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="48c13b9221db34d5ac7a06035688d231254dc30dbe87dd08127ed8b235b5a728" Namespace="kube-system" Pod="coredns-6f6b679f8f-6f26n" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--6f26n-eth0" Jan 17 12:21:07.142186 containerd[1544]: 2025-01-17 12:21:07.127 [INFO][4777] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="48c13b9221db34d5ac7a06035688d231254dc30dbe87dd08127ed8b235b5a728" Namespace="kube-system" Pod="coredns-6f6b679f8f-6f26n" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--6f26n-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--6f6b679f8f--6f26n-eth0", GenerateName:"coredns-6f6b679f8f-", Namespace:"kube-system", SelfLink:"", UID:"7756b0ac-5d50-42a5-ba21-9f884a8f9486", ResourceVersion:"798", Generation:0, CreationTimestamp:time.Date(2025, time.January, 17, 12, 20, 33, 0, time.Local), DeletionTimestamp:<nil>, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"6f6b679f8f", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"48c13b9221db34d5ac7a06035688d231254dc30dbe87dd08127ed8b235b5a728", Pod:"coredns-6f6b679f8f-6f26n", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali5874a32a051", MAC:"a6:b1:36:93:54:ae", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 17 12:21:07.142186 containerd[1544]: 2025-01-17 12:21:07.136 [INFO][4777] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="48c13b9221db34d5ac7a06035688d231254dc30dbe87dd08127ed8b235b5a728" Namespace="kube-system" Pod="coredns-6f6b679f8f-6f26n" WorkloadEndpoint="localhost-k8s-coredns--6f6b679f8f--6f26n-eth0" Jan 17 12:21:07.152256 containerd[1544]: time="2025-01-17T12:21:07.152023797Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 17 12:21:07.152256 containerd[1544]: time="2025-01-17T12:21:07.152064748Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 17 12:21:07.152256 containerd[1544]: time="2025-01-17T12:21:07.152083165Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 12:21:07.152256 containerd[1544]: time="2025-01-17T12:21:07.152153217Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 17 12:21:07.170810 systemd[1]: Started cri-containerd-48c13b9221db34d5ac7a06035688d231254dc30dbe87dd08127ed8b235b5a728.scope - libcontainer container 48c13b9221db34d5ac7a06035688d231254dc30dbe87dd08127ed8b235b5a728. Jan 17 12:21:07.179509 systemd-resolved[1454]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jan 17 12:21:07.198608 containerd[1544]: time="2025-01-17T12:21:07.198572815Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-6f26n,Uid:7756b0ac-5d50-42a5-ba21-9f884a8f9486,Namespace:kube-system,Attempt:1,} returns sandbox id \"48c13b9221db34d5ac7a06035688d231254dc30dbe87dd08127ed8b235b5a728\"" Jan 17 12:21:07.203380 containerd[1544]: time="2025-01-17T12:21:07.203316001Z" level=info msg="CreateContainer within sandbox \"48c13b9221db34d5ac7a06035688d231254dc30dbe87dd08127ed8b235b5a728\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jan 17 12:21:07.224463 containerd[1544]: time="2025-01-17T12:21:07.224237121Z" level=info msg="CreateContainer within sandbox \"48c13b9221db34d5ac7a06035688d231254dc30dbe87dd08127ed8b235b5a728\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"6f175b1175846681dc9748ae4546d2be37707213e4b96500f9138359afd2d4b4\"" Jan 17 12:21:07.225783 containerd[1544]: time="2025-01-17T12:21:07.225763501Z" level=info msg="StartContainer for \"6f175b1175846681dc9748ae4546d2be37707213e4b96500f9138359afd2d4b4\"" Jan 17 12:21:07.247768 systemd[1]: Started cri-containerd-6f175b1175846681dc9748ae4546d2be37707213e4b96500f9138359afd2d4b4.scope - libcontainer container 6f175b1175846681dc9748ae4546d2be37707213e4b96500f9138359afd2d4b4. Jan 17 12:21:07.265761 containerd[1544]: time="2025-01-17T12:21:07.265740165Z" level=info msg="StartContainer for \"6f175b1175846681dc9748ae4546d2be37707213e4b96500f9138359afd2d4b4\" returns successfully" Jan 17 12:21:07.559167 systemd-networkd[1453]: cali2818078aafc: Gained IPv6LL Jan 17 12:21:07.823857 kubelet[2723]: I0117 12:21:07.822945 2723 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-6f6b679f8f-6f26n" podStartSLOduration=34.822932076 podStartE2EDuration="34.822932076s" podCreationTimestamp="2025-01-17 12:20:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-17 12:21:07.79508472 +0000 UTC m=+40.278557313" watchObservedRunningTime="2025-01-17 12:21:07.822932076 +0000 UTC m=+40.306404663" Jan 17 12:21:08.307504 containerd[1544]: time="2025-01-17T12:21:08.307003004Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:21:08.307504 containerd[1544]: time="2025-01-17T12:21:08.307417513Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.29.1: active requests=0, bytes read=42001404" Jan 17 12:21:08.307504 containerd[1544]: time="2025-01-17T12:21:08.307479290Z" level=info msg="ImageCreate event name:\"sha256:421726ace5ed13894f7edf594dd3a462947aedc13d0f69d08525d7369477fb70\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:21:08.308649 containerd[1544]: time="2025-01-17T12:21:08.308632778Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:b8c43e264fe52e0c327b0bf3ac882a0224b33bdd7f4ff58a74242da7d9b00486\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:21:08.309785 containerd[1544]: time="2025-01-17T12:21:08.309210047Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.29.1\" with image id \"sha256:421726ace5ed13894f7edf594dd3a462947aedc13d0f69d08525d7369477fb70\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:b8c43e264fe52e0c327b0bf3ac882a0224b33bdd7f4ff58a74242da7d9b00486\", size \"43494504\" in 2.357892155s" Jan 17 12:21:08.309785 containerd[1544]: time="2025-01-17T12:21:08.309226234Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.1\" returns image reference \"sha256:421726ace5ed13894f7edf594dd3a462947aedc13d0f69d08525d7369477fb70\"" Jan 17 12:21:08.311298 containerd[1544]: time="2025-01-17T12:21:08.310761142Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.29.1\"" Jan 17 12:21:08.312115 containerd[1544]: time="2025-01-17T12:21:08.312101258Z" level=info msg="CreateContainer within sandbox \"c42d613ff2110abb6ca0e2360ab25abb8b1b10e7bcc7fdb0277c7f14f9b056bd\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Jan 17 12:21:08.327753 containerd[1544]: time="2025-01-17T12:21:08.327730118Z" level=info msg="CreateContainer within sandbox \"c42d613ff2110abb6ca0e2360ab25abb8b1b10e7bcc7fdb0277c7f14f9b056bd\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"46dfcd179e749f4d09f1e7243a9f2ac3144f05bb399587029c2f55ca80f877fa\"" Jan 17 12:21:08.328356 containerd[1544]: time="2025-01-17T12:21:08.328341879Z" level=info msg="StartContainer for \"46dfcd179e749f4d09f1e7243a9f2ac3144f05bb399587029c2f55ca80f877fa\"" Jan 17 12:21:08.354797 systemd[1]: Started cri-containerd-46dfcd179e749f4d09f1e7243a9f2ac3144f05bb399587029c2f55ca80f877fa.scope - libcontainer container 46dfcd179e749f4d09f1e7243a9f2ac3144f05bb399587029c2f55ca80f877fa. Jan 17 12:21:08.385073 containerd[1544]: time="2025-01-17T12:21:08.385042997Z" level=info msg="StartContainer for \"46dfcd179e749f4d09f1e7243a9f2ac3144f05bb399587029c2f55ca80f877fa\" returns successfully" Jan 17 12:21:08.390790 systemd-networkd[1453]: cali5874a32a051: Gained IPv6LL Jan 17 12:21:08.692324 kubelet[2723]: I0117 12:21:08.692093 2723 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 17 12:21:08.796165 kubelet[2723]: I0117 12:21:08.795509 2723 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-6b697d6776-wzkmk" podStartSLOduration=24.298618451 podStartE2EDuration="27.795494873s" podCreationTimestamp="2025-01-17 12:20:41 +0000 UTC" firstStartedPulling="2025-01-17 12:21:04.813785807 +0000 UTC m=+37.297258390" lastFinishedPulling="2025-01-17 12:21:08.310662229 +0000 UTC m=+40.794134812" observedRunningTime="2025-01-17 12:21:08.795108653 +0000 UTC m=+41.278581240" watchObservedRunningTime="2025-01-17 12:21:08.795494873 +0000 UTC m=+41.278967457" Jan 17 12:21:09.501735 kernel: bpftool[5002]: memfd_create() called without MFD_EXEC or MFD_NOEXEC_SEAL set Jan 17 12:21:09.781766 kubelet[2723]: I0117 12:21:09.781174 2723 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 17 12:21:09.902336 systemd-networkd[1453]: vxlan.calico: Link UP Jan 17 12:21:09.902340 systemd-networkd[1453]: vxlan.calico: Gained carrier Jan 17 12:21:10.431055 containerd[1544]: time="2025-01-17T12:21:10.431028273Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:21:10.431829 containerd[1544]: time="2025-01-17T12:21:10.431661607Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.29.1: active requests=0, bytes read=34141192" Jan 17 12:21:10.431829 containerd[1544]: time="2025-01-17T12:21:10.431806834Z" level=info msg="ImageCreate event name:\"sha256:6331715a2ae96b18a770a395cac108321d108e445e08b616e5bc9fbd1f9c21da\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:21:10.432942 containerd[1544]: time="2025-01-17T12:21:10.432918236Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:1072d6a98167a14ca361e9ce757733f9bae36d1f1c6a9621ea10934b6b1e10d9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:21:10.433558 containerd[1544]: time="2025-01-17T12:21:10.433307854Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.29.1\" with image id \"sha256:6331715a2ae96b18a770a395cac108321d108e445e08b616e5bc9fbd1f9c21da\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:1072d6a98167a14ca361e9ce757733f9bae36d1f1c6a9621ea10934b6b1e10d9\", size \"35634244\" in 2.122532597s" Jan 17 12:21:10.433558 containerd[1544]: time="2025-01-17T12:21:10.433328902Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.29.1\" returns image reference \"sha256:6331715a2ae96b18a770a395cac108321d108e445e08b616e5bc9fbd1f9c21da\"" Jan 17 12:21:10.434861 containerd[1544]: time="2025-01-17T12:21:10.434172836Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.1\"" Jan 17 12:21:10.444220 containerd[1544]: time="2025-01-17T12:21:10.444176337Z" level=info msg="CreateContainer within sandbox \"99dd9e4ee20f35388085cfc493ed39315999eee19d59ba8391cf9393e4a2f823\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Jan 17 12:21:10.452574 containerd[1544]: time="2025-01-17T12:21:10.452476690Z" level=info msg="CreateContainer within sandbox \"99dd9e4ee20f35388085cfc493ed39315999eee19d59ba8391cf9393e4a2f823\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"18cfb2061e3fadd3aa28522e0dbc5973dd923573767be84a3899f4553bd28ee7\"" Jan 17 12:21:10.454923 containerd[1544]: time="2025-01-17T12:21:10.454461123Z" level=info msg="StartContainer for \"18cfb2061e3fadd3aa28522e0dbc5973dd923573767be84a3899f4553bd28ee7\"" Jan 17 12:21:10.456776 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3713057290.mount: Deactivated successfully. Jan 17 12:21:10.485759 systemd[1]: Started cri-containerd-18cfb2061e3fadd3aa28522e0dbc5973dd923573767be84a3899f4553bd28ee7.scope - libcontainer container 18cfb2061e3fadd3aa28522e0dbc5973dd923573767be84a3899f4553bd28ee7. Jan 17 12:21:10.514389 containerd[1544]: time="2025-01-17T12:21:10.514317274Z" level=info msg="StartContainer for \"18cfb2061e3fadd3aa28522e0dbc5973dd923573767be84a3899f4553bd28ee7\" returns successfully" Jan 17 12:21:10.794006 kubelet[2723]: I0117 12:21:10.793877 2723 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-784b5c8b57-wvh2g" podStartSLOduration=24.272595094 podStartE2EDuration="29.79386209s" podCreationTimestamp="2025-01-17 12:20:41 +0000 UTC" firstStartedPulling="2025-01-17 12:21:04.912776131 +0000 UTC m=+37.396248714" lastFinishedPulling="2025-01-17 12:21:10.434043127 +0000 UTC m=+42.917515710" observedRunningTime="2025-01-17 12:21:10.791127697 +0000 UTC m=+43.274600289" watchObservedRunningTime="2025-01-17 12:21:10.79386209 +0000 UTC m=+43.277334683" Jan 17 12:21:11.013703 containerd[1544]: time="2025-01-17T12:21:11.013637428Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:21:11.013955 containerd[1544]: time="2025-01-17T12:21:11.013924275Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.29.1: active requests=0, bytes read=77" Jan 17 12:21:11.015772 containerd[1544]: time="2025-01-17T12:21:11.015749920Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.29.1\" with image id \"sha256:421726ace5ed13894f7edf594dd3a462947aedc13d0f69d08525d7369477fb70\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:b8c43e264fe52e0c327b0bf3ac882a0224b33bdd7f4ff58a74242da7d9b00486\", size \"43494504\" in 581.011582ms" Jan 17 12:21:11.015825 containerd[1544]: time="2025-01-17T12:21:11.015773723Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.1\" returns image reference \"sha256:421726ace5ed13894f7edf594dd3a462947aedc13d0f69d08525d7369477fb70\"" Jan 17 12:21:11.016695 containerd[1544]: time="2025-01-17T12:21:11.016573552Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\"" Jan 17 12:21:11.018024 containerd[1544]: time="2025-01-17T12:21:11.017971387Z" level=info msg="CreateContainer within sandbox \"899c759b7380f5fb3825df86319a1ff1551827e14d94f8418cf3528733e21bf4\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Jan 17 12:21:11.041026 containerd[1544]: time="2025-01-17T12:21:11.040997730Z" level=info msg="CreateContainer within sandbox \"899c759b7380f5fb3825df86319a1ff1551827e14d94f8418cf3528733e21bf4\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"64bf7fd4283c4242a2d544470207f637f464a3ae6e7e84d4a9b95781411f20a7\"" Jan 17 12:21:11.041989 containerd[1544]: time="2025-01-17T12:21:11.041367522Z" level=info msg="StartContainer for \"64bf7fd4283c4242a2d544470207f637f464a3ae6e7e84d4a9b95781411f20a7\"" Jan 17 12:21:11.061763 systemd[1]: Started cri-containerd-64bf7fd4283c4242a2d544470207f637f464a3ae6e7e84d4a9b95781411f20a7.scope - libcontainer container 64bf7fd4283c4242a2d544470207f637f464a3ae6e7e84d4a9b95781411f20a7. Jan 17 12:21:11.093593 containerd[1544]: time="2025-01-17T12:21:11.093563263Z" level=info msg="StartContainer for \"64bf7fd4283c4242a2d544470207f637f464a3ae6e7e84d4a9b95781411f20a7\" returns successfully" Jan 17 12:21:11.271227 systemd-networkd[1453]: vxlan.calico: Gained IPv6LL Jan 17 12:21:11.817276 kubelet[2723]: I0117 12:21:11.817121 2723 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-6b697d6776-n97bk" podStartSLOduration=25.703265434 podStartE2EDuration="30.817109055s" podCreationTimestamp="2025-01-17 12:20:41 +0000 UTC" firstStartedPulling="2025-01-17 12:21:05.902494856 +0000 UTC m=+38.385967442" lastFinishedPulling="2025-01-17 12:21:11.016338473 +0000 UTC m=+43.499811063" observedRunningTime="2025-01-17 12:21:11.803605801 +0000 UTC m=+44.287078389" watchObservedRunningTime="2025-01-17 12:21:11.817109055 +0000 UTC m=+44.300581644" Jan 17 12:21:12.386627 systemd[1]: run-containerd-runc-k8s.io-bdd3e8c0d8dc609233a7628780604ac329962b996666ed0d28c27db3140b9f1c-runc.S8gpyo.mount: Deactivated successfully. Jan 17 12:21:12.404906 kubelet[2723]: I0117 12:21:12.404650 2723 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 17 12:21:12.788463 kubelet[2723]: I0117 12:21:12.787737 2723 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 17 12:21:13.913495 containerd[1544]: time="2025-01-17T12:21:13.913461486Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:21:13.914147 containerd[1544]: time="2025-01-17T12:21:13.914126335Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1: active requests=0, bytes read=10501081" Jan 17 12:21:13.915186 containerd[1544]: time="2025-01-17T12:21:13.915167562Z" level=info msg="ImageCreate event name:\"sha256:8b7d18f262d5cf6a6343578ad0db68a140c4c9989d9e02c58c27cb5d2c70320f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:21:13.915955 containerd[1544]: time="2025-01-17T12:21:13.915930966Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:a338da9488cbaa83c78457c3d7354d84149969c0480e88dd768e036632ff5b76\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 17 12:21:13.916526 containerd[1544]: time="2025-01-17T12:21:13.916510537Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\" with image id \"sha256:8b7d18f262d5cf6a6343578ad0db68a140c4c9989d9e02c58c27cb5d2c70320f\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:a338da9488cbaa83c78457c3d7354d84149969c0480e88dd768e036632ff5b76\", size \"11994117\" in 2.899915342s" Jan 17 12:21:13.916556 containerd[1544]: time="2025-01-17T12:21:13.916528833Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\" returns image reference \"sha256:8b7d18f262d5cf6a6343578ad0db68a140c4c9989d9e02c58c27cb5d2c70320f\"" Jan 17 12:21:13.918375 containerd[1544]: time="2025-01-17T12:21:13.918309139Z" level=info msg="CreateContainer within sandbox \"775d8d469d2ab031a2239f0071ad078481fb2c0911d17de24c2d4477a636c3c0\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Jan 17 12:21:13.924666 containerd[1544]: time="2025-01-17T12:21:13.924643767Z" level=info msg="CreateContainer within sandbox \"775d8d469d2ab031a2239f0071ad078481fb2c0911d17de24c2d4477a636c3c0\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"e47d6a92416408776b3c36f47c50690b0b4cf8aa9a70825962b589c18cb8a567\"" Jan 17 12:21:13.925063 containerd[1544]: time="2025-01-17T12:21:13.925044856Z" level=info msg="StartContainer for \"e47d6a92416408776b3c36f47c50690b0b4cf8aa9a70825962b589c18cb8a567\"" Jan 17 12:21:13.944773 systemd[1]: Started cri-containerd-e47d6a92416408776b3c36f47c50690b0b4cf8aa9a70825962b589c18cb8a567.scope - libcontainer container e47d6a92416408776b3c36f47c50690b0b4cf8aa9a70825962b589c18cb8a567. Jan 17 12:21:13.985944 containerd[1544]: time="2025-01-17T12:21:13.985904558Z" level=info msg="StartContainer for \"e47d6a92416408776b3c36f47c50690b0b4cf8aa9a70825962b589c18cb8a567\" returns successfully" Jan 17 12:21:15.434850 kubelet[2723]: I0117 12:21:15.434737 2723 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Jan 17 12:21:15.442196 kubelet[2723]: I0117 12:21:15.442182 2723 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Jan 17 12:21:26.359466 kubelet[2723]: I0117 12:21:26.359170 2723 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 17 12:21:26.370925 kubelet[2723]: I0117 12:21:26.370887 2723 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-zj6v2" podStartSLOduration=35.523079336 podStartE2EDuration="45.370875102s" podCreationTimestamp="2025-01-17 12:20:41 +0000 UTC" firstStartedPulling="2025-01-17 12:21:04.069287561 +0000 UTC m=+36.552760145" lastFinishedPulling="2025-01-17 12:21:13.917083328 +0000 UTC m=+46.400555911" observedRunningTime="2025-01-17 12:21:14.833012698 +0000 UTC m=+47.316485290" watchObservedRunningTime="2025-01-17 12:21:26.370875102 +0000 UTC m=+58.854347688" Jan 17 12:21:27.599072 containerd[1544]: time="2025-01-17T12:21:27.599027061Z" level=info msg="StopPodSandbox for \"d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1\"" Jan 17 12:21:27.944750 containerd[1544]: 2025-01-17 12:21:27.917 [WARNING][5327] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--6f6b679f8f--6f26n-eth0", GenerateName:"coredns-6f6b679f8f-", Namespace:"kube-system", SelfLink:"", UID:"7756b0ac-5d50-42a5-ba21-9f884a8f9486", ResourceVersion:"816", Generation:0, CreationTimestamp:time.Date(2025, time.January, 17, 12, 20, 33, 0, time.Local), DeletionTimestamp:<nil>, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"6f6b679f8f", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"48c13b9221db34d5ac7a06035688d231254dc30dbe87dd08127ed8b235b5a728", Pod:"coredns-6f6b679f8f-6f26n", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali5874a32a051", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 17 12:21:27.944750 containerd[1544]: 2025-01-17 12:21:27.921 [INFO][5327] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" Jan 17 12:21:27.944750 containerd[1544]: 2025-01-17 12:21:27.921 [INFO][5327] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" iface="eth0" netns="" Jan 17 12:21:27.944750 containerd[1544]: 2025-01-17 12:21:27.921 [INFO][5327] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" Jan 17 12:21:27.944750 containerd[1544]: 2025-01-17 12:21:27.921 [INFO][5327] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" Jan 17 12:21:27.944750 containerd[1544]: 2025-01-17 12:21:27.937 [INFO][5333] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" HandleID="k8s-pod-network.d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" Workload="localhost-k8s-coredns--6f6b679f8f--6f26n-eth0" Jan 17 12:21:27.944750 containerd[1544]: 2025-01-17 12:21:27.937 [INFO][5333] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jan 17 12:21:27.944750 containerd[1544]: 2025-01-17 12:21:27.937 [INFO][5333] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jan 17 12:21:27.944750 containerd[1544]: 2025-01-17 12:21:27.941 [WARNING][5333] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" HandleID="k8s-pod-network.d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" Workload="localhost-k8s-coredns--6f6b679f8f--6f26n-eth0" Jan 17 12:21:27.944750 containerd[1544]: 2025-01-17 12:21:27.941 [INFO][5333] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" HandleID="k8s-pod-network.d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" Workload="localhost-k8s-coredns--6f6b679f8f--6f26n-eth0" Jan 17 12:21:27.944750 containerd[1544]: 2025-01-17 12:21:27.941 [INFO][5333] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jan 17 12:21:27.944750 containerd[1544]: 2025-01-17 12:21:27.943 [INFO][5327] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" Jan 17 12:21:27.944750 containerd[1544]: time="2025-01-17T12:21:27.944631195Z" level=info msg="TearDown network for sandbox \"d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1\" successfully" Jan 17 12:21:27.944750 containerd[1544]: time="2025-01-17T12:21:27.944647800Z" level=info msg="StopPodSandbox for \"d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1\" returns successfully" Jan 17 12:21:27.978277 containerd[1544]: time="2025-01-17T12:21:27.978235605Z" level=info msg="RemovePodSandbox for \"d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1\"" Jan 17 12:21:27.989102 containerd[1544]: time="2025-01-17T12:21:27.989069248Z" level=info msg="Forcibly stopping sandbox \"d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1\"" Jan 17 12:21:28.038657 containerd[1544]: 2025-01-17 12:21:28.017 [WARNING][5351] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--6f6b679f8f--6f26n-eth0", GenerateName:"coredns-6f6b679f8f-", Namespace:"kube-system", SelfLink:"", UID:"7756b0ac-5d50-42a5-ba21-9f884a8f9486", ResourceVersion:"816", Generation:0, CreationTimestamp:time.Date(2025, time.January, 17, 12, 20, 33, 0, time.Local), DeletionTimestamp:<nil>, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"6f6b679f8f", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"48c13b9221db34d5ac7a06035688d231254dc30dbe87dd08127ed8b235b5a728", Pod:"coredns-6f6b679f8f-6f26n", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.134/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali5874a32a051", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 17 12:21:28.038657 containerd[1544]: 2025-01-17 12:21:28.017 [INFO][5351] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" Jan 17 12:21:28.038657 containerd[1544]: 2025-01-17 12:21:28.017 [INFO][5351] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" iface="eth0" netns="" Jan 17 12:21:28.038657 containerd[1544]: 2025-01-17 12:21:28.017 [INFO][5351] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" Jan 17 12:21:28.038657 containerd[1544]: 2025-01-17 12:21:28.017 [INFO][5351] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" Jan 17 12:21:28.038657 containerd[1544]: 2025-01-17 12:21:28.032 [INFO][5357] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" HandleID="k8s-pod-network.d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" Workload="localhost-k8s-coredns--6f6b679f8f--6f26n-eth0" Jan 17 12:21:28.038657 containerd[1544]: 2025-01-17 12:21:28.032 [INFO][5357] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jan 17 12:21:28.038657 containerd[1544]: 2025-01-17 12:21:28.032 [INFO][5357] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jan 17 12:21:28.038657 containerd[1544]: 2025-01-17 12:21:28.035 [WARNING][5357] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" HandleID="k8s-pod-network.d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" Workload="localhost-k8s-coredns--6f6b679f8f--6f26n-eth0" Jan 17 12:21:28.038657 containerd[1544]: 2025-01-17 12:21:28.035 [INFO][5357] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" HandleID="k8s-pod-network.d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" Workload="localhost-k8s-coredns--6f6b679f8f--6f26n-eth0" Jan 17 12:21:28.038657 containerd[1544]: 2025-01-17 12:21:28.036 [INFO][5357] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jan 17 12:21:28.038657 containerd[1544]: 2025-01-17 12:21:28.037 [INFO][5351] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1" Jan 17 12:21:28.038991 containerd[1544]: time="2025-01-17T12:21:28.038700306Z" level=info msg="TearDown network for sandbox \"d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1\" successfully" Jan 17 12:21:28.045358 containerd[1544]: time="2025-01-17T12:21:28.045329661Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 17 12:21:28.050846 containerd[1544]: time="2025-01-17T12:21:28.050821477Z" level=info msg="RemovePodSandbox \"d39973ff083bd8c9735af0b7b40822bd0cc1c083ec1c78b7e58ebbbd720bb0e1\" returns successfully" Jan 17 12:21:28.057293 containerd[1544]: time="2025-01-17T12:21:28.057274123Z" level=info msg="StopPodSandbox for \"81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef\"" Jan 17 12:21:28.108898 containerd[1544]: 2025-01-17 12:21:28.086 [WARNING][5375] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--6f6b679f8f--bq2mf-eth0", GenerateName:"coredns-6f6b679f8f-", Namespace:"kube-system", SelfLink:"", UID:"95fe1c4f-c274-4a0d-aa79-7b5cb4353bdf", ResourceVersion:"812", Generation:0, CreationTimestamp:time.Date(2025, time.January, 17, 12, 20, 33, 0, time.Local), DeletionTimestamp:<nil>, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"6f6b679f8f", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"2979d448e752ef269fd63a5774b653a4d50161125cfc8a6dbc8abb731174203a", Pod:"coredns-6f6b679f8f-bq2mf", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali2818078aafc", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 17 12:21:28.108898 containerd[1544]: 2025-01-17 12:21:28.087 [INFO][5375] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" Jan 17 12:21:28.108898 containerd[1544]: 2025-01-17 12:21:28.087 [INFO][5375] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" iface="eth0" netns="" Jan 17 12:21:28.108898 containerd[1544]: 2025-01-17 12:21:28.087 [INFO][5375] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" Jan 17 12:21:28.108898 containerd[1544]: 2025-01-17 12:21:28.087 [INFO][5375] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" Jan 17 12:21:28.108898 containerd[1544]: 2025-01-17 12:21:28.102 [INFO][5381] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" HandleID="k8s-pod-network.81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" Workload="localhost-k8s-coredns--6f6b679f8f--bq2mf-eth0" Jan 17 12:21:28.108898 containerd[1544]: 2025-01-17 12:21:28.102 [INFO][5381] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jan 17 12:21:28.108898 containerd[1544]: 2025-01-17 12:21:28.102 [INFO][5381] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jan 17 12:21:28.108898 containerd[1544]: 2025-01-17 12:21:28.106 [WARNING][5381] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" HandleID="k8s-pod-network.81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" Workload="localhost-k8s-coredns--6f6b679f8f--bq2mf-eth0" Jan 17 12:21:28.108898 containerd[1544]: 2025-01-17 12:21:28.106 [INFO][5381] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" HandleID="k8s-pod-network.81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" Workload="localhost-k8s-coredns--6f6b679f8f--bq2mf-eth0" Jan 17 12:21:28.108898 containerd[1544]: 2025-01-17 12:21:28.106 [INFO][5381] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jan 17 12:21:28.108898 containerd[1544]: 2025-01-17 12:21:28.107 [INFO][5375] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" Jan 17 12:21:28.108898 containerd[1544]: time="2025-01-17T12:21:28.108872155Z" level=info msg="TearDown network for sandbox \"81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef\" successfully" Jan 17 12:21:28.108898 containerd[1544]: time="2025-01-17T12:21:28.108888206Z" level=info msg="StopPodSandbox for \"81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef\" returns successfully" Jan 17 12:21:28.110940 containerd[1544]: time="2025-01-17T12:21:28.109152400Z" level=info msg="RemovePodSandbox for \"81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef\"" Jan 17 12:21:28.110940 containerd[1544]: time="2025-01-17T12:21:28.109167297Z" level=info msg="Forcibly stopping sandbox \"81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef\"" Jan 17 12:21:28.152854 containerd[1544]: 2025-01-17 12:21:28.131 [WARNING][5399] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--6f6b679f8f--bq2mf-eth0", GenerateName:"coredns-6f6b679f8f-", Namespace:"kube-system", SelfLink:"", UID:"95fe1c4f-c274-4a0d-aa79-7b5cb4353bdf", ResourceVersion:"812", Generation:0, CreationTimestamp:time.Date(2025, time.January, 17, 12, 20, 33, 0, time.Local), DeletionTimestamp:<nil>, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"6f6b679f8f", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"2979d448e752ef269fd63a5774b653a4d50161125cfc8a6dbc8abb731174203a", Pod:"coredns-6f6b679f8f-bq2mf", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali2818078aafc", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 17 12:21:28.152854 containerd[1544]: 2025-01-17 12:21:28.131 [INFO][5399] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" Jan 17 12:21:28.152854 containerd[1544]: 2025-01-17 12:21:28.131 [INFO][5399] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" iface="eth0" netns="" Jan 17 12:21:28.152854 containerd[1544]: 2025-01-17 12:21:28.131 [INFO][5399] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" Jan 17 12:21:28.152854 containerd[1544]: 2025-01-17 12:21:28.132 [INFO][5399] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" Jan 17 12:21:28.152854 containerd[1544]: 2025-01-17 12:21:28.145 [INFO][5405] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" HandleID="k8s-pod-network.81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" Workload="localhost-k8s-coredns--6f6b679f8f--bq2mf-eth0" Jan 17 12:21:28.152854 containerd[1544]: 2025-01-17 12:21:28.145 [INFO][5405] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jan 17 12:21:28.152854 containerd[1544]: 2025-01-17 12:21:28.145 [INFO][5405] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jan 17 12:21:28.152854 containerd[1544]: 2025-01-17 12:21:28.149 [WARNING][5405] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" HandleID="k8s-pod-network.81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" Workload="localhost-k8s-coredns--6f6b679f8f--bq2mf-eth0" Jan 17 12:21:28.152854 containerd[1544]: 2025-01-17 12:21:28.149 [INFO][5405] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" HandleID="k8s-pod-network.81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" Workload="localhost-k8s-coredns--6f6b679f8f--bq2mf-eth0" Jan 17 12:21:28.152854 containerd[1544]: 2025-01-17 12:21:28.150 [INFO][5405] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jan 17 12:21:28.152854 containerd[1544]: 2025-01-17 12:21:28.151 [INFO][5399] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef" Jan 17 12:21:28.153457 containerd[1544]: time="2025-01-17T12:21:28.152873892Z" level=info msg="TearDown network for sandbox \"81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef\" successfully" Jan 17 12:21:28.154103 containerd[1544]: time="2025-01-17T12:21:28.154085824Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 17 12:21:28.154153 containerd[1544]: time="2025-01-17T12:21:28.154117786Z" level=info msg="RemovePodSandbox \"81f95f3de86761c591acf6653603749b60c185c281e781790658d16ec2de5eef\" returns successfully" Jan 17 12:21:28.154521 containerd[1544]: time="2025-01-17T12:21:28.154455581Z" level=info msg="StopPodSandbox for \"977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598\"" Jan 17 12:21:28.198607 containerd[1544]: 2025-01-17 12:21:28.179 [WARNING][5423] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--6b697d6776--wzkmk-eth0", GenerateName:"calico-apiserver-6b697d6776-", Namespace:"calico-apiserver", SelfLink:"", UID:"cb76a1e6-a555-4ec5-bb3d-c19efe7e0c24", ResourceVersion:"861", Generation:0, CreationTimestamp:time.Date(2025, time.January, 17, 12, 20, 41, 0, time.Local), DeletionTimestamp:<nil>, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6b697d6776", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"c42d613ff2110abb6ca0e2360ab25abb8b1b10e7bcc7fdb0277c7f14f9b056bd", Pod:"calico-apiserver-6b697d6776-wzkmk", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali4a2f335bdbd", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 17 12:21:28.198607 containerd[1544]: 2025-01-17 12:21:28.179 [INFO][5423] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" Jan 17 12:21:28.198607 containerd[1544]: 2025-01-17 12:21:28.179 [INFO][5423] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" iface="eth0" netns="" Jan 17 12:21:28.198607 containerd[1544]: 2025-01-17 12:21:28.179 [INFO][5423] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" Jan 17 12:21:28.198607 containerd[1544]: 2025-01-17 12:21:28.179 [INFO][5423] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" Jan 17 12:21:28.198607 containerd[1544]: 2025-01-17 12:21:28.191 [INFO][5429] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" HandleID="k8s-pod-network.977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" Workload="localhost-k8s-calico--apiserver--6b697d6776--wzkmk-eth0" Jan 17 12:21:28.198607 containerd[1544]: 2025-01-17 12:21:28.191 [INFO][5429] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jan 17 12:21:28.198607 containerd[1544]: 2025-01-17 12:21:28.191 [INFO][5429] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jan 17 12:21:28.198607 containerd[1544]: 2025-01-17 12:21:28.194 [WARNING][5429] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" HandleID="k8s-pod-network.977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" Workload="localhost-k8s-calico--apiserver--6b697d6776--wzkmk-eth0" Jan 17 12:21:28.198607 containerd[1544]: 2025-01-17 12:21:28.194 [INFO][5429] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" HandleID="k8s-pod-network.977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" Workload="localhost-k8s-calico--apiserver--6b697d6776--wzkmk-eth0" Jan 17 12:21:28.198607 containerd[1544]: 2025-01-17 12:21:28.196 [INFO][5429] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jan 17 12:21:28.198607 containerd[1544]: 2025-01-17 12:21:28.197 [INFO][5423] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" Jan 17 12:21:28.198607 containerd[1544]: time="2025-01-17T12:21:28.198381965Z" level=info msg="TearDown network for sandbox \"977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598\" successfully" Jan 17 12:21:28.198607 containerd[1544]: time="2025-01-17T12:21:28.198396456Z" level=info msg="StopPodSandbox for \"977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598\" returns successfully" Jan 17 12:21:28.199891 containerd[1544]: time="2025-01-17T12:21:28.199714829Z" level=info msg="RemovePodSandbox for \"977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598\"" Jan 17 12:21:28.199891 containerd[1544]: time="2025-01-17T12:21:28.199731394Z" level=info msg="Forcibly stopping sandbox \"977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598\"" Jan 17 12:21:28.247390 containerd[1544]: 2025-01-17 12:21:28.222 [WARNING][5448] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--6b697d6776--wzkmk-eth0", GenerateName:"calico-apiserver-6b697d6776-", Namespace:"calico-apiserver", SelfLink:"", UID:"cb76a1e6-a555-4ec5-bb3d-c19efe7e0c24", ResourceVersion:"861", Generation:0, CreationTimestamp:time.Date(2025, time.January, 17, 12, 20, 41, 0, time.Local), DeletionTimestamp:<nil>, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6b697d6776", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"c42d613ff2110abb6ca0e2360ab25abb8b1b10e7bcc7fdb0277c7f14f9b056bd", Pod:"calico-apiserver-6b697d6776-wzkmk", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali4a2f335bdbd", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 17 12:21:28.247390 containerd[1544]: 2025-01-17 12:21:28.222 [INFO][5448] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" Jan 17 12:21:28.247390 containerd[1544]: 2025-01-17 12:21:28.222 [INFO][5448] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" iface="eth0" netns="" Jan 17 12:21:28.247390 containerd[1544]: 2025-01-17 12:21:28.222 [INFO][5448] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" Jan 17 12:21:28.247390 containerd[1544]: 2025-01-17 12:21:28.222 [INFO][5448] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" Jan 17 12:21:28.247390 containerd[1544]: 2025-01-17 12:21:28.236 [INFO][5454] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" HandleID="k8s-pod-network.977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" Workload="localhost-k8s-calico--apiserver--6b697d6776--wzkmk-eth0" Jan 17 12:21:28.247390 containerd[1544]: 2025-01-17 12:21:28.236 [INFO][5454] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jan 17 12:21:28.247390 containerd[1544]: 2025-01-17 12:21:28.236 [INFO][5454] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jan 17 12:21:28.247390 containerd[1544]: 2025-01-17 12:21:28.241 [WARNING][5454] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" HandleID="k8s-pod-network.977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" Workload="localhost-k8s-calico--apiserver--6b697d6776--wzkmk-eth0" Jan 17 12:21:28.247390 containerd[1544]: 2025-01-17 12:21:28.241 [INFO][5454] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" HandleID="k8s-pod-network.977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" Workload="localhost-k8s-calico--apiserver--6b697d6776--wzkmk-eth0" Jan 17 12:21:28.247390 containerd[1544]: 2025-01-17 12:21:28.242 [INFO][5454] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jan 17 12:21:28.247390 containerd[1544]: 2025-01-17 12:21:28.245 [INFO][5448] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598" Jan 17 12:21:28.249848 containerd[1544]: time="2025-01-17T12:21:28.247407136Z" level=info msg="TearDown network for sandbox \"977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598\" successfully" Jan 17 12:21:28.249848 containerd[1544]: time="2025-01-17T12:21:28.249653654Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 17 12:21:28.249848 containerd[1544]: time="2025-01-17T12:21:28.249750931Z" level=info msg="RemovePodSandbox \"977a4d12ffe0cf7f328adeacef7296f7ee4d33b3124550ccad3d59071c8f8598\" returns successfully" Jan 17 12:21:28.250235 containerd[1544]: time="2025-01-17T12:21:28.250143764Z" level=info msg="StopPodSandbox for \"711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21\"" Jan 17 12:21:28.295904 containerd[1544]: 2025-01-17 12:21:28.275 [WARNING][5472] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--zj6v2-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"d5c6d7d3-a76d-4ec7-b919-828df6f31125", ResourceVersion:"875", Generation:0, CreationTimestamp:time.Date(2025, time.January, 17, 12, 20, 41, 0, time.Local), DeletionTimestamp:<nil>, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"56747c9949", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"775d8d469d2ab031a2239f0071ad078481fb2c0911d17de24c2d4477a636c3c0", Pod:"csi-node-driver-zj6v2", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calic5bb7ed3b64", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 17 12:21:28.295904 containerd[1544]: 2025-01-17 12:21:28.275 [INFO][5472] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" Jan 17 12:21:28.295904 containerd[1544]: 2025-01-17 12:21:28.275 [INFO][5472] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" iface="eth0" netns="" Jan 17 12:21:28.295904 containerd[1544]: 2025-01-17 12:21:28.275 [INFO][5472] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" Jan 17 12:21:28.295904 containerd[1544]: 2025-01-17 12:21:28.275 [INFO][5472] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" Jan 17 12:21:28.295904 containerd[1544]: 2025-01-17 12:21:28.288 [INFO][5478] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" HandleID="k8s-pod-network.711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" Workload="localhost-k8s-csi--node--driver--zj6v2-eth0" Jan 17 12:21:28.295904 containerd[1544]: 2025-01-17 12:21:28.288 [INFO][5478] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jan 17 12:21:28.295904 containerd[1544]: 2025-01-17 12:21:28.288 [INFO][5478] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jan 17 12:21:28.295904 containerd[1544]: 2025-01-17 12:21:28.292 [WARNING][5478] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" HandleID="k8s-pod-network.711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" Workload="localhost-k8s-csi--node--driver--zj6v2-eth0" Jan 17 12:21:28.295904 containerd[1544]: 2025-01-17 12:21:28.292 [INFO][5478] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" HandleID="k8s-pod-network.711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" Workload="localhost-k8s-csi--node--driver--zj6v2-eth0" Jan 17 12:21:28.295904 containerd[1544]: 2025-01-17 12:21:28.293 [INFO][5478] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jan 17 12:21:28.295904 containerd[1544]: 2025-01-17 12:21:28.294 [INFO][5472] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" Jan 17 12:21:28.295904 containerd[1544]: time="2025-01-17T12:21:28.295859545Z" level=info msg="TearDown network for sandbox \"711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21\" successfully" Jan 17 12:21:28.295904 containerd[1544]: time="2025-01-17T12:21:28.295874706Z" level=info msg="StopPodSandbox for \"711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21\" returns successfully" Jan 17 12:21:28.298125 containerd[1544]: time="2025-01-17T12:21:28.296463828Z" level=info msg="RemovePodSandbox for \"711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21\"" Jan 17 12:21:28.298125 containerd[1544]: time="2025-01-17T12:21:28.296479837Z" level=info msg="Forcibly stopping sandbox \"711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21\"" Jan 17 12:21:28.341179 containerd[1544]: 2025-01-17 12:21:28.321 [WARNING][5497] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--zj6v2-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"d5c6d7d3-a76d-4ec7-b919-828df6f31125", ResourceVersion:"875", Generation:0, CreationTimestamp:time.Date(2025, time.January, 17, 12, 20, 41, 0, time.Local), DeletionTimestamp:<nil>, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"56747c9949", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"775d8d469d2ab031a2239f0071ad078481fb2c0911d17de24c2d4477a636c3c0", Pod:"csi-node-driver-zj6v2", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calic5bb7ed3b64", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 17 12:21:28.341179 containerd[1544]: 2025-01-17 12:21:28.321 [INFO][5497] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" Jan 17 12:21:28.341179 containerd[1544]: 2025-01-17 12:21:28.321 [INFO][5497] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" iface="eth0" netns="" Jan 17 12:21:28.341179 containerd[1544]: 2025-01-17 12:21:28.321 [INFO][5497] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" Jan 17 12:21:28.341179 containerd[1544]: 2025-01-17 12:21:28.321 [INFO][5497] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" Jan 17 12:21:28.341179 containerd[1544]: 2025-01-17 12:21:28.334 [INFO][5503] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" HandleID="k8s-pod-network.711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" Workload="localhost-k8s-csi--node--driver--zj6v2-eth0" Jan 17 12:21:28.341179 containerd[1544]: 2025-01-17 12:21:28.334 [INFO][5503] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jan 17 12:21:28.341179 containerd[1544]: 2025-01-17 12:21:28.335 [INFO][5503] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jan 17 12:21:28.341179 containerd[1544]: 2025-01-17 12:21:28.338 [WARNING][5503] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" HandleID="k8s-pod-network.711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" Workload="localhost-k8s-csi--node--driver--zj6v2-eth0" Jan 17 12:21:28.341179 containerd[1544]: 2025-01-17 12:21:28.338 [INFO][5503] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" HandleID="k8s-pod-network.711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" Workload="localhost-k8s-csi--node--driver--zj6v2-eth0" Jan 17 12:21:28.341179 containerd[1544]: 2025-01-17 12:21:28.339 [INFO][5503] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jan 17 12:21:28.341179 containerd[1544]: 2025-01-17 12:21:28.340 [INFO][5497] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21" Jan 17 12:21:28.341601 containerd[1544]: time="2025-01-17T12:21:28.341201822Z" level=info msg="TearDown network for sandbox \"711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21\" successfully" Jan 17 12:21:28.342501 containerd[1544]: time="2025-01-17T12:21:28.342484239Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 17 12:21:28.342527 containerd[1544]: time="2025-01-17T12:21:28.342517810Z" level=info msg="RemovePodSandbox \"711252bed62345ce28a2d5afc3a378de94b83af93d6157ef979758eea2440f21\" returns successfully" Jan 17 12:21:28.342872 containerd[1544]: time="2025-01-17T12:21:28.342856633Z" level=info msg="StopPodSandbox for \"94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe\"" Jan 17 12:21:28.387206 containerd[1544]: 2025-01-17 12:21:28.365 [WARNING][5522] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--784b5c8b57--wvh2g-eth0", GenerateName:"calico-kube-controllers-784b5c8b57-", Namespace:"calico-system", SelfLink:"", UID:"0c7fe481-2df4-4da8-b215-b904f488836f", ResourceVersion:"847", Generation:0, CreationTimestamp:time.Date(2025, time.January, 17, 12, 20, 41, 0, time.Local), DeletionTimestamp:<nil>, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"784b5c8b57", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"99dd9e4ee20f35388085cfc493ed39315999eee19d59ba8391cf9393e4a2f823", Pod:"calico-kube-controllers-784b5c8b57-wvh2g", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali1dda6d62550", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 17 12:21:28.387206 containerd[1544]: 2025-01-17 12:21:28.365 [INFO][5522] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" Jan 17 12:21:28.387206 containerd[1544]: 2025-01-17 12:21:28.365 [INFO][5522] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" iface="eth0" netns="" Jan 17 12:21:28.387206 containerd[1544]: 2025-01-17 12:21:28.365 [INFO][5522] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" Jan 17 12:21:28.387206 containerd[1544]: 2025-01-17 12:21:28.366 [INFO][5522] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" Jan 17 12:21:28.387206 containerd[1544]: 2025-01-17 12:21:28.380 [INFO][5528] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" HandleID="k8s-pod-network.94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" Workload="localhost-k8s-calico--kube--controllers--784b5c8b57--wvh2g-eth0" Jan 17 12:21:28.387206 containerd[1544]: 2025-01-17 12:21:28.380 [INFO][5528] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jan 17 12:21:28.387206 containerd[1544]: 2025-01-17 12:21:28.380 [INFO][5528] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jan 17 12:21:28.387206 containerd[1544]: 2025-01-17 12:21:28.384 [WARNING][5528] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" HandleID="k8s-pod-network.94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" Workload="localhost-k8s-calico--kube--controllers--784b5c8b57--wvh2g-eth0" Jan 17 12:21:28.387206 containerd[1544]: 2025-01-17 12:21:28.384 [INFO][5528] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" HandleID="k8s-pod-network.94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" Workload="localhost-k8s-calico--kube--controllers--784b5c8b57--wvh2g-eth0" Jan 17 12:21:28.387206 containerd[1544]: 2025-01-17 12:21:28.384 [INFO][5528] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jan 17 12:21:28.387206 containerd[1544]: 2025-01-17 12:21:28.386 [INFO][5522] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" Jan 17 12:21:28.389532 containerd[1544]: time="2025-01-17T12:21:28.387230541Z" level=info msg="TearDown network for sandbox \"94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe\" successfully" Jan 17 12:21:28.389532 containerd[1544]: time="2025-01-17T12:21:28.387246557Z" level=info msg="StopPodSandbox for \"94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe\" returns successfully" Jan 17 12:21:28.389532 containerd[1544]: time="2025-01-17T12:21:28.387640029Z" level=info msg="RemovePodSandbox for \"94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe\"" Jan 17 12:21:28.389532 containerd[1544]: time="2025-01-17T12:21:28.387655040Z" level=info msg="Forcibly stopping sandbox \"94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe\"" Jan 17 12:21:28.432293 containerd[1544]: 2025-01-17 12:21:28.411 [WARNING][5547] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--784b5c8b57--wvh2g-eth0", GenerateName:"calico-kube-controllers-784b5c8b57-", Namespace:"calico-system", SelfLink:"", UID:"0c7fe481-2df4-4da8-b215-b904f488836f", ResourceVersion:"847", Generation:0, CreationTimestamp:time.Date(2025, time.January, 17, 12, 20, 41, 0, time.Local), DeletionTimestamp:<nil>, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"784b5c8b57", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"99dd9e4ee20f35388085cfc493ed39315999eee19d59ba8391cf9393e4a2f823", Pod:"calico-kube-controllers-784b5c8b57-wvh2g", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali1dda6d62550", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 17 12:21:28.432293 containerd[1544]: 2025-01-17 12:21:28.411 [INFO][5547] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" Jan 17 12:21:28.432293 containerd[1544]: 2025-01-17 12:21:28.411 [INFO][5547] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" iface="eth0" netns="" Jan 17 12:21:28.432293 containerd[1544]: 2025-01-17 12:21:28.411 [INFO][5547] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" Jan 17 12:21:28.432293 containerd[1544]: 2025-01-17 12:21:28.411 [INFO][5547] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" Jan 17 12:21:28.432293 containerd[1544]: 2025-01-17 12:21:28.424 [INFO][5553] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" HandleID="k8s-pod-network.94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" Workload="localhost-k8s-calico--kube--controllers--784b5c8b57--wvh2g-eth0" Jan 17 12:21:28.432293 containerd[1544]: 2025-01-17 12:21:28.424 [INFO][5553] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jan 17 12:21:28.432293 containerd[1544]: 2025-01-17 12:21:28.424 [INFO][5553] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jan 17 12:21:28.432293 containerd[1544]: 2025-01-17 12:21:28.429 [WARNING][5553] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" HandleID="k8s-pod-network.94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" Workload="localhost-k8s-calico--kube--controllers--784b5c8b57--wvh2g-eth0" Jan 17 12:21:28.432293 containerd[1544]: 2025-01-17 12:21:28.429 [INFO][5553] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" HandleID="k8s-pod-network.94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" Workload="localhost-k8s-calico--kube--controllers--784b5c8b57--wvh2g-eth0" Jan 17 12:21:28.432293 containerd[1544]: 2025-01-17 12:21:28.429 [INFO][5553] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jan 17 12:21:28.432293 containerd[1544]: 2025-01-17 12:21:28.431 [INFO][5547] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe" Jan 17 12:21:28.433101 containerd[1544]: time="2025-01-17T12:21:28.432319028Z" level=info msg="TearDown network for sandbox \"94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe\" successfully" Jan 17 12:21:28.443688 containerd[1544]: time="2025-01-17T12:21:28.443643113Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 17 12:21:28.443770 containerd[1544]: time="2025-01-17T12:21:28.443694786Z" level=info msg="RemovePodSandbox \"94a6d6e05bbf7088d006e415b5468a56aaf01bd2657985fc5ee824a05c75e3fe\" returns successfully" Jan 17 12:21:28.444176 containerd[1544]: time="2025-01-17T12:21:28.443993917Z" level=info msg="StopPodSandbox for \"fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4\"" Jan 17 12:21:28.487099 containerd[1544]: 2025-01-17 12:21:28.466 [WARNING][5572] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--6b697d6776--n97bk-eth0", GenerateName:"calico-apiserver-6b697d6776-", Namespace:"calico-apiserver", SelfLink:"", UID:"9c19e775-a72c-40f9-9f63-2947d850da10", ResourceVersion:"911", Generation:0, CreationTimestamp:time.Date(2025, time.January, 17, 12, 20, 41, 0, time.Local), DeletionTimestamp:<nil>, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6b697d6776", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"899c759b7380f5fb3825df86319a1ff1551827e14d94f8418cf3528733e21bf4", Pod:"calico-apiserver-6b697d6776-n97bk", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali1b1536c4fdb", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 17 12:21:28.487099 containerd[1544]: 2025-01-17 12:21:28.466 [INFO][5572] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" Jan 17 12:21:28.487099 containerd[1544]: 2025-01-17 12:21:28.466 [INFO][5572] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" iface="eth0" netns="" Jan 17 12:21:28.487099 containerd[1544]: 2025-01-17 12:21:28.466 [INFO][5572] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" Jan 17 12:21:28.487099 containerd[1544]: 2025-01-17 12:21:28.466 [INFO][5572] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" Jan 17 12:21:28.487099 containerd[1544]: 2025-01-17 12:21:28.480 [INFO][5578] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" HandleID="k8s-pod-network.fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" Workload="localhost-k8s-calico--apiserver--6b697d6776--n97bk-eth0" Jan 17 12:21:28.487099 containerd[1544]: 2025-01-17 12:21:28.480 [INFO][5578] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jan 17 12:21:28.487099 containerd[1544]: 2025-01-17 12:21:28.480 [INFO][5578] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jan 17 12:21:28.487099 containerd[1544]: 2025-01-17 12:21:28.484 [WARNING][5578] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" HandleID="k8s-pod-network.fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" Workload="localhost-k8s-calico--apiserver--6b697d6776--n97bk-eth0" Jan 17 12:21:28.487099 containerd[1544]: 2025-01-17 12:21:28.484 [INFO][5578] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" HandleID="k8s-pod-network.fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" Workload="localhost-k8s-calico--apiserver--6b697d6776--n97bk-eth0" Jan 17 12:21:28.487099 containerd[1544]: 2025-01-17 12:21:28.484 [INFO][5578] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jan 17 12:21:28.487099 containerd[1544]: 2025-01-17 12:21:28.486 [INFO][5572] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" Jan 17 12:21:28.487099 containerd[1544]: time="2025-01-17T12:21:28.487076091Z" level=info msg="TearDown network for sandbox \"fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4\" successfully" Jan 17 12:21:28.488794 containerd[1544]: time="2025-01-17T12:21:28.488515007Z" level=info msg="StopPodSandbox for \"fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4\" returns successfully" Jan 17 12:21:28.488836 containerd[1544]: time="2025-01-17T12:21:28.488819957Z" level=info msg="RemovePodSandbox for \"fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4\"" Jan 17 12:21:28.488857 containerd[1544]: time="2025-01-17T12:21:28.488839210Z" level=info msg="Forcibly stopping sandbox \"fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4\"" Jan 17 12:21:28.533367 containerd[1544]: 2025-01-17 12:21:28.510 [WARNING][5596] cni-plugin/k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--6b697d6776--n97bk-eth0", GenerateName:"calico-apiserver-6b697d6776-", Namespace:"calico-apiserver", SelfLink:"", UID:"9c19e775-a72c-40f9-9f63-2947d850da10", ResourceVersion:"911", Generation:0, CreationTimestamp:time.Date(2025, time.January, 17, 12, 20, 41, 0, time.Local), DeletionTimestamp:<nil>, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"6b697d6776", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"899c759b7380f5fb3825df86319a1ff1551827e14d94f8418cf3528733e21bf4", Pod:"calico-apiserver-6b697d6776-n97bk", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali1b1536c4fdb", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 17 12:21:28.533367 containerd[1544]: 2025-01-17 12:21:28.510 [INFO][5596] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" Jan 17 12:21:28.533367 containerd[1544]: 2025-01-17 12:21:28.510 [INFO][5596] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" iface="eth0" netns="" Jan 17 12:21:28.533367 containerd[1544]: 2025-01-17 12:21:28.510 [INFO][5596] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" Jan 17 12:21:28.533367 containerd[1544]: 2025-01-17 12:21:28.510 [INFO][5596] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" Jan 17 12:21:28.533367 containerd[1544]: 2025-01-17 12:21:28.527 [INFO][5602] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" HandleID="k8s-pod-network.fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" Workload="localhost-k8s-calico--apiserver--6b697d6776--n97bk-eth0" Jan 17 12:21:28.533367 containerd[1544]: 2025-01-17 12:21:28.527 [INFO][5602] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jan 17 12:21:28.533367 containerd[1544]: 2025-01-17 12:21:28.527 [INFO][5602] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jan 17 12:21:28.533367 containerd[1544]: 2025-01-17 12:21:28.530 [WARNING][5602] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" HandleID="k8s-pod-network.fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" Workload="localhost-k8s-calico--apiserver--6b697d6776--n97bk-eth0" Jan 17 12:21:28.533367 containerd[1544]: 2025-01-17 12:21:28.530 [INFO][5602] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" HandleID="k8s-pod-network.fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" Workload="localhost-k8s-calico--apiserver--6b697d6776--n97bk-eth0" Jan 17 12:21:28.533367 containerd[1544]: 2025-01-17 12:21:28.531 [INFO][5602] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jan 17 12:21:28.533367 containerd[1544]: 2025-01-17 12:21:28.532 [INFO][5596] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4" Jan 17 12:21:28.533749 containerd[1544]: time="2025-01-17T12:21:28.533388629Z" level=info msg="TearDown network for sandbox \"fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4\" successfully" Jan 17 12:21:28.546190 containerd[1544]: time="2025-01-17T12:21:28.546168160Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 17 12:21:28.549299 containerd[1544]: time="2025-01-17T12:21:28.546204786Z" level=info msg="RemovePodSandbox \"fa08fe5ad282bafd4af90c4f01da1e74b7f365016deb395ea1537206806aa6b4\" returns successfully" Jan 17 12:21:38.292362 systemd[1]: run-containerd-runc-k8s.io-18cfb2061e3fadd3aa28522e0dbc5973dd923573767be84a3899f4553bd28ee7-runc.rZSsHm.mount: Deactivated successfully. Jan 17 12:21:42.341156 systemd[1]: run-containerd-runc-k8s.io-bdd3e8c0d8dc609233a7628780604ac329962b996666ed0d28c27db3140b9f1c-runc.yxW6RR.mount: Deactivated successfully. Jan 17 12:21:43.872830 systemd[1]: Started sshd@7-139.178.70.100:22-147.75.109.163:40274.service - OpenSSH per-connection server daemon (147.75.109.163:40274). Jan 17 12:21:43.953086 sshd[5716]: Accepted publickey for core from 147.75.109.163 port 40274 ssh2: RSA SHA256:d86Zfld7pfipwDMCy9Zh9gJz3C7zt8CsQJU6anwQyxc Jan 17 12:21:43.954708 sshd[5716]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 12:21:43.957831 systemd-logind[1522]: New session 10 of user core. Jan 17 12:21:43.967758 systemd[1]: Started session-10.scope - Session 10 of User core. Jan 17 12:21:44.443891 sshd[5716]: pam_unix(sshd:session): session closed for user core Jan 17 12:21:44.446200 systemd[1]: sshd@7-139.178.70.100:22-147.75.109.163:40274.service: Deactivated successfully. Jan 17 12:21:44.447700 systemd[1]: session-10.scope: Deactivated successfully. Jan 17 12:21:44.448271 systemd-logind[1522]: Session 10 logged out. Waiting for processes to exit. Jan 17 12:21:44.448795 systemd-logind[1522]: Removed session 10. Jan 17 12:21:49.457316 systemd[1]: Started sshd@8-139.178.70.100:22-147.75.109.163:48544.service - OpenSSH per-connection server daemon (147.75.109.163:48544). Jan 17 12:21:49.542901 sshd[5730]: Accepted publickey for core from 147.75.109.163 port 48544 ssh2: RSA SHA256:d86Zfld7pfipwDMCy9Zh9gJz3C7zt8CsQJU6anwQyxc Jan 17 12:21:49.545394 sshd[5730]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 12:21:49.549471 systemd-logind[1522]: New session 11 of user core. Jan 17 12:21:49.556129 systemd[1]: Started session-11.scope - Session 11 of User core. Jan 17 12:21:49.693992 sshd[5730]: pam_unix(sshd:session): session closed for user core Jan 17 12:21:49.700124 systemd[1]: sshd@8-139.178.70.100:22-147.75.109.163:48544.service: Deactivated successfully. Jan 17 12:21:49.701838 systemd[1]: session-11.scope: Deactivated successfully. Jan 17 12:21:49.702377 systemd-logind[1522]: Session 11 logged out. Waiting for processes to exit. Jan 17 12:21:49.703217 systemd-logind[1522]: Removed session 11. Jan 17 12:21:54.705125 systemd[1]: Started sshd@9-139.178.70.100:22-147.75.109.163:48560.service - OpenSSH per-connection server daemon (147.75.109.163:48560). Jan 17 12:21:54.976654 sshd[5772]: Accepted publickey for core from 147.75.109.163 port 48560 ssh2: RSA SHA256:d86Zfld7pfipwDMCy9Zh9gJz3C7zt8CsQJU6anwQyxc Jan 17 12:21:54.977928 sshd[5772]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 12:21:54.985056 systemd-logind[1522]: New session 12 of user core. Jan 17 12:21:54.988804 systemd[1]: Started session-12.scope - Session 12 of User core. Jan 17 12:21:55.240047 sshd[5772]: pam_unix(sshd:session): session closed for user core Jan 17 12:21:55.243923 systemd[1]: sshd@9-139.178.70.100:22-147.75.109.163:48560.service: Deactivated successfully. Jan 17 12:21:55.245994 systemd[1]: session-12.scope: Deactivated successfully. Jan 17 12:21:55.247030 systemd-logind[1522]: Session 12 logged out. Waiting for processes to exit. Jan 17 12:21:55.248181 systemd-logind[1522]: Removed session 12. Jan 17 12:21:55.254015 systemd[1]: Started sshd@10-139.178.70.100:22-147.75.109.163:48576.service - OpenSSH per-connection server daemon (147.75.109.163:48576). Jan 17 12:21:55.305898 sshd[5786]: Accepted publickey for core from 147.75.109.163 port 48576 ssh2: RSA SHA256:d86Zfld7pfipwDMCy9Zh9gJz3C7zt8CsQJU6anwQyxc Jan 17 12:21:55.310211 sshd[5786]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 12:21:55.321371 systemd-logind[1522]: New session 13 of user core. Jan 17 12:21:55.328842 systemd[1]: Started session-13.scope - Session 13 of User core. Jan 17 12:21:55.535202 sshd[5786]: pam_unix(sshd:session): session closed for user core Jan 17 12:21:55.543464 systemd[1]: sshd@10-139.178.70.100:22-147.75.109.163:48576.service: Deactivated successfully. Jan 17 12:21:55.545997 systemd[1]: session-13.scope: Deactivated successfully. Jan 17 12:21:55.549777 systemd-logind[1522]: Session 13 logged out. Waiting for processes to exit. Jan 17 12:21:55.557953 systemd[1]: Started sshd@11-139.178.70.100:22-147.75.109.163:48578.service - OpenSSH per-connection server daemon (147.75.109.163:48578). Jan 17 12:21:55.561000 systemd-logind[1522]: Removed session 13. Jan 17 12:21:55.672556 sshd[5797]: Accepted publickey for core from 147.75.109.163 port 48578 ssh2: RSA SHA256:d86Zfld7pfipwDMCy9Zh9gJz3C7zt8CsQJU6anwQyxc Jan 17 12:21:55.674246 sshd[5797]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 12:21:55.677884 systemd-logind[1522]: New session 14 of user core. Jan 17 12:21:55.683897 systemd[1]: Started session-14.scope - Session 14 of User core. Jan 17 12:21:55.806836 sshd[5797]: pam_unix(sshd:session): session closed for user core Jan 17 12:21:55.809371 systemd[1]: sshd@11-139.178.70.100:22-147.75.109.163:48578.service: Deactivated successfully. Jan 17 12:21:55.810686 systemd[1]: session-14.scope: Deactivated successfully. Jan 17 12:21:55.811224 systemd-logind[1522]: Session 14 logged out. Waiting for processes to exit. Jan 17 12:21:55.812031 systemd-logind[1522]: Removed session 14. Jan 17 12:22:00.817194 systemd[1]: Started sshd@12-139.178.70.100:22-147.75.109.163:54170.service - OpenSSH per-connection server daemon (147.75.109.163:54170). Jan 17 12:22:00.947662 sshd[5814]: Accepted publickey for core from 147.75.109.163 port 54170 ssh2: RSA SHA256:d86Zfld7pfipwDMCy9Zh9gJz3C7zt8CsQJU6anwQyxc Jan 17 12:22:00.956283 sshd[5814]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 12:22:00.959401 systemd-logind[1522]: New session 15 of user core. Jan 17 12:22:00.967991 systemd[1]: Started session-15.scope - Session 15 of User core. Jan 17 12:22:01.102196 sshd[5814]: pam_unix(sshd:session): session closed for user core Jan 17 12:22:01.104797 systemd-logind[1522]: Session 15 logged out. Waiting for processes to exit. Jan 17 12:22:01.104963 systemd[1]: sshd@12-139.178.70.100:22-147.75.109.163:54170.service: Deactivated successfully. Jan 17 12:22:01.106256 systemd[1]: session-15.scope: Deactivated successfully. Jan 17 12:22:01.106892 systemd-logind[1522]: Removed session 15. Jan 17 12:22:06.108935 systemd[1]: Started sshd@13-139.178.70.100:22-147.75.109.163:54176.service - OpenSSH per-connection server daemon (147.75.109.163:54176). Jan 17 12:22:06.171687 sshd[5830]: Accepted publickey for core from 147.75.109.163 port 54176 ssh2: RSA SHA256:d86Zfld7pfipwDMCy9Zh9gJz3C7zt8CsQJU6anwQyxc Jan 17 12:22:06.172515 sshd[5830]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 12:22:06.175318 systemd-logind[1522]: New session 16 of user core. Jan 17 12:22:06.178770 systemd[1]: Started session-16.scope - Session 16 of User core. Jan 17 12:22:06.333721 sshd[5830]: pam_unix(sshd:session): session closed for user core Jan 17 12:22:06.339472 systemd-logind[1522]: Session 16 logged out. Waiting for processes to exit. Jan 17 12:22:06.339586 systemd[1]: sshd@13-139.178.70.100:22-147.75.109.163:54176.service: Deactivated successfully. Jan 17 12:22:06.340635 systemd[1]: session-16.scope: Deactivated successfully. Jan 17 12:22:06.341190 systemd-logind[1522]: Removed session 16. Jan 17 12:22:11.341470 systemd[1]: Started sshd@14-139.178.70.100:22-147.75.109.163:54276.service - OpenSSH per-connection server daemon (147.75.109.163:54276). Jan 17 12:22:11.662695 sshd[5862]: Accepted publickey for core from 147.75.109.163 port 54276 ssh2: RSA SHA256:d86Zfld7pfipwDMCy9Zh9gJz3C7zt8CsQJU6anwQyxc Jan 17 12:22:11.663924 sshd[5862]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 12:22:11.667872 systemd-logind[1522]: New session 17 of user core. Jan 17 12:22:11.674843 systemd[1]: Started session-17.scope - Session 17 of User core. Jan 17 12:22:11.830805 sshd[5862]: pam_unix(sshd:session): session closed for user core Jan 17 12:22:11.833166 systemd-logind[1522]: Session 17 logged out. Waiting for processes to exit. Jan 17 12:22:11.833939 systemd[1]: sshd@14-139.178.70.100:22-147.75.109.163:54276.service: Deactivated successfully. Jan 17 12:22:11.835113 systemd[1]: session-17.scope: Deactivated successfully. Jan 17 12:22:11.836029 systemd-logind[1522]: Removed session 17. Jan 17 12:22:12.335059 systemd[1]: run-containerd-runc-k8s.io-bdd3e8c0d8dc609233a7628780604ac329962b996666ed0d28c27db3140b9f1c-runc.L0kcUY.mount: Deactivated successfully. Jan 17 12:22:16.840553 systemd[1]: Started sshd@15-139.178.70.100:22-147.75.109.163:54292.service - OpenSSH per-connection server daemon (147.75.109.163:54292). Jan 17 12:22:16.904035 sshd[5896]: Accepted publickey for core from 147.75.109.163 port 54292 ssh2: RSA SHA256:d86Zfld7pfipwDMCy9Zh9gJz3C7zt8CsQJU6anwQyxc Jan 17 12:22:16.905235 sshd[5896]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 12:22:16.908714 systemd-logind[1522]: New session 18 of user core. Jan 17 12:22:16.916794 systemd[1]: Started session-18.scope - Session 18 of User core. Jan 17 12:22:17.048527 sshd[5896]: pam_unix(sshd:session): session closed for user core Jan 17 12:22:17.054281 systemd[1]: sshd@15-139.178.70.100:22-147.75.109.163:54292.service: Deactivated successfully. Jan 17 12:22:17.055250 systemd[1]: session-18.scope: Deactivated successfully. Jan 17 12:22:17.055663 systemd-logind[1522]: Session 18 logged out. Waiting for processes to exit. Jan 17 12:22:17.062738 systemd[1]: Started sshd@16-139.178.70.100:22-147.75.109.163:54306.service - OpenSSH per-connection server daemon (147.75.109.163:54306). Jan 17 12:22:17.063615 systemd-logind[1522]: Removed session 18. Jan 17 12:22:17.083974 sshd[5908]: Accepted publickey for core from 147.75.109.163 port 54306 ssh2: RSA SHA256:d86Zfld7pfipwDMCy9Zh9gJz3C7zt8CsQJU6anwQyxc Jan 17 12:22:17.084947 sshd[5908]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 12:22:17.087995 systemd-logind[1522]: New session 19 of user core. Jan 17 12:22:17.095766 systemd[1]: Started session-19.scope - Session 19 of User core. Jan 17 12:22:17.432283 sshd[5908]: pam_unix(sshd:session): session closed for user core Jan 17 12:22:17.438119 systemd[1]: sshd@16-139.178.70.100:22-147.75.109.163:54306.service: Deactivated successfully. Jan 17 12:22:17.439072 systemd[1]: session-19.scope: Deactivated successfully. Jan 17 12:22:17.439877 systemd-logind[1522]: Session 19 logged out. Waiting for processes to exit. Jan 17 12:22:17.440918 systemd[1]: Started sshd@17-139.178.70.100:22-147.75.109.163:38532.service - OpenSSH per-connection server daemon (147.75.109.163:38532). Jan 17 12:22:17.442438 systemd-logind[1522]: Removed session 19. Jan 17 12:22:17.538988 sshd[5919]: Accepted publickey for core from 147.75.109.163 port 38532 ssh2: RSA SHA256:d86Zfld7pfipwDMCy9Zh9gJz3C7zt8CsQJU6anwQyxc Jan 17 12:22:17.557269 sshd[5919]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 12:22:17.560784 systemd-logind[1522]: New session 20 of user core. Jan 17 12:22:17.566759 systemd[1]: Started session-20.scope - Session 20 of User core. Jan 17 12:22:19.117811 sshd[5919]: pam_unix(sshd:session): session closed for user core Jan 17 12:22:19.125633 systemd[1]: Started sshd@18-139.178.70.100:22-147.75.109.163:38534.service - OpenSSH per-connection server daemon (147.75.109.163:38534). Jan 17 12:22:19.131915 systemd[1]: sshd@17-139.178.70.100:22-147.75.109.163:38532.service: Deactivated successfully. Jan 17 12:22:19.135321 systemd[1]: session-20.scope: Deactivated successfully. Jan 17 12:22:19.137168 systemd-logind[1522]: Session 20 logged out. Waiting for processes to exit. Jan 17 12:22:19.138085 systemd-logind[1522]: Removed session 20. Jan 17 12:22:19.192594 sshd[5933]: Accepted publickey for core from 147.75.109.163 port 38534 ssh2: RSA SHA256:d86Zfld7pfipwDMCy9Zh9gJz3C7zt8CsQJU6anwQyxc Jan 17 12:22:19.194350 sshd[5933]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 12:22:19.198291 systemd-logind[1522]: New session 21 of user core. Jan 17 12:22:19.206857 systemd[1]: Started session-21.scope - Session 21 of User core. Jan 17 12:22:19.876547 sshd[5933]: pam_unix(sshd:session): session closed for user core Jan 17 12:22:19.882926 systemd[1]: sshd@18-139.178.70.100:22-147.75.109.163:38534.service: Deactivated successfully. Jan 17 12:22:19.883914 systemd[1]: session-21.scope: Deactivated successfully. Jan 17 12:22:19.884361 systemd-logind[1522]: Session 21 logged out. Waiting for processes to exit. Jan 17 12:22:19.888857 systemd[1]: Started sshd@19-139.178.70.100:22-147.75.109.163:38542.service - OpenSSH per-connection server daemon (147.75.109.163:38542). Jan 17 12:22:19.891541 systemd-logind[1522]: Removed session 21. Jan 17 12:22:19.948249 sshd[5948]: Accepted publickey for core from 147.75.109.163 port 38542 ssh2: RSA SHA256:d86Zfld7pfipwDMCy9Zh9gJz3C7zt8CsQJU6anwQyxc Jan 17 12:22:19.949193 sshd[5948]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 12:22:19.952950 systemd-logind[1522]: New session 22 of user core. Jan 17 12:22:19.956766 systemd[1]: Started session-22.scope - Session 22 of User core. Jan 17 12:22:20.064293 sshd[5948]: pam_unix(sshd:session): session closed for user core Jan 17 12:22:20.065905 systemd[1]: sshd@19-139.178.70.100:22-147.75.109.163:38542.service: Deactivated successfully. Jan 17 12:22:20.067121 systemd[1]: session-22.scope: Deactivated successfully. Jan 17 12:22:20.067877 systemd-logind[1522]: Session 22 logged out. Waiting for processes to exit. Jan 17 12:22:20.068488 systemd-logind[1522]: Removed session 22. Jan 17 12:22:25.073892 systemd[1]: Started sshd@20-139.178.70.100:22-147.75.109.163:38558.service - OpenSSH per-connection server daemon (147.75.109.163:38558). Jan 17 12:22:25.206870 sshd[5966]: Accepted publickey for core from 147.75.109.163 port 38558 ssh2: RSA SHA256:d86Zfld7pfipwDMCy9Zh9gJz3C7zt8CsQJU6anwQyxc Jan 17 12:22:25.208136 sshd[5966]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 12:22:25.210835 systemd-logind[1522]: New session 23 of user core. Jan 17 12:22:25.215762 systemd[1]: Started session-23.scope - Session 23 of User core. Jan 17 12:22:25.382579 sshd[5966]: pam_unix(sshd:session): session closed for user core Jan 17 12:22:25.384634 systemd-logind[1522]: Session 23 logged out. Waiting for processes to exit. Jan 17 12:22:25.385124 systemd[1]: sshd@20-139.178.70.100:22-147.75.109.163:38558.service: Deactivated successfully. Jan 17 12:22:25.386485 systemd[1]: session-23.scope: Deactivated successfully. Jan 17 12:22:25.387591 systemd-logind[1522]: Removed session 23. Jan 17 12:22:30.393198 systemd[1]: Started sshd@21-139.178.70.100:22-147.75.109.163:46226.service - OpenSSH per-connection server daemon (147.75.109.163:46226). Jan 17 12:22:30.429472 sshd[5989]: Accepted publickey for core from 147.75.109.163 port 46226 ssh2: RSA SHA256:d86Zfld7pfipwDMCy9Zh9gJz3C7zt8CsQJU6anwQyxc Jan 17 12:22:30.430461 sshd[5989]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 12:22:30.433364 systemd-logind[1522]: New session 24 of user core. Jan 17 12:22:30.438799 systemd[1]: Started session-24.scope - Session 24 of User core. Jan 17 12:22:30.532050 sshd[5989]: pam_unix(sshd:session): session closed for user core Jan 17 12:22:30.534310 systemd[1]: sshd@21-139.178.70.100:22-147.75.109.163:46226.service: Deactivated successfully. Jan 17 12:22:30.535536 systemd[1]: session-24.scope: Deactivated successfully. Jan 17 12:22:30.536252 systemd-logind[1522]: Session 24 logged out. Waiting for processes to exit. Jan 17 12:22:30.536989 systemd-logind[1522]: Removed session 24. Jan 17 12:22:35.542168 systemd[1]: Started sshd@22-139.178.70.100:22-147.75.109.163:46242.service - OpenSSH per-connection server daemon (147.75.109.163:46242). Jan 17 12:22:35.588710 sshd[6003]: Accepted publickey for core from 147.75.109.163 port 46242 ssh2: RSA SHA256:d86Zfld7pfipwDMCy9Zh9gJz3C7zt8CsQJU6anwQyxc Jan 17 12:22:35.589659 sshd[6003]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 17 12:22:35.593001 systemd-logind[1522]: New session 25 of user core. Jan 17 12:22:35.596767 systemd[1]: Started session-25.scope - Session 25 of User core. Jan 17 12:22:35.684841 sshd[6003]: pam_unix(sshd:session): session closed for user core Jan 17 12:22:35.686977 systemd[1]: sshd@22-139.178.70.100:22-147.75.109.163:46242.service: Deactivated successfully. Jan 17 12:22:35.688214 systemd[1]: session-25.scope: Deactivated successfully. Jan 17 12:22:35.688844 systemd-logind[1522]: Session 25 logged out. Waiting for processes to exit. Jan 17 12:22:35.689387 systemd-logind[1522]: Removed session 25.