Jul 9 23:56:11.750684 kernel: Linux version 6.6.95-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 14.2.1_p20241221 p7) 14.2.1 20241221, GNU ld (Gentoo 2.43 p3) 2.43.1) #1 SMP PREEMPT_DYNAMIC Wed Jul 9 22:08:48 -00 2025 Jul 9 23:56:11.750723 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=vmware flatcar.autologin verity.usrhash=c257b65f06e0ad68d969d5b3e057f031663dc29a4487d91a77595a40c4dc82d6 Jul 9 23:56:11.750742 kernel: Disabled fast string operations Jul 9 23:56:11.750749 kernel: BIOS-provided physical RAM map: Jul 9 23:56:11.750753 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ebff] usable Jul 9 23:56:11.750757 kernel: BIOS-e820: [mem 0x000000000009ec00-0x000000000009ffff] reserved Jul 9 23:56:11.750764 kernel: BIOS-e820: [mem 0x00000000000dc000-0x00000000000fffff] reserved Jul 9 23:56:11.750769 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007fedffff] usable Jul 9 23:56:11.750773 kernel: BIOS-e820: [mem 0x000000007fee0000-0x000000007fefefff] ACPI data Jul 9 23:56:11.750777 kernel: BIOS-e820: [mem 0x000000007feff000-0x000000007fefffff] ACPI NVS Jul 9 23:56:11.750781 kernel: BIOS-e820: [mem 0x000000007ff00000-0x000000007fffffff] usable Jul 9 23:56:11.750785 kernel: BIOS-e820: [mem 0x00000000f0000000-0x00000000f7ffffff] reserved Jul 9 23:56:11.750789 kernel: BIOS-e820: [mem 0x00000000fec00000-0x00000000fec0ffff] reserved Jul 9 23:56:11.750794 kernel: BIOS-e820: [mem 0x00000000fee00000-0x00000000fee00fff] reserved Jul 9 23:56:11.750800 kernel: BIOS-e820: [mem 0x00000000fffe0000-0x00000000ffffffff] reserved Jul 9 23:56:11.750805 kernel: NX (Execute Disable) protection: active Jul 9 23:56:11.750809 kernel: APIC: Static calls initialized Jul 9 23:56:11.750814 kernel: SMBIOS 2.7 present. Jul 9 23:56:11.750819 kernel: DMI: VMware, Inc. VMware Virtual Platform/440BX Desktop Reference Platform, BIOS 6.00 05/28/2020 Jul 9 23:56:11.750824 kernel: vmware: hypercall mode: 0x00 Jul 9 23:56:11.750828 kernel: Hypervisor detected: VMware Jul 9 23:56:11.750833 kernel: vmware: TSC freq read from hypervisor : 3408.000 MHz Jul 9 23:56:11.750839 kernel: vmware: Host bus clock speed read from hypervisor : 66000000 Hz Jul 9 23:56:11.750843 kernel: vmware: using clock offset of 3393798379 ns Jul 9 23:56:11.750848 kernel: tsc: Detected 3408.000 MHz processor Jul 9 23:56:11.750853 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Jul 9 23:56:11.750858 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Jul 9 23:56:11.750863 kernel: last_pfn = 0x80000 max_arch_pfn = 0x400000000 Jul 9 23:56:11.750868 kernel: total RAM covered: 3072M Jul 9 23:56:11.750873 kernel: Found optimal setting for mtrr clean up Jul 9 23:56:11.750879 kernel: gran_size: 64K chunk_size: 64K num_reg: 2 lose cover RAM: 0G Jul 9 23:56:11.750884 kernel: MTRR map: 6 entries (5 fixed + 1 variable; max 21), built from 8 variable MTRRs Jul 9 23:56:11.750890 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Jul 9 23:56:11.750895 kernel: Using GB pages for direct mapping Jul 9 23:56:11.750899 kernel: ACPI: Early table checksum verification disabled Jul 9 23:56:11.750904 kernel: ACPI: RSDP 0x00000000000F6A00 000024 (v02 PTLTD ) Jul 9 23:56:11.750909 kernel: ACPI: XSDT 0x000000007FEE965B 00005C (v01 INTEL 440BX 06040000 VMW 01324272) Jul 9 23:56:11.750914 kernel: ACPI: FACP 0x000000007FEFEE73 0000F4 (v04 INTEL 440BX 06040000 PTL 000F4240) Jul 9 23:56:11.750919 kernel: ACPI: DSDT 0x000000007FEEAD55 01411E (v01 PTLTD Custom 06040000 MSFT 03000001) Jul 9 23:56:11.750924 kernel: ACPI: FACS 0x000000007FEFFFC0 000040 Jul 9 23:56:11.750931 kernel: ACPI: FACS 0x000000007FEFFFC0 000040 Jul 9 23:56:11.750936 kernel: ACPI: BOOT 0x000000007FEEAD2D 000028 (v01 PTLTD $SBFTBL$ 06040000 LTP 00000001) Jul 9 23:56:11.750941 kernel: ACPI: APIC 0x000000007FEEA5EB 000742 (v01 PTLTD ? APIC 06040000 LTP 00000000) Jul 9 23:56:11.750946 kernel: ACPI: MCFG 0x000000007FEEA5AF 00003C (v01 PTLTD $PCITBL$ 06040000 LTP 00000001) Jul 9 23:56:11.750951 kernel: ACPI: SRAT 0x000000007FEE9757 0008A8 (v02 VMWARE MEMPLUG 06040000 VMW 00000001) Jul 9 23:56:11.750956 kernel: ACPI: HPET 0x000000007FEE971F 000038 (v01 VMWARE VMW HPET 06040000 VMW 00000001) Jul 9 23:56:11.750963 kernel: ACPI: WAET 0x000000007FEE96F7 000028 (v01 VMWARE VMW WAET 06040000 VMW 00000001) Jul 9 23:56:11.750968 kernel: ACPI: Reserving FACP table memory at [mem 0x7fefee73-0x7fefef66] Jul 9 23:56:11.750973 kernel: ACPI: Reserving DSDT table memory at [mem 0x7feead55-0x7fefee72] Jul 9 23:56:11.750978 kernel: ACPI: Reserving FACS table memory at [mem 0x7fefffc0-0x7fefffff] Jul 9 23:56:11.750983 kernel: ACPI: Reserving FACS table memory at [mem 0x7fefffc0-0x7fefffff] Jul 9 23:56:11.750988 kernel: ACPI: Reserving BOOT table memory at [mem 0x7feead2d-0x7feead54] Jul 9 23:56:11.750993 kernel: ACPI: Reserving APIC table memory at [mem 0x7feea5eb-0x7feead2c] Jul 9 23:56:11.750998 kernel: ACPI: Reserving MCFG table memory at [mem 0x7feea5af-0x7feea5ea] Jul 9 23:56:11.751003 kernel: ACPI: Reserving SRAT table memory at [mem 0x7fee9757-0x7fee9ffe] Jul 9 23:56:11.751009 kernel: ACPI: Reserving HPET table memory at [mem 0x7fee971f-0x7fee9756] Jul 9 23:56:11.751014 kernel: ACPI: Reserving WAET table memory at [mem 0x7fee96f7-0x7fee971e] Jul 9 23:56:11.751019 kernel: system APIC only can use physical flat Jul 9 23:56:11.751024 kernel: APIC: Switched APIC routing to: physical flat Jul 9 23:56:11.751029 kernel: SRAT: PXM 0 -> APIC 0x00 -> Node 0 Jul 9 23:56:11.751034 kernel: SRAT: PXM 0 -> APIC 0x02 -> Node 0 Jul 9 23:56:11.751039 kernel: SRAT: PXM 0 -> APIC 0x04 -> Node 0 Jul 9 23:56:11.751044 kernel: SRAT: PXM 0 -> APIC 0x06 -> Node 0 Jul 9 23:56:11.751049 kernel: SRAT: PXM 0 -> APIC 0x08 -> Node 0 Jul 9 23:56:11.751055 kernel: SRAT: PXM 0 -> APIC 0x0a -> Node 0 Jul 9 23:56:11.751059 kernel: SRAT: PXM 0 -> APIC 0x0c -> Node 0 Jul 9 23:56:11.751073 kernel: SRAT: PXM 0 -> APIC 0x0e -> Node 0 Jul 9 23:56:11.751090 kernel: SRAT: PXM 0 -> APIC 0x10 -> Node 0 Jul 9 23:56:11.751102 kernel: SRAT: PXM 0 -> APIC 0x12 -> Node 0 Jul 9 23:56:11.751108 kernel: SRAT: PXM 0 -> APIC 0x14 -> Node 0 Jul 9 23:56:11.751113 kernel: SRAT: PXM 0 -> APIC 0x16 -> Node 0 Jul 9 23:56:11.751117 kernel: SRAT: PXM 0 -> APIC 0x18 -> Node 0 Jul 9 23:56:11.751130 kernel: SRAT: PXM 0 -> APIC 0x1a -> Node 0 Jul 9 23:56:11.751135 kernel: SRAT: PXM 0 -> APIC 0x1c -> Node 0 Jul 9 23:56:11.751140 kernel: SRAT: PXM 0 -> APIC 0x1e -> Node 0 Jul 9 23:56:11.751148 kernel: SRAT: PXM 0 -> APIC 0x20 -> Node 0 Jul 9 23:56:11.751152 kernel: SRAT: PXM 0 -> APIC 0x22 -> Node 0 Jul 9 23:56:11.751157 kernel: SRAT: PXM 0 -> APIC 0x24 -> Node 0 Jul 9 23:56:11.751162 kernel: SRAT: PXM 0 -> APIC 0x26 -> Node 0 Jul 9 23:56:11.751167 kernel: SRAT: PXM 0 -> APIC 0x28 -> Node 0 Jul 9 23:56:11.751172 kernel: SRAT: PXM 0 -> APIC 0x2a -> Node 0 Jul 9 23:56:11.751177 kernel: SRAT: PXM 0 -> APIC 0x2c -> Node 0 Jul 9 23:56:11.751182 kernel: SRAT: PXM 0 -> APIC 0x2e -> Node 0 Jul 9 23:56:11.751186 kernel: SRAT: PXM 0 -> APIC 0x30 -> Node 0 Jul 9 23:56:11.751191 kernel: SRAT: PXM 0 -> APIC 0x32 -> Node 0 Jul 9 23:56:11.751198 kernel: SRAT: PXM 0 -> APIC 0x34 -> Node 0 Jul 9 23:56:11.751202 kernel: SRAT: PXM 0 -> APIC 0x36 -> Node 0 Jul 9 23:56:11.751207 kernel: SRAT: PXM 0 -> APIC 0x38 -> Node 0 Jul 9 23:56:11.751212 kernel: SRAT: PXM 0 -> APIC 0x3a -> Node 0 Jul 9 23:56:11.751217 kernel: SRAT: PXM 0 -> APIC 0x3c -> Node 0 Jul 9 23:56:11.751222 kernel: SRAT: PXM 0 -> APIC 0x3e -> Node 0 Jul 9 23:56:11.751227 kernel: SRAT: PXM 0 -> APIC 0x40 -> Node 0 Jul 9 23:56:11.751232 kernel: SRAT: PXM 0 -> APIC 0x42 -> Node 0 Jul 9 23:56:11.751237 kernel: SRAT: PXM 0 -> APIC 0x44 -> Node 0 Jul 9 23:56:11.751242 kernel: SRAT: PXM 0 -> APIC 0x46 -> Node 0 Jul 9 23:56:11.751247 kernel: SRAT: PXM 0 -> APIC 0x48 -> Node 0 Jul 9 23:56:11.751252 kernel: SRAT: PXM 0 -> APIC 0x4a -> Node 0 Jul 9 23:56:11.751257 kernel: SRAT: PXM 0 -> APIC 0x4c -> Node 0 Jul 9 23:56:11.751262 kernel: SRAT: PXM 0 -> APIC 0x4e -> Node 0 Jul 9 23:56:11.751267 kernel: SRAT: PXM 0 -> APIC 0x50 -> Node 0 Jul 9 23:56:11.751272 kernel: SRAT: PXM 0 -> APIC 0x52 -> Node 0 Jul 9 23:56:11.751277 kernel: SRAT: PXM 0 -> APIC 0x54 -> Node 0 Jul 9 23:56:11.751282 kernel: SRAT: PXM 0 -> APIC 0x56 -> Node 0 Jul 9 23:56:11.751287 kernel: SRAT: PXM 0 -> APIC 0x58 -> Node 0 Jul 9 23:56:11.751291 kernel: SRAT: PXM 0 -> APIC 0x5a -> Node 0 Jul 9 23:56:11.751297 kernel: SRAT: PXM 0 -> APIC 0x5c -> Node 0 Jul 9 23:56:11.751302 kernel: SRAT: PXM 0 -> APIC 0x5e -> Node 0 Jul 9 23:56:11.751307 kernel: SRAT: PXM 0 -> APIC 0x60 -> Node 0 Jul 9 23:56:11.751312 kernel: SRAT: PXM 0 -> APIC 0x62 -> Node 0 Jul 9 23:56:11.751317 kernel: SRAT: PXM 0 -> APIC 0x64 -> Node 0 Jul 9 23:56:11.751322 kernel: SRAT: PXM 0 -> APIC 0x66 -> Node 0 Jul 9 23:56:11.751327 kernel: SRAT: PXM 0 -> APIC 0x68 -> Node 0 Jul 9 23:56:11.751331 kernel: SRAT: PXM 0 -> APIC 0x6a -> Node 0 Jul 9 23:56:11.751336 kernel: SRAT: PXM 0 -> APIC 0x6c -> Node 0 Jul 9 23:56:11.751341 kernel: SRAT: PXM 0 -> APIC 0x6e -> Node 0 Jul 9 23:56:11.751347 kernel: SRAT: PXM 0 -> APIC 0x70 -> Node 0 Jul 9 23:56:11.751352 kernel: SRAT: PXM 0 -> APIC 0x72 -> Node 0 Jul 9 23:56:11.751357 kernel: SRAT: PXM 0 -> APIC 0x74 -> Node 0 Jul 9 23:56:11.751365 kernel: SRAT: PXM 0 -> APIC 0x76 -> Node 0 Jul 9 23:56:11.751382 kernel: SRAT: PXM 0 -> APIC 0x78 -> Node 0 Jul 9 23:56:11.751388 kernel: SRAT: PXM 0 -> APIC 0x7a -> Node 0 Jul 9 23:56:11.751394 kernel: SRAT: PXM 0 -> APIC 0x7c -> Node 0 Jul 9 23:56:11.751399 kernel: SRAT: PXM 0 -> APIC 0x7e -> Node 0 Jul 9 23:56:11.751404 kernel: SRAT: PXM 0 -> APIC 0x80 -> Node 0 Jul 9 23:56:11.751411 kernel: SRAT: PXM 0 -> APIC 0x82 -> Node 0 Jul 9 23:56:11.751417 kernel: SRAT: PXM 0 -> APIC 0x84 -> Node 0 Jul 9 23:56:11.751422 kernel: SRAT: PXM 0 -> APIC 0x86 -> Node 0 Jul 9 23:56:11.751427 kernel: SRAT: PXM 0 -> APIC 0x88 -> Node 0 Jul 9 23:56:11.751432 kernel: SRAT: PXM 0 -> APIC 0x8a -> Node 0 Jul 9 23:56:11.751437 kernel: SRAT: PXM 0 -> APIC 0x8c -> Node 0 Jul 9 23:56:11.751443 kernel: SRAT: PXM 0 -> APIC 0x8e -> Node 0 Jul 9 23:56:11.751448 kernel: SRAT: PXM 0 -> APIC 0x90 -> Node 0 Jul 9 23:56:11.751453 kernel: SRAT: PXM 0 -> APIC 0x92 -> Node 0 Jul 9 23:56:11.751458 kernel: SRAT: PXM 0 -> APIC 0x94 -> Node 0 Jul 9 23:56:11.751465 kernel: SRAT: PXM 0 -> APIC 0x96 -> Node 0 Jul 9 23:56:11.751470 kernel: SRAT: PXM 0 -> APIC 0x98 -> Node 0 Jul 9 23:56:11.751475 kernel: SRAT: PXM 0 -> APIC 0x9a -> Node 0 Jul 9 23:56:11.751480 kernel: SRAT: PXM 0 -> APIC 0x9c -> Node 0 Jul 9 23:56:11.751485 kernel: SRAT: PXM 0 -> APIC 0x9e -> Node 0 Jul 9 23:56:11.751490 kernel: SRAT: PXM 0 -> APIC 0xa0 -> Node 0 Jul 9 23:56:11.751495 kernel: SRAT: PXM 0 -> APIC 0xa2 -> Node 0 Jul 9 23:56:11.751501 kernel: SRAT: PXM 0 -> APIC 0xa4 -> Node 0 Jul 9 23:56:11.751506 kernel: SRAT: PXM 0 -> APIC 0xa6 -> Node 0 Jul 9 23:56:11.751511 kernel: SRAT: PXM 0 -> APIC 0xa8 -> Node 0 Jul 9 23:56:11.751517 kernel: SRAT: PXM 0 -> APIC 0xaa -> Node 0 Jul 9 23:56:11.751523 kernel: SRAT: PXM 0 -> APIC 0xac -> Node 0 Jul 9 23:56:11.751528 kernel: SRAT: PXM 0 -> APIC 0xae -> Node 0 Jul 9 23:56:11.751533 kernel: SRAT: PXM 0 -> APIC 0xb0 -> Node 0 Jul 9 23:56:11.751538 kernel: SRAT: PXM 0 -> APIC 0xb2 -> Node 0 Jul 9 23:56:11.751543 kernel: SRAT: PXM 0 -> APIC 0xb4 -> Node 0 Jul 9 23:56:11.751548 kernel: SRAT: PXM 0 -> APIC 0xb6 -> Node 0 Jul 9 23:56:11.751553 kernel: SRAT: PXM 0 -> APIC 0xb8 -> Node 0 Jul 9 23:56:11.751559 kernel: SRAT: PXM 0 -> APIC 0xba -> Node 0 Jul 9 23:56:11.751564 kernel: SRAT: PXM 0 -> APIC 0xbc -> Node 0 Jul 9 23:56:11.751570 kernel: SRAT: PXM 0 -> APIC 0xbe -> Node 0 Jul 9 23:56:11.751575 kernel: SRAT: PXM 0 -> APIC 0xc0 -> Node 0 Jul 9 23:56:11.751581 kernel: SRAT: PXM 0 -> APIC 0xc2 -> Node 0 Jul 9 23:56:11.751586 kernel: SRAT: PXM 0 -> APIC 0xc4 -> Node 0 Jul 9 23:56:11.751591 kernel: SRAT: PXM 0 -> APIC 0xc6 -> Node 0 Jul 9 23:56:11.751596 kernel: SRAT: PXM 0 -> APIC 0xc8 -> Node 0 Jul 9 23:56:11.751601 kernel: SRAT: PXM 0 -> APIC 0xca -> Node 0 Jul 9 23:56:11.751606 kernel: SRAT: PXM 0 -> APIC 0xcc -> Node 0 Jul 9 23:56:11.751611 kernel: SRAT: PXM 0 -> APIC 0xce -> Node 0 Jul 9 23:56:11.751617 kernel: SRAT: PXM 0 -> APIC 0xd0 -> Node 0 Jul 9 23:56:11.751623 kernel: SRAT: PXM 0 -> APIC 0xd2 -> Node 0 Jul 9 23:56:11.751628 kernel: SRAT: PXM 0 -> APIC 0xd4 -> Node 0 Jul 9 23:56:11.751633 kernel: SRAT: PXM 0 -> APIC 0xd6 -> Node 0 Jul 9 23:56:11.751638 kernel: SRAT: PXM 0 -> APIC 0xd8 -> Node 0 Jul 9 23:56:11.751644 kernel: SRAT: PXM 0 -> APIC 0xda -> Node 0 Jul 9 23:56:11.751649 kernel: SRAT: PXM 0 -> APIC 0xdc -> Node 0 Jul 9 23:56:11.751654 kernel: SRAT: PXM 0 -> APIC 0xde -> Node 0 Jul 9 23:56:11.751659 kernel: SRAT: PXM 0 -> APIC 0xe0 -> Node 0 Jul 9 23:56:11.751664 kernel: SRAT: PXM 0 -> APIC 0xe2 -> Node 0 Jul 9 23:56:11.751669 kernel: SRAT: PXM 0 -> APIC 0xe4 -> Node 0 Jul 9 23:56:11.751675 kernel: SRAT: PXM 0 -> APIC 0xe6 -> Node 0 Jul 9 23:56:11.751681 kernel: SRAT: PXM 0 -> APIC 0xe8 -> Node 0 Jul 9 23:56:11.751686 kernel: SRAT: PXM 0 -> APIC 0xea -> Node 0 Jul 9 23:56:11.751691 kernel: SRAT: PXM 0 -> APIC 0xec -> Node 0 Jul 9 23:56:11.751696 kernel: SRAT: PXM 0 -> APIC 0xee -> Node 0 Jul 9 23:56:11.751702 kernel: SRAT: PXM 0 -> APIC 0xf0 -> Node 0 Jul 9 23:56:11.751707 kernel: SRAT: PXM 0 -> APIC 0xf2 -> Node 0 Jul 9 23:56:11.751712 kernel: SRAT: PXM 0 -> APIC 0xf4 -> Node 0 Jul 9 23:56:11.751717 kernel: SRAT: PXM 0 -> APIC 0xf6 -> Node 0 Jul 9 23:56:11.751722 kernel: SRAT: PXM 0 -> APIC 0xf8 -> Node 0 Jul 9 23:56:11.751728 kernel: SRAT: PXM 0 -> APIC 0xfa -> Node 0 Jul 9 23:56:11.751734 kernel: SRAT: PXM 0 -> APIC 0xfc -> Node 0 Jul 9 23:56:11.751739 kernel: SRAT: PXM 0 -> APIC 0xfe -> Node 0 Jul 9 23:56:11.751744 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x0009ffff] Jul 9 23:56:11.751750 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00100000-0x7fffffff] Jul 9 23:56:11.751755 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x80000000-0xbfffffff] hotplug Jul 9 23:56:11.751761 kernel: NUMA: Node 0 [mem 0x00000000-0x0009ffff] + [mem 0x00100000-0x7fffffff] -> [mem 0x00000000-0x7fffffff] Jul 9 23:56:11.751766 kernel: NODE_DATA(0) allocated [mem 0x7fffa000-0x7fffffff] Jul 9 23:56:11.751772 kernel: Zone ranges: Jul 9 23:56:11.751778 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Jul 9 23:56:11.751784 kernel: DMA32 [mem 0x0000000001000000-0x000000007fffffff] Jul 9 23:56:11.751789 kernel: Normal empty Jul 9 23:56:11.751794 kernel: Movable zone start for each node Jul 9 23:56:11.751800 kernel: Early memory node ranges Jul 9 23:56:11.751805 kernel: node 0: [mem 0x0000000000001000-0x000000000009dfff] Jul 9 23:56:11.751810 kernel: node 0: [mem 0x0000000000100000-0x000000007fedffff] Jul 9 23:56:11.751816 kernel: node 0: [mem 0x000000007ff00000-0x000000007fffffff] Jul 9 23:56:11.751821 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007fffffff] Jul 9 23:56:11.751826 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Jul 9 23:56:11.751831 kernel: On node 0, zone DMA: 98 pages in unavailable ranges Jul 9 23:56:11.751838 kernel: On node 0, zone DMA32: 32 pages in unavailable ranges Jul 9 23:56:11.751843 kernel: ACPI: PM-Timer IO Port: 0x1008 Jul 9 23:56:11.751848 kernel: system APIC only can use physical flat Jul 9 23:56:11.751854 kernel: ACPI: LAPIC_NMI (acpi_id[0x00] high edge lint[0x1]) Jul 9 23:56:11.751859 kernel: ACPI: LAPIC_NMI (acpi_id[0x01] high edge lint[0x1]) Jul 9 23:56:11.751864 kernel: ACPI: LAPIC_NMI (acpi_id[0x02] high edge lint[0x1]) Jul 9 23:56:11.751869 kernel: ACPI: LAPIC_NMI (acpi_id[0x03] high edge lint[0x1]) Jul 9 23:56:11.751875 kernel: ACPI: LAPIC_NMI (acpi_id[0x04] high edge lint[0x1]) Jul 9 23:56:11.751880 kernel: ACPI: LAPIC_NMI (acpi_id[0x05] high edge lint[0x1]) Jul 9 23:56:11.751886 kernel: ACPI: LAPIC_NMI (acpi_id[0x06] high edge lint[0x1]) Jul 9 23:56:11.751892 kernel: ACPI: LAPIC_NMI (acpi_id[0x07] high edge lint[0x1]) Jul 9 23:56:11.751897 kernel: ACPI: LAPIC_NMI (acpi_id[0x08] high edge lint[0x1]) Jul 9 23:56:11.751903 kernel: ACPI: LAPIC_NMI (acpi_id[0x09] high edge lint[0x1]) Jul 9 23:56:11.751908 kernel: ACPI: LAPIC_NMI (acpi_id[0x0a] high edge lint[0x1]) Jul 9 23:56:11.751913 kernel: ACPI: LAPIC_NMI (acpi_id[0x0b] high edge lint[0x1]) Jul 9 23:56:11.751918 kernel: ACPI: LAPIC_NMI (acpi_id[0x0c] high edge lint[0x1]) Jul 9 23:56:11.751924 kernel: ACPI: LAPIC_NMI (acpi_id[0x0d] high edge lint[0x1]) Jul 9 23:56:11.751929 kernel: ACPI: LAPIC_NMI (acpi_id[0x0e] high edge lint[0x1]) Jul 9 23:56:11.751934 kernel: ACPI: LAPIC_NMI (acpi_id[0x0f] high edge lint[0x1]) Jul 9 23:56:11.751940 kernel: ACPI: LAPIC_NMI (acpi_id[0x10] high edge lint[0x1]) Jul 9 23:56:11.751946 kernel: ACPI: LAPIC_NMI (acpi_id[0x11] high edge lint[0x1]) Jul 9 23:56:11.751951 kernel: ACPI: LAPIC_NMI (acpi_id[0x12] high edge lint[0x1]) Jul 9 23:56:11.751956 kernel: ACPI: LAPIC_NMI (acpi_id[0x13] high edge lint[0x1]) Jul 9 23:56:11.751961 kernel: ACPI: LAPIC_NMI (acpi_id[0x14] high edge lint[0x1]) Jul 9 23:56:11.751967 kernel: ACPI: LAPIC_NMI (acpi_id[0x15] high edge lint[0x1]) Jul 9 23:56:11.751972 kernel: ACPI: LAPIC_NMI (acpi_id[0x16] high edge lint[0x1]) Jul 9 23:56:11.751977 kernel: ACPI: LAPIC_NMI (acpi_id[0x17] high edge lint[0x1]) Jul 9 23:56:11.751982 kernel: ACPI: LAPIC_NMI (acpi_id[0x18] high edge lint[0x1]) Jul 9 23:56:11.751987 kernel: ACPI: LAPIC_NMI (acpi_id[0x19] high edge lint[0x1]) Jul 9 23:56:11.751994 kernel: ACPI: LAPIC_NMI (acpi_id[0x1a] high edge lint[0x1]) Jul 9 23:56:11.751999 kernel: ACPI: LAPIC_NMI (acpi_id[0x1b] high edge lint[0x1]) Jul 9 23:56:11.752005 kernel: ACPI: LAPIC_NMI (acpi_id[0x1c] high edge lint[0x1]) Jul 9 23:56:11.752010 kernel: ACPI: LAPIC_NMI (acpi_id[0x1d] high edge lint[0x1]) Jul 9 23:56:11.752015 kernel: ACPI: LAPIC_NMI (acpi_id[0x1e] high edge lint[0x1]) Jul 9 23:56:11.752020 kernel: ACPI: LAPIC_NMI (acpi_id[0x1f] high edge lint[0x1]) Jul 9 23:56:11.752026 kernel: ACPI: LAPIC_NMI (acpi_id[0x20] high edge lint[0x1]) Jul 9 23:56:11.752031 kernel: ACPI: LAPIC_NMI (acpi_id[0x21] high edge lint[0x1]) Jul 9 23:56:11.752036 kernel: ACPI: LAPIC_NMI (acpi_id[0x22] high edge lint[0x1]) Jul 9 23:56:11.752042 kernel: ACPI: LAPIC_NMI (acpi_id[0x23] high edge lint[0x1]) Jul 9 23:56:11.752048 kernel: ACPI: LAPIC_NMI (acpi_id[0x24] high edge lint[0x1]) Jul 9 23:56:11.752053 kernel: ACPI: LAPIC_NMI (acpi_id[0x25] high edge lint[0x1]) Jul 9 23:56:11.752058 kernel: ACPI: LAPIC_NMI (acpi_id[0x26] high edge lint[0x1]) Jul 9 23:56:11.752063 kernel: ACPI: LAPIC_NMI (acpi_id[0x27] high edge lint[0x1]) Jul 9 23:56:11.752069 kernel: ACPI: LAPIC_NMI (acpi_id[0x28] high edge lint[0x1]) Jul 9 23:56:11.752074 kernel: ACPI: LAPIC_NMI (acpi_id[0x29] high edge lint[0x1]) Jul 9 23:56:11.752079 kernel: ACPI: LAPIC_NMI (acpi_id[0x2a] high edge lint[0x1]) Jul 9 23:56:11.752085 kernel: ACPI: LAPIC_NMI (acpi_id[0x2b] high edge lint[0x1]) Jul 9 23:56:11.752090 kernel: ACPI: LAPIC_NMI (acpi_id[0x2c] high edge lint[0x1]) Jul 9 23:56:11.752096 kernel: ACPI: LAPIC_NMI (acpi_id[0x2d] high edge lint[0x1]) Jul 9 23:56:11.752101 kernel: ACPI: LAPIC_NMI (acpi_id[0x2e] high edge lint[0x1]) Jul 9 23:56:11.752107 kernel: ACPI: LAPIC_NMI (acpi_id[0x2f] high edge lint[0x1]) Jul 9 23:56:11.752112 kernel: ACPI: LAPIC_NMI (acpi_id[0x30] high edge lint[0x1]) Jul 9 23:56:11.752117 kernel: ACPI: LAPIC_NMI (acpi_id[0x31] high edge lint[0x1]) Jul 9 23:56:11.752122 kernel: ACPI: LAPIC_NMI (acpi_id[0x32] high edge lint[0x1]) Jul 9 23:56:11.752128 kernel: ACPI: LAPIC_NMI (acpi_id[0x33] high edge lint[0x1]) Jul 9 23:56:11.752133 kernel: ACPI: LAPIC_NMI (acpi_id[0x34] high edge lint[0x1]) Jul 9 23:56:11.752138 kernel: ACPI: LAPIC_NMI (acpi_id[0x35] high edge lint[0x1]) Jul 9 23:56:11.752143 kernel: ACPI: LAPIC_NMI (acpi_id[0x36] high edge lint[0x1]) Jul 9 23:56:11.752150 kernel: ACPI: LAPIC_NMI (acpi_id[0x37] high edge lint[0x1]) Jul 9 23:56:11.752155 kernel: ACPI: LAPIC_NMI (acpi_id[0x38] high edge lint[0x1]) Jul 9 23:56:11.752160 kernel: ACPI: LAPIC_NMI (acpi_id[0x39] high edge lint[0x1]) Jul 9 23:56:11.752166 kernel: ACPI: LAPIC_NMI (acpi_id[0x3a] high edge lint[0x1]) Jul 9 23:56:11.752171 kernel: ACPI: LAPIC_NMI (acpi_id[0x3b] high edge lint[0x1]) Jul 9 23:56:11.752176 kernel: ACPI: LAPIC_NMI (acpi_id[0x3c] high edge lint[0x1]) Jul 9 23:56:11.752181 kernel: ACPI: LAPIC_NMI (acpi_id[0x3d] high edge lint[0x1]) Jul 9 23:56:11.752187 kernel: ACPI: LAPIC_NMI (acpi_id[0x3e] high edge lint[0x1]) Jul 9 23:56:11.752192 kernel: ACPI: LAPIC_NMI (acpi_id[0x3f] high edge lint[0x1]) Jul 9 23:56:11.752198 kernel: ACPI: LAPIC_NMI (acpi_id[0x40] high edge lint[0x1]) Jul 9 23:56:11.752204 kernel: ACPI: LAPIC_NMI (acpi_id[0x41] high edge lint[0x1]) Jul 9 23:56:11.752209 kernel: ACPI: LAPIC_NMI (acpi_id[0x42] high edge lint[0x1]) Jul 9 23:56:11.752214 kernel: ACPI: LAPIC_NMI (acpi_id[0x43] high edge lint[0x1]) Jul 9 23:56:11.752219 kernel: ACPI: LAPIC_NMI (acpi_id[0x44] high edge lint[0x1]) Jul 9 23:56:11.752225 kernel: ACPI: LAPIC_NMI (acpi_id[0x45] high edge lint[0x1]) Jul 9 23:56:11.752230 kernel: ACPI: LAPIC_NMI (acpi_id[0x46] high edge lint[0x1]) Jul 9 23:56:11.752235 kernel: ACPI: LAPIC_NMI (acpi_id[0x47] high edge lint[0x1]) Jul 9 23:56:11.752240 kernel: ACPI: LAPIC_NMI (acpi_id[0x48] high edge lint[0x1]) Jul 9 23:56:11.752246 kernel: ACPI: LAPIC_NMI (acpi_id[0x49] high edge lint[0x1]) Jul 9 23:56:11.752252 kernel: ACPI: LAPIC_NMI (acpi_id[0x4a] high edge lint[0x1]) Jul 9 23:56:11.752257 kernel: ACPI: LAPIC_NMI (acpi_id[0x4b] high edge lint[0x1]) Jul 9 23:56:11.752263 kernel: ACPI: LAPIC_NMI (acpi_id[0x4c] high edge lint[0x1]) Jul 9 23:56:11.752268 kernel: ACPI: LAPIC_NMI (acpi_id[0x4d] high edge lint[0x1]) Jul 9 23:56:11.752273 kernel: ACPI: LAPIC_NMI (acpi_id[0x4e] high edge lint[0x1]) Jul 9 23:56:11.752278 kernel: ACPI: LAPIC_NMI (acpi_id[0x4f] high edge lint[0x1]) Jul 9 23:56:11.752284 kernel: ACPI: LAPIC_NMI (acpi_id[0x50] high edge lint[0x1]) Jul 9 23:56:11.752289 kernel: ACPI: LAPIC_NMI (acpi_id[0x51] high edge lint[0x1]) Jul 9 23:56:11.752294 kernel: ACPI: LAPIC_NMI (acpi_id[0x52] high edge lint[0x1]) Jul 9 23:56:11.752299 kernel: ACPI: LAPIC_NMI (acpi_id[0x53] high edge lint[0x1]) Jul 9 23:56:11.752306 kernel: ACPI: LAPIC_NMI (acpi_id[0x54] high edge lint[0x1]) Jul 9 23:56:11.752311 kernel: ACPI: LAPIC_NMI (acpi_id[0x55] high edge lint[0x1]) Jul 9 23:56:11.752316 kernel: ACPI: LAPIC_NMI (acpi_id[0x56] high edge lint[0x1]) Jul 9 23:56:11.752321 kernel: ACPI: LAPIC_NMI (acpi_id[0x57] high edge lint[0x1]) Jul 9 23:56:11.752327 kernel: ACPI: LAPIC_NMI (acpi_id[0x58] high edge lint[0x1]) Jul 9 23:56:11.752332 kernel: ACPI: LAPIC_NMI (acpi_id[0x59] high edge lint[0x1]) Jul 9 23:56:11.752337 kernel: ACPI: LAPIC_NMI (acpi_id[0x5a] high edge lint[0x1]) Jul 9 23:56:11.752342 kernel: ACPI: LAPIC_NMI (acpi_id[0x5b] high edge lint[0x1]) Jul 9 23:56:11.752347 kernel: ACPI: LAPIC_NMI (acpi_id[0x5c] high edge lint[0x1]) Jul 9 23:56:11.752354 kernel: ACPI: LAPIC_NMI (acpi_id[0x5d] high edge lint[0x1]) Jul 9 23:56:11.752359 kernel: ACPI: LAPIC_NMI (acpi_id[0x5e] high edge lint[0x1]) Jul 9 23:56:11.752364 kernel: ACPI: LAPIC_NMI (acpi_id[0x5f] high edge lint[0x1]) Jul 9 23:56:11.754583 kernel: ACPI: LAPIC_NMI (acpi_id[0x60] high edge lint[0x1]) Jul 9 23:56:11.754596 kernel: ACPI: LAPIC_NMI (acpi_id[0x61] high edge lint[0x1]) Jul 9 23:56:11.754602 kernel: ACPI: LAPIC_NMI (acpi_id[0x62] high edge lint[0x1]) Jul 9 23:56:11.754615 kernel: ACPI: LAPIC_NMI (acpi_id[0x63] high edge lint[0x1]) Jul 9 23:56:11.754621 kernel: ACPI: LAPIC_NMI (acpi_id[0x64] high edge lint[0x1]) Jul 9 23:56:11.754627 kernel: ACPI: LAPIC_NMI (acpi_id[0x65] high edge lint[0x1]) Jul 9 23:56:11.754632 kernel: ACPI: LAPIC_NMI (acpi_id[0x66] high edge lint[0x1]) Jul 9 23:56:11.754640 kernel: ACPI: LAPIC_NMI (acpi_id[0x67] high edge lint[0x1]) Jul 9 23:56:11.754646 kernel: ACPI: LAPIC_NMI (acpi_id[0x68] high edge lint[0x1]) Jul 9 23:56:11.754653 kernel: ACPI: LAPIC_NMI (acpi_id[0x69] high edge lint[0x1]) Jul 9 23:56:11.754659 kernel: ACPI: LAPIC_NMI (acpi_id[0x6a] high edge lint[0x1]) Jul 9 23:56:11.754664 kernel: ACPI: LAPIC_NMI (acpi_id[0x6b] high edge lint[0x1]) Jul 9 23:56:11.754669 kernel: ACPI: LAPIC_NMI (acpi_id[0x6c] high edge lint[0x1]) Jul 9 23:56:11.754675 kernel: ACPI: LAPIC_NMI (acpi_id[0x6d] high edge lint[0x1]) Jul 9 23:56:11.754680 kernel: ACPI: LAPIC_NMI (acpi_id[0x6e] high edge lint[0x1]) Jul 9 23:56:11.754685 kernel: ACPI: LAPIC_NMI (acpi_id[0x6f] high edge lint[0x1]) Jul 9 23:56:11.754690 kernel: ACPI: LAPIC_NMI (acpi_id[0x70] high edge lint[0x1]) Jul 9 23:56:11.754697 kernel: ACPI: LAPIC_NMI (acpi_id[0x71] high edge lint[0x1]) Jul 9 23:56:11.754702 kernel: ACPI: LAPIC_NMI (acpi_id[0x72] high edge lint[0x1]) Jul 9 23:56:11.754707 kernel: ACPI: LAPIC_NMI (acpi_id[0x73] high edge lint[0x1]) Jul 9 23:56:11.754713 kernel: ACPI: LAPIC_NMI (acpi_id[0x74] high edge lint[0x1]) Jul 9 23:56:11.754718 kernel: ACPI: LAPIC_NMI (acpi_id[0x75] high edge lint[0x1]) Jul 9 23:56:11.754723 kernel: ACPI: LAPIC_NMI (acpi_id[0x76] high edge lint[0x1]) Jul 9 23:56:11.754728 kernel: ACPI: LAPIC_NMI (acpi_id[0x77] high edge lint[0x1]) Jul 9 23:56:11.754734 kernel: ACPI: LAPIC_NMI (acpi_id[0x78] high edge lint[0x1]) Jul 9 23:56:11.754739 kernel: ACPI: LAPIC_NMI (acpi_id[0x79] high edge lint[0x1]) Jul 9 23:56:11.754745 kernel: ACPI: LAPIC_NMI (acpi_id[0x7a] high edge lint[0x1]) Jul 9 23:56:11.754751 kernel: ACPI: LAPIC_NMI (acpi_id[0x7b] high edge lint[0x1]) Jul 9 23:56:11.754756 kernel: ACPI: LAPIC_NMI (acpi_id[0x7c] high edge lint[0x1]) Jul 9 23:56:11.754761 kernel: ACPI: LAPIC_NMI (acpi_id[0x7d] high edge lint[0x1]) Jul 9 23:56:11.754766 kernel: ACPI: LAPIC_NMI (acpi_id[0x7e] high edge lint[0x1]) Jul 9 23:56:11.754772 kernel: ACPI: LAPIC_NMI (acpi_id[0x7f] high edge lint[0x1]) Jul 9 23:56:11.754777 kernel: IOAPIC[0]: apic_id 1, version 17, address 0xfec00000, GSI 0-23 Jul 9 23:56:11.754782 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 high edge) Jul 9 23:56:11.754788 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Jul 9 23:56:11.754793 kernel: ACPI: HPET id: 0x8086af01 base: 0xfed00000 Jul 9 23:56:11.754799 kernel: TSC deadline timer available Jul 9 23:56:11.754805 kernel: smpboot: Allowing 128 CPUs, 126 hotplug CPUs Jul 9 23:56:11.754810 kernel: [mem 0x80000000-0xefffffff] available for PCI devices Jul 9 23:56:11.754816 kernel: Booting paravirtualized kernel on VMware hypervisor Jul 9 23:56:11.754821 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Jul 9 23:56:11.754827 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:128 nr_cpu_ids:128 nr_node_ids:1 Jul 9 23:56:11.754832 kernel: percpu: Embedded 58 pages/cpu s197096 r8192 d32280 u262144 Jul 9 23:56:11.754838 kernel: pcpu-alloc: s197096 r8192 d32280 u262144 alloc=1*2097152 Jul 9 23:56:11.754843 kernel: pcpu-alloc: [0] 000 001 002 003 004 005 006 007 Jul 9 23:56:11.754850 kernel: pcpu-alloc: [0] 008 009 010 011 012 013 014 015 Jul 9 23:56:11.754855 kernel: pcpu-alloc: [0] 016 017 018 019 020 021 022 023 Jul 9 23:56:11.754860 kernel: pcpu-alloc: [0] 024 025 026 027 028 029 030 031 Jul 9 23:56:11.754865 kernel: pcpu-alloc: [0] 032 033 034 035 036 037 038 039 Jul 9 23:56:11.754878 kernel: pcpu-alloc: [0] 040 041 042 043 044 045 046 047 Jul 9 23:56:11.754884 kernel: pcpu-alloc: [0] 048 049 050 051 052 053 054 055 Jul 9 23:56:11.754890 kernel: pcpu-alloc: [0] 056 057 058 059 060 061 062 063 Jul 9 23:56:11.754895 kernel: pcpu-alloc: [0] 064 065 066 067 068 069 070 071 Jul 9 23:56:11.754901 kernel: pcpu-alloc: [0] 072 073 074 075 076 077 078 079 Jul 9 23:56:11.754907 kernel: pcpu-alloc: [0] 080 081 082 083 084 085 086 087 Jul 9 23:56:11.754913 kernel: pcpu-alloc: [0] 088 089 090 091 092 093 094 095 Jul 9 23:56:11.754918 kernel: pcpu-alloc: [0] 096 097 098 099 100 101 102 103 Jul 9 23:56:11.754924 kernel: pcpu-alloc: [0] 104 105 106 107 108 109 110 111 Jul 9 23:56:11.754929 kernel: pcpu-alloc: [0] 112 113 114 115 116 117 118 119 Jul 9 23:56:11.754935 kernel: pcpu-alloc: [0] 120 121 122 123 124 125 126 127 Jul 9 23:56:11.754941 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=vmware flatcar.autologin verity.usrhash=c257b65f06e0ad68d969d5b3e057f031663dc29a4487d91a77595a40c4dc82d6 Jul 9 23:56:11.754947 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Jul 9 23:56:11.754954 kernel: random: crng init done Jul 9 23:56:11.754959 kernel: printk: log_buf_len individual max cpu contribution: 4096 bytes Jul 9 23:56:11.754965 kernel: printk: log_buf_len total cpu_extra contributions: 520192 bytes Jul 9 23:56:11.754970 kernel: printk: log_buf_len min size: 262144 bytes Jul 9 23:56:11.754976 kernel: printk: log_buf_len: 1048576 bytes Jul 9 23:56:11.754981 kernel: printk: early log buf free: 239648(91%) Jul 9 23:56:11.754987 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Jul 9 23:56:11.754993 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) Jul 9 23:56:11.754999 kernel: Fallback order for Node 0: 0 Jul 9 23:56:11.755005 kernel: Built 1 zonelists, mobility grouping on. Total pages: 515808 Jul 9 23:56:11.755011 kernel: Policy zone: DMA32 Jul 9 23:56:11.755017 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Jul 9 23:56:11.755023 kernel: Memory: 1934332K/2096628K available (14336K kernel code, 2295K rwdata, 22872K rodata, 43488K init, 1588K bss, 162036K reserved, 0K cma-reserved) Jul 9 23:56:11.755030 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=128, Nodes=1 Jul 9 23:56:11.755037 kernel: ftrace: allocating 37940 entries in 149 pages Jul 9 23:56:11.755042 kernel: ftrace: allocated 149 pages with 4 groups Jul 9 23:56:11.755048 kernel: Dynamic Preempt: voluntary Jul 9 23:56:11.755054 kernel: rcu: Preemptible hierarchical RCU implementation. Jul 9 23:56:11.755059 kernel: rcu: RCU event tracing is enabled. Jul 9 23:56:11.755065 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=128. Jul 9 23:56:11.755071 kernel: Trampoline variant of Tasks RCU enabled. Jul 9 23:56:11.755077 kernel: Rude variant of Tasks RCU enabled. Jul 9 23:56:11.755082 kernel: Tracing variant of Tasks RCU enabled. Jul 9 23:56:11.755088 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Jul 9 23:56:11.755095 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=128 Jul 9 23:56:11.755100 kernel: NR_IRQS: 33024, nr_irqs: 1448, preallocated irqs: 16 Jul 9 23:56:11.755106 kernel: rcu: srcu_init: Setting srcu_struct sizes to big. Jul 9 23:56:11.755111 kernel: Console: colour VGA+ 80x25 Jul 9 23:56:11.755117 kernel: printk: console [tty0] enabled Jul 9 23:56:11.755123 kernel: printk: console [ttyS0] enabled Jul 9 23:56:11.755128 kernel: ACPI: Core revision 20230628 Jul 9 23:56:11.755134 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 133484882848 ns Jul 9 23:56:11.755140 kernel: APIC: Switch to symmetric I/O mode setup Jul 9 23:56:11.755147 kernel: x2apic enabled Jul 9 23:56:11.755153 kernel: APIC: Switched APIC routing to: physical x2apic Jul 9 23:56:11.755158 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Jul 9 23:56:11.755164 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x311fd3cd494, max_idle_ns: 440795223879 ns Jul 9 23:56:11.755170 kernel: Calibrating delay loop (skipped) preset value.. 6816.00 BogoMIPS (lpj=3408000) Jul 9 23:56:11.755176 kernel: Disabled fast string operations Jul 9 23:56:11.755182 kernel: Last level iTLB entries: 4KB 64, 2MB 8, 4MB 8 Jul 9 23:56:11.755189 kernel: Last level dTLB entries: 4KB 64, 2MB 32, 4MB 32, 1GB 4 Jul 9 23:56:11.755195 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Jul 9 23:56:11.755202 kernel: Spectre V2 : Spectre BHI mitigation: SW BHB clearing on vm exit Jul 9 23:56:11.755208 kernel: Spectre V2 : Spectre BHI mitigation: SW BHB clearing on syscall Jul 9 23:56:11.755213 kernel: Spectre V2 : Mitigation: Enhanced / Automatic IBRS Jul 9 23:56:11.755219 kernel: Spectre V2 : Spectre v2 / PBRSB-eIBRS: Retire a single CALL on VMEXIT Jul 9 23:56:11.755225 kernel: RETBleed: Mitigation: Enhanced IBRS Jul 9 23:56:11.755231 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Jul 9 23:56:11.755237 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Jul 9 23:56:11.755243 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode Jul 9 23:56:11.755248 kernel: SRBDS: Unknown: Dependent on hypervisor status Jul 9 23:56:11.755255 kernel: GDS: Unknown: Dependent on hypervisor status Jul 9 23:56:11.755261 kernel: ITS: Mitigation: Aligned branch/return thunks Jul 9 23:56:11.755267 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Jul 9 23:56:11.755272 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Jul 9 23:56:11.755278 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Jul 9 23:56:11.755283 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Jul 9 23:56:11.755289 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Jul 9 23:56:11.755295 kernel: Freeing SMP alternatives memory: 32K Jul 9 23:56:11.755300 kernel: pid_max: default: 131072 minimum: 1024 Jul 9 23:56:11.755307 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Jul 9 23:56:11.755313 kernel: landlock: Up and running. Jul 9 23:56:11.755319 kernel: SELinux: Initializing. Jul 9 23:56:11.755324 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Jul 9 23:56:11.755330 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Jul 9 23:56:11.755336 kernel: smpboot: CPU0: Intel(R) Xeon(R) E-2278G CPU @ 3.40GHz (family: 0x6, model: 0x9e, stepping: 0xd) Jul 9 23:56:11.755342 kernel: RCU Tasks: Setting shift to 7 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=128. Jul 9 23:56:11.755347 kernel: RCU Tasks Rude: Setting shift to 7 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=128. Jul 9 23:56:11.755354 kernel: RCU Tasks Trace: Setting shift to 7 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=128. Jul 9 23:56:11.755360 kernel: Performance Events: Skylake events, core PMU driver. Jul 9 23:56:11.755366 kernel: core: CPUID marked event: 'cpu cycles' unavailable Jul 9 23:56:11.755378 kernel: core: CPUID marked event: 'instructions' unavailable Jul 9 23:56:11.755384 kernel: core: CPUID marked event: 'bus cycles' unavailable Jul 9 23:56:11.755389 kernel: core: CPUID marked event: 'cache references' unavailable Jul 9 23:56:11.755395 kernel: core: CPUID marked event: 'cache misses' unavailable Jul 9 23:56:11.755400 kernel: core: CPUID marked event: 'branch instructions' unavailable Jul 9 23:56:11.755409 kernel: core: CPUID marked event: 'branch misses' unavailable Jul 9 23:56:11.755419 kernel: ... version: 1 Jul 9 23:56:11.755425 kernel: ... bit width: 48 Jul 9 23:56:11.755431 kernel: ... generic registers: 4 Jul 9 23:56:11.755437 kernel: ... value mask: 0000ffffffffffff Jul 9 23:56:11.755442 kernel: ... max period: 000000007fffffff Jul 9 23:56:11.755448 kernel: ... fixed-purpose events: 0 Jul 9 23:56:11.755454 kernel: ... event mask: 000000000000000f Jul 9 23:56:11.755459 kernel: signal: max sigframe size: 1776 Jul 9 23:56:11.755465 kernel: rcu: Hierarchical SRCU implementation. Jul 9 23:56:11.755473 kernel: rcu: Max phase no-delay instances is 400. Jul 9 23:56:11.755478 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Jul 9 23:56:11.755484 kernel: smp: Bringing up secondary CPUs ... Jul 9 23:56:11.755490 kernel: smpboot: x86: Booting SMP configuration: Jul 9 23:56:11.755495 kernel: .... node #0, CPUs: #1 Jul 9 23:56:11.755501 kernel: Disabled fast string operations Jul 9 23:56:11.755507 kernel: smpboot: CPU 1 Converting physical 2 to logical package 1 Jul 9 23:56:11.755512 kernel: smpboot: CPU 1 Converting physical 0 to logical die 1 Jul 9 23:56:11.755518 kernel: smp: Brought up 1 node, 2 CPUs Jul 9 23:56:11.755523 kernel: smpboot: Max logical packages: 128 Jul 9 23:56:11.755530 kernel: smpboot: Total of 2 processors activated (13632.00 BogoMIPS) Jul 9 23:56:11.755536 kernel: devtmpfs: initialized Jul 9 23:56:11.755542 kernel: x86/mm: Memory block size: 128MB Jul 9 23:56:11.755549 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x7feff000-0x7fefffff] (4096 bytes) Jul 9 23:56:11.755555 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Jul 9 23:56:11.755561 kernel: futex hash table entries: 32768 (order: 9, 2097152 bytes, linear) Jul 9 23:56:11.755566 kernel: pinctrl core: initialized pinctrl subsystem Jul 9 23:56:11.755572 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Jul 9 23:56:11.755578 kernel: audit: initializing netlink subsys (disabled) Jul 9 23:56:11.755585 kernel: audit: type=2000 audit(1752105370.088:1): state=initialized audit_enabled=0 res=1 Jul 9 23:56:11.755591 kernel: thermal_sys: Registered thermal governor 'step_wise' Jul 9 23:56:11.755596 kernel: thermal_sys: Registered thermal governor 'user_space' Jul 9 23:56:11.755602 kernel: cpuidle: using governor menu Jul 9 23:56:11.755608 kernel: Simple Boot Flag at 0x36 set to 0x80 Jul 9 23:56:11.755613 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Jul 9 23:56:11.755619 kernel: dca service started, version 1.12.1 Jul 9 23:56:11.755625 kernel: PCI: MMCONFIG for domain 0000 [bus 00-7f] at [mem 0xf0000000-0xf7ffffff] (base 0xf0000000) Jul 9 23:56:11.755631 kernel: PCI: Using configuration type 1 for base access Jul 9 23:56:11.755637 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Jul 9 23:56:11.755643 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Jul 9 23:56:11.755649 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Jul 9 23:56:11.755655 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Jul 9 23:56:11.755660 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Jul 9 23:56:11.755666 kernel: ACPI: Added _OSI(Module Device) Jul 9 23:56:11.755672 kernel: ACPI: Added _OSI(Processor Device) Jul 9 23:56:11.755677 kernel: ACPI: Added _OSI(Processor Aggregator Device) Jul 9 23:56:11.755683 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Jul 9 23:56:11.755690 kernel: ACPI: [Firmware Bug]: BIOS _OSI(Linux) query ignored Jul 9 23:56:11.755696 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Jul 9 23:56:11.755701 kernel: ACPI: Interpreter enabled Jul 9 23:56:11.755707 kernel: ACPI: PM: (supports S0 S1 S5) Jul 9 23:56:11.755713 kernel: ACPI: Using IOAPIC for interrupt routing Jul 9 23:56:11.755722 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Jul 9 23:56:11.755733 kernel: PCI: Using E820 reservations for host bridge windows Jul 9 23:56:11.755739 kernel: ACPI: Enabled 4 GPEs in block 00 to 0F Jul 9 23:56:11.755750 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-7f]) Jul 9 23:56:11.756005 kernel: acpi PNP0A03:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Jul 9 23:56:11.756068 kernel: acpi PNP0A03:00: _OSC: platform does not support [AER LTR] Jul 9 23:56:11.756121 kernel: acpi PNP0A03:00: _OSC: OS now controls [PCIeHotplug PME PCIeCapability] Jul 9 23:56:11.756130 kernel: PCI host bridge to bus 0000:00 Jul 9 23:56:11.756183 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Jul 9 23:56:11.756232 kernel: pci_bus 0000:00: root bus resource [mem 0x000cc000-0x000dbfff window] Jul 9 23:56:11.756284 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] Jul 9 23:56:11.756331 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Jul 9 23:56:11.756750 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xfeff window] Jul 9 23:56:11.756806 kernel: pci_bus 0000:00: root bus resource [bus 00-7f] Jul 9 23:56:11.756869 kernel: pci 0000:00:00.0: [8086:7190] type 00 class 0x060000 Jul 9 23:56:11.756931 kernel: pci 0000:00:01.0: [8086:7191] type 01 class 0x060400 Jul 9 23:56:11.756991 kernel: pci 0000:00:07.0: [8086:7110] type 00 class 0x060100 Jul 9 23:56:11.757048 kernel: pci 0000:00:07.1: [8086:7111] type 00 class 0x01018a Jul 9 23:56:11.757101 kernel: pci 0000:00:07.1: reg 0x20: [io 0x1060-0x106f] Jul 9 23:56:11.757153 kernel: pci 0000:00:07.1: legacy IDE quirk: reg 0x10: [io 0x01f0-0x01f7] Jul 9 23:56:11.757204 kernel: pci 0000:00:07.1: legacy IDE quirk: reg 0x14: [io 0x03f6] Jul 9 23:56:11.757255 kernel: pci 0000:00:07.1: legacy IDE quirk: reg 0x18: [io 0x0170-0x0177] Jul 9 23:56:11.757307 kernel: pci 0000:00:07.1: legacy IDE quirk: reg 0x1c: [io 0x0376] Jul 9 23:56:11.757367 kernel: pci 0000:00:07.3: [8086:7113] type 00 class 0x068000 Jul 9 23:56:11.757485 kernel: pci 0000:00:07.3: quirk: [io 0x1000-0x103f] claimed by PIIX4 ACPI Jul 9 23:56:11.757538 kernel: pci 0000:00:07.3: quirk: [io 0x1040-0x104f] claimed by PIIX4 SMB Jul 9 23:56:11.757595 kernel: pci 0000:00:07.7: [15ad:0740] type 00 class 0x088000 Jul 9 23:56:11.757647 kernel: pci 0000:00:07.7: reg 0x10: [io 0x1080-0x10bf] Jul 9 23:56:11.757700 kernel: pci 0000:00:07.7: reg 0x14: [mem 0xfebfe000-0xfebfffff 64bit] Jul 9 23:56:11.757759 kernel: pci 0000:00:0f.0: [15ad:0405] type 00 class 0x030000 Jul 9 23:56:11.757812 kernel: pci 0000:00:0f.0: reg 0x10: [io 0x1070-0x107f] Jul 9 23:56:11.757863 kernel: pci 0000:00:0f.0: reg 0x14: [mem 0xe8000000-0xefffffff pref] Jul 9 23:56:11.757914 kernel: pci 0000:00:0f.0: reg 0x18: [mem 0xfe000000-0xfe7fffff] Jul 9 23:56:11.757965 kernel: pci 0000:00:0f.0: reg 0x30: [mem 0x00000000-0x00007fff pref] Jul 9 23:56:11.758015 kernel: pci 0000:00:0f.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Jul 9 23:56:11.758071 kernel: pci 0000:00:11.0: [15ad:0790] type 01 class 0x060401 Jul 9 23:56:11.758129 kernel: pci 0000:00:15.0: [15ad:07a0] type 01 class 0x060400 Jul 9 23:56:11.758181 kernel: pci 0000:00:15.0: PME# supported from D0 D3hot D3cold Jul 9 23:56:11.758240 kernel: pci 0000:00:15.1: [15ad:07a0] type 01 class 0x060400 Jul 9 23:56:11.758292 kernel: pci 0000:00:15.1: PME# supported from D0 D3hot D3cold Jul 9 23:56:11.758348 kernel: pci 0000:00:15.2: [15ad:07a0] type 01 class 0x060400 Jul 9 23:56:11.758417 kernel: pci 0000:00:15.2: PME# supported from D0 D3hot D3cold Jul 9 23:56:11.758478 kernel: pci 0000:00:15.3: [15ad:07a0] type 01 class 0x060400 Jul 9 23:56:11.758531 kernel: pci 0000:00:15.3: PME# supported from D0 D3hot D3cold Jul 9 23:56:11.758590 kernel: pci 0000:00:15.4: [15ad:07a0] type 01 class 0x060400 Jul 9 23:56:11.758642 kernel: pci 0000:00:15.4: PME# supported from D0 D3hot D3cold Jul 9 23:56:11.758697 kernel: pci 0000:00:15.5: [15ad:07a0] type 01 class 0x060400 Jul 9 23:56:11.758749 kernel: pci 0000:00:15.5: PME# supported from D0 D3hot D3cold Jul 9 23:56:11.758806 kernel: pci 0000:00:15.6: [15ad:07a0] type 01 class 0x060400 Jul 9 23:56:11.758861 kernel: pci 0000:00:15.6: PME# supported from D0 D3hot D3cold Jul 9 23:56:11.758935 kernel: pci 0000:00:15.7: [15ad:07a0] type 01 class 0x060400 Jul 9 23:56:11.758989 kernel: pci 0000:00:15.7: PME# supported from D0 D3hot D3cold Jul 9 23:56:11.759047 kernel: pci 0000:00:16.0: [15ad:07a0] type 01 class 0x060400 Jul 9 23:56:11.759116 kernel: pci 0000:00:16.0: PME# supported from D0 D3hot D3cold Jul 9 23:56:11.759174 kernel: pci 0000:00:16.1: [15ad:07a0] type 01 class 0x060400 Jul 9 23:56:11.759227 kernel: pci 0000:00:16.1: PME# supported from D0 D3hot D3cold Jul 9 23:56:11.759282 kernel: pci 0000:00:16.2: [15ad:07a0] type 01 class 0x060400 Jul 9 23:56:11.759334 kernel: pci 0000:00:16.2: PME# supported from D0 D3hot D3cold Jul 9 23:56:11.759515 kernel: pci 0000:00:16.3: [15ad:07a0] type 01 class 0x060400 Jul 9 23:56:11.759568 kernel: pci 0000:00:16.3: PME# supported from D0 D3hot D3cold Jul 9 23:56:11.759628 kernel: pci 0000:00:16.4: [15ad:07a0] type 01 class 0x060400 Jul 9 23:56:11.759680 kernel: pci 0000:00:16.4: PME# supported from D0 D3hot D3cold Jul 9 23:56:11.759737 kernel: pci 0000:00:16.5: [15ad:07a0] type 01 class 0x060400 Jul 9 23:56:11.759789 kernel: pci 0000:00:16.5: PME# supported from D0 D3hot D3cold Jul 9 23:56:11.759845 kernel: pci 0000:00:16.6: [15ad:07a0] type 01 class 0x060400 Jul 9 23:56:11.759897 kernel: pci 0000:00:16.6: PME# supported from D0 D3hot D3cold Jul 9 23:56:11.759954 kernel: pci 0000:00:16.7: [15ad:07a0] type 01 class 0x060400 Jul 9 23:56:11.760006 kernel: pci 0000:00:16.7: PME# supported from D0 D3hot D3cold Jul 9 23:56:11.760077 kernel: pci 0000:00:17.0: [15ad:07a0] type 01 class 0x060400 Jul 9 23:56:11.760131 kernel: pci 0000:00:17.0: PME# supported from D0 D3hot D3cold Jul 9 23:56:11.760187 kernel: pci 0000:00:17.1: [15ad:07a0] type 01 class 0x060400 Jul 9 23:56:11.760240 kernel: pci 0000:00:17.1: PME# supported from D0 D3hot D3cold Jul 9 23:56:11.760297 kernel: pci 0000:00:17.2: [15ad:07a0] type 01 class 0x060400 Jul 9 23:56:11.760353 kernel: pci 0000:00:17.2: PME# supported from D0 D3hot D3cold Jul 9 23:56:11.760473 kernel: pci 0000:00:17.3: [15ad:07a0] type 01 class 0x060400 Jul 9 23:56:11.760554 kernel: pci 0000:00:17.3: PME# supported from D0 D3hot D3cold Jul 9 23:56:11.760611 kernel: pci 0000:00:17.4: [15ad:07a0] type 01 class 0x060400 Jul 9 23:56:11.760663 kernel: pci 0000:00:17.4: PME# supported from D0 D3hot D3cold Jul 9 23:56:11.760721 kernel: pci 0000:00:17.5: [15ad:07a0] type 01 class 0x060400 Jul 9 23:56:11.760778 kernel: pci 0000:00:17.5: PME# supported from D0 D3hot D3cold Jul 9 23:56:11.760834 kernel: pci 0000:00:17.6: [15ad:07a0] type 01 class 0x060400 Jul 9 23:56:11.760886 kernel: pci 0000:00:17.6: PME# supported from D0 D3hot D3cold Jul 9 23:56:11.760942 kernel: pci 0000:00:17.7: [15ad:07a0] type 01 class 0x060400 Jul 9 23:56:11.760994 kernel: pci 0000:00:17.7: PME# supported from D0 D3hot D3cold Jul 9 23:56:11.761050 kernel: pci 0000:00:18.0: [15ad:07a0] type 01 class 0x060400 Jul 9 23:56:11.761106 kernel: pci 0000:00:18.0: PME# supported from D0 D3hot D3cold Jul 9 23:56:11.761161 kernel: pci 0000:00:18.1: [15ad:07a0] type 01 class 0x060400 Jul 9 23:56:11.761213 kernel: pci 0000:00:18.1: PME# supported from D0 D3hot D3cold Jul 9 23:56:11.761269 kernel: pci 0000:00:18.2: [15ad:07a0] type 01 class 0x060400 Jul 9 23:56:11.761321 kernel: pci 0000:00:18.2: PME# supported from D0 D3hot D3cold Jul 9 23:56:11.761429 kernel: pci 0000:00:18.3: [15ad:07a0] type 01 class 0x060400 Jul 9 23:56:11.761488 kernel: pci 0000:00:18.3: PME# supported from D0 D3hot D3cold Jul 9 23:56:11.761543 kernel: pci 0000:00:18.4: [15ad:07a0] type 01 class 0x060400 Jul 9 23:56:11.761595 kernel: pci 0000:00:18.4: PME# supported from D0 D3hot D3cold Jul 9 23:56:11.761654 kernel: pci 0000:00:18.5: [15ad:07a0] type 01 class 0x060400 Jul 9 23:56:11.761706 kernel: pci 0000:00:18.5: PME# supported from D0 D3hot D3cold Jul 9 23:56:11.761778 kernel: pci 0000:00:18.6: [15ad:07a0] type 01 class 0x060400 Jul 9 23:56:11.761848 kernel: pci 0000:00:18.6: PME# supported from D0 D3hot D3cold Jul 9 23:56:11.761904 kernel: pci 0000:00:18.7: [15ad:07a0] type 01 class 0x060400 Jul 9 23:56:11.761956 kernel: pci 0000:00:18.7: PME# supported from D0 D3hot D3cold Jul 9 23:56:11.762008 kernel: pci_bus 0000:01: extended config space not accessible Jul 9 23:56:11.762062 kernel: pci 0000:00:01.0: PCI bridge to [bus 01] Jul 9 23:56:11.762115 kernel: pci_bus 0000:02: extended config space not accessible Jul 9 23:56:11.762124 kernel: acpiphp: Slot [32] registered Jul 9 23:56:11.762132 kernel: acpiphp: Slot [33] registered Jul 9 23:56:11.762138 kernel: acpiphp: Slot [34] registered Jul 9 23:56:11.762144 kernel: acpiphp: Slot [35] registered Jul 9 23:56:11.762150 kernel: acpiphp: Slot [36] registered Jul 9 23:56:11.762156 kernel: acpiphp: Slot [37] registered Jul 9 23:56:11.762162 kernel: acpiphp: Slot [38] registered Jul 9 23:56:11.762167 kernel: acpiphp: Slot [39] registered Jul 9 23:56:11.762173 kernel: acpiphp: Slot [40] registered Jul 9 23:56:11.762179 kernel: acpiphp: Slot [41] registered Jul 9 23:56:11.762185 kernel: acpiphp: Slot [42] registered Jul 9 23:56:11.762191 kernel: acpiphp: Slot [43] registered Jul 9 23:56:11.762197 kernel: acpiphp: Slot [44] registered Jul 9 23:56:11.762202 kernel: acpiphp: Slot [45] registered Jul 9 23:56:11.762208 kernel: acpiphp: Slot [46] registered Jul 9 23:56:11.762214 kernel: acpiphp: Slot [47] registered Jul 9 23:56:11.762219 kernel: acpiphp: Slot [48] registered Jul 9 23:56:11.762225 kernel: acpiphp: Slot [49] registered Jul 9 23:56:11.762231 kernel: acpiphp: Slot [50] registered Jul 9 23:56:11.762236 kernel: acpiphp: Slot [51] registered Jul 9 23:56:11.762243 kernel: acpiphp: Slot [52] registered Jul 9 23:56:11.762249 kernel: acpiphp: Slot [53] registered Jul 9 23:56:11.762255 kernel: acpiphp: Slot [54] registered Jul 9 23:56:11.762260 kernel: acpiphp: Slot [55] registered Jul 9 23:56:11.762266 kernel: acpiphp: Slot [56] registered Jul 9 23:56:11.762272 kernel: acpiphp: Slot [57] registered Jul 9 23:56:11.762278 kernel: acpiphp: Slot [58] registered Jul 9 23:56:11.762284 kernel: acpiphp: Slot [59] registered Jul 9 23:56:11.762289 kernel: acpiphp: Slot [60] registered Jul 9 23:56:11.762296 kernel: acpiphp: Slot [61] registered Jul 9 23:56:11.762302 kernel: acpiphp: Slot [62] registered Jul 9 23:56:11.762307 kernel: acpiphp: Slot [63] registered Jul 9 23:56:11.762359 kernel: pci 0000:00:11.0: PCI bridge to [bus 02] (subtractive decode) Jul 9 23:56:11.762439 kernel: pci 0000:00:11.0: bridge window [io 0x2000-0x3fff] Jul 9 23:56:11.762521 kernel: pci 0000:00:11.0: bridge window [mem 0xfd600000-0xfdffffff] Jul 9 23:56:11.762572 kernel: pci 0000:00:11.0: bridge window [mem 0xe7b00000-0xe7ffffff 64bit pref] Jul 9 23:56:11.762623 kernel: pci 0000:00:11.0: bridge window [mem 0x000a0000-0x000bffff window] (subtractive decode) Jul 9 23:56:11.762677 kernel: pci 0000:00:11.0: bridge window [mem 0x000cc000-0x000dbfff window] (subtractive decode) Jul 9 23:56:11.762728 kernel: pci 0000:00:11.0: bridge window [mem 0xc0000000-0xfebfffff window] (subtractive decode) Jul 9 23:56:11.762779 kernel: pci 0000:00:11.0: bridge window [io 0x0000-0x0cf7 window] (subtractive decode) Jul 9 23:56:11.762829 kernel: pci 0000:00:11.0: bridge window [io 0x0d00-0xfeff window] (subtractive decode) Jul 9 23:56:11.762887 kernel: pci 0000:03:00.0: [15ad:07c0] type 00 class 0x010700 Jul 9 23:56:11.762940 kernel: pci 0000:03:00.0: reg 0x10: [io 0x4000-0x4007] Jul 9 23:56:11.762993 kernel: pci 0000:03:00.0: reg 0x14: [mem 0xfd5f8000-0xfd5fffff 64bit] Jul 9 23:56:11.763048 kernel: pci 0000:03:00.0: reg 0x30: [mem 0x00000000-0x0000ffff pref] Jul 9 23:56:11.763100 kernel: pci 0000:03:00.0: PME# supported from D0 D3hot D3cold Jul 9 23:56:11.763153 kernel: pci 0000:03:00.0: disabling ASPM on pre-1.1 PCIe device. You can enable it with 'pcie_aspm=force' Jul 9 23:56:11.763206 kernel: pci 0000:00:15.0: PCI bridge to [bus 03] Jul 9 23:56:11.763257 kernel: pci 0000:00:15.0: bridge window [io 0x4000-0x4fff] Jul 9 23:56:11.763308 kernel: pci 0000:00:15.0: bridge window [mem 0xfd500000-0xfd5fffff] Jul 9 23:56:11.763360 kernel: pci 0000:00:15.1: PCI bridge to [bus 04] Jul 9 23:56:11.763428 kernel: pci 0000:00:15.1: bridge window [io 0x8000-0x8fff] Jul 9 23:56:11.763484 kernel: pci 0000:00:15.1: bridge window [mem 0xfd100000-0xfd1fffff] Jul 9 23:56:11.763535 kernel: pci 0000:00:15.1: bridge window [mem 0xe7800000-0xe78fffff 64bit pref] Jul 9 23:56:11.763587 kernel: pci 0000:00:15.2: PCI bridge to [bus 05] Jul 9 23:56:11.763639 kernel: pci 0000:00:15.2: bridge window [io 0xc000-0xcfff] Jul 9 23:56:11.763690 kernel: pci 0000:00:15.2: bridge window [mem 0xfcd00000-0xfcdfffff] Jul 9 23:56:11.763740 kernel: pci 0000:00:15.2: bridge window [mem 0xe7400000-0xe74fffff 64bit pref] Jul 9 23:56:11.763793 kernel: pci 0000:00:15.3: PCI bridge to [bus 06] Jul 9 23:56:11.763843 kernel: pci 0000:00:15.3: bridge window [mem 0xfc900000-0xfc9fffff] Jul 9 23:56:11.763896 kernel: pci 0000:00:15.3: bridge window [mem 0xe7000000-0xe70fffff 64bit pref] Jul 9 23:56:11.763949 kernel: pci 0000:00:15.4: PCI bridge to [bus 07] Jul 9 23:56:11.763999 kernel: pci 0000:00:15.4: bridge window [mem 0xfc500000-0xfc5fffff] Jul 9 23:56:11.764051 kernel: pci 0000:00:15.4: bridge window [mem 0xe6c00000-0xe6cfffff 64bit pref] Jul 9 23:56:11.764105 kernel: pci 0000:00:15.5: PCI bridge to [bus 08] Jul 9 23:56:11.764157 kernel: pci 0000:00:15.5: bridge window [mem 0xfc100000-0xfc1fffff] Jul 9 23:56:11.764208 kernel: pci 0000:00:15.5: bridge window [mem 0xe6800000-0xe68fffff 64bit pref] Jul 9 23:56:11.764260 kernel: pci 0000:00:15.6: PCI bridge to [bus 09] Jul 9 23:56:11.764311 kernel: pci 0000:00:15.6: bridge window [mem 0xfbd00000-0xfbdfffff] Jul 9 23:56:11.764362 kernel: pci 0000:00:15.6: bridge window [mem 0xe6400000-0xe64fffff 64bit pref] Jul 9 23:56:11.764452 kernel: pci 0000:00:15.7: PCI bridge to [bus 0a] Jul 9 23:56:11.764505 kernel: pci 0000:00:15.7: bridge window [mem 0xfb900000-0xfb9fffff] Jul 9 23:56:11.764558 kernel: pci 0000:00:15.7: bridge window [mem 0xe6000000-0xe60fffff 64bit pref] Jul 9 23:56:11.764616 kernel: pci 0000:0b:00.0: [15ad:07b0] type 00 class 0x020000 Jul 9 23:56:11.764670 kernel: pci 0000:0b:00.0: reg 0x10: [mem 0xfd4fc000-0xfd4fcfff] Jul 9 23:56:11.764724 kernel: pci 0000:0b:00.0: reg 0x14: [mem 0xfd4fd000-0xfd4fdfff] Jul 9 23:56:11.764776 kernel: pci 0000:0b:00.0: reg 0x18: [mem 0xfd4fe000-0xfd4fffff] Jul 9 23:56:11.764828 kernel: pci 0000:0b:00.0: reg 0x1c: [io 0x5000-0x500f] Jul 9 23:56:11.764881 kernel: pci 0000:0b:00.0: reg 0x30: [mem 0x00000000-0x0000ffff pref] Jul 9 23:56:11.764936 kernel: pci 0000:0b:00.0: supports D1 D2 Jul 9 23:56:11.764990 kernel: pci 0000:0b:00.0: PME# supported from D0 D1 D2 D3hot D3cold Jul 9 23:56:11.765043 kernel: pci 0000:0b:00.0: disabling ASPM on pre-1.1 PCIe device. You can enable it with 'pcie_aspm=force' Jul 9 23:56:11.765095 kernel: pci 0000:00:16.0: PCI bridge to [bus 0b] Jul 9 23:56:11.765146 kernel: pci 0000:00:16.0: bridge window [io 0x5000-0x5fff] Jul 9 23:56:11.765198 kernel: pci 0000:00:16.0: bridge window [mem 0xfd400000-0xfd4fffff] Jul 9 23:56:11.765249 kernel: pci 0000:00:16.1: PCI bridge to [bus 0c] Jul 9 23:56:11.765301 kernel: pci 0000:00:16.1: bridge window [io 0x9000-0x9fff] Jul 9 23:56:11.765355 kernel: pci 0000:00:16.1: bridge window [mem 0xfd000000-0xfd0fffff] Jul 9 23:56:11.765650 kernel: pci 0000:00:16.1: bridge window [mem 0xe7700000-0xe77fffff 64bit pref] Jul 9 23:56:11.765707 kernel: pci 0000:00:16.2: PCI bridge to [bus 0d] Jul 9 23:56:11.765758 kernel: pci 0000:00:16.2: bridge window [io 0xd000-0xdfff] Jul 9 23:56:11.765809 kernel: pci 0000:00:16.2: bridge window [mem 0xfcc00000-0xfccfffff] Jul 9 23:56:11.765860 kernel: pci 0000:00:16.2: bridge window [mem 0xe7300000-0xe73fffff 64bit pref] Jul 9 23:56:11.765912 kernel: pci 0000:00:16.3: PCI bridge to [bus 0e] Jul 9 23:56:11.765963 kernel: pci 0000:00:16.3: bridge window [mem 0xfc800000-0xfc8fffff] Jul 9 23:56:11.766017 kernel: pci 0000:00:16.3: bridge window [mem 0xe6f00000-0xe6ffffff 64bit pref] Jul 9 23:56:11.766069 kernel: pci 0000:00:16.4: PCI bridge to [bus 0f] Jul 9 23:56:11.766119 kernel: pci 0000:00:16.4: bridge window [mem 0xfc400000-0xfc4fffff] Jul 9 23:56:11.766170 kernel: pci 0000:00:16.4: bridge window [mem 0xe6b00000-0xe6bfffff 64bit pref] Jul 9 23:56:11.766222 kernel: pci 0000:00:16.5: PCI bridge to [bus 10] Jul 9 23:56:11.766274 kernel: pci 0000:00:16.5: bridge window [mem 0xfc000000-0xfc0fffff] Jul 9 23:56:11.766324 kernel: pci 0000:00:16.5: bridge window [mem 0xe6700000-0xe67fffff 64bit pref] Jul 9 23:56:11.766387 kernel: pci 0000:00:16.6: PCI bridge to [bus 11] Jul 9 23:56:11.766465 kernel: pci 0000:00:16.6: bridge window [mem 0xfbc00000-0xfbcfffff] Jul 9 23:56:11.766519 kernel: pci 0000:00:16.6: bridge window [mem 0xe6300000-0xe63fffff 64bit pref] Jul 9 23:56:11.766586 kernel: pci 0000:00:16.7: PCI bridge to [bus 12] Jul 9 23:56:11.766637 kernel: pci 0000:00:16.7: bridge window [mem 0xfb800000-0xfb8fffff] Jul 9 23:56:11.766688 kernel: pci 0000:00:16.7: bridge window [mem 0xe5f00000-0xe5ffffff 64bit pref] Jul 9 23:56:11.766741 kernel: pci 0000:00:17.0: PCI bridge to [bus 13] Jul 9 23:56:11.766791 kernel: pci 0000:00:17.0: bridge window [io 0x6000-0x6fff] Jul 9 23:56:11.766842 kernel: pci 0000:00:17.0: bridge window [mem 0xfd300000-0xfd3fffff] Jul 9 23:56:11.766897 kernel: pci 0000:00:17.0: bridge window [mem 0xe7a00000-0xe7afffff 64bit pref] Jul 9 23:56:11.766949 kernel: pci 0000:00:17.1: PCI bridge to [bus 14] Jul 9 23:56:11.767000 kernel: pci 0000:00:17.1: bridge window [io 0xa000-0xafff] Jul 9 23:56:11.767051 kernel: pci 0000:00:17.1: bridge window [mem 0xfcf00000-0xfcffffff] Jul 9 23:56:11.767102 kernel: pci 0000:00:17.1: bridge window [mem 0xe7600000-0xe76fffff 64bit pref] Jul 9 23:56:11.767154 kernel: pci 0000:00:17.2: PCI bridge to [bus 15] Jul 9 23:56:11.767205 kernel: pci 0000:00:17.2: bridge window [io 0xe000-0xefff] Jul 9 23:56:11.767258 kernel: pci 0000:00:17.2: bridge window [mem 0xfcb00000-0xfcbfffff] Jul 9 23:56:11.767310 kernel: pci 0000:00:17.2: bridge window [mem 0xe7200000-0xe72fffff 64bit pref] Jul 9 23:56:11.767362 kernel: pci 0000:00:17.3: PCI bridge to [bus 16] Jul 9 23:56:11.767421 kernel: pci 0000:00:17.3: bridge window [mem 0xfc700000-0xfc7fffff] Jul 9 23:56:11.767472 kernel: pci 0000:00:17.3: bridge window [mem 0xe6e00000-0xe6efffff 64bit pref] Jul 9 23:56:11.767523 kernel: pci 0000:00:17.4: PCI bridge to [bus 17] Jul 9 23:56:11.767575 kernel: pci 0000:00:17.4: bridge window [mem 0xfc300000-0xfc3fffff] Jul 9 23:56:11.767627 kernel: pci 0000:00:17.4: bridge window [mem 0xe6a00000-0xe6afffff 64bit pref] Jul 9 23:56:11.767681 kernel: pci 0000:00:17.5: PCI bridge to [bus 18] Jul 9 23:56:11.767733 kernel: pci 0000:00:17.5: bridge window [mem 0xfbf00000-0xfbffffff] Jul 9 23:56:11.767784 kernel: pci 0000:00:17.5: bridge window [mem 0xe6600000-0xe66fffff 64bit pref] Jul 9 23:56:11.767836 kernel: pci 0000:00:17.6: PCI bridge to [bus 19] Jul 9 23:56:11.767888 kernel: pci 0000:00:17.6: bridge window [mem 0xfbb00000-0xfbbfffff] Jul 9 23:56:11.767938 kernel: pci 0000:00:17.6: bridge window [mem 0xe6200000-0xe62fffff 64bit pref] Jul 9 23:56:11.767991 kernel: pci 0000:00:17.7: PCI bridge to [bus 1a] Jul 9 23:56:11.768042 kernel: pci 0000:00:17.7: bridge window [mem 0xfb700000-0xfb7fffff] Jul 9 23:56:11.768095 kernel: pci 0000:00:17.7: bridge window [mem 0xe5e00000-0xe5efffff 64bit pref] Jul 9 23:56:11.768147 kernel: pci 0000:00:18.0: PCI bridge to [bus 1b] Jul 9 23:56:11.768199 kernel: pci 0000:00:18.0: bridge window [io 0x7000-0x7fff] Jul 9 23:56:11.768250 kernel: pci 0000:00:18.0: bridge window [mem 0xfd200000-0xfd2fffff] Jul 9 23:56:11.768301 kernel: pci 0000:00:18.0: bridge window [mem 0xe7900000-0xe79fffff 64bit pref] Jul 9 23:56:11.768354 kernel: pci 0000:00:18.1: PCI bridge to [bus 1c] Jul 9 23:56:11.768440 kernel: pci 0000:00:18.1: bridge window [io 0xb000-0xbfff] Jul 9 23:56:11.768522 kernel: pci 0000:00:18.1: bridge window [mem 0xfce00000-0xfcefffff] Jul 9 23:56:11.768577 kernel: pci 0000:00:18.1: bridge window [mem 0xe7500000-0xe75fffff 64bit pref] Jul 9 23:56:11.768629 kernel: pci 0000:00:18.2: PCI bridge to [bus 1d] Jul 9 23:56:11.768681 kernel: pci 0000:00:18.2: bridge window [mem 0xfca00000-0xfcafffff] Jul 9 23:56:11.768731 kernel: pci 0000:00:18.2: bridge window [mem 0xe7100000-0xe71fffff 64bit pref] Jul 9 23:56:11.768783 kernel: pci 0000:00:18.3: PCI bridge to [bus 1e] Jul 9 23:56:11.768835 kernel: pci 0000:00:18.3: bridge window [mem 0xfc600000-0xfc6fffff] Jul 9 23:56:11.768885 kernel: pci 0000:00:18.3: bridge window [mem 0xe6d00000-0xe6dfffff 64bit pref] Jul 9 23:56:11.768940 kernel: pci 0000:00:18.4: PCI bridge to [bus 1f] Jul 9 23:56:11.768991 kernel: pci 0000:00:18.4: bridge window [mem 0xfc200000-0xfc2fffff] Jul 9 23:56:11.769041 kernel: pci 0000:00:18.4: bridge window [mem 0xe6900000-0xe69fffff 64bit pref] Jul 9 23:56:11.769094 kernel: pci 0000:00:18.5: PCI bridge to [bus 20] Jul 9 23:56:11.769162 kernel: pci 0000:00:18.5: bridge window [mem 0xfbe00000-0xfbefffff] Jul 9 23:56:11.769227 kernel: pci 0000:00:18.5: bridge window [mem 0xe6500000-0xe65fffff 64bit pref] Jul 9 23:56:11.769279 kernel: pci 0000:00:18.6: PCI bridge to [bus 21] Jul 9 23:56:11.769330 kernel: pci 0000:00:18.6: bridge window [mem 0xfba00000-0xfbafffff] Jul 9 23:56:11.769415 kernel: pci 0000:00:18.6: bridge window [mem 0xe6100000-0xe61fffff 64bit pref] Jul 9 23:56:11.769474 kernel: pci 0000:00:18.7: PCI bridge to [bus 22] Jul 9 23:56:11.769524 kernel: pci 0000:00:18.7: bridge window [mem 0xfb600000-0xfb6fffff] Jul 9 23:56:11.769592 kernel: pci 0000:00:18.7: bridge window [mem 0xe5d00000-0xe5dfffff 64bit pref] Jul 9 23:56:11.769601 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 9 Jul 9 23:56:11.769608 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 0 Jul 9 23:56:11.769614 kernel: ACPI: PCI: Interrupt link LNKB disabled Jul 9 23:56:11.769620 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Jul 9 23:56:11.769626 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 10 Jul 9 23:56:11.769633 kernel: iommu: Default domain type: Translated Jul 9 23:56:11.769639 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Jul 9 23:56:11.769659 kernel: PCI: Using ACPI for IRQ routing Jul 9 23:56:11.769665 kernel: PCI: pci_cache_line_size set to 64 bytes Jul 9 23:56:11.769671 kernel: e820: reserve RAM buffer [mem 0x0009ec00-0x0009ffff] Jul 9 23:56:11.769677 kernel: e820: reserve RAM buffer [mem 0x7fee0000-0x7fffffff] Jul 9 23:56:11.769727 kernel: pci 0000:00:0f.0: vgaarb: setting as boot VGA device Jul 9 23:56:11.769777 kernel: pci 0000:00:0f.0: vgaarb: bridge control possible Jul 9 23:56:11.769827 kernel: pci 0000:00:0f.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Jul 9 23:56:11.769838 kernel: vgaarb: loaded Jul 9 23:56:11.769844 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 Jul 9 23:56:11.769850 kernel: hpet0: 16 comparators, 64-bit 14.318180 MHz counter Jul 9 23:56:11.769856 kernel: clocksource: Switched to clocksource tsc-early Jul 9 23:56:11.769862 kernel: VFS: Disk quotas dquot_6.6.0 Jul 9 23:56:11.769868 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Jul 9 23:56:11.769874 kernel: pnp: PnP ACPI init Jul 9 23:56:11.769930 kernel: system 00:00: [io 0x1000-0x103f] has been reserved Jul 9 23:56:11.769981 kernel: system 00:00: [io 0x1040-0x104f] has been reserved Jul 9 23:56:11.770028 kernel: system 00:00: [io 0x0cf0-0x0cf1] has been reserved Jul 9 23:56:11.770078 kernel: system 00:04: [mem 0xfed00000-0xfed003ff] has been reserved Jul 9 23:56:11.770129 kernel: pnp 00:06: [dma 2] Jul 9 23:56:11.770181 kernel: system 00:07: [io 0xfce0-0xfcff] has been reserved Jul 9 23:56:11.770230 kernel: system 00:07: [mem 0xf0000000-0xf7ffffff] has been reserved Jul 9 23:56:11.770277 kernel: system 00:07: [mem 0xfe800000-0xfe9fffff] has been reserved Jul 9 23:56:11.770287 kernel: pnp: PnP ACPI: found 8 devices Jul 9 23:56:11.770293 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Jul 9 23:56:11.770299 kernel: NET: Registered PF_INET protocol family Jul 9 23:56:11.770305 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) Jul 9 23:56:11.770311 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) Jul 9 23:56:11.770316 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Jul 9 23:56:11.770322 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) Jul 9 23:56:11.770328 kernel: TCP bind hash table entries: 16384 (order: 7, 524288 bytes, linear) Jul 9 23:56:11.770336 kernel: TCP: Hash tables configured (established 16384 bind 16384) Jul 9 23:56:11.770342 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) Jul 9 23:56:11.770347 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) Jul 9 23:56:11.770353 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Jul 9 23:56:11.770359 kernel: NET: Registered PF_XDP protocol family Jul 9 23:56:11.770523 kernel: pci 0000:00:15.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 03] add_size 200000 add_align 100000 Jul 9 23:56:11.770579 kernel: pci 0000:00:15.3: bridge window [io 0x1000-0x0fff] to [bus 06] add_size 1000 Jul 9 23:56:11.770633 kernel: pci 0000:00:15.4: bridge window [io 0x1000-0x0fff] to [bus 07] add_size 1000 Jul 9 23:56:11.770687 kernel: pci 0000:00:15.5: bridge window [io 0x1000-0x0fff] to [bus 08] add_size 1000 Jul 9 23:56:11.770741 kernel: pci 0000:00:15.6: bridge window [io 0x1000-0x0fff] to [bus 09] add_size 1000 Jul 9 23:56:11.770794 kernel: pci 0000:00:15.7: bridge window [io 0x1000-0x0fff] to [bus 0a] add_size 1000 Jul 9 23:56:11.770847 kernel: pci 0000:00:16.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 0b] add_size 200000 add_align 100000 Jul 9 23:56:11.770899 kernel: pci 0000:00:16.3: bridge window [io 0x1000-0x0fff] to [bus 0e] add_size 1000 Jul 9 23:56:11.770952 kernel: pci 0000:00:16.4: bridge window [io 0x1000-0x0fff] to [bus 0f] add_size 1000 Jul 9 23:56:11.771008 kernel: pci 0000:00:16.5: bridge window [io 0x1000-0x0fff] to [bus 10] add_size 1000 Jul 9 23:56:11.771060 kernel: pci 0000:00:16.6: bridge window [io 0x1000-0x0fff] to [bus 11] add_size 1000 Jul 9 23:56:11.771112 kernel: pci 0000:00:16.7: bridge window [io 0x1000-0x0fff] to [bus 12] add_size 1000 Jul 9 23:56:11.771164 kernel: pci 0000:00:17.3: bridge window [io 0x1000-0x0fff] to [bus 16] add_size 1000 Jul 9 23:56:11.771215 kernel: pci 0000:00:17.4: bridge window [io 0x1000-0x0fff] to [bus 17] add_size 1000 Jul 9 23:56:11.771266 kernel: pci 0000:00:17.5: bridge window [io 0x1000-0x0fff] to [bus 18] add_size 1000 Jul 9 23:56:11.771320 kernel: pci 0000:00:17.6: bridge window [io 0x1000-0x0fff] to [bus 19] add_size 1000 Jul 9 23:56:11.771378 kernel: pci 0000:00:17.7: bridge window [io 0x1000-0x0fff] to [bus 1a] add_size 1000 Jul 9 23:56:11.771462 kernel: pci 0000:00:18.2: bridge window [io 0x1000-0x0fff] to [bus 1d] add_size 1000 Jul 9 23:56:11.771514 kernel: pci 0000:00:18.3: bridge window [io 0x1000-0x0fff] to [bus 1e] add_size 1000 Jul 9 23:56:11.771564 kernel: pci 0000:00:18.4: bridge window [io 0x1000-0x0fff] to [bus 1f] add_size 1000 Jul 9 23:56:11.771619 kernel: pci 0000:00:18.5: bridge window [io 0x1000-0x0fff] to [bus 20] add_size 1000 Jul 9 23:56:11.771671 kernel: pci 0000:00:18.6: bridge window [io 0x1000-0x0fff] to [bus 21] add_size 1000 Jul 9 23:56:11.771722 kernel: pci 0000:00:18.7: bridge window [io 0x1000-0x0fff] to [bus 22] add_size 1000 Jul 9 23:56:11.771774 kernel: pci 0000:00:15.0: BAR 15: assigned [mem 0xc0000000-0xc01fffff 64bit pref] Jul 9 23:56:11.771826 kernel: pci 0000:00:16.0: BAR 15: assigned [mem 0xc0200000-0xc03fffff 64bit pref] Jul 9 23:56:11.771876 kernel: pci 0000:00:15.3: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.771928 kernel: pci 0000:00:15.3: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.771982 kernel: pci 0000:00:15.4: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.772033 kernel: pci 0000:00:15.4: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.772085 kernel: pci 0000:00:15.5: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.772135 kernel: pci 0000:00:15.5: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.772186 kernel: pci 0000:00:15.6: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.772238 kernel: pci 0000:00:15.6: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.772288 kernel: pci 0000:00:15.7: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.772339 kernel: pci 0000:00:15.7: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.772420 kernel: pci 0000:00:16.3: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.772501 kernel: pci 0000:00:16.3: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.772553 kernel: pci 0000:00:16.4: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.772604 kernel: pci 0000:00:16.4: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.772655 kernel: pci 0000:00:16.5: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.772706 kernel: pci 0000:00:16.5: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.772757 kernel: pci 0000:00:16.6: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.772808 kernel: pci 0000:00:16.6: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.772862 kernel: pci 0000:00:16.7: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.772913 kernel: pci 0000:00:16.7: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.772965 kernel: pci 0000:00:17.3: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.773015 kernel: pci 0000:00:17.3: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.773066 kernel: pci 0000:00:17.4: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.773117 kernel: pci 0000:00:17.4: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.773168 kernel: pci 0000:00:17.5: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.773219 kernel: pci 0000:00:17.5: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.773274 kernel: pci 0000:00:17.6: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.773326 kernel: pci 0000:00:17.6: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.773531 kernel: pci 0000:00:17.7: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.773588 kernel: pci 0000:00:17.7: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.773641 kernel: pci 0000:00:18.2: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.773704 kernel: pci 0000:00:18.2: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.773773 kernel: pci 0000:00:18.3: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.773839 kernel: pci 0000:00:18.3: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.773894 kernel: pci 0000:00:18.4: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.773946 kernel: pci 0000:00:18.4: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.773998 kernel: pci 0000:00:18.5: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.774112 kernel: pci 0000:00:18.5: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.774169 kernel: pci 0000:00:18.6: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.774223 kernel: pci 0000:00:18.6: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.774275 kernel: pci 0000:00:18.7: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.774326 kernel: pci 0000:00:18.7: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.774389 kernel: pci 0000:00:18.7: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.774443 kernel: pci 0000:00:18.7: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.774495 kernel: pci 0000:00:18.6: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.774547 kernel: pci 0000:00:18.6: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.774599 kernel: pci 0000:00:18.5: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.774650 kernel: pci 0000:00:18.5: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.774702 kernel: pci 0000:00:18.4: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.774763 kernel: pci 0000:00:18.4: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.774892 kernel: pci 0000:00:18.3: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.774949 kernel: pci 0000:00:18.3: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.775004 kernel: pci 0000:00:18.2: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.775057 kernel: pci 0000:00:18.2: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.775108 kernel: pci 0000:00:17.7: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.775159 kernel: pci 0000:00:17.7: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.775211 kernel: pci 0000:00:17.6: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.775262 kernel: pci 0000:00:17.6: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.775313 kernel: pci 0000:00:17.5: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.775365 kernel: pci 0000:00:17.5: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.775433 kernel: pci 0000:00:17.4: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.775489 kernel: pci 0000:00:17.4: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.775543 kernel: pci 0000:00:17.3: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.775594 kernel: pci 0000:00:17.3: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.775664 kernel: pci 0000:00:16.7: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.775723 kernel: pci 0000:00:16.7: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.775776 kernel: pci 0000:00:16.6: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.775827 kernel: pci 0000:00:16.6: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.775879 kernel: pci 0000:00:16.5: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.775930 kernel: pci 0000:00:16.5: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.775981 kernel: pci 0000:00:16.4: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.776036 kernel: pci 0000:00:16.4: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.776087 kernel: pci 0000:00:16.3: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.776140 kernel: pci 0000:00:16.3: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.776191 kernel: pci 0000:00:15.7: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.776243 kernel: pci 0000:00:15.7: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.776294 kernel: pci 0000:00:15.6: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.776346 kernel: pci 0000:00:15.6: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.776515 kernel: pci 0000:00:15.5: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.776567 kernel: pci 0000:00:15.5: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.776622 kernel: pci 0000:00:15.4: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.776672 kernel: pci 0000:00:15.4: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.776725 kernel: pci 0000:00:15.3: BAR 13: no space for [io size 0x1000] Jul 9 23:56:11.776776 kernel: pci 0000:00:15.3: BAR 13: failed to assign [io size 0x1000] Jul 9 23:56:11.776829 kernel: pci 0000:00:01.0: PCI bridge to [bus 01] Jul 9 23:56:11.776882 kernel: pci 0000:00:11.0: PCI bridge to [bus 02] Jul 9 23:56:11.776933 kernel: pci 0000:00:11.0: bridge window [io 0x2000-0x3fff] Jul 9 23:56:11.776984 kernel: pci 0000:00:11.0: bridge window [mem 0xfd600000-0xfdffffff] Jul 9 23:56:11.777035 kernel: pci 0000:00:11.0: bridge window [mem 0xe7b00000-0xe7ffffff 64bit pref] Jul 9 23:56:11.777094 kernel: pci 0000:03:00.0: BAR 6: assigned [mem 0xfd500000-0xfd50ffff pref] Jul 9 23:56:11.777146 kernel: pci 0000:00:15.0: PCI bridge to [bus 03] Jul 9 23:56:11.777197 kernel: pci 0000:00:15.0: bridge window [io 0x4000-0x4fff] Jul 9 23:56:11.777250 kernel: pci 0000:00:15.0: bridge window [mem 0xfd500000-0xfd5fffff] Jul 9 23:56:11.777301 kernel: pci 0000:00:15.0: bridge window [mem 0xc0000000-0xc01fffff 64bit pref] Jul 9 23:56:11.777353 kernel: pci 0000:00:15.1: PCI bridge to [bus 04] Jul 9 23:56:11.777432 kernel: pci 0000:00:15.1: bridge window [io 0x8000-0x8fff] Jul 9 23:56:11.777500 kernel: pci 0000:00:15.1: bridge window [mem 0xfd100000-0xfd1fffff] Jul 9 23:56:11.777550 kernel: pci 0000:00:15.1: bridge window [mem 0xe7800000-0xe78fffff 64bit pref] Jul 9 23:56:11.777606 kernel: pci 0000:00:15.2: PCI bridge to [bus 05] Jul 9 23:56:11.777657 kernel: pci 0000:00:15.2: bridge window [io 0xc000-0xcfff] Jul 9 23:56:11.777708 kernel: pci 0000:00:15.2: bridge window [mem 0xfcd00000-0xfcdfffff] Jul 9 23:56:11.777759 kernel: pci 0000:00:15.2: bridge window [mem 0xe7400000-0xe74fffff 64bit pref] Jul 9 23:56:11.777810 kernel: pci 0000:00:15.3: PCI bridge to [bus 06] Jul 9 23:56:11.777862 kernel: pci 0000:00:15.3: bridge window [mem 0xfc900000-0xfc9fffff] Jul 9 23:56:11.777913 kernel: pci 0000:00:15.3: bridge window [mem 0xe7000000-0xe70fffff 64bit pref] Jul 9 23:56:11.777964 kernel: pci 0000:00:15.4: PCI bridge to [bus 07] Jul 9 23:56:11.778015 kernel: pci 0000:00:15.4: bridge window [mem 0xfc500000-0xfc5fffff] Jul 9 23:56:11.778069 kernel: pci 0000:00:15.4: bridge window [mem 0xe6c00000-0xe6cfffff 64bit pref] Jul 9 23:56:11.778123 kernel: pci 0000:00:15.5: PCI bridge to [bus 08] Jul 9 23:56:11.778175 kernel: pci 0000:00:15.5: bridge window [mem 0xfc100000-0xfc1fffff] Jul 9 23:56:11.778226 kernel: pci 0000:00:15.5: bridge window [mem 0xe6800000-0xe68fffff 64bit pref] Jul 9 23:56:11.778338 kernel: pci 0000:00:15.6: PCI bridge to [bus 09] Jul 9 23:56:11.780265 kernel: pci 0000:00:15.6: bridge window [mem 0xfbd00000-0xfbdfffff] Jul 9 23:56:11.780328 kernel: pci 0000:00:15.6: bridge window [mem 0xe6400000-0xe64fffff 64bit pref] Jul 9 23:56:11.780391 kernel: pci 0000:00:15.7: PCI bridge to [bus 0a] Jul 9 23:56:11.780484 kernel: pci 0000:00:15.7: bridge window [mem 0xfb900000-0xfb9fffff] Jul 9 23:56:11.780537 kernel: pci 0000:00:15.7: bridge window [mem 0xe6000000-0xe60fffff 64bit pref] Jul 9 23:56:11.780593 kernel: pci 0000:0b:00.0: BAR 6: assigned [mem 0xfd400000-0xfd40ffff pref] Jul 9 23:56:11.780646 kernel: pci 0000:00:16.0: PCI bridge to [bus 0b] Jul 9 23:56:11.780698 kernel: pci 0000:00:16.0: bridge window [io 0x5000-0x5fff] Jul 9 23:56:11.780750 kernel: pci 0000:00:16.0: bridge window [mem 0xfd400000-0xfd4fffff] Jul 9 23:56:11.780802 kernel: pci 0000:00:16.0: bridge window [mem 0xc0200000-0xc03fffff 64bit pref] Jul 9 23:56:11.780859 kernel: pci 0000:00:16.1: PCI bridge to [bus 0c] Jul 9 23:56:11.780913 kernel: pci 0000:00:16.1: bridge window [io 0x9000-0x9fff] Jul 9 23:56:11.780965 kernel: pci 0000:00:16.1: bridge window [mem 0xfd000000-0xfd0fffff] Jul 9 23:56:11.781018 kernel: pci 0000:00:16.1: bridge window [mem 0xe7700000-0xe77fffff 64bit pref] Jul 9 23:56:11.781072 kernel: pci 0000:00:16.2: PCI bridge to [bus 0d] Jul 9 23:56:11.781124 kernel: pci 0000:00:16.2: bridge window [io 0xd000-0xdfff] Jul 9 23:56:11.781177 kernel: pci 0000:00:16.2: bridge window [mem 0xfcc00000-0xfccfffff] Jul 9 23:56:11.781229 kernel: pci 0000:00:16.2: bridge window [mem 0xe7300000-0xe73fffff 64bit pref] Jul 9 23:56:11.781281 kernel: pci 0000:00:16.3: PCI bridge to [bus 0e] Jul 9 23:56:11.781332 kernel: pci 0000:00:16.3: bridge window [mem 0xfc800000-0xfc8fffff] Jul 9 23:56:11.781399 kernel: pci 0000:00:16.3: bridge window [mem 0xe6f00000-0xe6ffffff 64bit pref] Jul 9 23:56:11.781453 kernel: pci 0000:00:16.4: PCI bridge to [bus 0f] Jul 9 23:56:11.781506 kernel: pci 0000:00:16.4: bridge window [mem 0xfc400000-0xfc4fffff] Jul 9 23:56:11.781558 kernel: pci 0000:00:16.4: bridge window [mem 0xe6b00000-0xe6bfffff 64bit pref] Jul 9 23:56:11.781611 kernel: pci 0000:00:16.5: PCI bridge to [bus 10] Jul 9 23:56:11.781663 kernel: pci 0000:00:16.5: bridge window [mem 0xfc000000-0xfc0fffff] Jul 9 23:56:11.781716 kernel: pci 0000:00:16.5: bridge window [mem 0xe6700000-0xe67fffff 64bit pref] Jul 9 23:56:11.781769 kernel: pci 0000:00:16.6: PCI bridge to [bus 11] Jul 9 23:56:11.781821 kernel: pci 0000:00:16.6: bridge window [mem 0xfbc00000-0xfbcfffff] Jul 9 23:56:11.781876 kernel: pci 0000:00:16.6: bridge window [mem 0xe6300000-0xe63fffff 64bit pref] Jul 9 23:56:11.781929 kernel: pci 0000:00:16.7: PCI bridge to [bus 12] Jul 9 23:56:11.781981 kernel: pci 0000:00:16.7: bridge window [mem 0xfb800000-0xfb8fffff] Jul 9 23:56:11.782033 kernel: pci 0000:00:16.7: bridge window [mem 0xe5f00000-0xe5ffffff 64bit pref] Jul 9 23:56:11.782086 kernel: pci 0000:00:17.0: PCI bridge to [bus 13] Jul 9 23:56:11.782139 kernel: pci 0000:00:17.0: bridge window [io 0x6000-0x6fff] Jul 9 23:56:11.782191 kernel: pci 0000:00:17.0: bridge window [mem 0xfd300000-0xfd3fffff] Jul 9 23:56:11.782243 kernel: pci 0000:00:17.0: bridge window [mem 0xe7a00000-0xe7afffff 64bit pref] Jul 9 23:56:11.782296 kernel: pci 0000:00:17.1: PCI bridge to [bus 14] Jul 9 23:56:11.782350 kernel: pci 0000:00:17.1: bridge window [io 0xa000-0xafff] Jul 9 23:56:11.782462 kernel: pci 0000:00:17.1: bridge window [mem 0xfcf00000-0xfcffffff] Jul 9 23:56:11.782516 kernel: pci 0000:00:17.1: bridge window [mem 0xe7600000-0xe76fffff 64bit pref] Jul 9 23:56:11.782570 kernel: pci 0000:00:17.2: PCI bridge to [bus 15] Jul 9 23:56:11.782622 kernel: pci 0000:00:17.2: bridge window [io 0xe000-0xefff] Jul 9 23:56:11.782676 kernel: pci 0000:00:17.2: bridge window [mem 0xfcb00000-0xfcbfffff] Jul 9 23:56:11.782729 kernel: pci 0000:00:17.2: bridge window [mem 0xe7200000-0xe72fffff 64bit pref] Jul 9 23:56:11.782781 kernel: pci 0000:00:17.3: PCI bridge to [bus 16] Jul 9 23:56:11.782834 kernel: pci 0000:00:17.3: bridge window [mem 0xfc700000-0xfc7fffff] Jul 9 23:56:11.782887 kernel: pci 0000:00:17.3: bridge window [mem 0xe6e00000-0xe6efffff 64bit pref] Jul 9 23:56:11.782958 kernel: pci 0000:00:17.4: PCI bridge to [bus 17] Jul 9 23:56:11.783014 kernel: pci 0000:00:17.4: bridge window [mem 0xfc300000-0xfc3fffff] Jul 9 23:56:11.783067 kernel: pci 0000:00:17.4: bridge window [mem 0xe6a00000-0xe6afffff 64bit pref] Jul 9 23:56:11.783119 kernel: pci 0000:00:17.5: PCI bridge to [bus 18] Jul 9 23:56:11.783173 kernel: pci 0000:00:17.5: bridge window [mem 0xfbf00000-0xfbffffff] Jul 9 23:56:11.783225 kernel: pci 0000:00:17.5: bridge window [mem 0xe6600000-0xe66fffff 64bit pref] Jul 9 23:56:11.783278 kernel: pci 0000:00:17.6: PCI bridge to [bus 19] Jul 9 23:56:11.783331 kernel: pci 0000:00:17.6: bridge window [mem 0xfbb00000-0xfbbfffff] Jul 9 23:56:11.783391 kernel: pci 0000:00:17.6: bridge window [mem 0xe6200000-0xe62fffff 64bit pref] Jul 9 23:56:11.783449 kernel: pci 0000:00:17.7: PCI bridge to [bus 1a] Jul 9 23:56:11.783505 kernel: pci 0000:00:17.7: bridge window [mem 0xfb700000-0xfb7fffff] Jul 9 23:56:11.783558 kernel: pci 0000:00:17.7: bridge window [mem 0xe5e00000-0xe5efffff 64bit pref] Jul 9 23:56:11.783612 kernel: pci 0000:00:18.0: PCI bridge to [bus 1b] Jul 9 23:56:11.783664 kernel: pci 0000:00:18.0: bridge window [io 0x7000-0x7fff] Jul 9 23:56:11.783717 kernel: pci 0000:00:18.0: bridge window [mem 0xfd200000-0xfd2fffff] Jul 9 23:56:11.783770 kernel: pci 0000:00:18.0: bridge window [mem 0xe7900000-0xe79fffff 64bit pref] Jul 9 23:56:11.783843 kernel: pci 0000:00:18.1: PCI bridge to [bus 1c] Jul 9 23:56:11.783912 kernel: pci 0000:00:18.1: bridge window [io 0xb000-0xbfff] Jul 9 23:56:11.783964 kernel: pci 0000:00:18.1: bridge window [mem 0xfce00000-0xfcefffff] Jul 9 23:56:11.784023 kernel: pci 0000:00:18.1: bridge window [mem 0xe7500000-0xe75fffff 64bit pref] Jul 9 23:56:11.784104 kernel: pci 0000:00:18.2: PCI bridge to [bus 1d] Jul 9 23:56:11.784166 kernel: pci 0000:00:18.2: bridge window [mem 0xfca00000-0xfcafffff] Jul 9 23:56:11.784236 kernel: pci 0000:00:18.2: bridge window [mem 0xe7100000-0xe71fffff 64bit pref] Jul 9 23:56:11.784292 kernel: pci 0000:00:18.3: PCI bridge to [bus 1e] Jul 9 23:56:11.784345 kernel: pci 0000:00:18.3: bridge window [mem 0xfc600000-0xfc6fffff] Jul 9 23:56:11.784423 kernel: pci 0000:00:18.3: bridge window [mem 0xe6d00000-0xe6dfffff 64bit pref] Jul 9 23:56:11.784492 kernel: pci 0000:00:18.4: PCI bridge to [bus 1f] Jul 9 23:56:11.784545 kernel: pci 0000:00:18.4: bridge window [mem 0xfc200000-0xfc2fffff] Jul 9 23:56:11.784597 kernel: pci 0000:00:18.4: bridge window [mem 0xe6900000-0xe69fffff 64bit pref] Jul 9 23:56:11.784654 kernel: pci 0000:00:18.5: PCI bridge to [bus 20] Jul 9 23:56:11.784707 kernel: pci 0000:00:18.5: bridge window [mem 0xfbe00000-0xfbefffff] Jul 9 23:56:11.784768 kernel: pci 0000:00:18.5: bridge window [mem 0xe6500000-0xe65fffff 64bit pref] Jul 9 23:56:11.784823 kernel: pci 0000:00:18.6: PCI bridge to [bus 21] Jul 9 23:56:11.784876 kernel: pci 0000:00:18.6: bridge window [mem 0xfba00000-0xfbafffff] Jul 9 23:56:11.784928 kernel: pci 0000:00:18.6: bridge window [mem 0xe6100000-0xe61fffff 64bit pref] Jul 9 23:56:11.784981 kernel: pci 0000:00:18.7: PCI bridge to [bus 22] Jul 9 23:56:11.785063 kernel: pci 0000:00:18.7: bridge window [mem 0xfb600000-0xfb6fffff] Jul 9 23:56:11.785118 kernel: pci 0000:00:18.7: bridge window [mem 0xe5d00000-0xe5dfffff 64bit pref] Jul 9 23:56:11.785173 kernel: pci_bus 0000:00: resource 4 [mem 0x000a0000-0x000bffff window] Jul 9 23:56:11.785221 kernel: pci_bus 0000:00: resource 5 [mem 0x000cc000-0x000dbfff window] Jul 9 23:56:11.785268 kernel: pci_bus 0000:00: resource 6 [mem 0xc0000000-0xfebfffff window] Jul 9 23:56:11.785315 kernel: pci_bus 0000:00: resource 7 [io 0x0000-0x0cf7 window] Jul 9 23:56:11.785361 kernel: pci_bus 0000:00: resource 8 [io 0x0d00-0xfeff window] Jul 9 23:56:11.785515 kernel: pci_bus 0000:02: resource 0 [io 0x2000-0x3fff] Jul 9 23:56:11.785565 kernel: pci_bus 0000:02: resource 1 [mem 0xfd600000-0xfdffffff] Jul 9 23:56:11.785613 kernel: pci_bus 0000:02: resource 2 [mem 0xe7b00000-0xe7ffffff 64bit pref] Jul 9 23:56:11.785664 kernel: pci_bus 0000:02: resource 4 [mem 0x000a0000-0x000bffff window] Jul 9 23:56:11.785712 kernel: pci_bus 0000:02: resource 5 [mem 0x000cc000-0x000dbfff window] Jul 9 23:56:11.785759 kernel: pci_bus 0000:02: resource 6 [mem 0xc0000000-0xfebfffff window] Jul 9 23:56:11.785806 kernel: pci_bus 0000:02: resource 7 [io 0x0000-0x0cf7 window] Jul 9 23:56:11.785853 kernel: pci_bus 0000:02: resource 8 [io 0x0d00-0xfeff window] Jul 9 23:56:11.785905 kernel: pci_bus 0000:03: resource 0 [io 0x4000-0x4fff] Jul 9 23:56:11.785953 kernel: pci_bus 0000:03: resource 1 [mem 0xfd500000-0xfd5fffff] Jul 9 23:56:11.786004 kernel: pci_bus 0000:03: resource 2 [mem 0xc0000000-0xc01fffff 64bit pref] Jul 9 23:56:11.786055 kernel: pci_bus 0000:04: resource 0 [io 0x8000-0x8fff] Jul 9 23:56:11.786103 kernel: pci_bus 0000:04: resource 1 [mem 0xfd100000-0xfd1fffff] Jul 9 23:56:11.786151 kernel: pci_bus 0000:04: resource 2 [mem 0xe7800000-0xe78fffff 64bit pref] Jul 9 23:56:11.786203 kernel: pci_bus 0000:05: resource 0 [io 0xc000-0xcfff] Jul 9 23:56:11.786251 kernel: pci_bus 0000:05: resource 1 [mem 0xfcd00000-0xfcdfffff] Jul 9 23:56:11.786298 kernel: pci_bus 0000:05: resource 2 [mem 0xe7400000-0xe74fffff 64bit pref] Jul 9 23:56:11.786355 kernel: pci_bus 0000:06: resource 1 [mem 0xfc900000-0xfc9fffff] Jul 9 23:56:11.786458 kernel: pci_bus 0000:06: resource 2 [mem 0xe7000000-0xe70fffff 64bit pref] Jul 9 23:56:11.786525 kernel: pci_bus 0000:07: resource 1 [mem 0xfc500000-0xfc5fffff] Jul 9 23:56:11.786572 kernel: pci_bus 0000:07: resource 2 [mem 0xe6c00000-0xe6cfffff 64bit pref] Jul 9 23:56:11.786624 kernel: pci_bus 0000:08: resource 1 [mem 0xfc100000-0xfc1fffff] Jul 9 23:56:11.786670 kernel: pci_bus 0000:08: resource 2 [mem 0xe6800000-0xe68fffff 64bit pref] Jul 9 23:56:11.786725 kernel: pci_bus 0000:09: resource 1 [mem 0xfbd00000-0xfbdfffff] Jul 9 23:56:11.786773 kernel: pci_bus 0000:09: resource 2 [mem 0xe6400000-0xe64fffff 64bit pref] Jul 9 23:56:11.786823 kernel: pci_bus 0000:0a: resource 1 [mem 0xfb900000-0xfb9fffff] Jul 9 23:56:11.786872 kernel: pci_bus 0000:0a: resource 2 [mem 0xe6000000-0xe60fffff 64bit pref] Jul 9 23:56:11.786934 kernel: pci_bus 0000:0b: resource 0 [io 0x5000-0x5fff] Jul 9 23:56:11.786984 kernel: pci_bus 0000:0b: resource 1 [mem 0xfd400000-0xfd4fffff] Jul 9 23:56:11.787031 kernel: pci_bus 0000:0b: resource 2 [mem 0xc0200000-0xc03fffff 64bit pref] Jul 9 23:56:11.787084 kernel: pci_bus 0000:0c: resource 0 [io 0x9000-0x9fff] Jul 9 23:56:11.787132 kernel: pci_bus 0000:0c: resource 1 [mem 0xfd000000-0xfd0fffff] Jul 9 23:56:11.787179 kernel: pci_bus 0000:0c: resource 2 [mem 0xe7700000-0xe77fffff 64bit pref] Jul 9 23:56:11.787229 kernel: pci_bus 0000:0d: resource 0 [io 0xd000-0xdfff] Jul 9 23:56:11.787295 kernel: pci_bus 0000:0d: resource 1 [mem 0xfcc00000-0xfccfffff] Jul 9 23:56:11.787359 kernel: pci_bus 0000:0d: resource 2 [mem 0xe7300000-0xe73fffff 64bit pref] Jul 9 23:56:11.791504 kernel: pci_bus 0000:0e: resource 1 [mem 0xfc800000-0xfc8fffff] Jul 9 23:56:11.791561 kernel: pci_bus 0000:0e: resource 2 [mem 0xe6f00000-0xe6ffffff 64bit pref] Jul 9 23:56:11.791621 kernel: pci_bus 0000:0f: resource 1 [mem 0xfc400000-0xfc4fffff] Jul 9 23:56:11.791672 kernel: pci_bus 0000:0f: resource 2 [mem 0xe6b00000-0xe6bfffff 64bit pref] Jul 9 23:56:11.791726 kernel: pci_bus 0000:10: resource 1 [mem 0xfc000000-0xfc0fffff] Jul 9 23:56:11.791776 kernel: pci_bus 0000:10: resource 2 [mem 0xe6700000-0xe67fffff 64bit pref] Jul 9 23:56:11.791834 kernel: pci_bus 0000:11: resource 1 [mem 0xfbc00000-0xfbcfffff] Jul 9 23:56:11.791886 kernel: pci_bus 0000:11: resource 2 [mem 0xe6300000-0xe63fffff 64bit pref] Jul 9 23:56:11.791940 kernel: pci_bus 0000:12: resource 1 [mem 0xfb800000-0xfb8fffff] Jul 9 23:56:11.791990 kernel: pci_bus 0000:12: resource 2 [mem 0xe5f00000-0xe5ffffff 64bit pref] Jul 9 23:56:11.792043 kernel: pci_bus 0000:13: resource 0 [io 0x6000-0x6fff] Jul 9 23:56:11.792093 kernel: pci_bus 0000:13: resource 1 [mem 0xfd300000-0xfd3fffff] Jul 9 23:56:11.792145 kernel: pci_bus 0000:13: resource 2 [mem 0xe7a00000-0xe7afffff 64bit pref] Jul 9 23:56:11.792199 kernel: pci_bus 0000:14: resource 0 [io 0xa000-0xafff] Jul 9 23:56:11.792249 kernel: pci_bus 0000:14: resource 1 [mem 0xfcf00000-0xfcffffff] Jul 9 23:56:11.792298 kernel: pci_bus 0000:14: resource 2 [mem 0xe7600000-0xe76fffff 64bit pref] Jul 9 23:56:11.792351 kernel: pci_bus 0000:15: resource 0 [io 0xe000-0xefff] Jul 9 23:56:11.792476 kernel: pci_bus 0000:15: resource 1 [mem 0xfcb00000-0xfcbfffff] Jul 9 23:56:11.792531 kernel: pci_bus 0000:15: resource 2 [mem 0xe7200000-0xe72fffff 64bit pref] Jul 9 23:56:11.792588 kernel: pci_bus 0000:16: resource 1 [mem 0xfc700000-0xfc7fffff] Jul 9 23:56:11.792637 kernel: pci_bus 0000:16: resource 2 [mem 0xe6e00000-0xe6efffff 64bit pref] Jul 9 23:56:11.792716 kernel: pci_bus 0000:17: resource 1 [mem 0xfc300000-0xfc3fffff] Jul 9 23:56:11.792776 kernel: pci_bus 0000:17: resource 2 [mem 0xe6a00000-0xe6afffff 64bit pref] Jul 9 23:56:11.792838 kernel: pci_bus 0000:18: resource 1 [mem 0xfbf00000-0xfbffffff] Jul 9 23:56:11.792900 kernel: pci_bus 0000:18: resource 2 [mem 0xe6600000-0xe66fffff 64bit pref] Jul 9 23:56:11.792967 kernel: pci_bus 0000:19: resource 1 [mem 0xfbb00000-0xfbbfffff] Jul 9 23:56:11.793029 kernel: pci_bus 0000:19: resource 2 [mem 0xe6200000-0xe62fffff 64bit pref] Jul 9 23:56:11.793090 kernel: pci_bus 0000:1a: resource 1 [mem 0xfb700000-0xfb7fffff] Jul 9 23:56:11.793153 kernel: pci_bus 0000:1a: resource 2 [mem 0xe5e00000-0xe5efffff 64bit pref] Jul 9 23:56:11.793208 kernel: pci_bus 0000:1b: resource 0 [io 0x7000-0x7fff] Jul 9 23:56:11.793258 kernel: pci_bus 0000:1b: resource 1 [mem 0xfd200000-0xfd2fffff] Jul 9 23:56:11.793310 kernel: pci_bus 0000:1b: resource 2 [mem 0xe7900000-0xe79fffff 64bit pref] Jul 9 23:56:11.793362 kernel: pci_bus 0000:1c: resource 0 [io 0xb000-0xbfff] Jul 9 23:56:11.793425 kernel: pci_bus 0000:1c: resource 1 [mem 0xfce00000-0xfcefffff] Jul 9 23:56:11.793475 kernel: pci_bus 0000:1c: resource 2 [mem 0xe7500000-0xe75fffff 64bit pref] Jul 9 23:56:11.793528 kernel: pci_bus 0000:1d: resource 1 [mem 0xfca00000-0xfcafffff] Jul 9 23:56:11.793578 kernel: pci_bus 0000:1d: resource 2 [mem 0xe7100000-0xe71fffff 64bit pref] Jul 9 23:56:11.793634 kernel: pci_bus 0000:1e: resource 1 [mem 0xfc600000-0xfc6fffff] Jul 9 23:56:11.793684 kernel: pci_bus 0000:1e: resource 2 [mem 0xe6d00000-0xe6dfffff 64bit pref] Jul 9 23:56:11.793739 kernel: pci_bus 0000:1f: resource 1 [mem 0xfc200000-0xfc2fffff] Jul 9 23:56:11.793788 kernel: pci_bus 0000:1f: resource 2 [mem 0xe6900000-0xe69fffff 64bit pref] Jul 9 23:56:11.793862 kernel: pci_bus 0000:20: resource 1 [mem 0xfbe00000-0xfbefffff] Jul 9 23:56:11.793921 kernel: pci_bus 0000:20: resource 2 [mem 0xe6500000-0xe65fffff 64bit pref] Jul 9 23:56:11.793985 kernel: pci_bus 0000:21: resource 1 [mem 0xfba00000-0xfbafffff] Jul 9 23:56:11.794047 kernel: pci_bus 0000:21: resource 2 [mem 0xe6100000-0xe61fffff 64bit pref] Jul 9 23:56:11.794101 kernel: pci_bus 0000:22: resource 1 [mem 0xfb600000-0xfb6fffff] Jul 9 23:56:11.794152 kernel: pci_bus 0000:22: resource 2 [mem 0xe5d00000-0xe5dfffff 64bit pref] Jul 9 23:56:11.794212 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers Jul 9 23:56:11.794222 kernel: PCI: CLS 32 bytes, default 64 Jul 9 23:56:11.794229 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Jul 9 23:56:11.794238 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x311fd3cd494, max_idle_ns: 440795223879 ns Jul 9 23:56:11.794245 kernel: clocksource: Switched to clocksource tsc Jul 9 23:56:11.794252 kernel: Initialise system trusted keyrings Jul 9 23:56:11.794259 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 Jul 9 23:56:11.794265 kernel: Key type asymmetric registered Jul 9 23:56:11.794271 kernel: Asymmetric key parser 'x509' registered Jul 9 23:56:11.794277 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Jul 9 23:56:11.794284 kernel: io scheduler mq-deadline registered Jul 9 23:56:11.794290 kernel: io scheduler kyber registered Jul 9 23:56:11.794298 kernel: io scheduler bfq registered Jul 9 23:56:11.794355 kernel: pcieport 0000:00:15.0: PME: Signaling with IRQ 24 Jul 9 23:56:11.795000 kernel: pcieport 0000:00:15.0: pciehp: Slot #160 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 9 23:56:11.795079 kernel: pcieport 0000:00:15.1: PME: Signaling with IRQ 25 Jul 9 23:56:11.795142 kernel: pcieport 0000:00:15.1: pciehp: Slot #161 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 9 23:56:11.795210 kernel: pcieport 0000:00:15.2: PME: Signaling with IRQ 26 Jul 9 23:56:11.795267 kernel: pcieport 0000:00:15.2: pciehp: Slot #162 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 9 23:56:11.795323 kernel: pcieport 0000:00:15.3: PME: Signaling with IRQ 27 Jul 9 23:56:11.795390 kernel: pcieport 0000:00:15.3: pciehp: Slot #163 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 9 23:56:11.795446 kernel: pcieport 0000:00:15.4: PME: Signaling with IRQ 28 Jul 9 23:56:11.795501 kernel: pcieport 0000:00:15.4: pciehp: Slot #164 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 9 23:56:11.795557 kernel: pcieport 0000:00:15.5: PME: Signaling with IRQ 29 Jul 9 23:56:11.795613 kernel: pcieport 0000:00:15.5: pciehp: Slot #165 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 9 23:56:11.795672 kernel: pcieport 0000:00:15.6: PME: Signaling with IRQ 30 Jul 9 23:56:11.795728 kernel: pcieport 0000:00:15.6: pciehp: Slot #166 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 9 23:56:11.795784 kernel: pcieport 0000:00:15.7: PME: Signaling with IRQ 31 Jul 9 23:56:11.795839 kernel: pcieport 0000:00:15.7: pciehp: Slot #167 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 9 23:56:11.795916 kernel: pcieport 0000:00:16.0: PME: Signaling with IRQ 32 Jul 9 23:56:11.795979 kernel: pcieport 0000:00:16.0: pciehp: Slot #192 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 9 23:56:11.796051 kernel: pcieport 0000:00:16.1: PME: Signaling with IRQ 33 Jul 9 23:56:11.796117 kernel: pcieport 0000:00:16.1: pciehp: Slot #193 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 9 23:56:11.796195 kernel: pcieport 0000:00:16.2: PME: Signaling with IRQ 34 Jul 9 23:56:11.796251 kernel: pcieport 0000:00:16.2: pciehp: Slot #194 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 9 23:56:11.796311 kernel: pcieport 0000:00:16.3: PME: Signaling with IRQ 35 Jul 9 23:56:11.796367 kernel: pcieport 0000:00:16.3: pciehp: Slot #195 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 9 23:56:11.796522 kernel: pcieport 0000:00:16.4: PME: Signaling with IRQ 36 Jul 9 23:56:11.796576 kernel: pcieport 0000:00:16.4: pciehp: Slot #196 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 9 23:56:11.796634 kernel: pcieport 0000:00:16.5: PME: Signaling with IRQ 37 Jul 9 23:56:11.796688 kernel: pcieport 0000:00:16.5: pciehp: Slot #197 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 9 23:56:11.796743 kernel: pcieport 0000:00:16.6: PME: Signaling with IRQ 38 Jul 9 23:56:11.796797 kernel: pcieport 0000:00:16.6: pciehp: Slot #198 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 9 23:56:11.796854 kernel: pcieport 0000:00:16.7: PME: Signaling with IRQ 39 Jul 9 23:56:11.796908 kernel: pcieport 0000:00:16.7: pciehp: Slot #199 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 9 23:56:11.796978 kernel: pcieport 0000:00:17.0: PME: Signaling with IRQ 40 Jul 9 23:56:11.797054 kernel: pcieport 0000:00:17.0: pciehp: Slot #224 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 9 23:56:11.797117 kernel: pcieport 0000:00:17.1: PME: Signaling with IRQ 41 Jul 9 23:56:11.797188 kernel: pcieport 0000:00:17.1: pciehp: Slot #225 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 9 23:56:11.797266 kernel: pcieport 0000:00:17.2: PME: Signaling with IRQ 42 Jul 9 23:56:11.797334 kernel: pcieport 0000:00:17.2: pciehp: Slot #226 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 9 23:56:11.797401 kernel: pcieport 0000:00:17.3: PME: Signaling with IRQ 43 Jul 9 23:56:11.797466 kernel: pcieport 0000:00:17.3: pciehp: Slot #227 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 9 23:56:11.797521 kernel: pcieport 0000:00:17.4: PME: Signaling with IRQ 44 Jul 9 23:56:11.797575 kernel: pcieport 0000:00:17.4: pciehp: Slot #228 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 9 23:56:11.797634 kernel: pcieport 0000:00:17.5: PME: Signaling with IRQ 45 Jul 9 23:56:11.797688 kernel: pcieport 0000:00:17.5: pciehp: Slot #229 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 9 23:56:11.797742 kernel: pcieport 0000:00:17.6: PME: Signaling with IRQ 46 Jul 9 23:56:11.797795 kernel: pcieport 0000:00:17.6: pciehp: Slot #230 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 9 23:56:11.797850 kernel: pcieport 0000:00:17.7: PME: Signaling with IRQ 47 Jul 9 23:56:11.797907 kernel: pcieport 0000:00:17.7: pciehp: Slot #231 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 9 23:56:11.797962 kernel: pcieport 0000:00:18.0: PME: Signaling with IRQ 48 Jul 9 23:56:11.798016 kernel: pcieport 0000:00:18.0: pciehp: Slot #256 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 9 23:56:11.798073 kernel: pcieport 0000:00:18.1: PME: Signaling with IRQ 49 Jul 9 23:56:11.798137 kernel: pcieport 0000:00:18.1: pciehp: Slot #257 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 9 23:56:11.798193 kernel: pcieport 0000:00:18.2: PME: Signaling with IRQ 50 Jul 9 23:56:11.798263 kernel: pcieport 0000:00:18.2: pciehp: Slot #258 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 9 23:56:11.798344 kernel: pcieport 0000:00:18.3: PME: Signaling with IRQ 51 Jul 9 23:56:11.798456 kernel: pcieport 0000:00:18.3: pciehp: Slot #259 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 9 23:56:11.798517 kernel: pcieport 0000:00:18.4: PME: Signaling with IRQ 52 Jul 9 23:56:11.798571 kernel: pcieport 0000:00:18.4: pciehp: Slot #260 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 9 23:56:11.798628 kernel: pcieport 0000:00:18.5: PME: Signaling with IRQ 53 Jul 9 23:56:11.798708 kernel: pcieport 0000:00:18.5: pciehp: Slot #261 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 9 23:56:11.798764 kernel: pcieport 0000:00:18.6: PME: Signaling with IRQ 54 Jul 9 23:56:11.798818 kernel: pcieport 0000:00:18.6: pciehp: Slot #262 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 9 23:56:11.798896 kernel: pcieport 0000:00:18.7: PME: Signaling with IRQ 55 Jul 9 23:56:11.798977 kernel: pcieport 0000:00:18.7: pciehp: Slot #263 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ Jul 9 23:56:11.798990 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Jul 9 23:56:11.798997 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Jul 9 23:56:11.799004 kernel: 00:05: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Jul 9 23:56:11.799010 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBC,PNP0f13:MOUS] at 0x60,0x64 irq 1,12 Jul 9 23:56:11.799017 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Jul 9 23:56:11.799023 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Jul 9 23:56:11.799086 kernel: rtc_cmos 00:01: registered as rtc0 Jul 9 23:56:11.799154 kernel: rtc_cmos 00:01: setting system clock to 2025-07-09T23:56:11 UTC (1752105371) Jul 9 23:56:11.799225 kernel: rtc_cmos 00:01: alarms up to one month, y3k, 114 bytes nvram Jul 9 23:56:11.799240 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Jul 9 23:56:11.799252 kernel: intel_pstate: CPU model not supported Jul 9 23:56:11.799259 kernel: NET: Registered PF_INET6 protocol family Jul 9 23:56:11.799265 kernel: Segment Routing with IPv6 Jul 9 23:56:11.799272 kernel: In-situ OAM (IOAM) with IPv6 Jul 9 23:56:11.799278 kernel: NET: Registered PF_PACKET protocol family Jul 9 23:56:11.799284 kernel: Key type dns_resolver registered Jul 9 23:56:11.799291 kernel: IPI shorthand broadcast: enabled Jul 9 23:56:11.799299 kernel: sched_clock: Marking stable (912228948, 228814812)->(1206063847, -65020087) Jul 9 23:56:11.799305 kernel: registered taskstats version 1 Jul 9 23:56:11.799312 kernel: Loading compiled-in X.509 certificates Jul 9 23:56:11.799318 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.95-flatcar: 50743221a03cbb928e294992219bf2bc20f6f14b' Jul 9 23:56:11.799324 kernel: Key type .fscrypt registered Jul 9 23:56:11.799331 kernel: Key type fscrypt-provisioning registered Jul 9 23:56:11.799337 kernel: ima: No TPM chip found, activating TPM-bypass! Jul 9 23:56:11.799343 kernel: ima: Allocated hash algorithm: sha1 Jul 9 23:56:11.799351 kernel: ima: No architecture policies found Jul 9 23:56:11.799357 kernel: clk: Disabling unused clocks Jul 9 23:56:11.799364 kernel: Freeing unused kernel image (initmem) memory: 43488K Jul 9 23:56:11.799378 kernel: Write protecting the kernel read-only data: 38912k Jul 9 23:56:11.799385 kernel: Freeing unused kernel image (rodata/data gap) memory: 1704K Jul 9 23:56:11.799392 kernel: Run /init as init process Jul 9 23:56:11.799398 kernel: with arguments: Jul 9 23:56:11.799405 kernel: /init Jul 9 23:56:11.799413 kernel: with environment: Jul 9 23:56:11.799423 kernel: HOME=/ Jul 9 23:56:11.799436 kernel: TERM=linux Jul 9 23:56:11.799448 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Jul 9 23:56:11.799458 systemd[1]: Successfully made /usr/ read-only. Jul 9 23:56:11.799467 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Jul 9 23:56:11.799474 systemd[1]: Detected virtualization vmware. Jul 9 23:56:11.799486 systemd[1]: Detected architecture x86-64. Jul 9 23:56:11.799494 systemd[1]: Running in initrd. Jul 9 23:56:11.799501 systemd[1]: No hostname configured, using default hostname. Jul 9 23:56:11.799510 systemd[1]: Hostname set to . Jul 9 23:56:11.799516 systemd[1]: Initializing machine ID from random generator. Jul 9 23:56:11.799523 systemd[1]: Queued start job for default target initrd.target. Jul 9 23:56:11.799529 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jul 9 23:56:11.799536 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jul 9 23:56:11.799544 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Jul 9 23:56:11.799550 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jul 9 23:56:11.799558 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Jul 9 23:56:11.799566 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Jul 9 23:56:11.799573 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Jul 9 23:56:11.799580 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Jul 9 23:56:11.799587 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jul 9 23:56:11.799599 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jul 9 23:56:11.799611 systemd[1]: Reached target paths.target - Path Units. Jul 9 23:56:11.799623 systemd[1]: Reached target slices.target - Slice Units. Jul 9 23:56:11.799631 systemd[1]: Reached target swap.target - Swaps. Jul 9 23:56:11.799638 systemd[1]: Reached target timers.target - Timer Units. Jul 9 23:56:11.799644 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Jul 9 23:56:11.799651 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jul 9 23:56:11.799657 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Jul 9 23:56:11.799664 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Jul 9 23:56:11.799671 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jul 9 23:56:11.799677 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jul 9 23:56:11.799689 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jul 9 23:56:11.799695 systemd[1]: Reached target sockets.target - Socket Units. Jul 9 23:56:11.799703 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Jul 9 23:56:11.799714 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jul 9 23:56:11.799726 systemd[1]: Finished network-cleanup.service - Network Cleanup. Jul 9 23:56:11.799738 systemd[1]: Starting systemd-fsck-usr.service... Jul 9 23:56:11.799750 systemd[1]: Starting systemd-journald.service - Journal Service... Jul 9 23:56:11.799758 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jul 9 23:56:11.799766 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jul 9 23:56:11.799789 systemd-journald[215]: Collecting audit messages is disabled. Jul 9 23:56:11.799806 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Jul 9 23:56:11.799813 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jul 9 23:56:11.799822 systemd[1]: Finished systemd-fsck-usr.service. Jul 9 23:56:11.799829 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Jul 9 23:56:11.799837 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jul 9 23:56:11.799844 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Jul 9 23:56:11.799851 kernel: Bridge firewalling registered Jul 9 23:56:11.799858 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jul 9 23:56:11.799865 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jul 9 23:56:11.799872 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jul 9 23:56:11.799883 systemd-journald[215]: Journal started Jul 9 23:56:11.799905 systemd-journald[215]: Runtime Journal (/run/log/journal/77b13b93103a4a87a53f747005f09748) is 4.8M, max 38.6M, 33.8M free. Jul 9 23:56:11.753744 systemd-modules-load[216]: Inserted module 'overlay' Jul 9 23:56:11.784771 systemd-modules-load[216]: Inserted module 'br_netfilter' Jul 9 23:56:11.802624 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jul 9 23:56:11.807034 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jul 9 23:56:11.807067 systemd[1]: Started systemd-journald.service - Journal Service. Jul 9 23:56:11.807638 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jul 9 23:56:11.808067 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jul 9 23:56:11.810238 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Jul 9 23:56:11.821482 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jul 9 23:56:11.821968 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jul 9 23:56:11.827823 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jul 9 23:56:11.829152 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jul 9 23:56:11.829903 dracut-cmdline[248]: dracut-dracut-053 Jul 9 23:56:11.833049 dracut-cmdline[248]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=vmware flatcar.autologin verity.usrhash=c257b65f06e0ad68d969d5b3e057f031663dc29a4487d91a77595a40c4dc82d6 Jul 9 23:56:11.853428 systemd-resolved[259]: Positive Trust Anchors: Jul 9 23:56:11.853438 systemd-resolved[259]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jul 9 23:56:11.853460 systemd-resolved[259]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jul 9 23:56:11.855560 systemd-resolved[259]: Defaulting to hostname 'linux'. Jul 9 23:56:11.856204 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jul 9 23:56:11.856601 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jul 9 23:56:11.885386 kernel: SCSI subsystem initialized Jul 9 23:56:11.894386 kernel: Loading iSCSI transport class v2.0-870. Jul 9 23:56:11.901383 kernel: iscsi: registered transport (tcp) Jul 9 23:56:11.915731 kernel: iscsi: registered transport (qla4xxx) Jul 9 23:56:11.915777 kernel: QLogic iSCSI HBA Driver Jul 9 23:56:11.936249 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Jul 9 23:56:11.943487 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Jul 9 23:56:11.959279 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Jul 9 23:56:11.959320 kernel: device-mapper: uevent: version 1.0.3 Jul 9 23:56:11.959329 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Jul 9 23:56:11.990385 kernel: raid6: avx2x4 gen() 47674 MB/s Jul 9 23:56:12.007386 kernel: raid6: avx2x2 gen() 53289 MB/s Jul 9 23:56:12.024583 kernel: raid6: avx2x1 gen() 44737 MB/s Jul 9 23:56:12.024634 kernel: raid6: using algorithm avx2x2 gen() 53289 MB/s Jul 9 23:56:12.042578 kernel: raid6: .... xor() 32337 MB/s, rmw enabled Jul 9 23:56:12.042606 kernel: raid6: using avx2x2 recovery algorithm Jul 9 23:56:12.056383 kernel: xor: automatically using best checksumming function avx Jul 9 23:56:12.146395 kernel: Btrfs loaded, zoned=no, fsverity=no Jul 9 23:56:12.151428 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Jul 9 23:56:12.155466 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jul 9 23:56:12.163948 systemd-udevd[436]: Using default interface naming scheme 'v255'. Jul 9 23:56:12.167062 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jul 9 23:56:12.176517 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Jul 9 23:56:12.183491 dracut-pre-trigger[440]: rd.md=0: removing MD RAID activation Jul 9 23:56:12.198818 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Jul 9 23:56:12.203469 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jul 9 23:56:12.275577 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jul 9 23:56:12.279458 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Jul 9 23:56:12.292057 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Jul 9 23:56:12.293760 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Jul 9 23:56:12.294009 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jul 9 23:56:12.294210 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jul 9 23:56:12.298468 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Jul 9 23:56:12.306493 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Jul 9 23:56:12.354381 kernel: VMware PVSCSI driver - version 1.0.7.0-k Jul 9 23:56:12.357716 kernel: vmw_pvscsi: using 64bit dma Jul 9 23:56:12.357747 kernel: vmw_pvscsi: max_id: 16 Jul 9 23:56:12.357755 kernel: vmw_pvscsi: setting ring_pages to 8 Jul 9 23:56:12.358382 kernel: VMware vmxnet3 virtual NIC driver - version 1.7.0.0-k-NAPI Jul 9 23:56:12.362482 kernel: vmw_pvscsi: enabling reqCallThreshold Jul 9 23:56:12.362501 kernel: vmw_pvscsi: driver-based request coalescing enabled Jul 9 23:56:12.362510 kernel: vmw_pvscsi: using MSI-X Jul 9 23:56:12.362518 kernel: vmxnet3 0000:0b:00.0: # of Tx queues : 2, # of Rx queues : 2 Jul 9 23:56:12.368636 kernel: scsi host0: VMware PVSCSI storage adapter rev 2, req/cmp/msg rings: 8/8/1 pages, cmd_per_lun=254 Jul 9 23:56:12.368757 kernel: vmw_pvscsi 0000:03:00.0: VMware PVSCSI rev 2 host #0 Jul 9 23:56:12.368841 kernel: scsi 0:0:0:0: Direct-Access VMware Virtual disk 2.0 PQ: 0 ANSI: 6 Jul 9 23:56:12.371385 kernel: vmxnet3 0000:0b:00.0 eth0: NIC Link is Up 10000 Mbps Jul 9 23:56:12.386384 kernel: cryptd: max_cpu_qlen set to 1000 Jul 9 23:56:12.392820 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jul 9 23:56:12.393377 kernel: vmxnet3 0000:0b:00.0 ens192: renamed from eth0 Jul 9 23:56:12.394656 kernel: AVX2 version of gcm_enc/dec engaged. Jul 9 23:56:12.394512 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jul 9 23:56:12.395898 kernel: AES CTR mode by8 optimization enabled Jul 9 23:56:12.395910 kernel: libata version 3.00 loaded. Jul 9 23:56:12.396481 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jul 9 23:56:12.396728 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jul 9 23:56:12.396918 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jul 9 23:56:12.398762 kernel: ata_piix 0000:00:07.1: version 2.13 Jul 9 23:56:12.397407 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Jul 9 23:56:12.401271 kernel: scsi host1: ata_piix Jul 9 23:56:12.401364 kernel: scsi host2: ata_piix Jul 9 23:56:12.401458 kernel: ata1: PATA max UDMA/33 cmd 0x1f0 ctl 0x3f6 bmdma 0x1060 irq 14 Jul 9 23:56:12.401468 kernel: ata2: PATA max UDMA/33 cmd 0x170 ctl 0x376 bmdma 0x1068 irq 15 Jul 9 23:56:12.406594 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jul 9 23:56:12.420319 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jul 9 23:56:12.423485 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jul 9 23:56:12.430281 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jul 9 23:56:12.570440 kernel: ata2.00: ATAPI: VMware Virtual IDE CDROM Drive, 00000001, max UDMA/33 Jul 9 23:56:12.578092 kernel: scsi 2:0:0:0: CD-ROM NECVMWar VMware IDE CDR10 1.00 PQ: 0 ANSI: 5 Jul 9 23:56:12.590908 kernel: sd 0:0:0:0: [sda] 17805312 512-byte logical blocks: (9.12 GB/8.49 GiB) Jul 9 23:56:12.591030 kernel: sd 0:0:0:0: [sda] Write Protect is off Jul 9 23:56:12.591101 kernel: sd 0:0:0:0: [sda] Mode Sense: 31 00 00 00 Jul 9 23:56:12.592524 kernel: sd 0:0:0:0: [sda] Cache data unavailable Jul 9 23:56:12.592604 kernel: sd 0:0:0:0: [sda] Assuming drive cache: write through Jul 9 23:56:12.596379 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 1x/1x writer dvd-ram cd/rw xa/form2 cdda tray Jul 9 23:56:12.596472 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jul 9 23:56:12.596486 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Jul 9 23:56:12.597425 kernel: sd 0:0:0:0: [sda] Attached SCSI disk Jul 9 23:56:12.611380 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Jul 9 23:56:12.649414 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/sda6 scanned by (udev-worker) (485) Jul 9 23:56:12.655394 kernel: BTRFS: device fsid 2ea7ed46-2399-4750-93a6-9faa0c83416c devid 1 transid 37 /dev/sda3 scanned by (udev-worker) (501) Jul 9 23:56:12.657436 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Virtual_disk ROOT. Jul 9 23:56:12.663620 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Virtual_disk EFI-SYSTEM. Jul 9 23:56:12.669332 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Virtual_disk OEM. Jul 9 23:56:12.673892 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Virtual_disk USR-A. Jul 9 23:56:12.674135 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Virtual_disk USR-A. Jul 9 23:56:12.682450 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Jul 9 23:56:12.706382 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jul 9 23:56:13.717414 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jul 9 23:56:13.717451 disk-uuid[597]: The operation has completed successfully. Jul 9 23:56:13.754753 systemd[1]: disk-uuid.service: Deactivated successfully. Jul 9 23:56:13.754816 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Jul 9 23:56:13.773449 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Jul 9 23:56:13.775071 sh[613]: Success Jul 9 23:56:13.783389 kernel: device-mapper: verity: sha256 using implementation "sha256-avx2" Jul 9 23:56:13.822553 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Jul 9 23:56:13.826862 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Jul 9 23:56:13.827633 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Jul 9 23:56:13.843392 kernel: BTRFS info (device dm-0): first mount of filesystem 2ea7ed46-2399-4750-93a6-9faa0c83416c Jul 9 23:56:13.843414 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Jul 9 23:56:13.843423 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Jul 9 23:56:13.844493 kernel: BTRFS info (device dm-0): disabling log replay at mount time Jul 9 23:56:13.844506 kernel: BTRFS info (device dm-0): using free space tree Jul 9 23:56:13.851380 kernel: BTRFS info (device dm-0): enabling ssd optimizations Jul 9 23:56:13.853310 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Jul 9 23:56:13.860465 systemd[1]: Starting afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments... Jul 9 23:56:13.861560 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Jul 9 23:56:13.879846 kernel: BTRFS info (device sda6): first mount of filesystem 8e2332fd-cd78-45f6-aab3-8af291a1450c Jul 9 23:56:13.879880 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Jul 9 23:56:13.879889 kernel: BTRFS info (device sda6): using free space tree Jul 9 23:56:13.882384 kernel: BTRFS info (device sda6): enabling ssd optimizations Jul 9 23:56:13.885385 kernel: BTRFS info (device sda6): last unmount of filesystem 8e2332fd-cd78-45f6-aab3-8af291a1450c Jul 9 23:56:13.886344 systemd[1]: Finished ignition-setup.service - Ignition (setup). Jul 9 23:56:13.890484 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Jul 9 23:56:13.939333 systemd[1]: Finished afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments. Jul 9 23:56:13.944405 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Jul 9 23:56:13.995545 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jul 9 23:56:14.001342 ignition[670]: Ignition 2.20.0 Jul 9 23:56:14.001349 ignition[670]: Stage: fetch-offline Jul 9 23:56:14.004455 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jul 9 23:56:14.001379 ignition[670]: no configs at "/usr/lib/ignition/base.d" Jul 9 23:56:14.004758 unknown[670]: fetched base config from "system" Jul 9 23:56:14.001392 ignition[670]: no config dir at "/usr/lib/ignition/base.platform.d/vmware" Jul 9 23:56:14.004762 unknown[670]: fetched user config from "vmware" Jul 9 23:56:14.001442 ignition[670]: parsed url from cmdline: "" Jul 9 23:56:14.006231 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Jul 9 23:56:14.001444 ignition[670]: no config URL provided Jul 9 23:56:14.001446 ignition[670]: reading system config file "/usr/lib/ignition/user.ign" Jul 9 23:56:14.001451 ignition[670]: no config at "/usr/lib/ignition/user.ign" Jul 9 23:56:14.001792 ignition[670]: config successfully fetched Jul 9 23:56:14.001808 ignition[670]: parsing config with SHA512: aa41d3989f89ca51a7c6448231bca6362d5a0256dfadc6c09f3b480557de1d5a761066296b78b5bf2ccf2fa83fa6c033fad4504bc6729e4c2189af4c28c51ed9 Jul 9 23:56:14.005076 ignition[670]: fetch-offline: fetch-offline passed Jul 9 23:56:14.005128 ignition[670]: Ignition finished successfully Jul 9 23:56:14.018770 systemd-networkd[802]: lo: Link UP Jul 9 23:56:14.018776 systemd-networkd[802]: lo: Gained carrier Jul 9 23:56:14.019551 systemd-networkd[802]: Enumeration completed Jul 9 23:56:14.019595 systemd[1]: Started systemd-networkd.service - Network Configuration. Jul 9 23:56:14.023528 kernel: vmxnet3 0000:0b:00.0 ens192: intr type 3, mode 0, 3 vectors allocated Jul 9 23:56:14.023630 kernel: vmxnet3 0000:0b:00.0 ens192: NIC Link is Up 10000 Mbps Jul 9 23:56:14.019757 systemd[1]: Reached target network.target - Network. Jul 9 23:56:14.020127 systemd-networkd[802]: ens192: Configuring with /etc/systemd/network/10-dracut-cmdline-99.network. Jul 9 23:56:14.023134 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Jul 9 23:56:14.023527 systemd-networkd[802]: ens192: Link UP Jul 9 23:56:14.023529 systemd-networkd[802]: ens192: Gained carrier Jul 9 23:56:14.028451 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Jul 9 23:56:14.036153 ignition[807]: Ignition 2.20.0 Jul 9 23:56:14.036160 ignition[807]: Stage: kargs Jul 9 23:56:14.036257 ignition[807]: no configs at "/usr/lib/ignition/base.d" Jul 9 23:56:14.036263 ignition[807]: no config dir at "/usr/lib/ignition/base.platform.d/vmware" Jul 9 23:56:14.036887 ignition[807]: kargs: kargs passed Jul 9 23:56:14.036914 ignition[807]: Ignition finished successfully Jul 9 23:56:14.038056 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Jul 9 23:56:14.041541 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Jul 9 23:56:14.051179 ignition[814]: Ignition 2.20.0 Jul 9 23:56:14.051186 ignition[814]: Stage: disks Jul 9 23:56:14.051293 ignition[814]: no configs at "/usr/lib/ignition/base.d" Jul 9 23:56:14.051299 ignition[814]: no config dir at "/usr/lib/ignition/base.platform.d/vmware" Jul 9 23:56:14.051793 ignition[814]: disks: disks passed Jul 9 23:56:14.051821 ignition[814]: Ignition finished successfully Jul 9 23:56:14.052487 systemd[1]: Finished ignition-disks.service - Ignition (disks). Jul 9 23:56:14.052812 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Jul 9 23:56:14.052932 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Jul 9 23:56:14.053105 systemd[1]: Reached target local-fs.target - Local File Systems. Jul 9 23:56:14.053274 systemd[1]: Reached target sysinit.target - System Initialization. Jul 9 23:56:14.053460 systemd[1]: Reached target basic.target - Basic System. Jul 9 23:56:14.057466 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Jul 9 23:56:14.067428 systemd-fsck[823]: ROOT: clean, 14/1628000 files, 120691/1617920 blocks Jul 9 23:56:14.068375 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Jul 9 23:56:14.075468 systemd[1]: Mounting sysroot.mount - /sysroot... Jul 9 23:56:14.130379 kernel: EXT4-fs (sda9): mounted filesystem 147af866-f15a-4a2f-aea7-d9959c235d2a r/w with ordered data mode. Quota mode: none. Jul 9 23:56:14.130648 systemd[1]: Mounted sysroot.mount - /sysroot. Jul 9 23:56:14.130985 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Jul 9 23:56:14.139417 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jul 9 23:56:14.140575 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Jul 9 23:56:14.140891 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Jul 9 23:56:14.140917 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Jul 9 23:56:14.140931 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Jul 9 23:56:14.143889 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Jul 9 23:56:14.144794 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Jul 9 23:56:14.147389 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/sda6 scanned by mount (831) Jul 9 23:56:14.149751 kernel: BTRFS info (device sda6): first mount of filesystem 8e2332fd-cd78-45f6-aab3-8af291a1450c Jul 9 23:56:14.149772 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Jul 9 23:56:14.149782 kernel: BTRFS info (device sda6): using free space tree Jul 9 23:56:14.154336 kernel: BTRFS info (device sda6): enabling ssd optimizations Jul 9 23:56:14.153982 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jul 9 23:56:14.176216 initrd-setup-root[855]: cut: /sysroot/etc/passwd: No such file or directory Jul 9 23:56:14.178683 initrd-setup-root[862]: cut: /sysroot/etc/group: No such file or directory Jul 9 23:56:14.181225 initrd-setup-root[869]: cut: /sysroot/etc/shadow: No such file or directory Jul 9 23:56:14.183219 initrd-setup-root[876]: cut: /sysroot/etc/gshadow: No such file or directory Jul 9 23:56:14.233805 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Jul 9 23:56:14.238412 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Jul 9 23:56:14.240916 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Jul 9 23:56:14.245391 kernel: BTRFS info (device sda6): last unmount of filesystem 8e2332fd-cd78-45f6-aab3-8af291a1450c Jul 9 23:56:14.258044 ignition[944]: INFO : Ignition 2.20.0 Jul 9 23:56:14.258044 ignition[944]: INFO : Stage: mount Jul 9 23:56:14.258337 ignition[944]: INFO : no configs at "/usr/lib/ignition/base.d" Jul 9 23:56:14.258337 ignition[944]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/vmware" Jul 9 23:56:14.259384 ignition[944]: INFO : mount: mount passed Jul 9 23:56:14.259384 ignition[944]: INFO : Ignition finished successfully Jul 9 23:56:14.260256 systemd[1]: Finished ignition-mount.service - Ignition (mount). Jul 9 23:56:14.260694 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Jul 9 23:56:14.265422 systemd[1]: Starting ignition-files.service - Ignition (files)... Jul 9 23:56:14.840930 systemd[1]: sysroot-oem.mount: Deactivated successfully. Jul 9 23:56:14.845468 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jul 9 23:56:14.854232 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/sda6 scanned by mount (955) Jul 9 23:56:14.854257 kernel: BTRFS info (device sda6): first mount of filesystem 8e2332fd-cd78-45f6-aab3-8af291a1450c Jul 9 23:56:14.854266 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Jul 9 23:56:14.855865 kernel: BTRFS info (device sda6): using free space tree Jul 9 23:56:14.859973 kernel: BTRFS info (device sda6): enabling ssd optimizations Jul 9 23:56:14.859611 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jul 9 23:56:14.877058 ignition[972]: INFO : Ignition 2.20.0 Jul 9 23:56:14.877058 ignition[972]: INFO : Stage: files Jul 9 23:56:14.877458 ignition[972]: INFO : no configs at "/usr/lib/ignition/base.d" Jul 9 23:56:14.877458 ignition[972]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/vmware" Jul 9 23:56:14.877846 ignition[972]: DEBUG : files: compiled without relabeling support, skipping Jul 9 23:56:14.878412 ignition[972]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Jul 9 23:56:14.878412 ignition[972]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Jul 9 23:56:14.880425 ignition[972]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Jul 9 23:56:14.880566 ignition[972]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Jul 9 23:56:14.880707 ignition[972]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Jul 9 23:56:14.880636 unknown[972]: wrote ssh authorized keys file for user: core Jul 9 23:56:14.882423 ignition[972]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Jul 9 23:56:14.882653 ignition[972]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.3-linux-amd64.tar.gz: attempt #1 Jul 9 23:56:14.936560 ignition[972]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Jul 9 23:56:15.189616 ignition[972]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-amd64.tar.gz" Jul 9 23:56:15.189616 ignition[972]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Jul 9 23:56:15.189616 ignition[972]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1 Jul 9 23:56:15.204600 systemd-networkd[802]: ens192: Gained IPv6LL Jul 9 23:56:15.642093 ignition[972]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Jul 9 23:56:15.711063 ignition[972]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Jul 9 23:56:15.711063 ignition[972]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Jul 9 23:56:15.711063 ignition[972]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Jul 9 23:56:15.711063 ignition[972]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Jul 9 23:56:15.711063 ignition[972]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Jul 9 23:56:15.711063 ignition[972]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Jul 9 23:56:15.711063 ignition[972]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Jul 9 23:56:15.711063 ignition[972]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Jul 9 23:56:15.711063 ignition[972]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Jul 9 23:56:15.711063 ignition[972]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Jul 9 23:56:15.711063 ignition[972]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Jul 9 23:56:15.711063 ignition[972]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Jul 9 23:56:15.711063 ignition[972]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Jul 9 23:56:15.711063 ignition[972]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Jul 9 23:56:15.711063 ignition[972]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://extensions.flatcar.org/extensions/kubernetes-v1.33.0-x86-64.raw: attempt #1 Jul 9 23:56:16.354281 ignition[972]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Jul 9 23:56:16.641572 ignition[972]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-x86-64.raw" Jul 9 23:56:16.641850 ignition[972]: INFO : files: createFilesystemsFiles: createFiles: op(c): [started] writing file "/sysroot/etc/systemd/network/00-vmware.network" Jul 9 23:56:16.641850 ignition[972]: INFO : files: createFilesystemsFiles: createFiles: op(c): [finished] writing file "/sysroot/etc/systemd/network/00-vmware.network" Jul 9 23:56:16.641850 ignition[972]: INFO : files: op(d): [started] processing unit "prepare-helm.service" Jul 9 23:56:16.642446 ignition[972]: INFO : files: op(d): op(e): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jul 9 23:56:16.642614 ignition[972]: INFO : files: op(d): op(e): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jul 9 23:56:16.642614 ignition[972]: INFO : files: op(d): [finished] processing unit "prepare-helm.service" Jul 9 23:56:16.642614 ignition[972]: INFO : files: op(f): [started] processing unit "coreos-metadata.service" Jul 9 23:56:16.642614 ignition[972]: INFO : files: op(f): op(10): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Jul 9 23:56:16.642614 ignition[972]: INFO : files: op(f): op(10): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Jul 9 23:56:16.642614 ignition[972]: INFO : files: op(f): [finished] processing unit "coreos-metadata.service" Jul 9 23:56:16.642614 ignition[972]: INFO : files: op(11): [started] setting preset to disabled for "coreos-metadata.service" Jul 9 23:56:16.664038 ignition[972]: INFO : files: op(11): op(12): [started] removing enablement symlink(s) for "coreos-metadata.service" Jul 9 23:56:16.666396 ignition[972]: INFO : files: op(11): op(12): [finished] removing enablement symlink(s) for "coreos-metadata.service" Jul 9 23:56:16.666396 ignition[972]: INFO : files: op(11): [finished] setting preset to disabled for "coreos-metadata.service" Jul 9 23:56:16.666396 ignition[972]: INFO : files: op(13): [started] setting preset to enabled for "prepare-helm.service" Jul 9 23:56:16.666396 ignition[972]: INFO : files: op(13): [finished] setting preset to enabled for "prepare-helm.service" Jul 9 23:56:16.666396 ignition[972]: INFO : files: createResultFile: createFiles: op(14): [started] writing file "/sysroot/etc/.ignition-result.json" Jul 9 23:56:16.667540 ignition[972]: INFO : files: createResultFile: createFiles: op(14): [finished] writing file "/sysroot/etc/.ignition-result.json" Jul 9 23:56:16.667540 ignition[972]: INFO : files: files passed Jul 9 23:56:16.667540 ignition[972]: INFO : Ignition finished successfully Jul 9 23:56:16.667401 systemd[1]: Finished ignition-files.service - Ignition (files). Jul 9 23:56:16.678461 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Jul 9 23:56:16.680035 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Jul 9 23:56:16.680617 systemd[1]: ignition-quench.service: Deactivated successfully. Jul 9 23:56:16.680801 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Jul 9 23:56:16.685955 initrd-setup-root-after-ignition[1002]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jul 9 23:56:16.685955 initrd-setup-root-after-ignition[1002]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Jul 9 23:56:16.686651 initrd-setup-root-after-ignition[1006]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jul 9 23:56:16.687351 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Jul 9 23:56:16.687675 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Jul 9 23:56:16.690466 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Jul 9 23:56:16.702342 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Jul 9 23:56:16.702420 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Jul 9 23:56:16.702752 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Jul 9 23:56:16.702962 systemd[1]: Reached target initrd.target - Initrd Default Target. Jul 9 23:56:16.703160 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Jul 9 23:56:16.703615 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Jul 9 23:56:16.712177 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jul 9 23:56:16.714476 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Jul 9 23:56:16.720793 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Jul 9 23:56:16.720969 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Jul 9 23:56:16.721179 systemd[1]: Stopped target timers.target - Timer Units. Jul 9 23:56:16.721365 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Jul 9 23:56:16.721442 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jul 9 23:56:16.721687 systemd[1]: Stopped target initrd.target - Initrd Default Target. Jul 9 23:56:16.721926 systemd[1]: Stopped target basic.target - Basic System. Jul 9 23:56:16.722101 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Jul 9 23:56:16.722433 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Jul 9 23:56:16.722630 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Jul 9 23:56:16.722835 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Jul 9 23:56:16.723024 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Jul 9 23:56:16.723232 systemd[1]: Stopped target sysinit.target - System Initialization. Jul 9 23:56:16.723448 systemd[1]: Stopped target local-fs.target - Local File Systems. Jul 9 23:56:16.723603 systemd[1]: Stopped target swap.target - Swaps. Jul 9 23:56:16.723760 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Jul 9 23:56:16.723830 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Jul 9 23:56:16.724075 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Jul 9 23:56:16.724311 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jul 9 23:56:16.724504 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Jul 9 23:56:16.724545 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jul 9 23:56:16.724714 systemd[1]: dracut-initqueue.service: Deactivated successfully. Jul 9 23:56:16.724774 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Jul 9 23:56:16.725029 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Jul 9 23:56:16.725092 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Jul 9 23:56:16.725324 systemd[1]: Stopped target paths.target - Path Units. Jul 9 23:56:16.725492 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Jul 9 23:56:16.729394 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jul 9 23:56:16.729553 systemd[1]: Stopped target slices.target - Slice Units. Jul 9 23:56:16.729747 systemd[1]: Stopped target sockets.target - Socket Units. Jul 9 23:56:16.729930 systemd[1]: iscsid.socket: Deactivated successfully. Jul 9 23:56:16.729995 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Jul 9 23:56:16.730202 systemd[1]: iscsiuio.socket: Deactivated successfully. Jul 9 23:56:16.730246 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jul 9 23:56:16.730493 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Jul 9 23:56:16.730556 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Jul 9 23:56:16.730794 systemd[1]: ignition-files.service: Deactivated successfully. Jul 9 23:56:16.730851 systemd[1]: Stopped ignition-files.service - Ignition (files). Jul 9 23:56:16.737479 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Jul 9 23:56:16.738785 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Jul 9 23:56:16.738883 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Jul 9 23:56:16.738973 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Jul 9 23:56:16.739153 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Jul 9 23:56:16.739234 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Jul 9 23:56:16.742772 systemd[1]: initrd-cleanup.service: Deactivated successfully. Jul 9 23:56:16.742824 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Jul 9 23:56:16.746339 ignition[1026]: INFO : Ignition 2.20.0 Jul 9 23:56:16.746339 ignition[1026]: INFO : Stage: umount Jul 9 23:56:16.746966 ignition[1026]: INFO : no configs at "/usr/lib/ignition/base.d" Jul 9 23:56:16.746966 ignition[1026]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/vmware" Jul 9 23:56:16.747786 ignition[1026]: INFO : umount: umount passed Jul 9 23:56:16.747786 ignition[1026]: INFO : Ignition finished successfully Jul 9 23:56:16.748324 systemd[1]: ignition-mount.service: Deactivated successfully. Jul 9 23:56:16.748391 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Jul 9 23:56:16.749170 systemd[1]: Stopped target network.target - Network. Jul 9 23:56:16.749279 systemd[1]: ignition-disks.service: Deactivated successfully. Jul 9 23:56:16.749314 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Jul 9 23:56:16.749832 systemd[1]: ignition-kargs.service: Deactivated successfully. Jul 9 23:56:16.749858 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Jul 9 23:56:16.749988 systemd[1]: ignition-setup.service: Deactivated successfully. Jul 9 23:56:16.750012 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Jul 9 23:56:16.750617 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Jul 9 23:56:16.750655 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Jul 9 23:56:16.750851 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Jul 9 23:56:16.751102 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Jul 9 23:56:16.753240 systemd[1]: sysroot-boot.mount: Deactivated successfully. Jul 9 23:56:16.756314 systemd[1]: systemd-resolved.service: Deactivated successfully. Jul 9 23:56:16.756384 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Jul 9 23:56:16.757736 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Jul 9 23:56:16.757859 systemd[1]: systemd-networkd.service: Deactivated successfully. Jul 9 23:56:16.757955 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Jul 9 23:56:16.758632 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Jul 9 23:56:16.759073 systemd[1]: systemd-networkd.socket: Deactivated successfully. Jul 9 23:56:16.759101 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Jul 9 23:56:16.762551 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Jul 9 23:56:16.762637 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Jul 9 23:56:16.762665 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jul 9 23:56:16.762786 systemd[1]: afterburn-network-kargs.service: Deactivated successfully. Jul 9 23:56:16.762809 systemd[1]: Stopped afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments. Jul 9 23:56:16.762920 systemd[1]: systemd-sysctl.service: Deactivated successfully. Jul 9 23:56:16.762942 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Jul 9 23:56:16.763106 systemd[1]: systemd-modules-load.service: Deactivated successfully. Jul 9 23:56:16.763130 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Jul 9 23:56:16.763235 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Jul 9 23:56:16.763257 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Jul 9 23:56:16.763502 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Jul 9 23:56:16.764477 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Jul 9 23:56:16.764514 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Jul 9 23:56:16.768797 systemd[1]: network-cleanup.service: Deactivated successfully. Jul 9 23:56:16.768855 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Jul 9 23:56:16.776780 systemd[1]: systemd-udevd.service: Deactivated successfully. Jul 9 23:56:16.776862 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Jul 9 23:56:16.777166 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Jul 9 23:56:16.777191 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Jul 9 23:56:16.777405 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Jul 9 23:56:16.777423 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Jul 9 23:56:16.777757 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Jul 9 23:56:16.777783 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Jul 9 23:56:16.778038 systemd[1]: dracut-cmdline.service: Deactivated successfully. Jul 9 23:56:16.778062 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Jul 9 23:56:16.778331 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jul 9 23:56:16.778354 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jul 9 23:56:16.782485 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Jul 9 23:56:16.782616 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Jul 9 23:56:16.782643 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jul 9 23:56:16.782890 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Jul 9 23:56:16.782914 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jul 9 23:56:16.783029 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Jul 9 23:56:16.783051 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Jul 9 23:56:16.783162 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jul 9 23:56:16.783183 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jul 9 23:56:16.783893 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Jul 9 23:56:16.783926 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Jul 9 23:56:16.786318 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Jul 9 23:56:16.786394 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Jul 9 23:56:16.920204 systemd[1]: sysroot-boot.service: Deactivated successfully. Jul 9 23:56:16.920271 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Jul 9 23:56:16.920556 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Jul 9 23:56:16.920686 systemd[1]: initrd-setup-root.service: Deactivated successfully. Jul 9 23:56:16.920716 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Jul 9 23:56:16.924465 systemd[1]: Starting initrd-switch-root.service - Switch Root... Jul 9 23:56:16.934003 systemd[1]: Switching root. Jul 9 23:56:16.968258 systemd-journald[215]: Journal stopped Jul 9 23:56:18.159511 systemd-journald[215]: Received SIGTERM from PID 1 (systemd). Jul 9 23:56:18.159537 kernel: SELinux: policy capability network_peer_controls=1 Jul 9 23:56:18.159545 kernel: SELinux: policy capability open_perms=1 Jul 9 23:56:18.159551 kernel: SELinux: policy capability extended_socket_class=1 Jul 9 23:56:18.159556 kernel: SELinux: policy capability always_check_network=0 Jul 9 23:56:18.159562 kernel: SELinux: policy capability cgroup_seclabel=1 Jul 9 23:56:18.159569 kernel: SELinux: policy capability nnp_nosuid_transition=1 Jul 9 23:56:18.159575 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Jul 9 23:56:18.159581 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Jul 9 23:56:18.159587 systemd[1]: Successfully loaded SELinux policy in 31.325ms. Jul 9 23:56:18.159594 kernel: audit: type=1403 audit(1752105377.611:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Jul 9 23:56:18.159600 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 7.051ms. Jul 9 23:56:18.159607 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Jul 9 23:56:18.159615 systemd[1]: Detected virtualization vmware. Jul 9 23:56:18.159622 systemd[1]: Detected architecture x86-64. Jul 9 23:56:18.159628 systemd[1]: Detected first boot. Jul 9 23:56:18.159635 systemd[1]: Initializing machine ID from random generator. Jul 9 23:56:18.159643 zram_generator::config[1072]: No configuration found. Jul 9 23:56:18.159726 kernel: vmw_vmci 0000:00:07.7: Using capabilities 0xc Jul 9 23:56:18.159737 kernel: Guest personality initialized and is active Jul 9 23:56:18.159744 kernel: VMCI host device registered (name=vmci, major=10, minor=125) Jul 9 23:56:18.159750 kernel: Initialized host personality Jul 9 23:56:18.159756 kernel: NET: Registered PF_VSOCK protocol family Jul 9 23:56:18.159762 systemd[1]: Populated /etc with preset unit settings. Jul 9 23:56:18.159772 systemd[1]: /etc/systemd/system/coreos-metadata.service:11: Ignoring unknown escape sequences: "echo "COREOS_CUSTOM_PRIVATE_IPV4=$(ip addr show ens192 | grep "inet 10." | grep -Po "inet \K[\d.]+") Jul 9 23:56:18.159779 systemd[1]: COREOS_CUSTOM_PUBLIC_IPV4=$(ip addr show ens192 | grep -v "inet 10." | grep -Po "inet \K[\d.]+")" > ${OUTPUT}" Jul 9 23:56:18.159786 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Jul 9 23:56:18.159793 systemd[1]: initrd-switch-root.service: Deactivated successfully. Jul 9 23:56:18.159799 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Jul 9 23:56:18.159806 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Jul 9 23:56:18.159814 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Jul 9 23:56:18.159821 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Jul 9 23:56:18.159828 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Jul 9 23:56:18.159834 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Jul 9 23:56:18.159841 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Jul 9 23:56:18.159847 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Jul 9 23:56:18.159854 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Jul 9 23:56:18.159861 systemd[1]: Created slice user.slice - User and Session Slice. Jul 9 23:56:18.159869 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jul 9 23:56:18.159876 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jul 9 23:56:18.159885 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Jul 9 23:56:18.159892 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Jul 9 23:56:18.159899 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Jul 9 23:56:18.159906 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jul 9 23:56:18.159912 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Jul 9 23:56:18.159919 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jul 9 23:56:18.159927 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Jul 9 23:56:18.159934 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Jul 9 23:56:18.159941 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Jul 9 23:56:18.159948 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Jul 9 23:56:18.159955 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jul 9 23:56:18.159962 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jul 9 23:56:18.159969 systemd[1]: Reached target slices.target - Slice Units. Jul 9 23:56:18.159975 systemd[1]: Reached target swap.target - Swaps. Jul 9 23:56:18.159983 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Jul 9 23:56:18.159990 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Jul 9 23:56:18.159997 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Jul 9 23:56:18.160004 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jul 9 23:56:18.160011 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jul 9 23:56:18.160019 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jul 9 23:56:18.160026 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Jul 9 23:56:18.160033 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Jul 9 23:56:18.160040 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Jul 9 23:56:18.160047 systemd[1]: Mounting media.mount - External Media Directory... Jul 9 23:56:18.160054 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jul 9 23:56:18.160061 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Jul 9 23:56:18.160068 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Jul 9 23:56:18.160076 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Jul 9 23:56:18.160085 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Jul 9 23:56:18.160092 systemd[1]: Reached target machines.target - Containers. Jul 9 23:56:18.160099 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Jul 9 23:56:18.160106 systemd[1]: Starting ignition-delete-config.service - Ignition (delete config)... Jul 9 23:56:18.160113 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jul 9 23:56:18.160119 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Jul 9 23:56:18.160126 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jul 9 23:56:18.160134 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jul 9 23:56:18.160142 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jul 9 23:56:18.160148 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Jul 9 23:56:18.160155 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jul 9 23:56:18.160162 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Jul 9 23:56:18.160169 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Jul 9 23:56:18.160176 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Jul 9 23:56:18.160183 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Jul 9 23:56:18.160190 systemd[1]: Stopped systemd-fsck-usr.service. Jul 9 23:56:18.160198 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Jul 9 23:56:18.160205 kernel: fuse: init (API version 7.39) Jul 9 23:56:18.160212 systemd[1]: Starting systemd-journald.service - Journal Service... Jul 9 23:56:18.160218 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jul 9 23:56:18.160225 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Jul 9 23:56:18.160233 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Jul 9 23:56:18.160240 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Jul 9 23:56:18.160247 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jul 9 23:56:18.160255 systemd[1]: verity-setup.service: Deactivated successfully. Jul 9 23:56:18.160262 systemd[1]: Stopped verity-setup.service. Jul 9 23:56:18.160269 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jul 9 23:56:18.160276 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Jul 9 23:56:18.160283 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Jul 9 23:56:18.160290 systemd[1]: Mounted media.mount - External Media Directory. Jul 9 23:56:18.160297 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Jul 9 23:56:18.160304 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Jul 9 23:56:18.160311 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Jul 9 23:56:18.160332 systemd-journald[1162]: Collecting audit messages is disabled. Jul 9 23:56:18.160349 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jul 9 23:56:18.160356 systemd[1]: modprobe@configfs.service: Deactivated successfully. Jul 9 23:56:18.160363 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Jul 9 23:56:18.160421 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jul 9 23:56:18.160432 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jul 9 23:56:18.160439 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jul 9 23:56:18.160446 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jul 9 23:56:18.160453 systemd[1]: modprobe@fuse.service: Deactivated successfully. Jul 9 23:56:18.160460 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Jul 9 23:56:18.160468 systemd-journald[1162]: Journal started Jul 9 23:56:18.160484 systemd-journald[1162]: Runtime Journal (/run/log/journal/71dadb66a4ac4100b7b4726d92269bb0) is 4.8M, max 38.6M, 33.8M free. Jul 9 23:56:18.003647 systemd[1]: Queued start job for default target multi-user.target. Jul 9 23:56:18.014508 systemd[1]: Unnecessary job was removed for dev-sda6.device - /dev/sda6. Jul 9 23:56:18.014740 systemd[1]: systemd-journald.service: Deactivated successfully. Jul 9 23:56:18.160973 jq[1142]: true Jul 9 23:56:18.162392 kernel: loop: module loaded Jul 9 23:56:18.162434 systemd[1]: Started systemd-journald.service - Journal Service. Jul 9 23:56:18.167904 systemd[1]: modprobe@loop.service: Deactivated successfully. Jul 9 23:56:18.168033 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jul 9 23:56:18.168295 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jul 9 23:56:18.168549 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Jul 9 23:56:18.168797 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Jul 9 23:56:18.177338 systemd[1]: Reached target network-pre.target - Preparation for Network. Jul 9 23:56:18.179492 jq[1181]: true Jul 9 23:56:18.182444 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Jul 9 23:56:18.186443 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Jul 9 23:56:18.186570 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Jul 9 23:56:18.186589 systemd[1]: Reached target local-fs.target - Local File Systems. Jul 9 23:56:18.187248 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Jul 9 23:56:18.189486 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Jul 9 23:56:18.191302 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Jul 9 23:56:18.191488 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jul 9 23:56:18.192202 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Jul 9 23:56:18.195798 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Jul 9 23:56:18.195931 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jul 9 23:56:18.196802 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Jul 9 23:56:18.196928 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jul 9 23:56:18.199362 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jul 9 23:56:18.200518 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Jul 9 23:56:18.202693 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Jul 9 23:56:18.204418 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Jul 9 23:56:18.205535 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Jul 9 23:56:18.205687 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Jul 9 23:56:18.206042 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Jul 9 23:56:18.212522 systemd-journald[1162]: Time spent on flushing to /var/log/journal/71dadb66a4ac4100b7b4726d92269bb0 is 17.183ms for 1849 entries. Jul 9 23:56:18.212522 systemd-journald[1162]: System Journal (/var/log/journal/71dadb66a4ac4100b7b4726d92269bb0) is 8M, max 584.8M, 576.8M free. Jul 9 23:56:18.239285 systemd-journald[1162]: Received client request to flush runtime journal. Jul 9 23:56:18.239308 kernel: ACPI: bus type drm_connector registered Jul 9 23:56:18.213729 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Jul 9 23:56:18.232016 systemd[1]: modprobe@drm.service: Deactivated successfully. Jul 9 23:56:18.236026 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jul 9 23:56:18.239946 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Jul 9 23:56:18.242291 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Jul 9 23:56:18.242547 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Jul 9 23:56:18.257696 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Jul 9 23:56:18.258930 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jul 9 23:56:18.269386 kernel: loop0: detected capacity change from 0 to 229808 Jul 9 23:56:18.333639 systemd-tmpfiles[1214]: ACLs are not supported, ignoring. Jul 9 23:56:18.333652 systemd-tmpfiles[1214]: ACLs are not supported, ignoring. Jul 9 23:56:18.338921 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Jul 9 23:56:18.340968 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jul 9 23:56:18.341623 ignition[1215]: Ignition 2.20.0 Jul 9 23:56:18.343972 ignition[1215]: deleting config from guestinfo properties Jul 9 23:56:18.349640 systemd[1]: Starting systemd-sysusers.service - Create System Users... Jul 9 23:56:18.350413 ignition[1215]: Successfully deleted config Jul 9 23:56:18.356113 systemd[1]: Finished ignition-delete-config.service - Ignition (delete config). Jul 9 23:56:18.360578 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jul 9 23:56:18.363512 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Jul 9 23:56:18.368938 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Jul 9 23:56:18.382686 udevadm[1246]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Jul 9 23:56:18.392531 kernel: loop1: detected capacity change from 0 to 147912 Jul 9 23:56:18.401866 systemd[1]: Finished systemd-sysusers.service - Create System Users. Jul 9 23:56:18.407656 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jul 9 23:56:18.423783 systemd-tmpfiles[1248]: ACLs are not supported, ignoring. Jul 9 23:56:18.423954 systemd-tmpfiles[1248]: ACLs are not supported, ignoring. Jul 9 23:56:18.427380 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jul 9 23:56:18.438406 kernel: loop2: detected capacity change from 0 to 138176 Jul 9 23:56:18.483689 kernel: loop3: detected capacity change from 0 to 2960 Jul 9 23:56:18.510385 kernel: loop4: detected capacity change from 0 to 229808 Jul 9 23:56:18.536023 kernel: loop5: detected capacity change from 0 to 147912 Jul 9 23:56:18.559412 kernel: loop6: detected capacity change from 0 to 138176 Jul 9 23:56:18.577467 kernel: loop7: detected capacity change from 0 to 2960 Jul 9 23:56:18.592909 (sd-merge)[1255]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-vmware'. Jul 9 23:56:18.593177 (sd-merge)[1255]: Merged extensions into '/usr'. Jul 9 23:56:18.596396 systemd[1]: Reload requested from client PID 1213 ('systemd-sysext') (unit systemd-sysext.service)... Jul 9 23:56:18.596408 systemd[1]: Reloading... Jul 9 23:56:18.652630 zram_generator::config[1284]: No configuration found. Jul 9 23:56:18.787657 systemd[1]: /etc/systemd/system/coreos-metadata.service:11: Ignoring unknown escape sequences: "echo "COREOS_CUSTOM_PRIVATE_IPV4=$(ip addr show ens192 | grep "inet 10." | grep -Po "inet \K[\d.]+") Jul 9 23:56:18.806860 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jul 9 23:56:18.848592 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Jul 9 23:56:18.848759 systemd[1]: Reloading finished in 252 ms. Jul 9 23:56:18.850817 ldconfig[1208]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Jul 9 23:56:18.860093 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Jul 9 23:56:18.860485 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Jul 9 23:56:18.867563 systemd[1]: Starting ensure-sysext.service... Jul 9 23:56:18.870181 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jul 9 23:56:18.875915 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Jul 9 23:56:18.881481 systemd-tmpfiles[1341]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Jul 9 23:56:18.881732 systemd-tmpfiles[1341]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Jul 9 23:56:18.882253 systemd-tmpfiles[1341]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Jul 9 23:56:18.882430 systemd-tmpfiles[1341]: ACLs are not supported, ignoring. Jul 9 23:56:18.882469 systemd-tmpfiles[1341]: ACLs are not supported, ignoring. Jul 9 23:56:18.884181 systemd-tmpfiles[1341]: Detected autofs mount point /boot during canonicalization of boot. Jul 9 23:56:18.884185 systemd-tmpfiles[1341]: Skipping /boot Jul 9 23:56:18.884564 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jul 9 23:56:18.884746 systemd[1]: Reload requested from client PID 1340 ('systemctl') (unit ensure-sysext.service)... Jul 9 23:56:18.884755 systemd[1]: Reloading... Jul 9 23:56:18.896078 systemd-tmpfiles[1341]: Detected autofs mount point /boot during canonicalization of boot. Jul 9 23:56:18.896084 systemd-tmpfiles[1341]: Skipping /boot Jul 9 23:56:18.912014 systemd-udevd[1344]: Using default interface naming scheme 'v255'. Jul 9 23:56:18.925461 zram_generator::config[1370]: No configuration found. Jul 9 23:56:19.032940 systemd[1]: /etc/systemd/system/coreos-metadata.service:11: Ignoring unknown escape sequences: "echo "COREOS_CUSTOM_PRIVATE_IPV4=$(ip addr show ens192 | grep "inet 10." | grep -Po "inet \K[\d.]+") Jul 9 23:56:19.036388 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 37 scanned by (udev-worker) (1398) Jul 9 23:56:19.047397 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input2 Jul 9 23:56:19.058183 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jul 9 23:56:19.062415 kernel: ACPI: button: Power Button [PWRF] Jul 9 23:56:19.122742 systemd[1]: Condition check resulted in dev-ttyS0.device - /dev/ttyS0 being skipped. Jul 9 23:56:19.123168 systemd[1]: Reloading finished in 238 ms. Jul 9 23:56:19.130643 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jul 9 23:56:19.137554 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jul 9 23:56:19.145380 kernel: piix4_smbus 0000:00:07.3: SMBus Host Controller not enabled! Jul 9 23:56:19.160099 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Virtual_disk OEM. Jul 9 23:56:19.161289 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jul 9 23:56:19.165481 systemd[1]: Starting audit-rules.service - Load Audit Rules... Jul 9 23:56:19.166984 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Jul 9 23:56:19.167688 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jul 9 23:56:19.175629 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jul 9 23:56:19.177231 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jul 9 23:56:19.178691 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jul 9 23:56:19.178882 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jul 9 23:56:19.182414 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Jul 9 23:56:19.182625 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Jul 9 23:56:19.184052 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Jul 9 23:56:19.186130 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jul 9 23:56:19.188265 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jul 9 23:56:19.197387 kernel: input: ImPS/2 Generic Wheel Mouse as /devices/platform/i8042/serio1/input/input3 Jul 9 23:56:19.194346 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Jul 9 23:56:19.194487 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jul 9 23:56:19.195144 systemd[1]: Finished ensure-sysext.service. Jul 9 23:56:19.196039 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jul 9 23:56:19.196144 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jul 9 23:56:19.196422 systemd[1]: modprobe@drm.service: Deactivated successfully. Jul 9 23:56:19.196521 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jul 9 23:56:19.210497 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Jul 9 23:56:19.210864 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jul 9 23:56:19.210986 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jul 9 23:56:19.211260 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Jul 9 23:56:19.211552 systemd[1]: modprobe@loop.service: Deactivated successfully. Jul 9 23:56:19.211648 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jul 9 23:56:19.216464 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Jul 9 23:56:19.216864 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jul 9 23:56:19.216907 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jul 9 23:56:19.219085 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Jul 9 23:56:19.249616 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jul 9 23:56:19.272246 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Jul 9 23:56:19.277384 kernel: mousedev: PS/2 mouse device common for all mice Jul 9 23:56:19.278774 (udev-worker)[1382]: id: Truncating stdout of 'dmi_memory_id' up to 16384 byte. Jul 9 23:56:19.280613 systemd[1]: Starting systemd-update-done.service - Update is Completed... Jul 9 23:56:19.280757 systemd[1]: Started systemd-userdbd.service - User Database Manager. Jul 9 23:56:19.291756 augenrules[1506]: No rules Jul 9 23:56:19.292476 systemd[1]: audit-rules.service: Deactivated successfully. Jul 9 23:56:19.292623 systemd[1]: Finished audit-rules.service - Load Audit Rules. Jul 9 23:56:19.298605 systemd[1]: Finished systemd-update-done.service - Update is Completed. Jul 9 23:56:19.312696 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Jul 9 23:56:19.313143 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Jul 9 23:56:19.317096 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Jul 9 23:56:19.321678 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Jul 9 23:56:19.333481 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jul 9 23:56:19.337456 lvm[1519]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Jul 9 23:56:19.354897 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Jul 9 23:56:19.355082 systemd[1]: Reached target time-set.target - System Time Set. Jul 9 23:56:19.356223 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Jul 9 23:56:19.356466 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jul 9 23:56:19.364438 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Jul 9 23:56:19.368323 lvm[1524]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Jul 9 23:56:19.369913 systemd-networkd[1465]: lo: Link UP Jul 9 23:56:19.370122 systemd-networkd[1465]: lo: Gained carrier Jul 9 23:56:19.370962 systemd-networkd[1465]: Enumeration completed Jul 9 23:56:19.371050 systemd[1]: Started systemd-networkd.service - Network Configuration. Jul 9 23:56:19.371260 systemd-networkd[1465]: ens192: Configuring with /etc/systemd/network/00-vmware.network. Jul 9 23:56:19.375462 kernel: vmxnet3 0000:0b:00.0 ens192: intr type 3, mode 0, 3 vectors allocated Jul 9 23:56:19.375602 kernel: vmxnet3 0000:0b:00.0 ens192: NIC Link is Up 10000 Mbps Jul 9 23:56:19.375465 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Jul 9 23:56:19.376248 systemd-networkd[1465]: ens192: Link UP Jul 9 23:56:19.376821 systemd-networkd[1465]: ens192: Gained carrier Jul 9 23:56:19.379468 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Jul 9 23:56:19.386535 systemd-timesyncd[1475]: Network configuration changed, trying to establish connection. Jul 9 23:56:19.395058 systemd-resolved[1467]: Positive Trust Anchors: Jul 9 23:56:19.395684 systemd-resolved[1467]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jul 9 23:56:19.395751 systemd-resolved[1467]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jul 9 23:56:19.396208 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Jul 9 23:56:19.398910 systemd-resolved[1467]: Defaulting to hostname 'linux'. Jul 9 23:56:19.400025 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jul 9 23:56:19.400174 systemd[1]: Reached target network.target - Network. Jul 9 23:56:19.400258 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jul 9 23:56:19.400365 systemd[1]: Reached target sysinit.target - System Initialization. Jul 9 23:56:19.400525 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Jul 9 23:56:19.400643 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Jul 9 23:56:19.400833 systemd[1]: Started logrotate.timer - Daily rotation of log files. Jul 9 23:56:19.400973 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Jul 9 23:56:19.401077 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Jul 9 23:56:19.401177 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Jul 9 23:56:19.401199 systemd[1]: Reached target paths.target - Path Units. Jul 9 23:56:19.401277 systemd[1]: Reached target timers.target - Timer Units. Jul 9 23:56:19.401846 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Jul 9 23:56:19.403358 systemd[1]: Starting docker.socket - Docker Socket for the API... Jul 9 23:56:19.404899 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Jul 9 23:56:19.405088 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Jul 9 23:56:19.405199 systemd[1]: Reached target ssh-access.target - SSH Access Available. Jul 9 23:56:19.407427 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Jul 9 23:56:19.407969 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Jul 9 23:56:19.408553 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Jul 9 23:56:19.408743 systemd[1]: Listening on docker.socket - Docker Socket for the API. Jul 9 23:56:19.409093 systemd[1]: Reached target sockets.target - Socket Units. Jul 9 23:56:19.409190 systemd[1]: Reached target basic.target - Basic System. Jul 9 23:56:19.409313 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Jul 9 23:56:19.409333 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Jul 9 23:56:19.410086 systemd[1]: Starting containerd.service - containerd container runtime... Jul 9 23:56:19.412584 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Jul 9 23:56:19.414456 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Jul 9 23:56:19.416391 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Jul 9 23:56:19.417223 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Jul 9 23:56:19.418810 jq[1534]: false Jul 9 23:56:19.418857 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Jul 9 23:56:19.420652 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Jul 9 23:56:19.424072 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Jul 9 23:56:19.426684 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Jul 9 23:56:19.434599 systemd[1]: Starting systemd-logind.service - User Login Management... Jul 9 23:56:19.435198 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Jul 9 23:56:19.435698 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Jul 9 23:56:19.438474 systemd[1]: Starting update-engine.service - Update Engine... Jul 9 23:56:19.440434 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Jul 9 23:56:19.443910 extend-filesystems[1535]: Found loop4 Jul 9 23:56:19.443910 extend-filesystems[1535]: Found loop5 Jul 9 23:56:19.443910 extend-filesystems[1535]: Found loop6 Jul 9 23:56:19.443910 extend-filesystems[1535]: Found loop7 Jul 9 23:56:19.443910 extend-filesystems[1535]: Found sda Jul 9 23:56:19.443910 extend-filesystems[1535]: Found sda1 Jul 9 23:56:19.443910 extend-filesystems[1535]: Found sda2 Jul 9 23:56:19.443910 extend-filesystems[1535]: Found sda3 Jul 9 23:56:19.443910 extend-filesystems[1535]: Found usr Jul 9 23:56:19.443910 extend-filesystems[1535]: Found sda4 Jul 9 23:56:19.443910 extend-filesystems[1535]: Found sda6 Jul 9 23:56:19.443910 extend-filesystems[1535]: Found sda7 Jul 9 23:56:19.443910 extend-filesystems[1535]: Found sda9 Jul 9 23:56:19.443910 extend-filesystems[1535]: Checking size of /dev/sda9 Jul 9 23:56:19.443430 systemd[1]: Starting vgauthd.service - VGAuth Service for open-vm-tools... Jul 9 23:56:19.446177 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Jul 9 23:56:19.458698 jq[1549]: true Jul 9 23:56:19.446506 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Jul 9 23:56:19.460719 dbus-daemon[1533]: [system] SELinux support is enabled Jul 9 23:56:19.459748 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Jul 9 23:56:19.460278 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Jul 9 23:56:19.460731 systemd[1]: motdgen.service: Deactivated successfully. Jul 9 23:56:19.461412 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Jul 9 23:56:19.462194 systemd[1]: Started dbus.service - D-Bus System Message Bus. Jul 9 23:56:19.470689 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Jul 9 23:56:19.471449 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Jul 9 23:56:19.471627 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Jul 9 23:56:19.471637 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Jul 9 23:56:19.472435 (ntainerd)[1562]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Jul 9 23:56:19.481869 jq[1554]: true Jul 9 23:56:19.494180 update_engine[1548]: I20250709 23:56:19.493434 1548 main.cc:92] Flatcar Update Engine starting Jul 9 23:56:19.494342 tar[1553]: linux-amd64/LICENSE Jul 9 23:56:19.494342 tar[1553]: linux-amd64/helm Jul 9 23:56:19.494492 extend-filesystems[1535]: Old size kept for /dev/sda9 Jul 9 23:56:19.494492 extend-filesystems[1535]: Found sr0 Jul 9 23:56:19.494756 systemd[1]: extend-filesystems.service: Deactivated successfully. Jul 9 23:56:19.494893 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Jul 9 23:56:19.502608 systemd-logind[1542]: Watching system buttons on /dev/input/event1 (Power Button) Jul 9 23:56:19.502621 systemd-logind[1542]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Jul 9 23:56:19.509531 update_engine[1548]: I20250709 23:56:19.504142 1548 update_check_scheduler.cc:74] Next update check in 6m3s Jul 9 23:56:19.504552 systemd-logind[1542]: New seat seat0. Jul 9 23:56:19.507521 systemd[1]: Started vgauthd.service - VGAuth Service for open-vm-tools. Jul 9 23:56:19.507713 systemd[1]: Started systemd-logind.service - User Login Management. Jul 9 23:56:19.507942 systemd[1]: Started update-engine.service - Update Engine. Jul 9 23:56:19.514471 systemd[1]: Starting vmtoolsd.service - Service for virtual machines hosted on VMware... Jul 9 23:56:19.517727 systemd[1]: Started locksmithd.service - Cluster reboot manager. Jul 9 23:56:19.557455 systemd[1]: Started vmtoolsd.service - Service for virtual machines hosted on VMware. Jul 9 23:56:19.566552 unknown[1575]: Pref_Init: Using '/etc/vmware-tools/vgauth.conf' as preferences filepath Jul 9 23:56:19.571242 unknown[1575]: Core dump limit set to -1 Jul 9 23:56:19.595422 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 37 scanned by (udev-worker) (1383) Jul 9 23:56:19.596169 bash[1592]: Updated "/home/core/.ssh/authorized_keys" Jul 9 23:56:19.596981 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Jul 9 23:56:19.597766 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Jul 9 23:56:19.744071 locksmithd[1578]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Jul 9 23:56:19.776616 containerd[1562]: time="2025-07-09T23:56:19.776580342Z" level=info msg="starting containerd" revision=9b2ad7760328148397346d10c7b2004271249db4 version=v1.7.23 Jul 9 23:56:19.808980 containerd[1562]: time="2025-07-09T23:56:19.808955396Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Jul 9 23:56:19.812621 containerd[1562]: time="2025-07-09T23:56:19.811507437Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.95-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Jul 9 23:56:19.812621 containerd[1562]: time="2025-07-09T23:56:19.811524102Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Jul 9 23:56:19.812621 containerd[1562]: time="2025-07-09T23:56:19.811533812Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Jul 9 23:56:19.812621 containerd[1562]: time="2025-07-09T23:56:19.811613656Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Jul 9 23:56:19.812621 containerd[1562]: time="2025-07-09T23:56:19.811623644Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Jul 9 23:56:19.812621 containerd[1562]: time="2025-07-09T23:56:19.811660639Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Jul 9 23:56:19.812621 containerd[1562]: time="2025-07-09T23:56:19.811668800Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Jul 9 23:56:19.812621 containerd[1562]: time="2025-07-09T23:56:19.811769411Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Jul 9 23:56:19.812621 containerd[1562]: time="2025-07-09T23:56:19.811778084Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Jul 9 23:56:19.812621 containerd[1562]: time="2025-07-09T23:56:19.811785045Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Jul 9 23:56:19.812621 containerd[1562]: time="2025-07-09T23:56:19.811790160Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Jul 9 23:56:19.812780 containerd[1562]: time="2025-07-09T23:56:19.811837843Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Jul 9 23:56:19.812780 containerd[1562]: time="2025-07-09T23:56:19.811951921Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Jul 9 23:56:19.812780 containerd[1562]: time="2025-07-09T23:56:19.812017700Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Jul 9 23:56:19.812780 containerd[1562]: time="2025-07-09T23:56:19.812025400Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Jul 9 23:56:19.812780 containerd[1562]: time="2025-07-09T23:56:19.812067167Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Jul 9 23:56:19.812780 containerd[1562]: time="2025-07-09T23:56:19.812092937Z" level=info msg="metadata content store policy set" policy=shared Jul 9 23:56:19.815887 containerd[1562]: time="2025-07-09T23:56:19.815303012Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Jul 9 23:56:19.815887 containerd[1562]: time="2025-07-09T23:56:19.815325284Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Jul 9 23:56:19.815887 containerd[1562]: time="2025-07-09T23:56:19.815335431Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Jul 9 23:56:19.815887 containerd[1562]: time="2025-07-09T23:56:19.815344276Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Jul 9 23:56:19.815887 containerd[1562]: time="2025-07-09T23:56:19.815351815Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Jul 9 23:56:19.815887 containerd[1562]: time="2025-07-09T23:56:19.815430221Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Jul 9 23:56:19.815887 containerd[1562]: time="2025-07-09T23:56:19.815543234Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Jul 9 23:56:19.815887 containerd[1562]: time="2025-07-09T23:56:19.815597563Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Jul 9 23:56:19.815887 containerd[1562]: time="2025-07-09T23:56:19.815610468Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Jul 9 23:56:19.815887 containerd[1562]: time="2025-07-09T23:56:19.815618198Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Jul 9 23:56:19.815887 containerd[1562]: time="2025-07-09T23:56:19.815625604Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Jul 9 23:56:19.815887 containerd[1562]: time="2025-07-09T23:56:19.815632503Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Jul 9 23:56:19.815887 containerd[1562]: time="2025-07-09T23:56:19.815638511Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Jul 9 23:56:19.815887 containerd[1562]: time="2025-07-09T23:56:19.815645727Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Jul 9 23:56:19.816080 containerd[1562]: time="2025-07-09T23:56:19.815653153Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Jul 9 23:56:19.816080 containerd[1562]: time="2025-07-09T23:56:19.815660179Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Jul 9 23:56:19.816080 containerd[1562]: time="2025-07-09T23:56:19.815666651Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Jul 9 23:56:19.816080 containerd[1562]: time="2025-07-09T23:56:19.815673282Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Jul 9 23:56:19.816080 containerd[1562]: time="2025-07-09T23:56:19.815684493Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Jul 9 23:56:19.816080 containerd[1562]: time="2025-07-09T23:56:19.815692364Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Jul 9 23:56:19.816080 containerd[1562]: time="2025-07-09T23:56:19.815698985Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Jul 9 23:56:19.816080 containerd[1562]: time="2025-07-09T23:56:19.815705570Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Jul 9 23:56:19.816080 containerd[1562]: time="2025-07-09T23:56:19.815712046Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Jul 9 23:56:19.816080 containerd[1562]: time="2025-07-09T23:56:19.815721283Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Jul 9 23:56:19.816080 containerd[1562]: time="2025-07-09T23:56:19.815729215Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Jul 9 23:56:19.816080 containerd[1562]: time="2025-07-09T23:56:19.815736012Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Jul 9 23:56:19.816080 containerd[1562]: time="2025-07-09T23:56:19.815742884Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Jul 9 23:56:19.816080 containerd[1562]: time="2025-07-09T23:56:19.815751643Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Jul 9 23:56:19.816256 containerd[1562]: time="2025-07-09T23:56:19.815758175Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Jul 9 23:56:19.816256 containerd[1562]: time="2025-07-09T23:56:19.815764252Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Jul 9 23:56:19.816256 containerd[1562]: time="2025-07-09T23:56:19.815770801Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Jul 9 23:56:19.816256 containerd[1562]: time="2025-07-09T23:56:19.815778096Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Jul 9 23:56:19.816256 containerd[1562]: time="2025-07-09T23:56:19.815797528Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Jul 9 23:56:19.816256 containerd[1562]: time="2025-07-09T23:56:19.815805515Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Jul 9 23:56:19.816256 containerd[1562]: time="2025-07-09T23:56:19.815811258Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Jul 9 23:56:19.818117 containerd[1562]: time="2025-07-09T23:56:19.817687414Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Jul 9 23:56:19.818117 containerd[1562]: time="2025-07-09T23:56:19.817710086Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Jul 9 23:56:19.818117 containerd[1562]: time="2025-07-09T23:56:19.817718161Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Jul 9 23:56:19.818117 containerd[1562]: time="2025-07-09T23:56:19.817726091Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Jul 9 23:56:19.818117 containerd[1562]: time="2025-07-09T23:56:19.817731492Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Jul 9 23:56:19.818117 containerd[1562]: time="2025-07-09T23:56:19.817738586Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Jul 9 23:56:19.818117 containerd[1562]: time="2025-07-09T23:56:19.817744017Z" level=info msg="NRI interface is disabled by configuration." Jul 9 23:56:19.818117 containerd[1562]: time="2025-07-09T23:56:19.817750190Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Jul 9 23:56:19.818242 containerd[1562]: time="2025-07-09T23:56:19.817915954Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Jul 9 23:56:19.818242 containerd[1562]: time="2025-07-09T23:56:19.817943779Z" level=info msg="Connect containerd service" Jul 9 23:56:19.818242 containerd[1562]: time="2025-07-09T23:56:19.817960752Z" level=info msg="using legacy CRI server" Jul 9 23:56:19.818242 containerd[1562]: time="2025-07-09T23:56:19.817964900Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Jul 9 23:56:19.818242 containerd[1562]: time="2025-07-09T23:56:19.818025436Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Jul 9 23:56:19.821556 containerd[1562]: time="2025-07-09T23:56:19.821539015Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Jul 9 23:56:19.822571 containerd[1562]: time="2025-07-09T23:56:19.822554582Z" level=info msg="Start subscribing containerd event" Jul 9 23:56:19.822746 containerd[1562]: time="2025-07-09T23:56:19.822734483Z" level=info msg="Start recovering state" Jul 9 23:56:19.823170 containerd[1562]: time="2025-07-09T23:56:19.823157351Z" level=info msg="Start event monitor" Jul 9 23:56:19.823240 containerd[1562]: time="2025-07-09T23:56:19.823233241Z" level=info msg="Start snapshots syncer" Jul 9 23:56:19.823277 containerd[1562]: time="2025-07-09T23:56:19.823270323Z" level=info msg="Start cni network conf syncer for default" Jul 9 23:56:19.823329 containerd[1562]: time="2025-07-09T23:56:19.823319749Z" level=info msg="Start streaming server" Jul 9 23:56:19.824082 containerd[1562]: time="2025-07-09T23:56:19.822811847Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Jul 9 23:56:19.824082 containerd[1562]: time="2025-07-09T23:56:19.823579039Z" level=info msg=serving... address=/run/containerd/containerd.sock Jul 9 23:56:19.823653 systemd[1]: Started containerd.service - containerd container runtime. Jul 9 23:56:19.825891 containerd[1562]: time="2025-07-09T23:56:19.824174499Z" level=info msg="containerd successfully booted in 0.048971s" Jul 9 23:56:19.877140 sshd_keygen[1568]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Jul 9 23:56:19.891310 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Jul 9 23:56:19.901246 systemd[1]: Starting issuegen.service - Generate /run/issue... Jul 9 23:56:19.903485 systemd[1]: issuegen.service: Deactivated successfully. Jul 9 23:56:19.903618 systemd[1]: Finished issuegen.service - Generate /run/issue. Jul 9 23:56:19.905654 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Jul 9 23:56:19.912745 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Jul 9 23:56:19.916612 systemd[1]: Started getty@tty1.service - Getty on tty1. Jul 9 23:56:19.918527 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Jul 9 23:56:19.919572 systemd[1]: Reached target getty.target - Login Prompts. Jul 9 23:56:20.004431 tar[1553]: linux-amd64/README.md Jul 9 23:56:20.016352 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Jul 9 23:56:21.220544 systemd-networkd[1465]: ens192: Gained IPv6LL Jul 9 23:56:21.220965 systemd-timesyncd[1475]: Network configuration changed, trying to establish connection. Jul 9 23:56:21.221876 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Jul 9 23:56:21.222764 systemd[1]: Reached target network-online.target - Network is Online. Jul 9 23:56:21.227593 systemd[1]: Starting coreos-metadata.service - VMware metadata agent... Jul 9 23:56:21.229086 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 9 23:56:21.230611 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Jul 9 23:56:21.249694 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Jul 9 23:56:21.259510 systemd[1]: coreos-metadata.service: Deactivated successfully. Jul 9 23:56:21.259652 systemd[1]: Finished coreos-metadata.service - VMware metadata agent. Jul 9 23:56:21.260274 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Jul 9 23:56:22.580102 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jul 9 23:56:22.580706 systemd[1]: Reached target multi-user.target - Multi-User System. Jul 9 23:56:22.581105 systemd[1]: Startup finished in 997ms (kernel) + 5.981s (initrd) + 4.999s (userspace) = 11.978s. Jul 9 23:56:22.586210 (kubelet)[1712]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jul 9 23:56:22.610215 login[1676]: pam_unix(login:session): session opened for user core(uid=500) by LOGIN(uid=0) Jul 9 23:56:22.610722 login[1677]: pam_unix(login:session): session opened for user core(uid=500) by LOGIN(uid=0) Jul 9 23:56:22.615152 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Jul 9 23:56:22.625520 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Jul 9 23:56:22.633841 systemd-logind[1542]: New session 2 of user core. Jul 9 23:56:22.636921 systemd-logind[1542]: New session 1 of user core. Jul 9 23:56:22.640579 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Jul 9 23:56:22.646643 systemd[1]: Starting user@500.service - User Manager for UID 500... Jul 9 23:56:22.648249 (systemd)[1719]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Jul 9 23:56:22.650030 systemd-logind[1542]: New session c1 of user core. Jul 9 23:56:22.741241 systemd[1719]: Queued start job for default target default.target. Jul 9 23:56:22.759167 systemd[1719]: Created slice app.slice - User Application Slice. Jul 9 23:56:22.759183 systemd[1719]: Reached target paths.target - Paths. Jul 9 23:56:22.759208 systemd[1719]: Reached target timers.target - Timers. Jul 9 23:56:22.762453 systemd[1719]: Starting dbus.socket - D-Bus User Message Bus Socket... Jul 9 23:56:22.766564 systemd[1719]: Listening on dbus.socket - D-Bus User Message Bus Socket. Jul 9 23:56:22.767008 systemd[1719]: Reached target sockets.target - Sockets. Jul 9 23:56:22.767032 systemd[1719]: Reached target basic.target - Basic System. Jul 9 23:56:22.767069 systemd[1719]: Reached target default.target - Main User Target. Jul 9 23:56:22.767085 systemd[1719]: Startup finished in 113ms. Jul 9 23:56:22.767375 systemd[1]: Started user@500.service - User Manager for UID 500. Jul 9 23:56:22.774456 systemd[1]: Started session-1.scope - Session 1 of User core. Jul 9 23:56:22.775768 systemd[1]: Started session-2.scope - Session 2 of User core. Jul 9 23:56:22.919142 systemd-timesyncd[1475]: Network configuration changed, trying to establish connection. Jul 9 23:56:23.124684 kubelet[1712]: E0709 23:56:23.124650 1712 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jul 9 23:56:23.126088 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jul 9 23:56:23.126174 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jul 9 23:56:23.126469 systemd[1]: kubelet.service: Consumed 689ms CPU time, 269.1M memory peak. Jul 9 23:56:33.376667 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Jul 9 23:56:33.392603 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 9 23:56:33.603515 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jul 9 23:56:33.606081 (kubelet)[1763]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jul 9 23:56:33.674135 kubelet[1763]: E0709 23:56:33.674051 1763 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jul 9 23:56:33.677312 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jul 9 23:56:33.677555 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jul 9 23:56:33.677946 systemd[1]: kubelet.service: Consumed 110ms CPU time, 109.8M memory peak. Jul 9 23:56:43.927864 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Jul 9 23:56:43.938543 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 9 23:56:44.278422 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jul 9 23:56:44.281746 (kubelet)[1777]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jul 9 23:56:44.312684 kubelet[1777]: E0709 23:56:44.312642 1777 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jul 9 23:56:44.313987 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jul 9 23:56:44.314063 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jul 9 23:56:44.314467 systemd[1]: kubelet.service: Consumed 97ms CPU time, 108.2M memory peak. Jul 9 23:56:49.738218 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Jul 9 23:56:49.748609 systemd[1]: Started sshd@0-139.178.70.101:22-139.178.89.65:50234.service - OpenSSH per-connection server daemon (139.178.89.65:50234). Jul 9 23:56:49.829531 sshd[1785]: Accepted publickey for core from 139.178.89.65 port 50234 ssh2: RSA SHA256:iW7N8ouL0MNZquiamslGIbLBE90/GD9BMvBNQ+/8OB0 Jul 9 23:56:49.830511 sshd-session[1785]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 9 23:56:49.834433 systemd-logind[1542]: New session 3 of user core. Jul 9 23:56:49.840500 systemd[1]: Started session-3.scope - Session 3 of User core. Jul 9 23:56:49.895514 systemd[1]: Started sshd@1-139.178.70.101:22-139.178.89.65:50244.service - OpenSSH per-connection server daemon (139.178.89.65:50244). Jul 9 23:56:49.925719 sshd[1790]: Accepted publickey for core from 139.178.89.65 port 50244 ssh2: RSA SHA256:iW7N8ouL0MNZquiamslGIbLBE90/GD9BMvBNQ+/8OB0 Jul 9 23:56:49.926574 sshd-session[1790]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 9 23:56:49.929588 systemd-logind[1542]: New session 4 of user core. Jul 9 23:56:49.935478 systemd[1]: Started session-4.scope - Session 4 of User core. Jul 9 23:56:49.985154 sshd[1792]: Connection closed by 139.178.89.65 port 50244 Jul 9 23:56:49.985095 sshd-session[1790]: pam_unix(sshd:session): session closed for user core Jul 9 23:56:50.001306 systemd[1]: sshd@1-139.178.70.101:22-139.178.89.65:50244.service: Deactivated successfully. Jul 9 23:56:50.003081 systemd[1]: session-4.scope: Deactivated successfully. Jul 9 23:56:50.003694 systemd-logind[1542]: Session 4 logged out. Waiting for processes to exit. Jul 9 23:56:50.007593 systemd[1]: Started sshd@2-139.178.70.101:22-139.178.89.65:50252.service - OpenSSH per-connection server daemon (139.178.89.65:50252). Jul 9 23:56:50.008475 systemd-logind[1542]: Removed session 4. Jul 9 23:56:50.039585 sshd[1797]: Accepted publickey for core from 139.178.89.65 port 50252 ssh2: RSA SHA256:iW7N8ouL0MNZquiamslGIbLBE90/GD9BMvBNQ+/8OB0 Jul 9 23:56:50.040559 sshd-session[1797]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 9 23:56:50.044052 systemd-logind[1542]: New session 5 of user core. Jul 9 23:56:50.049496 systemd[1]: Started session-5.scope - Session 5 of User core. Jul 9 23:56:50.095740 sshd[1800]: Connection closed by 139.178.89.65 port 50252 Jul 9 23:56:50.096715 sshd-session[1797]: pam_unix(sshd:session): session closed for user core Jul 9 23:56:50.108926 systemd[1]: sshd@2-139.178.70.101:22-139.178.89.65:50252.service: Deactivated successfully. Jul 9 23:56:50.110192 systemd[1]: session-5.scope: Deactivated successfully. Jul 9 23:56:50.111166 systemd-logind[1542]: Session 5 logged out. Waiting for processes to exit. Jul 9 23:56:50.114628 systemd[1]: Started sshd@3-139.178.70.101:22-139.178.89.65:50262.service - OpenSSH per-connection server daemon (139.178.89.65:50262). Jul 9 23:56:50.115792 systemd-logind[1542]: Removed session 5. Jul 9 23:56:50.145120 sshd[1805]: Accepted publickey for core from 139.178.89.65 port 50262 ssh2: RSA SHA256:iW7N8ouL0MNZquiamslGIbLBE90/GD9BMvBNQ+/8OB0 Jul 9 23:56:50.145965 sshd-session[1805]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 9 23:56:50.149570 systemd-logind[1542]: New session 6 of user core. Jul 9 23:56:50.159502 systemd[1]: Started session-6.scope - Session 6 of User core. Jul 9 23:56:50.209396 sshd[1808]: Connection closed by 139.178.89.65 port 50262 Jul 9 23:56:50.210317 sshd-session[1805]: pam_unix(sshd:session): session closed for user core Jul 9 23:56:50.222523 systemd[1]: sshd@3-139.178.70.101:22-139.178.89.65:50262.service: Deactivated successfully. Jul 9 23:56:50.223355 systemd[1]: session-6.scope: Deactivated successfully. Jul 9 23:56:50.224202 systemd-logind[1542]: Session 6 logged out. Waiting for processes to exit. Jul 9 23:56:50.224950 systemd[1]: Started sshd@4-139.178.70.101:22-139.178.89.65:50276.service - OpenSSH per-connection server daemon (139.178.89.65:50276). Jul 9 23:56:50.227556 systemd-logind[1542]: Removed session 6. Jul 9 23:56:50.257489 sshd[1813]: Accepted publickey for core from 139.178.89.65 port 50276 ssh2: RSA SHA256:iW7N8ouL0MNZquiamslGIbLBE90/GD9BMvBNQ+/8OB0 Jul 9 23:56:50.258669 sshd-session[1813]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 9 23:56:50.261723 systemd-logind[1542]: New session 7 of user core. Jul 9 23:56:50.268506 systemd[1]: Started session-7.scope - Session 7 of User core. Jul 9 23:56:50.324098 sudo[1817]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Jul 9 23:56:50.324512 sudo[1817]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jul 9 23:56:50.335915 sudo[1817]: pam_unix(sudo:session): session closed for user root Jul 9 23:56:50.336873 sshd[1816]: Connection closed by 139.178.89.65 port 50276 Jul 9 23:56:50.337205 sshd-session[1813]: pam_unix(sshd:session): session closed for user core Jul 9 23:56:50.348726 systemd[1]: sshd@4-139.178.70.101:22-139.178.89.65:50276.service: Deactivated successfully. Jul 9 23:56:50.349756 systemd[1]: session-7.scope: Deactivated successfully. Jul 9 23:56:50.350741 systemd-logind[1542]: Session 7 logged out. Waiting for processes to exit. Jul 9 23:56:50.351747 systemd[1]: Started sshd@5-139.178.70.101:22-139.178.89.65:50290.service - OpenSSH per-connection server daemon (139.178.89.65:50290). Jul 9 23:56:50.353558 systemd-logind[1542]: Removed session 7. Jul 9 23:56:50.383960 sshd[1822]: Accepted publickey for core from 139.178.89.65 port 50290 ssh2: RSA SHA256:iW7N8ouL0MNZquiamslGIbLBE90/GD9BMvBNQ+/8OB0 Jul 9 23:56:50.384786 sshd-session[1822]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 9 23:56:50.387832 systemd-logind[1542]: New session 8 of user core. Jul 9 23:56:50.397587 systemd[1]: Started session-8.scope - Session 8 of User core. Jul 9 23:56:50.445351 sudo[1827]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Jul 9 23:56:50.445521 sudo[1827]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jul 9 23:56:50.447356 sudo[1827]: pam_unix(sudo:session): session closed for user root Jul 9 23:56:50.450203 sudo[1826]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Jul 9 23:56:50.450349 sudo[1826]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jul 9 23:56:50.459635 systemd[1]: Starting audit-rules.service - Load Audit Rules... Jul 9 23:56:50.474518 augenrules[1849]: No rules Jul 9 23:56:50.475121 systemd[1]: audit-rules.service: Deactivated successfully. Jul 9 23:56:50.475250 systemd[1]: Finished audit-rules.service - Load Audit Rules. Jul 9 23:56:50.476013 sudo[1826]: pam_unix(sudo:session): session closed for user root Jul 9 23:56:50.476932 sshd[1825]: Connection closed by 139.178.89.65 port 50290 Jul 9 23:56:50.476890 sshd-session[1822]: pam_unix(sshd:session): session closed for user core Jul 9 23:56:50.482294 systemd[1]: sshd@5-139.178.70.101:22-139.178.89.65:50290.service: Deactivated successfully. Jul 9 23:56:50.483253 systemd[1]: session-8.scope: Deactivated successfully. Jul 9 23:56:50.484064 systemd-logind[1542]: Session 8 logged out. Waiting for processes to exit. Jul 9 23:56:50.484833 systemd[1]: Started sshd@6-139.178.70.101:22-139.178.89.65:50298.service - OpenSSH per-connection server daemon (139.178.89.65:50298). Jul 9 23:56:50.486646 systemd-logind[1542]: Removed session 8. Jul 9 23:56:50.525234 sshd[1857]: Accepted publickey for core from 139.178.89.65 port 50298 ssh2: RSA SHA256:iW7N8ouL0MNZquiamslGIbLBE90/GD9BMvBNQ+/8OB0 Jul 9 23:56:50.526066 sshd-session[1857]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 9 23:56:50.529491 systemd-logind[1542]: New session 9 of user core. Jul 9 23:56:50.534469 systemd[1]: Started session-9.scope - Session 9 of User core. Jul 9 23:56:50.583347 sudo[1861]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Jul 9 23:56:50.583724 sudo[1861]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jul 9 23:56:50.847515 systemd[1]: Starting docker.service - Docker Application Container Engine... Jul 9 23:56:50.847592 (dockerd)[1878]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Jul 9 23:56:51.095739 dockerd[1878]: time="2025-07-09T23:56:51.095578810Z" level=info msg="Starting up" Jul 9 23:56:51.148137 systemd[1]: var-lib-docker-check\x2doverlayfs\x2dsupport1871339874-merged.mount: Deactivated successfully. Jul 9 23:56:51.164921 dockerd[1878]: time="2025-07-09T23:56:51.164802339Z" level=info msg="Loading containers: start." Jul 9 23:56:51.369384 kernel: Initializing XFRM netlink socket Jul 9 23:56:51.434082 systemd-timesyncd[1475]: Network configuration changed, trying to establish connection. Jul 9 23:58:17.777965 systemd-resolved[1467]: Clock change detected. Flushing caches. Jul 9 23:58:17.778471 systemd-timesyncd[1475]: Contacted time server 66.42.71.197:123 (2.flatcar.pool.ntp.org). Jul 9 23:58:17.778501 systemd-timesyncd[1475]: Initial clock synchronization to Wed 2025-07-09 23:58:17.777926 UTC. Jul 9 23:58:17.808799 systemd-networkd[1465]: docker0: Link UP Jul 9 23:58:17.860879 dockerd[1878]: time="2025-07-09T23:58:17.860850517Z" level=info msg="Loading containers: done." Jul 9 23:58:17.868783 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck358767119-merged.mount: Deactivated successfully. Jul 9 23:58:17.879732 dockerd[1878]: time="2025-07-09T23:58:17.879705955Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Jul 9 23:58:17.879809 dockerd[1878]: time="2025-07-09T23:58:17.879776480Z" level=info msg="Docker daemon" commit=41ca978a0a5400cc24b274137efa9f25517fcc0b containerd-snapshotter=false storage-driver=overlay2 version=27.3.1 Jul 9 23:58:17.879878 dockerd[1878]: time="2025-07-09T23:58:17.879866818Z" level=info msg="Daemon has completed initialization" Jul 9 23:58:17.897045 dockerd[1878]: time="2025-07-09T23:58:17.896652839Z" level=info msg="API listen on /run/docker.sock" Jul 9 23:58:17.896835 systemd[1]: Started docker.service - Docker Application Container Engine. Jul 9 23:58:18.424564 containerd[1562]: time="2025-07-09T23:58:18.424328922Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.2\"" Jul 9 23:58:18.897355 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount465294413.mount: Deactivated successfully. Jul 9 23:58:19.861997 containerd[1562]: time="2025-07-09T23:58:19.861969091Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.33.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 9 23:58:19.863115 containerd[1562]: time="2025-07-09T23:58:19.862964214Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.33.2: active requests=0, bytes read=30079099" Jul 9 23:58:19.863395 containerd[1562]: time="2025-07-09T23:58:19.863383772Z" level=info msg="ImageCreate event name:\"sha256:ee794efa53d856b7e291320be3cd6390fa2e113c3f258a21290bc27fc214233e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 9 23:58:19.867799 containerd[1562]: time="2025-07-09T23:58:19.866910609Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:e8ae58675899e946fabe38425f2b3bfd33120b7930d05b5898de97c81a7f6137\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 9 23:58:19.867799 containerd[1562]: time="2025-07-09T23:58:19.867509259Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.33.2\" with image id \"sha256:ee794efa53d856b7e291320be3cd6390fa2e113c3f258a21290bc27fc214233e\", repo tag \"registry.k8s.io/kube-apiserver:v1.33.2\", repo digest \"registry.k8s.io/kube-apiserver@sha256:e8ae58675899e946fabe38425f2b3bfd33120b7930d05b5898de97c81a7f6137\", size \"30075899\" in 1.443149404s" Jul 9 23:58:19.867799 containerd[1562]: time="2025-07-09T23:58:19.867524633Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.2\" returns image reference \"sha256:ee794efa53d856b7e291320be3cd6390fa2e113c3f258a21290bc27fc214233e\"" Jul 9 23:58:19.868226 containerd[1562]: time="2025-07-09T23:58:19.868173304Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.2\"" Jul 9 23:58:20.841005 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Jul 9 23:58:20.847191 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 9 23:58:21.533356 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jul 9 23:58:21.535614 (kubelet)[2132]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jul 9 23:58:21.618448 kubelet[2132]: E0709 23:58:21.618378 2132 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jul 9 23:58:21.618935 containerd[1562]: time="2025-07-09T23:58:21.618896782Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.33.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 9 23:58:21.620741 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jul 9 23:58:21.620867 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jul 9 23:58:21.621379 systemd[1]: kubelet.service: Consumed 116ms CPU time, 108.5M memory peak. Jul 9 23:58:21.625591 containerd[1562]: time="2025-07-09T23:58:21.625556670Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.33.2: active requests=0, bytes read=26018946" Jul 9 23:58:21.630379 containerd[1562]: time="2025-07-09T23:58:21.630334509Z" level=info msg="ImageCreate event name:\"sha256:ff4f56c76b82d6cda0555115a0fe479d5dd612264b85efb9cc14b1b4b937bdf2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 9 23:58:21.638441 containerd[1562]: time="2025-07-09T23:58:21.638411066Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:2236e72a4be5dcc9c04600353ff8849db1557f5364947c520ff05471ae719081\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 9 23:58:21.639209 containerd[1562]: time="2025-07-09T23:58:21.639115608Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.33.2\" with image id \"sha256:ff4f56c76b82d6cda0555115a0fe479d5dd612264b85efb9cc14b1b4b937bdf2\", repo tag \"registry.k8s.io/kube-controller-manager:v1.33.2\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:2236e72a4be5dcc9c04600353ff8849db1557f5364947c520ff05471ae719081\", size \"27646507\" in 1.770884893s" Jul 9 23:58:21.639209 containerd[1562]: time="2025-07-09T23:58:21.639140774Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.2\" returns image reference \"sha256:ff4f56c76b82d6cda0555115a0fe479d5dd612264b85efb9cc14b1b4b937bdf2\"" Jul 9 23:58:21.639702 containerd[1562]: time="2025-07-09T23:58:21.639656672Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.2\"" Jul 9 23:58:22.928782 containerd[1562]: time="2025-07-09T23:58:22.928738460Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.33.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 9 23:58:22.929573 containerd[1562]: time="2025-07-09T23:58:22.929484910Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.33.2: active requests=0, bytes read=20155055" Jul 9 23:58:22.930322 containerd[1562]: time="2025-07-09T23:58:22.929999037Z" level=info msg="ImageCreate event name:\"sha256:cfed1ff7489289d4e8d796b0d95fd251990403510563cf843912f42ab9718a7b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 9 23:58:22.932471 containerd[1562]: time="2025-07-09T23:58:22.932174644Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:304c28303133be7d927973bc9bd6c83945b3735c59d283c25b63d5b9ed53bca3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 9 23:58:22.932892 containerd[1562]: time="2025-07-09T23:58:22.932875291Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.33.2\" with image id \"sha256:cfed1ff7489289d4e8d796b0d95fd251990403510563cf843912f42ab9718a7b\", repo tag \"registry.k8s.io/kube-scheduler:v1.33.2\", repo digest \"registry.k8s.io/kube-scheduler@sha256:304c28303133be7d927973bc9bd6c83945b3735c59d283c25b63d5b9ed53bca3\", size \"21782634\" in 1.293080418s" Jul 9 23:58:22.932939 containerd[1562]: time="2025-07-09T23:58:22.932892958Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.2\" returns image reference \"sha256:cfed1ff7489289d4e8d796b0d95fd251990403510563cf843912f42ab9718a7b\"" Jul 9 23:58:22.933202 containerd[1562]: time="2025-07-09T23:58:22.933185858Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.2\"" Jul 9 23:58:23.873973 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount423393373.mount: Deactivated successfully. Jul 9 23:58:24.483983 containerd[1562]: time="2025-07-09T23:58:24.483940086Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.33.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 9 23:58:24.484530 containerd[1562]: time="2025-07-09T23:58:24.484503943Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.33.2: active requests=0, bytes read=31892746" Jul 9 23:58:24.484822 containerd[1562]: time="2025-07-09T23:58:24.484787661Z" level=info msg="ImageCreate event name:\"sha256:661d404f36f01cd854403fd3540f18dcf0342d22bd9c6516bb9de234ac183b19\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 9 23:58:24.489037 containerd[1562]: time="2025-07-09T23:58:24.488911222Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:4796ef3e43efa5ed2a5b015c18f81d3c2fe3aea36f555ea643cc01827eb65e51\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 9 23:58:24.489402 containerd[1562]: time="2025-07-09T23:58:24.489316446Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.33.2\" with image id \"sha256:661d404f36f01cd854403fd3540f18dcf0342d22bd9c6516bb9de234ac183b19\", repo tag \"registry.k8s.io/kube-proxy:v1.33.2\", repo digest \"registry.k8s.io/kube-proxy@sha256:4796ef3e43efa5ed2a5b015c18f81d3c2fe3aea36f555ea643cc01827eb65e51\", size \"31891765\" in 1.556111903s" Jul 9 23:58:24.489402 containerd[1562]: time="2025-07-09T23:58:24.489337319Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.2\" returns image reference \"sha256:661d404f36f01cd854403fd3540f18dcf0342d22bd9c6516bb9de234ac183b19\"" Jul 9 23:58:24.489681 containerd[1562]: time="2025-07-09T23:58:24.489637683Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\"" Jul 9 23:58:24.945012 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount945597080.mount: Deactivated successfully. Jul 9 23:58:25.746210 containerd[1562]: time="2025-07-09T23:58:25.746163441Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.12.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 9 23:58:25.755885 containerd[1562]: time="2025-07-09T23:58:25.755842971Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.12.0: active requests=0, bytes read=20942238" Jul 9 23:58:25.771771 containerd[1562]: time="2025-07-09T23:58:25.771722851Z" level=info msg="ImageCreate event name:\"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 9 23:58:25.777009 containerd[1562]: time="2025-07-09T23:58:25.776974533Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 9 23:58:25.780684 containerd[1562]: time="2025-07-09T23:58:25.780135378Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.12.0\" with image id \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\", repo tag \"registry.k8s.io/coredns/coredns:v1.12.0\", repo digest \"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\", size \"20939036\" in 1.290474223s" Jul 9 23:58:25.780684 containerd[1562]: time="2025-07-09T23:58:25.780178104Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\" returns image reference \"sha256:1cf5f116067c67da67f97bff78c4bbc76913f59057c18627b96facaced73ea0b\"" Jul 9 23:58:25.781282 containerd[1562]: time="2025-07-09T23:58:25.781242066Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Jul 9 23:58:26.395900 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount964408868.mount: Deactivated successfully. Jul 9 23:58:26.546440 containerd[1562]: time="2025-07-09T23:58:26.546359820Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 9 23:58:26.551476 containerd[1562]: time="2025-07-09T23:58:26.551277360Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=321138" Jul 9 23:58:26.558458 containerd[1562]: time="2025-07-09T23:58:26.558413686Z" level=info msg="ImageCreate event name:\"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 9 23:58:26.563573 containerd[1562]: time="2025-07-09T23:58:26.563544117Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 9 23:58:26.564136 containerd[1562]: time="2025-07-09T23:58:26.564007678Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"320368\" in 782.74429ms" Jul 9 23:58:26.564136 containerd[1562]: time="2025-07-09T23:58:26.564029168Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Jul 9 23:58:26.564326 containerd[1562]: time="2025-07-09T23:58:26.564306976Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.21-0\"" Jul 9 23:58:27.000956 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2729927689.mount: Deactivated successfully. Jul 9 23:58:29.692495 containerd[1562]: time="2025-07-09T23:58:29.692460486Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.21-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 9 23:58:29.696143 containerd[1562]: time="2025-07-09T23:58:29.696108046Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.21-0: active requests=0, bytes read=58247175" Jul 9 23:58:29.701379 containerd[1562]: time="2025-07-09T23:58:29.701350244Z" level=info msg="ImageCreate event name:\"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 9 23:58:29.705913 containerd[1562]: time="2025-07-09T23:58:29.705889213Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:d58c035df557080a27387d687092e3fc2b64c6d0e3162dc51453a115f847d121\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 9 23:58:29.706757 containerd[1562]: time="2025-07-09T23:58:29.706737291Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.21-0\" with image id \"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\", repo tag \"registry.k8s.io/etcd:3.5.21-0\", repo digest \"registry.k8s.io/etcd@sha256:d58c035df557080a27387d687092e3fc2b64c6d0e3162dc51453a115f847d121\", size \"58938593\" in 3.142408993s" Jul 9 23:58:29.706801 containerd[1562]: time="2025-07-09T23:58:29.706760547Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.21-0\" returns image reference \"sha256:499038711c0816eda03a1ad96a8eb0440c005baa6949698223c6176b7f5077e1\"" Jul 9 23:58:30.884802 update_engine[1548]: I20250709 23:58:30.884421 1548 update_attempter.cc:509] Updating boot flags... Jul 9 23:58:30.915147 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 37 scanned by (udev-worker) (2293) Jul 9 23:58:30.965095 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 37 scanned by (udev-worker) (2292) Jul 9 23:58:31.871144 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 4. Jul 9 23:58:31.883221 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 9 23:58:32.315207 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jul 9 23:58:32.315781 (kubelet)[2309]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jul 9 23:58:32.365781 kubelet[2309]: E0709 23:58:32.364700 2309 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jul 9 23:58:32.365722 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jul 9 23:58:32.365811 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jul 9 23:58:32.366001 systemd[1]: kubelet.service: Consumed 76ms CPU time, 110.2M memory peak. Jul 9 23:58:32.588945 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jul 9 23:58:32.589046 systemd[1]: kubelet.service: Consumed 76ms CPU time, 110.2M memory peak. Jul 9 23:58:32.597219 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 9 23:58:32.614356 systemd[1]: Reload requested from client PID 2324 ('systemctl') (unit session-9.scope)... Jul 9 23:58:32.614367 systemd[1]: Reloading... Jul 9 23:58:32.676137 zram_generator::config[2371]: No configuration found. Jul 9 23:58:32.731655 systemd[1]: /etc/systemd/system/coreos-metadata.service:11: Ignoring unknown escape sequences: "echo "COREOS_CUSTOM_PRIVATE_IPV4=$(ip addr show ens192 | grep "inet 10." | grep -Po "inet \K[\d.]+") Jul 9 23:58:32.749616 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jul 9 23:58:32.814300 systemd[1]: Reloading finished in 199 ms. Jul 9 23:58:32.838382 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Jul 9 23:58:32.838456 systemd[1]: kubelet.service: Failed with result 'signal'. Jul 9 23:58:32.838737 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jul 9 23:58:32.844299 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 9 23:58:33.136267 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jul 9 23:58:33.139701 (kubelet)[2436]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jul 9 23:58:33.211418 kubelet[2436]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jul 9 23:58:33.211418 kubelet[2436]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Jul 9 23:58:33.211418 kubelet[2436]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jul 9 23:58:33.223265 kubelet[2436]: I0709 23:58:33.223230 2436 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jul 9 23:58:33.620723 kubelet[2436]: I0709 23:58:33.620525 2436 server.go:530] "Kubelet version" kubeletVersion="v1.33.0" Jul 9 23:58:33.620723 kubelet[2436]: I0709 23:58:33.620555 2436 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jul 9 23:58:33.620723 kubelet[2436]: I0709 23:58:33.620700 2436 server.go:956] "Client rotation is on, will bootstrap in background" Jul 9 23:58:33.653208 kubelet[2436]: E0709 23:58:33.653033 2436 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://139.178.70.101:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 139.178.70.101:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Jul 9 23:58:33.653408 kubelet[2436]: I0709 23:58:33.653324 2436 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jul 9 23:58:33.669363 kubelet[2436]: E0709 23:58:33.669196 2436 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Jul 9 23:58:33.669363 kubelet[2436]: I0709 23:58:33.669223 2436 server.go:1423] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Jul 9 23:58:33.675996 kubelet[2436]: I0709 23:58:33.675923 2436 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jul 9 23:58:33.678378 kubelet[2436]: I0709 23:58:33.678233 2436 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jul 9 23:58:33.680776 kubelet[2436]: I0709 23:58:33.678257 2436 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jul 9 23:58:33.681592 kubelet[2436]: I0709 23:58:33.681465 2436 topology_manager.go:138] "Creating topology manager with none policy" Jul 9 23:58:33.681592 kubelet[2436]: I0709 23:58:33.681478 2436 container_manager_linux.go:303] "Creating device plugin manager" Jul 9 23:58:33.682226 kubelet[2436]: I0709 23:58:33.682196 2436 state_mem.go:36] "Initialized new in-memory state store" Jul 9 23:58:33.685937 kubelet[2436]: I0709 23:58:33.685841 2436 kubelet.go:480] "Attempting to sync node with API server" Jul 9 23:58:33.685937 kubelet[2436]: I0709 23:58:33.685857 2436 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Jul 9 23:58:33.685937 kubelet[2436]: I0709 23:58:33.685875 2436 kubelet.go:386] "Adding apiserver pod source" Jul 9 23:58:33.685937 kubelet[2436]: I0709 23:58:33.685882 2436 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jul 9 23:58:33.691767 kubelet[2436]: E0709 23:58:33.691456 2436 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://139.178.70.101:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 139.178.70.101:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Jul 9 23:58:33.695311 kubelet[2436]: E0709 23:58:33.695045 2436 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://139.178.70.101:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 139.178.70.101:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Jul 9 23:58:33.695311 kubelet[2436]: I0709 23:58:33.695132 2436 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v1.7.23" apiVersion="v1" Jul 9 23:58:33.695471 kubelet[2436]: I0709 23:58:33.695409 2436 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Jul 9 23:58:33.697172 kubelet[2436]: W0709 23:58:33.696745 2436 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Jul 9 23:58:33.701825 kubelet[2436]: I0709 23:58:33.701383 2436 watchdog_linux.go:99] "Systemd watchdog is not enabled" Jul 9 23:58:33.701825 kubelet[2436]: I0709 23:58:33.701427 2436 server.go:1289] "Started kubelet" Jul 9 23:58:33.706161 kubelet[2436]: I0709 23:58:33.706109 2436 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Jul 9 23:58:33.708615 kubelet[2436]: I0709 23:58:33.708531 2436 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jul 9 23:58:33.709294 kubelet[2436]: I0709 23:58:33.709282 2436 server.go:317] "Adding debug handlers to kubelet server" Jul 9 23:58:33.713572 kubelet[2436]: I0709 23:58:33.713543 2436 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Jul 9 23:58:33.714602 kubelet[2436]: I0709 23:58:33.714482 2436 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jul 9 23:58:33.714736 kubelet[2436]: I0709 23:58:33.714729 2436 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jul 9 23:58:33.715356 kubelet[2436]: I0709 23:58:33.715339 2436 volume_manager.go:297] "Starting Kubelet Volume Manager" Jul 9 23:58:33.716353 kubelet[2436]: E0709 23:58:33.715484 2436 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"localhost\" not found" Jul 9 23:58:33.718428 kubelet[2436]: E0709 23:58:33.714915 2436 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://139.178.70.101:6443/api/v1/namespaces/default/events\": dial tcp 139.178.70.101:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.1850baae812cf7e5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-07-09 23:58:33.701406693 +0000 UTC m=+0.559336711,LastTimestamp:2025-07-09 23:58:33.701406693 +0000 UTC m=+0.559336711,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Jul 9 23:58:33.718428 kubelet[2436]: E0709 23:58:33.717786 2436 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://139.178.70.101:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 139.178.70.101:6443: connect: connection refused" interval="200ms" Jul 9 23:58:33.718428 kubelet[2436]: I0709 23:58:33.718062 2436 factory.go:223] Registration of the systemd container factory successfully Jul 9 23:58:33.718428 kubelet[2436]: I0709 23:58:33.718114 2436 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jul 9 23:58:33.718775 kubelet[2436]: I0709 23:58:33.718605 2436 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Jul 9 23:58:33.718775 kubelet[2436]: I0709 23:58:33.718648 2436 reconciler.go:26] "Reconciler: start to sync state" Jul 9 23:58:33.719099 kubelet[2436]: E0709 23:58:33.719087 2436 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://139.178.70.101:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 139.178.70.101:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Jul 9 23:58:33.722037 kubelet[2436]: I0709 23:58:33.722016 2436 factory.go:223] Registration of the containerd container factory successfully Jul 9 23:58:33.724973 kubelet[2436]: I0709 23:58:33.724940 2436 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Jul 9 23:58:33.725656 kubelet[2436]: I0709 23:58:33.725646 2436 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Jul 9 23:58:33.725716 kubelet[2436]: I0709 23:58:33.725711 2436 status_manager.go:230] "Starting to sync pod status with apiserver" Jul 9 23:58:33.725757 kubelet[2436]: I0709 23:58:33.725752 2436 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Jul 9 23:58:33.725788 kubelet[2436]: I0709 23:58:33.725784 2436 kubelet.go:2436] "Starting kubelet main sync loop" Jul 9 23:58:33.725843 kubelet[2436]: E0709 23:58:33.725834 2436 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jul 9 23:58:33.729864 kubelet[2436]: E0709 23:58:33.729851 2436 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://139.178.70.101:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 139.178.70.101:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Jul 9 23:58:33.752665 kubelet[2436]: I0709 23:58:33.752650 2436 cpu_manager.go:221] "Starting CPU manager" policy="none" Jul 9 23:58:33.752665 kubelet[2436]: I0709 23:58:33.752660 2436 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Jul 9 23:58:33.752665 kubelet[2436]: I0709 23:58:33.752672 2436 state_mem.go:36] "Initialized new in-memory state store" Jul 9 23:58:33.763362 kubelet[2436]: I0709 23:58:33.763330 2436 policy_none.go:49] "None policy: Start" Jul 9 23:58:33.763362 kubelet[2436]: I0709 23:58:33.763356 2436 memory_manager.go:186] "Starting memorymanager" policy="None" Jul 9 23:58:33.763362 kubelet[2436]: I0709 23:58:33.763368 2436 state_mem.go:35] "Initializing new in-memory state store" Jul 9 23:58:33.788792 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Jul 9 23:58:33.801618 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Jul 9 23:58:33.804123 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Jul 9 23:58:33.810701 kubelet[2436]: E0709 23:58:33.810678 2436 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Jul 9 23:58:33.810995 kubelet[2436]: I0709 23:58:33.810799 2436 eviction_manager.go:189] "Eviction manager: starting control loop" Jul 9 23:58:33.810995 kubelet[2436]: I0709 23:58:33.810814 2436 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jul 9 23:58:33.811353 kubelet[2436]: I0709 23:58:33.811131 2436 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jul 9 23:58:33.812081 kubelet[2436]: E0709 23:58:33.812052 2436 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Jul 9 23:58:33.812452 kubelet[2436]: E0709 23:58:33.812088 2436 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Jul 9 23:58:33.912884 kubelet[2436]: I0709 23:58:33.912155 2436 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jul 9 23:58:33.912884 kubelet[2436]: E0709 23:58:33.912374 2436 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://139.178.70.101:6443/api/v1/nodes\": dial tcp 139.178.70.101:6443: connect: connection refused" node="localhost" Jul 9 23:58:33.919480 kubelet[2436]: E0709 23:58:33.919418 2436 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://139.178.70.101:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 139.178.70.101:6443: connect: connection refused" interval="400ms" Jul 9 23:58:33.923026 systemd[1]: Created slice kubepods-burstable-poda9674aa579a9a6f3b4f62af6707ef78a.slice - libcontainer container kubepods-burstable-poda9674aa579a9a6f3b4f62af6707ef78a.slice. Jul 9 23:58:33.928613 kubelet[2436]: E0709 23:58:33.928586 2436 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jul 9 23:58:33.951032 systemd[1]: Created slice kubepods-burstable-pod84b858ec27c8b2738b1d9ff9927e0dcb.slice - libcontainer container kubepods-burstable-pod84b858ec27c8b2738b1d9ff9927e0dcb.slice. Jul 9 23:58:33.952468 kubelet[2436]: E0709 23:58:33.952450 2436 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jul 9 23:58:33.987921 systemd[1]: Created slice kubepods-burstable-pod834ee54f1daa06092e339273649eb5ea.slice - libcontainer container kubepods-burstable-pod834ee54f1daa06092e339273649eb5ea.slice. Jul 9 23:58:33.989024 kubelet[2436]: E0709 23:58:33.989008 2436 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jul 9 23:58:34.019332 kubelet[2436]: I0709 23:58:34.019280 2436 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/a9674aa579a9a6f3b4f62af6707ef78a-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"a9674aa579a9a6f3b4f62af6707ef78a\") " pod="kube-system/kube-apiserver-localhost" Jul 9 23:58:34.019332 kubelet[2436]: I0709 23:58:34.019309 2436 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/a9674aa579a9a6f3b4f62af6707ef78a-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"a9674aa579a9a6f3b4f62af6707ef78a\") " pod="kube-system/kube-apiserver-localhost" Jul 9 23:58:34.019332 kubelet[2436]: I0709 23:58:34.019337 2436 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/a9674aa579a9a6f3b4f62af6707ef78a-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"a9674aa579a9a6f3b4f62af6707ef78a\") " pod="kube-system/kube-apiserver-localhost" Jul 9 23:58:34.019473 kubelet[2436]: I0709 23:58:34.019347 2436 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/84b858ec27c8b2738b1d9ff9927e0dcb-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"84b858ec27c8b2738b1d9ff9927e0dcb\") " pod="kube-system/kube-controller-manager-localhost" Jul 9 23:58:34.019473 kubelet[2436]: I0709 23:58:34.019359 2436 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/84b858ec27c8b2738b1d9ff9927e0dcb-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"84b858ec27c8b2738b1d9ff9927e0dcb\") " pod="kube-system/kube-controller-manager-localhost" Jul 9 23:58:34.019473 kubelet[2436]: I0709 23:58:34.019368 2436 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/84b858ec27c8b2738b1d9ff9927e0dcb-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"84b858ec27c8b2738b1d9ff9927e0dcb\") " pod="kube-system/kube-controller-manager-localhost" Jul 9 23:58:34.019473 kubelet[2436]: I0709 23:58:34.019377 2436 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/84b858ec27c8b2738b1d9ff9927e0dcb-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"84b858ec27c8b2738b1d9ff9927e0dcb\") " pod="kube-system/kube-controller-manager-localhost" Jul 9 23:58:34.019473 kubelet[2436]: I0709 23:58:34.019385 2436 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/84b858ec27c8b2738b1d9ff9927e0dcb-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"84b858ec27c8b2738b1d9ff9927e0dcb\") " pod="kube-system/kube-controller-manager-localhost" Jul 9 23:58:34.113593 kubelet[2436]: I0709 23:58:34.113550 2436 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jul 9 23:58:34.113771 kubelet[2436]: E0709 23:58:34.113754 2436 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://139.178.70.101:6443/api/v1/nodes\": dial tcp 139.178.70.101:6443: connect: connection refused" node="localhost" Jul 9 23:58:34.120277 kubelet[2436]: I0709 23:58:34.120141 2436 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/834ee54f1daa06092e339273649eb5ea-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"834ee54f1daa06092e339273649eb5ea\") " pod="kube-system/kube-scheduler-localhost" Jul 9 23:58:34.229618 containerd[1562]: time="2025-07-09T23:58:34.229542484Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:a9674aa579a9a6f3b4f62af6707ef78a,Namespace:kube-system,Attempt:0,}" Jul 9 23:58:34.253255 containerd[1562]: time="2025-07-09T23:58:34.253049042Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:84b858ec27c8b2738b1d9ff9927e0dcb,Namespace:kube-system,Attempt:0,}" Jul 9 23:58:34.290241 containerd[1562]: time="2025-07-09T23:58:34.290001996Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:834ee54f1daa06092e339273649eb5ea,Namespace:kube-system,Attempt:0,}" Jul 9 23:58:34.320673 kubelet[2436]: E0709 23:58:34.320642 2436 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://139.178.70.101:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 139.178.70.101:6443: connect: connection refused" interval="800ms" Jul 9 23:58:34.515667 kubelet[2436]: I0709 23:58:34.515550 2436 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jul 9 23:58:34.515966 kubelet[2436]: E0709 23:58:34.515952 2436 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://139.178.70.101:6443/api/v1/nodes\": dial tcp 139.178.70.101:6443: connect: connection refused" node="localhost" Jul 9 23:58:34.733721 kubelet[2436]: E0709 23:58:34.733672 2436 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://139.178.70.101:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 139.178.70.101:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Jul 9 23:58:34.770380 kubelet[2436]: E0709 23:58:34.770242 2436 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://139.178.70.101:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 139.178.70.101:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Jul 9 23:58:34.833282 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1402366393.mount: Deactivated successfully. Jul 9 23:58:34.835636 containerd[1562]: time="2025-07-09T23:58:34.835552603Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jul 9 23:58:34.836232 containerd[1562]: time="2025-07-09T23:58:34.836205732Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jul 9 23:58:34.836727 containerd[1562]: time="2025-07-09T23:58:34.836702268Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Jul 9 23:58:34.836998 containerd[1562]: time="2025-07-09T23:58:34.836980110Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jul 9 23:58:34.837589 containerd[1562]: time="2025-07-09T23:58:34.837455314Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312056" Jul 9 23:58:34.838173 containerd[1562]: time="2025-07-09T23:58:34.838144948Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jul 9 23:58:34.838483 containerd[1562]: time="2025-07-09T23:58:34.838425856Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Jul 9 23:58:34.839821 containerd[1562]: time="2025-07-09T23:58:34.839793398Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jul 9 23:58:34.840940 containerd[1562]: time="2025-07-09T23:58:34.840797209Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 610.459382ms" Jul 9 23:58:34.842634 containerd[1562]: time="2025-07-09T23:58:34.842565464Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 589.424412ms" Jul 9 23:58:34.844005 containerd[1562]: time="2025-07-09T23:58:34.843844650Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 553.775182ms" Jul 9 23:58:34.976310 containerd[1562]: time="2025-07-09T23:58:34.974689222Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 9 23:58:34.976310 containerd[1562]: time="2025-07-09T23:58:34.976186262Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 9 23:58:34.976310 containerd[1562]: time="2025-07-09T23:58:34.976199456Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 9 23:58:34.976979 containerd[1562]: time="2025-07-09T23:58:34.976950946Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 9 23:58:34.983802 containerd[1562]: time="2025-07-09T23:58:34.982384829Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 9 23:58:34.983802 containerd[1562]: time="2025-07-09T23:58:34.982422866Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 9 23:58:34.983802 containerd[1562]: time="2025-07-09T23:58:34.982429900Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 9 23:58:34.983802 containerd[1562]: time="2025-07-09T23:58:34.982479701Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 9 23:58:34.985551 containerd[1562]: time="2025-07-09T23:58:34.985494502Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 9 23:58:34.986186 containerd[1562]: time="2025-07-09T23:58:34.986147453Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 9 23:58:34.986226 containerd[1562]: time="2025-07-09T23:58:34.986188943Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 9 23:58:34.986384 containerd[1562]: time="2025-07-09T23:58:34.986365545Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 9 23:58:34.998262 systemd[1]: Started cri-containerd-5a0b1ad6c20e8cdf2cb1c21386d94fcc2bb930e0d26dcc0e9852ef58924343fb.scope - libcontainer container 5a0b1ad6c20e8cdf2cb1c21386d94fcc2bb930e0d26dcc0e9852ef58924343fb. Jul 9 23:58:34.999523 systemd[1]: Started cri-containerd-682f874f620c16961175c9728ad45ee8bbc1ae5a9280882bb5f612ec78a84b7e.scope - libcontainer container 682f874f620c16961175c9728ad45ee8bbc1ae5a9280882bb5f612ec78a84b7e. Jul 9 23:58:35.021387 systemd[1]: Started cri-containerd-9452497784434066ce2598a03004130b853dbe24009ac3c29f18755d95c08a7e.scope - libcontainer container 9452497784434066ce2598a03004130b853dbe24009ac3c29f18755d95c08a7e. Jul 9 23:58:35.042703 containerd[1562]: time="2025-07-09T23:58:35.042677166Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:84b858ec27c8b2738b1d9ff9927e0dcb,Namespace:kube-system,Attempt:0,} returns sandbox id \"682f874f620c16961175c9728ad45ee8bbc1ae5a9280882bb5f612ec78a84b7e\"" Jul 9 23:58:35.063293 containerd[1562]: time="2025-07-09T23:58:35.063267795Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:a9674aa579a9a6f3b4f62af6707ef78a,Namespace:kube-system,Attempt:0,} returns sandbox id \"5a0b1ad6c20e8cdf2cb1c21386d94fcc2bb930e0d26dcc0e9852ef58924343fb\"" Jul 9 23:58:35.069212 containerd[1562]: time="2025-07-09T23:58:35.069189670Z" level=info msg="CreateContainer within sandbox \"682f874f620c16961175c9728ad45ee8bbc1ae5a9280882bb5f612ec78a84b7e\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Jul 9 23:58:35.082404 containerd[1562]: time="2025-07-09T23:58:35.082380243Z" level=info msg="CreateContainer within sandbox \"5a0b1ad6c20e8cdf2cb1c21386d94fcc2bb930e0d26dcc0e9852ef58924343fb\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Jul 9 23:58:35.090920 containerd[1562]: time="2025-07-09T23:58:35.090846276Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:834ee54f1daa06092e339273649eb5ea,Namespace:kube-system,Attempt:0,} returns sandbox id \"9452497784434066ce2598a03004130b853dbe24009ac3c29f18755d95c08a7e\"" Jul 9 23:58:35.093431 containerd[1562]: time="2025-07-09T23:58:35.093368958Z" level=info msg="CreateContainer within sandbox \"9452497784434066ce2598a03004130b853dbe24009ac3c29f18755d95c08a7e\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Jul 9 23:58:35.103099 containerd[1562]: time="2025-07-09T23:58:35.103051758Z" level=info msg="CreateContainer within sandbox \"682f874f620c16961175c9728ad45ee8bbc1ae5a9280882bb5f612ec78a84b7e\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"08d6f38a3a33911e476436bef8e5cca29e92a5c2026d9f2e89e28dad6afca688\"" Jul 9 23:58:35.104911 containerd[1562]: time="2025-07-09T23:58:35.104891072Z" level=info msg="CreateContainer within sandbox \"5a0b1ad6c20e8cdf2cb1c21386d94fcc2bb930e0d26dcc0e9852ef58924343fb\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"5738130e765920a168ea23160e3619bed64b61d19b2e8c092de1c74662b5d54d\"" Jul 9 23:58:35.106560 containerd[1562]: time="2025-07-09T23:58:35.106534272Z" level=info msg="CreateContainer within sandbox \"9452497784434066ce2598a03004130b853dbe24009ac3c29f18755d95c08a7e\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"faf59ce1092801270cf02d55ac9c65774e1a990f0a36fbe5d51863146cec2df6\"" Jul 9 23:58:35.108873 containerd[1562]: time="2025-07-09T23:58:35.108842871Z" level=info msg="StartContainer for \"08d6f38a3a33911e476436bef8e5cca29e92a5c2026d9f2e89e28dad6afca688\"" Jul 9 23:58:35.109985 containerd[1562]: time="2025-07-09T23:58:35.109966250Z" level=info msg="StartContainer for \"faf59ce1092801270cf02d55ac9c65774e1a990f0a36fbe5d51863146cec2df6\"" Jul 9 23:58:35.110629 containerd[1562]: time="2025-07-09T23:58:35.110100530Z" level=info msg="StartContainer for \"5738130e765920a168ea23160e3619bed64b61d19b2e8c092de1c74662b5d54d\"" Jul 9 23:58:35.121318 kubelet[2436]: E0709 23:58:35.121290 2436 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://139.178.70.101:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 139.178.70.101:6443: connect: connection refused" interval="1.6s" Jul 9 23:58:35.126647 kubelet[2436]: E0709 23:58:35.126605 2436 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://139.178.70.101:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0\": dial tcp 139.178.70.101:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Jul 9 23:58:35.135208 systemd[1]: Started cri-containerd-faf59ce1092801270cf02d55ac9c65774e1a990f0a36fbe5d51863146cec2df6.scope - libcontainer container faf59ce1092801270cf02d55ac9c65774e1a990f0a36fbe5d51863146cec2df6. Jul 9 23:58:35.138851 systemd[1]: Started cri-containerd-08d6f38a3a33911e476436bef8e5cca29e92a5c2026d9f2e89e28dad6afca688.scope - libcontainer container 08d6f38a3a33911e476436bef8e5cca29e92a5c2026d9f2e89e28dad6afca688. Jul 9 23:58:35.140988 systemd[1]: Started cri-containerd-5738130e765920a168ea23160e3619bed64b61d19b2e8c092de1c74662b5d54d.scope - libcontainer container 5738130e765920a168ea23160e3619bed64b61d19b2e8c092de1c74662b5d54d. Jul 9 23:58:35.167092 kubelet[2436]: E0709 23:58:35.167041 2436 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://139.178.70.101:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 139.178.70.101:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Jul 9 23:58:35.188937 containerd[1562]: time="2025-07-09T23:58:35.188905620Z" level=info msg="StartContainer for \"faf59ce1092801270cf02d55ac9c65774e1a990f0a36fbe5d51863146cec2df6\" returns successfully" Jul 9 23:58:35.189250 containerd[1562]: time="2025-07-09T23:58:35.188947564Z" level=info msg="StartContainer for \"5738130e765920a168ea23160e3619bed64b61d19b2e8c092de1c74662b5d54d\" returns successfully" Jul 9 23:58:35.195467 containerd[1562]: time="2025-07-09T23:58:35.195235098Z" level=info msg="StartContainer for \"08d6f38a3a33911e476436bef8e5cca29e92a5c2026d9f2e89e28dad6afca688\" returns successfully" Jul 9 23:58:35.317260 kubelet[2436]: I0709 23:58:35.317170 2436 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jul 9 23:58:35.318125 kubelet[2436]: E0709 23:58:35.317440 2436 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://139.178.70.101:6443/api/v1/nodes\": dial tcp 139.178.70.101:6443: connect: connection refused" node="localhost" Jul 9 23:58:35.757052 kubelet[2436]: E0709 23:58:35.757034 2436 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jul 9 23:58:35.758531 kubelet[2436]: E0709 23:58:35.758520 2436 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jul 9 23:58:35.759519 kubelet[2436]: E0709 23:58:35.759509 2436 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jul 9 23:58:35.852198 kubelet[2436]: E0709 23:58:35.852160 2436 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://139.178.70.101:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 139.178.70.101:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Jul 9 23:58:36.236854 kubelet[2436]: E0709 23:58:36.236781 2436 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://139.178.70.101:6443/api/v1/namespaces/default/events\": dial tcp 139.178.70.101:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.1850baae812cf7e5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-07-09 23:58:33.701406693 +0000 UTC m=+0.559336711,LastTimestamp:2025-07-09 23:58:33.701406693 +0000 UTC m=+0.559336711,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Jul 9 23:58:36.761716 kubelet[2436]: E0709 23:58:36.761624 2436 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jul 9 23:58:36.762379 kubelet[2436]: E0709 23:58:36.762190 2436 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jul 9 23:58:36.762379 kubelet[2436]: E0709 23:58:36.762324 2436 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"localhost\" not found" node="localhost" Jul 9 23:58:36.919358 kubelet[2436]: I0709 23:58:36.919184 2436 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jul 9 23:58:38.083002 kubelet[2436]: E0709 23:58:38.082979 2436 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Jul 9 23:58:38.243769 kubelet[2436]: I0709 23:58:38.243734 2436 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Jul 9 23:58:38.243769 kubelet[2436]: E0709 23:58:38.243771 2436 kubelet_node_status.go:548] "Error updating node status, will retry" err="error getting node \"localhost\": node \"localhost\" not found" Jul 9 23:58:38.316785 kubelet[2436]: I0709 23:58:38.316755 2436 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Jul 9 23:58:38.323597 kubelet[2436]: E0709 23:58:38.323565 2436 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-localhost" Jul 9 23:58:38.323597 kubelet[2436]: I0709 23:58:38.323592 2436 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Jul 9 23:58:38.324720 kubelet[2436]: E0709 23:58:38.324651 2436 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-localhost" Jul 9 23:58:38.324720 kubelet[2436]: I0709 23:58:38.324667 2436 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Jul 9 23:58:38.325541 kubelet[2436]: E0709 23:58:38.325524 2436 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-localhost" Jul 9 23:58:38.693392 kubelet[2436]: I0709 23:58:38.693329 2436 apiserver.go:52] "Watching apiserver" Jul 9 23:58:38.719100 kubelet[2436]: I0709 23:58:38.719046 2436 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Jul 9 23:58:40.790476 systemd[1]: Reload requested from client PID 2717 ('systemctl') (unit session-9.scope)... Jul 9 23:58:40.790488 systemd[1]: Reloading... Jul 9 23:58:40.859102 zram_generator::config[2765]: No configuration found. Jul 9 23:58:40.932023 systemd[1]: /etc/systemd/system/coreos-metadata.service:11: Ignoring unknown escape sequences: "echo "COREOS_CUSTOM_PRIVATE_IPV4=$(ip addr show ens192 | grep "inet 10." | grep -Po "inet \K[\d.]+") Jul 9 23:58:40.951677 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jul 9 23:58:41.034263 systemd[1]: Reloading finished in 243 ms. Jul 9 23:58:41.054899 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Jul 9 23:58:41.067844 systemd[1]: kubelet.service: Deactivated successfully. Jul 9 23:58:41.068047 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jul 9 23:58:41.068097 systemd[1]: kubelet.service: Consumed 721ms CPU time, 129M memory peak. Jul 9 23:58:41.073260 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jul 9 23:58:41.657317 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jul 9 23:58:41.668374 (kubelet)[2830]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jul 9 23:58:41.761768 kubelet[2830]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jul 9 23:58:41.761768 kubelet[2830]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Jul 9 23:58:41.761768 kubelet[2830]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jul 9 23:58:41.762008 kubelet[2830]: I0709 23:58:41.761785 2830 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jul 9 23:58:41.766091 kubelet[2830]: I0709 23:58:41.765438 2830 server.go:530] "Kubelet version" kubeletVersion="v1.33.0" Jul 9 23:58:41.766091 kubelet[2830]: I0709 23:58:41.765452 2830 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jul 9 23:58:41.766091 kubelet[2830]: I0709 23:58:41.765604 2830 server.go:956] "Client rotation is on, will bootstrap in background" Jul 9 23:58:41.766328 kubelet[2830]: I0709 23:58:41.766317 2830 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Jul 9 23:58:41.828096 kubelet[2830]: I0709 23:58:41.826311 2830 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jul 9 23:58:41.828817 kubelet[2830]: E0709 23:58:41.828798 2830 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Jul 9 23:58:41.828817 kubelet[2830]: I0709 23:58:41.828817 2830 server.go:1423] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Jul 9 23:58:41.830533 kubelet[2830]: I0709 23:58:41.830469 2830 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jul 9 23:58:41.830601 kubelet[2830]: I0709 23:58:41.830582 2830 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jul 9 23:58:41.830683 kubelet[2830]: I0709 23:58:41.830600 2830 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jul 9 23:58:41.830739 kubelet[2830]: I0709 23:58:41.830691 2830 topology_manager.go:138] "Creating topology manager with none policy" Jul 9 23:58:41.830739 kubelet[2830]: I0709 23:58:41.830701 2830 container_manager_linux.go:303] "Creating device plugin manager" Jul 9 23:58:41.830739 kubelet[2830]: I0709 23:58:41.830731 2830 state_mem.go:36] "Initialized new in-memory state store" Jul 9 23:58:41.830857 kubelet[2830]: I0709 23:58:41.830847 2830 kubelet.go:480] "Attempting to sync node with API server" Jul 9 23:58:41.830882 kubelet[2830]: I0709 23:58:41.830859 2830 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Jul 9 23:58:41.830882 kubelet[2830]: I0709 23:58:41.830874 2830 kubelet.go:386] "Adding apiserver pod source" Jul 9 23:58:41.831021 kubelet[2830]: I0709 23:58:41.830883 2830 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jul 9 23:58:41.840164 kubelet[2830]: I0709 23:58:41.840112 2830 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v1.7.23" apiVersion="v1" Jul 9 23:58:41.841086 kubelet[2830]: I0709 23:58:41.840509 2830 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Jul 9 23:58:41.845618 kubelet[2830]: I0709 23:58:41.845609 2830 watchdog_linux.go:99] "Systemd watchdog is not enabled" Jul 9 23:58:41.878199 kubelet[2830]: I0709 23:58:41.878185 2830 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jul 9 23:58:41.882863 kubelet[2830]: I0709 23:58:41.882829 2830 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jul 9 23:58:41.883788 kubelet[2830]: I0709 23:58:41.883777 2830 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jul 9 23:58:41.883881 kubelet[2830]: I0709 23:58:41.883869 2830 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Jul 9 23:58:41.898163 kubelet[2830]: I0709 23:58:41.898146 2830 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Jul 9 23:58:41.900879 kubelet[2830]: I0709 23:58:41.900869 2830 server.go:1289] "Started kubelet" Jul 9 23:58:41.900941 kubelet[2830]: I0709 23:58:41.900931 2830 volume_manager.go:297] "Starting Kubelet Volume Manager" Jul 9 23:58:41.903185 kubelet[2830]: I0709 23:58:41.903172 2830 server.go:317] "Adding debug handlers to kubelet server" Jul 9 23:58:41.913745 kubelet[2830]: I0709 23:58:41.912911 2830 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Jul 9 23:58:41.913745 kubelet[2830]: I0709 23:58:41.912991 2830 reconciler.go:26] "Reconciler: start to sync state" Jul 9 23:58:41.941437 kubelet[2830]: E0709 23:58:41.941399 2830 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jul 9 23:58:41.941817 kubelet[2830]: I0709 23:58:41.941805 2830 factory.go:223] Registration of the containerd container factory successfully Jul 9 23:58:41.942090 kubelet[2830]: I0709 23:58:41.941898 2830 factory.go:223] Registration of the systemd container factory successfully Jul 9 23:58:41.942090 kubelet[2830]: I0709 23:58:41.941943 2830 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jul 9 23:58:41.946758 kubelet[2830]: I0709 23:58:41.946740 2830 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Jul 9 23:58:41.947671 kubelet[2830]: I0709 23:58:41.947663 2830 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Jul 9 23:58:41.947966 kubelet[2830]: I0709 23:58:41.947787 2830 status_manager.go:230] "Starting to sync pod status with apiserver" Jul 9 23:58:41.947966 kubelet[2830]: I0709 23:58:41.947807 2830 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Jul 9 23:58:41.947966 kubelet[2830]: I0709 23:58:41.947811 2830 kubelet.go:2436] "Starting kubelet main sync loop" Jul 9 23:58:41.947966 kubelet[2830]: E0709 23:58:41.947837 2830 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jul 9 23:58:41.971982 kubelet[2830]: I0709 23:58:41.971946 2830 cpu_manager.go:221] "Starting CPU manager" policy="none" Jul 9 23:58:41.971982 kubelet[2830]: I0709 23:58:41.971975 2830 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Jul 9 23:58:41.971982 kubelet[2830]: I0709 23:58:41.971989 2830 state_mem.go:36] "Initialized new in-memory state store" Jul 9 23:58:41.972169 kubelet[2830]: I0709 23:58:41.972067 2830 state_mem.go:88] "Updated default CPUSet" cpuSet="" Jul 9 23:58:41.972169 kubelet[2830]: I0709 23:58:41.972085 2830 state_mem.go:96] "Updated CPUSet assignments" assignments={} Jul 9 23:58:41.972169 kubelet[2830]: I0709 23:58:41.972104 2830 policy_none.go:49] "None policy: Start" Jul 9 23:58:41.972169 kubelet[2830]: I0709 23:58:41.972116 2830 memory_manager.go:186] "Starting memorymanager" policy="None" Jul 9 23:58:41.972169 kubelet[2830]: I0709 23:58:41.972124 2830 state_mem.go:35] "Initializing new in-memory state store" Jul 9 23:58:41.972583 kubelet[2830]: I0709 23:58:41.972181 2830 state_mem.go:75] "Updated machine memory state" Jul 9 23:58:41.974676 kubelet[2830]: E0709 23:58:41.974661 2830 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Jul 9 23:58:41.974946 kubelet[2830]: I0709 23:58:41.974934 2830 eviction_manager.go:189] "Eviction manager: starting control loop" Jul 9 23:58:41.974972 kubelet[2830]: I0709 23:58:41.974944 2830 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jul 9 23:58:41.975798 kubelet[2830]: I0709 23:58:41.975790 2830 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jul 9 23:58:41.985094 kubelet[2830]: E0709 23:58:41.984607 2830 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Jul 9 23:58:42.038529 sudo[2865]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Jul 9 23:58:42.038772 sudo[2865]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) Jul 9 23:58:42.049978 kubelet[2830]: I0709 23:58:42.049899 2830 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Jul 9 23:58:42.050684 kubelet[2830]: I0709 23:58:42.050225 2830 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Jul 9 23:58:42.050684 kubelet[2830]: I0709 23:58:42.050400 2830 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-localhost" Jul 9 23:58:42.079802 kubelet[2830]: I0709 23:58:42.079056 2830 kubelet_node_status.go:75] "Attempting to register node" node="localhost" Jul 9 23:58:42.084523 kubelet[2830]: I0709 23:58:42.084438 2830 kubelet_node_status.go:124] "Node was previously registered" node="localhost" Jul 9 23:58:42.084523 kubelet[2830]: I0709 23:58:42.084495 2830 kubelet_node_status.go:78] "Successfully registered node" node="localhost" Jul 9 23:58:42.114966 kubelet[2830]: I0709 23:58:42.114905 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/a9674aa579a9a6f3b4f62af6707ef78a-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"a9674aa579a9a6f3b4f62af6707ef78a\") " pod="kube-system/kube-apiserver-localhost" Jul 9 23:58:42.114966 kubelet[2830]: I0709 23:58:42.114936 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/84b858ec27c8b2738b1d9ff9927e0dcb-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"84b858ec27c8b2738b1d9ff9927e0dcb\") " pod="kube-system/kube-controller-manager-localhost" Jul 9 23:58:42.114966 kubelet[2830]: I0709 23:58:42.114949 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/834ee54f1daa06092e339273649eb5ea-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"834ee54f1daa06092e339273649eb5ea\") " pod="kube-system/kube-scheduler-localhost" Jul 9 23:58:42.114966 kubelet[2830]: I0709 23:58:42.114959 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/a9674aa579a9a6f3b4f62af6707ef78a-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"a9674aa579a9a6f3b4f62af6707ef78a\") " pod="kube-system/kube-apiserver-localhost" Jul 9 23:58:42.114966 kubelet[2830]: I0709 23:58:42.114968 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/a9674aa579a9a6f3b4f62af6707ef78a-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"a9674aa579a9a6f3b4f62af6707ef78a\") " pod="kube-system/kube-apiserver-localhost" Jul 9 23:58:42.115203 kubelet[2830]: I0709 23:58:42.114979 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/84b858ec27c8b2738b1d9ff9927e0dcb-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"84b858ec27c8b2738b1d9ff9927e0dcb\") " pod="kube-system/kube-controller-manager-localhost" Jul 9 23:58:42.115203 kubelet[2830]: I0709 23:58:42.114987 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/84b858ec27c8b2738b1d9ff9927e0dcb-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"84b858ec27c8b2738b1d9ff9927e0dcb\") " pod="kube-system/kube-controller-manager-localhost" Jul 9 23:58:42.115203 kubelet[2830]: I0709 23:58:42.114996 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/84b858ec27c8b2738b1d9ff9927e0dcb-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"84b858ec27c8b2738b1d9ff9927e0dcb\") " pod="kube-system/kube-controller-manager-localhost" Jul 9 23:58:42.115203 kubelet[2830]: I0709 23:58:42.115012 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/84b858ec27c8b2738b1d9ff9927e0dcb-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"84b858ec27c8b2738b1d9ff9927e0dcb\") " pod="kube-system/kube-controller-manager-localhost" Jul 9 23:58:42.485056 sudo[2865]: pam_unix(sudo:session): session closed for user root Jul 9 23:58:42.831344 kubelet[2830]: I0709 23:58:42.831309 2830 apiserver.go:52] "Watching apiserver" Jul 9 23:58:42.913677 kubelet[2830]: I0709 23:58:42.913640 2830 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Jul 9 23:58:42.959946 kubelet[2830]: I0709 23:58:42.959918 2830 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-localhost" Jul 9 23:58:42.960189 kubelet[2830]: I0709 23:58:42.960168 2830 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-localhost" Jul 9 23:58:43.010439 kubelet[2830]: E0709 23:58:43.010395 2830 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Jul 9 23:58:43.012094 kubelet[2830]: E0709 23:58:43.011204 2830 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost" Jul 9 23:58:43.031562 kubelet[2830]: I0709 23:58:43.031519 2830 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=1.031497864 podStartE2EDuration="1.031497864s" podCreationTimestamp="2025-07-09 23:58:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-09 23:58:43.012519206 +0000 UTC m=+1.312659563" watchObservedRunningTime="2025-07-09 23:58:43.031497864 +0000 UTC m=+1.331638219" Jul 9 23:58:43.048790 kubelet[2830]: I0709 23:58:43.048653 2830 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.048641608 podStartE2EDuration="1.048641608s" podCreationTimestamp="2025-07-09 23:58:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-09 23:58:43.031908929 +0000 UTC m=+1.332049292" watchObservedRunningTime="2025-07-09 23:58:43.048641608 +0000 UTC m=+1.348781966" Jul 9 23:58:44.667055 sudo[1861]: pam_unix(sudo:session): session closed for user root Jul 9 23:58:44.667993 sshd[1860]: Connection closed by 139.178.89.65 port 50298 Jul 9 23:58:44.672898 sshd-session[1857]: pam_unix(sshd:session): session closed for user core Jul 9 23:58:44.675578 systemd[1]: sshd@6-139.178.70.101:22-139.178.89.65:50298.service: Deactivated successfully. Jul 9 23:58:44.677178 systemd[1]: session-9.scope: Deactivated successfully. Jul 9 23:58:44.677367 systemd[1]: session-9.scope: Consumed 4.165s CPU time, 206.6M memory peak. Jul 9 23:58:44.678303 systemd-logind[1542]: Session 9 logged out. Waiting for processes to exit. Jul 9 23:58:44.679200 systemd-logind[1542]: Removed session 9. Jul 9 23:58:45.047254 kubelet[2830]: I0709 23:58:45.047195 2830 kuberuntime_manager.go:1746] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Jul 9 23:58:45.048387 containerd[1562]: time="2025-07-09T23:58:45.047806805Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Jul 9 23:58:45.048575 kubelet[2830]: I0709 23:58:45.047921 2830 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Jul 9 23:58:45.829985 kubelet[2830]: I0709 23:58:45.829944 2830 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=3.829933422 podStartE2EDuration="3.829933422s" podCreationTimestamp="2025-07-09 23:58:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-09 23:58:43.048641175 +0000 UTC m=+1.348781539" watchObservedRunningTime="2025-07-09 23:58:45.829933422 +0000 UTC m=+4.130073785" Jul 9 23:58:45.838179 kubelet[2830]: I0709 23:58:45.837876 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-cilium-run\") pod \"cilium-n6jnq\" (UID: \"9efe80a0-3426-4d84-92fb-e466ff07f979\") " pod="kube-system/cilium-n6jnq" Jul 9 23:58:45.838179 kubelet[2830]: I0709 23:58:45.837898 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-bpf-maps\") pod \"cilium-n6jnq\" (UID: \"9efe80a0-3426-4d84-92fb-e466ff07f979\") " pod="kube-system/cilium-n6jnq" Jul 9 23:58:45.838179 kubelet[2830]: I0709 23:58:45.837909 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-lib-modules\") pod \"cilium-n6jnq\" (UID: \"9efe80a0-3426-4d84-92fb-e466ff07f979\") " pod="kube-system/cilium-n6jnq" Jul 9 23:58:45.838179 kubelet[2830]: I0709 23:58:45.837935 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/9efe80a0-3426-4d84-92fb-e466ff07f979-cilium-config-path\") pod \"cilium-n6jnq\" (UID: \"9efe80a0-3426-4d84-92fb-e466ff07f979\") " pod="kube-system/cilium-n6jnq" Jul 9 23:58:45.838179 kubelet[2830]: I0709 23:58:45.837951 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-xtables-lock\") pod \"cilium-n6jnq\" (UID: \"9efe80a0-3426-4d84-92fb-e466ff07f979\") " pod="kube-system/cilium-n6jnq" Jul 9 23:58:45.838179 kubelet[2830]: I0709 23:58:45.837961 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-host-proc-sys-net\") pod \"cilium-n6jnq\" (UID: \"9efe80a0-3426-4d84-92fb-e466ff07f979\") " pod="kube-system/cilium-n6jnq" Jul 9 23:58:45.838353 kubelet[2830]: I0709 23:58:45.837970 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-host-proc-sys-kernel\") pod \"cilium-n6jnq\" (UID: \"9efe80a0-3426-4d84-92fb-e466ff07f979\") " pod="kube-system/cilium-n6jnq" Jul 9 23:58:45.838353 kubelet[2830]: I0709 23:58:45.837979 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/9efe80a0-3426-4d84-92fb-e466ff07f979-hubble-tls\") pod \"cilium-n6jnq\" (UID: \"9efe80a0-3426-4d84-92fb-e466ff07f979\") " pod="kube-system/cilium-n6jnq" Jul 9 23:58:45.838353 kubelet[2830]: I0709 23:58:45.837989 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ndbt\" (UniqueName: \"kubernetes.io/projected/9efe80a0-3426-4d84-92fb-e466ff07f979-kube-api-access-4ndbt\") pod \"cilium-n6jnq\" (UID: \"9efe80a0-3426-4d84-92fb-e466ff07f979\") " pod="kube-system/cilium-n6jnq" Jul 9 23:58:45.838353 kubelet[2830]: I0709 23:58:45.838001 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/9efe80a0-3426-4d84-92fb-e466ff07f979-clustermesh-secrets\") pod \"cilium-n6jnq\" (UID: \"9efe80a0-3426-4d84-92fb-e466ff07f979\") " pod="kube-system/cilium-n6jnq" Jul 9 23:58:45.838353 kubelet[2830]: I0709 23:58:45.838009 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-hostproc\") pod \"cilium-n6jnq\" (UID: \"9efe80a0-3426-4d84-92fb-e466ff07f979\") " pod="kube-system/cilium-n6jnq" Jul 9 23:58:45.838353 kubelet[2830]: I0709 23:58:45.838017 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-cilium-cgroup\") pod \"cilium-n6jnq\" (UID: \"9efe80a0-3426-4d84-92fb-e466ff07f979\") " pod="kube-system/cilium-n6jnq" Jul 9 23:58:45.838451 kubelet[2830]: I0709 23:58:45.838028 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-cni-path\") pod \"cilium-n6jnq\" (UID: \"9efe80a0-3426-4d84-92fb-e466ff07f979\") " pod="kube-system/cilium-n6jnq" Jul 9 23:58:45.838451 kubelet[2830]: I0709 23:58:45.838037 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-etc-cni-netd\") pod \"cilium-n6jnq\" (UID: \"9efe80a0-3426-4d84-92fb-e466ff07f979\") " pod="kube-system/cilium-n6jnq" Jul 9 23:58:45.844469 systemd[1]: Created slice kubepods-besteffort-podbb0ae745_95a0_40d0_82dd_01c63c245c31.slice - libcontainer container kubepods-besteffort-podbb0ae745_95a0_40d0_82dd_01c63c245c31.slice. Jul 9 23:58:45.852098 systemd[1]: Created slice kubepods-burstable-pod9efe80a0_3426_4d84_92fb_e466ff07f979.slice - libcontainer container kubepods-burstable-pod9efe80a0_3426_4d84_92fb_e466ff07f979.slice. Jul 9 23:58:45.938332 kubelet[2830]: I0709 23:58:45.938137 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/bb0ae745-95a0-40d0-82dd-01c63c245c31-kube-proxy\") pod \"kube-proxy-v4mjn\" (UID: \"bb0ae745-95a0-40d0-82dd-01c63c245c31\") " pod="kube-system/kube-proxy-v4mjn" Jul 9 23:58:45.938332 kubelet[2830]: I0709 23:58:45.938160 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/bb0ae745-95a0-40d0-82dd-01c63c245c31-xtables-lock\") pod \"kube-proxy-v4mjn\" (UID: \"bb0ae745-95a0-40d0-82dd-01c63c245c31\") " pod="kube-system/kube-proxy-v4mjn" Jul 9 23:58:45.938332 kubelet[2830]: I0709 23:58:45.938221 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ll94\" (UniqueName: \"kubernetes.io/projected/bb0ae745-95a0-40d0-82dd-01c63c245c31-kube-api-access-2ll94\") pod \"kube-proxy-v4mjn\" (UID: \"bb0ae745-95a0-40d0-82dd-01c63c245c31\") " pod="kube-system/kube-proxy-v4mjn" Jul 9 23:58:45.938332 kubelet[2830]: I0709 23:58:45.938236 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bb0ae745-95a0-40d0-82dd-01c63c245c31-lib-modules\") pod \"kube-proxy-v4mjn\" (UID: \"bb0ae745-95a0-40d0-82dd-01c63c245c31\") " pod="kube-system/kube-proxy-v4mjn" Jul 9 23:58:45.953761 kubelet[2830]: E0709 23:58:45.953332 2830 projected.go:289] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found Jul 9 23:58:45.953761 kubelet[2830]: E0709 23:58:45.953352 2830 projected.go:194] Error preparing data for projected volume kube-api-access-4ndbt for pod kube-system/cilium-n6jnq: configmap "kube-root-ca.crt" not found Jul 9 23:58:45.953761 kubelet[2830]: E0709 23:58:45.953403 2830 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9efe80a0-3426-4d84-92fb-e466ff07f979-kube-api-access-4ndbt podName:9efe80a0-3426-4d84-92fb-e466ff07f979 nodeName:}" failed. No retries permitted until 2025-07-09 23:58:46.453389553 +0000 UTC m=+4.753529907 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-4ndbt" (UniqueName: "kubernetes.io/projected/9efe80a0-3426-4d84-92fb-e466ff07f979-kube-api-access-4ndbt") pod "cilium-n6jnq" (UID: "9efe80a0-3426-4d84-92fb-e466ff07f979") : configmap "kube-root-ca.crt" not found Jul 9 23:58:46.153015 containerd[1562]: time="2025-07-09T23:58:46.152689831Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-v4mjn,Uid:bb0ae745-95a0-40d0-82dd-01c63c245c31,Namespace:kube-system,Attempt:0,}" Jul 9 23:58:46.172105 systemd[1]: Created slice kubepods-besteffort-pod03702575_51ea_44b7_a61a_3a443ff54801.slice - libcontainer container kubepods-besteffort-pod03702575_51ea_44b7_a61a_3a443ff54801.slice. Jul 9 23:58:46.223962 containerd[1562]: time="2025-07-09T23:58:46.223876455Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 9 23:58:46.224272 containerd[1562]: time="2025-07-09T23:58:46.224171367Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 9 23:58:46.224362 containerd[1562]: time="2025-07-09T23:58:46.224232194Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 9 23:58:46.224991 containerd[1562]: time="2025-07-09T23:58:46.224919737Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 9 23:58:46.239878 kubelet[2830]: I0709 23:58:46.239847 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffklt\" (UniqueName: \"kubernetes.io/projected/03702575-51ea-44b7-a61a-3a443ff54801-kube-api-access-ffklt\") pod \"cilium-operator-6c4d7847fc-xmh74\" (UID: \"03702575-51ea-44b7-a61a-3a443ff54801\") " pod="kube-system/cilium-operator-6c4d7847fc-xmh74" Jul 9 23:58:46.241409 kubelet[2830]: I0709 23:58:46.239893 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/03702575-51ea-44b7-a61a-3a443ff54801-cilium-config-path\") pod \"cilium-operator-6c4d7847fc-xmh74\" (UID: \"03702575-51ea-44b7-a61a-3a443ff54801\") " pod="kube-system/cilium-operator-6c4d7847fc-xmh74" Jul 9 23:58:46.241183 systemd[1]: Started cri-containerd-8ae364ab6509723f384f07758589e552e2a857c581d2d16794faed82eb455e11.scope - libcontainer container 8ae364ab6509723f384f07758589e552e2a857c581d2d16794faed82eb455e11. Jul 9 23:58:46.256683 containerd[1562]: time="2025-07-09T23:58:46.256652834Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-v4mjn,Uid:bb0ae745-95a0-40d0-82dd-01c63c245c31,Namespace:kube-system,Attempt:0,} returns sandbox id \"8ae364ab6509723f384f07758589e552e2a857c581d2d16794faed82eb455e11\"" Jul 9 23:58:46.274596 containerd[1562]: time="2025-07-09T23:58:46.274564800Z" level=info msg="CreateContainer within sandbox \"8ae364ab6509723f384f07758589e552e2a857c581d2d16794faed82eb455e11\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Jul 9 23:58:46.474500 containerd[1562]: time="2025-07-09T23:58:46.474276312Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-xmh74,Uid:03702575-51ea-44b7-a61a-3a443ff54801,Namespace:kube-system,Attempt:0,}" Jul 9 23:58:46.620242 containerd[1562]: time="2025-07-09T23:58:46.620201275Z" level=info msg="CreateContainer within sandbox \"8ae364ab6509723f384f07758589e552e2a857c581d2d16794faed82eb455e11\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"d6f7c6ee8ec161a3fb7c35a1892ba0deb9990ded3b56cbbb80b7f734546a0879\"" Jul 9 23:58:46.620835 containerd[1562]: time="2025-07-09T23:58:46.620807902Z" level=info msg="StartContainer for \"d6f7c6ee8ec161a3fb7c35a1892ba0deb9990ded3b56cbbb80b7f734546a0879\"" Jul 9 23:58:46.644851 containerd[1562]: time="2025-07-09T23:58:46.644751963Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 9 23:58:46.644851 containerd[1562]: time="2025-07-09T23:58:46.644793554Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 9 23:58:46.644851 containerd[1562]: time="2025-07-09T23:58:46.644802317Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 9 23:58:46.645104 containerd[1562]: time="2025-07-09T23:58:46.644860930Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 9 23:58:46.645910 systemd[1]: Started cri-containerd-d6f7c6ee8ec161a3fb7c35a1892ba0deb9990ded3b56cbbb80b7f734546a0879.scope - libcontainer container d6f7c6ee8ec161a3fb7c35a1892ba0deb9990ded3b56cbbb80b7f734546a0879. Jul 9 23:58:46.660236 systemd[1]: Started cri-containerd-2951185f2d2f0f3fa6ecbc5a8060cf32c35ad336b49440efdede342adc870e00.scope - libcontainer container 2951185f2d2f0f3fa6ecbc5a8060cf32c35ad336b49440efdede342adc870e00. Jul 9 23:58:46.684383 containerd[1562]: time="2025-07-09T23:58:46.684352360Z" level=info msg="StartContainer for \"d6f7c6ee8ec161a3fb7c35a1892ba0deb9990ded3b56cbbb80b7f734546a0879\" returns successfully" Jul 9 23:58:46.699110 containerd[1562]: time="2025-07-09T23:58:46.699052165Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-xmh74,Uid:03702575-51ea-44b7-a61a-3a443ff54801,Namespace:kube-system,Attempt:0,} returns sandbox id \"2951185f2d2f0f3fa6ecbc5a8060cf32c35ad336b49440efdede342adc870e00\"" Jul 9 23:58:46.700409 containerd[1562]: time="2025-07-09T23:58:46.700380586Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Jul 9 23:58:46.755376 containerd[1562]: time="2025-07-09T23:58:46.755282226Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-n6jnq,Uid:9efe80a0-3426-4d84-92fb-e466ff07f979,Namespace:kube-system,Attempt:0,}" Jul 9 23:58:46.840446 containerd[1562]: time="2025-07-09T23:58:46.840308772Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 9 23:58:46.840446 containerd[1562]: time="2025-07-09T23:58:46.840381407Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 9 23:58:46.840628 containerd[1562]: time="2025-07-09T23:58:46.840444473Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 9 23:58:46.840628 containerd[1562]: time="2025-07-09T23:58:46.840523765Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 9 23:58:46.852179 systemd[1]: Started cri-containerd-f58f391fd85e24d2648281b4945bb4e7390d20b9d624e85bd712e8c100a4ec28.scope - libcontainer container f58f391fd85e24d2648281b4945bb4e7390d20b9d624e85bd712e8c100a4ec28. Jul 9 23:58:46.866484 containerd[1562]: time="2025-07-09T23:58:46.866455157Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-n6jnq,Uid:9efe80a0-3426-4d84-92fb-e466ff07f979,Namespace:kube-system,Attempt:0,} returns sandbox id \"f58f391fd85e24d2648281b4945bb4e7390d20b9d624e85bd712e8c100a4ec28\"" Jul 9 23:58:46.998180 kubelet[2830]: I0709 23:58:46.998138 2830 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-v4mjn" podStartSLOduration=1.998128905 podStartE2EDuration="1.998128905s" podCreationTimestamp="2025-07-09 23:58:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-09 23:58:46.997812768 +0000 UTC m=+5.297953126" watchObservedRunningTime="2025-07-09 23:58:46.998128905 +0000 UTC m=+5.298269268" Jul 9 23:58:48.702744 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4224728402.mount: Deactivated successfully. Jul 9 23:58:51.250818 containerd[1562]: time="2025-07-09T23:58:51.250777454Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 9 23:58:51.271229 containerd[1562]: time="2025-07-09T23:58:51.271028133Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=18904197" Jul 9 23:58:51.312738 containerd[1562]: time="2025-07-09T23:58:51.312677493Z" level=info msg="ImageCreate event name:\"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 9 23:58:51.313577 containerd[1562]: time="2025-07-09T23:58:51.313483050Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"18897442\" in 4.613077107s" Jul 9 23:58:51.313577 containerd[1562]: time="2025-07-09T23:58:51.313507273Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" Jul 9 23:58:51.410392 containerd[1562]: time="2025-07-09T23:58:51.410216857Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Jul 9 23:58:51.437116 containerd[1562]: time="2025-07-09T23:58:51.437095037Z" level=info msg="CreateContainer within sandbox \"2951185f2d2f0f3fa6ecbc5a8060cf32c35ad336b49440efdede342adc870e00\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Jul 9 23:58:51.646429 containerd[1562]: time="2025-07-09T23:58:51.646342466Z" level=info msg="CreateContainer within sandbox \"2951185f2d2f0f3fa6ecbc5a8060cf32c35ad336b49440efdede342adc870e00\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"a43a4694ee5216f4d4a2826d38fea0b0bbc43feadabf1494f475086735f473d8\"" Jul 9 23:58:51.647245 containerd[1562]: time="2025-07-09T23:58:51.646756165Z" level=info msg="StartContainer for \"a43a4694ee5216f4d4a2826d38fea0b0bbc43feadabf1494f475086735f473d8\"" Jul 9 23:58:51.674247 systemd[1]: Started cri-containerd-a43a4694ee5216f4d4a2826d38fea0b0bbc43feadabf1494f475086735f473d8.scope - libcontainer container a43a4694ee5216f4d4a2826d38fea0b0bbc43feadabf1494f475086735f473d8. Jul 9 23:58:51.706717 containerd[1562]: time="2025-07-09T23:58:51.706686259Z" level=info msg="StartContainer for \"a43a4694ee5216f4d4a2826d38fea0b0bbc43feadabf1494f475086735f473d8\" returns successfully" Jul 9 23:58:53.617801 kubelet[2830]: I0709 23:58:53.617763 2830 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-6c4d7847fc-xmh74" podStartSLOduration=2.908055547 podStartE2EDuration="7.61774709s" podCreationTimestamp="2025-07-09 23:58:46 +0000 UTC" firstStartedPulling="2025-07-09 23:58:46.699930172 +0000 UTC m=+5.000070524" lastFinishedPulling="2025-07-09 23:58:51.409621713 +0000 UTC m=+9.709762067" observedRunningTime="2025-07-09 23:58:52.174702934 +0000 UTC m=+10.474843298" watchObservedRunningTime="2025-07-09 23:58:53.61774709 +0000 UTC m=+11.917887447" Jul 9 23:59:00.913479 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3892486668.mount: Deactivated successfully. Jul 9 23:59:05.348116 containerd[1562]: time="2025-07-09T23:59:05.315732929Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 9 23:59:05.616952 containerd[1562]: time="2025-07-09T23:59:05.616791380Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=166730503" Jul 9 23:59:05.828205 containerd[1562]: time="2025-07-09T23:59:05.828161544Z" level=info msg="ImageCreate event name:\"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jul 9 23:59:06.153631 containerd[1562]: time="2025-07-09T23:59:06.153420026Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"166719855\" in 14.743166601s" Jul 9 23:59:06.153631 containerd[1562]: time="2025-07-09T23:59:06.153473086Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" Jul 9 23:59:06.240582 containerd[1562]: time="2025-07-09T23:59:06.240446915Z" level=info msg="CreateContainer within sandbox \"f58f391fd85e24d2648281b4945bb4e7390d20b9d624e85bd712e8c100a4ec28\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Jul 9 23:59:06.296285 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3069920998.mount: Deactivated successfully. Jul 9 23:59:06.304120 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1213285987.mount: Deactivated successfully. Jul 9 23:59:06.341848 containerd[1562]: time="2025-07-09T23:59:06.341814436Z" level=info msg="CreateContainer within sandbox \"f58f391fd85e24d2648281b4945bb4e7390d20b9d624e85bd712e8c100a4ec28\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"abbae4097b8ce16b67b85924c3b963388218cc676731b56a3385d272ef7088e9\"" Jul 9 23:59:06.342307 containerd[1562]: time="2025-07-09T23:59:06.342166382Z" level=info msg="StartContainer for \"abbae4097b8ce16b67b85924c3b963388218cc676731b56a3385d272ef7088e9\"" Jul 9 23:59:06.557185 systemd[1]: Started cri-containerd-abbae4097b8ce16b67b85924c3b963388218cc676731b56a3385d272ef7088e9.scope - libcontainer container abbae4097b8ce16b67b85924c3b963388218cc676731b56a3385d272ef7088e9. Jul 9 23:59:06.595572 containerd[1562]: time="2025-07-09T23:59:06.595544288Z" level=info msg="StartContainer for \"abbae4097b8ce16b67b85924c3b963388218cc676731b56a3385d272ef7088e9\" returns successfully" Jul 9 23:59:06.648409 systemd[1]: cri-containerd-abbae4097b8ce16b67b85924c3b963388218cc676731b56a3385d272ef7088e9.scope: Deactivated successfully. Jul 9 23:59:07.292012 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-abbae4097b8ce16b67b85924c3b963388218cc676731b56a3385d272ef7088e9-rootfs.mount: Deactivated successfully. Jul 9 23:59:07.681593 containerd[1562]: time="2025-07-09T23:59:07.681538501Z" level=info msg="shim disconnected" id=abbae4097b8ce16b67b85924c3b963388218cc676731b56a3385d272ef7088e9 namespace=k8s.io Jul 9 23:59:07.682169 containerd[1562]: time="2025-07-09T23:59:07.682001867Z" level=warning msg="cleaning up after shim disconnected" id=abbae4097b8ce16b67b85924c3b963388218cc676731b56a3385d272ef7088e9 namespace=k8s.io Jul 9 23:59:07.682169 containerd[1562]: time="2025-07-09T23:59:07.682017412Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jul 9 23:59:08.317278 containerd[1562]: time="2025-07-09T23:59:08.317172592Z" level=info msg="CreateContainer within sandbox \"f58f391fd85e24d2648281b4945bb4e7390d20b9d624e85bd712e8c100a4ec28\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Jul 9 23:59:08.397997 containerd[1562]: time="2025-07-09T23:59:08.397959931Z" level=info msg="CreateContainer within sandbox \"f58f391fd85e24d2648281b4945bb4e7390d20b9d624e85bd712e8c100a4ec28\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"85452c6ec0b3f70f0b3549025557efb2e2eec40926df16ecff76c9276cc308d4\"" Jul 9 23:59:08.399102 containerd[1562]: time="2025-07-09T23:59:08.398390523Z" level=info msg="StartContainer for \"85452c6ec0b3f70f0b3549025557efb2e2eec40926df16ecff76c9276cc308d4\"" Jul 9 23:59:08.420190 systemd[1]: Started cri-containerd-85452c6ec0b3f70f0b3549025557efb2e2eec40926df16ecff76c9276cc308d4.scope - libcontainer container 85452c6ec0b3f70f0b3549025557efb2e2eec40926df16ecff76c9276cc308d4. Jul 9 23:59:08.444757 containerd[1562]: time="2025-07-09T23:59:08.444726562Z" level=info msg="StartContainer for \"85452c6ec0b3f70f0b3549025557efb2e2eec40926df16ecff76c9276cc308d4\" returns successfully" Jul 9 23:59:08.595698 systemd[1]: systemd-sysctl.service: Deactivated successfully. Jul 9 23:59:08.595854 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Jul 9 23:59:08.596310 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... Jul 9 23:59:08.601556 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jul 9 23:59:08.603184 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Jul 9 23:59:08.603813 systemd[1]: cri-containerd-85452c6ec0b3f70f0b3549025557efb2e2eec40926df16ecff76c9276cc308d4.scope: Deactivated successfully. Jul 9 23:59:08.785692 containerd[1562]: time="2025-07-09T23:59:08.785653673Z" level=info msg="shim disconnected" id=85452c6ec0b3f70f0b3549025557efb2e2eec40926df16ecff76c9276cc308d4 namespace=k8s.io Jul 9 23:59:08.785692 containerd[1562]: time="2025-07-09T23:59:08.785688936Z" level=warning msg="cleaning up after shim disconnected" id=85452c6ec0b3f70f0b3549025557efb2e2eec40926df16ecff76c9276cc308d4 namespace=k8s.io Jul 9 23:59:08.785692 containerd[1562]: time="2025-07-09T23:59:08.785697335Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jul 9 23:59:08.787339 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jul 9 23:59:09.271707 containerd[1562]: time="2025-07-09T23:59:09.271351252Z" level=info msg="CreateContainer within sandbox \"f58f391fd85e24d2648281b4945bb4e7390d20b9d624e85bd712e8c100a4ec28\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Jul 9 23:59:09.340457 containerd[1562]: time="2025-07-09T23:59:09.340425595Z" level=info msg="CreateContainer within sandbox \"f58f391fd85e24d2648281b4945bb4e7390d20b9d624e85bd712e8c100a4ec28\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"0bb7627fe8e47d5a2abdcda9b06dca14ee9749ad1b159e59b387545f92d8e706\"" Jul 9 23:59:09.341302 containerd[1562]: time="2025-07-09T23:59:09.340889626Z" level=info msg="StartContainer for \"0bb7627fe8e47d5a2abdcda9b06dca14ee9749ad1b159e59b387545f92d8e706\"" Jul 9 23:59:09.356252 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-85452c6ec0b3f70f0b3549025557efb2e2eec40926df16ecff76c9276cc308d4-rootfs.mount: Deactivated successfully. Jul 9 23:59:09.360197 systemd[1]: Started cri-containerd-0bb7627fe8e47d5a2abdcda9b06dca14ee9749ad1b159e59b387545f92d8e706.scope - libcontainer container 0bb7627fe8e47d5a2abdcda9b06dca14ee9749ad1b159e59b387545f92d8e706. Jul 9 23:59:09.429348 containerd[1562]: time="2025-07-09T23:59:09.429294236Z" level=info msg="StartContainer for \"0bb7627fe8e47d5a2abdcda9b06dca14ee9749ad1b159e59b387545f92d8e706\" returns successfully" Jul 9 23:59:09.506961 systemd[1]: cri-containerd-0bb7627fe8e47d5a2abdcda9b06dca14ee9749ad1b159e59b387545f92d8e706.scope: Deactivated successfully. Jul 9 23:59:09.507334 systemd[1]: cri-containerd-0bb7627fe8e47d5a2abdcda9b06dca14ee9749ad1b159e59b387545f92d8e706.scope: Consumed 13ms CPU time, 5.5M memory peak, 1.2M read from disk. Jul 9 23:59:09.523317 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-0bb7627fe8e47d5a2abdcda9b06dca14ee9749ad1b159e59b387545f92d8e706-rootfs.mount: Deactivated successfully. Jul 9 23:59:09.615449 containerd[1562]: time="2025-07-09T23:59:09.615373706Z" level=info msg="shim disconnected" id=0bb7627fe8e47d5a2abdcda9b06dca14ee9749ad1b159e59b387545f92d8e706 namespace=k8s.io Jul 9 23:59:09.615449 containerd[1562]: time="2025-07-09T23:59:09.615441630Z" level=warning msg="cleaning up after shim disconnected" id=0bb7627fe8e47d5a2abdcda9b06dca14ee9749ad1b159e59b387545f92d8e706 namespace=k8s.io Jul 9 23:59:09.615449 containerd[1562]: time="2025-07-09T23:59:09.615451950Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jul 9 23:59:10.304153 containerd[1562]: time="2025-07-09T23:59:10.304112021Z" level=info msg="CreateContainer within sandbox \"f58f391fd85e24d2648281b4945bb4e7390d20b9d624e85bd712e8c100a4ec28\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Jul 9 23:59:10.371908 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3773742419.mount: Deactivated successfully. Jul 9 23:59:10.433625 containerd[1562]: time="2025-07-09T23:59:10.433590045Z" level=info msg="CreateContainer within sandbox \"f58f391fd85e24d2648281b4945bb4e7390d20b9d624e85bd712e8c100a4ec28\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"a419a7595cf002be07974654bec25192ff733dda08bac28ec25a4c624402a4df\"" Jul 9 23:59:10.434221 containerd[1562]: time="2025-07-09T23:59:10.434104204Z" level=info msg="StartContainer for \"a419a7595cf002be07974654bec25192ff733dda08bac28ec25a4c624402a4df\"" Jul 9 23:59:10.457166 systemd[1]: Started cri-containerd-a419a7595cf002be07974654bec25192ff733dda08bac28ec25a4c624402a4df.scope - libcontainer container a419a7595cf002be07974654bec25192ff733dda08bac28ec25a4c624402a4df. Jul 9 23:59:10.494096 containerd[1562]: time="2025-07-09T23:59:10.494039973Z" level=info msg="StartContainer for \"a419a7595cf002be07974654bec25192ff733dda08bac28ec25a4c624402a4df\" returns successfully" Jul 9 23:59:10.506475 systemd[1]: cri-containerd-a419a7595cf002be07974654bec25192ff733dda08bac28ec25a4c624402a4df.scope: Deactivated successfully. Jul 9 23:59:10.519213 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a419a7595cf002be07974654bec25192ff733dda08bac28ec25a4c624402a4df-rootfs.mount: Deactivated successfully. Jul 9 23:59:10.586615 containerd[1562]: time="2025-07-09T23:59:10.586574516Z" level=info msg="shim disconnected" id=a419a7595cf002be07974654bec25192ff733dda08bac28ec25a4c624402a4df namespace=k8s.io Jul 9 23:59:10.586615 containerd[1562]: time="2025-07-09T23:59:10.586609215Z" level=warning msg="cleaning up after shim disconnected" id=a419a7595cf002be07974654bec25192ff733dda08bac28ec25a4c624402a4df namespace=k8s.io Jul 9 23:59:10.586615 containerd[1562]: time="2025-07-09T23:59:10.586614497Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jul 9 23:59:11.266978 containerd[1562]: time="2025-07-09T23:59:11.266928020Z" level=info msg="CreateContainer within sandbox \"f58f391fd85e24d2648281b4945bb4e7390d20b9d624e85bd712e8c100a4ec28\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Jul 9 23:59:11.276284 containerd[1562]: time="2025-07-09T23:59:11.276198026Z" level=info msg="CreateContainer within sandbox \"f58f391fd85e24d2648281b4945bb4e7390d20b9d624e85bd712e8c100a4ec28\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"caf02a3680d5145fd262b9738a10659ded2a04de94f9c3db1c4a81285572e9ad\"" Jul 9 23:59:11.276681 containerd[1562]: time="2025-07-09T23:59:11.276662103Z" level=info msg="StartContainer for \"caf02a3680d5145fd262b9738a10659ded2a04de94f9c3db1c4a81285572e9ad\"" Jul 9 23:59:11.305263 systemd[1]: Started cri-containerd-caf02a3680d5145fd262b9738a10659ded2a04de94f9c3db1c4a81285572e9ad.scope - libcontainer container caf02a3680d5145fd262b9738a10659ded2a04de94f9c3db1c4a81285572e9ad. Jul 9 23:59:11.328400 containerd[1562]: time="2025-07-09T23:59:11.328226480Z" level=info msg="StartContainer for \"caf02a3680d5145fd262b9738a10659ded2a04de94f9c3db1c4a81285572e9ad\" returns successfully" Jul 9 23:59:11.516188 kubelet[2830]: I0709 23:59:11.510199 2830 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Jul 9 23:59:11.670530 systemd[1]: Created slice kubepods-burstable-podb58d55a4_0144_408b_af4c_f07b59cfbb81.slice - libcontainer container kubepods-burstable-podb58d55a4_0144_408b_af4c_f07b59cfbb81.slice. Jul 9 23:59:11.674783 systemd[1]: Created slice kubepods-burstable-pode8aeaca1_e1a7_4003_a88c_3cc747dcd8b5.slice - libcontainer container kubepods-burstable-pode8aeaca1_e1a7_4003_a88c_3cc747dcd8b5.slice. Jul 9 23:59:11.783426 kubelet[2830]: I0709 23:59:11.783254 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5hz6\" (UniqueName: \"kubernetes.io/projected/b58d55a4-0144-408b-af4c-f07b59cfbb81-kube-api-access-p5hz6\") pod \"coredns-674b8bbfcf-w9rnc\" (UID: \"b58d55a4-0144-408b-af4c-f07b59cfbb81\") " pod="kube-system/coredns-674b8bbfcf-w9rnc" Jul 9 23:59:11.783426 kubelet[2830]: I0709 23:59:11.783329 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b58d55a4-0144-408b-af4c-f07b59cfbb81-config-volume\") pod \"coredns-674b8bbfcf-w9rnc\" (UID: \"b58d55a4-0144-408b-af4c-f07b59cfbb81\") " pod="kube-system/coredns-674b8bbfcf-w9rnc" Jul 9 23:59:11.783426 kubelet[2830]: I0709 23:59:11.783347 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmlbb\" (UniqueName: \"kubernetes.io/projected/e8aeaca1-e1a7-4003-a88c-3cc747dcd8b5-kube-api-access-fmlbb\") pod \"coredns-674b8bbfcf-7lgtq\" (UID: \"e8aeaca1-e1a7-4003-a88c-3cc747dcd8b5\") " pod="kube-system/coredns-674b8bbfcf-7lgtq" Jul 9 23:59:11.783426 kubelet[2830]: I0709 23:59:11.783359 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e8aeaca1-e1a7-4003-a88c-3cc747dcd8b5-config-volume\") pod \"coredns-674b8bbfcf-7lgtq\" (UID: \"e8aeaca1-e1a7-4003-a88c-3cc747dcd8b5\") " pod="kube-system/coredns-674b8bbfcf-7lgtq" Jul 9 23:59:11.974767 containerd[1562]: time="2025-07-09T23:59:11.974706771Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-w9rnc,Uid:b58d55a4-0144-408b-af4c-f07b59cfbb81,Namespace:kube-system,Attempt:0,}" Jul 9 23:59:11.977879 containerd[1562]: time="2025-07-09T23:59:11.977510534Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-7lgtq,Uid:e8aeaca1-e1a7-4003-a88c-3cc747dcd8b5,Namespace:kube-system,Attempt:0,}" Jul 9 23:59:13.671565 systemd-networkd[1465]: cilium_host: Link UP Jul 9 23:59:13.672529 systemd-networkd[1465]: cilium_net: Link UP Jul 9 23:59:13.672648 systemd-networkd[1465]: cilium_net: Gained carrier Jul 9 23:59:13.672738 systemd-networkd[1465]: cilium_host: Gained carrier Jul 9 23:59:13.760167 systemd-networkd[1465]: cilium_host: Gained IPv6LL Jul 9 23:59:13.792156 systemd-networkd[1465]: cilium_net: Gained IPv6LL Jul 9 23:59:13.793062 systemd-networkd[1465]: cilium_vxlan: Link UP Jul 9 23:59:13.793184 systemd-networkd[1465]: cilium_vxlan: Gained carrier Jul 9 23:59:14.180152 kernel: NET: Registered PF_ALG protocol family Jul 9 23:59:14.622033 systemd-networkd[1465]: lxc_health: Link UP Jul 9 23:59:14.641144 systemd-networkd[1465]: lxc_health: Gained carrier Jul 9 23:59:14.788850 kubelet[2830]: I0709 23:59:14.778973 2830 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-n6jnq" podStartSLOduration=10.481389605 podStartE2EDuration="29.77050658s" podCreationTimestamp="2025-07-09 23:58:45 +0000 UTC" firstStartedPulling="2025-07-09 23:58:46.867053316 +0000 UTC m=+5.167193670" lastFinishedPulling="2025-07-09 23:59:06.15617029 +0000 UTC m=+24.456310645" observedRunningTime="2025-07-09 23:59:12.279313306 +0000 UTC m=+30.579453671" watchObservedRunningTime="2025-07-09 23:59:14.77050658 +0000 UTC m=+33.070646936" Jul 9 23:59:15.040094 kernel: eth0: renamed from tmpf73e7 Jul 9 23:59:15.059092 kernel: eth0: renamed from tmpf214a Jul 9 23:59:15.066662 systemd-networkd[1465]: lxc85b3c8f3d569: Link UP Jul 9 23:59:15.067256 systemd-networkd[1465]: lxc9a8ed2415913: Link UP Jul 9 23:59:15.067436 systemd-networkd[1465]: lxc85b3c8f3d569: Gained carrier Jul 9 23:59:15.067523 systemd-networkd[1465]: lxc9a8ed2415913: Gained carrier Jul 9 23:59:15.304191 systemd-networkd[1465]: cilium_vxlan: Gained IPv6LL Jul 9 23:59:15.944243 systemd-networkd[1465]: lxc_health: Gained IPv6LL Jul 9 23:59:16.264229 systemd-networkd[1465]: lxc9a8ed2415913: Gained IPv6LL Jul 9 23:59:16.328203 systemd-networkd[1465]: lxc85b3c8f3d569: Gained IPv6LL Jul 9 23:59:17.759888 containerd[1562]: time="2025-07-09T23:59:17.759487121Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 9 23:59:17.759888 containerd[1562]: time="2025-07-09T23:59:17.759653692Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 9 23:59:17.759888 containerd[1562]: time="2025-07-09T23:59:17.759676732Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 9 23:59:17.759888 containerd[1562]: time="2025-07-09T23:59:17.759832024Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 9 23:59:17.792259 systemd[1]: Started cri-containerd-f214a72b3c24ec0f9f7efcfb2d817b23125f94a032ce3a215d2b6f5e62dad22e.scope - libcontainer container f214a72b3c24ec0f9f7efcfb2d817b23125f94a032ce3a215d2b6f5e62dad22e. Jul 9 23:59:17.810070 systemd-resolved[1467]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jul 9 23:59:17.836331 containerd[1562]: time="2025-07-09T23:59:17.836290433Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-7lgtq,Uid:e8aeaca1-e1a7-4003-a88c-3cc747dcd8b5,Namespace:kube-system,Attempt:0,} returns sandbox id \"f214a72b3c24ec0f9f7efcfb2d817b23125f94a032ce3a215d2b6f5e62dad22e\"" Jul 9 23:59:17.839963 containerd[1562]: time="2025-07-09T23:59:17.839912539Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 9 23:59:17.840683 containerd[1562]: time="2025-07-09T23:59:17.840574685Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 9 23:59:17.841298 containerd[1562]: time="2025-07-09T23:59:17.840608207Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 9 23:59:17.841298 containerd[1562]: time="2025-07-09T23:59:17.841263403Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 9 23:59:17.857385 containerd[1562]: time="2025-07-09T23:59:17.857211087Z" level=info msg="CreateContainer within sandbox \"f214a72b3c24ec0f9f7efcfb2d817b23125f94a032ce3a215d2b6f5e62dad22e\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jul 9 23:59:17.860187 systemd[1]: Started cri-containerd-f73e706d10875d68b00e3095ffef3c4dafe649b2ac207f26c404e06e50e0ee03.scope - libcontainer container f73e706d10875d68b00e3095ffef3c4dafe649b2ac207f26c404e06e50e0ee03. Jul 9 23:59:17.869539 systemd-resolved[1467]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jul 9 23:59:17.890719 containerd[1562]: time="2025-07-09T23:59:17.890690688Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-w9rnc,Uid:b58d55a4-0144-408b-af4c-f07b59cfbb81,Namespace:kube-system,Attempt:0,} returns sandbox id \"f73e706d10875d68b00e3095ffef3c4dafe649b2ac207f26c404e06e50e0ee03\"" Jul 9 23:59:17.903384 containerd[1562]: time="2025-07-09T23:59:17.903292132Z" level=info msg="CreateContainer within sandbox \"f73e706d10875d68b00e3095ffef3c4dafe649b2ac207f26c404e06e50e0ee03\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jul 9 23:59:17.932757 containerd[1562]: time="2025-07-09T23:59:17.932702362Z" level=info msg="CreateContainer within sandbox \"f73e706d10875d68b00e3095ffef3c4dafe649b2ac207f26c404e06e50e0ee03\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"75819d7016666ab2dcd9632665f0bb461dceea1c7946ba8e5072ef6cb62a98e7\"" Jul 9 23:59:17.934519 containerd[1562]: time="2025-07-09T23:59:17.934502432Z" level=info msg="StartContainer for \"75819d7016666ab2dcd9632665f0bb461dceea1c7946ba8e5072ef6cb62a98e7\"" Jul 9 23:59:17.937460 containerd[1562]: time="2025-07-09T23:59:17.936670417Z" level=info msg="CreateContainer within sandbox \"f214a72b3c24ec0f9f7efcfb2d817b23125f94a032ce3a215d2b6f5e62dad22e\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"da99368faa907a24d097c26abdea40c75f1660792aa0063c77eb04c5393bf338\"" Jul 9 23:59:17.940253 containerd[1562]: time="2025-07-09T23:59:17.940227341Z" level=info msg="StartContainer for \"da99368faa907a24d097c26abdea40c75f1660792aa0063c77eb04c5393bf338\"" Jul 9 23:59:17.960193 systemd[1]: Started cri-containerd-da99368faa907a24d097c26abdea40c75f1660792aa0063c77eb04c5393bf338.scope - libcontainer container da99368faa907a24d097c26abdea40c75f1660792aa0063c77eb04c5393bf338. Jul 9 23:59:17.964121 systemd[1]: Started cri-containerd-75819d7016666ab2dcd9632665f0bb461dceea1c7946ba8e5072ef6cb62a98e7.scope - libcontainer container 75819d7016666ab2dcd9632665f0bb461dceea1c7946ba8e5072ef6cb62a98e7. Jul 9 23:59:17.993334 containerd[1562]: time="2025-07-09T23:59:17.993311586Z" level=info msg="StartContainer for \"75819d7016666ab2dcd9632665f0bb461dceea1c7946ba8e5072ef6cb62a98e7\" returns successfully" Jul 9 23:59:17.993877 containerd[1562]: time="2025-07-09T23:59:17.993311573Z" level=info msg="StartContainer for \"da99368faa907a24d097c26abdea40c75f1660792aa0063c77eb04c5393bf338\" returns successfully" Jul 9 23:59:18.304390 kubelet[2830]: I0709 23:59:18.304346 2830 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-7lgtq" podStartSLOduration=32.304329934 podStartE2EDuration="32.304329934s" podCreationTimestamp="2025-07-09 23:58:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-09 23:59:18.296410431 +0000 UTC m=+36.596550802" watchObservedRunningTime="2025-07-09 23:59:18.304329934 +0000 UTC m=+36.604470305" Jul 9 23:59:18.306692 kubelet[2830]: I0709 23:59:18.306448 2830 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-w9rnc" podStartSLOduration=32.306436202 podStartE2EDuration="32.306436202s" podCreationTimestamp="2025-07-09 23:58:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-09 23:59:18.305735416 +0000 UTC m=+36.605875781" watchObservedRunningTime="2025-07-09 23:59:18.306436202 +0000 UTC m=+36.606576567" Jul 9 23:59:18.768496 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3947545061.mount: Deactivated successfully. Jul 9 23:59:57.536626 systemd[1]: Started sshd@7-139.178.70.101:22-139.178.89.65:59742.service - OpenSSH per-connection server daemon (139.178.89.65:59742). Jul 9 23:59:57.597389 sshd[4229]: Accepted publickey for core from 139.178.89.65 port 59742 ssh2: RSA SHA256:iW7N8ouL0MNZquiamslGIbLBE90/GD9BMvBNQ+/8OB0 Jul 9 23:59:57.598358 sshd-session[4229]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 9 23:59:57.602924 systemd-logind[1542]: New session 10 of user core. Jul 9 23:59:57.608232 systemd[1]: Started session-10.scope - Session 10 of User core. Jul 9 23:59:58.726722 sshd[4231]: Connection closed by 139.178.89.65 port 59742 Jul 9 23:59:58.727234 sshd-session[4229]: pam_unix(sshd:session): session closed for user core Jul 9 23:59:58.729709 systemd[1]: sshd@7-139.178.70.101:22-139.178.89.65:59742.service: Deactivated successfully. Jul 9 23:59:58.731064 systemd[1]: session-10.scope: Deactivated successfully. Jul 9 23:59:58.731634 systemd-logind[1542]: Session 10 logged out. Waiting for processes to exit. Jul 9 23:59:58.732384 systemd-logind[1542]: Removed session 10. Jul 10 00:00:03.743357 systemd[1]: Started logrotate.service - Rotate and Compress System Logs. Jul 10 00:00:03.746170 systemd[1]: Started sshd@8-139.178.70.101:22-139.178.89.65:55204.service - OpenSSH per-connection server daemon (139.178.89.65:55204). Jul 10 00:00:03.756414 systemd[1]: logrotate.service: Deactivated successfully. Jul 10 00:00:03.797492 sshd[4245]: Accepted publickey for core from 139.178.89.65 port 55204 ssh2: RSA SHA256:iW7N8ouL0MNZquiamslGIbLBE90/GD9BMvBNQ+/8OB0 Jul 10 00:00:03.798624 sshd-session[4245]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:00:03.802458 systemd-logind[1542]: New session 11 of user core. Jul 10 00:00:03.808213 systemd[1]: Started session-11.scope - Session 11 of User core. Jul 10 00:00:03.902327 sshd[4248]: Connection closed by 139.178.89.65 port 55204 Jul 10 00:00:03.902757 sshd-session[4245]: pam_unix(sshd:session): session closed for user core Jul 10 00:00:03.904870 systemd[1]: sshd@8-139.178.70.101:22-139.178.89.65:55204.service: Deactivated successfully. Jul 10 00:00:03.906002 systemd[1]: session-11.scope: Deactivated successfully. Jul 10 00:00:03.906503 systemd-logind[1542]: Session 11 logged out. Waiting for processes to exit. Jul 10 00:00:03.907006 systemd-logind[1542]: Removed session 11. Jul 10 00:00:08.912933 systemd[1]: Started sshd@9-139.178.70.101:22-139.178.89.65:55218.service - OpenSSH per-connection server daemon (139.178.89.65:55218). Jul 10 00:00:08.972344 sshd[4260]: Accepted publickey for core from 139.178.89.65 port 55218 ssh2: RSA SHA256:iW7N8ouL0MNZquiamslGIbLBE90/GD9BMvBNQ+/8OB0 Jul 10 00:00:08.973374 sshd-session[4260]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:00:08.977198 systemd-logind[1542]: New session 12 of user core. Jul 10 00:00:08.985268 systemd[1]: Started session-12.scope - Session 12 of User core. Jul 10 00:00:09.093100 sshd[4262]: Connection closed by 139.178.89.65 port 55218 Jul 10 00:00:09.093421 sshd-session[4260]: pam_unix(sshd:session): session closed for user core Jul 10 00:00:09.095772 systemd[1]: sshd@9-139.178.70.101:22-139.178.89.65:55218.service: Deactivated successfully. Jul 10 00:00:09.096906 systemd[1]: session-12.scope: Deactivated successfully. Jul 10 00:00:09.097442 systemd-logind[1542]: Session 12 logged out. Waiting for processes to exit. Jul 10 00:00:09.098605 systemd-logind[1542]: Removed session 12. Jul 10 00:00:14.110421 systemd[1]: Started sshd@10-139.178.70.101:22-139.178.89.65:44780.service - OpenSSH per-connection server daemon (139.178.89.65:44780). Jul 10 00:00:14.146259 sshd[4275]: Accepted publickey for core from 139.178.89.65 port 44780 ssh2: RSA SHA256:iW7N8ouL0MNZquiamslGIbLBE90/GD9BMvBNQ+/8OB0 Jul 10 00:00:14.147626 sshd-session[4275]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:00:14.153723 systemd-logind[1542]: New session 13 of user core. Jul 10 00:00:14.159279 systemd[1]: Started session-13.scope - Session 13 of User core. Jul 10 00:00:14.266277 sshd[4277]: Connection closed by 139.178.89.65 port 44780 Jul 10 00:00:14.267667 sshd-session[4275]: pam_unix(sshd:session): session closed for user core Jul 10 00:00:14.273428 systemd[1]: sshd@10-139.178.70.101:22-139.178.89.65:44780.service: Deactivated successfully. Jul 10 00:00:14.274558 systemd[1]: session-13.scope: Deactivated successfully. Jul 10 00:00:14.275161 systemd-logind[1542]: Session 13 logged out. Waiting for processes to exit. Jul 10 00:00:14.279282 systemd[1]: Started sshd@11-139.178.70.101:22-139.178.89.65:44786.service - OpenSSH per-connection server daemon (139.178.89.65:44786). Jul 10 00:00:14.280542 systemd-logind[1542]: Removed session 13. Jul 10 00:00:14.320548 sshd[4288]: Accepted publickey for core from 139.178.89.65 port 44786 ssh2: RSA SHA256:iW7N8ouL0MNZquiamslGIbLBE90/GD9BMvBNQ+/8OB0 Jul 10 00:00:14.321686 sshd-session[4288]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:00:14.325633 systemd-logind[1542]: New session 14 of user core. Jul 10 00:00:14.330246 systemd[1]: Started session-14.scope - Session 14 of User core. Jul 10 00:00:14.467178 sshd[4291]: Connection closed by 139.178.89.65 port 44786 Jul 10 00:00:14.477228 sshd-session[4288]: pam_unix(sshd:session): session closed for user core Jul 10 00:00:14.485576 systemd[1]: Started sshd@12-139.178.70.101:22-139.178.89.65:44798.service - OpenSSH per-connection server daemon (139.178.89.65:44798). Jul 10 00:00:14.486471 systemd[1]: sshd@11-139.178.70.101:22-139.178.89.65:44786.service: Deactivated successfully. Jul 10 00:00:14.488401 systemd[1]: session-14.scope: Deactivated successfully. Jul 10 00:00:14.489875 systemd-logind[1542]: Session 14 logged out. Waiting for processes to exit. Jul 10 00:00:14.490678 systemd-logind[1542]: Removed session 14. Jul 10 00:00:14.736827 sshd[4298]: Accepted publickey for core from 139.178.89.65 port 44798 ssh2: RSA SHA256:iW7N8ouL0MNZquiamslGIbLBE90/GD9BMvBNQ+/8OB0 Jul 10 00:00:14.737654 sshd-session[4298]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:00:14.741168 systemd-logind[1542]: New session 15 of user core. Jul 10 00:00:14.746281 systemd[1]: Started session-15.scope - Session 15 of User core. Jul 10 00:00:14.916765 sshd[4303]: Connection closed by 139.178.89.65 port 44798 Jul 10 00:00:14.917266 sshd-session[4298]: pam_unix(sshd:session): session closed for user core Jul 10 00:00:14.921267 systemd[1]: sshd@12-139.178.70.101:22-139.178.89.65:44798.service: Deactivated successfully. Jul 10 00:00:14.923233 systemd[1]: session-15.scope: Deactivated successfully. Jul 10 00:00:14.923930 systemd-logind[1542]: Session 15 logged out. Waiting for processes to exit. Jul 10 00:00:14.925763 systemd-logind[1542]: Removed session 15. Jul 10 00:00:19.932329 systemd[1]: Started sshd@13-139.178.70.101:22-139.178.89.65:50892.service - OpenSSH per-connection server daemon (139.178.89.65:50892). Jul 10 00:00:19.973086 sshd[4321]: Accepted publickey for core from 139.178.89.65 port 50892 ssh2: RSA SHA256:iW7N8ouL0MNZquiamslGIbLBE90/GD9BMvBNQ+/8OB0 Jul 10 00:00:19.973303 sshd-session[4321]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:00:19.976663 systemd-logind[1542]: New session 16 of user core. Jul 10 00:00:19.981210 systemd[1]: Started session-16.scope - Session 16 of User core. Jul 10 00:00:20.078287 sshd[4323]: Connection closed by 139.178.89.65 port 50892 Jul 10 00:00:20.079089 sshd-session[4321]: pam_unix(sshd:session): session closed for user core Jul 10 00:00:20.086019 systemd[1]: sshd@13-139.178.70.101:22-139.178.89.65:50892.service: Deactivated successfully. Jul 10 00:00:20.087492 systemd[1]: session-16.scope: Deactivated successfully. Jul 10 00:00:20.088020 systemd-logind[1542]: Session 16 logged out. Waiting for processes to exit. Jul 10 00:00:20.088885 systemd-logind[1542]: Removed session 16. Jul 10 00:00:25.088428 systemd[1]: Started sshd@14-139.178.70.101:22-139.178.89.65:50900.service - OpenSSH per-connection server daemon (139.178.89.65:50900). Jul 10 00:00:25.326037 sshd[4335]: Accepted publickey for core from 139.178.89.65 port 50900 ssh2: RSA SHA256:iW7N8ouL0MNZquiamslGIbLBE90/GD9BMvBNQ+/8OB0 Jul 10 00:00:25.327169 sshd-session[4335]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:00:25.333464 systemd-logind[1542]: New session 17 of user core. Jul 10 00:00:25.339198 systemd[1]: Started session-17.scope - Session 17 of User core. Jul 10 00:00:25.431000 sshd[4337]: Connection closed by 139.178.89.65 port 50900 Jul 10 00:00:25.432174 sshd-session[4335]: pam_unix(sshd:session): session closed for user core Jul 10 00:00:25.434303 systemd[1]: sshd@14-139.178.70.101:22-139.178.89.65:50900.service: Deactivated successfully. Jul 10 00:00:25.435763 systemd[1]: session-17.scope: Deactivated successfully. Jul 10 00:00:25.436471 systemd-logind[1542]: Session 17 logged out. Waiting for processes to exit. Jul 10 00:00:25.437349 systemd-logind[1542]: Removed session 17. Jul 10 00:00:30.448525 systemd[1]: Started sshd@15-139.178.70.101:22-139.178.89.65:42996.service - OpenSSH per-connection server daemon (139.178.89.65:42996). Jul 10 00:00:30.478194 sshd[4349]: Accepted publickey for core from 139.178.89.65 port 42996 ssh2: RSA SHA256:iW7N8ouL0MNZquiamslGIbLBE90/GD9BMvBNQ+/8OB0 Jul 10 00:00:30.479108 sshd-session[4349]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:00:30.482327 systemd-logind[1542]: New session 18 of user core. Jul 10 00:00:30.489253 systemd[1]: Started session-18.scope - Session 18 of User core. Jul 10 00:00:30.656659 sshd[4351]: Connection closed by 139.178.89.65 port 42996 Jul 10 00:00:30.673227 systemd[1]: Started sshd@16-139.178.70.101:22-139.178.89.65:43008.service - OpenSSH per-connection server daemon (139.178.89.65:43008). Jul 10 00:00:30.683570 sshd-session[4349]: pam_unix(sshd:session): session closed for user core Jul 10 00:00:30.704051 systemd[1]: sshd@15-139.178.70.101:22-139.178.89.65:42996.service: Deactivated successfully. Jul 10 00:00:30.705236 systemd[1]: session-18.scope: Deactivated successfully. Jul 10 00:00:30.705720 systemd-logind[1542]: Session 18 logged out. Waiting for processes to exit. Jul 10 00:00:30.706301 systemd-logind[1542]: Removed session 18. Jul 10 00:00:30.813491 sshd[4360]: Accepted publickey for core from 139.178.89.65 port 43008 ssh2: RSA SHA256:iW7N8ouL0MNZquiamslGIbLBE90/GD9BMvBNQ+/8OB0 Jul 10 00:00:30.814328 sshd-session[4360]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:00:30.817401 systemd-logind[1542]: New session 19 of user core. Jul 10 00:00:30.821177 systemd[1]: Started session-19.scope - Session 19 of User core. Jul 10 00:00:31.937901 sshd[4365]: Connection closed by 139.178.89.65 port 43008 Jul 10 00:00:31.938617 sshd-session[4360]: pam_unix(sshd:session): session closed for user core Jul 10 00:00:31.945589 systemd[1]: sshd@16-139.178.70.101:22-139.178.89.65:43008.service: Deactivated successfully. Jul 10 00:00:31.947183 systemd[1]: session-19.scope: Deactivated successfully. Jul 10 00:00:31.948262 systemd-logind[1542]: Session 19 logged out. Waiting for processes to exit. Jul 10 00:00:31.951626 systemd[1]: Started sshd@17-139.178.70.101:22-139.178.89.65:43022.service - OpenSSH per-connection server daemon (139.178.89.65:43022). Jul 10 00:00:31.955017 systemd-logind[1542]: Removed session 19. Jul 10 00:00:32.010557 sshd[4374]: Accepted publickey for core from 139.178.89.65 port 43022 ssh2: RSA SHA256:iW7N8ouL0MNZquiamslGIbLBE90/GD9BMvBNQ+/8OB0 Jul 10 00:00:32.011521 sshd-session[4374]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:00:32.015050 systemd-logind[1542]: New session 20 of user core. Jul 10 00:00:32.029260 systemd[1]: Started session-20.scope - Session 20 of User core. Jul 10 00:00:33.523789 sshd[4377]: Connection closed by 139.178.89.65 port 43022 Jul 10 00:00:33.526256 sshd-session[4374]: pam_unix(sshd:session): session closed for user core Jul 10 00:00:33.540364 systemd[1]: Started sshd@18-139.178.70.101:22-139.178.89.65:43032.service - OpenSSH per-connection server daemon (139.178.89.65:43032). Jul 10 00:00:33.543722 systemd[1]: sshd@17-139.178.70.101:22-139.178.89.65:43022.service: Deactivated successfully. Jul 10 00:00:33.546378 systemd[1]: session-20.scope: Deactivated successfully. Jul 10 00:00:33.549281 systemd-logind[1542]: Session 20 logged out. Waiting for processes to exit. Jul 10 00:00:33.550704 systemd-logind[1542]: Removed session 20. Jul 10 00:00:33.578429 sshd[4391]: Accepted publickey for core from 139.178.89.65 port 43032 ssh2: RSA SHA256:iW7N8ouL0MNZquiamslGIbLBE90/GD9BMvBNQ+/8OB0 Jul 10 00:00:33.577937 sshd-session[4391]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:00:33.586949 systemd-logind[1542]: New session 21 of user core. Jul 10 00:00:33.591283 systemd[1]: Started session-21.scope - Session 21 of User core. Jul 10 00:00:33.852408 sshd[4396]: Connection closed by 139.178.89.65 port 43032 Jul 10 00:00:33.854370 sshd-session[4391]: pam_unix(sshd:session): session closed for user core Jul 10 00:00:33.864035 systemd[1]: sshd@18-139.178.70.101:22-139.178.89.65:43032.service: Deactivated successfully. Jul 10 00:00:33.866737 systemd[1]: session-21.scope: Deactivated successfully. Jul 10 00:00:33.868833 systemd-logind[1542]: Session 21 logged out. Waiting for processes to exit. Jul 10 00:00:33.874924 systemd[1]: Started sshd@19-139.178.70.101:22-139.178.89.65:43042.service - OpenSSH per-connection server daemon (139.178.89.65:43042). Jul 10 00:00:33.876031 systemd-logind[1542]: Removed session 21. Jul 10 00:00:33.910679 sshd[4405]: Accepted publickey for core from 139.178.89.65 port 43042 ssh2: RSA SHA256:iW7N8ouL0MNZquiamslGIbLBE90/GD9BMvBNQ+/8OB0 Jul 10 00:00:33.911712 sshd-session[4405]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:00:33.915251 systemd-logind[1542]: New session 22 of user core. Jul 10 00:00:33.922266 systemd[1]: Started session-22.scope - Session 22 of User core. Jul 10 00:00:34.074864 sshd[4408]: Connection closed by 139.178.89.65 port 43042 Jul 10 00:00:34.075981 sshd-session[4405]: pam_unix(sshd:session): session closed for user core Jul 10 00:00:34.078753 systemd[1]: sshd@19-139.178.70.101:22-139.178.89.65:43042.service: Deactivated successfully. Jul 10 00:00:34.080869 systemd[1]: session-22.scope: Deactivated successfully. Jul 10 00:00:34.081584 systemd-logind[1542]: Session 22 logged out. Waiting for processes to exit. Jul 10 00:00:34.082978 systemd-logind[1542]: Removed session 22. Jul 10 00:00:39.084932 systemd[1]: Started sshd@20-139.178.70.101:22-139.178.89.65:43048.service - OpenSSH per-connection server daemon (139.178.89.65:43048). Jul 10 00:00:39.120490 sshd[4423]: Accepted publickey for core from 139.178.89.65 port 43048 ssh2: RSA SHA256:iW7N8ouL0MNZquiamslGIbLBE90/GD9BMvBNQ+/8OB0 Jul 10 00:00:39.121404 sshd-session[4423]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:00:39.124228 systemd-logind[1542]: New session 23 of user core. Jul 10 00:00:39.138453 systemd[1]: Started session-23.scope - Session 23 of User core. Jul 10 00:00:39.226545 sshd[4425]: Connection closed by 139.178.89.65 port 43048 Jul 10 00:00:39.226891 sshd-session[4423]: pam_unix(sshd:session): session closed for user core Jul 10 00:00:39.229030 systemd[1]: sshd@20-139.178.70.101:22-139.178.89.65:43048.service: Deactivated successfully. Jul 10 00:00:39.230182 systemd[1]: session-23.scope: Deactivated successfully. Jul 10 00:00:39.230613 systemd-logind[1542]: Session 23 logged out. Waiting for processes to exit. Jul 10 00:00:39.231192 systemd-logind[1542]: Removed session 23. Jul 10 00:00:44.241161 systemd[1]: Started sshd@21-139.178.70.101:22-139.178.89.65:42956.service - OpenSSH per-connection server daemon (139.178.89.65:42956). Jul 10 00:00:44.274754 sshd[4439]: Accepted publickey for core from 139.178.89.65 port 42956 ssh2: RSA SHA256:iW7N8ouL0MNZquiamslGIbLBE90/GD9BMvBNQ+/8OB0 Jul 10 00:00:44.275650 sshd-session[4439]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:00:44.278595 systemd-logind[1542]: New session 24 of user core. Jul 10 00:00:44.290245 systemd[1]: Started session-24.scope - Session 24 of User core. Jul 10 00:00:44.379102 sshd[4441]: Connection closed by 139.178.89.65 port 42956 Jul 10 00:00:44.379446 sshd-session[4439]: pam_unix(sshd:session): session closed for user core Jul 10 00:00:44.381666 systemd-logind[1542]: Session 24 logged out. Waiting for processes to exit. Jul 10 00:00:44.381966 systemd[1]: sshd@21-139.178.70.101:22-139.178.89.65:42956.service: Deactivated successfully. Jul 10 00:00:44.383205 systemd[1]: session-24.scope: Deactivated successfully. Jul 10 00:00:44.383832 systemd-logind[1542]: Removed session 24. Jul 10 00:00:49.389068 systemd[1]: Started sshd@22-139.178.70.101:22-139.178.89.65:42972.service - OpenSSH per-connection server daemon (139.178.89.65:42972). Jul 10 00:00:49.421181 sshd[4456]: Accepted publickey for core from 139.178.89.65 port 42972 ssh2: RSA SHA256:iW7N8ouL0MNZquiamslGIbLBE90/GD9BMvBNQ+/8OB0 Jul 10 00:00:49.422211 sshd-session[4456]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:00:49.426125 systemd-logind[1542]: New session 25 of user core. Jul 10 00:00:49.432206 systemd[1]: Started session-25.scope - Session 25 of User core. Jul 10 00:00:49.525511 sshd[4458]: Connection closed by 139.178.89.65 port 42972 Jul 10 00:00:49.525947 sshd-session[4456]: pam_unix(sshd:session): session closed for user core Jul 10 00:00:49.532211 systemd[1]: sshd@22-139.178.70.101:22-139.178.89.65:42972.service: Deactivated successfully. Jul 10 00:00:49.533182 systemd[1]: session-25.scope: Deactivated successfully. Jul 10 00:00:49.533897 systemd-logind[1542]: Session 25 logged out. Waiting for processes to exit. Jul 10 00:00:49.538231 systemd[1]: Started sshd@23-139.178.70.101:22-139.178.89.65:47254.service - OpenSSH per-connection server daemon (139.178.89.65:47254). Jul 10 00:00:49.539509 systemd-logind[1542]: Removed session 25. Jul 10 00:00:49.567396 sshd[4468]: Accepted publickey for core from 139.178.89.65 port 47254 ssh2: RSA SHA256:iW7N8ouL0MNZquiamslGIbLBE90/GD9BMvBNQ+/8OB0 Jul 10 00:00:49.568226 sshd-session[4468]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:00:49.571141 systemd-logind[1542]: New session 26 of user core. Jul 10 00:00:49.576172 systemd[1]: Started session-26.scope - Session 26 of User core. Jul 10 00:00:51.126352 containerd[1562]: time="2025-07-10T00:00:51.125761287Z" level=info msg="StopContainer for \"a43a4694ee5216f4d4a2826d38fea0b0bbc43feadabf1494f475086735f473d8\" with timeout 30 (s)" Jul 10 00:00:51.142524 containerd[1562]: time="2025-07-10T00:00:51.142425749Z" level=info msg="Stop container \"a43a4694ee5216f4d4a2826d38fea0b0bbc43feadabf1494f475086735f473d8\" with signal terminated" Jul 10 00:00:51.162453 containerd[1562]: time="2025-07-10T00:00:51.162414347Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Jul 10 00:00:51.166448 containerd[1562]: time="2025-07-10T00:00:51.166421522Z" level=info msg="StopContainer for \"caf02a3680d5145fd262b9738a10659ded2a04de94f9c3db1c4a81285572e9ad\" with timeout 2 (s)" Jul 10 00:00:51.166975 containerd[1562]: time="2025-07-10T00:00:51.166690627Z" level=info msg="Stop container \"caf02a3680d5145fd262b9738a10659ded2a04de94f9c3db1c4a81285572e9ad\" with signal terminated" Jul 10 00:00:51.167214 systemd[1]: cri-containerd-a43a4694ee5216f4d4a2826d38fea0b0bbc43feadabf1494f475086735f473d8.scope: Deactivated successfully. Jul 10 00:00:51.167408 systemd[1]: cri-containerd-a43a4694ee5216f4d4a2826d38fea0b0bbc43feadabf1494f475086735f473d8.scope: Consumed 238ms CPU time, 38.9M memory peak, 14.9M read from disk, 4K written to disk. Jul 10 00:00:51.173332 systemd-networkd[1465]: lxc_health: Link DOWN Jul 10 00:00:51.174221 systemd-networkd[1465]: lxc_health: Lost carrier Jul 10 00:00:51.189635 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a43a4694ee5216f4d4a2826d38fea0b0bbc43feadabf1494f475086735f473d8-rootfs.mount: Deactivated successfully. Jul 10 00:00:51.190799 systemd[1]: cri-containerd-caf02a3680d5145fd262b9738a10659ded2a04de94f9c3db1c4a81285572e9ad.scope: Deactivated successfully. Jul 10 00:00:51.191171 systemd[1]: cri-containerd-caf02a3680d5145fd262b9738a10659ded2a04de94f9c3db1c4a81285572e9ad.scope: Consumed 4.684s CPU time, 195.1M memory peak, 71.7M read from disk, 13.3M written to disk. Jul 10 00:00:51.200222 containerd[1562]: time="2025-07-10T00:00:51.200167867Z" level=info msg="shim disconnected" id=a43a4694ee5216f4d4a2826d38fea0b0bbc43feadabf1494f475086735f473d8 namespace=k8s.io Jul 10 00:00:51.200222 containerd[1562]: time="2025-07-10T00:00:51.200208623Z" level=warning msg="cleaning up after shim disconnected" id=a43a4694ee5216f4d4a2826d38fea0b0bbc43feadabf1494f475086735f473d8 namespace=k8s.io Jul 10 00:00:51.200222 containerd[1562]: time="2025-07-10T00:00:51.200214575Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jul 10 00:00:51.215062 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-caf02a3680d5145fd262b9738a10659ded2a04de94f9c3db1c4a81285572e9ad-rootfs.mount: Deactivated successfully. Jul 10 00:00:51.222502 containerd[1562]: time="2025-07-10T00:00:51.222471036Z" level=info msg="shim disconnected" id=caf02a3680d5145fd262b9738a10659ded2a04de94f9c3db1c4a81285572e9ad namespace=k8s.io Jul 10 00:00:51.223036 containerd[1562]: time="2025-07-10T00:00:51.222732999Z" level=warning msg="cleaning up after shim disconnected" id=caf02a3680d5145fd262b9738a10659ded2a04de94f9c3db1c4a81285572e9ad namespace=k8s.io Jul 10 00:00:51.223036 containerd[1562]: time="2025-07-10T00:00:51.222749507Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jul 10 00:00:51.229955 containerd[1562]: time="2025-07-10T00:00:51.229919327Z" level=warning msg="cleanup warnings time=\"2025-07-10T00:00:51Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Jul 10 00:00:51.232138 containerd[1562]: time="2025-07-10T00:00:51.232064943Z" level=info msg="StopContainer for \"a43a4694ee5216f4d4a2826d38fea0b0bbc43feadabf1494f475086735f473d8\" returns successfully" Jul 10 00:00:51.234469 containerd[1562]: time="2025-07-10T00:00:51.234428571Z" level=info msg="StopContainer for \"caf02a3680d5145fd262b9738a10659ded2a04de94f9c3db1c4a81285572e9ad\" returns successfully" Jul 10 00:00:51.236363 containerd[1562]: time="2025-07-10T00:00:51.236308272Z" level=info msg="StopPodSandbox for \"2951185f2d2f0f3fa6ecbc5a8060cf32c35ad336b49440efdede342adc870e00\"" Jul 10 00:00:51.236924 containerd[1562]: time="2025-07-10T00:00:51.236792704Z" level=info msg="StopPodSandbox for \"f58f391fd85e24d2648281b4945bb4e7390d20b9d624e85bd712e8c100a4ec28\"" Jul 10 00:00:51.246705 containerd[1562]: time="2025-07-10T00:00:51.238038787Z" level=info msg="Container to stop \"85452c6ec0b3f70f0b3549025557efb2e2eec40926df16ecff76c9276cc308d4\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jul 10 00:00:51.246705 containerd[1562]: time="2025-07-10T00:00:51.246619370Z" level=info msg="Container to stop \"0bb7627fe8e47d5a2abdcda9b06dca14ee9749ad1b159e59b387545f92d8e706\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jul 10 00:00:51.246705 containerd[1562]: time="2025-07-10T00:00:51.246629947Z" level=info msg="Container to stop \"a419a7595cf002be07974654bec25192ff733dda08bac28ec25a4c624402a4df\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jul 10 00:00:51.246705 containerd[1562]: time="2025-07-10T00:00:51.246635856Z" level=info msg="Container to stop \"abbae4097b8ce16b67b85924c3b963388218cc676731b56a3385d272ef7088e9\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jul 10 00:00:51.246705 containerd[1562]: time="2025-07-10T00:00:51.246640738Z" level=info msg="Container to stop \"caf02a3680d5145fd262b9738a10659ded2a04de94f9c3db1c4a81285572e9ad\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jul 10 00:00:51.247373 containerd[1562]: time="2025-07-10T00:00:51.238048084Z" level=info msg="Container to stop \"a43a4694ee5216f4d4a2826d38fea0b0bbc43feadabf1494f475086735f473d8\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jul 10 00:00:51.249693 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-f58f391fd85e24d2648281b4945bb4e7390d20b9d624e85bd712e8c100a4ec28-shm.mount: Deactivated successfully. Jul 10 00:00:51.249772 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-2951185f2d2f0f3fa6ecbc5a8060cf32c35ad336b49440efdede342adc870e00-shm.mount: Deactivated successfully. Jul 10 00:00:51.257956 systemd[1]: cri-containerd-f58f391fd85e24d2648281b4945bb4e7390d20b9d624e85bd712e8c100a4ec28.scope: Deactivated successfully. Jul 10 00:00:51.259071 systemd[1]: cri-containerd-2951185f2d2f0f3fa6ecbc5a8060cf32c35ad336b49440efdede342adc870e00.scope: Deactivated successfully. Jul 10 00:00:51.273849 containerd[1562]: time="2025-07-10T00:00:51.273722425Z" level=info msg="shim disconnected" id=f58f391fd85e24d2648281b4945bb4e7390d20b9d624e85bd712e8c100a4ec28 namespace=k8s.io Jul 10 00:00:51.273849 containerd[1562]: time="2025-07-10T00:00:51.273754509Z" level=warning msg="cleaning up after shim disconnected" id=f58f391fd85e24d2648281b4945bb4e7390d20b9d624e85bd712e8c100a4ec28 namespace=k8s.io Jul 10 00:00:51.273849 containerd[1562]: time="2025-07-10T00:00:51.273760393Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jul 10 00:00:51.276575 containerd[1562]: time="2025-07-10T00:00:51.276448978Z" level=info msg="shim disconnected" id=2951185f2d2f0f3fa6ecbc5a8060cf32c35ad336b49440efdede342adc870e00 namespace=k8s.io Jul 10 00:00:51.276575 containerd[1562]: time="2025-07-10T00:00:51.276523951Z" level=warning msg="cleaning up after shim disconnected" id=2951185f2d2f0f3fa6ecbc5a8060cf32c35ad336b49440efdede342adc870e00 namespace=k8s.io Jul 10 00:00:51.276575 containerd[1562]: time="2025-07-10T00:00:51.276529567Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jul 10 00:00:51.285686 containerd[1562]: time="2025-07-10T00:00:51.285597416Z" level=info msg="TearDown network for sandbox \"f58f391fd85e24d2648281b4945bb4e7390d20b9d624e85bd712e8c100a4ec28\" successfully" Jul 10 00:00:51.285686 containerd[1562]: time="2025-07-10T00:00:51.285617243Z" level=info msg="StopPodSandbox for \"f58f391fd85e24d2648281b4945bb4e7390d20b9d624e85bd712e8c100a4ec28\" returns successfully" Jul 10 00:00:51.286609 containerd[1562]: time="2025-07-10T00:00:51.286592730Z" level=info msg="TearDown network for sandbox \"2951185f2d2f0f3fa6ecbc5a8060cf32c35ad336b49440efdede342adc870e00\" successfully" Jul 10 00:00:51.286737 containerd[1562]: time="2025-07-10T00:00:51.286655375Z" level=info msg="StopPodSandbox for \"2951185f2d2f0f3fa6ecbc5a8060cf32c35ad336b49440efdede342adc870e00\" returns successfully" Jul 10 00:00:51.355942 kubelet[2830]: I0710 00:00:51.355554 2830 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-hostproc\") pod \"9efe80a0-3426-4d84-92fb-e466ff07f979\" (UID: \"9efe80a0-3426-4d84-92fb-e466ff07f979\") " Jul 10 00:00:51.355942 kubelet[2830]: I0710 00:00:51.355599 2830 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffklt\" (UniqueName: \"kubernetes.io/projected/03702575-51ea-44b7-a61a-3a443ff54801-kube-api-access-ffklt\") pod \"03702575-51ea-44b7-a61a-3a443ff54801\" (UID: \"03702575-51ea-44b7-a61a-3a443ff54801\") " Jul 10 00:00:51.355942 kubelet[2830]: I0710 00:00:51.355616 2830 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ndbt\" (UniqueName: \"kubernetes.io/projected/9efe80a0-3426-4d84-92fb-e466ff07f979-kube-api-access-4ndbt\") pod \"9efe80a0-3426-4d84-92fb-e466ff07f979\" (UID: \"9efe80a0-3426-4d84-92fb-e466ff07f979\") " Jul 10 00:00:51.355942 kubelet[2830]: I0710 00:00:51.355631 2830 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/9efe80a0-3426-4d84-92fb-e466ff07f979-cilium-config-path\") pod \"9efe80a0-3426-4d84-92fb-e466ff07f979\" (UID: \"9efe80a0-3426-4d84-92fb-e466ff07f979\") " Jul 10 00:00:51.355942 kubelet[2830]: I0710 00:00:51.355643 2830 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/9efe80a0-3426-4d84-92fb-e466ff07f979-hubble-tls\") pod \"9efe80a0-3426-4d84-92fb-e466ff07f979\" (UID: \"9efe80a0-3426-4d84-92fb-e466ff07f979\") " Jul 10 00:00:51.355942 kubelet[2830]: I0710 00:00:51.355656 2830 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/03702575-51ea-44b7-a61a-3a443ff54801-cilium-config-path\") pod \"03702575-51ea-44b7-a61a-3a443ff54801\" (UID: \"03702575-51ea-44b7-a61a-3a443ff54801\") " Jul 10 00:00:51.357181 kubelet[2830]: I0710 00:00:51.355668 2830 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-host-proc-sys-kernel\") pod \"9efe80a0-3426-4d84-92fb-e466ff07f979\" (UID: \"9efe80a0-3426-4d84-92fb-e466ff07f979\") " Jul 10 00:00:51.357181 kubelet[2830]: I0710 00:00:51.355681 2830 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/9efe80a0-3426-4d84-92fb-e466ff07f979-clustermesh-secrets\") pod \"9efe80a0-3426-4d84-92fb-e466ff07f979\" (UID: \"9efe80a0-3426-4d84-92fb-e466ff07f979\") " Jul 10 00:00:51.357181 kubelet[2830]: I0710 00:00:51.355693 2830 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-lib-modules\") pod \"9efe80a0-3426-4d84-92fb-e466ff07f979\" (UID: \"9efe80a0-3426-4d84-92fb-e466ff07f979\") " Jul 10 00:00:51.357181 kubelet[2830]: I0710 00:00:51.355708 2830 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-etc-cni-netd\") pod \"9efe80a0-3426-4d84-92fb-e466ff07f979\" (UID: \"9efe80a0-3426-4d84-92fb-e466ff07f979\") " Jul 10 00:00:51.357181 kubelet[2830]: I0710 00:00:51.355718 2830 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-cni-path\") pod \"9efe80a0-3426-4d84-92fb-e466ff07f979\" (UID: \"9efe80a0-3426-4d84-92fb-e466ff07f979\") " Jul 10 00:00:51.357181 kubelet[2830]: I0710 00:00:51.355729 2830 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-host-proc-sys-net\") pod \"9efe80a0-3426-4d84-92fb-e466ff07f979\" (UID: \"9efe80a0-3426-4d84-92fb-e466ff07f979\") " Jul 10 00:00:51.357839 kubelet[2830]: I0710 00:00:51.355740 2830 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-cilium-cgroup\") pod \"9efe80a0-3426-4d84-92fb-e466ff07f979\" (UID: \"9efe80a0-3426-4d84-92fb-e466ff07f979\") " Jul 10 00:00:51.357839 kubelet[2830]: I0710 00:00:51.355751 2830 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-xtables-lock\") pod \"9efe80a0-3426-4d84-92fb-e466ff07f979\" (UID: \"9efe80a0-3426-4d84-92fb-e466ff07f979\") " Jul 10 00:00:51.357839 kubelet[2830]: I0710 00:00:51.355761 2830 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-bpf-maps\") pod \"9efe80a0-3426-4d84-92fb-e466ff07f979\" (UID: \"9efe80a0-3426-4d84-92fb-e466ff07f979\") " Jul 10 00:00:51.357839 kubelet[2830]: I0710 00:00:51.355771 2830 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-cilium-run\") pod \"9efe80a0-3426-4d84-92fb-e466ff07f979\" (UID: \"9efe80a0-3426-4d84-92fb-e466ff07f979\") " Jul 10 00:00:51.359082 kubelet[2830]: I0710 00:00:51.358416 2830 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-hostproc" (OuterVolumeSpecName: "hostproc") pod "9efe80a0-3426-4d84-92fb-e466ff07f979" (UID: "9efe80a0-3426-4d84-92fb-e466ff07f979"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Jul 10 00:00:51.360027 kubelet[2830]: I0710 00:00:51.357678 2830 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "9efe80a0-3426-4d84-92fb-e466ff07f979" (UID: "9efe80a0-3426-4d84-92fb-e466ff07f979"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Jul 10 00:00:51.360218 kubelet[2830]: I0710 00:00:51.360194 2830 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "9efe80a0-3426-4d84-92fb-e466ff07f979" (UID: "9efe80a0-3426-4d84-92fb-e466ff07f979"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Jul 10 00:00:51.360266 kubelet[2830]: I0710 00:00:51.360219 2830 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-cni-path" (OuterVolumeSpecName: "cni-path") pod "9efe80a0-3426-4d84-92fb-e466ff07f979" (UID: "9efe80a0-3426-4d84-92fb-e466ff07f979"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Jul 10 00:00:51.360266 kubelet[2830]: I0710 00:00:51.360233 2830 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "9efe80a0-3426-4d84-92fb-e466ff07f979" (UID: "9efe80a0-3426-4d84-92fb-e466ff07f979"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Jul 10 00:00:51.360384 kubelet[2830]: I0710 00:00:51.360244 2830 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "9efe80a0-3426-4d84-92fb-e466ff07f979" (UID: "9efe80a0-3426-4d84-92fb-e466ff07f979"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Jul 10 00:00:51.360384 kubelet[2830]: I0710 00:00:51.360341 2830 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "9efe80a0-3426-4d84-92fb-e466ff07f979" (UID: "9efe80a0-3426-4d84-92fb-e466ff07f979"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Jul 10 00:00:51.360384 kubelet[2830]: I0710 00:00:51.360353 2830 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "9efe80a0-3426-4d84-92fb-e466ff07f979" (UID: "9efe80a0-3426-4d84-92fb-e466ff07f979"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Jul 10 00:00:51.360384 kubelet[2830]: I0710 00:00:51.360377 2830 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "9efe80a0-3426-4d84-92fb-e466ff07f979" (UID: "9efe80a0-3426-4d84-92fb-e466ff07f979"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Jul 10 00:00:51.365113 kubelet[2830]: I0710 00:00:51.364798 2830 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "9efe80a0-3426-4d84-92fb-e466ff07f979" (UID: "9efe80a0-3426-4d84-92fb-e466ff07f979"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Jul 10 00:00:51.368957 kubelet[2830]: I0710 00:00:51.368933 2830 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9efe80a0-3426-4d84-92fb-e466ff07f979-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "9efe80a0-3426-4d84-92fb-e466ff07f979" (UID: "9efe80a0-3426-4d84-92fb-e466ff07f979"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" Jul 10 00:00:51.370619 kubelet[2830]: I0710 00:00:51.370578 2830 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03702575-51ea-44b7-a61a-3a443ff54801-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "03702575-51ea-44b7-a61a-3a443ff54801" (UID: "03702575-51ea-44b7-a61a-3a443ff54801"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Jul 10 00:00:51.370934 kubelet[2830]: I0710 00:00:51.370604 2830 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9efe80a0-3426-4d84-92fb-e466ff07f979-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "9efe80a0-3426-4d84-92fb-e466ff07f979" (UID: "9efe80a0-3426-4d84-92fb-e466ff07f979"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" Jul 10 00:00:51.371143 kubelet[2830]: I0710 00:00:51.371045 2830 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03702575-51ea-44b7-a61a-3a443ff54801-kube-api-access-ffklt" (OuterVolumeSpecName: "kube-api-access-ffklt") pod "03702575-51ea-44b7-a61a-3a443ff54801" (UID: "03702575-51ea-44b7-a61a-3a443ff54801"). InnerVolumeSpecName "kube-api-access-ffklt". PluginName "kubernetes.io/projected", VolumeGIDValue "" Jul 10 00:00:51.371464 kubelet[2830]: I0710 00:00:51.371442 2830 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9efe80a0-3426-4d84-92fb-e466ff07f979-kube-api-access-4ndbt" (OuterVolumeSpecName: "kube-api-access-4ndbt") pod "9efe80a0-3426-4d84-92fb-e466ff07f979" (UID: "9efe80a0-3426-4d84-92fb-e466ff07f979"). InnerVolumeSpecName "kube-api-access-4ndbt". PluginName "kubernetes.io/projected", VolumeGIDValue "" Jul 10 00:00:51.371899 kubelet[2830]: I0710 00:00:51.371876 2830 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9efe80a0-3426-4d84-92fb-e466ff07f979-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "9efe80a0-3426-4d84-92fb-e466ff07f979" (UID: "9efe80a0-3426-4d84-92fb-e466ff07f979"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Jul 10 00:00:51.442144 systemd[1]: Removed slice kubepods-burstable-pod9efe80a0_3426_4d84_92fb_e466ff07f979.slice - libcontainer container kubepods-burstable-pod9efe80a0_3426_4d84_92fb_e466ff07f979.slice. Jul 10 00:00:51.443041 systemd[1]: kubepods-burstable-pod9efe80a0_3426_4d84_92fb_e466ff07f979.slice: Consumed 4.738s CPU time, 196M memory peak, 73.8M read from disk, 13.3M written to disk. Jul 10 00:00:51.456989 kubelet[2830]: I0710 00:00:51.456565 2830 scope.go:117] "RemoveContainer" containerID="caf02a3680d5145fd262b9738a10659ded2a04de94f9c3db1c4a81285572e9ad" Jul 10 00:00:51.457953 kubelet[2830]: I0710 00:00:51.457943 2830 reconciler_common.go:299] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/9efe80a0-3426-4d84-92fb-e466ff07f979-clustermesh-secrets\") on node \"localhost\" DevicePath \"\"" Jul 10 00:00:51.458145 kubelet[2830]: I0710 00:00:51.458137 2830 reconciler_common.go:299] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-lib-modules\") on node \"localhost\" DevicePath \"\"" Jul 10 00:00:51.458365 kubelet[2830]: I0710 00:00:51.458197 2830 reconciler_common.go:299] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-etc-cni-netd\") on node \"localhost\" DevicePath \"\"" Jul 10 00:00:51.458365 kubelet[2830]: I0710 00:00:51.458206 2830 reconciler_common.go:299] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-cni-path\") on node \"localhost\" DevicePath \"\"" Jul 10 00:00:51.458365 kubelet[2830]: I0710 00:00:51.458213 2830 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-host-proc-sys-net\") on node \"localhost\" DevicePath \"\"" Jul 10 00:00:51.458365 kubelet[2830]: I0710 00:00:51.458219 2830 reconciler_common.go:299] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-cilium-cgroup\") on node \"localhost\" DevicePath \"\"" Jul 10 00:00:51.458365 kubelet[2830]: I0710 00:00:51.458225 2830 reconciler_common.go:299] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-xtables-lock\") on node \"localhost\" DevicePath \"\"" Jul 10 00:00:51.458365 kubelet[2830]: I0710 00:00:51.458231 2830 reconciler_common.go:299] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-bpf-maps\") on node \"localhost\" DevicePath \"\"" Jul 10 00:00:51.458365 kubelet[2830]: I0710 00:00:51.458236 2830 reconciler_common.go:299] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-cilium-run\") on node \"localhost\" DevicePath \"\"" Jul 10 00:00:51.458365 kubelet[2830]: I0710 00:00:51.458242 2830 reconciler_common.go:299] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-hostproc\") on node \"localhost\" DevicePath \"\"" Jul 10 00:00:51.458537 kubelet[2830]: I0710 00:00:51.458248 2830 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-ffklt\" (UniqueName: \"kubernetes.io/projected/03702575-51ea-44b7-a61a-3a443ff54801-kube-api-access-ffklt\") on node \"localhost\" DevicePath \"\"" Jul 10 00:00:51.458537 kubelet[2830]: I0710 00:00:51.458254 2830 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-4ndbt\" (UniqueName: \"kubernetes.io/projected/9efe80a0-3426-4d84-92fb-e466ff07f979-kube-api-access-4ndbt\") on node \"localhost\" DevicePath \"\"" Jul 10 00:00:51.458537 kubelet[2830]: I0710 00:00:51.458259 2830 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/9efe80a0-3426-4d84-92fb-e466ff07f979-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Jul 10 00:00:51.458537 kubelet[2830]: I0710 00:00:51.458265 2830 reconciler_common.go:299] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/9efe80a0-3426-4d84-92fb-e466ff07f979-hubble-tls\") on node \"localhost\" DevicePath \"\"" Jul 10 00:00:51.458537 kubelet[2830]: I0710 00:00:51.458271 2830 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/03702575-51ea-44b7-a61a-3a443ff54801-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Jul 10 00:00:51.458537 kubelet[2830]: I0710 00:00:51.458278 2830 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/9efe80a0-3426-4d84-92fb-e466ff07f979-host-proc-sys-kernel\") on node \"localhost\" DevicePath \"\"" Jul 10 00:00:51.462313 containerd[1562]: time="2025-07-10T00:00:51.461862195Z" level=info msg="RemoveContainer for \"caf02a3680d5145fd262b9738a10659ded2a04de94f9c3db1c4a81285572e9ad\"" Jul 10 00:00:51.464332 systemd[1]: Removed slice kubepods-besteffort-pod03702575_51ea_44b7_a61a_3a443ff54801.slice - libcontainer container kubepods-besteffort-pod03702575_51ea_44b7_a61a_3a443ff54801.slice. Jul 10 00:00:51.464408 systemd[1]: kubepods-besteffort-pod03702575_51ea_44b7_a61a_3a443ff54801.slice: Consumed 260ms CPU time, 39.5M memory peak, 14.9M read from disk, 4K written to disk. Jul 10 00:00:51.469923 containerd[1562]: time="2025-07-10T00:00:51.469894615Z" level=info msg="RemoveContainer for \"caf02a3680d5145fd262b9738a10659ded2a04de94f9c3db1c4a81285572e9ad\" returns successfully" Jul 10 00:00:51.472934 kubelet[2830]: I0710 00:00:51.472910 2830 scope.go:117] "RemoveContainer" containerID="a419a7595cf002be07974654bec25192ff733dda08bac28ec25a4c624402a4df" Jul 10 00:00:51.474307 containerd[1562]: time="2025-07-10T00:00:51.474033750Z" level=info msg="RemoveContainer for \"a419a7595cf002be07974654bec25192ff733dda08bac28ec25a4c624402a4df\"" Jul 10 00:00:51.476855 containerd[1562]: time="2025-07-10T00:00:51.476482705Z" level=info msg="RemoveContainer for \"a419a7595cf002be07974654bec25192ff733dda08bac28ec25a4c624402a4df\" returns successfully" Jul 10 00:00:51.477154 kubelet[2830]: I0710 00:00:51.477140 2830 scope.go:117] "RemoveContainer" containerID="0bb7627fe8e47d5a2abdcda9b06dca14ee9749ad1b159e59b387545f92d8e706" Jul 10 00:00:51.477874 containerd[1562]: time="2025-07-10T00:00:51.477823931Z" level=info msg="RemoveContainer for \"0bb7627fe8e47d5a2abdcda9b06dca14ee9749ad1b159e59b387545f92d8e706\"" Jul 10 00:00:51.480651 containerd[1562]: time="2025-07-10T00:00:51.480632350Z" level=info msg="RemoveContainer for \"0bb7627fe8e47d5a2abdcda9b06dca14ee9749ad1b159e59b387545f92d8e706\" returns successfully" Jul 10 00:00:51.480781 kubelet[2830]: I0710 00:00:51.480766 2830 scope.go:117] "RemoveContainer" containerID="85452c6ec0b3f70f0b3549025557efb2e2eec40926df16ecff76c9276cc308d4" Jul 10 00:00:51.481455 containerd[1562]: time="2025-07-10T00:00:51.481437721Z" level=info msg="RemoveContainer for \"85452c6ec0b3f70f0b3549025557efb2e2eec40926df16ecff76c9276cc308d4\"" Jul 10 00:00:51.482845 containerd[1562]: time="2025-07-10T00:00:51.482831335Z" level=info msg="RemoveContainer for \"85452c6ec0b3f70f0b3549025557efb2e2eec40926df16ecff76c9276cc308d4\" returns successfully" Jul 10 00:00:51.482931 kubelet[2830]: I0710 00:00:51.482903 2830 scope.go:117] "RemoveContainer" containerID="abbae4097b8ce16b67b85924c3b963388218cc676731b56a3385d272ef7088e9" Jul 10 00:00:51.483654 containerd[1562]: time="2025-07-10T00:00:51.483569224Z" level=info msg="RemoveContainer for \"abbae4097b8ce16b67b85924c3b963388218cc676731b56a3385d272ef7088e9\"" Jul 10 00:00:51.484751 containerd[1562]: time="2025-07-10T00:00:51.484714884Z" level=info msg="RemoveContainer for \"abbae4097b8ce16b67b85924c3b963388218cc676731b56a3385d272ef7088e9\" returns successfully" Jul 10 00:00:51.484840 kubelet[2830]: I0710 00:00:51.484826 2830 scope.go:117] "RemoveContainer" containerID="caf02a3680d5145fd262b9738a10659ded2a04de94f9c3db1c4a81285572e9ad" Jul 10 00:00:51.485007 containerd[1562]: time="2025-07-10T00:00:51.484988358Z" level=error msg="ContainerStatus for \"caf02a3680d5145fd262b9738a10659ded2a04de94f9c3db1c4a81285572e9ad\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"caf02a3680d5145fd262b9738a10659ded2a04de94f9c3db1c4a81285572e9ad\": not found" Jul 10 00:00:51.493565 kubelet[2830]: E0710 00:00:51.493482 2830 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"caf02a3680d5145fd262b9738a10659ded2a04de94f9c3db1c4a81285572e9ad\": not found" containerID="caf02a3680d5145fd262b9738a10659ded2a04de94f9c3db1c4a81285572e9ad" Jul 10 00:00:51.496661 kubelet[2830]: I0710 00:00:51.493515 2830 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"caf02a3680d5145fd262b9738a10659ded2a04de94f9c3db1c4a81285572e9ad"} err="failed to get container status \"caf02a3680d5145fd262b9738a10659ded2a04de94f9c3db1c4a81285572e9ad\": rpc error: code = NotFound desc = an error occurred when try to find container \"caf02a3680d5145fd262b9738a10659ded2a04de94f9c3db1c4a81285572e9ad\": not found" Jul 10 00:00:51.496661 kubelet[2830]: I0710 00:00:51.496591 2830 scope.go:117] "RemoveContainer" containerID="a419a7595cf002be07974654bec25192ff733dda08bac28ec25a4c624402a4df" Jul 10 00:00:51.496800 containerd[1562]: time="2025-07-10T00:00:51.496762366Z" level=error msg="ContainerStatus for \"a419a7595cf002be07974654bec25192ff733dda08bac28ec25a4c624402a4df\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"a419a7595cf002be07974654bec25192ff733dda08bac28ec25a4c624402a4df\": not found" Jul 10 00:00:51.496941 kubelet[2830]: E0710 00:00:51.496848 2830 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"a419a7595cf002be07974654bec25192ff733dda08bac28ec25a4c624402a4df\": not found" containerID="a419a7595cf002be07974654bec25192ff733dda08bac28ec25a4c624402a4df" Jul 10 00:00:51.496941 kubelet[2830]: I0710 00:00:51.496861 2830 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"a419a7595cf002be07974654bec25192ff733dda08bac28ec25a4c624402a4df"} err="failed to get container status \"a419a7595cf002be07974654bec25192ff733dda08bac28ec25a4c624402a4df\": rpc error: code = NotFound desc = an error occurred when try to find container \"a419a7595cf002be07974654bec25192ff733dda08bac28ec25a4c624402a4df\": not found" Jul 10 00:00:51.496941 kubelet[2830]: I0710 00:00:51.496871 2830 scope.go:117] "RemoveContainer" containerID="0bb7627fe8e47d5a2abdcda9b06dca14ee9749ad1b159e59b387545f92d8e706" Jul 10 00:00:51.497033 containerd[1562]: time="2025-07-10T00:00:51.496941800Z" level=error msg="ContainerStatus for \"0bb7627fe8e47d5a2abdcda9b06dca14ee9749ad1b159e59b387545f92d8e706\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"0bb7627fe8e47d5a2abdcda9b06dca14ee9749ad1b159e59b387545f92d8e706\": not found" Jul 10 00:00:51.497142 kubelet[2830]: E0710 00:00:51.497095 2830 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"0bb7627fe8e47d5a2abdcda9b06dca14ee9749ad1b159e59b387545f92d8e706\": not found" containerID="0bb7627fe8e47d5a2abdcda9b06dca14ee9749ad1b159e59b387545f92d8e706" Jul 10 00:00:51.497142 kubelet[2830]: I0710 00:00:51.497107 2830 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"0bb7627fe8e47d5a2abdcda9b06dca14ee9749ad1b159e59b387545f92d8e706"} err="failed to get container status \"0bb7627fe8e47d5a2abdcda9b06dca14ee9749ad1b159e59b387545f92d8e706\": rpc error: code = NotFound desc = an error occurred when try to find container \"0bb7627fe8e47d5a2abdcda9b06dca14ee9749ad1b159e59b387545f92d8e706\": not found" Jul 10 00:00:51.497142 kubelet[2830]: I0710 00:00:51.497114 2830 scope.go:117] "RemoveContainer" containerID="85452c6ec0b3f70f0b3549025557efb2e2eec40926df16ecff76c9276cc308d4" Jul 10 00:00:51.498011 kubelet[2830]: E0710 00:00:51.497259 2830 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"85452c6ec0b3f70f0b3549025557efb2e2eec40926df16ecff76c9276cc308d4\": not found" containerID="85452c6ec0b3f70f0b3549025557efb2e2eec40926df16ecff76c9276cc308d4" Jul 10 00:00:51.498011 kubelet[2830]: I0710 00:00:51.497268 2830 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"85452c6ec0b3f70f0b3549025557efb2e2eec40926df16ecff76c9276cc308d4"} err="failed to get container status \"85452c6ec0b3f70f0b3549025557efb2e2eec40926df16ecff76c9276cc308d4\": rpc error: code = NotFound desc = an error occurred when try to find container \"85452c6ec0b3f70f0b3549025557efb2e2eec40926df16ecff76c9276cc308d4\": not found" Jul 10 00:00:51.498011 kubelet[2830]: I0710 00:00:51.497274 2830 scope.go:117] "RemoveContainer" containerID="abbae4097b8ce16b67b85924c3b963388218cc676731b56a3385d272ef7088e9" Jul 10 00:00:51.498011 kubelet[2830]: E0710 00:00:51.497555 2830 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"abbae4097b8ce16b67b85924c3b963388218cc676731b56a3385d272ef7088e9\": not found" containerID="abbae4097b8ce16b67b85924c3b963388218cc676731b56a3385d272ef7088e9" Jul 10 00:00:51.498011 kubelet[2830]: I0710 00:00:51.497565 2830 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"abbae4097b8ce16b67b85924c3b963388218cc676731b56a3385d272ef7088e9"} err="failed to get container status \"abbae4097b8ce16b67b85924c3b963388218cc676731b56a3385d272ef7088e9\": rpc error: code = NotFound desc = an error occurred when try to find container \"abbae4097b8ce16b67b85924c3b963388218cc676731b56a3385d272ef7088e9\": not found" Jul 10 00:00:51.498011 kubelet[2830]: I0710 00:00:51.497576 2830 scope.go:117] "RemoveContainer" containerID="a43a4694ee5216f4d4a2826d38fea0b0bbc43feadabf1494f475086735f473d8" Jul 10 00:00:51.498181 containerd[1562]: time="2025-07-10T00:00:51.497191526Z" level=error msg="ContainerStatus for \"85452c6ec0b3f70f0b3549025557efb2e2eec40926df16ecff76c9276cc308d4\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"85452c6ec0b3f70f0b3549025557efb2e2eec40926df16ecff76c9276cc308d4\": not found" Jul 10 00:00:51.498181 containerd[1562]: time="2025-07-10T00:00:51.497336649Z" level=error msg="ContainerStatus for \"abbae4097b8ce16b67b85924c3b963388218cc676731b56a3385d272ef7088e9\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"abbae4097b8ce16b67b85924c3b963388218cc676731b56a3385d272ef7088e9\": not found" Jul 10 00:00:51.498365 containerd[1562]: time="2025-07-10T00:00:51.498349362Z" level=info msg="RemoveContainer for \"a43a4694ee5216f4d4a2826d38fea0b0bbc43feadabf1494f475086735f473d8\"" Jul 10 00:00:51.499637 containerd[1562]: time="2025-07-10T00:00:51.499613537Z" level=info msg="RemoveContainer for \"a43a4694ee5216f4d4a2826d38fea0b0bbc43feadabf1494f475086735f473d8\" returns successfully" Jul 10 00:00:51.499867 kubelet[2830]: I0710 00:00:51.499757 2830 scope.go:117] "RemoveContainer" containerID="a43a4694ee5216f4d4a2826d38fea0b0bbc43feadabf1494f475086735f473d8" Jul 10 00:00:51.500086 containerd[1562]: time="2025-07-10T00:00:51.499970621Z" level=error msg="ContainerStatus for \"a43a4694ee5216f4d4a2826d38fea0b0bbc43feadabf1494f475086735f473d8\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"a43a4694ee5216f4d4a2826d38fea0b0bbc43feadabf1494f475086735f473d8\": not found" Jul 10 00:00:51.500119 kubelet[2830]: E0710 00:00:51.500040 2830 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"a43a4694ee5216f4d4a2826d38fea0b0bbc43feadabf1494f475086735f473d8\": not found" containerID="a43a4694ee5216f4d4a2826d38fea0b0bbc43feadabf1494f475086735f473d8" Jul 10 00:00:51.500119 kubelet[2830]: I0710 00:00:51.500051 2830 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"a43a4694ee5216f4d4a2826d38fea0b0bbc43feadabf1494f475086735f473d8"} err="failed to get container status \"a43a4694ee5216f4d4a2826d38fea0b0bbc43feadabf1494f475086735f473d8\": rpc error: code = NotFound desc = an error occurred when try to find container \"a43a4694ee5216f4d4a2826d38fea0b0bbc43feadabf1494f475086735f473d8\": not found" Jul 10 00:00:51.950122 kubelet[2830]: I0710 00:00:51.950065 2830 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03702575-51ea-44b7-a61a-3a443ff54801" path="/var/lib/kubelet/pods/03702575-51ea-44b7-a61a-3a443ff54801/volumes" Jul 10 00:00:51.968493 kubelet[2830]: I0710 00:00:51.968233 2830 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9efe80a0-3426-4d84-92fb-e466ff07f979" path="/var/lib/kubelet/pods/9efe80a0-3426-4d84-92fb-e466ff07f979/volumes" Jul 10 00:00:52.111134 kubelet[2830]: E0710 00:00:52.111110 2830 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Jul 10 00:00:52.140313 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-f58f391fd85e24d2648281b4945bb4e7390d20b9d624e85bd712e8c100a4ec28-rootfs.mount: Deactivated successfully. Jul 10 00:00:52.140407 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-2951185f2d2f0f3fa6ecbc5a8060cf32c35ad336b49440efdede342adc870e00-rootfs.mount: Deactivated successfully. Jul 10 00:00:52.140473 systemd[1]: var-lib-kubelet-pods-9efe80a0\x2d3426\x2d4d84\x2d92fb\x2de466ff07f979-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d4ndbt.mount: Deactivated successfully. Jul 10 00:00:52.140535 systemd[1]: var-lib-kubelet-pods-03702575\x2d51ea\x2d44b7\x2da61a\x2d3a443ff54801-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dffklt.mount: Deactivated successfully. Jul 10 00:00:52.140590 systemd[1]: var-lib-kubelet-pods-9efe80a0\x2d3426\x2d4d84\x2d92fb\x2de466ff07f979-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Jul 10 00:00:52.140654 systemd[1]: var-lib-kubelet-pods-9efe80a0\x2d3426\x2d4d84\x2d92fb\x2de466ff07f979-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Jul 10 00:00:53.068258 sshd[4471]: Connection closed by 139.178.89.65 port 47254 Jul 10 00:00:53.069197 sshd-session[4468]: pam_unix(sshd:session): session closed for user core Jul 10 00:00:53.078188 systemd[1]: sshd@23-139.178.70.101:22-139.178.89.65:47254.service: Deactivated successfully. Jul 10 00:00:53.078331 systemd-logind[1542]: Session 26 logged out. Waiting for processes to exit. Jul 10 00:00:53.079773 systemd[1]: session-26.scope: Deactivated successfully. Jul 10 00:00:53.086369 systemd[1]: Started sshd@24-139.178.70.101:22-139.178.89.65:47258.service - OpenSSH per-connection server daemon (139.178.89.65:47258). Jul 10 00:00:53.087518 systemd-logind[1542]: Removed session 26. Jul 10 00:00:53.128941 sshd[4635]: Accepted publickey for core from 139.178.89.65 port 47258 ssh2: RSA SHA256:iW7N8ouL0MNZquiamslGIbLBE90/GD9BMvBNQ+/8OB0 Jul 10 00:00:53.129743 sshd-session[4635]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:00:53.133036 systemd-logind[1542]: New session 27 of user core. Jul 10 00:00:53.142163 systemd[1]: Started session-27.scope - Session 27 of User core. Jul 10 00:00:53.506098 sshd[4638]: Connection closed by 139.178.89.65 port 47258 Jul 10 00:00:53.507046 sshd-session[4635]: pam_unix(sshd:session): session closed for user core Jul 10 00:00:53.517516 systemd[1]: sshd@24-139.178.70.101:22-139.178.89.65:47258.service: Deactivated successfully. Jul 10 00:00:53.520420 systemd[1]: session-27.scope: Deactivated successfully. Jul 10 00:00:53.522901 systemd-logind[1542]: Session 27 logged out. Waiting for processes to exit. Jul 10 00:00:53.531345 systemd[1]: Started sshd@25-139.178.70.101:22-139.178.89.65:47262.service - OpenSSH per-connection server daemon (139.178.89.65:47262). Jul 10 00:00:53.533863 systemd-logind[1542]: Removed session 27. Jul 10 00:00:53.564433 sshd[4647]: Accepted publickey for core from 139.178.89.65 port 47262 ssh2: RSA SHA256:iW7N8ouL0MNZquiamslGIbLBE90/GD9BMvBNQ+/8OB0 Jul 10 00:00:53.565624 sshd-session[4647]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:00:53.568798 systemd-logind[1542]: New session 28 of user core. Jul 10 00:00:53.575295 systemd[1]: Started session-28.scope - Session 28 of User core. Jul 10 00:00:53.579911 systemd[1]: Created slice kubepods-burstable-pod98051d0c_1f50_4395_ab05_5d71186f8d32.slice - libcontainer container kubepods-burstable-pod98051d0c_1f50_4395_ab05_5d71186f8d32.slice. Jul 10 00:00:53.650283 sshd[4650]: Connection closed by 139.178.89.65 port 47262 Jul 10 00:00:53.650752 sshd-session[4647]: pam_unix(sshd:session): session closed for user core Jul 10 00:00:53.657402 systemd[1]: sshd@25-139.178.70.101:22-139.178.89.65:47262.service: Deactivated successfully. Jul 10 00:00:53.658510 systemd[1]: session-28.scope: Deactivated successfully. Jul 10 00:00:53.659452 systemd-logind[1542]: Session 28 logged out. Waiting for processes to exit. Jul 10 00:00:53.664326 systemd[1]: Started sshd@26-139.178.70.101:22-139.178.89.65:47274.service - OpenSSH per-connection server daemon (139.178.89.65:47274). Jul 10 00:00:53.665454 systemd-logind[1542]: Removed session 28. Jul 10 00:00:53.667894 kubelet[2830]: I0710 00:00:53.667875 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/98051d0c-1f50-4395-ab05-5d71186f8d32-cilium-cgroup\") pod \"cilium-xzs8p\" (UID: \"98051d0c-1f50-4395-ab05-5d71186f8d32\") " pod="kube-system/cilium-xzs8p" Jul 10 00:00:53.668507 kubelet[2830]: I0710 00:00:53.668112 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/98051d0c-1f50-4395-ab05-5d71186f8d32-lib-modules\") pod \"cilium-xzs8p\" (UID: \"98051d0c-1f50-4395-ab05-5d71186f8d32\") " pod="kube-system/cilium-xzs8p" Jul 10 00:00:53.668507 kubelet[2830]: I0710 00:00:53.668134 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/98051d0c-1f50-4395-ab05-5d71186f8d32-host-proc-sys-kernel\") pod \"cilium-xzs8p\" (UID: \"98051d0c-1f50-4395-ab05-5d71186f8d32\") " pod="kube-system/cilium-xzs8p" Jul 10 00:00:53.668507 kubelet[2830]: I0710 00:00:53.668148 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jm4j6\" (UniqueName: \"kubernetes.io/projected/98051d0c-1f50-4395-ab05-5d71186f8d32-kube-api-access-jm4j6\") pod \"cilium-xzs8p\" (UID: \"98051d0c-1f50-4395-ab05-5d71186f8d32\") " pod="kube-system/cilium-xzs8p" Jul 10 00:00:53.668507 kubelet[2830]: I0710 00:00:53.668169 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/98051d0c-1f50-4395-ab05-5d71186f8d32-xtables-lock\") pod \"cilium-xzs8p\" (UID: \"98051d0c-1f50-4395-ab05-5d71186f8d32\") " pod="kube-system/cilium-xzs8p" Jul 10 00:00:53.668507 kubelet[2830]: I0710 00:00:53.668184 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/98051d0c-1f50-4395-ab05-5d71186f8d32-cilium-config-path\") pod \"cilium-xzs8p\" (UID: \"98051d0c-1f50-4395-ab05-5d71186f8d32\") " pod="kube-system/cilium-xzs8p" Jul 10 00:00:53.668612 kubelet[2830]: I0710 00:00:53.668196 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/98051d0c-1f50-4395-ab05-5d71186f8d32-cilium-run\") pod \"cilium-xzs8p\" (UID: \"98051d0c-1f50-4395-ab05-5d71186f8d32\") " pod="kube-system/cilium-xzs8p" Jul 10 00:00:53.668612 kubelet[2830]: I0710 00:00:53.668206 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/98051d0c-1f50-4395-ab05-5d71186f8d32-cni-path\") pod \"cilium-xzs8p\" (UID: \"98051d0c-1f50-4395-ab05-5d71186f8d32\") " pod="kube-system/cilium-xzs8p" Jul 10 00:00:53.668612 kubelet[2830]: I0710 00:00:53.668222 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/98051d0c-1f50-4395-ab05-5d71186f8d32-etc-cni-netd\") pod \"cilium-xzs8p\" (UID: \"98051d0c-1f50-4395-ab05-5d71186f8d32\") " pod="kube-system/cilium-xzs8p" Jul 10 00:00:53.668612 kubelet[2830]: I0710 00:00:53.668230 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/98051d0c-1f50-4395-ab05-5d71186f8d32-clustermesh-secrets\") pod \"cilium-xzs8p\" (UID: \"98051d0c-1f50-4395-ab05-5d71186f8d32\") " pod="kube-system/cilium-xzs8p" Jul 10 00:00:53.668612 kubelet[2830]: I0710 00:00:53.668240 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/98051d0c-1f50-4395-ab05-5d71186f8d32-cilium-ipsec-secrets\") pod \"cilium-xzs8p\" (UID: \"98051d0c-1f50-4395-ab05-5d71186f8d32\") " pod="kube-system/cilium-xzs8p" Jul 10 00:00:53.668612 kubelet[2830]: I0710 00:00:53.668248 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/98051d0c-1f50-4395-ab05-5d71186f8d32-hubble-tls\") pod \"cilium-xzs8p\" (UID: \"98051d0c-1f50-4395-ab05-5d71186f8d32\") " pod="kube-system/cilium-xzs8p" Jul 10 00:00:53.668726 kubelet[2830]: I0710 00:00:53.668256 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/98051d0c-1f50-4395-ab05-5d71186f8d32-hostproc\") pod \"cilium-xzs8p\" (UID: \"98051d0c-1f50-4395-ab05-5d71186f8d32\") " pod="kube-system/cilium-xzs8p" Jul 10 00:00:53.668726 kubelet[2830]: I0710 00:00:53.668265 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/98051d0c-1f50-4395-ab05-5d71186f8d32-bpf-maps\") pod \"cilium-xzs8p\" (UID: \"98051d0c-1f50-4395-ab05-5d71186f8d32\") " pod="kube-system/cilium-xzs8p" Jul 10 00:00:53.668726 kubelet[2830]: I0710 00:00:53.668281 2830 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/98051d0c-1f50-4395-ab05-5d71186f8d32-host-proc-sys-net\") pod \"cilium-xzs8p\" (UID: \"98051d0c-1f50-4395-ab05-5d71186f8d32\") " pod="kube-system/cilium-xzs8p" Jul 10 00:00:53.698009 sshd[4656]: Accepted publickey for core from 139.178.89.65 port 47274 ssh2: RSA SHA256:iW7N8ouL0MNZquiamslGIbLBE90/GD9BMvBNQ+/8OB0 Jul 10 00:00:53.698886 sshd-session[4656]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jul 10 00:00:53.702058 systemd-logind[1542]: New session 29 of user core. Jul 10 00:00:53.713475 systemd[1]: Started session-29.scope - Session 29 of User core. Jul 10 00:00:53.884509 containerd[1562]: time="2025-07-10T00:00:53.884477491Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-xzs8p,Uid:98051d0c-1f50-4395-ab05-5d71186f8d32,Namespace:kube-system,Attempt:0,}" Jul 10 00:00:53.952348 containerd[1562]: time="2025-07-10T00:00:53.952133045Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 10 00:00:53.952348 containerd[1562]: time="2025-07-10T00:00:53.952185791Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 10 00:00:53.952348 containerd[1562]: time="2025-07-10T00:00:53.952200550Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 10 00:00:53.952348 containerd[1562]: time="2025-07-10T00:00:53.952269002Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 10 00:00:53.968221 systemd[1]: Started cri-containerd-93912764a0aaa067c59dbe981065b462ac9e0bf77f45c239781bd034fc785fc2.scope - libcontainer container 93912764a0aaa067c59dbe981065b462ac9e0bf77f45c239781bd034fc785fc2. Jul 10 00:00:53.990486 containerd[1562]: time="2025-07-10T00:00:53.990460363Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-xzs8p,Uid:98051d0c-1f50-4395-ab05-5d71186f8d32,Namespace:kube-system,Attempt:0,} returns sandbox id \"93912764a0aaa067c59dbe981065b462ac9e0bf77f45c239781bd034fc785fc2\"" Jul 10 00:00:53.994743 containerd[1562]: time="2025-07-10T00:00:53.994715183Z" level=info msg="CreateContainer within sandbox \"93912764a0aaa067c59dbe981065b462ac9e0bf77f45c239781bd034fc785fc2\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Jul 10 00:00:53.998624 containerd[1562]: time="2025-07-10T00:00:53.998565450Z" level=info msg="CreateContainer within sandbox \"93912764a0aaa067c59dbe981065b462ac9e0bf77f45c239781bd034fc785fc2\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"fa6b0f5262d1ef27ffd9977a95a542f417697d0e40c88e820d6f51b71ab9a5cc\"" Jul 10 00:00:54.000040 containerd[1562]: time="2025-07-10T00:00:53.999511755Z" level=info msg="StartContainer for \"fa6b0f5262d1ef27ffd9977a95a542f417697d0e40c88e820d6f51b71ab9a5cc\"" Jul 10 00:00:54.022254 systemd[1]: Started cri-containerd-fa6b0f5262d1ef27ffd9977a95a542f417697d0e40c88e820d6f51b71ab9a5cc.scope - libcontainer container fa6b0f5262d1ef27ffd9977a95a542f417697d0e40c88e820d6f51b71ab9a5cc. Jul 10 00:00:54.039862 containerd[1562]: time="2025-07-10T00:00:54.039834595Z" level=info msg="StartContainer for \"fa6b0f5262d1ef27ffd9977a95a542f417697d0e40c88e820d6f51b71ab9a5cc\" returns successfully" Jul 10 00:00:54.056819 systemd[1]: cri-containerd-fa6b0f5262d1ef27ffd9977a95a542f417697d0e40c88e820d6f51b71ab9a5cc.scope: Deactivated successfully. Jul 10 00:00:54.057028 systemd[1]: cri-containerd-fa6b0f5262d1ef27ffd9977a95a542f417697d0e40c88e820d6f51b71ab9a5cc.scope: Consumed 14ms CPU time, 9.4M memory peak, 3M read from disk. Jul 10 00:00:54.081053 containerd[1562]: time="2025-07-10T00:00:54.081003161Z" level=info msg="shim disconnected" id=fa6b0f5262d1ef27ffd9977a95a542f417697d0e40c88e820d6f51b71ab9a5cc namespace=k8s.io Jul 10 00:00:54.081053 containerd[1562]: time="2025-07-10T00:00:54.081041402Z" level=warning msg="cleaning up after shim disconnected" id=fa6b0f5262d1ef27ffd9977a95a542f417697d0e40c88e820d6f51b71ab9a5cc namespace=k8s.io Jul 10 00:00:54.081053 containerd[1562]: time="2025-07-10T00:00:54.081048973Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jul 10 00:00:54.483426 containerd[1562]: time="2025-07-10T00:00:54.483399168Z" level=info msg="CreateContainer within sandbox \"93912764a0aaa067c59dbe981065b462ac9e0bf77f45c239781bd034fc785fc2\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Jul 10 00:00:54.491404 containerd[1562]: time="2025-07-10T00:00:54.491334301Z" level=info msg="CreateContainer within sandbox \"93912764a0aaa067c59dbe981065b462ac9e0bf77f45c239781bd034fc785fc2\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"4db56d6073457548edeee9bf3a69d90c7911063f5adeaaac5a4033715b552ed1\"" Jul 10 00:00:54.491805 containerd[1562]: time="2025-07-10T00:00:54.491789744Z" level=info msg="StartContainer for \"4db56d6073457548edeee9bf3a69d90c7911063f5adeaaac5a4033715b552ed1\"" Jul 10 00:00:54.512362 kubelet[2830]: I0710 00:00:54.511709 2830 setters.go:618] "Node became not ready" node="localhost" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-07-10T00:00:54Z","lastTransitionTime":"2025-07-10T00:00:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Jul 10 00:00:54.512219 systemd[1]: Started cri-containerd-4db56d6073457548edeee9bf3a69d90c7911063f5adeaaac5a4033715b552ed1.scope - libcontainer container 4db56d6073457548edeee9bf3a69d90c7911063f5adeaaac5a4033715b552ed1. Jul 10 00:00:54.543435 containerd[1562]: time="2025-07-10T00:00:54.543409477Z" level=info msg="StartContainer for \"4db56d6073457548edeee9bf3a69d90c7911063f5adeaaac5a4033715b552ed1\" returns successfully" Jul 10 00:00:54.552135 systemd[1]: cri-containerd-4db56d6073457548edeee9bf3a69d90c7911063f5adeaaac5a4033715b552ed1.scope: Deactivated successfully. Jul 10 00:00:54.552312 systemd[1]: cri-containerd-4db56d6073457548edeee9bf3a69d90c7911063f5adeaaac5a4033715b552ed1.scope: Consumed 12ms CPU time, 7.2M memory peak, 1.9M read from disk. Jul 10 00:00:54.565769 containerd[1562]: time="2025-07-10T00:00:54.565732525Z" level=info msg="shim disconnected" id=4db56d6073457548edeee9bf3a69d90c7911063f5adeaaac5a4033715b552ed1 namespace=k8s.io Jul 10 00:00:54.565769 containerd[1562]: time="2025-07-10T00:00:54.565764305Z" level=warning msg="cleaning up after shim disconnected" id=4db56d6073457548edeee9bf3a69d90c7911063f5adeaaac5a4033715b552ed1 namespace=k8s.io Jul 10 00:00:54.565769 containerd[1562]: time="2025-07-10T00:00:54.565769632Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jul 10 00:00:55.488250 containerd[1562]: time="2025-07-10T00:00:55.488224039Z" level=info msg="CreateContainer within sandbox \"93912764a0aaa067c59dbe981065b462ac9e0bf77f45c239781bd034fc785fc2\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Jul 10 00:00:55.536756 containerd[1562]: time="2025-07-10T00:00:55.536667415Z" level=info msg="CreateContainer within sandbox \"93912764a0aaa067c59dbe981065b462ac9e0bf77f45c239781bd034fc785fc2\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"46cdd595dca10a286027bb67331333f69a6dd8ead25e8d75950904036e2a0a08\"" Jul 10 00:00:55.537354 containerd[1562]: time="2025-07-10T00:00:55.537146124Z" level=info msg="StartContainer for \"46cdd595dca10a286027bb67331333f69a6dd8ead25e8d75950904036e2a0a08\"" Jul 10 00:00:55.566213 systemd[1]: Started cri-containerd-46cdd595dca10a286027bb67331333f69a6dd8ead25e8d75950904036e2a0a08.scope - libcontainer container 46cdd595dca10a286027bb67331333f69a6dd8ead25e8d75950904036e2a0a08. Jul 10 00:00:55.587678 containerd[1562]: time="2025-07-10T00:00:55.587646266Z" level=info msg="StartContainer for \"46cdd595dca10a286027bb67331333f69a6dd8ead25e8d75950904036e2a0a08\" returns successfully" Jul 10 00:00:55.602772 systemd[1]: cri-containerd-46cdd595dca10a286027bb67331333f69a6dd8ead25e8d75950904036e2a0a08.scope: Deactivated successfully. Jul 10 00:00:55.623040 containerd[1562]: time="2025-07-10T00:00:55.622997270Z" level=info msg="shim disconnected" id=46cdd595dca10a286027bb67331333f69a6dd8ead25e8d75950904036e2a0a08 namespace=k8s.io Jul 10 00:00:55.623040 containerd[1562]: time="2025-07-10T00:00:55.623030132Z" level=warning msg="cleaning up after shim disconnected" id=46cdd595dca10a286027bb67331333f69a6dd8ead25e8d75950904036e2a0a08 namespace=k8s.io Jul 10 00:00:55.623286 containerd[1562]: time="2025-07-10T00:00:55.623047404Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jul 10 00:00:55.772712 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-46cdd595dca10a286027bb67331333f69a6dd8ead25e8d75950904036e2a0a08-rootfs.mount: Deactivated successfully. Jul 10 00:00:56.488101 containerd[1562]: time="2025-07-10T00:00:56.487904004Z" level=info msg="CreateContainer within sandbox \"93912764a0aaa067c59dbe981065b462ac9e0bf77f45c239781bd034fc785fc2\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Jul 10 00:00:56.501552 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount920088117.mount: Deactivated successfully. Jul 10 00:00:56.508787 containerd[1562]: time="2025-07-10T00:00:56.508750281Z" level=info msg="CreateContainer within sandbox \"93912764a0aaa067c59dbe981065b462ac9e0bf77f45c239781bd034fc785fc2\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"ba76a41f04aa262716d0e5083a5116d85646dc747314e30e8a9232fd9997198c\"" Jul 10 00:00:56.509642 containerd[1562]: time="2025-07-10T00:00:56.509619603Z" level=info msg="StartContainer for \"ba76a41f04aa262716d0e5083a5116d85646dc747314e30e8a9232fd9997198c\"" Jul 10 00:00:56.532253 systemd[1]: Started cri-containerd-ba76a41f04aa262716d0e5083a5116d85646dc747314e30e8a9232fd9997198c.scope - libcontainer container ba76a41f04aa262716d0e5083a5116d85646dc747314e30e8a9232fd9997198c. Jul 10 00:00:56.546624 systemd[1]: cri-containerd-ba76a41f04aa262716d0e5083a5116d85646dc747314e30e8a9232fd9997198c.scope: Deactivated successfully. Jul 10 00:00:56.578597 containerd[1562]: time="2025-07-10T00:00:56.578569264Z" level=info msg="StartContainer for \"ba76a41f04aa262716d0e5083a5116d85646dc747314e30e8a9232fd9997198c\" returns successfully" Jul 10 00:00:56.640018 containerd[1562]: time="2025-07-10T00:00:56.639022496Z" level=info msg="shim disconnected" id=ba76a41f04aa262716d0e5083a5116d85646dc747314e30e8a9232fd9997198c namespace=k8s.io Jul 10 00:00:56.640018 containerd[1562]: time="2025-07-10T00:00:56.639053034Z" level=warning msg="cleaning up after shim disconnected" id=ba76a41f04aa262716d0e5083a5116d85646dc747314e30e8a9232fd9997198c namespace=k8s.io Jul 10 00:00:56.640018 containerd[1562]: time="2025-07-10T00:00:56.639057963Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jul 10 00:00:56.772104 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-ba76a41f04aa262716d0e5083a5116d85646dc747314e30e8a9232fd9997198c-rootfs.mount: Deactivated successfully. Jul 10 00:00:57.112251 kubelet[2830]: E0710 00:00:57.112182 2830 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Jul 10 00:00:57.495179 containerd[1562]: time="2025-07-10T00:00:57.495097180Z" level=info msg="CreateContainer within sandbox \"93912764a0aaa067c59dbe981065b462ac9e0bf77f45c239781bd034fc785fc2\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Jul 10 00:00:57.513366 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3996845586.mount: Deactivated successfully. Jul 10 00:00:57.525575 containerd[1562]: time="2025-07-10T00:00:57.525501312Z" level=info msg="CreateContainer within sandbox \"93912764a0aaa067c59dbe981065b462ac9e0bf77f45c239781bd034fc785fc2\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"a8991cda0e6e5dfee9abf2a4416e118b7656ada5298d2d4630222c453b1966cb\"" Jul 10 00:00:57.526886 containerd[1562]: time="2025-07-10T00:00:57.526334461Z" level=info msg="StartContainer for \"a8991cda0e6e5dfee9abf2a4416e118b7656ada5298d2d4630222c453b1966cb\"" Jul 10 00:00:57.546196 systemd[1]: Started cri-containerd-a8991cda0e6e5dfee9abf2a4416e118b7656ada5298d2d4630222c453b1966cb.scope - libcontainer container a8991cda0e6e5dfee9abf2a4416e118b7656ada5298d2d4630222c453b1966cb. Jul 10 00:00:57.563403 containerd[1562]: time="2025-07-10T00:00:57.563382068Z" level=info msg="StartContainer for \"a8991cda0e6e5dfee9abf2a4416e118b7656ada5298d2d4630222c453b1966cb\" returns successfully" Jul 10 00:00:58.336154 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni)) Jul 10 00:00:58.508503 kubelet[2830]: I0710 00:00:58.508353 2830 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-xzs8p" podStartSLOduration=5.508338628 podStartE2EDuration="5.508338628s" podCreationTimestamp="2025-07-10 00:00:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-07-10 00:00:58.507879111 +0000 UTC m=+136.808019474" watchObservedRunningTime="2025-07-10 00:00:58.508338628 +0000 UTC m=+136.808478992" Jul 10 00:01:00.325943 kubelet[2830]: E0710 00:01:00.325883 2830 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 127.0.0.1:52538->127.0.0.1:45313: write tcp 127.0.0.1:52538->127.0.0.1:45313: write: broken pipe Jul 10 00:01:01.104872 systemd-networkd[1465]: lxc_health: Link UP Jul 10 00:01:01.110611 systemd-networkd[1465]: lxc_health: Gained carrier Jul 10 00:01:02.387382 systemd[1]: run-containerd-runc-k8s.io-a8991cda0e6e5dfee9abf2a4416e118b7656ada5298d2d4630222c453b1966cb-runc.Dt7oiY.mount: Deactivated successfully. Jul 10 00:01:03.080235 systemd-networkd[1465]: lxc_health: Gained IPv6LL Jul 10 00:01:06.651369 systemd[1]: run-containerd-runc-k8s.io-a8991cda0e6e5dfee9abf2a4416e118b7656ada5298d2d4630222c453b1966cb-runc.sh2Dwk.mount: Deactivated successfully. Jul 10 00:01:06.701824 sshd[4659]: Connection closed by 139.178.89.65 port 47274 Jul 10 00:01:06.702068 sshd-session[4656]: pam_unix(sshd:session): session closed for user core Jul 10 00:01:06.704743 systemd[1]: sshd@26-139.178.70.101:22-139.178.89.65:47274.service: Deactivated successfully. Jul 10 00:01:06.706389 systemd[1]: session-29.scope: Deactivated successfully. Jul 10 00:01:06.707413 systemd-logind[1542]: Session 29 logged out. Waiting for processes to exit. Jul 10 00:01:06.708500 systemd-logind[1542]: Removed session 29.