May 10 00:40:03.647460 kernel: Linux version 5.15.181-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 11.3.1_p20221209 p3) 11.3.1 20221209, GNU ld (Gentoo 2.39 p5) 2.39.0) #1 SMP Fri May 9 23:12:23 -00 2025 May 10 00:40:03.647474 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=vmware flatcar.autologin verity.usrhash=39569409b30be1967efab22b453b92a780dcf0fe8e1448a18bf235b5cf33e54a May 10 00:40:03.647480 kernel: Disabled fast string operations May 10 00:40:03.647484 kernel: BIOS-provided physical RAM map: May 10 00:40:03.647487 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ebff] usable May 10 00:40:03.647491 kernel: BIOS-e820: [mem 0x000000000009ec00-0x000000000009ffff] reserved May 10 00:40:03.647497 kernel: BIOS-e820: [mem 0x00000000000dc000-0x00000000000fffff] reserved May 10 00:40:03.647501 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007fedffff] usable May 10 00:40:03.647504 kernel: BIOS-e820: [mem 0x000000007fee0000-0x000000007fefefff] ACPI data May 10 00:40:03.647508 kernel: BIOS-e820: [mem 0x000000007feff000-0x000000007fefffff] ACPI NVS May 10 00:40:03.647512 kernel: BIOS-e820: [mem 0x000000007ff00000-0x000000007fffffff] usable May 10 00:40:03.647516 kernel: BIOS-e820: [mem 0x00000000f0000000-0x00000000f7ffffff] reserved May 10 00:40:03.647520 kernel: BIOS-e820: [mem 0x00000000fec00000-0x00000000fec0ffff] reserved May 10 00:40:03.647524 kernel: BIOS-e820: [mem 0x00000000fee00000-0x00000000fee00fff] reserved May 10 00:40:03.647530 kernel: BIOS-e820: [mem 0x00000000fffe0000-0x00000000ffffffff] reserved May 10 00:40:03.647534 kernel: NX (Execute Disable) protection: active May 10 00:40:03.647538 kernel: SMBIOS 2.7 present. May 10 00:40:03.647543 kernel: DMI: VMware, Inc. VMware Virtual Platform/440BX Desktop Reference Platform, BIOS 6.00 05/28/2020 May 10 00:40:03.647547 kernel: vmware: hypercall mode: 0x00 May 10 00:40:03.647551 kernel: Hypervisor detected: VMware May 10 00:40:03.647556 kernel: vmware: TSC freq read from hypervisor : 3408.000 MHz May 10 00:40:03.647560 kernel: vmware: Host bus clock speed read from hypervisor : 66000000 Hz May 10 00:40:03.647564 kernel: vmware: using clock offset of 2717547870 ns May 10 00:40:03.647569 kernel: tsc: Detected 3408.000 MHz processor May 10 00:40:03.647573 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved May 10 00:40:03.647578 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable May 10 00:40:03.647582 kernel: last_pfn = 0x80000 max_arch_pfn = 0x400000000 May 10 00:40:03.647587 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT May 10 00:40:03.647591 kernel: total RAM covered: 3072M May 10 00:40:03.647596 kernel: Found optimal setting for mtrr clean up May 10 00:40:03.647601 kernel: gran_size: 64K chunk_size: 64K num_reg: 2 lose cover RAM: 0G May 10 00:40:03.647605 kernel: Using GB pages for direct mapping May 10 00:40:03.647609 kernel: ACPI: Early table checksum verification disabled May 10 00:40:03.647614 kernel: ACPI: RSDP 0x00000000000F6A00 000024 (v02 PTLTD ) May 10 00:40:03.647618 kernel: ACPI: XSDT 0x000000007FEE965B 00005C (v01 INTEL 440BX 06040000 VMW 01324272) May 10 00:40:03.647622 kernel: ACPI: FACP 0x000000007FEFEE73 0000F4 (v04 INTEL 440BX 06040000 PTL 000F4240) May 10 00:40:03.647627 kernel: ACPI: DSDT 0x000000007FEEAD55 01411E (v01 PTLTD Custom 06040000 MSFT 03000001) May 10 00:40:03.647631 kernel: ACPI: FACS 0x000000007FEFFFC0 000040 May 10 00:40:03.647635 kernel: ACPI: FACS 0x000000007FEFFFC0 000040 May 10 00:40:03.647640 kernel: ACPI: BOOT 0x000000007FEEAD2D 000028 (v01 PTLTD $SBFTBL$ 06040000 LTP 00000001) May 10 00:40:03.647646 kernel: ACPI: APIC 0x000000007FEEA5EB 000742 (v01 PTLTD ? APIC 06040000 LTP 00000000) May 10 00:40:03.647651 kernel: ACPI: MCFG 0x000000007FEEA5AF 00003C (v01 PTLTD $PCITBL$ 06040000 LTP 00000001) May 10 00:40:03.647656 kernel: ACPI: SRAT 0x000000007FEE9757 0008A8 (v02 VMWARE MEMPLUG 06040000 VMW 00000001) May 10 00:40:03.647660 kernel: ACPI: HPET 0x000000007FEE971F 000038 (v01 VMWARE VMW HPET 06040000 VMW 00000001) May 10 00:40:03.647666 kernel: ACPI: WAET 0x000000007FEE96F7 000028 (v01 VMWARE VMW WAET 06040000 VMW 00000001) May 10 00:40:03.647683 kernel: ACPI: Reserving FACP table memory at [mem 0x7fefee73-0x7fefef66] May 10 00:40:03.647699 kernel: ACPI: Reserving DSDT table memory at [mem 0x7feead55-0x7fefee72] May 10 00:40:03.647705 kernel: ACPI: Reserving FACS table memory at [mem 0x7fefffc0-0x7fefffff] May 10 00:40:03.647709 kernel: ACPI: Reserving FACS table memory at [mem 0x7fefffc0-0x7fefffff] May 10 00:40:03.647720 kernel: ACPI: Reserving BOOT table memory at [mem 0x7feead2d-0x7feead54] May 10 00:40:03.647725 kernel: ACPI: Reserving APIC table memory at [mem 0x7feea5eb-0x7feead2c] May 10 00:40:03.647730 kernel: ACPI: Reserving MCFG table memory at [mem 0x7feea5af-0x7feea5ea] May 10 00:40:03.647735 kernel: ACPI: Reserving SRAT table memory at [mem 0x7fee9757-0x7fee9ffe] May 10 00:40:03.647741 kernel: ACPI: Reserving HPET table memory at [mem 0x7fee971f-0x7fee9756] May 10 00:40:03.647745 kernel: ACPI: Reserving WAET table memory at [mem 0x7fee96f7-0x7fee971e] May 10 00:40:03.647750 kernel: system APIC only can use physical flat May 10 00:40:03.647755 kernel: Setting APIC routing to physical flat. May 10 00:40:03.647759 kernel: SRAT: PXM 0 -> APIC 0x00 -> Node 0 May 10 00:40:03.647764 kernel: SRAT: PXM 0 -> APIC 0x02 -> Node 0 May 10 00:40:03.647768 kernel: SRAT: PXM 0 -> APIC 0x04 -> Node 0 May 10 00:40:03.647773 kernel: SRAT: PXM 0 -> APIC 0x06 -> Node 0 May 10 00:40:03.647777 kernel: SRAT: PXM 0 -> APIC 0x08 -> Node 0 May 10 00:40:03.647783 kernel: SRAT: PXM 0 -> APIC 0x0a -> Node 0 May 10 00:40:03.647787 kernel: SRAT: PXM 0 -> APIC 0x0c -> Node 0 May 10 00:40:03.647792 kernel: SRAT: PXM 0 -> APIC 0x0e -> Node 0 May 10 00:40:03.647796 kernel: SRAT: PXM 0 -> APIC 0x10 -> Node 0 May 10 00:40:03.647804 kernel: SRAT: PXM 0 -> APIC 0x12 -> Node 0 May 10 00:40:03.647809 kernel: SRAT: PXM 0 -> APIC 0x14 -> Node 0 May 10 00:40:03.647816 kernel: SRAT: PXM 0 -> APIC 0x16 -> Node 0 May 10 00:40:03.647821 kernel: SRAT: PXM 0 -> APIC 0x18 -> Node 0 May 10 00:40:03.647825 kernel: SRAT: PXM 0 -> APIC 0x1a -> Node 0 May 10 00:40:03.647830 kernel: SRAT: PXM 0 -> APIC 0x1c -> Node 0 May 10 00:40:03.647835 kernel: SRAT: PXM 0 -> APIC 0x1e -> Node 0 May 10 00:40:03.647840 kernel: SRAT: PXM 0 -> APIC 0x20 -> Node 0 May 10 00:40:03.647844 kernel: SRAT: PXM 0 -> APIC 0x22 -> Node 0 May 10 00:40:03.647849 kernel: SRAT: PXM 0 -> APIC 0x24 -> Node 0 May 10 00:40:03.647853 kernel: SRAT: PXM 0 -> APIC 0x26 -> Node 0 May 10 00:40:03.647858 kernel: SRAT: PXM 0 -> APIC 0x28 -> Node 0 May 10 00:40:03.647863 kernel: SRAT: PXM 0 -> APIC 0x2a -> Node 0 May 10 00:40:03.647867 kernel: SRAT: PXM 0 -> APIC 0x2c -> Node 0 May 10 00:40:03.647872 kernel: SRAT: PXM 0 -> APIC 0x2e -> Node 0 May 10 00:40:03.647876 kernel: SRAT: PXM 0 -> APIC 0x30 -> Node 0 May 10 00:40:03.647882 kernel: SRAT: PXM 0 -> APIC 0x32 -> Node 0 May 10 00:40:03.647886 kernel: SRAT: PXM 0 -> APIC 0x34 -> Node 0 May 10 00:40:03.647891 kernel: SRAT: PXM 0 -> APIC 0x36 -> Node 0 May 10 00:40:03.647895 kernel: SRAT: PXM 0 -> APIC 0x38 -> Node 0 May 10 00:40:03.647900 kernel: SRAT: PXM 0 -> APIC 0x3a -> Node 0 May 10 00:40:03.647914 kernel: SRAT: PXM 0 -> APIC 0x3c -> Node 0 May 10 00:40:03.647919 kernel: SRAT: PXM 0 -> APIC 0x3e -> Node 0 May 10 00:40:03.647924 kernel: SRAT: PXM 0 -> APIC 0x40 -> Node 0 May 10 00:40:03.647929 kernel: SRAT: PXM 0 -> APIC 0x42 -> Node 0 May 10 00:40:03.647933 kernel: SRAT: PXM 0 -> APIC 0x44 -> Node 0 May 10 00:40:03.647939 kernel: SRAT: PXM 0 -> APIC 0x46 -> Node 0 May 10 00:40:03.647944 kernel: SRAT: PXM 0 -> APIC 0x48 -> Node 0 May 10 00:40:03.647949 kernel: SRAT: PXM 0 -> APIC 0x4a -> Node 0 May 10 00:40:03.647953 kernel: SRAT: PXM 0 -> APIC 0x4c -> Node 0 May 10 00:40:03.647958 kernel: SRAT: PXM 0 -> APIC 0x4e -> Node 0 May 10 00:40:03.647962 kernel: SRAT: PXM 0 -> APIC 0x50 -> Node 0 May 10 00:40:03.647967 kernel: SRAT: PXM 0 -> APIC 0x52 -> Node 0 May 10 00:40:03.647971 kernel: SRAT: PXM 0 -> APIC 0x54 -> Node 0 May 10 00:40:03.647976 kernel: SRAT: PXM 0 -> APIC 0x56 -> Node 0 May 10 00:40:03.647981 kernel: SRAT: PXM 0 -> APIC 0x58 -> Node 0 May 10 00:40:03.647986 kernel: SRAT: PXM 0 -> APIC 0x5a -> Node 0 May 10 00:40:03.647991 kernel: SRAT: PXM 0 -> APIC 0x5c -> Node 0 May 10 00:40:03.647995 kernel: SRAT: PXM 0 -> APIC 0x5e -> Node 0 May 10 00:40:03.648000 kernel: SRAT: PXM 0 -> APIC 0x60 -> Node 0 May 10 00:40:03.648004 kernel: SRAT: PXM 0 -> APIC 0x62 -> Node 0 May 10 00:40:03.648009 kernel: SRAT: PXM 0 -> APIC 0x64 -> Node 0 May 10 00:40:03.648013 kernel: SRAT: PXM 0 -> APIC 0x66 -> Node 0 May 10 00:40:03.648018 kernel: SRAT: PXM 0 -> APIC 0x68 -> Node 0 May 10 00:40:03.648022 kernel: SRAT: PXM 0 -> APIC 0x6a -> Node 0 May 10 00:40:03.648027 kernel: SRAT: PXM 0 -> APIC 0x6c -> Node 0 May 10 00:40:03.648032 kernel: SRAT: PXM 0 -> APIC 0x6e -> Node 0 May 10 00:40:03.648037 kernel: SRAT: PXM 0 -> APIC 0x70 -> Node 0 May 10 00:40:03.648041 kernel: SRAT: PXM 0 -> APIC 0x72 -> Node 0 May 10 00:40:03.648046 kernel: SRAT: PXM 0 -> APIC 0x74 -> Node 0 May 10 00:40:03.648050 kernel: SRAT: PXM 0 -> APIC 0x76 -> Node 0 May 10 00:40:03.648055 kernel: SRAT: PXM 0 -> APIC 0x78 -> Node 0 May 10 00:40:03.648063 kernel: SRAT: PXM 0 -> APIC 0x7a -> Node 0 May 10 00:40:03.648069 kernel: SRAT: PXM 0 -> APIC 0x7c -> Node 0 May 10 00:40:03.648074 kernel: SRAT: PXM 0 -> APIC 0x7e -> Node 0 May 10 00:40:03.648078 kernel: SRAT: PXM 0 -> APIC 0x80 -> Node 0 May 10 00:40:03.648083 kernel: SRAT: PXM 0 -> APIC 0x82 -> Node 0 May 10 00:40:03.648089 kernel: SRAT: PXM 0 -> APIC 0x84 -> Node 0 May 10 00:40:03.648094 kernel: SRAT: PXM 0 -> APIC 0x86 -> Node 0 May 10 00:40:03.648099 kernel: SRAT: PXM 0 -> APIC 0x88 -> Node 0 May 10 00:40:03.648104 kernel: SRAT: PXM 0 -> APIC 0x8a -> Node 0 May 10 00:40:03.648109 kernel: SRAT: PXM 0 -> APIC 0x8c -> Node 0 May 10 00:40:03.648113 kernel: SRAT: PXM 0 -> APIC 0x8e -> Node 0 May 10 00:40:03.648118 kernel: SRAT: PXM 0 -> APIC 0x90 -> Node 0 May 10 00:40:03.648124 kernel: SRAT: PXM 0 -> APIC 0x92 -> Node 0 May 10 00:40:03.648129 kernel: SRAT: PXM 0 -> APIC 0x94 -> Node 0 May 10 00:40:03.648134 kernel: SRAT: PXM 0 -> APIC 0x96 -> Node 0 May 10 00:40:03.648139 kernel: SRAT: PXM 0 -> APIC 0x98 -> Node 0 May 10 00:40:03.648144 kernel: SRAT: PXM 0 -> APIC 0x9a -> Node 0 May 10 00:40:03.648149 kernel: SRAT: PXM 0 -> APIC 0x9c -> Node 0 May 10 00:40:03.648153 kernel: SRAT: PXM 0 -> APIC 0x9e -> Node 0 May 10 00:40:03.648158 kernel: SRAT: PXM 0 -> APIC 0xa0 -> Node 0 May 10 00:40:03.648163 kernel: SRAT: PXM 0 -> APIC 0xa2 -> Node 0 May 10 00:40:03.648168 kernel: SRAT: PXM 0 -> APIC 0xa4 -> Node 0 May 10 00:40:03.648174 kernel: SRAT: PXM 0 -> APIC 0xa6 -> Node 0 May 10 00:40:03.648179 kernel: SRAT: PXM 0 -> APIC 0xa8 -> Node 0 May 10 00:40:03.648202 kernel: SRAT: PXM 0 -> APIC 0xaa -> Node 0 May 10 00:40:03.648207 kernel: SRAT: PXM 0 -> APIC 0xac -> Node 0 May 10 00:40:03.648212 kernel: SRAT: PXM 0 -> APIC 0xae -> Node 0 May 10 00:40:03.648217 kernel: SRAT: PXM 0 -> APIC 0xb0 -> Node 0 May 10 00:40:03.648222 kernel: SRAT: PXM 0 -> APIC 0xb2 -> Node 0 May 10 00:40:03.648227 kernel: SRAT: PXM 0 -> APIC 0xb4 -> Node 0 May 10 00:40:03.648232 kernel: SRAT: PXM 0 -> APIC 0xb6 -> Node 0 May 10 00:40:03.648238 kernel: SRAT: PXM 0 -> APIC 0xb8 -> Node 0 May 10 00:40:03.648243 kernel: SRAT: PXM 0 -> APIC 0xba -> Node 0 May 10 00:40:03.648262 kernel: SRAT: PXM 0 -> APIC 0xbc -> Node 0 May 10 00:40:03.648267 kernel: SRAT: PXM 0 -> APIC 0xbe -> Node 0 May 10 00:40:03.648272 kernel: SRAT: PXM 0 -> APIC 0xc0 -> Node 0 May 10 00:40:03.648277 kernel: SRAT: PXM 0 -> APIC 0xc2 -> Node 0 May 10 00:40:03.648282 kernel: SRAT: PXM 0 -> APIC 0xc4 -> Node 0 May 10 00:40:03.648286 kernel: SRAT: PXM 0 -> APIC 0xc6 -> Node 0 May 10 00:40:03.648291 kernel: SRAT: PXM 0 -> APIC 0xc8 -> Node 0 May 10 00:40:03.648296 kernel: SRAT: PXM 0 -> APIC 0xca -> Node 0 May 10 00:40:03.648302 kernel: SRAT: PXM 0 -> APIC 0xcc -> Node 0 May 10 00:40:03.648307 kernel: SRAT: PXM 0 -> APIC 0xce -> Node 0 May 10 00:40:03.648312 kernel: SRAT: PXM 0 -> APIC 0xd0 -> Node 0 May 10 00:40:03.648316 kernel: SRAT: PXM 0 -> APIC 0xd2 -> Node 0 May 10 00:40:03.648321 kernel: SRAT: PXM 0 -> APIC 0xd4 -> Node 0 May 10 00:40:03.648326 kernel: SRAT: PXM 0 -> APIC 0xd6 -> Node 0 May 10 00:40:03.648331 kernel: SRAT: PXM 0 -> APIC 0xd8 -> Node 0 May 10 00:40:03.648336 kernel: SRAT: PXM 0 -> APIC 0xda -> Node 0 May 10 00:40:03.648341 kernel: SRAT: PXM 0 -> APIC 0xdc -> Node 0 May 10 00:40:03.648346 kernel: SRAT: PXM 0 -> APIC 0xde -> Node 0 May 10 00:40:03.648351 kernel: SRAT: PXM 0 -> APIC 0xe0 -> Node 0 May 10 00:40:03.648356 kernel: SRAT: PXM 0 -> APIC 0xe2 -> Node 0 May 10 00:40:03.648361 kernel: SRAT: PXM 0 -> APIC 0xe4 -> Node 0 May 10 00:40:03.648366 kernel: SRAT: PXM 0 -> APIC 0xe6 -> Node 0 May 10 00:40:03.648371 kernel: SRAT: PXM 0 -> APIC 0xe8 -> Node 0 May 10 00:40:03.648376 kernel: SRAT: PXM 0 -> APIC 0xea -> Node 0 May 10 00:40:03.648380 kernel: SRAT: PXM 0 -> APIC 0xec -> Node 0 May 10 00:40:03.648385 kernel: SRAT: PXM 0 -> APIC 0xee -> Node 0 May 10 00:40:03.648390 kernel: SRAT: PXM 0 -> APIC 0xf0 -> Node 0 May 10 00:40:03.648395 kernel: SRAT: PXM 0 -> APIC 0xf2 -> Node 0 May 10 00:40:03.648401 kernel: SRAT: PXM 0 -> APIC 0xf4 -> Node 0 May 10 00:40:03.648406 kernel: SRAT: PXM 0 -> APIC 0xf6 -> Node 0 May 10 00:40:03.648411 kernel: SRAT: PXM 0 -> APIC 0xf8 -> Node 0 May 10 00:40:03.648416 kernel: SRAT: PXM 0 -> APIC 0xfa -> Node 0 May 10 00:40:03.648421 kernel: SRAT: PXM 0 -> APIC 0xfc -> Node 0 May 10 00:40:03.648426 kernel: SRAT: PXM 0 -> APIC 0xfe -> Node 0 May 10 00:40:03.648431 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x0009ffff] May 10 00:40:03.648436 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00100000-0x7fffffff] May 10 00:40:03.648441 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x80000000-0xbfffffff] hotplug May 10 00:40:03.648447 kernel: NUMA: Node 0 [mem 0x00000000-0x0009ffff] + [mem 0x00100000-0x7fffffff] -> [mem 0x00000000-0x7fffffff] May 10 00:40:03.648452 kernel: NODE_DATA(0) allocated [mem 0x7fffa000-0x7fffffff] May 10 00:40:03.648457 kernel: Zone ranges: May 10 00:40:03.648462 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] May 10 00:40:03.648467 kernel: DMA32 [mem 0x0000000001000000-0x000000007fffffff] May 10 00:40:03.648472 kernel: Normal empty May 10 00:40:03.648477 kernel: Movable zone start for each node May 10 00:40:03.648482 kernel: Early memory node ranges May 10 00:40:03.648487 kernel: node 0: [mem 0x0000000000001000-0x000000000009dfff] May 10 00:40:03.648492 kernel: node 0: [mem 0x0000000000100000-0x000000007fedffff] May 10 00:40:03.648498 kernel: node 0: [mem 0x000000007ff00000-0x000000007fffffff] May 10 00:40:03.648503 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007fffffff] May 10 00:40:03.648507 kernel: On node 0, zone DMA: 1 pages in unavailable ranges May 10 00:40:03.648512 kernel: On node 0, zone DMA: 98 pages in unavailable ranges May 10 00:40:03.648517 kernel: On node 0, zone DMA32: 32 pages in unavailable ranges May 10 00:40:03.648522 kernel: ACPI: PM-Timer IO Port: 0x1008 May 10 00:40:03.648527 kernel: system APIC only can use physical flat May 10 00:40:03.648532 kernel: ACPI: LAPIC_NMI (acpi_id[0x00] high edge lint[0x1]) May 10 00:40:03.648537 kernel: ACPI: LAPIC_NMI (acpi_id[0x01] high edge lint[0x1]) May 10 00:40:03.648543 kernel: ACPI: LAPIC_NMI (acpi_id[0x02] high edge lint[0x1]) May 10 00:40:03.648548 kernel: ACPI: LAPIC_NMI (acpi_id[0x03] high edge lint[0x1]) May 10 00:40:03.648553 kernel: ACPI: LAPIC_NMI (acpi_id[0x04] high edge lint[0x1]) May 10 00:40:03.648558 kernel: ACPI: LAPIC_NMI (acpi_id[0x05] high edge lint[0x1]) May 10 00:40:03.648563 kernel: ACPI: LAPIC_NMI (acpi_id[0x06] high edge lint[0x1]) May 10 00:40:03.648568 kernel: ACPI: LAPIC_NMI (acpi_id[0x07] high edge lint[0x1]) May 10 00:40:03.648573 kernel: ACPI: LAPIC_NMI (acpi_id[0x08] high edge lint[0x1]) May 10 00:40:03.648577 kernel: ACPI: LAPIC_NMI (acpi_id[0x09] high edge lint[0x1]) May 10 00:40:03.648582 kernel: ACPI: LAPIC_NMI (acpi_id[0x0a] high edge lint[0x1]) May 10 00:40:03.648587 kernel: ACPI: LAPIC_NMI (acpi_id[0x0b] high edge lint[0x1]) May 10 00:40:03.648593 kernel: ACPI: LAPIC_NMI (acpi_id[0x0c] high edge lint[0x1]) May 10 00:40:03.648598 kernel: ACPI: LAPIC_NMI (acpi_id[0x0d] high edge lint[0x1]) May 10 00:40:03.648603 kernel: ACPI: LAPIC_NMI (acpi_id[0x0e] high edge lint[0x1]) May 10 00:40:03.648608 kernel: ACPI: LAPIC_NMI (acpi_id[0x0f] high edge lint[0x1]) May 10 00:40:03.648613 kernel: ACPI: LAPIC_NMI (acpi_id[0x10] high edge lint[0x1]) May 10 00:40:03.648617 kernel: ACPI: LAPIC_NMI (acpi_id[0x11] high edge lint[0x1]) May 10 00:40:03.648622 kernel: ACPI: LAPIC_NMI (acpi_id[0x12] high edge lint[0x1]) May 10 00:40:03.648627 kernel: ACPI: LAPIC_NMI (acpi_id[0x13] high edge lint[0x1]) May 10 00:40:03.648632 kernel: ACPI: LAPIC_NMI (acpi_id[0x14] high edge lint[0x1]) May 10 00:40:03.648638 kernel: ACPI: LAPIC_NMI (acpi_id[0x15] high edge lint[0x1]) May 10 00:40:03.648643 kernel: ACPI: LAPIC_NMI (acpi_id[0x16] high edge lint[0x1]) May 10 00:40:03.648648 kernel: ACPI: LAPIC_NMI (acpi_id[0x17] high edge lint[0x1]) May 10 00:40:03.648652 kernel: ACPI: LAPIC_NMI (acpi_id[0x18] high edge lint[0x1]) May 10 00:40:03.648657 kernel: ACPI: LAPIC_NMI (acpi_id[0x19] high edge lint[0x1]) May 10 00:40:03.648662 kernel: ACPI: LAPIC_NMI (acpi_id[0x1a] high edge lint[0x1]) May 10 00:40:03.648667 kernel: ACPI: LAPIC_NMI (acpi_id[0x1b] high edge lint[0x1]) May 10 00:40:03.648672 kernel: ACPI: LAPIC_NMI (acpi_id[0x1c] high edge lint[0x1]) May 10 00:40:03.648677 kernel: ACPI: LAPIC_NMI (acpi_id[0x1d] high edge lint[0x1]) May 10 00:40:03.648683 kernel: ACPI: LAPIC_NMI (acpi_id[0x1e] high edge lint[0x1]) May 10 00:40:03.648688 kernel: ACPI: LAPIC_NMI (acpi_id[0x1f] high edge lint[0x1]) May 10 00:40:03.648693 kernel: ACPI: LAPIC_NMI (acpi_id[0x20] high edge lint[0x1]) May 10 00:40:03.648698 kernel: ACPI: LAPIC_NMI (acpi_id[0x21] high edge lint[0x1]) May 10 00:40:03.648703 kernel: ACPI: LAPIC_NMI (acpi_id[0x22] high edge lint[0x1]) May 10 00:40:03.648708 kernel: ACPI: LAPIC_NMI (acpi_id[0x23] high edge lint[0x1]) May 10 00:40:03.648712 kernel: ACPI: LAPIC_NMI (acpi_id[0x24] high edge lint[0x1]) May 10 00:40:03.648717 kernel: ACPI: LAPIC_NMI (acpi_id[0x25] high edge lint[0x1]) May 10 00:40:03.648722 kernel: ACPI: LAPIC_NMI (acpi_id[0x26] high edge lint[0x1]) May 10 00:40:03.648727 kernel: ACPI: LAPIC_NMI (acpi_id[0x27] high edge lint[0x1]) May 10 00:40:03.648733 kernel: ACPI: LAPIC_NMI (acpi_id[0x28] high edge lint[0x1]) May 10 00:40:03.648738 kernel: ACPI: LAPIC_NMI (acpi_id[0x29] high edge lint[0x1]) May 10 00:40:03.648743 kernel: ACPI: LAPIC_NMI (acpi_id[0x2a] high edge lint[0x1]) May 10 00:40:03.648748 kernel: ACPI: LAPIC_NMI (acpi_id[0x2b] high edge lint[0x1]) May 10 00:40:03.648753 kernel: ACPI: LAPIC_NMI (acpi_id[0x2c] high edge lint[0x1]) May 10 00:40:03.648758 kernel: ACPI: LAPIC_NMI (acpi_id[0x2d] high edge lint[0x1]) May 10 00:40:03.648763 kernel: ACPI: LAPIC_NMI (acpi_id[0x2e] high edge lint[0x1]) May 10 00:40:03.648768 kernel: ACPI: LAPIC_NMI (acpi_id[0x2f] high edge lint[0x1]) May 10 00:40:03.648773 kernel: ACPI: LAPIC_NMI (acpi_id[0x30] high edge lint[0x1]) May 10 00:40:03.648779 kernel: ACPI: LAPIC_NMI (acpi_id[0x31] high edge lint[0x1]) May 10 00:40:03.648784 kernel: ACPI: LAPIC_NMI (acpi_id[0x32] high edge lint[0x1]) May 10 00:40:03.648788 kernel: ACPI: LAPIC_NMI (acpi_id[0x33] high edge lint[0x1]) May 10 00:40:03.648793 kernel: ACPI: LAPIC_NMI (acpi_id[0x34] high edge lint[0x1]) May 10 00:40:03.648798 kernel: ACPI: LAPIC_NMI (acpi_id[0x35] high edge lint[0x1]) May 10 00:40:03.648803 kernel: ACPI: LAPIC_NMI (acpi_id[0x36] high edge lint[0x1]) May 10 00:40:03.648808 kernel: ACPI: LAPIC_NMI (acpi_id[0x37] high edge lint[0x1]) May 10 00:40:03.648813 kernel: ACPI: LAPIC_NMI (acpi_id[0x38] high edge lint[0x1]) May 10 00:40:03.648818 kernel: ACPI: LAPIC_NMI (acpi_id[0x39] high edge lint[0x1]) May 10 00:40:03.648824 kernel: ACPI: LAPIC_NMI (acpi_id[0x3a] high edge lint[0x1]) May 10 00:40:03.648828 kernel: ACPI: LAPIC_NMI (acpi_id[0x3b] high edge lint[0x1]) May 10 00:40:03.648833 kernel: ACPI: LAPIC_NMI (acpi_id[0x3c] high edge lint[0x1]) May 10 00:40:03.648838 kernel: ACPI: LAPIC_NMI (acpi_id[0x3d] high edge lint[0x1]) May 10 00:40:03.648843 kernel: ACPI: LAPIC_NMI (acpi_id[0x3e] high edge lint[0x1]) May 10 00:40:03.648848 kernel: ACPI: LAPIC_NMI (acpi_id[0x3f] high edge lint[0x1]) May 10 00:40:03.648853 kernel: ACPI: LAPIC_NMI (acpi_id[0x40] high edge lint[0x1]) May 10 00:40:03.648858 kernel: ACPI: LAPIC_NMI (acpi_id[0x41] high edge lint[0x1]) May 10 00:40:03.648863 kernel: ACPI: LAPIC_NMI (acpi_id[0x42] high edge lint[0x1]) May 10 00:40:03.648868 kernel: ACPI: LAPIC_NMI (acpi_id[0x43] high edge lint[0x1]) May 10 00:40:03.648873 kernel: ACPI: LAPIC_NMI (acpi_id[0x44] high edge lint[0x1]) May 10 00:40:03.648878 kernel: ACPI: LAPIC_NMI (acpi_id[0x45] high edge lint[0x1]) May 10 00:40:03.648883 kernel: ACPI: LAPIC_NMI (acpi_id[0x46] high edge lint[0x1]) May 10 00:40:03.648888 kernel: ACPI: LAPIC_NMI (acpi_id[0x47] high edge lint[0x1]) May 10 00:40:03.648893 kernel: ACPI: LAPIC_NMI (acpi_id[0x48] high edge lint[0x1]) May 10 00:40:03.648898 kernel: ACPI: LAPIC_NMI (acpi_id[0x49] high edge lint[0x1]) May 10 00:40:03.648903 kernel: ACPI: LAPIC_NMI (acpi_id[0x4a] high edge lint[0x1]) May 10 00:40:03.648942 kernel: ACPI: LAPIC_NMI (acpi_id[0x4b] high edge lint[0x1]) May 10 00:40:03.648949 kernel: ACPI: LAPIC_NMI (acpi_id[0x4c] high edge lint[0x1]) May 10 00:40:03.648956 kernel: ACPI: LAPIC_NMI (acpi_id[0x4d] high edge lint[0x1]) May 10 00:40:03.648961 kernel: ACPI: LAPIC_NMI (acpi_id[0x4e] high edge lint[0x1]) May 10 00:40:03.648966 kernel: ACPI: LAPIC_NMI (acpi_id[0x4f] high edge lint[0x1]) May 10 00:40:03.648971 kernel: ACPI: LAPIC_NMI (acpi_id[0x50] high edge lint[0x1]) May 10 00:40:03.648976 kernel: ACPI: LAPIC_NMI (acpi_id[0x51] high edge lint[0x1]) May 10 00:40:03.648981 kernel: ACPI: LAPIC_NMI (acpi_id[0x52] high edge lint[0x1]) May 10 00:40:03.648986 kernel: ACPI: LAPIC_NMI (acpi_id[0x53] high edge lint[0x1]) May 10 00:40:03.648991 kernel: ACPI: LAPIC_NMI (acpi_id[0x54] high edge lint[0x1]) May 10 00:40:03.648996 kernel: ACPI: LAPIC_NMI (acpi_id[0x55] high edge lint[0x1]) May 10 00:40:03.649002 kernel: ACPI: LAPIC_NMI (acpi_id[0x56] high edge lint[0x1]) May 10 00:40:03.649006 kernel: ACPI: LAPIC_NMI (acpi_id[0x57] high edge lint[0x1]) May 10 00:40:03.649012 kernel: ACPI: LAPIC_NMI (acpi_id[0x58] high edge lint[0x1]) May 10 00:40:03.649017 kernel: ACPI: LAPIC_NMI (acpi_id[0x59] high edge lint[0x1]) May 10 00:40:03.649021 kernel: ACPI: LAPIC_NMI (acpi_id[0x5a] high edge lint[0x1]) May 10 00:40:03.649026 kernel: ACPI: LAPIC_NMI (acpi_id[0x5b] high edge lint[0x1]) May 10 00:40:03.649031 kernel: ACPI: LAPIC_NMI (acpi_id[0x5c] high edge lint[0x1]) May 10 00:40:03.649036 kernel: ACPI: LAPIC_NMI (acpi_id[0x5d] high edge lint[0x1]) May 10 00:40:03.649041 kernel: ACPI: LAPIC_NMI (acpi_id[0x5e] high edge lint[0x1]) May 10 00:40:03.649052 kernel: ACPI: LAPIC_NMI (acpi_id[0x5f] high edge lint[0x1]) May 10 00:40:03.649058 kernel: ACPI: LAPIC_NMI (acpi_id[0x60] high edge lint[0x1]) May 10 00:40:03.649063 kernel: ACPI: LAPIC_NMI (acpi_id[0x61] high edge lint[0x1]) May 10 00:40:03.649068 kernel: ACPI: LAPIC_NMI (acpi_id[0x62] high edge lint[0x1]) May 10 00:40:03.649072 kernel: ACPI: LAPIC_NMI (acpi_id[0x63] high edge lint[0x1]) May 10 00:40:03.649078 kernel: ACPI: LAPIC_NMI (acpi_id[0x64] high edge lint[0x1]) May 10 00:40:03.649082 kernel: ACPI: LAPIC_NMI (acpi_id[0x65] high edge lint[0x1]) May 10 00:40:03.649087 kernel: ACPI: LAPIC_NMI (acpi_id[0x66] high edge lint[0x1]) May 10 00:40:03.649092 kernel: ACPI: LAPIC_NMI (acpi_id[0x67] high edge lint[0x1]) May 10 00:40:03.649097 kernel: ACPI: LAPIC_NMI (acpi_id[0x68] high edge lint[0x1]) May 10 00:40:03.649103 kernel: ACPI: LAPIC_NMI (acpi_id[0x69] high edge lint[0x1]) May 10 00:40:03.649108 kernel: ACPI: LAPIC_NMI (acpi_id[0x6a] high edge lint[0x1]) May 10 00:40:03.649113 kernel: ACPI: LAPIC_NMI (acpi_id[0x6b] high edge lint[0x1]) May 10 00:40:03.649118 kernel: ACPI: LAPIC_NMI (acpi_id[0x6c] high edge lint[0x1]) May 10 00:40:03.649123 kernel: ACPI: LAPIC_NMI (acpi_id[0x6d] high edge lint[0x1]) May 10 00:40:03.649128 kernel: ACPI: LAPIC_NMI (acpi_id[0x6e] high edge lint[0x1]) May 10 00:40:03.649133 kernel: ACPI: LAPIC_NMI (acpi_id[0x6f] high edge lint[0x1]) May 10 00:40:03.649138 kernel: ACPI: LAPIC_NMI (acpi_id[0x70] high edge lint[0x1]) May 10 00:40:03.649143 kernel: ACPI: LAPIC_NMI (acpi_id[0x71] high edge lint[0x1]) May 10 00:40:03.649148 kernel: ACPI: LAPIC_NMI (acpi_id[0x72] high edge lint[0x1]) May 10 00:40:03.649153 kernel: ACPI: LAPIC_NMI (acpi_id[0x73] high edge lint[0x1]) May 10 00:40:03.649158 kernel: ACPI: LAPIC_NMI (acpi_id[0x74] high edge lint[0x1]) May 10 00:40:03.649163 kernel: ACPI: LAPIC_NMI (acpi_id[0x75] high edge lint[0x1]) May 10 00:40:03.649168 kernel: ACPI: LAPIC_NMI (acpi_id[0x76] high edge lint[0x1]) May 10 00:40:03.649173 kernel: ACPI: LAPIC_NMI (acpi_id[0x77] high edge lint[0x1]) May 10 00:40:03.649178 kernel: ACPI: LAPIC_NMI (acpi_id[0x78] high edge lint[0x1]) May 10 00:40:03.649183 kernel: ACPI: LAPIC_NMI (acpi_id[0x79] high edge lint[0x1]) May 10 00:40:03.649188 kernel: ACPI: LAPIC_NMI (acpi_id[0x7a] high edge lint[0x1]) May 10 00:40:03.649194 kernel: ACPI: LAPIC_NMI (acpi_id[0x7b] high edge lint[0x1]) May 10 00:40:03.649199 kernel: ACPI: LAPIC_NMI (acpi_id[0x7c] high edge lint[0x1]) May 10 00:40:03.649204 kernel: ACPI: LAPIC_NMI (acpi_id[0x7d] high edge lint[0x1]) May 10 00:40:03.649209 kernel: ACPI: LAPIC_NMI (acpi_id[0x7e] high edge lint[0x1]) May 10 00:40:03.649214 kernel: ACPI: LAPIC_NMI (acpi_id[0x7f] high edge lint[0x1]) May 10 00:40:03.649219 kernel: IOAPIC[0]: apic_id 1, version 17, address 0xfec00000, GSI 0-23 May 10 00:40:03.649224 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 high edge) May 10 00:40:03.649229 kernel: ACPI: Using ACPI (MADT) for SMP configuration information May 10 00:40:03.649234 kernel: ACPI: HPET id: 0x8086af01 base: 0xfed00000 May 10 00:40:03.649239 kernel: TSC deadline timer available May 10 00:40:03.649245 kernel: smpboot: Allowing 128 CPUs, 126 hotplug CPUs May 10 00:40:03.649250 kernel: [mem 0x80000000-0xefffffff] available for PCI devices May 10 00:40:03.649255 kernel: Booting paravirtualized kernel on VMware hypervisor May 10 00:40:03.649260 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns May 10 00:40:03.649265 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:512 nr_cpu_ids:128 nr_node_ids:1 May 10 00:40:03.649270 kernel: percpu: Embedded 56 pages/cpu s188696 r8192 d32488 u262144 May 10 00:40:03.649275 kernel: pcpu-alloc: s188696 r8192 d32488 u262144 alloc=1*2097152 May 10 00:40:03.649280 kernel: pcpu-alloc: [0] 000 001 002 003 004 005 006 007 May 10 00:40:03.649286 kernel: pcpu-alloc: [0] 008 009 010 011 012 013 014 015 May 10 00:40:03.649291 kernel: pcpu-alloc: [0] 016 017 018 019 020 021 022 023 May 10 00:40:03.649296 kernel: pcpu-alloc: [0] 024 025 026 027 028 029 030 031 May 10 00:40:03.649300 kernel: pcpu-alloc: [0] 032 033 034 035 036 037 038 039 May 10 00:40:03.649305 kernel: pcpu-alloc: [0] 040 041 042 043 044 045 046 047 May 10 00:40:03.649310 kernel: pcpu-alloc: [0] 048 049 050 051 052 053 054 055 May 10 00:40:03.649322 kernel: pcpu-alloc: [0] 056 057 058 059 060 061 062 063 May 10 00:40:03.649328 kernel: pcpu-alloc: [0] 064 065 066 067 068 069 070 071 May 10 00:40:03.649333 kernel: pcpu-alloc: [0] 072 073 074 075 076 077 078 079 May 10 00:40:03.649339 kernel: pcpu-alloc: [0] 080 081 082 083 084 085 086 087 May 10 00:40:03.649344 kernel: pcpu-alloc: [0] 088 089 090 091 092 093 094 095 May 10 00:40:03.649365 kernel: pcpu-alloc: [0] 096 097 098 099 100 101 102 103 May 10 00:40:03.649370 kernel: pcpu-alloc: [0] 104 105 106 107 108 109 110 111 May 10 00:40:03.649376 kernel: pcpu-alloc: [0] 112 113 114 115 116 117 118 119 May 10 00:40:03.649381 kernel: pcpu-alloc: [0] 120 121 122 123 124 125 126 127 May 10 00:40:03.649386 kernel: Built 1 zonelists, mobility grouping on. Total pages: 515808 May 10 00:40:03.649392 kernel: Policy zone: DMA32 May 10 00:40:03.649398 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=vmware flatcar.autologin verity.usrhash=39569409b30be1967efab22b453b92a780dcf0fe8e1448a18bf235b5cf33e54a May 10 00:40:03.649405 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. May 10 00:40:03.649410 kernel: printk: log_buf_len individual max cpu contribution: 4096 bytes May 10 00:40:03.649415 kernel: printk: log_buf_len total cpu_extra contributions: 520192 bytes May 10 00:40:03.649421 kernel: printk: log_buf_len min size: 262144 bytes May 10 00:40:03.649426 kernel: printk: log_buf_len: 1048576 bytes May 10 00:40:03.649431 kernel: printk: early log buf free: 239728(91%) May 10 00:40:03.649437 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) May 10 00:40:03.649442 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) May 10 00:40:03.649448 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off May 10 00:40:03.649454 kernel: Memory: 1940392K/2096628K available (12294K kernel code, 2276K rwdata, 13724K rodata, 47456K init, 4124K bss, 155976K reserved, 0K cma-reserved) May 10 00:40:03.649460 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=128, Nodes=1 May 10 00:40:03.649465 kernel: ftrace: allocating 34584 entries in 136 pages May 10 00:40:03.649471 kernel: ftrace: allocated 136 pages with 2 groups May 10 00:40:03.649478 kernel: rcu: Hierarchical RCU implementation. May 10 00:40:03.649483 kernel: rcu: RCU event tracing is enabled. May 10 00:40:03.649489 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=128. May 10 00:40:03.649495 kernel: Rude variant of Tasks RCU enabled. May 10 00:40:03.649500 kernel: Tracing variant of Tasks RCU enabled. May 10 00:40:03.649506 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. May 10 00:40:03.649512 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=128 May 10 00:40:03.649517 kernel: NR_IRQS: 33024, nr_irqs: 1448, preallocated irqs: 16 May 10 00:40:03.649523 kernel: random: crng init done May 10 00:40:03.649528 kernel: Console: colour VGA+ 80x25 May 10 00:40:03.649534 kernel: printk: console [tty0] enabled May 10 00:40:03.649540 kernel: printk: console [ttyS0] enabled May 10 00:40:03.649545 kernel: ACPI: Core revision 20210730 May 10 00:40:03.649551 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 133484882848 ns May 10 00:40:03.649556 kernel: APIC: Switch to symmetric I/O mode setup May 10 00:40:03.649562 kernel: x2apic enabled May 10 00:40:03.649567 kernel: Switched APIC routing to physical x2apic. May 10 00:40:03.649573 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 May 10 00:40:03.649578 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x311fd3cd494, max_idle_ns: 440795223879 ns May 10 00:40:03.649585 kernel: Calibrating delay loop (skipped) preset value.. 6816.00 BogoMIPS (lpj=3408000) May 10 00:40:03.649590 kernel: Disabled fast string operations May 10 00:40:03.649596 kernel: Last level iTLB entries: 4KB 64, 2MB 8, 4MB 8 May 10 00:40:03.649601 kernel: Last level dTLB entries: 4KB 64, 2MB 32, 4MB 32, 1GB 4 May 10 00:40:03.649621 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization May 10 00:40:03.649626 kernel: Spectre V2 : WARNING: Unprivileged eBPF is enabled with eIBRS on, data leaks possible via Spectre v2 BHB attacks! May 10 00:40:03.649632 kernel: Spectre V2 : Spectre BHI mitigation: SW BHB clearing on vm exit May 10 00:40:03.649651 kernel: Spectre V2 : Spectre BHI mitigation: SW BHB clearing on syscall May 10 00:40:03.649658 kernel: Spectre V2 : Mitigation: Enhanced / Automatic IBRS May 10 00:40:03.649670 kernel: Spectre V2 : Spectre v2 / PBRSB-eIBRS: Retire a single CALL on VMEXIT May 10 00:40:03.649676 kernel: RETBleed: Mitigation: Enhanced IBRS May 10 00:40:03.649682 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier May 10 00:40:03.649687 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl and seccomp May 10 00:40:03.649692 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode May 10 00:40:03.649698 kernel: SRBDS: Unknown: Dependent on hypervisor status May 10 00:40:03.649703 kernel: GDS: Unknown: Dependent on hypervisor status May 10 00:40:03.649708 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' May 10 00:40:03.649715 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' May 10 00:40:03.649721 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' May 10 00:40:03.649726 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 May 10 00:40:03.649731 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. May 10 00:40:03.649737 kernel: Freeing SMP alternatives memory: 32K May 10 00:40:03.649742 kernel: pid_max: default: 131072 minimum: 1024 May 10 00:40:03.649747 kernel: LSM: Security Framework initializing May 10 00:40:03.649753 kernel: SELinux: Initializing. May 10 00:40:03.649758 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) May 10 00:40:03.649765 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) May 10 00:40:03.649770 kernel: smpboot: CPU0: Intel(R) Xeon(R) E-2278G CPU @ 3.40GHz (family: 0x6, model: 0x9e, stepping: 0xd) May 10 00:40:03.649775 kernel: Performance Events: Skylake events, core PMU driver. May 10 00:40:03.649781 kernel: core: CPUID marked event: 'cpu cycles' unavailable May 10 00:40:03.649789 kernel: core: CPUID marked event: 'instructions' unavailable May 10 00:40:03.649795 kernel: core: CPUID marked event: 'bus cycles' unavailable May 10 00:40:03.649800 kernel: core: CPUID marked event: 'cache references' unavailable May 10 00:40:03.649805 kernel: core: CPUID marked event: 'cache misses' unavailable May 10 00:40:03.649810 kernel: core: CPUID marked event: 'branch instructions' unavailable May 10 00:40:03.649816 kernel: core: CPUID marked event: 'branch misses' unavailable May 10 00:40:03.649822 kernel: ... version: 1 May 10 00:40:03.649827 kernel: ... bit width: 48 May 10 00:40:03.649832 kernel: ... generic registers: 4 May 10 00:40:03.649838 kernel: ... value mask: 0000ffffffffffff May 10 00:40:03.649843 kernel: ... max period: 000000007fffffff May 10 00:40:03.649848 kernel: ... fixed-purpose events: 0 May 10 00:40:03.649854 kernel: ... event mask: 000000000000000f May 10 00:40:03.649859 kernel: signal: max sigframe size: 1776 May 10 00:40:03.649865 kernel: rcu: Hierarchical SRCU implementation. May 10 00:40:03.649870 kernel: NMI watchdog: Perf NMI watchdog permanently disabled May 10 00:40:03.649875 kernel: smp: Bringing up secondary CPUs ... May 10 00:40:03.649881 kernel: x86: Booting SMP configuration: May 10 00:40:03.649886 kernel: .... node #0, CPUs: #1 May 10 00:40:03.649892 kernel: Disabled fast string operations May 10 00:40:03.649897 kernel: smpboot: CPU 1 Converting physical 2 to logical package 1 May 10 00:40:03.649902 kernel: smpboot: CPU 1 Converting physical 0 to logical die 1 May 10 00:40:03.649917 kernel: smp: Brought up 1 node, 2 CPUs May 10 00:40:03.649924 kernel: smpboot: Max logical packages: 128 May 10 00:40:03.649930 kernel: smpboot: Total of 2 processors activated (13632.00 BogoMIPS) May 10 00:40:03.649935 kernel: devtmpfs: initialized May 10 00:40:03.649941 kernel: x86/mm: Memory block size: 128MB May 10 00:40:03.649946 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x7feff000-0x7fefffff] (4096 bytes) May 10 00:40:03.649952 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns May 10 00:40:03.649958 kernel: futex hash table entries: 32768 (order: 9, 2097152 bytes, linear) May 10 00:40:03.649963 kernel: pinctrl core: initialized pinctrl subsystem May 10 00:40:03.649968 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family May 10 00:40:03.649974 kernel: audit: initializing netlink subsys (disabled) May 10 00:40:03.649980 kernel: audit: type=2000 audit(1746837602.057:1): state=initialized audit_enabled=0 res=1 May 10 00:40:03.649985 kernel: thermal_sys: Registered thermal governor 'step_wise' May 10 00:40:03.649990 kernel: thermal_sys: Registered thermal governor 'user_space' May 10 00:40:03.649996 kernel: cpuidle: using governor menu May 10 00:40:03.650001 kernel: Simple Boot Flag at 0x36 set to 0x80 May 10 00:40:03.650007 kernel: ACPI: bus type PCI registered May 10 00:40:03.650012 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 May 10 00:40:03.650017 kernel: dca service started, version 1.12.1 May 10 00:40:03.650023 kernel: PCI: MMCONFIG for domain 0000 [bus 00-7f] at [mem 0xf0000000-0xf7ffffff] (base 0xf0000000) May 10 00:40:03.650029 kernel: PCI: MMCONFIG at [mem 0xf0000000-0xf7ffffff] reserved in E820 May 10 00:40:03.650034 kernel: PCI: Using configuration type 1 for base access May 10 00:40:03.650040 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. May 10 00:40:03.650045 kernel: HugeTLB registered 1.00 GiB page size, pre-allocated 0 pages May 10 00:40:03.650050 kernel: HugeTLB registered 2.00 MiB page size, pre-allocated 0 pages May 10 00:40:03.650056 kernel: ACPI: Added _OSI(Module Device) May 10 00:40:03.650061 kernel: ACPI: Added _OSI(Processor Device) May 10 00:40:03.650066 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) May 10 00:40:03.650072 kernel: ACPI: Added _OSI(Processor Aggregator Device) May 10 00:40:03.650078 kernel: ACPI: Added _OSI(Linux-Dell-Video) May 10 00:40:03.650084 kernel: ACPI: Added _OSI(Linux-Lenovo-NV-HDMI-Audio) May 10 00:40:03.650089 kernel: ACPI: Added _OSI(Linux-HPI-Hybrid-Graphics) May 10 00:40:03.650094 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded May 10 00:40:03.650100 kernel: ACPI: [Firmware Bug]: BIOS _OSI(Linux) query ignored May 10 00:40:03.650105 kernel: ACPI: Interpreter enabled May 10 00:40:03.650110 kernel: ACPI: PM: (supports S0 S1 S5) May 10 00:40:03.650115 kernel: ACPI: Using IOAPIC for interrupt routing May 10 00:40:03.650121 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug May 10 00:40:03.650127 kernel: ACPI: Enabled 4 GPEs in block 00 to 0F May 10 00:40:03.650132 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-7f]) May 10 00:40:03.650203 kernel: acpi PNP0A03:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] May 10 00:40:03.650250 kernel: acpi PNP0A03:00: _OSC: platform does not support [AER LTR] May 10 00:40:03.650299 kernel: acpi PNP0A03:00: _OSC: OS now controls [PCIeHotplug PME PCIeCapability] May 10 00:40:03.650308 kernel: PCI host bridge to bus 0000:00 May 10 00:40:03.650354 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] May 10 00:40:03.650397 kernel: pci_bus 0000:00: root bus resource [mem 0x000cc000-0x000dbfff window] May 10 00:40:03.650435 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] May 10 00:40:03.650492 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] May 10 00:40:03.650531 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xfeff window] May 10 00:40:03.650570 kernel: pci_bus 0000:00: root bus resource [bus 00-7f] May 10 00:40:03.650637 kernel: pci 0000:00:00.0: [8086:7190] type 00 class 0x060000 May 10 00:40:03.650693 kernel: pci 0000:00:01.0: [8086:7191] type 01 class 0x060400 May 10 00:40:03.650742 kernel: pci 0000:00:07.0: [8086:7110] type 00 class 0x060100 May 10 00:40:03.650790 kernel: pci 0000:00:07.1: [8086:7111] type 00 class 0x01018a May 10 00:40:03.650834 kernel: pci 0000:00:07.1: reg 0x20: [io 0x1060-0x106f] May 10 00:40:03.650878 kernel: pci 0000:00:07.1: legacy IDE quirk: reg 0x10: [io 0x01f0-0x01f7] May 10 00:40:03.656970 kernel: pci 0000:00:07.1: legacy IDE quirk: reg 0x14: [io 0x03f6] May 10 00:40:03.657025 kernel: pci 0000:00:07.1: legacy IDE quirk: reg 0x18: [io 0x0170-0x0177] May 10 00:40:03.657075 kernel: pci 0000:00:07.1: legacy IDE quirk: reg 0x1c: [io 0x0376] May 10 00:40:03.657125 kernel: pci 0000:00:07.3: [8086:7113] type 00 class 0x068000 May 10 00:40:03.657171 kernel: pci 0000:00:07.3: quirk: [io 0x1000-0x103f] claimed by PIIX4 ACPI May 10 00:40:03.657214 kernel: pci 0000:00:07.3: quirk: [io 0x1040-0x104f] claimed by PIIX4 SMB May 10 00:40:03.657262 kernel: pci 0000:00:07.7: [15ad:0740] type 00 class 0x088000 May 10 00:40:03.657307 kernel: pci 0000:00:07.7: reg 0x10: [io 0x1080-0x10bf] May 10 00:40:03.657354 kernel: pci 0000:00:07.7: reg 0x14: [mem 0xfebfe000-0xfebfffff 64bit] May 10 00:40:03.657401 kernel: pci 0000:00:0f.0: [15ad:0405] type 00 class 0x030000 May 10 00:40:03.657445 kernel: pci 0000:00:0f.0: reg 0x10: [io 0x1070-0x107f] May 10 00:40:03.657489 kernel: pci 0000:00:0f.0: reg 0x14: [mem 0xe8000000-0xefffffff pref] May 10 00:40:03.657532 kernel: pci 0000:00:0f.0: reg 0x18: [mem 0xfe000000-0xfe7fffff] May 10 00:40:03.657576 kernel: pci 0000:00:0f.0: reg 0x30: [mem 0x00000000-0x00007fff pref] May 10 00:40:03.657619 kernel: pci 0000:00:0f.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] May 10 00:40:03.657667 kernel: pci 0000:00:11.0: [15ad:0790] type 01 class 0x060401 May 10 00:40:03.657716 kernel: pci 0000:00:15.0: [15ad:07a0] type 01 class 0x060400 May 10 00:40:03.657761 kernel: pci 0000:00:15.0: PME# supported from D0 D3hot D3cold May 10 00:40:03.657809 kernel: pci 0000:00:15.1: [15ad:07a0] type 01 class 0x060400 May 10 00:40:03.657856 kernel: pci 0000:00:15.1: PME# supported from D0 D3hot D3cold May 10 00:40:03.657911 kernel: pci 0000:00:15.2: [15ad:07a0] type 01 class 0x060400 May 10 00:40:03.657965 kernel: pci 0000:00:15.2: PME# supported from D0 D3hot D3cold May 10 00:40:03.658015 kernel: pci 0000:00:15.3: [15ad:07a0] type 01 class 0x060400 May 10 00:40:03.658059 kernel: pci 0000:00:15.3: PME# supported from D0 D3hot D3cold May 10 00:40:03.658106 kernel: pci 0000:00:15.4: [15ad:07a0] type 01 class 0x060400 May 10 00:40:03.658150 kernel: pci 0000:00:15.4: PME# supported from D0 D3hot D3cold May 10 00:40:03.658197 kernel: pci 0000:00:15.5: [15ad:07a0] type 01 class 0x060400 May 10 00:40:03.658244 kernel: pci 0000:00:15.5: PME# supported from D0 D3hot D3cold May 10 00:40:03.658299 kernel: pci 0000:00:15.6: [15ad:07a0] type 01 class 0x060400 May 10 00:40:03.658347 kernel: pci 0000:00:15.6: PME# supported from D0 D3hot D3cold May 10 00:40:03.658395 kernel: pci 0000:00:15.7: [15ad:07a0] type 01 class 0x060400 May 10 00:40:03.658439 kernel: pci 0000:00:15.7: PME# supported from D0 D3hot D3cold May 10 00:40:03.658486 kernel: pci 0000:00:16.0: [15ad:07a0] type 01 class 0x060400 May 10 00:40:03.658533 kernel: pci 0000:00:16.0: PME# supported from D0 D3hot D3cold May 10 00:40:03.658581 kernel: pci 0000:00:16.1: [15ad:07a0] type 01 class 0x060400 May 10 00:40:03.658625 kernel: pci 0000:00:16.1: PME# supported from D0 D3hot D3cold May 10 00:40:03.658671 kernel: pci 0000:00:16.2: [15ad:07a0] type 01 class 0x060400 May 10 00:40:03.658714 kernel: pci 0000:00:16.2: PME# supported from D0 D3hot D3cold May 10 00:40:03.658762 kernel: pci 0000:00:16.3: [15ad:07a0] type 01 class 0x060400 May 10 00:40:03.658809 kernel: pci 0000:00:16.3: PME# supported from D0 D3hot D3cold May 10 00:40:03.658856 kernel: pci 0000:00:16.4: [15ad:07a0] type 01 class 0x060400 May 10 00:40:03.658901 kernel: pci 0000:00:16.4: PME# supported from D0 D3hot D3cold May 10 00:40:03.658963 kernel: pci 0000:00:16.5: [15ad:07a0] type 01 class 0x060400 May 10 00:40:03.659006 kernel: pci 0000:00:16.5: PME# supported from D0 D3hot D3cold May 10 00:40:03.659055 kernel: pci 0000:00:16.6: [15ad:07a0] type 01 class 0x060400 May 10 00:40:03.659101 kernel: pci 0000:00:16.6: PME# supported from D0 D3hot D3cold May 10 00:40:03.659147 kernel: pci 0000:00:16.7: [15ad:07a0] type 01 class 0x060400 May 10 00:40:03.659192 kernel: pci 0000:00:16.7: PME# supported from D0 D3hot D3cold May 10 00:40:03.659239 kernel: pci 0000:00:17.0: [15ad:07a0] type 01 class 0x060400 May 10 00:40:03.659283 kernel: pci 0000:00:17.0: PME# supported from D0 D3hot D3cold May 10 00:40:03.659330 kernel: pci 0000:00:17.1: [15ad:07a0] type 01 class 0x060400 May 10 00:40:03.659376 kernel: pci 0000:00:17.1: PME# supported from D0 D3hot D3cold May 10 00:40:03.659423 kernel: pci 0000:00:17.2: [15ad:07a0] type 01 class 0x060400 May 10 00:40:03.659468 kernel: pci 0000:00:17.2: PME# supported from D0 D3hot D3cold May 10 00:40:03.659516 kernel: pci 0000:00:17.3: [15ad:07a0] type 01 class 0x060400 May 10 00:40:03.659560 kernel: pci 0000:00:17.3: PME# supported from D0 D3hot D3cold May 10 00:40:03.659607 kernel: pci 0000:00:17.4: [15ad:07a0] type 01 class 0x060400 May 10 00:40:03.659654 kernel: pci 0000:00:17.4: PME# supported from D0 D3hot D3cold May 10 00:40:03.659700 kernel: pci 0000:00:17.5: [15ad:07a0] type 01 class 0x060400 May 10 00:40:03.659745 kernel: pci 0000:00:17.5: PME# supported from D0 D3hot D3cold May 10 00:40:03.659792 kernel: pci 0000:00:17.6: [15ad:07a0] type 01 class 0x060400 May 10 00:40:03.659836 kernel: pci 0000:00:17.6: PME# supported from D0 D3hot D3cold May 10 00:40:03.659885 kernel: pci 0000:00:17.7: [15ad:07a0] type 01 class 0x060400 May 10 00:40:03.659937 kernel: pci 0000:00:17.7: PME# supported from D0 D3hot D3cold May 10 00:40:03.659988 kernel: pci 0000:00:18.0: [15ad:07a0] type 01 class 0x060400 May 10 00:40:03.660032 kernel: pci 0000:00:18.0: PME# supported from D0 D3hot D3cold May 10 00:40:03.660080 kernel: pci 0000:00:18.1: [15ad:07a0] type 01 class 0x060400 May 10 00:40:03.660124 kernel: pci 0000:00:18.1: PME# supported from D0 D3hot D3cold May 10 00:40:03.660172 kernel: pci 0000:00:18.2: [15ad:07a0] type 01 class 0x060400 May 10 00:40:03.660216 kernel: pci 0000:00:18.2: PME# supported from D0 D3hot D3cold May 10 00:40:03.660264 kernel: pci 0000:00:18.3: [15ad:07a0] type 01 class 0x060400 May 10 00:40:03.660309 kernel: pci 0000:00:18.3: PME# supported from D0 D3hot D3cold May 10 00:40:03.660358 kernel: pci 0000:00:18.4: [15ad:07a0] type 01 class 0x060400 May 10 00:40:03.660402 kernel: pci 0000:00:18.4: PME# supported from D0 D3hot D3cold May 10 00:40:03.660450 kernel: pci 0000:00:18.5: [15ad:07a0] type 01 class 0x060400 May 10 00:40:03.660494 kernel: pci 0000:00:18.5: PME# supported from D0 D3hot D3cold May 10 00:40:03.660543 kernel: pci 0000:00:18.6: [15ad:07a0] type 01 class 0x060400 May 10 00:40:03.660587 kernel: pci 0000:00:18.6: PME# supported from D0 D3hot D3cold May 10 00:40:03.660635 kernel: pci 0000:00:18.7: [15ad:07a0] type 01 class 0x060400 May 10 00:40:03.660679 kernel: pci 0000:00:18.7: PME# supported from D0 D3hot D3cold May 10 00:40:03.660725 kernel: pci_bus 0000:01: extended config space not accessible May 10 00:40:03.660771 kernel: pci 0000:00:01.0: PCI bridge to [bus 01] May 10 00:40:03.660817 kernel: pci_bus 0000:02: extended config space not accessible May 10 00:40:03.660826 kernel: acpiphp: Slot [32] registered May 10 00:40:03.660832 kernel: acpiphp: Slot [33] registered May 10 00:40:03.660837 kernel: acpiphp: Slot [34] registered May 10 00:40:03.660843 kernel: acpiphp: Slot [35] registered May 10 00:40:03.660848 kernel: acpiphp: Slot [36] registered May 10 00:40:03.660853 kernel: acpiphp: Slot [37] registered May 10 00:40:03.660858 kernel: acpiphp: Slot [38] registered May 10 00:40:03.660864 kernel: acpiphp: Slot [39] registered May 10 00:40:03.660870 kernel: acpiphp: Slot [40] registered May 10 00:40:03.660875 kernel: acpiphp: Slot [41] registered May 10 00:40:03.660881 kernel: acpiphp: Slot [42] registered May 10 00:40:03.660886 kernel: acpiphp: Slot [43] registered May 10 00:40:03.660891 kernel: acpiphp: Slot [44] registered May 10 00:40:03.660897 kernel: acpiphp: Slot [45] registered May 10 00:40:03.660902 kernel: acpiphp: Slot [46] registered May 10 00:40:03.660914 kernel: acpiphp: Slot [47] registered May 10 00:40:03.660920 kernel: acpiphp: Slot [48] registered May 10 00:40:03.660925 kernel: acpiphp: Slot [49] registered May 10 00:40:03.660932 kernel: acpiphp: Slot [50] registered May 10 00:40:03.660937 kernel: acpiphp: Slot [51] registered May 10 00:40:03.660942 kernel: acpiphp: Slot [52] registered May 10 00:40:03.660948 kernel: acpiphp: Slot [53] registered May 10 00:40:03.660953 kernel: acpiphp: Slot [54] registered May 10 00:40:03.660959 kernel: acpiphp: Slot [55] registered May 10 00:40:03.660964 kernel: acpiphp: Slot [56] registered May 10 00:40:03.660969 kernel: acpiphp: Slot [57] registered May 10 00:40:03.660975 kernel: acpiphp: Slot [58] registered May 10 00:40:03.660981 kernel: acpiphp: Slot [59] registered May 10 00:40:03.660986 kernel: acpiphp: Slot [60] registered May 10 00:40:03.660992 kernel: acpiphp: Slot [61] registered May 10 00:40:03.660997 kernel: acpiphp: Slot [62] registered May 10 00:40:03.661002 kernel: acpiphp: Slot [63] registered May 10 00:40:03.661048 kernel: pci 0000:00:11.0: PCI bridge to [bus 02] (subtractive decode) May 10 00:40:03.661092 kernel: pci 0000:00:11.0: bridge window [io 0x2000-0x3fff] May 10 00:40:03.661135 kernel: pci 0000:00:11.0: bridge window [mem 0xfd600000-0xfdffffff] May 10 00:40:03.661178 kernel: pci 0000:00:11.0: bridge window [mem 0xe7b00000-0xe7ffffff 64bit pref] May 10 00:40:03.661223 kernel: pci 0000:00:11.0: bridge window [mem 0x000a0000-0x000bffff window] (subtractive decode) May 10 00:40:03.661267 kernel: pci 0000:00:11.0: bridge window [mem 0x000cc000-0x000dbfff window] (subtractive decode) May 10 00:40:03.661314 kernel: pci 0000:00:11.0: bridge window [mem 0xc0000000-0xfebfffff window] (subtractive decode) May 10 00:40:03.661358 kernel: pci 0000:00:11.0: bridge window [io 0x0000-0x0cf7 window] (subtractive decode) May 10 00:40:03.661401 kernel: pci 0000:00:11.0: bridge window [io 0x0d00-0xfeff window] (subtractive decode) May 10 00:40:03.661451 kernel: pci 0000:03:00.0: [15ad:07c0] type 00 class 0x010700 May 10 00:40:03.661497 kernel: pci 0000:03:00.0: reg 0x10: [io 0x4000-0x4007] May 10 00:40:03.661544 kernel: pci 0000:03:00.0: reg 0x14: [mem 0xfd5f8000-0xfd5fffff 64bit] May 10 00:40:03.661589 kernel: pci 0000:03:00.0: reg 0x30: [mem 0x00000000-0x0000ffff pref] May 10 00:40:03.661633 kernel: pci 0000:03:00.0: PME# supported from D0 D3hot D3cold May 10 00:40:03.661678 kernel: pci 0000:03:00.0: disabling ASPM on pre-1.1 PCIe device. You can enable it with 'pcie_aspm=force' May 10 00:40:03.661723 kernel: pci 0000:00:15.0: PCI bridge to [bus 03] May 10 00:40:03.661767 kernel: pci 0000:00:15.0: bridge window [io 0x4000-0x4fff] May 10 00:40:03.661811 kernel: pci 0000:00:15.0: bridge window [mem 0xfd500000-0xfd5fffff] May 10 00:40:03.661858 kernel: pci 0000:00:15.1: PCI bridge to [bus 04] May 10 00:40:03.661901 kernel: pci 0000:00:15.1: bridge window [io 0x8000-0x8fff] May 10 00:40:03.669986 kernel: pci 0000:00:15.1: bridge window [mem 0xfd100000-0xfd1fffff] May 10 00:40:03.670037 kernel: pci 0000:00:15.1: bridge window [mem 0xe7800000-0xe78fffff 64bit pref] May 10 00:40:03.670083 kernel: pci 0000:00:15.2: PCI bridge to [bus 05] May 10 00:40:03.670147 kernel: pci 0000:00:15.2: bridge window [io 0xc000-0xcfff] May 10 00:40:03.670207 kernel: pci 0000:00:15.2: bridge window [mem 0xfcd00000-0xfcdfffff] May 10 00:40:03.670250 kernel: pci 0000:00:15.2: bridge window [mem 0xe7400000-0xe74fffff 64bit pref] May 10 00:40:03.670301 kernel: pci 0000:00:15.3: PCI bridge to [bus 06] May 10 00:40:03.670347 kernel: pci 0000:00:15.3: bridge window [mem 0xfc900000-0xfc9fffff] May 10 00:40:03.670390 kernel: pci 0000:00:15.3: bridge window [mem 0xe7000000-0xe70fffff 64bit pref] May 10 00:40:03.670434 kernel: pci 0000:00:15.4: PCI bridge to [bus 07] May 10 00:40:03.670477 kernel: pci 0000:00:15.4: bridge window [mem 0xfc500000-0xfc5fffff] May 10 00:40:03.670520 kernel: pci 0000:00:15.4: bridge window [mem 0xe6c00000-0xe6cfffff 64bit pref] May 10 00:40:03.670565 kernel: pci 0000:00:15.5: PCI bridge to [bus 08] May 10 00:40:03.670609 kernel: pci 0000:00:15.5: bridge window [mem 0xfc100000-0xfc1fffff] May 10 00:40:03.670651 kernel: pci 0000:00:15.5: bridge window [mem 0xe6800000-0xe68fffff 64bit pref] May 10 00:40:03.670695 kernel: pci 0000:00:15.6: PCI bridge to [bus 09] May 10 00:40:03.670738 kernel: pci 0000:00:15.6: bridge window [mem 0xfbd00000-0xfbdfffff] May 10 00:40:03.670782 kernel: pci 0000:00:15.6: bridge window [mem 0xe6400000-0xe64fffff 64bit pref] May 10 00:40:03.670827 kernel: pci 0000:00:15.7: PCI bridge to [bus 0a] May 10 00:40:03.670871 kernel: pci 0000:00:15.7: bridge window [mem 0xfb900000-0xfb9fffff] May 10 00:40:03.670923 kernel: pci 0000:00:15.7: bridge window [mem 0xe6000000-0xe60fffff 64bit pref] May 10 00:40:03.670975 kernel: pci 0000:0b:00.0: [15ad:07b0] type 00 class 0x020000 May 10 00:40:03.671021 kernel: pci 0000:0b:00.0: reg 0x10: [mem 0xfd4fc000-0xfd4fcfff] May 10 00:40:03.671066 kernel: pci 0000:0b:00.0: reg 0x14: [mem 0xfd4fd000-0xfd4fdfff] May 10 00:40:03.671110 kernel: pci 0000:0b:00.0: reg 0x18: [mem 0xfd4fe000-0xfd4fffff] May 10 00:40:03.671155 kernel: pci 0000:0b:00.0: reg 0x1c: [io 0x5000-0x500f] May 10 00:40:03.671203 kernel: pci 0000:0b:00.0: reg 0x30: [mem 0x00000000-0x0000ffff pref] May 10 00:40:03.671249 kernel: pci 0000:0b:00.0: supports D1 D2 May 10 00:40:03.671295 kernel: pci 0000:0b:00.0: PME# supported from D0 D1 D2 D3hot D3cold May 10 00:40:03.671339 kernel: pci 0000:0b:00.0: disabling ASPM on pre-1.1 PCIe device. You can enable it with 'pcie_aspm=force' May 10 00:40:03.671383 kernel: pci 0000:00:16.0: PCI bridge to [bus 0b] May 10 00:40:03.671427 kernel: pci 0000:00:16.0: bridge window [io 0x5000-0x5fff] May 10 00:40:03.671470 kernel: pci 0000:00:16.0: bridge window [mem 0xfd400000-0xfd4fffff] May 10 00:40:03.671518 kernel: pci 0000:00:16.1: PCI bridge to [bus 0c] May 10 00:40:03.671561 kernel: pci 0000:00:16.1: bridge window [io 0x9000-0x9fff] May 10 00:40:03.671624 kernel: pci 0000:00:16.1: bridge window [mem 0xfd000000-0xfd0fffff] May 10 00:40:03.671682 kernel: pci 0000:00:16.1: bridge window [mem 0xe7700000-0xe77fffff 64bit pref] May 10 00:40:03.671726 kernel: pci 0000:00:16.2: PCI bridge to [bus 0d] May 10 00:40:03.671769 kernel: pci 0000:00:16.2: bridge window [io 0xd000-0xdfff] May 10 00:40:03.671812 kernel: pci 0000:00:16.2: bridge window [mem 0xfcc00000-0xfccfffff] May 10 00:40:03.671855 kernel: pci 0000:00:16.2: bridge window [mem 0xe7300000-0xe73fffff 64bit pref] May 10 00:40:03.671901 kernel: pci 0000:00:16.3: PCI bridge to [bus 0e] May 10 00:40:03.674033 kernel: pci 0000:00:16.3: bridge window [mem 0xfc800000-0xfc8fffff] May 10 00:40:03.674083 kernel: pci 0000:00:16.3: bridge window [mem 0xe6f00000-0xe6ffffff 64bit pref] May 10 00:40:03.674131 kernel: pci 0000:00:16.4: PCI bridge to [bus 0f] May 10 00:40:03.674176 kernel: pci 0000:00:16.4: bridge window [mem 0xfc400000-0xfc4fffff] May 10 00:40:03.674220 kernel: pci 0000:00:16.4: bridge window [mem 0xe6b00000-0xe6bfffff 64bit pref] May 10 00:40:03.674265 kernel: pci 0000:00:16.5: PCI bridge to [bus 10] May 10 00:40:03.674338 kernel: pci 0000:00:16.5: bridge window [mem 0xfc000000-0xfc0fffff] May 10 00:40:03.674417 kernel: pci 0000:00:16.5: bridge window [mem 0xe6700000-0xe67fffff 64bit pref] May 10 00:40:03.674462 kernel: pci 0000:00:16.6: PCI bridge to [bus 11] May 10 00:40:03.674505 kernel: pci 0000:00:16.6: bridge window [mem 0xfbc00000-0xfbcfffff] May 10 00:40:03.674548 kernel: pci 0000:00:16.6: bridge window [mem 0xe6300000-0xe63fffff 64bit pref] May 10 00:40:03.674592 kernel: pci 0000:00:16.7: PCI bridge to [bus 12] May 10 00:40:03.674634 kernel: pci 0000:00:16.7: bridge window [mem 0xfb800000-0xfb8fffff] May 10 00:40:03.674677 kernel: pci 0000:00:16.7: bridge window [mem 0xe5f00000-0xe5ffffff 64bit pref] May 10 00:40:03.674721 kernel: pci 0000:00:17.0: PCI bridge to [bus 13] May 10 00:40:03.674767 kernel: pci 0000:00:17.0: bridge window [io 0x6000-0x6fff] May 10 00:40:03.674809 kernel: pci 0000:00:17.0: bridge window [mem 0xfd300000-0xfd3fffff] May 10 00:40:03.674852 kernel: pci 0000:00:17.0: bridge window [mem 0xe7a00000-0xe7afffff 64bit pref] May 10 00:40:03.674896 kernel: pci 0000:00:17.1: PCI bridge to [bus 14] May 10 00:40:03.674947 kernel: pci 0000:00:17.1: bridge window [io 0xa000-0xafff] May 10 00:40:03.674991 kernel: pci 0000:00:17.1: bridge window [mem 0xfcf00000-0xfcffffff] May 10 00:40:03.675035 kernel: pci 0000:00:17.1: bridge window [mem 0xe7600000-0xe76fffff 64bit pref] May 10 00:40:03.675080 kernel: pci 0000:00:17.2: PCI bridge to [bus 15] May 10 00:40:03.675127 kernel: pci 0000:00:17.2: bridge window [io 0xe000-0xefff] May 10 00:40:03.675170 kernel: pci 0000:00:17.2: bridge window [mem 0xfcb00000-0xfcbfffff] May 10 00:40:03.675213 kernel: pci 0000:00:17.2: bridge window [mem 0xe7200000-0xe72fffff 64bit pref] May 10 00:40:03.675276 kernel: pci 0000:00:17.3: PCI bridge to [bus 16] May 10 00:40:03.675335 kernel: pci 0000:00:17.3: bridge window [mem 0xfc700000-0xfc7fffff] May 10 00:40:03.675446 kernel: pci 0000:00:17.3: bridge window [mem 0xe6e00000-0xe6efffff 64bit pref] May 10 00:40:03.675491 kernel: pci 0000:00:17.4: PCI bridge to [bus 17] May 10 00:40:03.675538 kernel: pci 0000:00:17.4: bridge window [mem 0xfc300000-0xfc3fffff] May 10 00:40:03.675581 kernel: pci 0000:00:17.4: bridge window [mem 0xe6a00000-0xe6afffff 64bit pref] May 10 00:40:03.675642 kernel: pci 0000:00:17.5: PCI bridge to [bus 18] May 10 00:40:03.675687 kernel: pci 0000:00:17.5: bridge window [mem 0xfbf00000-0xfbffffff] May 10 00:40:03.675730 kernel: pci 0000:00:17.5: bridge window [mem 0xe6600000-0xe66fffff 64bit pref] May 10 00:40:03.675776 kernel: pci 0000:00:17.6: PCI bridge to [bus 19] May 10 00:40:03.675820 kernel: pci 0000:00:17.6: bridge window [mem 0xfbb00000-0xfbbfffff] May 10 00:40:03.675863 kernel: pci 0000:00:17.6: bridge window [mem 0xe6200000-0xe62fffff 64bit pref] May 10 00:40:03.675936 kernel: pci 0000:00:17.7: PCI bridge to [bus 1a] May 10 00:40:03.675983 kernel: pci 0000:00:17.7: bridge window [mem 0xfb700000-0xfb7fffff] May 10 00:40:03.676028 kernel: pci 0000:00:17.7: bridge window [mem 0xe5e00000-0xe5efffff 64bit pref] May 10 00:40:03.676074 kernel: pci 0000:00:18.0: PCI bridge to [bus 1b] May 10 00:40:03.676119 kernel: pci 0000:00:18.0: bridge window [io 0x7000-0x7fff] May 10 00:40:03.676164 kernel: pci 0000:00:18.0: bridge window [mem 0xfd200000-0xfd2fffff] May 10 00:40:03.676208 kernel: pci 0000:00:18.0: bridge window [mem 0xe7900000-0xe79fffff 64bit pref] May 10 00:40:03.676255 kernel: pci 0000:00:18.1: PCI bridge to [bus 1c] May 10 00:40:03.676303 kernel: pci 0000:00:18.1: bridge window [io 0xb000-0xbfff] May 10 00:40:03.676348 kernel: pci 0000:00:18.1: bridge window [mem 0xfce00000-0xfcefffff] May 10 00:40:03.676393 kernel: pci 0000:00:18.1: bridge window [mem 0xe7500000-0xe75fffff 64bit pref] May 10 00:40:03.676440 kernel: pci 0000:00:18.2: PCI bridge to [bus 1d] May 10 00:40:03.676485 kernel: pci 0000:00:18.2: bridge window [mem 0xfca00000-0xfcafffff] May 10 00:40:03.676530 kernel: pci 0000:00:18.2: bridge window [mem 0xe7100000-0xe71fffff 64bit pref] May 10 00:40:03.676577 kernel: pci 0000:00:18.3: PCI bridge to [bus 1e] May 10 00:40:03.676622 kernel: pci 0000:00:18.3: bridge window [mem 0xfc600000-0xfc6fffff] May 10 00:40:03.687825 kernel: pci 0000:00:18.3: bridge window [mem 0xe6d00000-0xe6dfffff 64bit pref] May 10 00:40:03.687882 kernel: pci 0000:00:18.4: PCI bridge to [bus 1f] May 10 00:40:03.687941 kernel: pci 0000:00:18.4: bridge window [mem 0xfc200000-0xfc2fffff] May 10 00:40:03.687988 kernel: pci 0000:00:18.4: bridge window [mem 0xe6900000-0xe69fffff 64bit pref] May 10 00:40:03.688035 kernel: pci 0000:00:18.5: PCI bridge to [bus 20] May 10 00:40:03.688082 kernel: pci 0000:00:18.5: bridge window [mem 0xfbe00000-0xfbefffff] May 10 00:40:03.688128 kernel: pci 0000:00:18.5: bridge window [mem 0xe6500000-0xe65fffff 64bit pref] May 10 00:40:03.688176 kernel: pci 0000:00:18.6: PCI bridge to [bus 21] May 10 00:40:03.688226 kernel: pci 0000:00:18.6: bridge window [mem 0xfba00000-0xfbafffff] May 10 00:40:03.688272 kernel: pci 0000:00:18.6: bridge window [mem 0xe6100000-0xe61fffff 64bit pref] May 10 00:40:03.688320 kernel: pci 0000:00:18.7: PCI bridge to [bus 22] May 10 00:40:03.688365 kernel: pci 0000:00:18.7: bridge window [mem 0xfb600000-0xfb6fffff] May 10 00:40:03.688411 kernel: pci 0000:00:18.7: bridge window [mem 0xe5d00000-0xe5dfffff 64bit pref] May 10 00:40:03.688420 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 9 May 10 00:40:03.688426 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 0 May 10 00:40:03.688432 kernel: ACPI: PCI: Interrupt link LNKB disabled May 10 00:40:03.688440 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 May 10 00:40:03.688446 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 10 May 10 00:40:03.688451 kernel: iommu: Default domain type: Translated May 10 00:40:03.688458 kernel: iommu: DMA domain TLB invalidation policy: lazy mode May 10 00:40:03.688819 kernel: pci 0000:00:0f.0: vgaarb: setting as boot VGA device May 10 00:40:03.688871 kernel: pci 0000:00:0f.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none May 10 00:40:03.688991 kernel: pci 0000:00:0f.0: vgaarb: bridge control possible May 10 00:40:03.689001 kernel: vgaarb: loaded May 10 00:40:03.689007 kernel: pps_core: LinuxPPS API ver. 1 registered May 10 00:40:03.689015 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti May 10 00:40:03.689021 kernel: PTP clock support registered May 10 00:40:03.689027 kernel: PCI: Using ACPI for IRQ routing May 10 00:40:03.689033 kernel: PCI: pci_cache_line_size set to 64 bytes May 10 00:40:03.689039 kernel: e820: reserve RAM buffer [mem 0x0009ec00-0x0009ffff] May 10 00:40:03.689044 kernel: e820: reserve RAM buffer [mem 0x7fee0000-0x7fffffff] May 10 00:40:03.689050 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 May 10 00:40:03.689056 kernel: hpet0: 16 comparators, 64-bit 14.318180 MHz counter May 10 00:40:03.689062 kernel: clocksource: Switched to clocksource tsc-early May 10 00:40:03.689069 kernel: VFS: Disk quotas dquot_6.6.0 May 10 00:40:03.689075 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) May 10 00:40:03.689080 kernel: pnp: PnP ACPI init May 10 00:40:03.689132 kernel: system 00:00: [io 0x1000-0x103f] has been reserved May 10 00:40:03.689176 kernel: system 00:00: [io 0x1040-0x104f] has been reserved May 10 00:40:03.689218 kernel: system 00:00: [io 0x0cf0-0x0cf1] has been reserved May 10 00:40:03.689263 kernel: system 00:04: [mem 0xfed00000-0xfed003ff] has been reserved May 10 00:40:03.689310 kernel: pnp 00:06: [dma 2] May 10 00:40:03.689356 kernel: system 00:07: [io 0xfce0-0xfcff] has been reserved May 10 00:40:03.689398 kernel: system 00:07: [mem 0xf0000000-0xf7ffffff] has been reserved May 10 00:40:03.689440 kernel: system 00:07: [mem 0xfe800000-0xfe9fffff] has been reserved May 10 00:40:03.689448 kernel: pnp: PnP ACPI: found 8 devices May 10 00:40:03.689454 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns May 10 00:40:03.689472 kernel: NET: Registered PF_INET protocol family May 10 00:40:03.689478 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) May 10 00:40:03.689490 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) May 10 00:40:03.689496 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) May 10 00:40:03.689502 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) May 10 00:40:03.689511 kernel: TCP bind hash table entries: 16384 (order: 6, 262144 bytes, linear) May 10 00:40:03.689518 kernel: TCP: Hash tables configured (established 16384 bind 16384) May 10 00:40:03.689523 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) May 10 00:40:03.689529 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) May 10 00:40:03.689535 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family May 10 00:40:03.689542 kernel: NET: Registered PF_XDP protocol family May 10 00:40:03.689598 kernel: pci 0000:00:15.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 03] add_size 200000 add_align 100000 May 10 00:40:03.689648 kernel: pci 0000:00:15.3: bridge window [io 0x1000-0x0fff] to [bus 06] add_size 1000 May 10 00:40:03.689696 kernel: pci 0000:00:15.4: bridge window [io 0x1000-0x0fff] to [bus 07] add_size 1000 May 10 00:40:03.689745 kernel: pci 0000:00:15.5: bridge window [io 0x1000-0x0fff] to [bus 08] add_size 1000 May 10 00:40:03.689793 kernel: pci 0000:00:15.6: bridge window [io 0x1000-0x0fff] to [bus 09] add_size 1000 May 10 00:40:03.689841 kernel: pci 0000:00:15.7: bridge window [io 0x1000-0x0fff] to [bus 0a] add_size 1000 May 10 00:40:03.689892 kernel: pci 0000:00:16.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 0b] add_size 200000 add_align 100000 May 10 00:40:03.689956 kernel: pci 0000:00:16.3: bridge window [io 0x1000-0x0fff] to [bus 0e] add_size 1000 May 10 00:40:03.690006 kernel: pci 0000:00:16.4: bridge window [io 0x1000-0x0fff] to [bus 0f] add_size 1000 May 10 00:40:03.690054 kernel: pci 0000:00:16.5: bridge window [io 0x1000-0x0fff] to [bus 10] add_size 1000 May 10 00:40:03.690101 kernel: pci 0000:00:16.6: bridge window [io 0x1000-0x0fff] to [bus 11] add_size 1000 May 10 00:40:03.690152 kernel: pci 0000:00:16.7: bridge window [io 0x1000-0x0fff] to [bus 12] add_size 1000 May 10 00:40:03.690199 kernel: pci 0000:00:17.3: bridge window [io 0x1000-0x0fff] to [bus 16] add_size 1000 May 10 00:40:03.690247 kernel: pci 0000:00:17.4: bridge window [io 0x1000-0x0fff] to [bus 17] add_size 1000 May 10 00:40:03.690304 kernel: pci 0000:00:17.5: bridge window [io 0x1000-0x0fff] to [bus 18] add_size 1000 May 10 00:40:03.690359 kernel: pci 0000:00:17.6: bridge window [io 0x1000-0x0fff] to [bus 19] add_size 1000 May 10 00:40:03.690408 kernel: pci 0000:00:17.7: bridge window [io 0x1000-0x0fff] to [bus 1a] add_size 1000 May 10 00:40:03.690458 kernel: pci 0000:00:18.2: bridge window [io 0x1000-0x0fff] to [bus 1d] add_size 1000 May 10 00:40:03.690505 kernel: pci 0000:00:18.3: bridge window [io 0x1000-0x0fff] to [bus 1e] add_size 1000 May 10 00:40:03.690551 kernel: pci 0000:00:18.4: bridge window [io 0x1000-0x0fff] to [bus 1f] add_size 1000 May 10 00:40:03.690597 kernel: pci 0000:00:18.5: bridge window [io 0x1000-0x0fff] to [bus 20] add_size 1000 May 10 00:40:03.690643 kernel: pci 0000:00:18.6: bridge window [io 0x1000-0x0fff] to [bus 21] add_size 1000 May 10 00:40:03.690690 kernel: pci 0000:00:18.7: bridge window [io 0x1000-0x0fff] to [bus 22] add_size 1000 May 10 00:40:03.690738 kernel: pci 0000:00:15.0: BAR 15: assigned [mem 0xc0000000-0xc01fffff 64bit pref] May 10 00:40:03.690786 kernel: pci 0000:00:16.0: BAR 15: assigned [mem 0xc0200000-0xc03fffff 64bit pref] May 10 00:40:03.690830 kernel: pci 0000:00:15.3: BAR 13: no space for [io size 0x1000] May 10 00:40:03.690875 kernel: pci 0000:00:15.3: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.690927 kernel: pci 0000:00:15.4: BAR 13: no space for [io size 0x1000] May 10 00:40:03.690973 kernel: pci 0000:00:15.4: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.691018 kernel: pci 0000:00:15.5: BAR 13: no space for [io size 0x1000] May 10 00:40:03.691064 kernel: pci 0000:00:15.5: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.691113 kernel: pci 0000:00:15.6: BAR 13: no space for [io size 0x1000] May 10 00:40:03.691158 kernel: pci 0000:00:15.6: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.691202 kernel: pci 0000:00:15.7: BAR 13: no space for [io size 0x1000] May 10 00:40:03.691247 kernel: pci 0000:00:15.7: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.691292 kernel: pci 0000:00:16.3: BAR 13: no space for [io size 0x1000] May 10 00:40:03.691337 kernel: pci 0000:00:16.3: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.691382 kernel: pci 0000:00:16.4: BAR 13: no space for [io size 0x1000] May 10 00:40:03.691427 kernel: pci 0000:00:16.4: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.691475 kernel: pci 0000:00:16.5: BAR 13: no space for [io size 0x1000] May 10 00:40:03.691520 kernel: pci 0000:00:16.5: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.691565 kernel: pci 0000:00:16.6: BAR 13: no space for [io size 0x1000] May 10 00:40:03.691609 kernel: pci 0000:00:16.6: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.691654 kernel: pci 0000:00:16.7: BAR 13: no space for [io size 0x1000] May 10 00:40:03.691700 kernel: pci 0000:00:16.7: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.691745 kernel: pci 0000:00:17.3: BAR 13: no space for [io size 0x1000] May 10 00:40:03.691791 kernel: pci 0000:00:17.3: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.691837 kernel: pci 0000:00:17.4: BAR 13: no space for [io size 0x1000] May 10 00:40:03.691884 kernel: pci 0000:00:17.4: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.691944 kernel: pci 0000:00:17.5: BAR 13: no space for [io size 0x1000] May 10 00:40:03.691990 kernel: pci 0000:00:17.5: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.692037 kernel: pci 0000:00:17.6: BAR 13: no space for [io size 0x1000] May 10 00:40:03.692082 kernel: pci 0000:00:17.6: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.692128 kernel: pci 0000:00:17.7: BAR 13: no space for [io size 0x1000] May 10 00:40:03.692173 kernel: pci 0000:00:17.7: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.692222 kernel: pci 0000:00:18.2: BAR 13: no space for [io size 0x1000] May 10 00:40:03.692267 kernel: pci 0000:00:18.2: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.692312 kernel: pci 0000:00:18.3: BAR 13: no space for [io size 0x1000] May 10 00:40:03.692357 kernel: pci 0000:00:18.3: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.692403 kernel: pci 0000:00:18.4: BAR 13: no space for [io size 0x1000] May 10 00:40:03.692448 kernel: pci 0000:00:18.4: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.692493 kernel: pci 0000:00:18.5: BAR 13: no space for [io size 0x1000] May 10 00:40:03.692538 kernel: pci 0000:00:18.5: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.692585 kernel: pci 0000:00:18.6: BAR 13: no space for [io size 0x1000] May 10 00:40:03.692630 kernel: pci 0000:00:18.6: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.692676 kernel: pci 0000:00:18.7: BAR 13: no space for [io size 0x1000] May 10 00:40:03.692722 kernel: pci 0000:00:18.7: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.692767 kernel: pci 0000:00:18.7: BAR 13: no space for [io size 0x1000] May 10 00:40:03.692812 kernel: pci 0000:00:18.7: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.692857 kernel: pci 0000:00:18.6: BAR 13: no space for [io size 0x1000] May 10 00:40:03.692902 kernel: pci 0000:00:18.6: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.693345 kernel: pci 0000:00:18.5: BAR 13: no space for [io size 0x1000] May 10 00:40:03.693398 kernel: pci 0000:00:18.5: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.693445 kernel: pci 0000:00:18.4: BAR 13: no space for [io size 0x1000] May 10 00:40:03.693491 kernel: pci 0000:00:18.4: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.693535 kernel: pci 0000:00:18.3: BAR 13: no space for [io size 0x1000] May 10 00:40:03.693580 kernel: pci 0000:00:18.3: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.693857 kernel: pci 0000:00:18.2: BAR 13: no space for [io size 0x1000] May 10 00:40:03.693912 kernel: pci 0000:00:18.2: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.693961 kernel: pci 0000:00:17.7: BAR 13: no space for [io size 0x1000] May 10 00:40:03.694031 kernel: pci 0000:00:17.7: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.694303 kernel: pci 0000:00:17.6: BAR 13: no space for [io size 0x1000] May 10 00:40:03.694376 kernel: pci 0000:00:17.6: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.694424 kernel: pci 0000:00:17.5: BAR 13: no space for [io size 0x1000] May 10 00:40:03.694474 kernel: pci 0000:00:17.5: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.694519 kernel: pci 0000:00:17.4: BAR 13: no space for [io size 0x1000] May 10 00:40:03.694563 kernel: pci 0000:00:17.4: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.694607 kernel: pci 0000:00:17.3: BAR 13: no space for [io size 0x1000] May 10 00:40:03.694651 kernel: pci 0000:00:17.3: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.694695 kernel: pci 0000:00:16.7: BAR 13: no space for [io size 0x1000] May 10 00:40:03.694739 kernel: pci 0000:00:16.7: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.694785 kernel: pci 0000:00:16.6: BAR 13: no space for [io size 0x1000] May 10 00:40:03.694829 kernel: pci 0000:00:16.6: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.694873 kernel: pci 0000:00:16.5: BAR 13: no space for [io size 0x1000] May 10 00:40:03.694934 kernel: pci 0000:00:16.5: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.694983 kernel: pci 0000:00:16.4: BAR 13: no space for [io size 0x1000] May 10 00:40:03.695027 kernel: pci 0000:00:16.4: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.695071 kernel: pci 0000:00:16.3: BAR 13: no space for [io size 0x1000] May 10 00:40:03.695115 kernel: pci 0000:00:16.3: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.695159 kernel: pci 0000:00:15.7: BAR 13: no space for [io size 0x1000] May 10 00:40:03.695206 kernel: pci 0000:00:15.7: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.695250 kernel: pci 0000:00:15.6: BAR 13: no space for [io size 0x1000] May 10 00:40:03.695294 kernel: pci 0000:00:15.6: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.695339 kernel: pci 0000:00:15.5: BAR 13: no space for [io size 0x1000] May 10 00:40:03.695383 kernel: pci 0000:00:15.5: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.695429 kernel: pci 0000:00:15.4: BAR 13: no space for [io size 0x1000] May 10 00:40:03.695473 kernel: pci 0000:00:15.4: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.695517 kernel: pci 0000:00:15.3: BAR 13: no space for [io size 0x1000] May 10 00:40:03.695577 kernel: pci 0000:00:15.3: BAR 13: failed to assign [io size 0x1000] May 10 00:40:03.695621 kernel: pci 0000:00:01.0: PCI bridge to [bus 01] May 10 00:40:03.695668 kernel: pci 0000:00:11.0: PCI bridge to [bus 02] May 10 00:40:03.695712 kernel: pci 0000:00:11.0: bridge window [io 0x2000-0x3fff] May 10 00:40:03.695755 kernel: pci 0000:00:11.0: bridge window [mem 0xfd600000-0xfdffffff] May 10 00:40:03.695798 kernel: pci 0000:00:11.0: bridge window [mem 0xe7b00000-0xe7ffffff 64bit pref] May 10 00:40:03.695846 kernel: pci 0000:03:00.0: BAR 6: assigned [mem 0xfd500000-0xfd50ffff pref] May 10 00:40:03.695890 kernel: pci 0000:00:15.0: PCI bridge to [bus 03] May 10 00:40:03.696163 kernel: pci 0000:00:15.0: bridge window [io 0x4000-0x4fff] May 10 00:40:03.696213 kernel: pci 0000:00:15.0: bridge window [mem 0xfd500000-0xfd5fffff] May 10 00:40:03.696262 kernel: pci 0000:00:15.0: bridge window [mem 0xc0000000-0xc01fffff 64bit pref] May 10 00:40:03.696366 kernel: pci 0000:00:15.1: PCI bridge to [bus 04] May 10 00:40:03.696681 kernel: pci 0000:00:15.1: bridge window [io 0x8000-0x8fff] May 10 00:40:03.696733 kernel: pci 0000:00:15.1: bridge window [mem 0xfd100000-0xfd1fffff] May 10 00:40:03.696779 kernel: pci 0000:00:15.1: bridge window [mem 0xe7800000-0xe78fffff 64bit pref] May 10 00:40:03.696841 kernel: pci 0000:00:15.2: PCI bridge to [bus 05] May 10 00:40:03.697101 kernel: pci 0000:00:15.2: bridge window [io 0xc000-0xcfff] May 10 00:40:03.697154 kernel: pci 0000:00:15.2: bridge window [mem 0xfcd00000-0xfcdfffff] May 10 00:40:03.697201 kernel: pci 0000:00:15.2: bridge window [mem 0xe7400000-0xe74fffff 64bit pref] May 10 00:40:03.697461 kernel: pci 0000:00:15.3: PCI bridge to [bus 06] May 10 00:40:03.697512 kernel: pci 0000:00:15.3: bridge window [mem 0xfc900000-0xfc9fffff] May 10 00:40:03.697576 kernel: pci 0000:00:15.3: bridge window [mem 0xe7000000-0xe70fffff 64bit pref] May 10 00:40:03.697647 kernel: pci 0000:00:15.4: PCI bridge to [bus 07] May 10 00:40:03.697695 kernel: pci 0000:00:15.4: bridge window [mem 0xfc500000-0xfc5fffff] May 10 00:40:03.697738 kernel: pci 0000:00:15.4: bridge window [mem 0xe6c00000-0xe6cfffff 64bit pref] May 10 00:40:03.697785 kernel: pci 0000:00:15.5: PCI bridge to [bus 08] May 10 00:40:03.697829 kernel: pci 0000:00:15.5: bridge window [mem 0xfc100000-0xfc1fffff] May 10 00:40:03.697875 kernel: pci 0000:00:15.5: bridge window [mem 0xe6800000-0xe68fffff 64bit pref] May 10 00:40:03.698195 kernel: pci 0000:00:15.6: PCI bridge to [bus 09] May 10 00:40:03.698453 kernel: pci 0000:00:15.6: bridge window [mem 0xfbd00000-0xfbdfffff] May 10 00:40:03.698504 kernel: pci 0000:00:15.6: bridge window [mem 0xe6400000-0xe64fffff 64bit pref] May 10 00:40:03.698568 kernel: pci 0000:00:15.7: PCI bridge to [bus 0a] May 10 00:40:03.698633 kernel: pci 0000:00:15.7: bridge window [mem 0xfb900000-0xfb9fffff] May 10 00:40:03.698678 kernel: pci 0000:00:15.7: bridge window [mem 0xe6000000-0xe60fffff 64bit pref] May 10 00:40:03.698728 kernel: pci 0000:0b:00.0: BAR 6: assigned [mem 0xfd400000-0xfd40ffff pref] May 10 00:40:03.698774 kernel: pci 0000:00:16.0: PCI bridge to [bus 0b] May 10 00:40:03.698818 kernel: pci 0000:00:16.0: bridge window [io 0x5000-0x5fff] May 10 00:40:03.698880 kernel: pci 0000:00:16.0: bridge window [mem 0xfd400000-0xfd4fffff] May 10 00:40:03.698946 kernel: pci 0000:00:16.0: bridge window [mem 0xc0200000-0xc03fffff 64bit pref] May 10 00:40:03.698992 kernel: pci 0000:00:16.1: PCI bridge to [bus 0c] May 10 00:40:03.699036 kernel: pci 0000:00:16.1: bridge window [io 0x9000-0x9fff] May 10 00:40:03.699079 kernel: pci 0000:00:16.1: bridge window [mem 0xfd000000-0xfd0fffff] May 10 00:40:03.699122 kernel: pci 0000:00:16.1: bridge window [mem 0xe7700000-0xe77fffff 64bit pref] May 10 00:40:03.699168 kernel: pci 0000:00:16.2: PCI bridge to [bus 0d] May 10 00:40:03.699211 kernel: pci 0000:00:16.2: bridge window [io 0xd000-0xdfff] May 10 00:40:03.699254 kernel: pci 0000:00:16.2: bridge window [mem 0xfcc00000-0xfccfffff] May 10 00:40:03.699297 kernel: pci 0000:00:16.2: bridge window [mem 0xe7300000-0xe73fffff 64bit pref] May 10 00:40:03.699341 kernel: pci 0000:00:16.3: PCI bridge to [bus 0e] May 10 00:40:03.699384 kernel: pci 0000:00:16.3: bridge window [mem 0xfc800000-0xfc8fffff] May 10 00:40:03.699427 kernel: pci 0000:00:16.3: bridge window [mem 0xe6f00000-0xe6ffffff 64bit pref] May 10 00:40:03.699470 kernel: pci 0000:00:16.4: PCI bridge to [bus 0f] May 10 00:40:03.699513 kernel: pci 0000:00:16.4: bridge window [mem 0xfc400000-0xfc4fffff] May 10 00:40:03.699573 kernel: pci 0000:00:16.4: bridge window [mem 0xe6b00000-0xe6bfffff 64bit pref] May 10 00:40:03.699633 kernel: pci 0000:00:16.5: PCI bridge to [bus 10] May 10 00:40:03.699694 kernel: pci 0000:00:16.5: bridge window [mem 0xfc000000-0xfc0fffff] May 10 00:40:03.699752 kernel: pci 0000:00:16.5: bridge window [mem 0xe6700000-0xe67fffff 64bit pref] May 10 00:40:03.699796 kernel: pci 0000:00:16.6: PCI bridge to [bus 11] May 10 00:40:03.699839 kernel: pci 0000:00:16.6: bridge window [mem 0xfbc00000-0xfbcfffff] May 10 00:40:03.699882 kernel: pci 0000:00:16.6: bridge window [mem 0xe6300000-0xe63fffff 64bit pref] May 10 00:40:03.700195 kernel: pci 0000:00:16.7: PCI bridge to [bus 12] May 10 00:40:03.700246 kernel: pci 0000:00:16.7: bridge window [mem 0xfb800000-0xfb8fffff] May 10 00:40:03.700295 kernel: pci 0000:00:16.7: bridge window [mem 0xe5f00000-0xe5ffffff 64bit pref] May 10 00:40:03.700350 kernel: pci 0000:00:17.0: PCI bridge to [bus 13] May 10 00:40:03.700396 kernel: pci 0000:00:17.0: bridge window [io 0x6000-0x6fff] May 10 00:40:03.700440 kernel: pci 0000:00:17.0: bridge window [mem 0xfd300000-0xfd3fffff] May 10 00:40:03.700484 kernel: pci 0000:00:17.0: bridge window [mem 0xe7a00000-0xe7afffff 64bit pref] May 10 00:40:03.700528 kernel: pci 0000:00:17.1: PCI bridge to [bus 14] May 10 00:40:03.700607 kernel: pci 0000:00:17.1: bridge window [io 0xa000-0xafff] May 10 00:40:03.700650 kernel: pci 0000:00:17.1: bridge window [mem 0xfcf00000-0xfcffffff] May 10 00:40:03.700694 kernel: pci 0000:00:17.1: bridge window [mem 0xe7600000-0xe76fffff 64bit pref] May 10 00:40:03.700738 kernel: pci 0000:00:17.2: PCI bridge to [bus 15] May 10 00:40:03.700782 kernel: pci 0000:00:17.2: bridge window [io 0xe000-0xefff] May 10 00:40:03.700827 kernel: pci 0000:00:17.2: bridge window [mem 0xfcb00000-0xfcbfffff] May 10 00:40:03.700872 kernel: pci 0000:00:17.2: bridge window [mem 0xe7200000-0xe72fffff 64bit pref] May 10 00:40:03.701124 kernel: pci 0000:00:17.3: PCI bridge to [bus 16] May 10 00:40:03.701179 kernel: pci 0000:00:17.3: bridge window [mem 0xfc700000-0xfc7fffff] May 10 00:40:03.701226 kernel: pci 0000:00:17.3: bridge window [mem 0xe6e00000-0xe6efffff 64bit pref] May 10 00:40:03.701270 kernel: pci 0000:00:17.4: PCI bridge to [bus 17] May 10 00:40:03.701332 kernel: pci 0000:00:17.4: bridge window [mem 0xfc300000-0xfc3fffff] May 10 00:40:03.701603 kernel: pci 0000:00:17.4: bridge window [mem 0xe6a00000-0xe6afffff 64bit pref] May 10 00:40:03.701659 kernel: pci 0000:00:17.5: PCI bridge to [bus 18] May 10 00:40:03.701930 kernel: pci 0000:00:17.5: bridge window [mem 0xfbf00000-0xfbffffff] May 10 00:40:03.701983 kernel: pci 0000:00:17.5: bridge window [mem 0xe6600000-0xe66fffff 64bit pref] May 10 00:40:03.702030 kernel: pci 0000:00:17.6: PCI bridge to [bus 19] May 10 00:40:03.702076 kernel: pci 0000:00:17.6: bridge window [mem 0xfbb00000-0xfbbfffff] May 10 00:40:03.702122 kernel: pci 0000:00:17.6: bridge window [mem 0xe6200000-0xe62fffff 64bit pref] May 10 00:40:03.702166 kernel: pci 0000:00:17.7: PCI bridge to [bus 1a] May 10 00:40:03.702437 kernel: pci 0000:00:17.7: bridge window [mem 0xfb700000-0xfb7fffff] May 10 00:40:03.702497 kernel: pci 0000:00:17.7: bridge window [mem 0xe5e00000-0xe5efffff 64bit pref] May 10 00:40:03.702547 kernel: pci 0000:00:18.0: PCI bridge to [bus 1b] May 10 00:40:03.702593 kernel: pci 0000:00:18.0: bridge window [io 0x7000-0x7fff] May 10 00:40:03.702885 kernel: pci 0000:00:18.0: bridge window [mem 0xfd200000-0xfd2fffff] May 10 00:40:03.702973 kernel: pci 0000:00:18.0: bridge window [mem 0xe7900000-0xe79fffff 64bit pref] May 10 00:40:03.703022 kernel: pci 0000:00:18.1: PCI bridge to [bus 1c] May 10 00:40:03.703074 kernel: pci 0000:00:18.1: bridge window [io 0xb000-0xbfff] May 10 00:40:03.703150 kernel: pci 0000:00:18.1: bridge window [mem 0xfce00000-0xfcefffff] May 10 00:40:03.703197 kernel: pci 0000:00:18.1: bridge window [mem 0xe7500000-0xe75fffff 64bit pref] May 10 00:40:03.703242 kernel: pci 0000:00:18.2: PCI bridge to [bus 1d] May 10 00:40:03.703287 kernel: pci 0000:00:18.2: bridge window [mem 0xfca00000-0xfcafffff] May 10 00:40:03.703564 kernel: pci 0000:00:18.2: bridge window [mem 0xe7100000-0xe71fffff 64bit pref] May 10 00:40:03.703623 kernel: pci 0000:00:18.3: PCI bridge to [bus 1e] May 10 00:40:03.703674 kernel: pci 0000:00:18.3: bridge window [mem 0xfc600000-0xfc6fffff] May 10 00:40:03.704007 kernel: pci 0000:00:18.3: bridge window [mem 0xe6d00000-0xe6dfffff 64bit pref] May 10 00:40:03.704061 kernel: pci 0000:00:18.4: PCI bridge to [bus 1f] May 10 00:40:03.704108 kernel: pci 0000:00:18.4: bridge window [mem 0xfc200000-0xfc2fffff] May 10 00:40:03.704154 kernel: pci 0000:00:18.4: bridge window [mem 0xe6900000-0xe69fffff 64bit pref] May 10 00:40:03.704199 kernel: pci 0000:00:18.5: PCI bridge to [bus 20] May 10 00:40:03.704244 kernel: pci 0000:00:18.5: bridge window [mem 0xfbe00000-0xfbefffff] May 10 00:40:03.704289 kernel: pci 0000:00:18.5: bridge window [mem 0xe6500000-0xe65fffff 64bit pref] May 10 00:40:03.704391 kernel: pci 0000:00:18.6: PCI bridge to [bus 21] May 10 00:40:03.704437 kernel: pci 0000:00:18.6: bridge window [mem 0xfba00000-0xfbafffff] May 10 00:40:03.704749 kernel: pci 0000:00:18.6: bridge window [mem 0xe6100000-0xe61fffff 64bit pref] May 10 00:40:03.704804 kernel: pci 0000:00:18.7: PCI bridge to [bus 22] May 10 00:40:03.704850 kernel: pci 0000:00:18.7: bridge window [mem 0xfb600000-0xfb6fffff] May 10 00:40:03.704895 kernel: pci 0000:00:18.7: bridge window [mem 0xe5d00000-0xe5dfffff 64bit pref] May 10 00:40:03.704962 kernel: pci_bus 0000:00: resource 4 [mem 0x000a0000-0x000bffff window] May 10 00:40:03.705004 kernel: pci_bus 0000:00: resource 5 [mem 0x000cc000-0x000dbfff window] May 10 00:40:03.705042 kernel: pci_bus 0000:00: resource 6 [mem 0xc0000000-0xfebfffff window] May 10 00:40:03.705081 kernel: pci_bus 0000:00: resource 7 [io 0x0000-0x0cf7 window] May 10 00:40:03.705121 kernel: pci_bus 0000:00: resource 8 [io 0x0d00-0xfeff window] May 10 00:40:03.705163 kernel: pci_bus 0000:02: resource 0 [io 0x2000-0x3fff] May 10 00:40:03.705204 kernel: pci_bus 0000:02: resource 1 [mem 0xfd600000-0xfdffffff] May 10 00:40:03.705244 kernel: pci_bus 0000:02: resource 2 [mem 0xe7b00000-0xe7ffffff 64bit pref] May 10 00:40:03.705283 kernel: pci_bus 0000:02: resource 4 [mem 0x000a0000-0x000bffff window] May 10 00:40:03.705323 kernel: pci_bus 0000:02: resource 5 [mem 0x000cc000-0x000dbfff window] May 10 00:40:03.705363 kernel: pci_bus 0000:02: resource 6 [mem 0xc0000000-0xfebfffff window] May 10 00:40:03.705405 kernel: pci_bus 0000:02: resource 7 [io 0x0000-0x0cf7 window] May 10 00:40:03.705444 kernel: pci_bus 0000:02: resource 8 [io 0x0d00-0xfeff window] May 10 00:40:03.705697 kernel: pci_bus 0000:03: resource 0 [io 0x4000-0x4fff] May 10 00:40:03.705747 kernel: pci_bus 0000:03: resource 1 [mem 0xfd500000-0xfd5fffff] May 10 00:40:03.705788 kernel: pci_bus 0000:03: resource 2 [mem 0xc0000000-0xc01fffff 64bit pref] May 10 00:40:03.705832 kernel: pci_bus 0000:04: resource 0 [io 0x8000-0x8fff] May 10 00:40:03.705878 kernel: pci_bus 0000:04: resource 1 [mem 0xfd100000-0xfd1fffff] May 10 00:40:03.705929 kernel: pci_bus 0000:04: resource 2 [mem 0xe7800000-0xe78fffff 64bit pref] May 10 00:40:03.705977 kernel: pci_bus 0000:05: resource 0 [io 0xc000-0xcfff] May 10 00:40:03.706018 kernel: pci_bus 0000:05: resource 1 [mem 0xfcd00000-0xfcdfffff] May 10 00:40:03.706058 kernel: pci_bus 0000:05: resource 2 [mem 0xe7400000-0xe74fffff 64bit pref] May 10 00:40:03.706103 kernel: pci_bus 0000:06: resource 1 [mem 0xfc900000-0xfc9fffff] May 10 00:40:03.706145 kernel: pci_bus 0000:06: resource 2 [mem 0xe7000000-0xe70fffff 64bit pref] May 10 00:40:03.706189 kernel: pci_bus 0000:07: resource 1 [mem 0xfc500000-0xfc5fffff] May 10 00:40:03.706233 kernel: pci_bus 0000:07: resource 2 [mem 0xe6c00000-0xe6cfffff 64bit pref] May 10 00:40:03.706279 kernel: pci_bus 0000:08: resource 1 [mem 0xfc100000-0xfc1fffff] May 10 00:40:03.706361 kernel: pci_bus 0000:08: resource 2 [mem 0xe6800000-0xe68fffff 64bit pref] May 10 00:40:03.706423 kernel: pci_bus 0000:09: resource 1 [mem 0xfbd00000-0xfbdfffff] May 10 00:40:03.706463 kernel: pci_bus 0000:09: resource 2 [mem 0xe6400000-0xe64fffff 64bit pref] May 10 00:40:03.706508 kernel: pci_bus 0000:0a: resource 1 [mem 0xfb900000-0xfb9fffff] May 10 00:40:03.706550 kernel: pci_bus 0000:0a: resource 2 [mem 0xe6000000-0xe60fffff 64bit pref] May 10 00:40:03.706597 kernel: pci_bus 0000:0b: resource 0 [io 0x5000-0x5fff] May 10 00:40:03.706639 kernel: pci_bus 0000:0b: resource 1 [mem 0xfd400000-0xfd4fffff] May 10 00:40:03.706678 kernel: pci_bus 0000:0b: resource 2 [mem 0xc0200000-0xc03fffff 64bit pref] May 10 00:40:03.706723 kernel: pci_bus 0000:0c: resource 0 [io 0x9000-0x9fff] May 10 00:40:03.706765 kernel: pci_bus 0000:0c: resource 1 [mem 0xfd000000-0xfd0fffff] May 10 00:40:03.706807 kernel: pci_bus 0000:0c: resource 2 [mem 0xe7700000-0xe77fffff 64bit pref] May 10 00:40:03.706860 kernel: pci_bus 0000:0d: resource 0 [io 0xd000-0xdfff] May 10 00:40:03.707112 kernel: pci_bus 0000:0d: resource 1 [mem 0xfcc00000-0xfccfffff] May 10 00:40:03.707168 kernel: pci_bus 0000:0d: resource 2 [mem 0xe7300000-0xe73fffff 64bit pref] May 10 00:40:03.707216 kernel: pci_bus 0000:0e: resource 1 [mem 0xfc800000-0xfc8fffff] May 10 00:40:03.707258 kernel: pci_bus 0000:0e: resource 2 [mem 0xe6f00000-0xe6ffffff 64bit pref] May 10 00:40:03.707362 kernel: pci_bus 0000:0f: resource 1 [mem 0xfc400000-0xfc4fffff] May 10 00:40:03.707618 kernel: pci_bus 0000:0f: resource 2 [mem 0xe6b00000-0xe6bfffff 64bit pref] May 10 00:40:03.707672 kernel: pci_bus 0000:10: resource 1 [mem 0xfc000000-0xfc0fffff] May 10 00:40:03.707930 kernel: pci_bus 0000:10: resource 2 [mem 0xe6700000-0xe67fffff 64bit pref] May 10 00:40:03.707985 kernel: pci_bus 0000:11: resource 1 [mem 0xfbc00000-0xfbcfffff] May 10 00:40:03.708306 kernel: pci_bus 0000:11: resource 2 [mem 0xe6300000-0xe63fffff 64bit pref] May 10 00:40:03.708361 kernel: pci_bus 0000:12: resource 1 [mem 0xfb800000-0xfb8fffff] May 10 00:40:03.708404 kernel: pci_bus 0000:12: resource 2 [mem 0xe5f00000-0xe5ffffff 64bit pref] May 10 00:40:03.708449 kernel: pci_bus 0000:13: resource 0 [io 0x6000-0x6fff] May 10 00:40:03.708590 kernel: pci_bus 0000:13: resource 1 [mem 0xfd300000-0xfd3fffff] May 10 00:40:03.708637 kernel: pci_bus 0000:13: resource 2 [mem 0xe7a00000-0xe7afffff 64bit pref] May 10 00:40:03.708682 kernel: pci_bus 0000:14: resource 0 [io 0xa000-0xafff] May 10 00:40:03.709027 kernel: pci_bus 0000:14: resource 1 [mem 0xfcf00000-0xfcffffff] May 10 00:40:03.709077 kernel: pci_bus 0000:14: resource 2 [mem 0xe7600000-0xe76fffff 64bit pref] May 10 00:40:03.709123 kernel: pci_bus 0000:15: resource 0 [io 0xe000-0xefff] May 10 00:40:03.709165 kernel: pci_bus 0000:15: resource 1 [mem 0xfcb00000-0xfcbfffff] May 10 00:40:03.709338 kernel: pci_bus 0000:15: resource 2 [mem 0xe7200000-0xe72fffff 64bit pref] May 10 00:40:03.709579 kernel: pci_bus 0000:16: resource 1 [mem 0xfc700000-0xfc7fffff] May 10 00:40:03.709626 kernel: pci_bus 0000:16: resource 2 [mem 0xe6e00000-0xe6efffff 64bit pref] May 10 00:40:03.709959 kernel: pci_bus 0000:17: resource 1 [mem 0xfc300000-0xfc3fffff] May 10 00:40:03.710013 kernel: pci_bus 0000:17: resource 2 [mem 0xe6a00000-0xe6afffff 64bit pref] May 10 00:40:03.710061 kernel: pci_bus 0000:18: resource 1 [mem 0xfbf00000-0xfbffffff] May 10 00:40:03.710103 kernel: pci_bus 0000:18: resource 2 [mem 0xe6600000-0xe66fffff 64bit pref] May 10 00:40:03.710147 kernel: pci_bus 0000:19: resource 1 [mem 0xfbb00000-0xfbbfffff] May 10 00:40:03.710369 kernel: pci_bus 0000:19: resource 2 [mem 0xe6200000-0xe62fffff 64bit pref] May 10 00:40:03.710417 kernel: pci_bus 0000:1a: resource 1 [mem 0xfb700000-0xfb7fffff] May 10 00:40:03.710734 kernel: pci_bus 0000:1a: resource 2 [mem 0xe5e00000-0xe5efffff 64bit pref] May 10 00:40:03.710787 kernel: pci_bus 0000:1b: resource 0 [io 0x7000-0x7fff] May 10 00:40:03.710831 kernel: pci_bus 0000:1b: resource 1 [mem 0xfd200000-0xfd2fffff] May 10 00:40:03.710875 kernel: pci_bus 0000:1b: resource 2 [mem 0xe7900000-0xe79fffff 64bit pref] May 10 00:40:03.711103 kernel: pci_bus 0000:1c: resource 0 [io 0xb000-0xbfff] May 10 00:40:03.711343 kernel: pci_bus 0000:1c: resource 1 [mem 0xfce00000-0xfcefffff] May 10 00:40:03.711391 kernel: pci_bus 0000:1c: resource 2 [mem 0xe7500000-0xe75fffff 64bit pref] May 10 00:40:03.711697 kernel: pci_bus 0000:1d: resource 1 [mem 0xfca00000-0xfcafffff] May 10 00:40:03.711752 kernel: pci_bus 0000:1d: resource 2 [mem 0xe7100000-0xe71fffff 64bit pref] May 10 00:40:03.711805 kernel: pci_bus 0000:1e: resource 1 [mem 0xfc600000-0xfc6fffff] May 10 00:40:03.711847 kernel: pci_bus 0000:1e: resource 2 [mem 0xe6d00000-0xe6dfffff 64bit pref] May 10 00:40:03.712002 kernel: pci_bus 0000:1f: resource 1 [mem 0xfc200000-0xfc2fffff] May 10 00:40:03.712057 kernel: pci_bus 0000:1f: resource 2 [mem 0xe6900000-0xe69fffff 64bit pref] May 10 00:40:03.712103 kernel: pci_bus 0000:20: resource 1 [mem 0xfbe00000-0xfbefffff] May 10 00:40:03.712146 kernel: pci_bus 0000:20: resource 2 [mem 0xe6500000-0xe65fffff 64bit pref] May 10 00:40:03.712190 kernel: pci_bus 0000:21: resource 1 [mem 0xfba00000-0xfbafffff] May 10 00:40:03.712232 kernel: pci_bus 0000:21: resource 2 [mem 0xe6100000-0xe61fffff 64bit pref] May 10 00:40:03.712276 kernel: pci_bus 0000:22: resource 1 [mem 0xfb600000-0xfb6fffff] May 10 00:40:03.712319 kernel: pci_bus 0000:22: resource 2 [mem 0xe5d00000-0xe5dfffff 64bit pref] May 10 00:40:03.712368 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers May 10 00:40:03.712377 kernel: PCI: CLS 32 bytes, default 64 May 10 00:40:03.712383 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer May 10 00:40:03.712389 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x311fd3cd494, max_idle_ns: 440795223879 ns May 10 00:40:03.712395 kernel: clocksource: Switched to clocksource tsc May 10 00:40:03.712401 kernel: Initialise system trusted keyrings May 10 00:40:03.712407 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 May 10 00:40:03.712414 kernel: Key type asymmetric registered May 10 00:40:03.712420 kernel: Asymmetric key parser 'x509' registered May 10 00:40:03.712426 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 249) May 10 00:40:03.712432 kernel: io scheduler mq-deadline registered May 10 00:40:03.712438 kernel: io scheduler kyber registered May 10 00:40:03.712444 kernel: io scheduler bfq registered May 10 00:40:03.712489 kernel: pcieport 0000:00:15.0: PME: Signaling with IRQ 24 May 10 00:40:03.712536 kernel: pcieport 0000:00:15.0: pciehp: Slot #160 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ May 10 00:40:03.712581 kernel: pcieport 0000:00:15.1: PME: Signaling with IRQ 25 May 10 00:40:03.712627 kernel: pcieport 0000:00:15.1: pciehp: Slot #161 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ May 10 00:40:03.712672 kernel: pcieport 0000:00:15.2: PME: Signaling with IRQ 26 May 10 00:40:03.712716 kernel: pcieport 0000:00:15.2: pciehp: Slot #162 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ May 10 00:40:03.712760 kernel: pcieport 0000:00:15.3: PME: Signaling with IRQ 27 May 10 00:40:03.712804 kernel: pcieport 0000:00:15.3: pciehp: Slot #163 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ May 10 00:40:03.712850 kernel: pcieport 0000:00:15.4: PME: Signaling with IRQ 28 May 10 00:40:03.712896 kernel: pcieport 0000:00:15.4: pciehp: Slot #164 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ May 10 00:40:03.713228 kernel: pcieport 0000:00:15.5: PME: Signaling with IRQ 29 May 10 00:40:03.713278 kernel: pcieport 0000:00:15.5: pciehp: Slot #165 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ May 10 00:40:03.713355 kernel: pcieport 0000:00:15.6: PME: Signaling with IRQ 30 May 10 00:40:03.713403 kernel: pcieport 0000:00:15.6: pciehp: Slot #166 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ May 10 00:40:03.713452 kernel: pcieport 0000:00:15.7: PME: Signaling with IRQ 31 May 10 00:40:03.713498 kernel: pcieport 0000:00:15.7: pciehp: Slot #167 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ May 10 00:40:03.713544 kernel: pcieport 0000:00:16.0: PME: Signaling with IRQ 32 May 10 00:40:03.713589 kernel: pcieport 0000:00:16.0: pciehp: Slot #192 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ May 10 00:40:03.713764 kernel: pcieport 0000:00:16.1: PME: Signaling with IRQ 33 May 10 00:40:03.714047 kernel: pcieport 0000:00:16.1: pciehp: Slot #193 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ May 10 00:40:03.714102 kernel: pcieport 0000:00:16.2: PME: Signaling with IRQ 34 May 10 00:40:03.714154 kernel: pcieport 0000:00:16.2: pciehp: Slot #194 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ May 10 00:40:03.714200 kernel: pcieport 0000:00:16.3: PME: Signaling with IRQ 35 May 10 00:40:03.714251 kernel: pcieport 0000:00:16.3: pciehp: Slot #195 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ May 10 00:40:03.714323 kernel: pcieport 0000:00:16.4: PME: Signaling with IRQ 36 May 10 00:40:03.714372 kernel: pcieport 0000:00:16.4: pciehp: Slot #196 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ May 10 00:40:03.714419 kernel: pcieport 0000:00:16.5: PME: Signaling with IRQ 37 May 10 00:40:03.714468 kernel: pcieport 0000:00:16.5: pciehp: Slot #197 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ May 10 00:40:03.714513 kernel: pcieport 0000:00:16.6: PME: Signaling with IRQ 38 May 10 00:40:03.714560 kernel: pcieport 0000:00:16.6: pciehp: Slot #198 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ May 10 00:40:03.714606 kernel: pcieport 0000:00:16.7: PME: Signaling with IRQ 39 May 10 00:40:03.714653 kernel: pcieport 0000:00:16.7: pciehp: Slot #199 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ May 10 00:40:03.714701 kernel: pcieport 0000:00:17.0: PME: Signaling with IRQ 40 May 10 00:40:03.714747 kernel: pcieport 0000:00:17.0: pciehp: Slot #224 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ May 10 00:40:03.714793 kernel: pcieport 0000:00:17.1: PME: Signaling with IRQ 41 May 10 00:40:03.714839 kernel: pcieport 0000:00:17.1: pciehp: Slot #225 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ May 10 00:40:03.715169 kernel: pcieport 0000:00:17.2: PME: Signaling with IRQ 42 May 10 00:40:03.715230 kernel: pcieport 0000:00:17.2: pciehp: Slot #226 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ May 10 00:40:03.715280 kernel: pcieport 0000:00:17.3: PME: Signaling with IRQ 43 May 10 00:40:03.715330 kernel: pcieport 0000:00:17.3: pciehp: Slot #227 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ May 10 00:40:03.715377 kernel: pcieport 0000:00:17.4: PME: Signaling with IRQ 44 May 10 00:40:03.715424 kernel: pcieport 0000:00:17.4: pciehp: Slot #228 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ May 10 00:40:03.715470 kernel: pcieport 0000:00:17.5: PME: Signaling with IRQ 45 May 10 00:40:03.715517 kernel: pcieport 0000:00:17.5: pciehp: Slot #229 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ May 10 00:40:03.715566 kernel: pcieport 0000:00:17.6: PME: Signaling with IRQ 46 May 10 00:40:03.715612 kernel: pcieport 0000:00:17.6: pciehp: Slot #230 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ May 10 00:40:03.715658 kernel: pcieport 0000:00:17.7: PME: Signaling with IRQ 47 May 10 00:40:03.715705 kernel: pcieport 0000:00:17.7: pciehp: Slot #231 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ May 10 00:40:03.715751 kernel: pcieport 0000:00:18.0: PME: Signaling with IRQ 48 May 10 00:40:03.715795 kernel: pcieport 0000:00:18.0: pciehp: Slot #256 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ May 10 00:40:03.715844 kernel: pcieport 0000:00:18.1: PME: Signaling with IRQ 49 May 10 00:40:03.715888 kernel: pcieport 0000:00:18.1: pciehp: Slot #257 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ May 10 00:40:03.715947 kernel: pcieport 0000:00:18.2: PME: Signaling with IRQ 50 May 10 00:40:03.715996 kernel: pcieport 0000:00:18.2: pciehp: Slot #258 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ May 10 00:40:03.716042 kernel: pcieport 0000:00:18.3: PME: Signaling with IRQ 51 May 10 00:40:03.716089 kernel: pcieport 0000:00:18.3: pciehp: Slot #259 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ May 10 00:40:03.716137 kernel: pcieport 0000:00:18.4: PME: Signaling with IRQ 52 May 10 00:40:03.716184 kernel: pcieport 0000:00:18.4: pciehp: Slot #260 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ May 10 00:40:03.716230 kernel: pcieport 0000:00:18.5: PME: Signaling with IRQ 53 May 10 00:40:03.716277 kernel: pcieport 0000:00:18.5: pciehp: Slot #261 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ May 10 00:40:03.716323 kernel: pcieport 0000:00:18.6: PME: Signaling with IRQ 54 May 10 00:40:03.716371 kernel: pcieport 0000:00:18.6: pciehp: Slot #262 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ May 10 00:40:03.716417 kernel: pcieport 0000:00:18.7: PME: Signaling with IRQ 55 May 10 00:40:03.716464 kernel: pcieport 0000:00:18.7: pciehp: Slot #263 AttnBtn+ PwrCtrl+ MRL- AttnInd- PwrInd- HotPlug+ Surprise- Interlock- NoCompl+ IbPresDis- LLActRep+ May 10 00:40:03.716473 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 May 10 00:40:03.716480 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled May 10 00:40:03.716487 kernel: 00:05: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A May 10 00:40:03.716493 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBC,PNP0f13:MOUS] at 0x60,0x64 irq 1,12 May 10 00:40:03.716501 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 May 10 00:40:03.716507 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 May 10 00:40:03.716818 kernel: rtc_cmos 00:01: registered as rtc0 May 10 00:40:03.716868 kernel: rtc_cmos 00:01: setting system clock to 2025-05-10T00:40:03 UTC (1746837603) May 10 00:40:03.716938 kernel: rtc_cmos 00:01: alarms up to one month, y3k, 114 bytes nvram May 10 00:40:03.716948 kernel: intel_pstate: CPU model not supported May 10 00:40:03.716955 kernel: NET: Registered PF_INET6 protocol family May 10 00:40:03.716961 kernel: Segment Routing with IPv6 May 10 00:40:03.716969 kernel: In-situ OAM (IOAM) with IPv6 May 10 00:40:03.716975 kernel: NET: Registered PF_PACKET protocol family May 10 00:40:03.716981 kernel: Key type dns_resolver registered May 10 00:40:03.716987 kernel: IPI shorthand broadcast: enabled May 10 00:40:03.716992 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 May 10 00:40:03.716999 kernel: sched_clock: Marking stable (813464930, 220090680)->(1094257219, -60701609) May 10 00:40:03.717005 kernel: registered taskstats version 1 May 10 00:40:03.717011 kernel: Loading compiled-in X.509 certificates May 10 00:40:03.717017 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 5.15.181-flatcar: 0c62a22cd9157131d2e97d5a2e1bd9023e187117' May 10 00:40:03.717024 kernel: Key type .fscrypt registered May 10 00:40:03.717029 kernel: Key type fscrypt-provisioning registered May 10 00:40:03.717035 kernel: ima: No TPM chip found, activating TPM-bypass! May 10 00:40:03.717041 kernel: ima: Allocated hash algorithm: sha1 May 10 00:40:03.717047 kernel: ima: No architecture policies found May 10 00:40:03.717053 kernel: clk: Disabling unused clocks May 10 00:40:03.717059 kernel: Freeing unused kernel image (initmem) memory: 47456K May 10 00:40:03.717065 kernel: Write protecting the kernel read-only data: 28672k May 10 00:40:03.717073 kernel: Freeing unused kernel image (text/rodata gap) memory: 2040K May 10 00:40:03.717080 kernel: Freeing unused kernel image (rodata/data gap) memory: 612K May 10 00:40:03.717086 kernel: Run /init as init process May 10 00:40:03.717092 kernel: with arguments: May 10 00:40:03.717098 kernel: /init May 10 00:40:03.717103 kernel: with environment: May 10 00:40:03.717109 kernel: HOME=/ May 10 00:40:03.717115 kernel: TERM=linux May 10 00:40:03.717121 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a May 10 00:40:03.717128 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) May 10 00:40:03.717137 systemd[1]: Detected virtualization vmware. May 10 00:40:03.717143 systemd[1]: Detected architecture x86-64. May 10 00:40:03.717149 systemd[1]: Running in initrd. May 10 00:40:03.717155 systemd[1]: No hostname configured, using default hostname. May 10 00:40:03.717161 systemd[1]: Hostname set to . May 10 00:40:03.717168 systemd[1]: Initializing machine ID from random generator. May 10 00:40:03.717174 systemd[1]: Queued start job for default target initrd.target. May 10 00:40:03.717180 systemd[1]: Started systemd-ask-password-console.path. May 10 00:40:03.717186 systemd[1]: Reached target cryptsetup.target. May 10 00:40:03.717192 systemd[1]: Reached target paths.target. May 10 00:40:03.717198 systemd[1]: Reached target slices.target. May 10 00:40:03.717204 systemd[1]: Reached target swap.target. May 10 00:40:03.717210 systemd[1]: Reached target timers.target. May 10 00:40:03.717216 systemd[1]: Listening on iscsid.socket. May 10 00:40:03.717222 systemd[1]: Listening on iscsiuio.socket. May 10 00:40:03.717229 systemd[1]: Listening on systemd-journald-audit.socket. May 10 00:40:03.717235 systemd[1]: Listening on systemd-journald-dev-log.socket. May 10 00:40:03.717241 systemd[1]: Listening on systemd-journald.socket. May 10 00:40:03.717247 systemd[1]: Listening on systemd-networkd.socket. May 10 00:40:03.717253 systemd[1]: Listening on systemd-udevd-control.socket. May 10 00:40:03.717259 systemd[1]: Listening on systemd-udevd-kernel.socket. May 10 00:40:03.717266 systemd[1]: Reached target sockets.target. May 10 00:40:03.717272 systemd[1]: Starting kmod-static-nodes.service... May 10 00:40:03.717278 systemd[1]: Finished network-cleanup.service. May 10 00:40:03.717285 systemd[1]: Starting systemd-fsck-usr.service... May 10 00:40:03.717294 systemd[1]: Starting systemd-journald.service... May 10 00:40:03.717301 systemd[1]: Starting systemd-modules-load.service... May 10 00:40:03.717307 systemd[1]: Starting systemd-resolved.service... May 10 00:40:03.717313 systemd[1]: Starting systemd-vconsole-setup.service... May 10 00:40:03.717319 systemd[1]: Finished kmod-static-nodes.service. May 10 00:40:03.717325 systemd[1]: Finished systemd-fsck-usr.service. May 10 00:40:03.717331 systemd[1]: Starting systemd-tmpfiles-setup-dev.service... May 10 00:40:03.717338 systemd[1]: Finished systemd-vconsole-setup.service. May 10 00:40:03.717344 kernel: audit: type=1130 audit(1746837603.651:2): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:03.717350 systemd[1]: Finished systemd-tmpfiles-setup-dev.service. May 10 00:40:03.717356 kernel: audit: type=1130 audit(1746837603.660:3): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:03.717362 systemd[1]: Starting dracut-cmdline-ask.service... May 10 00:40:03.717369 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. May 10 00:40:03.717375 systemd[1]: Finished dracut-cmdline-ask.service. May 10 00:40:03.717381 systemd[1]: Starting dracut-cmdline.service... May 10 00:40:03.717402 kernel: audit: type=1130 audit(1746837603.676:4): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:03.717413 systemd[1]: Started systemd-resolved.service. May 10 00:40:03.717420 systemd[1]: Reached target nss-lookup.target. May 10 00:40:03.717427 kernel: audit: type=1130 audit(1746837603.683:5): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:03.717433 kernel: Bridge firewalling registered May 10 00:40:03.717440 kernel: SCSI subsystem initialized May 10 00:40:03.717446 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. May 10 00:40:03.717664 kernel: device-mapper: uevent: version 1.0.3 May 10 00:40:03.717677 systemd-journald[216]: Journal started May 10 00:40:03.717710 systemd-journald[216]: Runtime Journal (/run/log/journal/9ac8b8621a4f477f9866d476204fbb4e) is 4.8M, max 38.8M, 34.0M free. May 10 00:40:03.720865 systemd[1]: Started systemd-journald.service. May 10 00:40:03.720880 kernel: audit: type=1130 audit(1746837603.717:6): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:03.651000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:03.660000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:03.676000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:03.683000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:03.717000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:03.647825 systemd-modules-load[217]: Inserted module 'overlay' May 10 00:40:03.681637 systemd-resolved[218]: Positive Trust Anchors: May 10 00:40:03.681641 systemd-resolved[218]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d May 10 00:40:03.681661 systemd-resolved[218]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test May 10 00:40:03.723272 kernel: device-mapper: ioctl: 4.45.0-ioctl (2021-03-22) initialised: dm-devel@redhat.com May 10 00:40:03.683392 systemd-resolved[218]: Defaulting to hostname 'linux'. May 10 00:40:03.688511 systemd-modules-load[217]: Inserted module 'br_netfilter' May 10 00:40:03.724258 dracut-cmdline[231]: dracut-dracut-053 May 10 00:40:03.724258 dracut-cmdline[231]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LA May 10 00:40:03.724258 dracut-cmdline[231]: BEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=vmware flatcar.autologin verity.usrhash=39569409b30be1967efab22b453b92a780dcf0fe8e1448a18bf235b5cf33e54a May 10 00:40:03.725227 systemd-modules-load[217]: Inserted module 'dm_multipath' May 10 00:40:03.726057 systemd[1]: Finished systemd-modules-load.service. May 10 00:40:03.728956 kernel: audit: type=1130 audit(1746837603.724:7): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:03.724000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:03.728711 systemd[1]: Starting systemd-sysctl.service... May 10 00:40:03.733009 systemd[1]: Finished systemd-sysctl.service. May 10 00:40:03.735497 kernel: audit: type=1130 audit(1746837603.731:8): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:03.731000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:03.739939 kernel: Loading iSCSI transport class v2.0-870. May 10 00:40:03.751917 kernel: iscsi: registered transport (tcp) May 10 00:40:03.766380 kernel: iscsi: registered transport (qla4xxx) May 10 00:40:03.766404 kernel: QLogic iSCSI HBA Driver May 10 00:40:03.782263 systemd[1]: Finished dracut-cmdline.service. May 10 00:40:03.780000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:03.782858 systemd[1]: Starting dracut-pre-udev.service... May 10 00:40:03.785928 kernel: audit: type=1130 audit(1746837603.780:9): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:03.818949 kernel: raid6: avx2x4 gen() 49272 MB/s May 10 00:40:03.835946 kernel: raid6: avx2x4 xor() 21124 MB/s May 10 00:40:03.852941 kernel: raid6: avx2x2 gen() 54298 MB/s May 10 00:40:03.869946 kernel: raid6: avx2x2 xor() 32938 MB/s May 10 00:40:03.886949 kernel: raid6: avx2x1 gen() 46136 MB/s May 10 00:40:03.903947 kernel: raid6: avx2x1 xor() 28729 MB/s May 10 00:40:03.920947 kernel: raid6: sse2x4 gen() 21833 MB/s May 10 00:40:03.937950 kernel: raid6: sse2x4 xor() 11956 MB/s May 10 00:40:03.954917 kernel: raid6: sse2x2 gen() 22255 MB/s May 10 00:40:03.971948 kernel: raid6: sse2x2 xor() 13765 MB/s May 10 00:40:03.988946 kernel: raid6: sse2x1 gen() 18808 MB/s May 10 00:40:04.006068 kernel: raid6: sse2x1 xor() 9011 MB/s May 10 00:40:04.006087 kernel: raid6: using algorithm avx2x2 gen() 54298 MB/s May 10 00:40:04.006096 kernel: raid6: .... xor() 32938 MB/s, rmw enabled May 10 00:40:04.007268 kernel: raid6: using avx2x2 recovery algorithm May 10 00:40:04.015916 kernel: xor: automatically using best checksumming function avx May 10 00:40:04.073283 kernel: Btrfs loaded, crc32c=crc32c-intel, zoned=no, fsverity=no May 10 00:40:04.075000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:04.076989 systemd[1]: Finished dracut-pre-udev.service. May 10 00:40:04.077568 systemd[1]: Starting systemd-udevd.service... May 10 00:40:04.075000 audit: BPF prog-id=7 op=LOAD May 10 00:40:04.075000 audit: BPF prog-id=8 op=LOAD May 10 00:40:04.080919 kernel: audit: type=1130 audit(1746837604.075:10): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:04.087371 systemd-udevd[414]: Using default interface naming scheme 'v252'. May 10 00:40:04.089980 systemd[1]: Started systemd-udevd.service. May 10 00:40:04.090469 systemd[1]: Starting dracut-pre-trigger.service... May 10 00:40:04.088000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:04.098070 dracut-pre-trigger[419]: rd.md=0: removing MD RAID activation May 10 00:40:04.112000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:04.114758 systemd[1]: Finished dracut-pre-trigger.service. May 10 00:40:04.115252 systemd[1]: Starting systemd-udev-trigger.service... May 10 00:40:04.173000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:04.175363 systemd[1]: Finished systemd-udev-trigger.service. May 10 00:40:04.230921 kernel: VMware PVSCSI driver - version 1.0.7.0-k May 10 00:40:04.237342 kernel: VMware vmxnet3 virtual NIC driver - version 1.6.0.0-k-NAPI May 10 00:40:04.237363 kernel: vmxnet3 0000:0b:00.0: # of Tx queues : 2, # of Rx queues : 2 May 10 00:40:04.263132 kernel: vmw_pvscsi: using 64bit dma May 10 00:40:04.263144 kernel: vmw_pvscsi: max_id: 16 May 10 00:40:04.263152 kernel: vmw_pvscsi: setting ring_pages to 8 May 10 00:40:04.263162 kernel: vmw_pvscsi: enabling reqCallThreshold May 10 00:40:04.263169 kernel: vmw_pvscsi: driver-based request coalescing enabled May 10 00:40:04.263176 kernel: vmw_pvscsi: using MSI-X May 10 00:40:04.263183 kernel: scsi host0: VMware PVSCSI storage adapter rev 2, req/cmp/msg rings: 8/8/1 pages, cmd_per_lun=254 May 10 00:40:04.263252 kernel: vmw_pvscsi 0000:03:00.0: VMware PVSCSI rev 2 host #0 May 10 00:40:04.263311 kernel: scsi 0:0:0:0: Direct-Access VMware Virtual disk 2.0 PQ: 0 ANSI: 6 May 10 00:40:04.263374 kernel: libata version 3.00 loaded. May 10 00:40:04.263381 kernel: cryptd: max_cpu_qlen set to 1000 May 10 00:40:04.263390 kernel: ata_piix 0000:00:07.1: version 2.13 May 10 00:40:04.267198 kernel: scsi host1: ata_piix May 10 00:40:04.267268 kernel: vmxnet3 0000:0b:00.0 eth0: NIC Link is Up 10000 Mbps May 10 00:40:04.267327 kernel: scsi host2: ata_piix May 10 00:40:04.267384 kernel: ata1: PATA max UDMA/33 cmd 0x1f0 ctl 0x3f6 bmdma 0x1060 irq 14 May 10 00:40:04.267393 kernel: ata2: PATA max UDMA/33 cmd 0x170 ctl 0x376 bmdma 0x1068 irq 15 May 10 00:40:04.269915 kernel: vmxnet3 0000:0b:00.0 ens192: renamed from eth0 May 10 00:40:04.275914 kernel: AVX2 version of gcm_enc/dec engaged. May 10 00:40:04.275931 kernel: AES CTR mode by8 optimization enabled May 10 00:40:04.435985 kernel: ata2.00: ATAPI: VMware Virtual IDE CDROM Drive, 00000001, max UDMA/33 May 10 00:40:04.439920 kernel: scsi 2:0:0:0: CD-ROM NECVMWar VMware IDE CDR10 1.00 PQ: 0 ANSI: 5 May 10 00:40:04.451740 kernel: sd 0:0:0:0: [sda] 17805312 512-byte logical blocks: (9.12 GB/8.49 GiB) May 10 00:40:04.475756 kernel: sd 0:0:0:0: [sda] Write Protect is off May 10 00:40:04.475848 kernel: sd 0:0:0:0: [sda] Mode Sense: 31 00 00 00 May 10 00:40:04.475940 kernel: sd 0:0:0:0: [sda] Cache data unavailable May 10 00:40:04.476021 kernel: sd 0:0:0:0: [sda] Assuming drive cache: write through May 10 00:40:04.476099 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 May 10 00:40:04.476110 kernel: sd 0:0:0:0: [sda] Attached SCSI disk May 10 00:40:04.497401 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 1x/1x writer dvd-ram cd/rw xa/form2 cdda tray May 10 00:40:04.514426 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 May 10 00:40:04.514438 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/sda6 scanned by (udev-worker) (470) May 10 00:40:04.514446 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 May 10 00:40:04.506310 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device. May 10 00:40:04.510369 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device. May 10 00:40:04.510489 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device. May 10 00:40:04.511032 systemd[1]: Starting disk-uuid.service... May 10 00:40:04.523994 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. May 10 00:40:04.531445 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device. May 10 00:40:05.528711 disk-uuid[547]: The operation has completed successfully. May 10 00:40:05.529008 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 May 10 00:40:05.561737 systemd[1]: disk-uuid.service: Deactivated successfully. May 10 00:40:05.562053 systemd[1]: Finished disk-uuid.service. May 10 00:40:05.560000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:05.560000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:05.562827 systemd[1]: Starting verity-setup.service... May 10 00:40:05.572946 kernel: device-mapper: verity: sha256 using implementation "sha256-avx2" May 10 00:40:05.612598 systemd[1]: Found device dev-mapper-usr.device. May 10 00:40:05.612000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=verity-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:05.613710 systemd[1]: Mounting sysusr-usr.mount... May 10 00:40:05.613924 systemd[1]: Finished verity-setup.service. May 10 00:40:05.665285 systemd[1]: Mounted sysusr-usr.mount. May 10 00:40:05.667020 kernel: EXT4-fs (dm-0): mounted filesystem without journal. Opts: norecovery. Quota mode: none. May 10 00:40:05.665891 systemd[1]: Starting afterburn-network-kargs.service... May 10 00:40:05.666376 systemd[1]: Starting ignition-setup.service... May 10 00:40:05.680515 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm May 10 00:40:05.680546 kernel: BTRFS info (device sda6): using free space tree May 10 00:40:05.680555 kernel: BTRFS info (device sda6): has skinny extents May 10 00:40:05.685921 kernel: BTRFS info (device sda6): enabling ssd optimizations May 10 00:40:05.692198 systemd[1]: mnt-oem.mount: Deactivated successfully. May 10 00:40:05.695000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:05.697501 systemd[1]: Finished ignition-setup.service. May 10 00:40:05.698040 systemd[1]: Starting ignition-fetch-offline.service... May 10 00:40:05.751938 systemd[1]: Finished afterburn-network-kargs.service. May 10 00:40:05.752939 systemd[1]: Starting parse-ip-for-networkd.service... May 10 00:40:05.750000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=afterburn-network-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:05.802341 systemd[1]: Finished parse-ip-for-networkd.service. May 10 00:40:05.803222 systemd[1]: Starting systemd-networkd.service... May 10 00:40:05.800000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:05.801000 audit: BPF prog-id=9 op=LOAD May 10 00:40:05.817014 systemd-networkd[732]: lo: Link UP May 10 00:40:05.817019 systemd-networkd[732]: lo: Gained carrier May 10 00:40:05.817295 systemd-networkd[732]: Enumeration completed May 10 00:40:05.817559 systemd[1]: Started systemd-networkd.service. May 10 00:40:05.821230 kernel: vmxnet3 0000:0b:00.0 ens192: intr type 3, mode 0, 3 vectors allocated May 10 00:40:05.821332 kernel: vmxnet3 0000:0b:00.0 ens192: NIC Link is Up 10000 Mbps May 10 00:40:05.819000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:05.817676 systemd-networkd[732]: ens192: Configuring with /etc/systemd/network/10-dracut-cmdline-99.network. May 10 00:40:05.821081 systemd[1]: Reached target network.target. May 10 00:40:05.821695 systemd[1]: Starting iscsiuio.service... May 10 00:40:05.824861 systemd[1]: Started iscsiuio.service. May 10 00:40:05.823000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:05.825592 systemd[1]: Starting iscsid.service... May 10 00:40:05.826136 systemd-networkd[732]: ens192: Link UP May 10 00:40:05.826000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:05.828946 iscsid[737]: iscsid: can't open InitiatorName configuration file /etc/iscsi/initiatorname.iscsi May 10 00:40:05.828946 iscsid[737]: iscsid: Warning: InitiatorName file /etc/iscsi/initiatorname.iscsi does not exist or does not contain a properly formatted InitiatorName. If using software iscsi (iscsi_tcp or ib_iser) or partial offload (bnx2i or cxgbi iscsi), you may not be able to log May 10 00:40:05.828946 iscsid[737]: into or discover targets. Please create a file /etc/iscsi/initiatorname.iscsi that contains a sting with the format: InitiatorName=iqn.yyyy-mm.[:identifier]. May 10 00:40:05.828946 iscsid[737]: Example: InitiatorName=iqn.2001-04.com.redhat:fc6. May 10 00:40:05.828946 iscsid[737]: If using hardware iscsi like qla4xxx this message can be ignored. May 10 00:40:05.828946 iscsid[737]: iscsid: can't open InitiatorAlias configuration file /etc/iscsi/initiatorname.iscsi May 10 00:40:05.828946 iscsid[737]: iscsid: can't open iscsid.safe_logout configuration file /etc/iscsi/iscsid.conf May 10 00:40:05.826138 systemd-networkd[732]: ens192: Gained carrier May 10 00:40:05.828404 systemd[1]: Started iscsid.service. May 10 00:40:05.829224 systemd[1]: Starting dracut-initqueue.service... May 10 00:40:05.837194 systemd[1]: Finished dracut-initqueue.service. May 10 00:40:05.835000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:05.837546 systemd[1]: Reached target remote-fs-pre.target. May 10 00:40:05.837777 systemd[1]: Reached target remote-cryptsetup.target. May 10 00:40:05.837995 systemd[1]: Reached target remote-fs.target. May 10 00:40:05.838681 systemd[1]: Starting dracut-pre-mount.service... May 10 00:40:05.845699 systemd[1]: Finished dracut-pre-mount.service. May 10 00:40:05.843000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:05.884418 ignition[604]: Ignition 2.14.0 May 10 00:40:05.884429 ignition[604]: Stage: fetch-offline May 10 00:40:05.884490 ignition[604]: reading system config file "/usr/lib/ignition/base.d/base.ign" May 10 00:40:05.884514 ignition[604]: parsing config with SHA512: bd85a898f7da4744ff98e02742aa4854e1ceea8026a4e95cb6fb599b39b54cff0db353847df13d3c55ae196a9dc5d648977228d55e5da3ea20cd600fa7cec8ed May 10 00:40:05.890998 ignition[604]: no config dir at "/usr/lib/ignition/base.platform.d/vmware" May 10 00:40:05.891106 ignition[604]: parsed url from cmdline: "" May 10 00:40:05.891109 ignition[604]: no config URL provided May 10 00:40:05.891113 ignition[604]: reading system config file "/usr/lib/ignition/user.ign" May 10 00:40:05.891119 ignition[604]: no config at "/usr/lib/ignition/user.ign" May 10 00:40:05.898398 ignition[604]: config successfully fetched May 10 00:40:05.898439 ignition[604]: parsing config with SHA512: c0e125013b2714bbdf7d39d5c69d0a22a49eeb91a104dece750ee4a0828a1595fe7745764f2b0655b612481d78ff662cf4933d6ad8948bac5923702b30762203 May 10 00:40:05.901165 unknown[604]: fetched base config from "system" May 10 00:40:05.901173 unknown[604]: fetched user config from "vmware" May 10 00:40:05.901637 ignition[604]: fetch-offline: fetch-offline passed May 10 00:40:05.901703 ignition[604]: Ignition finished successfully May 10 00:40:05.902376 systemd[1]: Finished ignition-fetch-offline.service. May 10 00:40:05.900000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:05.902601 systemd[1]: ignition-fetch.service was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). May 10 00:40:05.903207 systemd[1]: Starting ignition-kargs.service... May 10 00:40:05.910158 ignition[752]: Ignition 2.14.0 May 10 00:40:05.910166 ignition[752]: Stage: kargs May 10 00:40:05.910235 ignition[752]: reading system config file "/usr/lib/ignition/base.d/base.ign" May 10 00:40:05.910246 ignition[752]: parsing config with SHA512: bd85a898f7da4744ff98e02742aa4854e1ceea8026a4e95cb6fb599b39b54cff0db353847df13d3c55ae196a9dc5d648977228d55e5da3ea20cd600fa7cec8ed May 10 00:40:05.911734 ignition[752]: no config dir at "/usr/lib/ignition/base.platform.d/vmware" May 10 00:40:05.913360 ignition[752]: kargs: kargs passed May 10 00:40:05.913388 ignition[752]: Ignition finished successfully May 10 00:40:05.912000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:05.914483 systemd[1]: Finished ignition-kargs.service. May 10 00:40:05.915114 systemd[1]: Starting ignition-disks.service... May 10 00:40:05.919526 ignition[758]: Ignition 2.14.0 May 10 00:40:05.919533 ignition[758]: Stage: disks May 10 00:40:05.919591 ignition[758]: reading system config file "/usr/lib/ignition/base.d/base.ign" May 10 00:40:05.919600 ignition[758]: parsing config with SHA512: bd85a898f7da4744ff98e02742aa4854e1ceea8026a4e95cb6fb599b39b54cff0db353847df13d3c55ae196a9dc5d648977228d55e5da3ea20cd600fa7cec8ed May 10 00:40:05.921298 ignition[758]: no config dir at "/usr/lib/ignition/base.platform.d/vmware" May 10 00:40:05.923163 ignition[758]: disks: disks passed May 10 00:40:05.923779 systemd[1]: Finished ignition-disks.service. May 10 00:40:05.923191 ignition[758]: Ignition finished successfully May 10 00:40:05.922000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:05.924214 systemd[1]: Reached target initrd-root-device.target. May 10 00:40:05.924456 systemd[1]: Reached target local-fs-pre.target. May 10 00:40:05.924663 systemd[1]: Reached target local-fs.target. May 10 00:40:05.924873 systemd[1]: Reached target sysinit.target. May 10 00:40:05.925060 systemd[1]: Reached target basic.target. May 10 00:40:05.925864 systemd[1]: Starting systemd-fsck-root.service... May 10 00:40:05.939724 systemd-fsck[766]: ROOT: clean, 623/1628000 files, 124060/1617920 blocks May 10 00:40:05.941606 systemd[1]: Finished systemd-fsck-root.service. May 10 00:40:05.939000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:05.942794 systemd[1]: Mounting sysroot.mount... May 10 00:40:05.947811 systemd[1]: Mounted sysroot.mount. May 10 00:40:05.947959 kernel: EXT4-fs (sda9): mounted filesystem with ordered data mode. Opts: (null). Quota mode: none. May 10 00:40:05.948092 systemd[1]: Reached target initrd-root-fs.target. May 10 00:40:05.949064 systemd[1]: Mounting sysroot-usr.mount... May 10 00:40:05.949563 systemd[1]: flatcar-metadata-hostname.service was skipped because no trigger condition checks were met. May 10 00:40:05.949748 systemd[1]: ignition-remount-sysroot.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). May 10 00:40:05.949999 systemd[1]: Reached target ignition-diskful.target. May 10 00:40:05.950758 systemd[1]: Mounted sysroot-usr.mount. May 10 00:40:05.951347 systemd[1]: Starting initrd-setup-root.service... May 10 00:40:05.954048 initrd-setup-root[776]: cut: /sysroot/etc/passwd: No such file or directory May 10 00:40:05.957698 initrd-setup-root[784]: cut: /sysroot/etc/group: No such file or directory May 10 00:40:05.959991 initrd-setup-root[792]: cut: /sysroot/etc/shadow: No such file or directory May 10 00:40:05.962328 initrd-setup-root[800]: cut: /sysroot/etc/gshadow: No such file or directory May 10 00:40:05.991004 systemd[1]: Finished initrd-setup-root.service. May 10 00:40:05.989000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:05.991563 systemd[1]: Starting ignition-mount.service... May 10 00:40:05.992008 systemd[1]: Starting sysroot-boot.service... May 10 00:40:05.995132 bash[817]: umount: /sysroot/usr/share/oem: not mounted. May 10 00:40:06.000095 ignition[818]: INFO : Ignition 2.14.0 May 10 00:40:06.000344 ignition[818]: INFO : Stage: mount May 10 00:40:06.000516 ignition[818]: INFO : reading system config file "/usr/lib/ignition/base.d/base.ign" May 10 00:40:06.000667 ignition[818]: DEBUG : parsing config with SHA512: bd85a898f7da4744ff98e02742aa4854e1ceea8026a4e95cb6fb599b39b54cff0db353847df13d3c55ae196a9dc5d648977228d55e5da3ea20cd600fa7cec8ed May 10 00:40:06.002192 ignition[818]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/vmware" May 10 00:40:06.003798 ignition[818]: INFO : mount: mount passed May 10 00:40:06.003946 ignition[818]: INFO : Ignition finished successfully May 10 00:40:06.004440 systemd[1]: Finished ignition-mount.service. May 10 00:40:06.002000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:06.008000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:06.010603 systemd[1]: Finished sysroot-boot.service. May 10 00:40:06.627439 systemd[1]: Mounting sysroot-usr-share-oem.mount... May 10 00:40:06.636945 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/sda6 scanned by mount (827) May 10 00:40:06.636971 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm May 10 00:40:06.636982 kernel: BTRFS info (device sda6): using free space tree May 10 00:40:06.638569 kernel: BTRFS info (device sda6): has skinny extents May 10 00:40:06.643356 kernel: BTRFS info (device sda6): enabling ssd optimizations May 10 00:40:06.642797 systemd[1]: Mounted sysroot-usr-share-oem.mount. May 10 00:40:06.643259 systemd[1]: Starting ignition-files.service... May 10 00:40:06.651867 ignition[847]: INFO : Ignition 2.14.0 May 10 00:40:06.651867 ignition[847]: INFO : Stage: files May 10 00:40:06.652197 ignition[847]: INFO : reading system config file "/usr/lib/ignition/base.d/base.ign" May 10 00:40:06.652197 ignition[847]: DEBUG : parsing config with SHA512: bd85a898f7da4744ff98e02742aa4854e1ceea8026a4e95cb6fb599b39b54cff0db353847df13d3c55ae196a9dc5d648977228d55e5da3ea20cd600fa7cec8ed May 10 00:40:06.653311 ignition[847]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/vmware" May 10 00:40:06.655443 ignition[847]: DEBUG : files: compiled without relabeling support, skipping May 10 00:40:06.655842 ignition[847]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" May 10 00:40:06.655842 ignition[847]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" May 10 00:40:06.658178 ignition[847]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" May 10 00:40:06.658395 ignition[847]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" May 10 00:40:06.658899 unknown[847]: wrote ssh authorized keys file for user: core May 10 00:40:06.659113 ignition[847]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" May 10 00:40:06.659888 ignition[847]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/etc/flatcar-cgroupv1" May 10 00:40:06.659888 ignition[847]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/etc/flatcar-cgroupv1" May 10 00:40:06.659888 ignition[847]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" May 10 00:40:06.659888 ignition[847]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://get.helm.sh/helm-v3.13.2-linux-amd64.tar.gz: attempt #1 May 10 00:40:06.698086 ignition[847]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK May 10 00:40:06.843469 ignition[847]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" May 10 00:40:06.844341 ignition[847]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" May 10 00:40:06.844570 ignition[847]: INFO : files: createFilesystemsFiles: createFiles: op(5): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1 May 10 00:40:07.338594 ignition[847]: INFO : files: createFilesystemsFiles: createFiles: op(5): GET result: OK May 10 00:40:07.387022 ignition[847]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" May 10 00:40:07.387263 ignition[847]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/install.sh" May 10 00:40:07.387263 ignition[847]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/install.sh" May 10 00:40:07.387263 ignition[847]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nginx.yaml" May 10 00:40:07.387263 ignition[847]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nginx.yaml" May 10 00:40:07.387871 ignition[847]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pod.yaml" May 10 00:40:07.387871 ignition[847]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" May 10 00:40:07.387871 ignition[847]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" May 10 00:40:07.388848 ignition[847]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" May 10 00:40:07.394124 ignition[847]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/etc/flatcar/update.conf" May 10 00:40:07.394309 ignition[847]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/etc/flatcar/update.conf" May 10 00:40:07.394309 ignition[847]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" May 10 00:40:07.394309 ignition[847]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" May 10 00:40:07.398763 ignition[847]: INFO : files: createFilesystemsFiles: createFiles: op(c): [started] writing file "/sysroot/etc/systemd/system/vmtoolsd.service" May 10 00:40:07.398970 ignition[847]: INFO : files: createFilesystemsFiles: createFiles: op(c): oem config not found in "/usr/share/oem", looking on oem partition May 10 00:40:07.412713 ignition[847]: INFO : files: createFilesystemsFiles: createFiles: op(c): op(d): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem1265997792" May 10 00:40:07.412976 ignition[847]: CRITICAL : files: createFilesystemsFiles: createFiles: op(c): op(d): [failed] mounting "/dev/disk/by-label/OEM" at "/mnt/oem1265997792": device or resource busy May 10 00:40:07.412976 ignition[847]: ERROR : files: createFilesystemsFiles: createFiles: op(c): failed to mount ext4 device "/dev/disk/by-label/OEM" at "/mnt/oem1265997792", trying btrfs: device or resource busy May 10 00:40:07.412976 ignition[847]: INFO : files: createFilesystemsFiles: createFiles: op(c): op(e): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem1265997792" May 10 00:40:07.414700 ignition[847]: INFO : files: createFilesystemsFiles: createFiles: op(c): op(e): [finished] mounting "/dev/disk/by-label/OEM" at "/mnt/oem1265997792" May 10 00:40:07.418834 ignition[847]: INFO : files: createFilesystemsFiles: createFiles: op(c): op(f): [started] unmounting "/mnt/oem1265997792" May 10 00:40:07.419660 systemd[1]: mnt-oem1265997792.mount: Deactivated successfully. May 10 00:40:07.420001 ignition[847]: INFO : files: createFilesystemsFiles: createFiles: op(c): op(f): [finished] unmounting "/mnt/oem1265997792" May 10 00:40:07.420218 ignition[847]: INFO : files: createFilesystemsFiles: createFiles: op(c): [finished] writing file "/sysroot/etc/systemd/system/vmtoolsd.service" May 10 00:40:07.420218 ignition[847]: INFO : files: createFilesystemsFiles: createFiles: op(10): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" May 10 00:40:07.420218 ignition[847]: INFO : files: createFilesystemsFiles: createFiles: op(10): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.30.1-x86-64.raw: attempt #1 May 10 00:40:07.751248 systemd-networkd[732]: ens192: Gained IPv6LL May 10 00:40:07.843302 ignition[847]: INFO : files: createFilesystemsFiles: createFiles: op(10): GET result: OK May 10 00:40:08.102123 ignition[847]: INFO : files: createFilesystemsFiles: createFiles: op(10): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-x86-64.raw" May 10 00:40:08.102492 ignition[847]: INFO : files: createFilesystemsFiles: createFiles: op(11): [started] writing file "/sysroot/etc/systemd/network/00-vmware.network" May 10 00:40:08.103018 ignition[847]: INFO : files: createFilesystemsFiles: createFiles: op(11): [finished] writing file "/sysroot/etc/systemd/network/00-vmware.network" May 10 00:40:08.103018 ignition[847]: INFO : files: op(12): [started] processing unit "vmtoolsd.service" May 10 00:40:08.103018 ignition[847]: INFO : files: op(12): [finished] processing unit "vmtoolsd.service" May 10 00:40:08.103018 ignition[847]: INFO : files: op(13): [started] processing unit "containerd.service" May 10 00:40:08.103018 ignition[847]: INFO : files: op(13): op(14): [started] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" May 10 00:40:08.103018 ignition[847]: INFO : files: op(13): op(14): [finished] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" May 10 00:40:08.103018 ignition[847]: INFO : files: op(13): [finished] processing unit "containerd.service" May 10 00:40:08.103018 ignition[847]: INFO : files: op(15): [started] processing unit "prepare-helm.service" May 10 00:40:08.103018 ignition[847]: INFO : files: op(15): op(16): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" May 10 00:40:08.103018 ignition[847]: INFO : files: op(15): op(16): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" May 10 00:40:08.103018 ignition[847]: INFO : files: op(15): [finished] processing unit "prepare-helm.service" May 10 00:40:08.103018 ignition[847]: INFO : files: op(17): [started] processing unit "coreos-metadata.service" May 10 00:40:08.103018 ignition[847]: INFO : files: op(17): op(18): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" May 10 00:40:08.103018 ignition[847]: INFO : files: op(17): op(18): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" May 10 00:40:08.103018 ignition[847]: INFO : files: op(17): [finished] processing unit "coreos-metadata.service" May 10 00:40:08.103018 ignition[847]: INFO : files: op(19): [started] setting preset to enabled for "prepare-helm.service" May 10 00:40:08.106661 ignition[847]: INFO : files: op(19): [finished] setting preset to enabled for "prepare-helm.service" May 10 00:40:08.106661 ignition[847]: INFO : files: op(1a): [started] setting preset to disabled for "coreos-metadata.service" May 10 00:40:08.106661 ignition[847]: INFO : files: op(1a): op(1b): [started] removing enablement symlink(s) for "coreos-metadata.service" May 10 00:40:08.184300 ignition[847]: INFO : files: op(1a): op(1b): [finished] removing enablement symlink(s) for "coreos-metadata.service" May 10 00:40:08.184603 ignition[847]: INFO : files: op(1a): [finished] setting preset to disabled for "coreos-metadata.service" May 10 00:40:08.184603 ignition[847]: INFO : files: op(1c): [started] setting preset to enabled for "vmtoolsd.service" May 10 00:40:08.184603 ignition[847]: INFO : files: op(1c): [finished] setting preset to enabled for "vmtoolsd.service" May 10 00:40:08.184603 ignition[847]: INFO : files: createResultFile: createFiles: op(1d): [started] writing file "/sysroot/etc/.ignition-result.json" May 10 00:40:08.184603 ignition[847]: INFO : files: createResultFile: createFiles: op(1d): [finished] writing file "/sysroot/etc/.ignition-result.json" May 10 00:40:08.184603 ignition[847]: INFO : files: files passed May 10 00:40:08.184603 ignition[847]: INFO : Ignition finished successfully May 10 00:40:08.191848 kernel: kauditd_printk_skb: 24 callbacks suppressed May 10 00:40:08.191872 kernel: audit: type=1130 audit(1746837608.184:35): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.184000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.186111 systemd[1]: Finished ignition-files.service. May 10 00:40:08.187792 systemd[1]: Starting initrd-setup-root-after-ignition.service... May 10 00:40:08.191713 systemd[1]: torcx-profile-populate.service was skipped because of an unmet condition check (ConditionPathExists=/sysroot/etc/torcx/next-profile). May 10 00:40:08.193056 systemd[1]: Starting ignition-quench.service... May 10 00:40:08.195514 systemd[1]: ignition-quench.service: Deactivated successfully. May 10 00:40:08.195602 systemd[1]: Finished ignition-quench.service. May 10 00:40:08.196574 initrd-setup-root-after-ignition[873]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory May 10 00:40:08.203034 kernel: audit: type=1130 audit(1746837608.194:36): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.203057 kernel: audit: type=1131 audit(1746837608.194:37): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.194000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.194000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.197079 systemd[1]: Finished initrd-setup-root-after-ignition.service. May 10 00:40:08.201000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.203500 systemd[1]: Reached target ignition-complete.target. May 10 00:40:08.206882 systemd[1]: Starting initrd-parse-etc.service... May 10 00:40:08.207024 kernel: audit: type=1130 audit(1746837608.201:38): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.215874 systemd[1]: initrd-parse-etc.service: Deactivated successfully. May 10 00:40:08.216133 systemd[1]: Finished initrd-parse-etc.service. May 10 00:40:08.214000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.216486 systemd[1]: Reached target initrd-fs.target. May 10 00:40:08.221331 kernel: audit: type=1130 audit(1746837608.214:39): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.221351 kernel: audit: type=1131 audit(1746837608.214:40): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.214000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.221493 systemd[1]: Reached target initrd.target. May 10 00:40:08.221736 systemd[1]: dracut-mount.service was skipped because no trigger condition checks were met. May 10 00:40:08.222447 systemd[1]: Starting dracut-pre-pivot.service... May 10 00:40:08.228925 systemd[1]: Finished dracut-pre-pivot.service. May 10 00:40:08.227000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.229559 systemd[1]: Starting initrd-cleanup.service... May 10 00:40:08.232033 kernel: audit: type=1130 audit(1746837608.227:41): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.236344 systemd[1]: initrd-cleanup.service: Deactivated successfully. May 10 00:40:08.236394 systemd[1]: Finished initrd-cleanup.service. May 10 00:40:08.234000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.236932 systemd[1]: Stopped target nss-lookup.target. May 10 00:40:08.241383 kernel: audit: type=1130 audit(1746837608.234:42): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.241395 kernel: audit: type=1131 audit(1746837608.234:43): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.234000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.241293 systemd[1]: Stopped target remote-cryptsetup.target. May 10 00:40:08.241454 systemd[1]: Stopped target timers.target. May 10 00:40:08.241641 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. May 10 00:40:08.244151 kernel: audit: type=1131 audit(1746837608.239:44): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.239000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.241667 systemd[1]: Stopped dracut-pre-pivot.service. May 10 00:40:08.241792 systemd[1]: Stopped target initrd.target. May 10 00:40:08.244206 systemd[1]: Stopped target basic.target. May 10 00:40:08.244375 systemd[1]: Stopped target ignition-complete.target. May 10 00:40:08.244539 systemd[1]: Stopped target ignition-diskful.target. May 10 00:40:08.244694 systemd[1]: Stopped target initrd-root-device.target. May 10 00:40:08.244866 systemd[1]: Stopped target remote-fs.target. May 10 00:40:08.245042 systemd[1]: Stopped target remote-fs-pre.target. May 10 00:40:08.245212 systemd[1]: Stopped target sysinit.target. May 10 00:40:08.245366 systemd[1]: Stopped target local-fs.target. May 10 00:40:08.245526 systemd[1]: Stopped target local-fs-pre.target. May 10 00:40:08.245687 systemd[1]: Stopped target swap.target. May 10 00:40:08.244000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.245844 systemd[1]: dracut-pre-mount.service: Deactivated successfully. May 10 00:40:08.245869 systemd[1]: Stopped dracut-pre-mount.service. May 10 00:40:08.244000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.246036 systemd[1]: Stopped target cryptsetup.target. May 10 00:40:08.244000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.246173 systemd[1]: dracut-initqueue.service: Deactivated successfully. May 10 00:40:08.246195 systemd[1]: Stopped dracut-initqueue.service. May 10 00:40:08.246367 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. May 10 00:40:08.246387 systemd[1]: Stopped ignition-fetch-offline.service. May 10 00:40:08.246518 systemd[1]: Stopped target paths.target. May 10 00:40:08.246657 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. May 10 00:40:08.251051 systemd[1]: Stopped systemd-ask-password-console.path. May 10 00:40:08.251174 systemd[1]: Stopped target slices.target. May 10 00:40:08.251513 systemd[1]: Stopped target sockets.target. May 10 00:40:08.251707 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. May 10 00:40:08.249000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.251733 systemd[1]: Stopped initrd-setup-root-after-ignition.service. May 10 00:40:08.251866 systemd[1]: ignition-files.service: Deactivated successfully. May 10 00:40:08.251886 systemd[1]: Stopped ignition-files.service. May 10 00:40:08.252430 systemd[1]: Stopping ignition-mount.service... May 10 00:40:08.252627 systemd[1]: Stopping iscsid.service... May 10 00:40:08.254427 iscsid[737]: iscsid shutting down. May 10 00:40:08.253055 systemd[1]: Stopping sysroot-boot.service... May 10 00:40:08.253165 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. May 10 00:40:08.253202 systemd[1]: Stopped systemd-udev-trigger.service. May 10 00:40:08.253329 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. May 10 00:40:08.253350 systemd[1]: Stopped dracut-pre-trigger.service. May 10 00:40:08.255591 systemd[1]: iscsid.service: Deactivated successfully. May 10 00:40:08.255872 systemd[1]: Stopped iscsid.service. May 10 00:40:08.250000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.251000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.251000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.254000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.256049 systemd[1]: iscsid.socket: Deactivated successfully. May 10 00:40:08.256065 systemd[1]: Closed iscsid.socket. May 10 00:40:08.258011 systemd[1]: Stopping iscsiuio.service... May 10 00:40:08.258000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.260443 systemd[1]: iscsiuio.service: Deactivated successfully. May 10 00:40:08.260492 systemd[1]: Stopped iscsiuio.service. May 10 00:40:08.262393 ignition[886]: INFO : Ignition 2.14.0 May 10 00:40:08.262393 ignition[886]: INFO : Stage: umount May 10 00:40:08.262393 ignition[886]: INFO : reading system config file "/usr/lib/ignition/base.d/base.ign" May 10 00:40:08.262393 ignition[886]: DEBUG : parsing config with SHA512: bd85a898f7da4744ff98e02742aa4854e1ceea8026a4e95cb6fb599b39b54cff0db353847df13d3c55ae196a9dc5d648977228d55e5da3ea20cd600fa7cec8ed May 10 00:40:08.260753 systemd[1]: iscsiuio.socket: Deactivated successfully. May 10 00:40:08.265114 ignition[886]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/vmware" May 10 00:40:08.265114 ignition[886]: INFO : umount: umount passed May 10 00:40:08.265114 ignition[886]: INFO : Ignition finished successfully May 10 00:40:08.260772 systemd[1]: Closed iscsiuio.socket. May 10 00:40:08.264000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.264000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.264000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.264000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.266029 systemd[1]: ignition-mount.service: Deactivated successfully. May 10 00:40:08.266085 systemd[1]: Stopped ignition-mount.service. May 10 00:40:08.266231 systemd[1]: Stopped target network.target. May 10 00:40:08.266317 systemd[1]: ignition-disks.service: Deactivated successfully. May 10 00:40:08.266342 systemd[1]: Stopped ignition-disks.service. May 10 00:40:08.266452 systemd[1]: ignition-kargs.service: Deactivated successfully. May 10 00:40:08.266471 systemd[1]: Stopped ignition-kargs.service. May 10 00:40:08.266573 systemd[1]: ignition-setup.service: Deactivated successfully. May 10 00:40:08.266592 systemd[1]: Stopped ignition-setup.service. May 10 00:40:08.267880 systemd[1]: Stopping systemd-networkd.service... May 10 00:40:08.268019 systemd[1]: Stopping systemd-resolved.service... May 10 00:40:08.268696 systemd[1]: sysroot-boot.mount: Deactivated successfully. May 10 00:40:08.270000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.272617 systemd[1]: systemd-networkd.service: Deactivated successfully. May 10 00:40:08.272666 systemd[1]: Stopped systemd-networkd.service. May 10 00:40:08.272840 systemd[1]: systemd-networkd.socket: Deactivated successfully. May 10 00:40:08.272858 systemd[1]: Closed systemd-networkd.socket. May 10 00:40:08.273359 systemd[1]: Stopping network-cleanup.service... May 10 00:40:08.275043 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. May 10 00:40:08.275082 systemd[1]: Stopped parse-ip-for-networkd.service. May 10 00:40:08.275209 systemd[1]: afterburn-network-kargs.service: Deactivated successfully. May 10 00:40:08.275238 systemd[1]: Stopped afterburn-network-kargs.service. May 10 00:40:08.275351 systemd[1]: systemd-sysctl.service: Deactivated successfully. May 10 00:40:08.273000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.273000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=afterburn-network-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.273000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.273000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.276000 audit: BPF prog-id=9 op=UNLOAD May 10 00:40:08.275375 systemd[1]: Stopped systemd-sysctl.service. May 10 00:40:08.275515 systemd[1]: systemd-modules-load.service: Deactivated successfully. May 10 00:40:08.275537 systemd[1]: Stopped systemd-modules-load.service. May 10 00:40:08.277776 systemd[1]: Stopping systemd-udevd.service... May 10 00:40:08.278852 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. May 10 00:40:08.279189 systemd[1]: systemd-resolved.service: Deactivated successfully. May 10 00:40:08.279244 systemd[1]: Stopped systemd-resolved.service. May 10 00:40:08.277000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.279000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.280886 systemd[1]: systemd-udevd.service: Deactivated successfully. May 10 00:40:08.281022 systemd[1]: Stopped systemd-udevd.service. May 10 00:40:08.280000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.280000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.281411 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. May 10 00:40:08.280000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.281439 systemd[1]: Closed systemd-udevd-control.socket. May 10 00:40:08.281561 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. May 10 00:40:08.281580 systemd[1]: Closed systemd-udevd-kernel.socket. May 10 00:40:08.282456 systemd[1]: dracut-pre-udev.service: Deactivated successfully. May 10 00:40:08.282485 systemd[1]: Stopped dracut-pre-udev.service. May 10 00:40:08.281000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.282667 systemd[1]: dracut-cmdline.service: Deactivated successfully. May 10 00:40:08.282689 systemd[1]: Stopped dracut-cmdline.service. May 10 00:40:08.282000 audit: BPF prog-id=6 op=UNLOAD May 10 00:40:08.282000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.282798 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. May 10 00:40:08.282000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.282817 systemd[1]: Stopped dracut-cmdline-ask.service. May 10 00:40:08.283325 systemd[1]: Starting initrd-udevadm-cleanup-db.service... May 10 00:40:08.283464 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. May 10 00:40:08.283493 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service. May 10 00:40:08.283747 systemd[1]: kmod-static-nodes.service: Deactivated successfully. May 10 00:40:08.283775 systemd[1]: Stopped kmod-static-nodes.service. May 10 00:40:08.284122 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 10 00:40:08.284153 systemd[1]: Stopped systemd-vconsole-setup.service. May 10 00:40:08.285024 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. May 10 00:40:08.285320 systemd[1]: network-cleanup.service: Deactivated successfully. May 10 00:40:08.285382 systemd[1]: Stopped network-cleanup.service. May 10 00:40:08.287000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=network-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.289122 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. May 10 00:40:08.289175 systemd[1]: Finished initrd-udevadm-cleanup-db.service. May 10 00:40:08.287000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.287000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.370000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.370000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:08.372327 systemd[1]: sysroot-boot.service: Deactivated successfully. May 10 00:40:08.372419 systemd[1]: Stopped sysroot-boot.service. May 10 00:40:08.372631 systemd[1]: Reached target initrd-switch-root.target. May 10 00:40:08.372751 systemd[1]: initrd-setup-root.service: Deactivated successfully. May 10 00:40:08.372784 systemd[1]: Stopped initrd-setup-root.service. May 10 00:40:08.373490 systemd[1]: Starting initrd-switch-root.service... May 10 00:40:08.396648 systemd[1]: Switching root. May 10 00:40:08.398000 audit: BPF prog-id=8 op=UNLOAD May 10 00:40:08.398000 audit: BPF prog-id=7 op=UNLOAD May 10 00:40:08.399000 audit: BPF prog-id=5 op=UNLOAD May 10 00:40:08.399000 audit: BPF prog-id=4 op=UNLOAD May 10 00:40:08.399000 audit: BPF prog-id=3 op=UNLOAD May 10 00:40:08.414190 systemd-journald[216]: Journal stopped May 10 00:40:11.706896 systemd-journald[216]: Received SIGTERM from PID 1 (systemd). May 10 00:40:11.706938 kernel: SELinux: Class mctp_socket not defined in policy. May 10 00:40:11.707399 kernel: SELinux: Class anon_inode not defined in policy. May 10 00:40:11.707407 kernel: SELinux: the above unknown classes and permissions will be allowed May 10 00:40:11.707412 kernel: SELinux: policy capability network_peer_controls=1 May 10 00:40:11.707422 kernel: SELinux: policy capability open_perms=1 May 10 00:40:11.707428 kernel: SELinux: policy capability extended_socket_class=1 May 10 00:40:11.707434 kernel: SELinux: policy capability always_check_network=0 May 10 00:40:11.707440 kernel: SELinux: policy capability cgroup_seclabel=1 May 10 00:40:11.707446 kernel: SELinux: policy capability nnp_nosuid_transition=1 May 10 00:40:11.707451 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 May 10 00:40:11.707457 kernel: SELinux: policy capability ioctl_skip_cloexec=0 May 10 00:40:11.707465 systemd[1]: Successfully loaded SELinux policy in 65.019ms. May 10 00:40:11.707473 systemd[1]: Relabelled /dev, /dev/shm, /run, /sys/fs/cgroup in 7.786ms. May 10 00:40:11.707481 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) May 10 00:40:11.707487 systemd[1]: Detected virtualization vmware. May 10 00:40:11.707495 systemd[1]: Detected architecture x86-64. May 10 00:40:11.707501 systemd[1]: Detected first boot. May 10 00:40:11.707508 systemd[1]: Initializing machine ID from random generator. May 10 00:40:11.707514 kernel: SELinux: Context system_u:object_r:container_file_t:s0:c1022,c1023 is not valid (left unmapped). May 10 00:40:11.707520 systemd[1]: Populated /etc with preset unit settings. May 10 00:40:11.707527 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. May 10 00:40:11.707534 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. May 10 00:40:11.707541 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 10 00:40:11.707549 systemd[1]: Queued start job for default target multi-user.target. May 10 00:40:11.707555 systemd[1]: Unnecessary job was removed for dev-sda6.device. May 10 00:40:11.707563 systemd[1]: Created slice system-addon\x2dconfig.slice. May 10 00:40:11.707570 systemd[1]: Created slice system-addon\x2drun.slice. May 10 00:40:11.707576 systemd[1]: Created slice system-getty.slice. May 10 00:40:11.707583 systemd[1]: Created slice system-modprobe.slice. May 10 00:40:11.707589 systemd[1]: Created slice system-serial\x2dgetty.slice. May 10 00:40:11.707597 systemd[1]: Created slice system-system\x2dcloudinit.slice. May 10 00:40:11.707604 systemd[1]: Created slice system-systemd\x2dfsck.slice. May 10 00:40:11.707611 systemd[1]: Created slice user.slice. May 10 00:40:11.707617 systemd[1]: Started systemd-ask-password-console.path. May 10 00:40:11.707624 systemd[1]: Started systemd-ask-password-wall.path. May 10 00:40:11.707630 systemd[1]: Set up automount boot.automount. May 10 00:40:11.707637 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount. May 10 00:40:11.707643 systemd[1]: Reached target integritysetup.target. May 10 00:40:11.707650 systemd[1]: Reached target remote-cryptsetup.target. May 10 00:40:11.707659 systemd[1]: Reached target remote-fs.target. May 10 00:40:11.707666 systemd[1]: Reached target slices.target. May 10 00:40:11.707673 systemd[1]: Reached target swap.target. May 10 00:40:11.707680 systemd[1]: Reached target torcx.target. May 10 00:40:11.707687 systemd[1]: Reached target veritysetup.target. May 10 00:40:11.707694 systemd[1]: Listening on systemd-coredump.socket. May 10 00:40:11.707700 systemd[1]: Listening on systemd-initctl.socket. May 10 00:40:11.707708 systemd[1]: Listening on systemd-journald-audit.socket. May 10 00:40:11.707716 systemd[1]: Listening on systemd-journald-dev-log.socket. May 10 00:40:11.707723 systemd[1]: Listening on systemd-journald.socket. May 10 00:40:11.707729 systemd[1]: Listening on systemd-networkd.socket. May 10 00:40:11.707736 systemd[1]: Listening on systemd-udevd-control.socket. May 10 00:40:11.707743 systemd[1]: Listening on systemd-udevd-kernel.socket. May 10 00:40:11.707750 systemd[1]: Listening on systemd-userdbd.socket. May 10 00:40:11.707758 systemd[1]: Mounting dev-hugepages.mount... May 10 00:40:11.707765 systemd[1]: Mounting dev-mqueue.mount... May 10 00:40:11.707772 systemd[1]: Mounting media.mount... May 10 00:40:11.707779 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). May 10 00:40:11.707786 systemd[1]: Mounting sys-kernel-debug.mount... May 10 00:40:11.707793 systemd[1]: Mounting sys-kernel-tracing.mount... May 10 00:40:11.707800 systemd[1]: Mounting tmp.mount... May 10 00:40:11.707808 systemd[1]: Starting flatcar-tmpfiles.service... May 10 00:40:11.707815 systemd[1]: Starting ignition-delete-config.service... May 10 00:40:11.707821 systemd[1]: Starting kmod-static-nodes.service... May 10 00:40:11.707828 systemd[1]: Starting modprobe@configfs.service... May 10 00:40:11.707835 systemd[1]: Starting modprobe@dm_mod.service... May 10 00:40:11.707842 systemd[1]: Starting modprobe@drm.service... May 10 00:40:11.707849 systemd[1]: Starting modprobe@efi_pstore.service... May 10 00:40:11.707857 systemd[1]: Starting modprobe@fuse.service... May 10 00:40:11.707864 systemd[1]: Starting modprobe@loop.service... May 10 00:40:11.707872 systemd[1]: setup-nsswitch.service was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). May 10 00:40:11.707879 systemd[1]: systemd-journald.service: unit configures an IP firewall, but the local system does not support BPF/cgroup firewalling. May 10 00:40:11.707886 systemd[1]: (This warning is only shown for the first unit using IP firewalling.) May 10 00:40:11.707893 systemd[1]: Starting systemd-journald.service... May 10 00:40:11.707900 kernel: fuse: init (API version 7.34) May 10 00:40:11.707917 systemd[1]: Starting systemd-modules-load.service... May 10 00:40:11.707926 kernel: loop: module loaded May 10 00:40:11.707933 systemd[1]: Starting systemd-network-generator.service... May 10 00:40:11.707958 systemd[1]: Starting systemd-remount-fs.service... May 10 00:40:11.707966 systemd[1]: Starting systemd-udev-trigger.service... May 10 00:40:11.707973 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). May 10 00:40:11.707980 systemd[1]: Mounted dev-hugepages.mount. May 10 00:40:11.707987 systemd[1]: Mounted dev-mqueue.mount. May 10 00:40:11.707994 systemd[1]: Mounted media.mount. May 10 00:40:11.708001 systemd[1]: Mounted sys-kernel-debug.mount. May 10 00:40:11.708008 systemd[1]: Mounted sys-kernel-tracing.mount. May 10 00:40:11.708015 systemd[1]: Mounted tmp.mount. May 10 00:40:11.708023 systemd[1]: Finished kmod-static-nodes.service. May 10 00:40:11.708031 systemd[1]: modprobe@configfs.service: Deactivated successfully. May 10 00:40:11.708039 systemd[1]: Finished modprobe@configfs.service. May 10 00:40:11.708046 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 10 00:40:11.708053 systemd[1]: Finished modprobe@dm_mod.service. May 10 00:40:11.708060 systemd[1]: modprobe@drm.service: Deactivated successfully. May 10 00:40:11.708067 systemd[1]: Finished modprobe@drm.service. May 10 00:40:11.708074 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 10 00:40:11.708081 systemd[1]: Finished modprobe@efi_pstore.service. May 10 00:40:11.708090 systemd[1]: Finished flatcar-tmpfiles.service. May 10 00:40:11.708097 systemd[1]: modprobe@fuse.service: Deactivated successfully. May 10 00:40:11.708103 systemd[1]: Finished modprobe@fuse.service. May 10 00:40:11.708111 systemd[1]: modprobe@loop.service: Deactivated successfully. May 10 00:40:11.708118 systemd[1]: Finished modprobe@loop.service. May 10 00:40:11.708129 systemd-journald[1049]: Journal started May 10 00:40:11.708158 systemd-journald[1049]: Runtime Journal (/run/log/journal/c1218c1b94cf4d568e5757d7d8c85bd0) is 4.8M, max 38.8M, 34.0M free. May 10 00:40:11.689000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:11.691000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:11.691000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:11.693000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:11.693000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:11.697000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:11.697000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:11.701000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:11.701000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:11.702000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=flatcar-tmpfiles comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:11.703000 audit: CONFIG_CHANGE op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 res=1 May 10 00:40:11.703000 audit[1049]: SYSCALL arch=c000003e syscall=46 success=yes exit=60 a0=6 a1=7ffeef331180 a2=4000 a3=7ffeef33121c items=0 ppid=1 pid=1049 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" subj=system_u:system_r:kernel_t:s0 key=(null) May 10 00:40:11.703000 audit: PROCTITLE proctitle="/usr/lib/systemd/systemd-journald" May 10 00:40:11.703000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:11.703000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:11.706000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:11.706000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:11.708771 jq[1017]: true May 10 00:40:11.709459 systemd[1]: Started systemd-journald.service. May 10 00:40:11.708000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:11.710366 systemd[1]: Finished systemd-network-generator.service. May 10 00:40:11.708000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:11.710611 systemd[1]: Finished systemd-remount-fs.service. May 10 00:40:11.708000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-remount-fs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:11.711041 systemd[1]: Reached target network-pre.target. May 10 00:40:11.711970 systemd[1]: Mounting sys-fs-fuse-connections.mount... May 10 00:40:11.714259 systemd[1]: Mounting sys-kernel-config.mount... May 10 00:40:11.714392 systemd[1]: remount-root.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). May 10 00:40:11.719259 jq[1068]: true May 10 00:40:11.729133 systemd[1]: Starting systemd-hwdb-update.service... May 10 00:40:11.730073 systemd[1]: Starting systemd-journal-flush.service... May 10 00:40:11.730200 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 10 00:40:11.731286 systemd[1]: Starting systemd-random-seed.service... May 10 00:40:11.731411 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. May 10 00:40:11.732258 systemd[1]: Starting systemd-sysusers.service... May 10 00:40:11.734266 systemd[1]: Mounted sys-fs-fuse-connections.mount. May 10 00:40:11.734424 systemd[1]: Mounted sys-kernel-config.mount. May 10 00:40:11.746865 systemd-journald[1049]: Time spent on flushing to /var/log/journal/c1218c1b94cf4d568e5757d7d8c85bd0 is 53.634ms for 1941 entries. May 10 00:40:11.746865 systemd-journald[1049]: System Journal (/var/log/journal/c1218c1b94cf4d568e5757d7d8c85bd0) is 8.0M, max 584.8M, 576.8M free. May 10 00:40:11.853194 systemd-journald[1049]: Received client request to flush runtime journal. May 10 00:40:11.747000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:11.754000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-random-seed comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:11.782000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:11.789000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysusers comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:11.823000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:11.748873 systemd[1]: Finished systemd-modules-load.service. May 10 00:40:11.749957 systemd[1]: Starting systemd-sysctl.service... May 10 00:40:11.755993 systemd[1]: Finished systemd-random-seed.service. May 10 00:40:11.756152 systemd[1]: Reached target first-boot-complete.target. May 10 00:40:11.853871 udevadm[1104]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. May 10 00:40:11.784464 systemd[1]: Finished systemd-sysctl.service. May 10 00:40:11.791078 systemd[1]: Finished systemd-sysusers.service. May 10 00:40:11.792240 systemd[1]: Starting systemd-tmpfiles-setup-dev.service... May 10 00:40:11.825451 systemd[1]: Finished systemd-udev-trigger.service. May 10 00:40:11.826526 systemd[1]: Starting systemd-udev-settle.service... May 10 00:40:11.854645 systemd[1]: Finished systemd-journal-flush.service. May 10 00:40:11.852000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-flush comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:11.888969 systemd[1]: Finished systemd-tmpfiles-setup-dev.service. May 10 00:40:11.887000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:11.966403 ignition[1088]: Ignition 2.14.0 May 10 00:40:11.966831 ignition[1088]: deleting config from guestinfo properties May 10 00:40:11.969211 ignition[1088]: Successfully deleted config May 10 00:40:11.969895 systemd[1]: Finished ignition-delete-config.service. May 10 00:40:11.968000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=ignition-delete-config comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:12.341622 systemd[1]: Finished systemd-hwdb-update.service. May 10 00:40:12.343405 systemd[1]: Starting systemd-udevd.service... May 10 00:40:12.339000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hwdb-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:12.357519 systemd-udevd[1109]: Using default interface naming scheme 'v252'. May 10 00:40:12.376547 systemd[1]: Started systemd-udevd.service. May 10 00:40:12.374000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:12.377794 systemd[1]: Starting systemd-networkd.service... May 10 00:40:12.384009 systemd[1]: Starting systemd-userdbd.service... May 10 00:40:12.408447 systemd[1]: Found device dev-ttyS0.device. May 10 00:40:12.415965 systemd[1]: Started systemd-userdbd.service. May 10 00:40:12.414000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-userdbd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:12.445930 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input2 May 10 00:40:12.451018 kernel: ACPI: button: Power Button [PWRF] May 10 00:40:12.471744 systemd-networkd[1110]: lo: Link UP May 10 00:40:12.471750 systemd-networkd[1110]: lo: Gained carrier May 10 00:40:12.472428 systemd-networkd[1110]: Enumeration completed May 10 00:40:12.472496 systemd-networkd[1110]: ens192: Configuring with /etc/systemd/network/00-vmware.network. May 10 00:40:12.472502 systemd[1]: Started systemd-networkd.service. May 10 00:40:12.470000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:12.475440 kernel: vmxnet3 0000:0b:00.0 ens192: intr type 3, mode 0, 3 vectors allocated May 10 00:40:12.475556 kernel: vmxnet3 0000:0b:00.0 ens192: NIC Link is Up 10000 Mbps May 10 00:40:12.476642 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): ens192: link becomes ready May 10 00:40:12.476922 systemd-networkd[1110]: ens192: Link UP May 10 00:40:12.477049 systemd-networkd[1110]: ens192: Gained carrier May 10 00:40:12.523000 audit[1122]: AVC avc: denied { confidentiality } for pid=1122 comm="(udev-worker)" lockdown_reason="use of tracefs" scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=lockdown permissive=1 May 10 00:40:12.523000 audit[1122]: SYSCALL arch=c000003e syscall=175 success=yes exit=0 a0=56521fdabbb0 a1=338ac a2=7fd7915f4bc5 a3=5 items=110 ppid=1109 pid=1122 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="(udev-worker)" exe="/usr/bin/udevadm" subj=system_u:system_r:kernel_t:s0 key=(null) May 10 00:40:12.523000 audit: CWD cwd="/" May 10 00:40:12.523000 audit: PATH item=0 name=(null) inode=45 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=1 name=(null) inode=25182 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=2 name=(null) inode=25182 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=3 name=(null) inode=25183 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=4 name=(null) inode=25182 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=5 name=(null) inode=25184 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=6 name=(null) inode=25182 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=7 name=(null) inode=25185 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=8 name=(null) inode=25185 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=9 name=(null) inode=25186 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=10 name=(null) inode=25185 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=11 name=(null) inode=25187 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=12 name=(null) inode=25185 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=13 name=(null) inode=25188 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=14 name=(null) inode=25185 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=15 name=(null) inode=25189 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=16 name=(null) inode=25185 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=17 name=(null) inode=25190 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=18 name=(null) inode=25182 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=19 name=(null) inode=25191 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=20 name=(null) inode=25191 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=21 name=(null) inode=25192 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=22 name=(null) inode=25191 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=23 name=(null) inode=25193 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=24 name=(null) inode=25191 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=25 name=(null) inode=25194 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=26 name=(null) inode=25191 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=27 name=(null) inode=25195 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=28 name=(null) inode=25191 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=29 name=(null) inode=25196 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=30 name=(null) inode=25182 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=31 name=(null) inode=25197 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=32 name=(null) inode=25197 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=33 name=(null) inode=25198 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=34 name=(null) inode=25197 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=35 name=(null) inode=25199 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=36 name=(null) inode=25197 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=37 name=(null) inode=25200 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=38 name=(null) inode=25197 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=39 name=(null) inode=25201 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=40 name=(null) inode=25197 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=41 name=(null) inode=25202 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=42 name=(null) inode=25182 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=43 name=(null) inode=25203 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=44 name=(null) inode=25203 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=45 name=(null) inode=25204 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=46 name=(null) inode=25203 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.529396 kernel: vmw_vmci 0000:00:07.7: Found VMCI PCI device at 0x11080, irq 16 May 10 00:40:12.530092 kernel: vmw_vmci 0000:00:07.7: Using capabilities 0xc May 10 00:40:12.530199 kernel: Guest personality initialized and is active May 10 00:40:12.523000 audit: PATH item=47 name=(null) inode=25205 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=48 name=(null) inode=25203 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=49 name=(null) inode=25206 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=50 name=(null) inode=25203 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=51 name=(null) inode=25207 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=52 name=(null) inode=25203 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=53 name=(null) inode=25208 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=54 name=(null) inode=45 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=55 name=(null) inode=25209 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=56 name=(null) inode=25209 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=57 name=(null) inode=25210 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=58 name=(null) inode=25209 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=59 name=(null) inode=25211 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=60 name=(null) inode=25209 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=61 name=(null) inode=25212 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=62 name=(null) inode=25212 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=63 name=(null) inode=25213 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=64 name=(null) inode=25212 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=65 name=(null) inode=25214 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=66 name=(null) inode=25212 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=67 name=(null) inode=25215 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=68 name=(null) inode=25212 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=69 name=(null) inode=25216 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=70 name=(null) inode=25212 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=71 name=(null) inode=25217 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=72 name=(null) inode=25209 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=73 name=(null) inode=25218 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=74 name=(null) inode=25218 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=75 name=(null) inode=25219 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=76 name=(null) inode=25218 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=77 name=(null) inode=25220 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=78 name=(null) inode=25218 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=79 name=(null) inode=25221 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=80 name=(null) inode=25218 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=81 name=(null) inode=25222 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=82 name=(null) inode=25218 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=83 name=(null) inode=25223 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=84 name=(null) inode=25209 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=85 name=(null) inode=25224 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=86 name=(null) inode=25224 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=87 name=(null) inode=25225 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=88 name=(null) inode=25224 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=89 name=(null) inode=25226 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=90 name=(null) inode=25224 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=91 name=(null) inode=25227 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=92 name=(null) inode=25224 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=93 name=(null) inode=25228 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=94 name=(null) inode=25224 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=95 name=(null) inode=25229 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=96 name=(null) inode=25209 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=97 name=(null) inode=25230 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=98 name=(null) inode=25230 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=99 name=(null) inode=25231 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=100 name=(null) inode=25230 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=101 name=(null) inode=25232 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=102 name=(null) inode=25230 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=103 name=(null) inode=25233 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=104 name=(null) inode=25230 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=105 name=(null) inode=25234 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=106 name=(null) inode=25230 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=107 name=(null) inode=25235 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=108 name=(null) inode=1 dev=00:07 mode=040700 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:debugfs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PATH item=109 name=(null) inode=25236 dev=00:07 mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:debugfs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 10 00:40:12.523000 audit: PROCTITLE proctitle="(udev-worker)" May 10 00:40:12.536934 kernel: piix4_smbus 0000:00:07.3: SMBus Host Controller not enabled! May 10 00:40:12.542842 kernel: VMCI host device registered (name=vmci, major=10, minor=125) May 10 00:40:12.542860 kernel: Initialized host personality May 10 00:40:12.554938 kernel: input: ImPS/2 Generic Wheel Mouse as /devices/platform/i8042/serio1/input/input3 May 10 00:40:12.557006 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. May 10 00:40:12.585924 kernel: mousedev: PS/2 mouse device common for all mice May 10 00:40:12.596754 (udev-worker)[1116]: id: Truncating stdout of 'dmi_memory_id' up to 16384 byte. May 10 00:40:12.607177 systemd[1]: Finished systemd-udev-settle.service. May 10 00:40:12.605000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-settle comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:12.608244 systemd[1]: Starting lvm2-activation-early.service... May 10 00:40:12.655082 lvm[1143]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. May 10 00:40:12.681504 systemd[1]: Finished lvm2-activation-early.service. May 10 00:40:12.679000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:12.681719 systemd[1]: Reached target cryptsetup.target. May 10 00:40:12.682874 systemd[1]: Starting lvm2-activation.service... May 10 00:40:12.685814 lvm[1145]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. May 10 00:40:12.714582 systemd[1]: Finished lvm2-activation.service. May 10 00:40:12.712000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:12.714751 systemd[1]: Reached target local-fs-pre.target. May 10 00:40:12.714850 systemd[1]: var-lib-machines.mount was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). May 10 00:40:12.714865 systemd[1]: Reached target local-fs.target. May 10 00:40:12.714966 systemd[1]: Reached target machines.target. May 10 00:40:12.715932 systemd[1]: Starting ldconfig.service... May 10 00:40:12.723648 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. May 10 00:40:12.723675 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). May 10 00:40:12.724426 systemd[1]: Starting systemd-boot-update.service... May 10 00:40:12.725160 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service... May 10 00:40:12.726098 systemd[1]: Starting systemd-machine-id-commit.service... May 10 00:40:12.727064 systemd[1]: Starting systemd-sysext.service... May 10 00:40:12.733978 systemd[1]: boot.automount: Got automount request for /boot, triggered by 1148 (bootctl) May 10 00:40:12.734782 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service... May 10 00:40:12.739810 systemd[1]: Unmounting usr-share-oem.mount... May 10 00:40:12.743051 systemd[1]: usr-share-oem.mount: Deactivated successfully. May 10 00:40:12.743196 systemd[1]: Unmounted usr-share-oem.mount. May 10 00:40:12.763921 kernel: loop0: detected capacity change from 0 to 210664 May 10 00:40:12.762000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-OEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:12.764137 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service. May 10 00:40:12.818000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-machine-id-commit comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:12.819992 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. May 10 00:40:12.820540 systemd[1]: Finished systemd-machine-id-commit.service. May 10 00:40:12.853929 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher May 10 00:40:12.873926 kernel: loop1: detected capacity change from 0 to 210664 May 10 00:40:12.935258 (sd-sysext)[1165]: Using extensions 'kubernetes'. May 10 00:40:12.935918 (sd-sysext)[1165]: Merged extensions into '/usr'. May 10 00:40:12.950613 systemd-fsck[1162]: fsck.fat 4.2 (2021-01-31) May 10 00:40:12.950613 systemd-fsck[1162]: /dev/sda1: 790 files, 120688/258078 clusters May 10 00:40:12.954007 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). May 10 00:40:12.955476 systemd[1]: Mounting usr-share-oem.mount... May 10 00:40:12.956676 systemd[1]: Starting modprobe@dm_mod.service... May 10 00:40:12.957512 systemd[1]: Starting modprobe@efi_pstore.service... May 10 00:40:12.958640 systemd[1]: Starting modprobe@loop.service... May 10 00:40:12.959262 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. May 10 00:40:12.959351 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). May 10 00:40:12.959439 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). May 10 00:40:12.960673 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service. May 10 00:40:12.959000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:12.961331 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 10 00:40:12.961415 systemd[1]: Finished modprobe@efi_pstore.service. May 10 00:40:12.959000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:12.959000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:12.963398 systemd[1]: Mounting boot.mount... May 10 00:40:12.964078 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 10 00:40:12.965000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:12.965000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:12.966851 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 10 00:40:12.966959 systemd[1]: Finished modprobe@dm_mod.service. May 10 00:40:12.970000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:12.970000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:12.970000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:12.971950 systemd[1]: Mounted usr-share-oem.mount. May 10 00:40:12.972212 systemd[1]: modprobe@loop.service: Deactivated successfully. May 10 00:40:12.972297 systemd[1]: Finished modprobe@loop.service. May 10 00:40:12.972551 systemd[1]: Finished systemd-sysext.service. May 10 00:40:12.974340 systemd[1]: Starting ensure-sysext.service... May 10 00:40:12.974523 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. May 10 00:40:12.975421 systemd[1]: Starting systemd-tmpfiles-setup.service... May 10 00:40:12.982145 systemd[1]: Reloading. May 10 00:40:12.988318 systemd-tmpfiles[1182]: /usr/lib/tmpfiles.d/legacy.conf:13: Duplicate line for path "/run/lock", ignoring. May 10 00:40:12.998058 systemd-tmpfiles[1182]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. May 10 00:40:13.001421 systemd-tmpfiles[1182]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. May 10 00:40:13.020028 /usr/lib/systemd/system-generators/torcx-generator[1201]: time="2025-05-10T00:40:13Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.7 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.7 /var/lib/torcx/store]" May 10 00:40:13.020254 /usr/lib/systemd/system-generators/torcx-generator[1201]: time="2025-05-10T00:40:13Z" level=info msg="torcx already run" May 10 00:40:13.092198 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. May 10 00:40:13.092210 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. May 10 00:40:13.105633 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 10 00:40:13.146143 systemd[1]: Mounted boot.mount. May 10 00:40:13.154006 systemd[1]: Starting modprobe@dm_mod.service... May 10 00:40:13.154970 systemd[1]: Starting modprobe@efi_pstore.service... May 10 00:40:13.156468 systemd[1]: Starting modprobe@loop.service... May 10 00:40:13.156647 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. May 10 00:40:13.156722 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). May 10 00:40:13.157315 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 10 00:40:13.157406 systemd[1]: Finished modprobe@dm_mod.service. May 10 00:40:13.155000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:13.155000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:13.157968 systemd[1]: modprobe@loop.service: Deactivated successfully. May 10 00:40:13.158274 systemd[1]: Finished modprobe@loop.service. May 10 00:40:13.156000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:13.156000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:13.158692 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. May 10 00:40:13.160241 systemd[1]: Starting modprobe@dm_mod.service... May 10 00:40:13.162485 systemd[1]: Starting modprobe@loop.service... May 10 00:40:13.162667 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. May 10 00:40:13.162758 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). May 10 00:40:13.163426 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 10 00:40:13.163514 systemd[1]: Finished modprobe@efi_pstore.service. May 10 00:40:13.161000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:13.161000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:13.164052 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 10 00:40:13.164131 systemd[1]: Finished modprobe@dm_mod.service. May 10 00:40:13.162000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:13.162000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:13.165044 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 10 00:40:13.169286 systemd[1]: Starting modprobe@dm_mod.service... May 10 00:40:13.170717 systemd[1]: Starting modprobe@drm.service... May 10 00:40:13.171730 systemd[1]: Starting modprobe@efi_pstore.service... May 10 00:40:13.172143 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. May 10 00:40:13.172271 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). May 10 00:40:13.173277 systemd[1]: Starting systemd-networkd-wait-online.service... May 10 00:40:13.175300 systemd[1]: modprobe@loop.service: Deactivated successfully. May 10 00:40:13.175386 systemd[1]: Finished modprobe@loop.service. May 10 00:40:13.173000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:13.173000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:13.175916 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 10 00:40:13.175993 systemd[1]: Finished modprobe@dm_mod.service. May 10 00:40:13.175000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:13.175000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:13.177250 systemd[1]: modprobe@drm.service: Deactivated successfully. May 10 00:40:13.177342 systemd[1]: Finished modprobe@drm.service. May 10 00:40:13.175000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:13.175000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:13.177841 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 10 00:40:13.178182 systemd[1]: Finished modprobe@efi_pstore.service. May 10 00:40:13.176000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:13.176000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:13.178712 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 10 00:40:13.178813 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. May 10 00:40:13.179643 systemd[1]: Finished ensure-sysext.service. May 10 00:40:13.177000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=ensure-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:13.224000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-boot-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:13.226687 systemd[1]: Finished systemd-boot-update.service. May 10 00:40:13.241200 kernel: kauditd_printk_skb: 227 callbacks suppressed May 10 00:40:13.241240 kernel: audit: type=1130 audit(1746837613.224:150): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-boot-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:13.357385 systemd[1]: Finished systemd-tmpfiles-setup.service. May 10 00:40:13.358667 systemd[1]: Starting audit-rules.service... May 10 00:40:13.355000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:13.364242 kernel: audit: type=1130 audit(1746837613.355:151): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:13.362935 systemd[1]: Starting clean-ca-certificates.service... May 10 00:40:13.364428 systemd[1]: Starting systemd-journal-catalog-update.service... May 10 00:40:13.366062 systemd[1]: Starting systemd-resolved.service... May 10 00:40:13.367563 systemd[1]: Starting systemd-timesyncd.service... May 10 00:40:13.368733 systemd[1]: Starting systemd-update-utmp.service... May 10 00:40:13.373289 systemd[1]: Finished clean-ca-certificates.service. May 10 00:40:13.371000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=clean-ca-certificates comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:13.376847 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). May 10 00:40:13.376991 kernel: audit: type=1130 audit(1746837613.371:152): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=clean-ca-certificates comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:13.382000 audit[1296]: SYSTEM_BOOT pid=1296 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg=' comm="systemd-update-utmp" exe="/usr/lib/systemd/systemd-update-utmp" hostname=? addr=? terminal=? res=success' May 10 00:40:13.389769 kernel: audit: type=1127 audit(1746837613.382:153): pid=1296 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg=' comm="systemd-update-utmp" exe="/usr/lib/systemd/systemd-update-utmp" hostname=? addr=? terminal=? res=success' May 10 00:40:13.394345 systemd[1]: Finished systemd-update-utmp.service. May 10 00:40:13.392000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-update-utmp comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:13.398049 kernel: audit: type=1130 audit(1746837613.392:154): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-update-utmp comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:13.447530 systemd[1]: Started systemd-timesyncd.service. May 10 00:40:13.445000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-timesyncd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:13.447785 systemd[1]: Reached target time-set.target. May 10 00:40:13.450923 kernel: audit: type=1130 audit(1746837613.445:155): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-timesyncd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:13.511061 systemd-networkd[1110]: ens192: Gained IPv6LL May 10 00:40:13.510000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-networkd-wait-online comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:40:13.511923 systemd[1]: Finished systemd-networkd-wait-online.service. May 10 00:40:13.512236 systemd-resolved[1294]: Positive Trust Anchors: May 10 00:40:13.512405 systemd-resolved[1294]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d May 10 00:40:13.512467 systemd-resolved[1294]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test May 10 00:40:13.514925 kernel: audit: type=1130 audit(1746837613.510:156): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-networkd-wait-online comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 10 00:41:34.603623 systemd-timesyncd[1295]: Contacted time server 72.14.183.239:123 (0.flatcar.pool.ntp.org). May 10 00:41:34.603838 systemd-timesyncd[1295]: Initial clock synchronization to Sat 2025-05-10 00:41:34.603564 UTC. May 10 00:41:34.670809 kernel: audit: type=1305 audit(1746837694.664:157): auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=add_rule key=(null) list=5 res=1 May 10 00:41:34.670865 kernel: audit: type=1300 audit(1746837694.664:157): arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7fffb7c3e3a0 a2=420 a3=0 items=0 ppid=1291 pid=1315 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/sbin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) May 10 00:41:34.664000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=add_rule key=(null) list=5 res=1 May 10 00:41:34.664000 audit[1315]: SYSCALL arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7fffb7c3e3a0 a2=420 a3=0 items=0 ppid=1291 pid=1315 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/sbin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) May 10 00:41:34.666062 systemd[1]: Finished audit-rules.service. May 10 00:41:34.671012 augenrules[1315]: No rules May 10 00:41:34.672637 kernel: audit: type=1327 audit(1746837694.664:157): proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 May 10 00:41:34.664000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 May 10 00:41:34.671575 systemd[1]: Finished systemd-journal-catalog-update.service. May 10 00:41:34.754143 systemd-resolved[1294]: Defaulting to hostname 'linux'. May 10 00:41:34.755602 systemd[1]: Started systemd-resolved.service. May 10 00:41:34.755772 systemd[1]: Reached target network.target. May 10 00:41:34.755869 systemd[1]: Reached target network-online.target. May 10 00:41:34.755965 systemd[1]: Reached target nss-lookup.target. May 10 00:41:34.799972 ldconfig[1147]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. May 10 00:41:34.802814 systemd[1]: Finished ldconfig.service. May 10 00:41:34.804278 systemd[1]: Starting systemd-update-done.service... May 10 00:41:34.811668 systemd[1]: Finished systemd-update-done.service. May 10 00:41:34.811887 systemd[1]: Reached target sysinit.target. May 10 00:41:34.812073 systemd[1]: Started motdgen.path. May 10 00:41:34.812206 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path. May 10 00:41:34.812440 systemd[1]: Started logrotate.timer. May 10 00:41:34.812609 systemd[1]: Started mdadm.timer. May 10 00:41:34.812719 systemd[1]: Started systemd-tmpfiles-clean.timer. May 10 00:41:34.812833 systemd[1]: update-engine-stub.timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). May 10 00:41:34.812867 systemd[1]: Reached target paths.target. May 10 00:41:34.812971 systemd[1]: Reached target timers.target. May 10 00:41:34.813291 systemd[1]: Listening on dbus.socket. May 10 00:41:34.814584 systemd[1]: Starting docker.socket... May 10 00:41:34.822892 systemd[1]: Listening on sshd.socket. May 10 00:41:34.823099 systemd[1]: systemd-pcrphase-sysinit.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). May 10 00:41:34.823511 systemd[1]: Listening on docker.socket. May 10 00:41:34.823646 systemd[1]: Reached target sockets.target. May 10 00:41:34.823767 systemd[1]: Reached target basic.target. May 10 00:41:34.823979 systemd[1]: System is tainted: cgroupsv1 May 10 00:41:34.824007 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). May 10 00:41:34.824037 systemd[1]: addon-config@usr-share-oem.service was skipped because no trigger condition checks were met. May 10 00:41:34.824059 systemd[1]: addon-run@usr-share-oem.service was skipped because no trigger condition checks were met. May 10 00:41:34.825196 systemd[1]: Starting containerd.service... May 10 00:41:34.826347 systemd[1]: Starting dbus.service... May 10 00:41:34.827597 systemd[1]: Starting enable-oem-cloudinit.service... May 10 00:41:34.828928 systemd[1]: Starting extend-filesystems.service... May 10 00:41:34.829083 systemd[1]: flatcar-setup-environment.service was skipped because of an unmet condition check (ConditionPathExists=/usr/share/oem/bin/flatcar-setup-environment). May 10 00:41:34.832249 jq[1330]: false May 10 00:41:34.835128 systemd[1]: Starting kubelet.service... May 10 00:41:34.836317 systemd[1]: Starting motdgen.service... May 10 00:41:34.837392 systemd[1]: Starting prepare-helm.service... May 10 00:41:34.839861 systemd[1]: Starting ssh-key-proc-cmdline.service... May 10 00:41:34.841133 systemd[1]: Starting sshd-keygen.service... May 10 00:41:34.842929 systemd[1]: Starting systemd-logind.service... May 10 00:41:34.843146 systemd[1]: systemd-pcrphase.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). May 10 00:41:34.843250 systemd[1]: tcsd.service was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). May 10 00:41:34.844606 systemd[1]: Starting update-engine.service... May 10 00:41:34.852024 systemd[1]: Starting update-ssh-keys-after-ignition.service... May 10 00:41:34.853338 systemd[1]: Starting vmtoolsd.service... May 10 00:41:34.853848 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). May 10 00:41:34.856649 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. May 10 00:41:34.864733 jq[1346]: true May 10 00:41:34.858682 systemd[1]: Condition check resulted in enable-oem-cloudinit.service being skipped. May 10 00:41:34.859875 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. May 10 00:41:34.860022 systemd[1]: Finished ssh-key-proc-cmdline.service. May 10 00:41:34.866290 jq[1352]: true May 10 00:41:34.880848 systemd[1]: Started vmtoolsd.service. May 10 00:41:34.890975 extend-filesystems[1331]: Found loop1 May 10 00:41:34.891272 extend-filesystems[1331]: Found sda May 10 00:41:34.891749 extend-filesystems[1331]: Found sda1 May 10 00:41:34.891897 extend-filesystems[1331]: Found sda2 May 10 00:41:34.892037 extend-filesystems[1331]: Found sda3 May 10 00:41:34.892177 extend-filesystems[1331]: Found usr May 10 00:41:34.892474 extend-filesystems[1331]: Found sda4 May 10 00:41:34.892615 extend-filesystems[1331]: Found sda6 May 10 00:41:34.892774 extend-filesystems[1331]: Found sda7 May 10 00:41:34.892929 extend-filesystems[1331]: Found sda9 May 10 00:41:34.893061 extend-filesystems[1331]: Checking size of /dev/sda9 May 10 00:41:34.894389 systemd[1]: motdgen.service: Deactivated successfully. May 10 00:41:34.894521 systemd[1]: Finished motdgen.service. May 10 00:41:34.931952 env[1375]: time="2025-05-10T00:41:34.931435668Z" level=info msg="starting containerd" revision=92b3a9d6f1b3bcc6dc74875cfdea653fe39f09c2 version=1.6.16 May 10 00:41:34.941947 extend-filesystems[1331]: Old size kept for /dev/sda9 May 10 00:41:34.941947 extend-filesystems[1331]: Found sr0 May 10 00:41:34.941467 systemd[1]: extend-filesystems.service: Deactivated successfully. May 10 00:41:34.941654 systemd[1]: Finished extend-filesystems.service. May 10 00:41:34.947838 tar[1351]: linux-amd64/helm May 10 00:41:34.973567 systemd-logind[1342]: Watching system buttons on /dev/input/event1 (Power Button) May 10 00:41:34.973845 systemd-logind[1342]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) May 10 00:41:34.974031 systemd-logind[1342]: New seat seat0. May 10 00:41:34.976650 bash[1373]: Updated "/home/core/.ssh/authorized_keys" May 10 00:41:34.977206 systemd[1]: Finished update-ssh-keys-after-ignition.service. May 10 00:41:34.999337 env[1375]: time="2025-05-10T00:41:34.999030323Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 May 10 00:41:34.999446 env[1375]: time="2025-05-10T00:41:34.999376944Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 May 10 00:41:35.004413 env[1375]: time="2025-05-10T00:41:35.004384019Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/5.15.181-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 May 10 00:41:35.004516 env[1375]: time="2025-05-10T00:41:35.004504256Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 May 10 00:41:35.004790 env[1375]: time="2025-05-10T00:41:35.004777332Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 May 10 00:41:35.004854 env[1375]: time="2025-05-10T00:41:35.004843607Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 May 10 00:41:35.004921 env[1375]: time="2025-05-10T00:41:35.004898985Z" level=warning msg="failed to load plugin io.containerd.snapshotter.v1.devmapper" error="devmapper not configured" May 10 00:41:35.004969 env[1375]: time="2025-05-10T00:41:35.004958909Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 May 10 00:41:35.005068 env[1375]: time="2025-05-10T00:41:35.005058268Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 May 10 00:41:35.005291 env[1375]: time="2025-05-10T00:41:35.005278863Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 May 10 00:41:35.005498 env[1375]: time="2025-05-10T00:41:35.005477778Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 May 10 00:41:35.005559 env[1375]: time="2025-05-10T00:41:35.005546414Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 May 10 00:41:35.005858 env[1375]: time="2025-05-10T00:41:35.005847092Z" level=warning msg="could not use snapshotter devmapper in metadata plugin" error="devmapper not configured" May 10 00:41:35.005913 env[1375]: time="2025-05-10T00:41:35.005897859Z" level=info msg="metadata content store policy set" policy=shared May 10 00:41:35.036381 env[1375]: time="2025-05-10T00:41:35.035261758Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 May 10 00:41:35.036381 env[1375]: time="2025-05-10T00:41:35.035322788Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 May 10 00:41:35.036381 env[1375]: time="2025-05-10T00:41:35.035333921Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 May 10 00:41:35.036381 env[1375]: time="2025-05-10T00:41:35.035364324Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 May 10 00:41:35.036381 env[1375]: time="2025-05-10T00:41:35.035383076Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 May 10 00:41:35.036381 env[1375]: time="2025-05-10T00:41:35.035407722Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 May 10 00:41:35.036381 env[1375]: time="2025-05-10T00:41:35.035420688Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 May 10 00:41:35.036381 env[1375]: time="2025-05-10T00:41:35.035429854Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 May 10 00:41:35.036381 env[1375]: time="2025-05-10T00:41:35.035438565Z" level=info msg="loading plugin \"io.containerd.service.v1.leases-service\"..." type=io.containerd.service.v1 May 10 00:41:35.036381 env[1375]: time="2025-05-10T00:41:35.035446163Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 May 10 00:41:35.036381 env[1375]: time="2025-05-10T00:41:35.035453764Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 May 10 00:41:35.036381 env[1375]: time="2025-05-10T00:41:35.035472605Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 May 10 00:41:35.036381 env[1375]: time="2025-05-10T00:41:35.036012795Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 May 10 00:41:35.036381 env[1375]: time="2025-05-10T00:41:35.036107946Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 May 10 00:41:35.037410 env[1375]: time="2025-05-10T00:41:35.037203528Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 May 10 00:41:35.037581 env[1375]: time="2025-05-10T00:41:35.037232284Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 May 10 00:41:35.037581 env[1375]: time="2025-05-10T00:41:35.037520824Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 May 10 00:41:35.037663 env[1375]: time="2025-05-10T00:41:35.037570613Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 May 10 00:41:35.038462 env[1375]: time="2025-05-10T00:41:35.038450049Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 May 10 00:41:35.038653 env[1375]: time="2025-05-10T00:41:35.038528822Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 May 10 00:41:35.039127 env[1375]: time="2025-05-10T00:41:35.039114806Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 May 10 00:41:35.039204 env[1375]: time="2025-05-10T00:41:35.039191444Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 May 10 00:41:35.039284 env[1375]: time="2025-05-10T00:41:35.039272606Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 May 10 00:41:35.039377 env[1375]: time="2025-05-10T00:41:35.039367147Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 May 10 00:41:35.039447 env[1375]: time="2025-05-10T00:41:35.039420477Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 May 10 00:41:35.039982 env[1375]: time="2025-05-10T00:41:35.039967863Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 May 10 00:41:35.040175 env[1375]: time="2025-05-10T00:41:35.040161309Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 May 10 00:41:35.040276 env[1375]: time="2025-05-10T00:41:35.040238904Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 May 10 00:41:35.040424 env[1375]: time="2025-05-10T00:41:35.040394720Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 May 10 00:41:35.040814 env[1375]: time="2025-05-10T00:41:35.040801674Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 May 10 00:41:35.040886 env[1375]: time="2025-05-10T00:41:35.040871680Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="no OpenTelemetry endpoint: skip plugin" type=io.containerd.tracing.processor.v1 May 10 00:41:35.040943 env[1375]: time="2025-05-10T00:41:35.040926422Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 May 10 00:41:35.041012 env[1375]: time="2025-05-10T00:41:35.040999251Z" level=error msg="failed to initialize a tracing processor \"otlp\"" error="no OpenTelemetry endpoint: skip plugin" May 10 00:41:35.041111 env[1375]: time="2025-05-10T00:41:35.041100069Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 May 10 00:41:35.041392 env[1375]: time="2025-05-10T00:41:35.041341857Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:false] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:false SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.6 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" May 10 00:41:35.047392 env[1375]: time="2025-05-10T00:41:35.041493428Z" level=info msg="Connect containerd service" May 10 00:41:35.047392 env[1375]: time="2025-05-10T00:41:35.041981459Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" May 10 00:41:35.047392 env[1375]: time="2025-05-10T00:41:35.044120263Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" May 10 00:41:35.047392 env[1375]: time="2025-05-10T00:41:35.044302150Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc May 10 00:41:35.047392 env[1375]: time="2025-05-10T00:41:35.044331325Z" level=info msg=serving... address=/run/containerd/containerd.sock May 10 00:41:35.047392 env[1375]: time="2025-05-10T00:41:35.044364853Z" level=info msg="containerd successfully booted in 0.115072s" May 10 00:41:35.047392 env[1375]: time="2025-05-10T00:41:35.045036323Z" level=info msg="Start subscribing containerd event" May 10 00:41:35.047392 env[1375]: time="2025-05-10T00:41:35.045076995Z" level=info msg="Start recovering state" May 10 00:41:35.047392 env[1375]: time="2025-05-10T00:41:35.045132003Z" level=info msg="Start event monitor" May 10 00:41:35.047392 env[1375]: time="2025-05-10T00:41:35.045151667Z" level=info msg="Start snapshots syncer" May 10 00:41:35.047392 env[1375]: time="2025-05-10T00:41:35.045161162Z" level=info msg="Start cni network conf syncer for default" May 10 00:41:35.047392 env[1375]: time="2025-05-10T00:41:35.045170349Z" level=info msg="Start streaming server" May 10 00:41:35.044445 systemd[1]: Started containerd.service. May 10 00:41:35.063797 kernel: NET: Registered PF_VSOCK protocol family May 10 00:41:35.064266 dbus-daemon[1328]: [system] SELinux support is enabled May 10 00:41:35.064388 systemd[1]: Started dbus.service. May 10 00:41:35.065876 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). May 10 00:41:35.065890 systemd[1]: Reached target system-config.target. May 10 00:41:35.066015 systemd[1]: user-cloudinit-proc-cmdline.service was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). May 10 00:41:35.066026 systemd[1]: Reached target user-config.target. May 10 00:41:35.068158 systemd[1]: Started systemd-logind.service. May 10 00:41:35.068821 dbus-daemon[1328]: [system] Successfully activated service 'org.freedesktop.systemd1' May 10 00:41:35.131724 update_engine[1343]: I0510 00:41:35.126482 1343 main.cc:92] Flatcar Update Engine starting May 10 00:41:35.140784 systemd[1]: Started update-engine.service. May 10 00:41:35.142895 update_engine[1343]: I0510 00:41:35.140801 1343 update_check_scheduler.cc:74] Next update check in 11m51s May 10 00:41:35.142533 systemd[1]: Started locksmithd.service. May 10 00:41:35.401650 tar[1351]: linux-amd64/LICENSE May 10 00:41:35.401786 tar[1351]: linux-amd64/README.md May 10 00:41:35.407203 systemd[1]: Finished prepare-helm.service. May 10 00:41:35.622895 sshd_keygen[1360]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 May 10 00:41:35.641383 systemd[1]: Finished sshd-keygen.service. May 10 00:41:35.642872 systemd[1]: Starting issuegen.service... May 10 00:41:35.648324 systemd[1]: issuegen.service: Deactivated successfully. May 10 00:41:35.648488 systemd[1]: Finished issuegen.service. May 10 00:41:35.649814 systemd[1]: Starting systemd-user-sessions.service... May 10 00:41:35.658353 systemd[1]: Finished systemd-user-sessions.service. May 10 00:41:35.659328 systemd[1]: Started getty@tty1.service. May 10 00:41:35.660138 systemd[1]: Started serial-getty@ttyS0.service. May 10 00:41:35.660336 systemd[1]: Reached target getty.target. May 10 00:41:35.794189 locksmithd[1405]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" May 10 00:41:37.024847 systemd[1]: Started kubelet.service. May 10 00:41:37.025265 systemd[1]: Reached target multi-user.target. May 10 00:41:37.026770 systemd[1]: Starting systemd-update-utmp-runlevel.service... May 10 00:41:37.035109 systemd[1]: systemd-update-utmp-runlevel.service: Deactivated successfully. May 10 00:41:37.035263 systemd[1]: Finished systemd-update-utmp-runlevel.service. May 10 00:41:37.035500 systemd[1]: Startup finished in 6.397s (kernel) + 6.868s (userspace) = 13.265s. May 10 00:41:37.240201 login[1473]: pam_unix(login:session): session opened for user core(uid=500) by LOGIN(uid=0) May 10 00:41:37.241966 login[1474]: pam_unix(login:session): session opened for user core(uid=500) by LOGIN(uid=0) May 10 00:41:37.279127 systemd[1]: Created slice user-500.slice. May 10 00:41:37.279880 systemd[1]: Starting user-runtime-dir@500.service... May 10 00:41:37.281584 systemd-logind[1342]: New session 1 of user core. May 10 00:41:37.285736 systemd-logind[1342]: New session 2 of user core. May 10 00:41:37.289190 systemd[1]: Finished user-runtime-dir@500.service. May 10 00:41:37.290108 systemd[1]: Starting user@500.service... May 10 00:41:37.293297 (systemd)[1486]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) May 10 00:41:37.399224 systemd[1486]: Queued start job for default target default.target. May 10 00:41:37.399386 systemd[1486]: Reached target paths.target. May 10 00:41:37.399401 systemd[1486]: Reached target sockets.target. May 10 00:41:37.399413 systemd[1486]: Reached target timers.target. May 10 00:41:37.399435 systemd[1486]: Reached target basic.target. May 10 00:41:37.399528 systemd[1]: Started user@500.service. May 10 00:41:37.400271 systemd[1]: Started session-1.scope. May 10 00:41:37.400795 systemd[1]: Started session-2.scope. May 10 00:41:37.401015 systemd[1486]: Reached target default.target. May 10 00:41:37.401146 systemd[1486]: Startup finished in 101ms. May 10 00:41:38.759164 kubelet[1481]: E0510 00:41:38.759124 1481 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 10 00:41:38.761115 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 10 00:41:38.761211 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 10 00:41:48.982526 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. May 10 00:41:48.982656 systemd[1]: Stopped kubelet.service. May 10 00:41:48.983719 systemd[1]: Starting kubelet.service... May 10 00:41:49.302744 systemd[1]: Started kubelet.service. May 10 00:41:49.490720 kubelet[1523]: E0510 00:41:49.490678 1523 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 10 00:41:49.492991 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 10 00:41:49.493073 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 10 00:41:59.732575 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. May 10 00:41:59.732798 systemd[1]: Stopped kubelet.service. May 10 00:41:59.734137 systemd[1]: Starting kubelet.service... May 10 00:42:00.048402 systemd[1]: Started kubelet.service. May 10 00:42:00.082097 kubelet[1538]: E0510 00:42:00.082062 1538 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 10 00:42:00.083066 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 10 00:42:00.083157 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 10 00:42:05.258535 systemd[1]: Created slice system-sshd.slice. May 10 00:42:05.259285 systemd[1]: Started sshd@0-139.178.70.106:22-139.178.89.65:44340.service. May 10 00:42:05.302174 sshd[1545]: Accepted publickey for core from 139.178.89.65 port 44340 ssh2: RSA SHA256:XIiY2M1wm02XXDRKVwO0PLWdmJcZoEVRw8Nio1wxBSQ May 10 00:42:05.303315 sshd[1545]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 10 00:42:05.306328 systemd[1]: Started session-3.scope. May 10 00:42:05.306637 systemd-logind[1342]: New session 3 of user core. May 10 00:42:05.354805 systemd[1]: Started sshd@1-139.178.70.106:22-139.178.89.65:44352.service. May 10 00:42:05.392936 sshd[1550]: Accepted publickey for core from 139.178.89.65 port 44352 ssh2: RSA SHA256:XIiY2M1wm02XXDRKVwO0PLWdmJcZoEVRw8Nio1wxBSQ May 10 00:42:05.394019 sshd[1550]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 10 00:42:05.396828 systemd[1]: Started session-4.scope. May 10 00:42:05.397100 systemd-logind[1342]: New session 4 of user core. May 10 00:42:05.447473 sshd[1550]: pam_unix(sshd:session): session closed for user core May 10 00:42:05.448985 systemd[1]: Started sshd@2-139.178.70.106:22-139.178.89.65:44356.service. May 10 00:42:05.451557 systemd[1]: sshd@1-139.178.70.106:22-139.178.89.65:44352.service: Deactivated successfully. May 10 00:42:05.451955 systemd[1]: session-4.scope: Deactivated successfully. May 10 00:42:05.452511 systemd-logind[1342]: Session 4 logged out. Waiting for processes to exit. May 10 00:42:05.452936 systemd-logind[1342]: Removed session 4. May 10 00:42:05.479239 sshd[1555]: Accepted publickey for core from 139.178.89.65 port 44356 ssh2: RSA SHA256:XIiY2M1wm02XXDRKVwO0PLWdmJcZoEVRw8Nio1wxBSQ May 10 00:42:05.480237 sshd[1555]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 10 00:42:05.483000 systemd[1]: Started session-5.scope. May 10 00:42:05.483627 systemd-logind[1342]: New session 5 of user core. May 10 00:42:05.531239 sshd[1555]: pam_unix(sshd:session): session closed for user core May 10 00:42:05.532875 systemd[1]: Started sshd@3-139.178.70.106:22-139.178.89.65:44370.service. May 10 00:42:05.536918 systemd[1]: sshd@2-139.178.70.106:22-139.178.89.65:44356.service: Deactivated successfully. May 10 00:42:05.537334 systemd[1]: session-5.scope: Deactivated successfully. May 10 00:42:05.538142 systemd-logind[1342]: Session 5 logged out. Waiting for processes to exit. May 10 00:42:05.538696 systemd-logind[1342]: Removed session 5. May 10 00:42:05.562264 sshd[1562]: Accepted publickey for core from 139.178.89.65 port 44370 ssh2: RSA SHA256:XIiY2M1wm02XXDRKVwO0PLWdmJcZoEVRw8Nio1wxBSQ May 10 00:42:05.563031 sshd[1562]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 10 00:42:05.565948 systemd[1]: Started session-6.scope. May 10 00:42:05.566128 systemd-logind[1342]: New session 6 of user core. May 10 00:42:05.617573 sshd[1562]: pam_unix(sshd:session): session closed for user core May 10 00:42:05.619199 systemd[1]: Started sshd@4-139.178.70.106:22-139.178.89.65:44374.service. May 10 00:42:05.623107 systemd[1]: sshd@3-139.178.70.106:22-139.178.89.65:44370.service: Deactivated successfully. May 10 00:42:05.623449 systemd[1]: session-6.scope: Deactivated successfully. May 10 00:42:05.624159 systemd-logind[1342]: Session 6 logged out. Waiting for processes to exit. May 10 00:42:05.624692 systemd-logind[1342]: Removed session 6. May 10 00:42:05.648507 sshd[1569]: Accepted publickey for core from 139.178.89.65 port 44374 ssh2: RSA SHA256:XIiY2M1wm02XXDRKVwO0PLWdmJcZoEVRw8Nio1wxBSQ May 10 00:42:05.649308 sshd[1569]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 10 00:42:05.652187 systemd[1]: Started session-7.scope. May 10 00:42:05.652358 systemd-logind[1342]: New session 7 of user core. May 10 00:42:05.710636 sudo[1575]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh May 10 00:42:05.710787 sudo[1575]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) May 10 00:42:05.724323 systemd[1]: Starting docker.service... May 10 00:42:05.746463 env[1586]: time="2025-05-10T00:42:05.746417994Z" level=info msg="Starting up" May 10 00:42:05.747453 env[1586]: time="2025-05-10T00:42:05.747442966Z" level=info msg="parsed scheme: \"unix\"" module=grpc May 10 00:42:05.747507 env[1586]: time="2025-05-10T00:42:05.747497242Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc May 10 00:42:05.747562 env[1586]: time="2025-05-10T00:42:05.747551463Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/libcontainerd/docker-containerd.sock 0 }] }" module=grpc May 10 00:42:05.747605 env[1586]: time="2025-05-10T00:42:05.747596629Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc May 10 00:42:05.752610 env[1586]: time="2025-05-10T00:42:05.752589587Z" level=info msg="parsed scheme: \"unix\"" module=grpc May 10 00:42:05.752610 env[1586]: time="2025-05-10T00:42:05.752602541Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc May 10 00:42:05.752681 env[1586]: time="2025-05-10T00:42:05.752611312Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/libcontainerd/docker-containerd.sock 0 }] }" module=grpc May 10 00:42:05.752681 env[1586]: time="2025-05-10T00:42:05.752616354Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc May 10 00:42:05.756164 systemd[1]: var-lib-docker-check\x2doverlayfs\x2dsupport1051730350-merged.mount: Deactivated successfully. May 10 00:42:05.916334 env[1586]: time="2025-05-10T00:42:05.915793541Z" level=warning msg="Your kernel does not support cgroup blkio weight" May 10 00:42:05.916440 env[1586]: time="2025-05-10T00:42:05.916430946Z" level=warning msg="Your kernel does not support cgroup blkio weight_device" May 10 00:42:05.916574 env[1586]: time="2025-05-10T00:42:05.916565448Z" level=info msg="Loading containers: start." May 10 00:42:06.000723 kernel: Initializing XFRM netlink socket May 10 00:42:06.023594 env[1586]: time="2025-05-10T00:42:06.023571844Z" level=info msg="Default bridge (docker0) is assigned with an IP address 172.17.0.0/16. Daemon option --bip can be used to set a preferred IP address" May 10 00:42:06.072368 systemd-networkd[1110]: docker0: Link UP May 10 00:42:06.083515 env[1586]: time="2025-05-10T00:42:06.083494784Z" level=info msg="Loading containers: done." May 10 00:42:06.090998 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck176532602-merged.mount: Deactivated successfully. May 10 00:42:06.093931 env[1586]: time="2025-05-10T00:42:06.093911046Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 May 10 00:42:06.094124 env[1586]: time="2025-05-10T00:42:06.094114261Z" level=info msg="Docker daemon" commit=112bdf3343 graphdriver(s)=overlay2 version=20.10.23 May 10 00:42:06.094217 env[1586]: time="2025-05-10T00:42:06.094208312Z" level=info msg="Daemon has completed initialization" May 10 00:42:06.100257 systemd[1]: Started docker.service. May 10 00:42:06.104647 env[1586]: time="2025-05-10T00:42:06.104617606Z" level=info msg="API listen on /run/docker.sock" May 10 00:42:07.056808 env[1375]: time="2025-05-10T00:42:07.056772526Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.30.12\"" May 10 00:42:07.631669 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1559634094.mount: Deactivated successfully. May 10 00:42:08.893616 env[1375]: time="2025-05-10T00:42:08.893566235Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-apiserver:v1.30.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:08.894502 env[1375]: time="2025-05-10T00:42:08.894480158Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:e113c59aa22f0650435e2a3ed64aadb01e87f3d2835aa3825fe078cd39699bfb,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:08.895615 env[1375]: time="2025-05-10T00:42:08.895595135Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-apiserver:v1.30.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:08.896624 env[1375]: time="2025-05-10T00:42:08.896593490Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-apiserver@sha256:4878682f7a044274d42399a6316ef452c5411aafd4ad99cc57de7235ca490e4e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:08.897115 env[1375]: time="2025-05-10T00:42:08.897100572Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.30.12\" returns image reference \"sha256:e113c59aa22f0650435e2a3ed64aadb01e87f3d2835aa3825fe078cd39699bfb\"" May 10 00:42:08.903807 env[1375]: time="2025-05-10T00:42:08.903781825Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.30.12\"" May 10 00:42:10.232460 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. May 10 00:42:10.232593 systemd[1]: Stopped kubelet.service. May 10 00:42:10.233833 systemd[1]: Starting kubelet.service... May 10 00:42:10.744149 env[1375]: time="2025-05-10T00:42:10.744120036Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-controller-manager:v1.30.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:10.761222 env[1375]: time="2025-05-10T00:42:10.760627366Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:70742b7b7d90a618a1fa06d89248dbe2c291c19d7f75f4ad60a69d0454dbbac8,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:10.765833 env[1375]: time="2025-05-10T00:42:10.765810146Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-controller-manager:v1.30.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:10.775430 env[1375]: time="2025-05-10T00:42:10.775401433Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-controller-manager@sha256:3a36711d0409d565b370a18d0c19339e93d4f1b1f2b3fd382eb31c714c463b74,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:10.775829 env[1375]: time="2025-05-10T00:42:10.775813682Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.30.12\" returns image reference \"sha256:70742b7b7d90a618a1fa06d89248dbe2c291c19d7f75f4ad60a69d0454dbbac8\"" May 10 00:42:10.781796 env[1375]: time="2025-05-10T00:42:10.781692883Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.30.12\"" May 10 00:42:10.804918 systemd[1]: Started kubelet.service. May 10 00:42:10.875322 kubelet[1733]: E0510 00:42:10.875300 1733 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 10 00:42:10.876399 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 10 00:42:10.876483 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 10 00:42:12.453302 env[1375]: time="2025-05-10T00:42:12.453248348Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-scheduler:v1.30.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:12.454441 env[1375]: time="2025-05-10T00:42:12.454418852Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:c0b91cfea9f9a1c09fc5d056f3a015e52604fd0d63671ff5bf31e642402ef05d,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:12.455902 env[1375]: time="2025-05-10T00:42:12.455884353Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-scheduler:v1.30.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:12.457218 env[1375]: time="2025-05-10T00:42:12.457197101Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-scheduler@sha256:521c843d01025be7d4e246ddee8cde74556eb9813c606d6db9f0f03236f6d029,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:12.457924 env[1375]: time="2025-05-10T00:42:12.457904974Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.30.12\" returns image reference \"sha256:c0b91cfea9f9a1c09fc5d056f3a015e52604fd0d63671ff5bf31e642402ef05d\"" May 10 00:42:12.465241 env[1375]: time="2025-05-10T00:42:12.465215718Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.12\"" May 10 00:42:13.804654 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4108544264.mount: Deactivated successfully. May 10 00:42:14.264528 env[1375]: time="2025-05-10T00:42:14.264499832Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy:v1.30.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:14.276334 env[1375]: time="2025-05-10T00:42:14.276317355Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:c9356fea5d151501039907c3ba870272461396117eabc74063632616f4e31b2b,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:14.284768 env[1375]: time="2025-05-10T00:42:14.284747940Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-proxy:v1.30.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:14.289905 env[1375]: time="2025-05-10T00:42:14.289880861Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy@sha256:ea8c7d5392acf6b0c11ebba78301e1a6c2dc6abcd7544102ed578e49d1c82f15,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:14.290301 env[1375]: time="2025-05-10T00:42:14.290276495Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.12\" returns image reference \"sha256:c9356fea5d151501039907c3ba870272461396117eabc74063632616f4e31b2b\"" May 10 00:42:14.297434 env[1375]: time="2025-05-10T00:42:14.297399228Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\"" May 10 00:42:14.777109 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount763166.mount: Deactivated successfully. May 10 00:42:15.688061 env[1375]: time="2025-05-10T00:42:15.688036590Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/coredns/coredns:v1.11.1,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:15.696143 env[1375]: time="2025-05-10T00:42:15.696130255Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:15.704037 env[1375]: time="2025-05-10T00:42:15.704025581Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/coredns/coredns:v1.11.1,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:15.706146 env[1375]: time="2025-05-10T00:42:15.706134596Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:15.706602 env[1375]: time="2025-05-10T00:42:15.706588546Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\" returns image reference \"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\"" May 10 00:42:15.714345 env[1375]: time="2025-05-10T00:42:15.714321633Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\"" May 10 00:42:16.262407 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2744395591.mount: Deactivated successfully. May 10 00:42:16.264750 env[1375]: time="2025-05-10T00:42:16.264727047Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause:3.9,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:16.265260 env[1375]: time="2025-05-10T00:42:16.265247995Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:e6f1816883972d4be47bd48879a08919b96afcd344132622e4d444987919323c,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:16.266015 env[1375]: time="2025-05-10T00:42:16.266000535Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.9,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:16.266778 env[1375]: time="2025-05-10T00:42:16.266766003Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:16.267099 env[1375]: time="2025-05-10T00:42:16.267082760Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\" returns image reference \"sha256:e6f1816883972d4be47bd48879a08919b96afcd344132622e4d444987919323c\"" May 10 00:42:16.272679 env[1375]: time="2025-05-10T00:42:16.272659988Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.12-0\"" May 10 00:42:16.780208 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount360771008.mount: Deactivated successfully. May 10 00:42:18.817930 env[1375]: time="2025-05-10T00:42:18.817889118Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/etcd:3.5.12-0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:18.819160 env[1375]: time="2025-05-10T00:42:18.819137801Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:3861cfcd7c04ccac1f062788eca39487248527ef0c0cfd477a83d7691a75a899,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:18.820542 env[1375]: time="2025-05-10T00:42:18.820515286Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/etcd:3.5.12-0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:18.822011 env[1375]: time="2025-05-10T00:42:18.821985677Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/etcd@sha256:44a8e24dcbba3470ee1fee21d5e88d128c936e9b55d4bc51fbef8086f8ed123b,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:18.822697 env[1375]: time="2025-05-10T00:42:18.822670929Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.12-0\" returns image reference \"sha256:3861cfcd7c04ccac1f062788eca39487248527ef0c0cfd477a83d7691a75a899\"" May 10 00:42:20.423809 systemd[1]: Stopped kubelet.service. May 10 00:42:20.425340 systemd[1]: Starting kubelet.service... May 10 00:42:20.437219 systemd[1]: Reloading. May 10 00:42:20.490023 /usr/lib/systemd/system-generators/torcx-generator[1846]: time="2025-05-10T00:42:20Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.7 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.7 /var/lib/torcx/store]" May 10 00:42:20.490276 /usr/lib/systemd/system-generators/torcx-generator[1846]: time="2025-05-10T00:42:20Z" level=info msg="torcx already run" May 10 00:42:20.559895 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. May 10 00:42:20.559999 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. May 10 00:42:20.572570 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 10 00:42:20.624923 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM May 10 00:42:20.625054 systemd[1]: kubelet.service: Failed with result 'signal'. May 10 00:42:20.625309 systemd[1]: Stopped kubelet.service. May 10 00:42:20.626912 systemd[1]: Starting kubelet.service... May 10 00:42:20.784729 update_engine[1343]: I0510 00:42:20.784627 1343 update_attempter.cc:509] Updating boot flags... May 10 00:42:21.587347 systemd[1]: Started kubelet.service. May 10 00:42:21.674992 kubelet[1939]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 10 00:42:21.674992 kubelet[1939]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. May 10 00:42:21.674992 kubelet[1939]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 10 00:42:21.786362 kubelet[1939]: I0510 00:42:21.786319 1939 server.go:205] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" May 10 00:42:22.017940 kubelet[1939]: I0510 00:42:22.017924 1939 server.go:484] "Kubelet version" kubeletVersion="v1.30.1" May 10 00:42:22.018024 kubelet[1939]: I0510 00:42:22.018016 1939 server.go:486] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" May 10 00:42:22.018186 kubelet[1939]: I0510 00:42:22.018179 1939 server.go:927] "Client rotation is on, will bootstrap in background" May 10 00:42:22.030637 kubelet[1939]: I0510 00:42:22.030628 1939 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" May 10 00:42:22.037779 kubelet[1939]: E0510 00:42:22.037767 1939 certificate_manager.go:562] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://139.178.70.106:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 139.178.70.106:6443: connect: connection refused May 10 00:42:22.042546 kubelet[1939]: I0510 00:42:22.042534 1939 server.go:742] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" May 10 00:42:22.044437 kubelet[1939]: I0510 00:42:22.044416 1939 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] May 10 00:42:22.044537 kubelet[1939]: I0510 00:42:22.044438 1939 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} May 10 00:42:22.044604 kubelet[1939]: I0510 00:42:22.044543 1939 topology_manager.go:138] "Creating topology manager with none policy" May 10 00:42:22.044604 kubelet[1939]: I0510 00:42:22.044550 1939 container_manager_linux.go:301] "Creating device plugin manager" May 10 00:42:22.045301 kubelet[1939]: I0510 00:42:22.045289 1939 state_mem.go:36] "Initialized new in-memory state store" May 10 00:42:22.045959 kubelet[1939]: I0510 00:42:22.045948 1939 kubelet.go:400] "Attempting to sync node with API server" May 10 00:42:22.046145 kubelet[1939]: I0510 00:42:22.046135 1939 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" May 10 00:42:22.046176 kubelet[1939]: I0510 00:42:22.046151 1939 kubelet.go:312] "Adding apiserver pod source" May 10 00:42:22.046176 kubelet[1939]: I0510 00:42:22.046161 1939 apiserver.go:42] "Waiting for node sync before watching apiserver pods" May 10 00:42:22.046271 kubelet[1939]: W0510 00:42:22.046250 1939 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://139.178.70.106:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 139.178.70.106:6443: connect: connection refused May 10 00:42:22.046329 kubelet[1939]: E0510 00:42:22.046321 1939 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://139.178.70.106:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 139.178.70.106:6443: connect: connection refused May 10 00:42:22.052064 kubelet[1939]: I0510 00:42:22.052050 1939 kuberuntime_manager.go:261] "Container runtime initialized" containerRuntime="containerd" version="1.6.16" apiVersion="v1" May 10 00:42:22.056335 kubelet[1939]: I0510 00:42:22.056322 1939 kubelet.go:815] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" May 10 00:42:22.056370 kubelet[1939]: W0510 00:42:22.056352 1939 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. May 10 00:42:22.056607 kubelet[1939]: I0510 00:42:22.056596 1939 server.go:1264] "Started kubelet" May 10 00:42:22.057076 kubelet[1939]: W0510 00:42:22.056651 1939 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://139.178.70.106:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 139.178.70.106:6443: connect: connection refused May 10 00:42:22.057076 kubelet[1939]: E0510 00:42:22.056684 1939 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://139.178.70.106:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 139.178.70.106:6443: connect: connection refused May 10 00:42:22.063395 kubelet[1939]: I0510 00:42:22.063382 1939 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 May 10 00:42:22.063959 kubelet[1939]: I0510 00:42:22.063952 1939 server.go:455] "Adding debug handlers to kubelet server" May 10 00:42:22.066034 kubelet[1939]: I0510 00:42:22.066007 1939 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 May 10 00:42:22.066206 kubelet[1939]: I0510 00:42:22.066138 1939 server.go:227] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" May 10 00:42:22.067864 kernel: SELinux: Context system_u:object_r:container_file_t:s0 is not valid (left unmapped). May 10 00:42:22.067900 kubelet[1939]: E0510 00:42:22.067693 1939 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://139.178.70.106:6443/api/v1/namespaces/default/events\": dial tcp 139.178.70.106:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.183e03abd7873bd9 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-05-10 00:42:22.056586201 +0000 UTC m=+0.465812965,LastTimestamp:2025-05-10 00:42:22.056586201 +0000 UTC m=+0.465812965,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" May 10 00:42:22.067995 kubelet[1939]: I0510 00:42:22.067988 1939 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" May 10 00:42:22.068398 kubelet[1939]: E0510 00:42:22.068380 1939 kubelet.go:1467] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" May 10 00:42:22.070641 kubelet[1939]: E0510 00:42:22.070629 1939 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 10 00:42:22.070680 kubelet[1939]: I0510 00:42:22.070647 1939 volume_manager.go:291] "Starting Kubelet Volume Manager" May 10 00:42:22.070716 kubelet[1939]: I0510 00:42:22.070707 1939 desired_state_of_world_populator.go:149] "Desired state populator starts to run" May 10 00:42:22.070745 kubelet[1939]: I0510 00:42:22.070731 1939 reconciler.go:26] "Reconciler: start to sync state" May 10 00:42:22.070888 kubelet[1939]: W0510 00:42:22.070867 1939 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://139.178.70.106:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 139.178.70.106:6443: connect: connection refused May 10 00:42:22.070917 kubelet[1939]: E0510 00:42:22.070892 1939 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://139.178.70.106:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 139.178.70.106:6443: connect: connection refused May 10 00:42:22.071138 kubelet[1939]: E0510 00:42:22.071122 1939 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://139.178.70.106:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 139.178.70.106:6443: connect: connection refused" interval="200ms" May 10 00:42:22.071208 kubelet[1939]: I0510 00:42:22.071197 1939 factory.go:221] Registration of the systemd container factory successfully May 10 00:42:22.071237 kubelet[1939]: I0510 00:42:22.071228 1939 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory May 10 00:42:22.071882 kubelet[1939]: I0510 00:42:22.071871 1939 factory.go:221] Registration of the containerd container factory successfully May 10 00:42:22.078679 kubelet[1939]: I0510 00:42:22.078662 1939 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" May 10 00:42:22.079300 kubelet[1939]: I0510 00:42:22.079293 1939 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" May 10 00:42:22.079346 kubelet[1939]: I0510 00:42:22.079340 1939 status_manager.go:217] "Starting to sync pod status with apiserver" May 10 00:42:22.079397 kubelet[1939]: I0510 00:42:22.079391 1939 kubelet.go:2337] "Starting kubelet main sync loop" May 10 00:42:22.079459 kubelet[1939]: E0510 00:42:22.079450 1939 kubelet.go:2361] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" May 10 00:42:22.082747 kubelet[1939]: W0510 00:42:22.082728 1939 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://139.178.70.106:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 139.178.70.106:6443: connect: connection refused May 10 00:42:22.082804 kubelet[1939]: E0510 00:42:22.082796 1939 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://139.178.70.106:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 139.178.70.106:6443: connect: connection refused May 10 00:42:22.095640 kubelet[1939]: I0510 00:42:22.095624 1939 cpu_manager.go:214] "Starting CPU manager" policy="none" May 10 00:42:22.095640 kubelet[1939]: I0510 00:42:22.095636 1939 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" May 10 00:42:22.095754 kubelet[1939]: I0510 00:42:22.095653 1939 state_mem.go:36] "Initialized new in-memory state store" May 10 00:42:22.096911 kubelet[1939]: I0510 00:42:22.096897 1939 policy_none.go:49] "None policy: Start" May 10 00:42:22.097306 kubelet[1939]: I0510 00:42:22.097291 1939 memory_manager.go:170] "Starting memorymanager" policy="None" May 10 00:42:22.097306 kubelet[1939]: I0510 00:42:22.097307 1939 state_mem.go:35] "Initializing new in-memory state store" May 10 00:42:22.100910 kubelet[1939]: I0510 00:42:22.100897 1939 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" May 10 00:42:22.101012 kubelet[1939]: I0510 00:42:22.100982 1939 container_log_manager.go:186] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" May 10 00:42:22.101057 kubelet[1939]: I0510 00:42:22.101046 1939 plugin_manager.go:118] "Starting Kubelet Plugin Manager" May 10 00:42:22.101860 kubelet[1939]: E0510 00:42:22.101851 1939 eviction_manager.go:282] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" May 10 00:42:22.172275 kubelet[1939]: I0510 00:42:22.172251 1939 kubelet_node_status.go:73] "Attempting to register node" node="localhost" May 10 00:42:22.172622 kubelet[1939]: E0510 00:42:22.172606 1939 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://139.178.70.106:6443/api/v1/nodes\": dial tcp 139.178.70.106:6443: connect: connection refused" node="localhost" May 10 00:42:22.179799 kubelet[1939]: I0510 00:42:22.179776 1939 topology_manager.go:215] "Topology Admit Handler" podUID="6ece95f10dbffa04b25ec3439a115512" podNamespace="kube-system" podName="kube-scheduler-localhost" May 10 00:42:22.180536 kubelet[1939]: I0510 00:42:22.180513 1939 topology_manager.go:215] "Topology Admit Handler" podUID="b00b54c5589197641a7ec2945ad6b746" podNamespace="kube-system" podName="kube-apiserver-localhost" May 10 00:42:22.181305 kubelet[1939]: I0510 00:42:22.181292 1939 topology_manager.go:215] "Topology Admit Handler" podUID="b20b39a8540dba87b5883a6f0f602dba" podNamespace="kube-system" podName="kube-controller-manager-localhost" May 10 00:42:22.271848 kubelet[1939]: E0510 00:42:22.271764 1939 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://139.178.70.106:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 139.178.70.106:6443: connect: connection refused" interval="400ms" May 10 00:42:22.372318 kubelet[1939]: I0510 00:42:22.372297 1939 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b00b54c5589197641a7ec2945ad6b746-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"b00b54c5589197641a7ec2945ad6b746\") " pod="kube-system/kube-apiserver-localhost" May 10 00:42:22.372394 kubelet[1939]: I0510 00:42:22.372324 1939 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 10 00:42:22.372394 kubelet[1939]: I0510 00:42:22.372340 1939 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 10 00:42:22.372394 kubelet[1939]: I0510 00:42:22.372368 1939 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 10 00:42:22.372394 kubelet[1939]: I0510 00:42:22.372382 1939 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/6ece95f10dbffa04b25ec3439a115512-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"6ece95f10dbffa04b25ec3439a115512\") " pod="kube-system/kube-scheduler-localhost" May 10 00:42:22.372394 kubelet[1939]: I0510 00:42:22.372393 1939 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/b00b54c5589197641a7ec2945ad6b746-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"b00b54c5589197641a7ec2945ad6b746\") " pod="kube-system/kube-apiserver-localhost" May 10 00:42:22.372525 kubelet[1939]: I0510 00:42:22.372405 1939 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/b00b54c5589197641a7ec2945ad6b746-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"b00b54c5589197641a7ec2945ad6b746\") " pod="kube-system/kube-apiserver-localhost" May 10 00:42:22.372525 kubelet[1939]: I0510 00:42:22.372416 1939 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 10 00:42:22.372525 kubelet[1939]: I0510 00:42:22.372443 1939 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 10 00:42:22.374176 kubelet[1939]: I0510 00:42:22.374160 1939 kubelet_node_status.go:73] "Attempting to register node" node="localhost" May 10 00:42:22.374398 kubelet[1939]: E0510 00:42:22.374383 1939 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://139.178.70.106:6443/api/v1/nodes\": dial tcp 139.178.70.106:6443: connect: connection refused" node="localhost" May 10 00:42:22.485643 env[1375]: time="2025-05-10T00:42:22.485606401Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:6ece95f10dbffa04b25ec3439a115512,Namespace:kube-system,Attempt:0,}" May 10 00:42:22.486250 env[1375]: time="2025-05-10T00:42:22.486048740Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:b00b54c5589197641a7ec2945ad6b746,Namespace:kube-system,Attempt:0,}" May 10 00:42:22.487965 env[1375]: time="2025-05-10T00:42:22.487801243Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:b20b39a8540dba87b5883a6f0f602dba,Namespace:kube-system,Attempt:0,}" May 10 00:42:22.672943 kubelet[1939]: E0510 00:42:22.672862 1939 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://139.178.70.106:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 139.178.70.106:6443: connect: connection refused" interval="800ms" May 10 00:42:22.776065 kubelet[1939]: I0510 00:42:22.776031 1939 kubelet_node_status.go:73] "Attempting to register node" node="localhost" May 10 00:42:22.776511 kubelet[1939]: E0510 00:42:22.776493 1939 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://139.178.70.106:6443/api/v1/nodes\": dial tcp 139.178.70.106:6443: connect: connection refused" node="localhost" May 10 00:42:22.941476 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3729127048.mount: Deactivated successfully. May 10 00:42:22.944153 env[1375]: time="2025-05-10T00:42:22.944121115Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:22.946123 env[1375]: time="2025-05-10T00:42:22.946101768Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:6270bb605e12e581514ada5fd5b3216f727db55dc87d5889c790e4c760683fee,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:22.946801 env[1375]: time="2025-05-10T00:42:22.946785010Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:22.947456 env[1375]: time="2025-05-10T00:42:22.947440813Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:22.948032 env[1375]: time="2025-05-10T00:42:22.948016865Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:6270bb605e12e581514ada5fd5b3216f727db55dc87d5889c790e4c760683fee,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:22.950579 env[1375]: time="2025-05-10T00:42:22.950564178Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:22.953736 env[1375]: time="2025-05-10T00:42:22.953163799Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:6270bb605e12e581514ada5fd5b3216f727db55dc87d5889c790e4c760683fee,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:22.954294 env[1375]: time="2025-05-10T00:42:22.954279362Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:22.954874 env[1375]: time="2025-05-10T00:42:22.954857303Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:22.955441 env[1375]: time="2025-05-10T00:42:22.955425845Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:22.956394 env[1375]: time="2025-05-10T00:42:22.956366068Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:22.960985 env[1375]: time="2025-05-10T00:42:22.960969850Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:22.974084 env[1375]: time="2025-05-10T00:42:22.974038074Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 10 00:42:22.974192 env[1375]: time="2025-05-10T00:42:22.974179363Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 10 00:42:22.974250 env[1375]: time="2025-05-10T00:42:22.974236163Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 10 00:42:22.974291 env[1375]: time="2025-05-10T00:42:22.974238198Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 10 00:42:22.974291 env[1375]: time="2025-05-10T00:42:22.974266168Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 10 00:42:22.974291 env[1375]: time="2025-05-10T00:42:22.974273226Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 10 00:42:22.974372 env[1375]: time="2025-05-10T00:42:22.974346934Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/c37ce95aadb694eea3e5f5fd7073dd489c8fd0d166aedc51fd8a94c24091d168 pid=1977 runtime=io.containerd.runc.v2 May 10 00:42:22.975098 env[1375]: time="2025-05-10T00:42:22.975072642Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/402272734216d83b04e0ba94d00a9bbbc906b7aed9af3d215105b5dd275e1672 pid=1999 runtime=io.containerd.runc.v2 May 10 00:42:22.977793 env[1375]: time="2025-05-10T00:42:22.977761504Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 10 00:42:22.977838 env[1375]: time="2025-05-10T00:42:22.977785325Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 10 00:42:22.977838 env[1375]: time="2025-05-10T00:42:22.977802599Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 10 00:42:22.977939 env[1375]: time="2025-05-10T00:42:22.977922558Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/03f5f99ae166d7fce4f91247fbc1176e0dcdf3784de438eb6aaa8f49cf330385 pid=2003 runtime=io.containerd.runc.v2 May 10 00:42:23.027107 kubelet[1939]: W0510 00:42:23.027047 1939 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://139.178.70.106:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 139.178.70.106:6443: connect: connection refused May 10 00:42:23.027107 kubelet[1939]: E0510 00:42:23.027092 1939 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://139.178.70.106:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 139.178.70.106:6443: connect: connection refused May 10 00:42:23.031130 env[1375]: time="2025-05-10T00:42:23.031108447Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:b00b54c5589197641a7ec2945ad6b746,Namespace:kube-system,Attempt:0,} returns sandbox id \"c37ce95aadb694eea3e5f5fd7073dd489c8fd0d166aedc51fd8a94c24091d168\"" May 10 00:42:23.033742 env[1375]: time="2025-05-10T00:42:23.033661582Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:6ece95f10dbffa04b25ec3439a115512,Namespace:kube-system,Attempt:0,} returns sandbox id \"402272734216d83b04e0ba94d00a9bbbc906b7aed9af3d215105b5dd275e1672\"" May 10 00:42:23.034520 env[1375]: time="2025-05-10T00:42:23.034504360Z" level=info msg="CreateContainer within sandbox \"c37ce95aadb694eea3e5f5fd7073dd489c8fd0d166aedc51fd8a94c24091d168\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" May 10 00:42:23.035909 env[1375]: time="2025-05-10T00:42:23.035896063Z" level=info msg="CreateContainer within sandbox \"402272734216d83b04e0ba94d00a9bbbc906b7aed9af3d215105b5dd275e1672\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" May 10 00:42:23.045903 env[1375]: time="2025-05-10T00:42:23.045887334Z" level=info msg="CreateContainer within sandbox \"402272734216d83b04e0ba94d00a9bbbc906b7aed9af3d215105b5dd275e1672\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"6936607234f90871c5351c35dbf27b8be7ec45ca3d30d5c94f43d4f270922a97\"" May 10 00:42:23.046762 env[1375]: time="2025-05-10T00:42:23.046745979Z" level=info msg="StartContainer for \"6936607234f90871c5351c35dbf27b8be7ec45ca3d30d5c94f43d4f270922a97\"" May 10 00:42:23.046816 env[1375]: time="2025-05-10T00:42:23.046742643Z" level=info msg="CreateContainer within sandbox \"c37ce95aadb694eea3e5f5fd7073dd489c8fd0d166aedc51fd8a94c24091d168\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"4fcf494c6e4b69cc0092d537ba80fea600dda66ca29a999b7311ef8736f41dcd\"" May 10 00:42:23.047039 env[1375]: time="2025-05-10T00:42:23.047028339Z" level=info msg="StartContainer for \"4fcf494c6e4b69cc0092d537ba80fea600dda66ca29a999b7311ef8736f41dcd\"" May 10 00:42:23.068713 env[1375]: time="2025-05-10T00:42:23.068673116Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:b20b39a8540dba87b5883a6f0f602dba,Namespace:kube-system,Attempt:0,} returns sandbox id \"03f5f99ae166d7fce4f91247fbc1176e0dcdf3784de438eb6aaa8f49cf330385\"" May 10 00:42:23.075408 env[1375]: time="2025-05-10T00:42:23.075390058Z" level=info msg="CreateContainer within sandbox \"03f5f99ae166d7fce4f91247fbc1176e0dcdf3784de438eb6aaa8f49cf330385\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" May 10 00:42:23.087696 env[1375]: time="2025-05-10T00:42:23.087671470Z" level=info msg="CreateContainer within sandbox \"03f5f99ae166d7fce4f91247fbc1176e0dcdf3784de438eb6aaa8f49cf330385\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"75ae76e8ea2bb78e2cb64098c9ecee9ed83e817201467d8cb518b5e23eb4ec9b\"" May 10 00:42:23.094423 env[1375]: time="2025-05-10T00:42:23.094350879Z" level=info msg="StartContainer for \"75ae76e8ea2bb78e2cb64098c9ecee9ed83e817201467d8cb518b5e23eb4ec9b\"" May 10 00:42:23.122956 env[1375]: time="2025-05-10T00:42:23.122846806Z" level=info msg="StartContainer for \"6936607234f90871c5351c35dbf27b8be7ec45ca3d30d5c94f43d4f270922a97\" returns successfully" May 10 00:42:23.134008 env[1375]: time="2025-05-10T00:42:23.133987750Z" level=info msg="StartContainer for \"4fcf494c6e4b69cc0092d537ba80fea600dda66ca29a999b7311ef8736f41dcd\" returns successfully" May 10 00:42:23.155628 env[1375]: time="2025-05-10T00:42:23.155607593Z" level=info msg="StartContainer for \"75ae76e8ea2bb78e2cb64098c9ecee9ed83e817201467d8cb518b5e23eb4ec9b\" returns successfully" May 10 00:42:23.163346 kubelet[1939]: W0510 00:42:23.163274 1939 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://139.178.70.106:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 139.178.70.106:6443: connect: connection refused May 10 00:42:23.163346 kubelet[1939]: E0510 00:42:23.163327 1939 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://139.178.70.106:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 139.178.70.106:6443: connect: connection refused May 10 00:42:23.228032 kubelet[1939]: E0510 00:42:23.227912 1939 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://139.178.70.106:6443/api/v1/namespaces/default/events\": dial tcp 139.178.70.106:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.183e03abd7873bd9 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-05-10 00:42:22.056586201 +0000 UTC m=+0.465812965,LastTimestamp:2025-05-10 00:42:22.056586201 +0000 UTC m=+0.465812965,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" May 10 00:42:23.281635 kubelet[1939]: W0510 00:42:23.281573 1939 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://139.178.70.106:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 139.178.70.106:6443: connect: connection refused May 10 00:42:23.281635 kubelet[1939]: E0510 00:42:23.281614 1939 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://139.178.70.106:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 139.178.70.106:6443: connect: connection refused May 10 00:42:23.326963 kubelet[1939]: W0510 00:42:23.326917 1939 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://139.178.70.106:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 139.178.70.106:6443: connect: connection refused May 10 00:42:23.326963 kubelet[1939]: E0510 00:42:23.326949 1939 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://139.178.70.106:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 139.178.70.106:6443: connect: connection refused May 10 00:42:23.473794 kubelet[1939]: E0510 00:42:23.473764 1939 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://139.178.70.106:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 139.178.70.106:6443: connect: connection refused" interval="1.6s" May 10 00:42:23.578303 kubelet[1939]: I0510 00:42:23.578123 1939 kubelet_node_status.go:73] "Attempting to register node" node="localhost" May 10 00:42:23.578303 kubelet[1939]: E0510 00:42:23.578287 1939 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://139.178.70.106:6443/api/v1/nodes\": dial tcp 139.178.70.106:6443: connect: connection refused" node="localhost" May 10 00:42:24.794768 kubelet[1939]: E0510 00:42:24.794743 1939 csi_plugin.go:308] Failed to initialize CSINode: error updating CSINode annotation: timed out waiting for the condition; caused by: nodes "localhost" not found May 10 00:42:25.077002 kubelet[1939]: E0510 00:42:25.076915 1939 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" May 10 00:42:25.146687 kubelet[1939]: E0510 00:42:25.146664 1939 csi_plugin.go:308] Failed to initialize CSINode: error updating CSINode annotation: timed out waiting for the condition; caused by: nodes "localhost" not found May 10 00:42:25.179738 kubelet[1939]: I0510 00:42:25.179723 1939 kubelet_node_status.go:73] "Attempting to register node" node="localhost" May 10 00:42:25.188776 kubelet[1939]: I0510 00:42:25.188764 1939 kubelet_node_status.go:76] "Successfully registered node" node="localhost" May 10 00:42:25.193569 kubelet[1939]: E0510 00:42:25.193549 1939 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 10 00:42:25.293771 kubelet[1939]: E0510 00:42:25.293744 1939 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 10 00:42:25.394453 kubelet[1939]: E0510 00:42:25.394385 1939 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" May 10 00:42:26.048630 kubelet[1939]: I0510 00:42:26.048598 1939 apiserver.go:52] "Watching apiserver" May 10 00:42:26.071019 kubelet[1939]: I0510 00:42:26.070997 1939 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" May 10 00:42:26.266032 systemd[1]: Reloading. May 10 00:42:26.317036 /usr/lib/systemd/system-generators/torcx-generator[2230]: time="2025-05-10T00:42:26Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.7 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.7 /var/lib/torcx/store]" May 10 00:42:26.317415 /usr/lib/systemd/system-generators/torcx-generator[2230]: time="2025-05-10T00:42:26Z" level=info msg="torcx already run" May 10 00:42:26.374581 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. May 10 00:42:26.374592 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. May 10 00:42:26.388779 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 10 00:42:26.436269 systemd[1]: Stopping kubelet.service... May 10 00:42:26.436811 kubelet[1939]: E0510 00:42:26.436215 1939 event.go:319] "Unable to write event (broadcaster is shut down)" event="&Event{ObjectMeta:{localhost.183e03abd7873bd9 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-05-10 00:42:22.056586201 +0000 UTC m=+0.465812965,LastTimestamp:2025-05-10 00:42:22.056586201 +0000 UTC m=+0.465812965,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" May 10 00:42:26.451005 systemd[1]: kubelet.service: Deactivated successfully. May 10 00:42:26.451226 systemd[1]: Stopped kubelet.service. May 10 00:42:26.453217 systemd[1]: Starting kubelet.service... May 10 00:42:27.130800 systemd[1]: Started kubelet.service. May 10 00:42:27.216593 kubelet[2305]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 10 00:42:27.216593 kubelet[2305]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. May 10 00:42:27.216593 kubelet[2305]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 10 00:42:27.216851 kubelet[2305]: I0510 00:42:27.216616 2305 server.go:205] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" May 10 00:42:27.219202 kubelet[2305]: I0510 00:42:27.219192 2305 server.go:484] "Kubelet version" kubeletVersion="v1.30.1" May 10 00:42:27.219270 kubelet[2305]: I0510 00:42:27.219263 2305 server.go:486] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" May 10 00:42:27.219432 kubelet[2305]: I0510 00:42:27.219424 2305 server.go:927] "Client rotation is on, will bootstrap in background" May 10 00:42:27.221316 kubelet[2305]: I0510 00:42:27.221305 2305 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". May 10 00:42:27.222022 kubelet[2305]: I0510 00:42:27.222009 2305 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" May 10 00:42:27.225099 kubelet[2305]: I0510 00:42:27.225089 2305 server.go:742] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" May 10 00:42:27.226244 kubelet[2305]: I0510 00:42:27.226230 2305 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] May 10 00:42:27.226438 kubelet[2305]: I0510 00:42:27.226287 2305 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} May 10 00:42:27.226542 kubelet[2305]: I0510 00:42:27.226534 2305 topology_manager.go:138] "Creating topology manager with none policy" May 10 00:42:27.226593 kubelet[2305]: I0510 00:42:27.226587 2305 container_manager_linux.go:301] "Creating device plugin manager" May 10 00:42:27.226657 kubelet[2305]: I0510 00:42:27.226651 2305 state_mem.go:36] "Initialized new in-memory state store" May 10 00:42:27.226754 kubelet[2305]: I0510 00:42:27.226748 2305 kubelet.go:400] "Attempting to sync node with API server" May 10 00:42:27.226804 kubelet[2305]: I0510 00:42:27.226797 2305 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" May 10 00:42:27.226854 kubelet[2305]: I0510 00:42:27.226848 2305 kubelet.go:312] "Adding apiserver pod source" May 10 00:42:27.226902 kubelet[2305]: I0510 00:42:27.226895 2305 apiserver.go:42] "Waiting for node sync before watching apiserver pods" May 10 00:42:27.228111 kubelet[2305]: I0510 00:42:27.228102 2305 kuberuntime_manager.go:261] "Container runtime initialized" containerRuntime="containerd" version="1.6.16" apiVersion="v1" May 10 00:42:27.228330 kubelet[2305]: I0510 00:42:27.228321 2305 kubelet.go:815] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" May 10 00:42:27.228570 kubelet[2305]: I0510 00:42:27.228563 2305 server.go:1264] "Started kubelet" May 10 00:42:27.229684 kubelet[2305]: I0510 00:42:27.229676 2305 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" May 10 00:42:27.236165 kubelet[2305]: E0510 00:42:27.232027 2305 kubelet.go:1467] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" May 10 00:42:27.236165 kubelet[2305]: I0510 00:42:27.233537 2305 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 May 10 00:42:27.236165 kubelet[2305]: I0510 00:42:27.234096 2305 server.go:455] "Adding debug handlers to kubelet server" May 10 00:42:27.236165 kubelet[2305]: I0510 00:42:27.234502 2305 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 May 10 00:42:27.236165 kubelet[2305]: I0510 00:42:27.234618 2305 server.go:227] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" May 10 00:42:27.245897 kubelet[2305]: I0510 00:42:27.244463 2305 volume_manager.go:291] "Starting Kubelet Volume Manager" May 10 00:42:27.245897 kubelet[2305]: I0510 00:42:27.244531 2305 desired_state_of_world_populator.go:149] "Desired state populator starts to run" May 10 00:42:27.245897 kubelet[2305]: I0510 00:42:27.244606 2305 reconciler.go:26] "Reconciler: start to sync state" May 10 00:42:27.245897 kubelet[2305]: I0510 00:42:27.245105 2305 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory May 10 00:42:27.245897 kubelet[2305]: I0510 00:42:27.245801 2305 factory.go:221] Registration of the containerd container factory successfully May 10 00:42:27.245897 kubelet[2305]: I0510 00:42:27.245808 2305 factory.go:221] Registration of the systemd container factory successfully May 10 00:42:27.258490 kubelet[2305]: I0510 00:42:27.258432 2305 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" May 10 00:42:27.259393 kubelet[2305]: I0510 00:42:27.259380 2305 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" May 10 00:42:27.259447 kubelet[2305]: I0510 00:42:27.259397 2305 status_manager.go:217] "Starting to sync pod status with apiserver" May 10 00:42:27.259447 kubelet[2305]: I0510 00:42:27.259410 2305 kubelet.go:2337] "Starting kubelet main sync loop" May 10 00:42:27.259447 kubelet[2305]: E0510 00:42:27.259433 2305 kubelet.go:2361] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" May 10 00:42:27.267617 sudo[2329]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin May 10 00:42:27.267767 sudo[2329]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=0) May 10 00:42:27.299553 kubelet[2305]: I0510 00:42:27.299535 2305 cpu_manager.go:214] "Starting CPU manager" policy="none" May 10 00:42:27.299553 kubelet[2305]: I0510 00:42:27.299547 2305 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" May 10 00:42:27.299657 kubelet[2305]: I0510 00:42:27.299568 2305 state_mem.go:36] "Initialized new in-memory state store" May 10 00:42:27.299657 kubelet[2305]: I0510 00:42:27.299648 2305 state_mem.go:88] "Updated default CPUSet" cpuSet="" May 10 00:42:27.299710 kubelet[2305]: I0510 00:42:27.299654 2305 state_mem.go:96] "Updated CPUSet assignments" assignments={} May 10 00:42:27.299710 kubelet[2305]: I0510 00:42:27.299665 2305 policy_none.go:49] "None policy: Start" May 10 00:42:27.299986 kubelet[2305]: I0510 00:42:27.299972 2305 memory_manager.go:170] "Starting memorymanager" policy="None" May 10 00:42:27.299986 kubelet[2305]: I0510 00:42:27.299987 2305 state_mem.go:35] "Initializing new in-memory state store" May 10 00:42:27.300084 kubelet[2305]: I0510 00:42:27.300073 2305 state_mem.go:75] "Updated machine memory state" May 10 00:42:27.300671 kubelet[2305]: I0510 00:42:27.300659 2305 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" May 10 00:42:27.300786 kubelet[2305]: I0510 00:42:27.300766 2305 container_log_manager.go:186] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" May 10 00:42:27.300834 kubelet[2305]: I0510 00:42:27.300825 2305 plugin_manager.go:118] "Starting Kubelet Plugin Manager" May 10 00:42:27.345904 kubelet[2305]: I0510 00:42:27.345881 2305 kubelet_node_status.go:73] "Attempting to register node" node="localhost" May 10 00:42:27.350587 kubelet[2305]: I0510 00:42:27.349444 2305 kubelet_node_status.go:112] "Node was previously registered" node="localhost" May 10 00:42:27.350587 kubelet[2305]: I0510 00:42:27.349483 2305 kubelet_node_status.go:76] "Successfully registered node" node="localhost" May 10 00:42:27.361789 kubelet[2305]: I0510 00:42:27.361763 2305 topology_manager.go:215] "Topology Admit Handler" podUID="b00b54c5589197641a7ec2945ad6b746" podNamespace="kube-system" podName="kube-apiserver-localhost" May 10 00:42:27.362333 kubelet[2305]: I0510 00:42:27.362318 2305 topology_manager.go:215] "Topology Admit Handler" podUID="b20b39a8540dba87b5883a6f0f602dba" podNamespace="kube-system" podName="kube-controller-manager-localhost" May 10 00:42:27.362639 kubelet[2305]: I0510 00:42:27.362626 2305 topology_manager.go:215] "Topology Admit Handler" podUID="6ece95f10dbffa04b25ec3439a115512" podNamespace="kube-system" podName="kube-scheduler-localhost" May 10 00:42:27.366065 kubelet[2305]: E0510 00:42:27.366050 2305 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-localhost\" already exists" pod="kube-system/kube-controller-manager-localhost" May 10 00:42:27.368425 kubelet[2305]: E0510 00:42:27.368411 2305 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" May 10 00:42:27.446626 kubelet[2305]: I0510 00:42:27.446562 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 10 00:42:27.446626 kubelet[2305]: I0510 00:42:27.446596 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 10 00:42:27.446626 kubelet[2305]: I0510 00:42:27.446611 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/b00b54c5589197641a7ec2945ad6b746-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"b00b54c5589197641a7ec2945ad6b746\") " pod="kube-system/kube-apiserver-localhost" May 10 00:42:27.446626 kubelet[2305]: I0510 00:42:27.446621 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b00b54c5589197641a7ec2945ad6b746-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"b00b54c5589197641a7ec2945ad6b746\") " pod="kube-system/kube-apiserver-localhost" May 10 00:42:27.446779 kubelet[2305]: I0510 00:42:27.446632 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 10 00:42:27.446779 kubelet[2305]: I0510 00:42:27.446640 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 10 00:42:27.446779 kubelet[2305]: I0510 00:42:27.446649 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b20b39a8540dba87b5883a6f0f602dba-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"b20b39a8540dba87b5883a6f0f602dba\") " pod="kube-system/kube-controller-manager-localhost" May 10 00:42:27.446779 kubelet[2305]: I0510 00:42:27.446658 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/6ece95f10dbffa04b25ec3439a115512-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"6ece95f10dbffa04b25ec3439a115512\") " pod="kube-system/kube-scheduler-localhost" May 10 00:42:27.446779 kubelet[2305]: I0510 00:42:27.446667 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/b00b54c5589197641a7ec2945ad6b746-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"b00b54c5589197641a7ec2945ad6b746\") " pod="kube-system/kube-apiserver-localhost" May 10 00:42:27.747836 sudo[2329]: pam_unix(sudo:session): session closed for user root May 10 00:42:28.228022 kubelet[2305]: I0510 00:42:28.227940 2305 apiserver.go:52] "Watching apiserver" May 10 00:42:28.244895 kubelet[2305]: I0510 00:42:28.244882 2305 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" May 10 00:42:28.283915 kubelet[2305]: E0510 00:42:28.283892 2305 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" May 10 00:42:28.295498 kubelet[2305]: I0510 00:42:28.295400 2305 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.295385995 podStartE2EDuration="1.295385995s" podCreationTimestamp="2025-05-10 00:42:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-10 00:42:28.295271996 +0000 UTC m=+1.143887510" watchObservedRunningTime="2025-05-10 00:42:28.295385995 +0000 UTC m=+1.144001498" May 10 00:42:28.304985 kubelet[2305]: I0510 00:42:28.304954 2305 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=2.304942584 podStartE2EDuration="2.304942584s" podCreationTimestamp="2025-05-10 00:42:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-10 00:42:28.300462838 +0000 UTC m=+1.149078357" watchObservedRunningTime="2025-05-10 00:42:28.304942584 +0000 UTC m=+1.153558096" May 10 00:42:29.106370 sudo[1575]: pam_unix(sudo:session): session closed for user root May 10 00:42:29.110974 sshd[1569]: pam_unix(sshd:session): session closed for user core May 10 00:42:29.112424 systemd[1]: sshd@4-139.178.70.106:22-139.178.89.65:44374.service: Deactivated successfully. May 10 00:42:29.113276 systemd[1]: session-7.scope: Deactivated successfully. May 10 00:42:29.113512 systemd-logind[1342]: Session 7 logged out. Waiting for processes to exit. May 10 00:42:29.114197 systemd-logind[1342]: Removed session 7. May 10 00:42:33.762793 kubelet[2305]: I0510 00:42:33.762747 2305 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=7.7627196099999995 podStartE2EDuration="7.76271961s" podCreationTimestamp="2025-05-10 00:42:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-10 00:42:28.305157783 +0000 UTC m=+1.153773287" watchObservedRunningTime="2025-05-10 00:42:33.76271961 +0000 UTC m=+6.611335119" May 10 00:42:41.304552 kubelet[2305]: I0510 00:42:41.304524 2305 kuberuntime_manager.go:1523] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" May 10 00:42:41.304960 env[1375]: time="2025-05-10T00:42:41.304830498Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." May 10 00:42:41.305209 kubelet[2305]: I0510 00:42:41.305104 2305 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" May 10 00:42:41.693023 kubelet[2305]: I0510 00:42:41.692936 2305 topology_manager.go:215] "Topology Admit Handler" podUID="283d8ef4-0834-4513-82df-d4bafd798112" podNamespace="kube-system" podName="kube-proxy-pssgg" May 10 00:42:41.694478 kubelet[2305]: I0510 00:42:41.694459 2305 topology_manager.go:215] "Topology Admit Handler" podUID="088fca65-9cfd-456a-ab25-9f5687c194bf" podNamespace="kube-system" podName="cilium-qcw9t" May 10 00:42:41.696597 kubelet[2305]: W0510 00:42:41.696565 2305 reflector.go:547] object-"kube-system"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:localhost" cannot list resource "configmaps" in API group "" in the namespace "kube-system": no relationship found between node 'localhost' and this object May 10 00:42:41.696597 kubelet[2305]: E0510 00:42:41.696596 2305 reflector.go:150] object-"kube-system"/"kube-root-ca.crt": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:localhost" cannot list resource "configmaps" in API group "" in the namespace "kube-system": no relationship found between node 'localhost' and this object May 10 00:42:41.696743 kubelet[2305]: W0510 00:42:41.696622 2305 reflector.go:547] object-"kube-system"/"kube-proxy": failed to list *v1.ConfigMap: configmaps "kube-proxy" is forbidden: User "system:node:localhost" cannot list resource "configmaps" in API group "" in the namespace "kube-system": no relationship found between node 'localhost' and this object May 10 00:42:41.696743 kubelet[2305]: E0510 00:42:41.696629 2305 reflector.go:150] object-"kube-system"/"kube-proxy": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps "kube-proxy" is forbidden: User "system:node:localhost" cannot list resource "configmaps" in API group "" in the namespace "kube-system": no relationship found between node 'localhost' and this object May 10 00:42:41.743771 kubelet[2305]: I0510 00:42:41.743733 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/283d8ef4-0834-4513-82df-d4bafd798112-kube-proxy\") pod \"kube-proxy-pssgg\" (UID: \"283d8ef4-0834-4513-82df-d4bafd798112\") " pod="kube-system/kube-proxy-pssgg" May 10 00:42:41.743880 kubelet[2305]: I0510 00:42:41.743786 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/283d8ef4-0834-4513-82df-d4bafd798112-lib-modules\") pod \"kube-proxy-pssgg\" (UID: \"283d8ef4-0834-4513-82df-d4bafd798112\") " pod="kube-system/kube-proxy-pssgg" May 10 00:42:41.743880 kubelet[2305]: I0510 00:42:41.743804 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-cilium-cgroup\") pod \"cilium-qcw9t\" (UID: \"088fca65-9cfd-456a-ab25-9f5687c194bf\") " pod="kube-system/cilium-qcw9t" May 10 00:42:41.743880 kubelet[2305]: I0510 00:42:41.743824 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-host-proc-sys-kernel\") pod \"cilium-qcw9t\" (UID: \"088fca65-9cfd-456a-ab25-9f5687c194bf\") " pod="kube-system/cilium-qcw9t" May 10 00:42:41.743880 kubelet[2305]: I0510 00:42:41.743843 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/283d8ef4-0834-4513-82df-d4bafd798112-xtables-lock\") pod \"kube-proxy-pssgg\" (UID: \"283d8ef4-0834-4513-82df-d4bafd798112\") " pod="kube-system/kube-proxy-pssgg" May 10 00:42:41.743880 kubelet[2305]: I0510 00:42:41.743858 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-cni-path\") pod \"cilium-qcw9t\" (UID: \"088fca65-9cfd-456a-ab25-9f5687c194bf\") " pod="kube-system/cilium-qcw9t" May 10 00:42:41.743880 kubelet[2305]: I0510 00:42:41.743868 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-etc-cni-netd\") pod \"cilium-qcw9t\" (UID: \"088fca65-9cfd-456a-ab25-9f5687c194bf\") " pod="kube-system/cilium-qcw9t" May 10 00:42:41.743999 kubelet[2305]: I0510 00:42:41.743880 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-xtables-lock\") pod \"cilium-qcw9t\" (UID: \"088fca65-9cfd-456a-ab25-9f5687c194bf\") " pod="kube-system/cilium-qcw9t" May 10 00:42:41.743999 kubelet[2305]: I0510 00:42:41.743892 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gzlf\" (UniqueName: \"kubernetes.io/projected/088fca65-9cfd-456a-ab25-9f5687c194bf-kube-api-access-6gzlf\") pod \"cilium-qcw9t\" (UID: \"088fca65-9cfd-456a-ab25-9f5687c194bf\") " pod="kube-system/cilium-qcw9t" May 10 00:42:41.743999 kubelet[2305]: I0510 00:42:41.743903 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-hostproc\") pod \"cilium-qcw9t\" (UID: \"088fca65-9cfd-456a-ab25-9f5687c194bf\") " pod="kube-system/cilium-qcw9t" May 10 00:42:41.743999 kubelet[2305]: I0510 00:42:41.743911 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/088fca65-9cfd-456a-ab25-9f5687c194bf-cilium-config-path\") pod \"cilium-qcw9t\" (UID: \"088fca65-9cfd-456a-ab25-9f5687c194bf\") " pod="kube-system/cilium-qcw9t" May 10 00:42:41.743999 kubelet[2305]: I0510 00:42:41.743920 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-host-proc-sys-net\") pod \"cilium-qcw9t\" (UID: \"088fca65-9cfd-456a-ab25-9f5687c194bf\") " pod="kube-system/cilium-qcw9t" May 10 00:42:41.744091 kubelet[2305]: I0510 00:42:41.743934 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdp87\" (UniqueName: \"kubernetes.io/projected/283d8ef4-0834-4513-82df-d4bafd798112-kube-api-access-qdp87\") pod \"kube-proxy-pssgg\" (UID: \"283d8ef4-0834-4513-82df-d4bafd798112\") " pod="kube-system/kube-proxy-pssgg" May 10 00:42:41.744091 kubelet[2305]: I0510 00:42:41.743946 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-bpf-maps\") pod \"cilium-qcw9t\" (UID: \"088fca65-9cfd-456a-ab25-9f5687c194bf\") " pod="kube-system/cilium-qcw9t" May 10 00:42:41.744091 kubelet[2305]: I0510 00:42:41.743954 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-lib-modules\") pod \"cilium-qcw9t\" (UID: \"088fca65-9cfd-456a-ab25-9f5687c194bf\") " pod="kube-system/cilium-qcw9t" May 10 00:42:41.744091 kubelet[2305]: I0510 00:42:41.743962 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/088fca65-9cfd-456a-ab25-9f5687c194bf-hubble-tls\") pod \"cilium-qcw9t\" (UID: \"088fca65-9cfd-456a-ab25-9f5687c194bf\") " pod="kube-system/cilium-qcw9t" May 10 00:42:41.744091 kubelet[2305]: I0510 00:42:41.743975 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/088fca65-9cfd-456a-ab25-9f5687c194bf-clustermesh-secrets\") pod \"cilium-qcw9t\" (UID: \"088fca65-9cfd-456a-ab25-9f5687c194bf\") " pod="kube-system/cilium-qcw9t" May 10 00:42:41.744091 kubelet[2305]: I0510 00:42:41.743993 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-cilium-run\") pod \"cilium-qcw9t\" (UID: \"088fca65-9cfd-456a-ab25-9f5687c194bf\") " pod="kube-system/cilium-qcw9t" May 10 00:42:41.795371 kubelet[2305]: I0510 00:42:41.795333 2305 topology_manager.go:215] "Topology Admit Handler" podUID="1fdaf146-04ca-4d78-a309-770e5f747397" podNamespace="kube-system" podName="cilium-operator-599987898-wrlnd" May 10 00:42:41.844695 kubelet[2305]: I0510 00:42:41.844665 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/1fdaf146-04ca-4d78-a309-770e5f747397-cilium-config-path\") pod \"cilium-operator-599987898-wrlnd\" (UID: \"1fdaf146-04ca-4d78-a309-770e5f747397\") " pod="kube-system/cilium-operator-599987898-wrlnd" May 10 00:42:41.844936 kubelet[2305]: I0510 00:42:41.844922 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vf5t\" (UniqueName: \"kubernetes.io/projected/1fdaf146-04ca-4d78-a309-770e5f747397-kube-api-access-6vf5t\") pod \"cilium-operator-599987898-wrlnd\" (UID: \"1fdaf146-04ca-4d78-a309-770e5f747397\") " pod="kube-system/cilium-operator-599987898-wrlnd" May 10 00:42:42.865007 kubelet[2305]: E0510 00:42:42.864982 2305 projected.go:294] Couldn't get configMap kube-system/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition May 10 00:42:42.865282 kubelet[2305]: E0510 00:42:42.865272 2305 projected.go:200] Error preparing data for projected volume kube-api-access-6gzlf for pod kube-system/cilium-qcw9t: failed to sync configmap cache: timed out waiting for the condition May 10 00:42:42.865356 kubelet[2305]: E0510 00:42:42.864988 2305 projected.go:294] Couldn't get configMap kube-system/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition May 10 00:42:42.865389 kubelet[2305]: E0510 00:42:42.865360 2305 projected.go:200] Error preparing data for projected volume kube-api-access-qdp87 for pod kube-system/kube-proxy-pssgg: failed to sync configmap cache: timed out waiting for the condition May 10 00:42:42.865446 kubelet[2305]: E0510 00:42:42.865438 2305 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/088fca65-9cfd-456a-ab25-9f5687c194bf-kube-api-access-6gzlf podName:088fca65-9cfd-456a-ab25-9f5687c194bf nodeName:}" failed. No retries permitted until 2025-05-10 00:42:43.365422568 +0000 UTC m=+16.214038072 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-6gzlf" (UniqueName: "kubernetes.io/projected/088fca65-9cfd-456a-ab25-9f5687c194bf-kube-api-access-6gzlf") pod "cilium-qcw9t" (UID: "088fca65-9cfd-456a-ab25-9f5687c194bf") : failed to sync configmap cache: timed out waiting for the condition May 10 00:42:42.865524 kubelet[2305]: E0510 00:42:42.865517 2305 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/283d8ef4-0834-4513-82df-d4bafd798112-kube-api-access-qdp87 podName:283d8ef4-0834-4513-82df-d4bafd798112 nodeName:}" failed. No retries permitted until 2025-05-10 00:42:43.365509441 +0000 UTC m=+16.214124943 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-qdp87" (UniqueName: "kubernetes.io/projected/283d8ef4-0834-4513-82df-d4bafd798112-kube-api-access-qdp87") pod "kube-proxy-pssgg" (UID: "283d8ef4-0834-4513-82df-d4bafd798112") : failed to sync configmap cache: timed out waiting for the condition May 10 00:42:42.950268 kubelet[2305]: E0510 00:42:42.950247 2305 projected.go:294] Couldn't get configMap kube-system/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition May 10 00:42:42.950401 kubelet[2305]: E0510 00:42:42.950390 2305 projected.go:200] Error preparing data for projected volume kube-api-access-6vf5t for pod kube-system/cilium-operator-599987898-wrlnd: failed to sync configmap cache: timed out waiting for the condition May 10 00:42:42.950496 kubelet[2305]: E0510 00:42:42.950485 2305 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1fdaf146-04ca-4d78-a309-770e5f747397-kube-api-access-6vf5t podName:1fdaf146-04ca-4d78-a309-770e5f747397 nodeName:}" failed. No retries permitted until 2025-05-10 00:42:43.450472901 +0000 UTC m=+16.299088412 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-6vf5t" (UniqueName: "kubernetes.io/projected/1fdaf146-04ca-4d78-a309-770e5f747397-kube-api-access-6vf5t") pod "cilium-operator-599987898-wrlnd" (UID: "1fdaf146-04ca-4d78-a309-770e5f747397") : failed to sync configmap cache: timed out waiting for the condition May 10 00:42:43.501650 env[1375]: time="2025-05-10T00:42:43.501617487Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-pssgg,Uid:283d8ef4-0834-4513-82df-d4bafd798112,Namespace:kube-system,Attempt:0,}" May 10 00:42:43.503323 env[1375]: time="2025-05-10T00:42:43.503305332Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-qcw9t,Uid:088fca65-9cfd-456a-ab25-9f5687c194bf,Namespace:kube-system,Attempt:0,}" May 10 00:42:43.563311 env[1375]: time="2025-05-10T00:42:43.563271127Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 10 00:42:43.563409 env[1375]: time="2025-05-10T00:42:43.563321971Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 10 00:42:43.563409 env[1375]: time="2025-05-10T00:42:43.563339393Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 10 00:42:43.563509 env[1375]: time="2025-05-10T00:42:43.563484379Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/6f49f71072fad4a81d5be8c44e9308a0d235941528aafe83b406ca7d718cfc1e pid=2388 runtime=io.containerd.runc.v2 May 10 00:42:43.575213 env[1375]: time="2025-05-10T00:42:43.575147882Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 10 00:42:43.578024 env[1375]: time="2025-05-10T00:42:43.575199096Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 10 00:42:43.578024 env[1375]: time="2025-05-10T00:42:43.575353441Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 10 00:42:43.578024 env[1375]: time="2025-05-10T00:42:43.575446611Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/c0d7c385e734730f32baf6689a7fee7fd110fc809d10710842ffdc2b05b9bb22 pid=2412 runtime=io.containerd.runc.v2 May 10 00:42:43.599879 env[1375]: time="2025-05-10T00:42:43.599850991Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-599987898-wrlnd,Uid:1fdaf146-04ca-4d78-a309-770e5f747397,Namespace:kube-system,Attempt:0,}" May 10 00:42:43.619351 env[1375]: time="2025-05-10T00:42:43.619317668Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-qcw9t,Uid:088fca65-9cfd-456a-ab25-9f5687c194bf,Namespace:kube-system,Attempt:0,} returns sandbox id \"6f49f71072fad4a81d5be8c44e9308a0d235941528aafe83b406ca7d718cfc1e\"" May 10 00:42:43.621628 env[1375]: time="2025-05-10T00:42:43.621602622Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" May 10 00:42:43.629988 env[1375]: time="2025-05-10T00:42:43.629961206Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-pssgg,Uid:283d8ef4-0834-4513-82df-d4bafd798112,Namespace:kube-system,Attempt:0,} returns sandbox id \"c0d7c385e734730f32baf6689a7fee7fd110fc809d10710842ffdc2b05b9bb22\"" May 10 00:42:43.632530 env[1375]: time="2025-05-10T00:42:43.632506308Z" level=info msg="CreateContainer within sandbox \"c0d7c385e734730f32baf6689a7fee7fd110fc809d10710842ffdc2b05b9bb22\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" May 10 00:42:43.634628 env[1375]: time="2025-05-10T00:42:43.634514929Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 10 00:42:43.634628 env[1375]: time="2025-05-10T00:42:43.634536450Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 10 00:42:43.634628 env[1375]: time="2025-05-10T00:42:43.634543350Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 10 00:42:43.634768 env[1375]: time="2025-05-10T00:42:43.634628790Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/b8c6df627179679384c908bd3dd39337b6a0fd4438207707b3eb0517ddeaff99 pid=2472 runtime=io.containerd.runc.v2 May 10 00:42:43.670192 env[1375]: time="2025-05-10T00:42:43.669481215Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-599987898-wrlnd,Uid:1fdaf146-04ca-4d78-a309-770e5f747397,Namespace:kube-system,Attempt:0,} returns sandbox id \"b8c6df627179679384c908bd3dd39337b6a0fd4438207707b3eb0517ddeaff99\"" May 10 00:42:43.704872 env[1375]: time="2025-05-10T00:42:43.704828412Z" level=info msg="CreateContainer within sandbox \"c0d7c385e734730f32baf6689a7fee7fd110fc809d10710842ffdc2b05b9bb22\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"9a908467088d240d3230e6dcc7bc7a0bd8c6bd2d1c576235db2491740f00c147\"" May 10 00:42:43.705923 env[1375]: time="2025-05-10T00:42:43.705903513Z" level=info msg="StartContainer for \"9a908467088d240d3230e6dcc7bc7a0bd8c6bd2d1c576235db2491740f00c147\"" May 10 00:42:43.750355 env[1375]: time="2025-05-10T00:42:43.750319553Z" level=info msg="StartContainer for \"9a908467088d240d3230e6dcc7bc7a0bd8c6bd2d1c576235db2491740f00c147\" returns successfully" May 10 00:42:47.392706 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3439657658.mount: Deactivated successfully. May 10 00:42:50.949000 env[1375]: time="2025-05-10T00:42:50.948968218Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:51.152351 env[1375]: time="2025-05-10T00:42:51.152322268Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:51.173266 env[1375]: time="2025-05-10T00:42:51.173226664Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:51.173609 env[1375]: time="2025-05-10T00:42:51.173583656Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" May 10 00:42:51.174960 env[1375]: time="2025-05-10T00:42:51.174936298Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" May 10 00:42:51.175742 env[1375]: time="2025-05-10T00:42:51.175692669Z" level=info msg="CreateContainer within sandbox \"6f49f71072fad4a81d5be8c44e9308a0d235941528aafe83b406ca7d718cfc1e\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" May 10 00:42:51.184614 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount219781147.mount: Deactivated successfully. May 10 00:42:51.193978 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2322736877.mount: Deactivated successfully. May 10 00:42:51.205722 env[1375]: time="2025-05-10T00:42:51.205645200Z" level=info msg="CreateContainer within sandbox \"6f49f71072fad4a81d5be8c44e9308a0d235941528aafe83b406ca7d718cfc1e\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"8952d1d23ebbe81c5ea39e82c7d33900aaf2b73296c926c67b68908119759b6d\"" May 10 00:42:51.210562 env[1375]: time="2025-05-10T00:42:51.206253706Z" level=info msg="StartContainer for \"8952d1d23ebbe81c5ea39e82c7d33900aaf2b73296c926c67b68908119759b6d\"" May 10 00:42:51.254048 env[1375]: time="2025-05-10T00:42:51.254012156Z" level=info msg="StartContainer for \"8952d1d23ebbe81c5ea39e82c7d33900aaf2b73296c926c67b68908119759b6d\" returns successfully" May 10 00:42:51.369285 kubelet[2305]: I0510 00:42:51.369235 2305 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-pssgg" podStartSLOduration=10.369219306 podStartE2EDuration="10.369219306s" podCreationTimestamp="2025-05-10 00:42:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-10 00:42:44.309217128 +0000 UTC m=+17.157832634" watchObservedRunningTime="2025-05-10 00:42:51.369219306 +0000 UTC m=+24.217834813" May 10 00:42:51.677640 env[1375]: time="2025-05-10T00:42:51.677609858Z" level=info msg="shim disconnected" id=8952d1d23ebbe81c5ea39e82c7d33900aaf2b73296c926c67b68908119759b6d May 10 00:42:51.677822 env[1375]: time="2025-05-10T00:42:51.677811330Z" level=warning msg="cleaning up after shim disconnected" id=8952d1d23ebbe81c5ea39e82c7d33900aaf2b73296c926c67b68908119759b6d namespace=k8s.io May 10 00:42:51.681571 env[1375]: time="2025-05-10T00:42:51.677938317Z" level=info msg="cleaning up dead shim" May 10 00:42:51.682569 env[1375]: time="2025-05-10T00:42:51.682406582Z" level=warning msg="cleanup warnings time=\"2025-05-10T00:42:51Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2711 runtime=io.containerd.runc.v2\n" May 10 00:42:52.182365 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-8952d1d23ebbe81c5ea39e82c7d33900aaf2b73296c926c67b68908119759b6d-rootfs.mount: Deactivated successfully. May 10 00:42:52.324751 env[1375]: time="2025-05-10T00:42:52.321088520Z" level=info msg="CreateContainer within sandbox \"6f49f71072fad4a81d5be8c44e9308a0d235941528aafe83b406ca7d718cfc1e\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" May 10 00:42:52.349716 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3398513740.mount: Deactivated successfully. May 10 00:42:52.354317 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1081846677.mount: Deactivated successfully. May 10 00:42:52.367649 env[1375]: time="2025-05-10T00:42:52.367630280Z" level=info msg="CreateContainer within sandbox \"6f49f71072fad4a81d5be8c44e9308a0d235941528aafe83b406ca7d718cfc1e\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"3ff79049377529728cc46f9178bcc19976126f729934984404722c2f0b27b0b3\"" May 10 00:42:52.372573 env[1375]: time="2025-05-10T00:42:52.368040836Z" level=info msg="StartContainer for \"3ff79049377529728cc46f9178bcc19976126f729934984404722c2f0b27b0b3\"" May 10 00:42:52.404859 env[1375]: time="2025-05-10T00:42:52.404833575Z" level=info msg="StartContainer for \"3ff79049377529728cc46f9178bcc19976126f729934984404722c2f0b27b0b3\" returns successfully" May 10 00:42:52.407838 systemd[1]: systemd-sysctl.service: Deactivated successfully. May 10 00:42:52.407994 systemd[1]: Stopped systemd-sysctl.service. May 10 00:42:52.408201 systemd[1]: Stopping systemd-sysctl.service... May 10 00:42:52.411671 systemd[1]: Starting systemd-sysctl.service... May 10 00:42:52.424811 env[1375]: time="2025-05-10T00:42:52.424762951Z" level=info msg="shim disconnected" id=3ff79049377529728cc46f9178bcc19976126f729934984404722c2f0b27b0b3 May 10 00:42:52.425287 env[1375]: time="2025-05-10T00:42:52.425274257Z" level=warning msg="cleaning up after shim disconnected" id=3ff79049377529728cc46f9178bcc19976126f729934984404722c2f0b27b0b3 namespace=k8s.io May 10 00:42:52.425365 env[1375]: time="2025-05-10T00:42:52.425354632Z" level=info msg="cleaning up dead shim" May 10 00:42:52.426304 systemd[1]: Finished systemd-sysctl.service. May 10 00:42:52.433088 env[1375]: time="2025-05-10T00:42:52.433025742Z" level=warning msg="cleanup warnings time=\"2025-05-10T00:42:52Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2777 runtime=io.containerd.runc.v2\n" May 10 00:42:53.322582 env[1375]: time="2025-05-10T00:42:53.322555795Z" level=info msg="CreateContainer within sandbox \"6f49f71072fad4a81d5be8c44e9308a0d235941528aafe83b406ca7d718cfc1e\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" May 10 00:42:53.423758 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount450647366.mount: Deactivated successfully. May 10 00:42:53.428374 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3181106034.mount: Deactivated successfully. May 10 00:42:53.459561 env[1375]: time="2025-05-10T00:42:53.459529923Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:53.865057 env[1375]: time="2025-05-10T00:42:53.865008838Z" level=info msg="CreateContainer within sandbox \"6f49f71072fad4a81d5be8c44e9308a0d235941528aafe83b406ca7d718cfc1e\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"6a757ac046ea2030a54db32acf4c62e0680307d74c4b09ea6ff15c1e84c74796\"" May 10 00:42:53.865345 env[1375]: time="2025-05-10T00:42:53.865317048Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:53.867290 env[1375]: time="2025-05-10T00:42:53.867266828Z" level=info msg="StartContainer for \"6a757ac046ea2030a54db32acf4c62e0680307d74c4b09ea6ff15c1e84c74796\"" May 10 00:42:53.894219 env[1375]: time="2025-05-10T00:42:53.894011766Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 10 00:42:53.894219 env[1375]: time="2025-05-10T00:42:53.894142085Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" May 10 00:42:53.896368 env[1375]: time="2025-05-10T00:42:53.896348224Z" level=info msg="CreateContainer within sandbox \"b8c6df627179679384c908bd3dd39337b6a0fd4438207707b3eb0517ddeaff99\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" May 10 00:42:53.917661 env[1375]: time="2025-05-10T00:42:53.917639436Z" level=info msg="StartContainer for \"6a757ac046ea2030a54db32acf4c62e0680307d74c4b09ea6ff15c1e84c74796\" returns successfully" May 10 00:42:54.092064 env[1375]: time="2025-05-10T00:42:54.092028255Z" level=info msg="CreateContainer within sandbox \"b8c6df627179679384c908bd3dd39337b6a0fd4438207707b3eb0517ddeaff99\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"adcc7525bb609b9139126018dae923fcdf79eff98c995541fada4214f7b5b834\"" May 10 00:42:54.093823 env[1375]: time="2025-05-10T00:42:54.093792900Z" level=info msg="StartContainer for \"adcc7525bb609b9139126018dae923fcdf79eff98c995541fada4214f7b5b834\"" May 10 00:42:54.102578 env[1375]: time="2025-05-10T00:42:54.102543533Z" level=info msg="shim disconnected" id=6a757ac046ea2030a54db32acf4c62e0680307d74c4b09ea6ff15c1e84c74796 May 10 00:42:54.102750 env[1375]: time="2025-05-10T00:42:54.102739689Z" level=warning msg="cleaning up after shim disconnected" id=6a757ac046ea2030a54db32acf4c62e0680307d74c4b09ea6ff15c1e84c74796 namespace=k8s.io May 10 00:42:54.102804 env[1375]: time="2025-05-10T00:42:54.102794478Z" level=info msg="cleaning up dead shim" May 10 00:42:54.110600 env[1375]: time="2025-05-10T00:42:54.110574600Z" level=warning msg="cleanup warnings time=\"2025-05-10T00:42:54Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2842 runtime=io.containerd.runc.v2\n" May 10 00:42:54.141273 env[1375]: time="2025-05-10T00:42:54.139386520Z" level=info msg="StartContainer for \"adcc7525bb609b9139126018dae923fcdf79eff98c995541fada4214f7b5b834\" returns successfully" May 10 00:42:54.325757 env[1375]: time="2025-05-10T00:42:54.325723426Z" level=info msg="CreateContainer within sandbox \"6f49f71072fad4a81d5be8c44e9308a0d235941528aafe83b406ca7d718cfc1e\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" May 10 00:42:54.334264 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3490950449.mount: Deactivated successfully. May 10 00:42:54.341147 env[1375]: time="2025-05-10T00:42:54.341111152Z" level=info msg="CreateContainer within sandbox \"6f49f71072fad4a81d5be8c44e9308a0d235941528aafe83b406ca7d718cfc1e\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"905227d6256131b22c1d3908dfb084bc598d8af705a928c6d2409fc309dfe62e\"" May 10 00:42:54.341888 env[1375]: time="2025-05-10T00:42:54.341573117Z" level=info msg="StartContainer for \"905227d6256131b22c1d3908dfb084bc598d8af705a928c6d2409fc309dfe62e\"" May 10 00:42:54.343869 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1530620153.mount: Deactivated successfully. May 10 00:42:54.437866 env[1375]: time="2025-05-10T00:42:54.437800703Z" level=info msg="StartContainer for \"905227d6256131b22c1d3908dfb084bc598d8af705a928c6d2409fc309dfe62e\" returns successfully" May 10 00:42:54.512799 env[1375]: time="2025-05-10T00:42:54.512771649Z" level=info msg="shim disconnected" id=905227d6256131b22c1d3908dfb084bc598d8af705a928c6d2409fc309dfe62e May 10 00:42:54.513132 env[1375]: time="2025-05-10T00:42:54.513119978Z" level=warning msg="cleaning up after shim disconnected" id=905227d6256131b22c1d3908dfb084bc598d8af705a928c6d2409fc309dfe62e namespace=k8s.io May 10 00:42:54.513187 env[1375]: time="2025-05-10T00:42:54.513172714Z" level=info msg="cleaning up dead shim" May 10 00:42:54.521928 env[1375]: time="2025-05-10T00:42:54.521895335Z" level=warning msg="cleanup warnings time=\"2025-05-10T00:42:54Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2933 runtime=io.containerd.runc.v2\n" May 10 00:42:55.329939 env[1375]: time="2025-05-10T00:42:55.329913635Z" level=info msg="CreateContainer within sandbox \"6f49f71072fad4a81d5be8c44e9308a0d235941528aafe83b406ca7d718cfc1e\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" May 10 00:42:55.351273 kubelet[2305]: I0510 00:42:55.351237 2305 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-599987898-wrlnd" podStartSLOduration=4.126505048 podStartE2EDuration="14.351224998s" podCreationTimestamp="2025-05-10 00:42:41 +0000 UTC" firstStartedPulling="2025-05-10 00:42:43.670080297 +0000 UTC m=+16.518695801" lastFinishedPulling="2025-05-10 00:42:53.894800249 +0000 UTC m=+26.743415751" observedRunningTime="2025-05-10 00:42:54.429108587 +0000 UTC m=+27.277724099" watchObservedRunningTime="2025-05-10 00:42:55.351224998 +0000 UTC m=+28.199840504" May 10 00:42:55.381675 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount538224298.mount: Deactivated successfully. May 10 00:42:55.385722 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4279794653.mount: Deactivated successfully. May 10 00:42:55.415596 env[1375]: time="2025-05-10T00:42:55.415572598Z" level=info msg="CreateContainer within sandbox \"6f49f71072fad4a81d5be8c44e9308a0d235941528aafe83b406ca7d718cfc1e\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"84c2df00a2d9ae51142c7e672a69177305be8f0719f57135a0e1f6047515e011\"" May 10 00:42:55.416668 env[1375]: time="2025-05-10T00:42:55.416647907Z" level=info msg="StartContainer for \"84c2df00a2d9ae51142c7e672a69177305be8f0719f57135a0e1f6047515e011\"" May 10 00:42:55.466326 env[1375]: time="2025-05-10T00:42:55.466292293Z" level=info msg="StartContainer for \"84c2df00a2d9ae51142c7e672a69177305be8f0719f57135a0e1f6047515e011\" returns successfully" May 10 00:42:55.624217 kubelet[2305]: I0510 00:42:55.624110 2305 kubelet_node_status.go:497] "Fast updating node status as it just became ready" May 10 00:42:55.751869 kubelet[2305]: I0510 00:42:55.751557 2305 topology_manager.go:215] "Topology Admit Handler" podUID="1e354242-44d8-4ef0-a5a3-096c2e405a05" podNamespace="kube-system" podName="coredns-7db6d8ff4d-97bdc" May 10 00:42:55.760828 kubelet[2305]: I0510 00:42:55.760797 2305 topology_manager.go:215] "Topology Admit Handler" podUID="8631dae2-eea4-4527-8929-a9e5a46a55f0" podNamespace="kube-system" podName="coredns-7db6d8ff4d-gmdhp" May 10 00:42:55.835854 kubelet[2305]: I0510 00:42:55.835829 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bj5wm\" (UniqueName: \"kubernetes.io/projected/1e354242-44d8-4ef0-a5a3-096c2e405a05-kube-api-access-bj5wm\") pod \"coredns-7db6d8ff4d-97bdc\" (UID: \"1e354242-44d8-4ef0-a5a3-096c2e405a05\") " pod="kube-system/coredns-7db6d8ff4d-97bdc" May 10 00:42:55.836010 kubelet[2305]: I0510 00:42:55.835997 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8631dae2-eea4-4527-8929-a9e5a46a55f0-config-volume\") pod \"coredns-7db6d8ff4d-gmdhp\" (UID: \"8631dae2-eea4-4527-8929-a9e5a46a55f0\") " pod="kube-system/coredns-7db6d8ff4d-gmdhp" May 10 00:42:55.836089 kubelet[2305]: I0510 00:42:55.836069 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ndgq\" (UniqueName: \"kubernetes.io/projected/8631dae2-eea4-4527-8929-a9e5a46a55f0-kube-api-access-8ndgq\") pod \"coredns-7db6d8ff4d-gmdhp\" (UID: \"8631dae2-eea4-4527-8929-a9e5a46a55f0\") " pod="kube-system/coredns-7db6d8ff4d-gmdhp" May 10 00:42:55.836129 kubelet[2305]: I0510 00:42:55.836108 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e354242-44d8-4ef0-a5a3-096c2e405a05-config-volume\") pod \"coredns-7db6d8ff4d-97bdc\" (UID: \"1e354242-44d8-4ef0-a5a3-096c2e405a05\") " pod="kube-system/coredns-7db6d8ff4d-97bdc" May 10 00:42:56.083912 env[1375]: time="2025-05-10T00:42:56.083864521Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-97bdc,Uid:1e354242-44d8-4ef0-a5a3-096c2e405a05,Namespace:kube-system,Attempt:0,}" May 10 00:42:56.084559 env[1375]: time="2025-05-10T00:42:56.084434941Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-gmdhp,Uid:8631dae2-eea4-4527-8929-a9e5a46a55f0,Namespace:kube-system,Attempt:0,}" May 10 00:42:56.343985 kubelet[2305]: I0510 00:42:56.343652 2305 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-qcw9t" podStartSLOduration=7.789351945 podStartE2EDuration="15.343641472s" podCreationTimestamp="2025-05-10 00:42:41 +0000 UTC" firstStartedPulling="2025-05-10 00:42:43.620341233 +0000 UTC m=+16.468956735" lastFinishedPulling="2025-05-10 00:42:51.174630761 +0000 UTC m=+24.023246262" observedRunningTime="2025-05-10 00:42:56.342501028 +0000 UTC m=+29.191116534" watchObservedRunningTime="2025-05-10 00:42:56.343641472 +0000 UTC m=+29.192256978" May 10 00:42:56.398724 kernel: Spectre V2 : WARNING: Unprivileged eBPF is enabled with eIBRS on, data leaks possible via Spectre v2 BHB attacks! May 10 00:42:56.714717 kernel: Spectre V2 : WARNING: Unprivileged eBPF is enabled with eIBRS on, data leaks possible via Spectre v2 BHB attacks! May 10 00:42:58.565183 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cilium_net: link becomes ready May 10 00:42:58.565674 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cilium_host: link becomes ready May 10 00:42:58.569073 systemd-networkd[1110]: cilium_host: Link UP May 10 00:42:58.569571 systemd-networkd[1110]: cilium_net: Link UP May 10 00:42:58.569859 systemd-networkd[1110]: cilium_net: Gained carrier May 10 00:42:58.570017 systemd-networkd[1110]: cilium_host: Gained carrier May 10 00:42:58.699318 systemd-networkd[1110]: cilium_vxlan: Link UP May 10 00:42:58.699322 systemd-networkd[1110]: cilium_vxlan: Gained carrier May 10 00:42:58.986991 systemd-networkd[1110]: cilium_net: Gained IPv6LL May 10 00:42:59.515970 systemd-networkd[1110]: cilium_host: Gained IPv6LL May 10 00:42:59.681777 kernel: NET: Registered PF_ALG protocol family May 10 00:43:00.282793 systemd-networkd[1110]: cilium_vxlan: Gained IPv6LL May 10 00:43:00.613729 systemd-networkd[1110]: lxc_health: Link UP May 10 00:43:00.621005 systemd-networkd[1110]: lxc_health: Gained carrier May 10 00:43:00.621733 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc_health: link becomes ready May 10 00:43:01.134637 systemd-networkd[1110]: lxcab43190eaf5b: Link UP May 10 00:43:01.140722 kernel: eth0: renamed from tmp39d5c May 10 00:43:01.153878 systemd-networkd[1110]: lxcab43190eaf5b: Gained carrier May 10 00:43:01.154911 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxcab43190eaf5b: link becomes ready May 10 00:43:01.167167 systemd-networkd[1110]: lxc816d5243b75a: Link UP May 10 00:43:01.173764 kernel: eth0: renamed from tmp7ec07 May 10 00:43:01.179048 systemd-networkd[1110]: lxc816d5243b75a: Gained carrier May 10 00:43:01.187955 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc816d5243b75a: link becomes ready May 10 00:43:02.394810 systemd-networkd[1110]: lxcab43190eaf5b: Gained IPv6LL May 10 00:43:02.522810 systemd-networkd[1110]: lxc816d5243b75a: Gained IPv6LL May 10 00:43:02.650832 systemd-networkd[1110]: lxc_health: Gained IPv6LL May 10 00:43:03.767125 env[1375]: time="2025-05-10T00:43:03.767082473Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 10 00:43:03.767125 env[1375]: time="2025-05-10T00:43:03.767108201Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 10 00:43:03.767440 env[1375]: time="2025-05-10T00:43:03.767410023Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 10 00:43:03.767575 env[1375]: time="2025-05-10T00:43:03.767551945Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/7ec0730ca0a00f5b72f18aa3b1e23c8234484496e947d2d1ece579ae4a13e50a pid=3477 runtime=io.containerd.runc.v2 May 10 00:43:03.790816 env[1375]: time="2025-05-10T00:43:03.790751086Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 10 00:43:03.791224 env[1375]: time="2025-05-10T00:43:03.790976730Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 10 00:43:03.791224 env[1375]: time="2025-05-10T00:43:03.790996032Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 10 00:43:03.791224 env[1375]: time="2025-05-10T00:43:03.791106325Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/39d5c0ce2760efe426a530b2ad4c0eeb1347640c3922d8f38d9aafc195b237d0 pid=3503 runtime=io.containerd.runc.v2 May 10 00:43:03.807289 systemd-resolved[1294]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address May 10 00:43:03.857938 env[1375]: time="2025-05-10T00:43:03.857901568Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-gmdhp,Uid:8631dae2-eea4-4527-8929-a9e5a46a55f0,Namespace:kube-system,Attempt:0,} returns sandbox id \"7ec0730ca0a00f5b72f18aa3b1e23c8234484496e947d2d1ece579ae4a13e50a\"" May 10 00:43:03.859000 systemd-resolved[1294]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address May 10 00:43:03.861831 env[1375]: time="2025-05-10T00:43:03.861803922Z" level=info msg="CreateContainer within sandbox \"7ec0730ca0a00f5b72f18aa3b1e23c8234484496e947d2d1ece579ae4a13e50a\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" May 10 00:43:03.876022 env[1375]: time="2025-05-10T00:43:03.875988488Z" level=info msg="CreateContainer within sandbox \"7ec0730ca0a00f5b72f18aa3b1e23c8234484496e947d2d1ece579ae4a13e50a\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"6bcb008136ea7e546217e2de171dab0dc9baece6ea1f691bfe090f03678ade2a\"" May 10 00:43:03.876526 env[1375]: time="2025-05-10T00:43:03.876505997Z" level=info msg="StartContainer for \"6bcb008136ea7e546217e2de171dab0dc9baece6ea1f691bfe090f03678ade2a\"" May 10 00:43:03.897617 env[1375]: time="2025-05-10T00:43:03.897591910Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-97bdc,Uid:1e354242-44d8-4ef0-a5a3-096c2e405a05,Namespace:kube-system,Attempt:0,} returns sandbox id \"39d5c0ce2760efe426a530b2ad4c0eeb1347640c3922d8f38d9aafc195b237d0\"" May 10 00:43:03.901448 env[1375]: time="2025-05-10T00:43:03.901144065Z" level=info msg="CreateContainer within sandbox \"39d5c0ce2760efe426a530b2ad4c0eeb1347640c3922d8f38d9aafc195b237d0\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" May 10 00:43:03.932519 env[1375]: time="2025-05-10T00:43:03.932491351Z" level=info msg="CreateContainer within sandbox \"39d5c0ce2760efe426a530b2ad4c0eeb1347640c3922d8f38d9aafc195b237d0\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"18061a5ac38328ebc5ec633f9e4806b9eb12f7cbc85e3e63f2d68a3b995e2409\"" May 10 00:43:03.935737 env[1375]: time="2025-05-10T00:43:03.933270275Z" level=info msg="StartContainer for \"18061a5ac38328ebc5ec633f9e4806b9eb12f7cbc85e3e63f2d68a3b995e2409\"" May 10 00:43:03.954262 env[1375]: time="2025-05-10T00:43:03.954236335Z" level=info msg="StartContainer for \"6bcb008136ea7e546217e2de171dab0dc9baece6ea1f691bfe090f03678ade2a\" returns successfully" May 10 00:43:03.979563 env[1375]: time="2025-05-10T00:43:03.979531530Z" level=info msg="StartContainer for \"18061a5ac38328ebc5ec633f9e4806b9eb12f7cbc85e3e63f2d68a3b995e2409\" returns successfully" May 10 00:43:04.383788 kubelet[2305]: I0510 00:43:04.383730 2305 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7db6d8ff4d-97bdc" podStartSLOduration=23.383709084 podStartE2EDuration="23.383709084s" podCreationTimestamp="2025-05-10 00:42:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-10 00:43:04.382827295 +0000 UTC m=+37.231442800" watchObservedRunningTime="2025-05-10 00:43:04.383709084 +0000 UTC m=+37.232324586" May 10 00:43:04.384294 kubelet[2305]: I0510 00:43:04.384263 2305 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7db6d8ff4d-gmdhp" podStartSLOduration=23.384252347 podStartE2EDuration="23.384252347s" podCreationTimestamp="2025-05-10 00:42:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-10 00:43:04.368009723 +0000 UTC m=+37.216625229" watchObservedRunningTime="2025-05-10 00:43:04.384252347 +0000 UTC m=+37.232867858" May 10 00:43:04.771486 systemd[1]: run-containerd-runc-k8s.io-39d5c0ce2760efe426a530b2ad4c0eeb1347640c3922d8f38d9aafc195b237d0-runc.TxevCx.mount: Deactivated successfully. May 10 00:43:26.734179 systemd[1]: Started sshd@5-139.178.70.106:22-27.115.42.62:48109.service. May 10 00:43:28.511331 sshd[3636]: Invalid user samba from 27.115.42.62 port 48109 May 10 00:43:28.519539 sshd[3636]: pam_faillock(sshd:auth): User unknown May 10 00:43:28.520637 sshd[3636]: pam_unix(sshd:auth): check pass; user unknown May 10 00:43:28.520741 sshd[3636]: pam_unix(sshd:auth): authentication failure; logname= uid=0 euid=0 tty=ssh ruser= rhost=27.115.42.62 May 10 00:43:28.521173 sshd[3636]: pam_faillock(sshd:auth): User unknown May 10 00:43:30.792485 sshd[3636]: Failed password for invalid user samba from 27.115.42.62 port 48109 ssh2 May 10 00:43:31.238590 sshd[3640]: pam_faillock(sshd:auth): User unknown May 10 00:43:31.246851 sshd[3636]: Postponed keyboard-interactive for invalid user samba from 27.115.42.62 port 48109 ssh2 [preauth] May 10 00:43:31.701233 sshd[3640]: pam_unix(sshd:auth): check pass; user unknown May 10 00:43:31.701584 sshd[3640]: pam_faillock(sshd:auth): User unknown May 10 00:43:33.717655 sshd[3636]: PAM: Permission denied for illegal user samba from 27.115.42.62 May 10 00:43:33.717909 sshd[3636]: Failed keyboard-interactive/pam for invalid user samba from 27.115.42.62 port 48109 ssh2 May 10 00:43:34.140657 sshd[3636]: Connection closed by invalid user samba 27.115.42.62 port 48109 [preauth] May 10 00:43:34.141599 systemd[1]: sshd@5-139.178.70.106:22-27.115.42.62:48109.service: Deactivated successfully. May 10 00:43:39.301192 systemd[1]: Started sshd@6-139.178.70.106:22-139.178.89.65:47498.service. May 10 00:43:39.333215 sshd[3643]: Accepted publickey for core from 139.178.89.65 port 47498 ssh2: RSA SHA256:XIiY2M1wm02XXDRKVwO0PLWdmJcZoEVRw8Nio1wxBSQ May 10 00:43:39.334834 sshd[3643]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 10 00:43:39.339783 systemd-logind[1342]: New session 8 of user core. May 10 00:43:39.340234 systemd[1]: Started session-8.scope. May 10 00:43:39.537599 sshd[3643]: pam_unix(sshd:session): session closed for user core May 10 00:43:39.540121 systemd[1]: sshd@6-139.178.70.106:22-139.178.89.65:47498.service: Deactivated successfully. May 10 00:43:39.540782 systemd-logind[1342]: Session 8 logged out. Waiting for processes to exit. May 10 00:43:39.540817 systemd[1]: session-8.scope: Deactivated successfully. May 10 00:43:39.541619 systemd-logind[1342]: Removed session 8. May 10 00:43:44.540080 systemd[1]: Started sshd@7-139.178.70.106:22-139.178.89.65:47504.service. May 10 00:43:44.647371 sshd[3659]: Accepted publickey for core from 139.178.89.65 port 47504 ssh2: RSA SHA256:XIiY2M1wm02XXDRKVwO0PLWdmJcZoEVRw8Nio1wxBSQ May 10 00:43:44.648391 sshd[3659]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 10 00:43:44.652076 systemd[1]: Started session-9.scope. May 10 00:43:44.652305 systemd-logind[1342]: New session 9 of user core. May 10 00:43:44.760560 sshd[3659]: pam_unix(sshd:session): session closed for user core May 10 00:43:44.761958 systemd[1]: sshd@7-139.178.70.106:22-139.178.89.65:47504.service: Deactivated successfully. May 10 00:43:44.762545 systemd[1]: session-9.scope: Deactivated successfully. May 10 00:43:44.762654 systemd-logind[1342]: Session 9 logged out. Waiting for processes to exit. May 10 00:43:44.763411 systemd-logind[1342]: Removed session 9. May 10 00:43:49.763393 systemd[1]: Started sshd@8-139.178.70.106:22-139.178.89.65:41798.service. May 10 00:43:49.794106 sshd[3673]: Accepted publickey for core from 139.178.89.65 port 41798 ssh2: RSA SHA256:XIiY2M1wm02XXDRKVwO0PLWdmJcZoEVRw8Nio1wxBSQ May 10 00:43:49.795240 sshd[3673]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 10 00:43:49.798018 systemd[1]: Started session-10.scope. May 10 00:43:49.798255 systemd-logind[1342]: New session 10 of user core. May 10 00:43:49.916949 sshd[3673]: pam_unix(sshd:session): session closed for user core May 10 00:43:49.918357 systemd[1]: sshd@8-139.178.70.106:22-139.178.89.65:41798.service: Deactivated successfully. May 10 00:43:49.919151 systemd[1]: session-10.scope: Deactivated successfully. May 10 00:43:49.919382 systemd-logind[1342]: Session 10 logged out. Waiting for processes to exit. May 10 00:43:49.919924 systemd-logind[1342]: Removed session 10. May 10 00:43:54.921006 systemd[1]: Started sshd@9-139.178.70.106:22-139.178.89.65:41802.service. May 10 00:43:55.162889 sshd[3687]: Accepted publickey for core from 139.178.89.65 port 41802 ssh2: RSA SHA256:XIiY2M1wm02XXDRKVwO0PLWdmJcZoEVRw8Nio1wxBSQ May 10 00:43:55.164459 sshd[3687]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 10 00:43:55.167969 systemd[1]: Started session-11.scope. May 10 00:43:55.168085 systemd-logind[1342]: New session 11 of user core. May 10 00:43:55.315686 sshd[3687]: pam_unix(sshd:session): session closed for user core May 10 00:43:55.317943 systemd[1]: Started sshd@10-139.178.70.106:22-139.178.89.65:41814.service. May 10 00:43:55.321952 systemd[1]: sshd@9-139.178.70.106:22-139.178.89.65:41802.service: Deactivated successfully. May 10 00:43:55.322431 systemd[1]: session-11.scope: Deactivated successfully. May 10 00:43:55.322948 systemd-logind[1342]: Session 11 logged out. Waiting for processes to exit. May 10 00:43:55.323434 systemd-logind[1342]: Removed session 11. May 10 00:43:55.354950 sshd[3698]: Accepted publickey for core from 139.178.89.65 port 41814 ssh2: RSA SHA256:XIiY2M1wm02XXDRKVwO0PLWdmJcZoEVRw8Nio1wxBSQ May 10 00:43:55.356335 sshd[3698]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 10 00:43:55.359761 systemd-logind[1342]: New session 12 of user core. May 10 00:43:55.360587 systemd[1]: Started session-12.scope. May 10 00:43:55.548294 systemd[1]: Started sshd@11-139.178.70.106:22-139.178.89.65:41824.service. May 10 00:43:55.553878 sshd[3698]: pam_unix(sshd:session): session closed for user core May 10 00:43:55.557263 systemd[1]: sshd@10-139.178.70.106:22-139.178.89.65:41814.service: Deactivated successfully. May 10 00:43:55.558212 systemd[1]: session-12.scope: Deactivated successfully. May 10 00:43:55.558546 systemd-logind[1342]: Session 12 logged out. Waiting for processes to exit. May 10 00:43:55.562083 systemd-logind[1342]: Removed session 12. May 10 00:43:55.636125 sshd[3709]: Accepted publickey for core from 139.178.89.65 port 41824 ssh2: RSA SHA256:XIiY2M1wm02XXDRKVwO0PLWdmJcZoEVRw8Nio1wxBSQ May 10 00:43:55.636624 sshd[3709]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 10 00:43:55.639504 systemd[1]: Started session-13.scope. May 10 00:43:55.639619 systemd-logind[1342]: New session 13 of user core. May 10 00:43:55.733836 sshd[3709]: pam_unix(sshd:session): session closed for user core May 10 00:43:55.735349 systemd[1]: sshd@11-139.178.70.106:22-139.178.89.65:41824.service: Deactivated successfully. May 10 00:43:55.736080 systemd[1]: session-13.scope: Deactivated successfully. May 10 00:43:55.736329 systemd-logind[1342]: Session 13 logged out. Waiting for processes to exit. May 10 00:43:55.736896 systemd-logind[1342]: Removed session 13. May 10 00:44:00.737292 systemd[1]: Started sshd@12-139.178.70.106:22-139.178.89.65:55706.service. May 10 00:44:00.769706 sshd[3725]: Accepted publickey for core from 139.178.89.65 port 55706 ssh2: RSA SHA256:XIiY2M1wm02XXDRKVwO0PLWdmJcZoEVRw8Nio1wxBSQ May 10 00:44:00.770878 sshd[3725]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 10 00:44:00.773872 systemd[1]: Started session-14.scope. May 10 00:44:00.774201 systemd-logind[1342]: New session 14 of user core. May 10 00:44:00.880741 sshd[3725]: pam_unix(sshd:session): session closed for user core May 10 00:44:00.882397 systemd[1]: sshd@12-139.178.70.106:22-139.178.89.65:55706.service: Deactivated successfully. May 10 00:44:00.883400 systemd[1]: session-14.scope: Deactivated successfully. May 10 00:44:00.883880 systemd-logind[1342]: Session 14 logged out. Waiting for processes to exit. May 10 00:44:00.884477 systemd-logind[1342]: Removed session 14. May 10 00:44:05.883118 systemd[1]: Started sshd@13-139.178.70.106:22-139.178.89.65:55716.service. May 10 00:44:05.912841 sshd[3738]: Accepted publickey for core from 139.178.89.65 port 55716 ssh2: RSA SHA256:XIiY2M1wm02XXDRKVwO0PLWdmJcZoEVRw8Nio1wxBSQ May 10 00:44:05.914036 sshd[3738]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 10 00:44:05.916939 systemd[1]: Started session-15.scope. May 10 00:44:05.917087 systemd-logind[1342]: New session 15 of user core. May 10 00:44:06.011084 sshd[3738]: pam_unix(sshd:session): session closed for user core May 10 00:44:06.012917 systemd[1]: Started sshd@14-139.178.70.106:22-139.178.89.65:55724.service. May 10 00:44:06.015710 systemd[1]: sshd@13-139.178.70.106:22-139.178.89.65:55716.service: Deactivated successfully. May 10 00:44:06.016111 systemd[1]: session-15.scope: Deactivated successfully. May 10 00:44:06.016331 systemd-logind[1342]: Session 15 logged out. Waiting for processes to exit. May 10 00:44:06.016929 systemd-logind[1342]: Removed session 15. May 10 00:44:06.043984 sshd[3749]: Accepted publickey for core from 139.178.89.65 port 55724 ssh2: RSA SHA256:XIiY2M1wm02XXDRKVwO0PLWdmJcZoEVRw8Nio1wxBSQ May 10 00:44:06.045176 sshd[3749]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 10 00:44:06.047878 systemd[1]: Started session-16.scope. May 10 00:44:06.048060 systemd-logind[1342]: New session 16 of user core. May 10 00:44:06.456322 sshd[3749]: pam_unix(sshd:session): session closed for user core May 10 00:44:06.458163 systemd[1]: Started sshd@15-139.178.70.106:22-139.178.89.65:55728.service. May 10 00:44:06.461116 systemd[1]: sshd@14-139.178.70.106:22-139.178.89.65:55724.service: Deactivated successfully. May 10 00:44:06.461854 systemd-logind[1342]: Session 16 logged out. Waiting for processes to exit. May 10 00:44:06.462230 systemd[1]: session-16.scope: Deactivated successfully. May 10 00:44:06.462692 systemd-logind[1342]: Removed session 16. May 10 00:44:06.498235 sshd[3761]: Accepted publickey for core from 139.178.89.65 port 55728 ssh2: RSA SHA256:XIiY2M1wm02XXDRKVwO0PLWdmJcZoEVRw8Nio1wxBSQ May 10 00:44:06.499250 sshd[3761]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 10 00:44:06.504317 systemd[1]: Started session-17.scope. May 10 00:44:06.504620 systemd-logind[1342]: New session 17 of user core. May 10 00:44:07.732437 systemd[1]: Started sshd@16-139.178.70.106:22-139.178.89.65:43376.service. May 10 00:44:07.734268 sshd[3761]: pam_unix(sshd:session): session closed for user core May 10 00:44:07.744284 systemd[1]: sshd@15-139.178.70.106:22-139.178.89.65:55728.service: Deactivated successfully. May 10 00:44:07.745035 systemd[1]: session-17.scope: Deactivated successfully. May 10 00:44:07.745126 systemd-logind[1342]: Session 17 logged out. Waiting for processes to exit. May 10 00:44:07.746689 systemd-logind[1342]: Removed session 17. May 10 00:44:07.787463 sshd[3777]: Accepted publickey for core from 139.178.89.65 port 43376 ssh2: RSA SHA256:XIiY2M1wm02XXDRKVwO0PLWdmJcZoEVRw8Nio1wxBSQ May 10 00:44:07.788614 sshd[3777]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 10 00:44:07.791792 systemd[1]: Started session-18.scope. May 10 00:44:07.792098 systemd-logind[1342]: New session 18 of user core. May 10 00:44:08.015888 sshd[3777]: pam_unix(sshd:session): session closed for user core May 10 00:44:08.017433 systemd[1]: Started sshd@17-139.178.70.106:22-139.178.89.65:43386.service. May 10 00:44:08.020506 systemd[1]: sshd@16-139.178.70.106:22-139.178.89.65:43376.service: Deactivated successfully. May 10 00:44:08.021406 systemd[1]: session-18.scope: Deactivated successfully. May 10 00:44:08.024003 systemd-logind[1342]: Session 18 logged out. Waiting for processes to exit. May 10 00:44:08.026859 systemd-logind[1342]: Removed session 18. May 10 00:44:08.057247 sshd[3789]: Accepted publickey for core from 139.178.89.65 port 43386 ssh2: RSA SHA256:XIiY2M1wm02XXDRKVwO0PLWdmJcZoEVRw8Nio1wxBSQ May 10 00:44:08.058122 sshd[3789]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 10 00:44:08.060974 systemd[1]: Started session-19.scope. May 10 00:44:08.061175 systemd-logind[1342]: New session 19 of user core. May 10 00:44:08.158669 sshd[3789]: pam_unix(sshd:session): session closed for user core May 10 00:44:08.160481 systemd[1]: sshd@17-139.178.70.106:22-139.178.89.65:43386.service: Deactivated successfully. May 10 00:44:08.160927 systemd[1]: session-19.scope: Deactivated successfully. May 10 00:44:08.161210 systemd-logind[1342]: Session 19 logged out. Waiting for processes to exit. May 10 00:44:08.161751 systemd-logind[1342]: Removed session 19. May 10 00:44:13.160850 systemd[1]: Started sshd@18-139.178.70.106:22-139.178.89.65:43402.service. May 10 00:44:13.190261 sshd[3807]: Accepted publickey for core from 139.178.89.65 port 43402 ssh2: RSA SHA256:XIiY2M1wm02XXDRKVwO0PLWdmJcZoEVRw8Nio1wxBSQ May 10 00:44:13.191566 sshd[3807]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 10 00:44:13.194110 systemd-logind[1342]: New session 20 of user core. May 10 00:44:13.194477 systemd[1]: Started session-20.scope. May 10 00:44:13.280820 sshd[3807]: pam_unix(sshd:session): session closed for user core May 10 00:44:13.282321 systemd-logind[1342]: Session 20 logged out. Waiting for processes to exit. May 10 00:44:13.282412 systemd[1]: sshd@18-139.178.70.106:22-139.178.89.65:43402.service: Deactivated successfully. May 10 00:44:13.282903 systemd[1]: session-20.scope: Deactivated successfully. May 10 00:44:13.283201 systemd-logind[1342]: Removed session 20. May 10 00:44:18.283266 systemd[1]: Started sshd@19-139.178.70.106:22-139.178.89.65:56616.service. May 10 00:44:18.312985 sshd[3822]: Accepted publickey for core from 139.178.89.65 port 56616 ssh2: RSA SHA256:XIiY2M1wm02XXDRKVwO0PLWdmJcZoEVRw8Nio1wxBSQ May 10 00:44:18.314031 sshd[3822]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 10 00:44:18.316496 systemd-logind[1342]: New session 21 of user core. May 10 00:44:18.316827 systemd[1]: Started session-21.scope. May 10 00:44:18.400282 sshd[3822]: pam_unix(sshd:session): session closed for user core May 10 00:44:18.402103 systemd[1]: sshd@19-139.178.70.106:22-139.178.89.65:56616.service: Deactivated successfully. May 10 00:44:18.402954 systemd[1]: session-21.scope: Deactivated successfully. May 10 00:44:18.403205 systemd-logind[1342]: Session 21 logged out. Waiting for processes to exit. May 10 00:44:18.403636 systemd-logind[1342]: Removed session 21. May 10 00:44:23.403211 systemd[1]: Started sshd@20-139.178.70.106:22-139.178.89.65:56624.service. May 10 00:44:23.433975 sshd[3835]: Accepted publickey for core from 139.178.89.65 port 56624 ssh2: RSA SHA256:XIiY2M1wm02XXDRKVwO0PLWdmJcZoEVRw8Nio1wxBSQ May 10 00:44:23.435191 sshd[3835]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 10 00:44:23.437989 systemd-logind[1342]: New session 22 of user core. May 10 00:44:23.438375 systemd[1]: Started session-22.scope. May 10 00:44:23.536826 sshd[3835]: pam_unix(sshd:session): session closed for user core May 10 00:44:23.538707 systemd[1]: Started sshd@21-139.178.70.106:22-139.178.89.65:56626.service. May 10 00:44:23.540528 systemd[1]: sshd@20-139.178.70.106:22-139.178.89.65:56624.service: Deactivated successfully. May 10 00:44:23.541217 systemd-logind[1342]: Session 22 logged out. Waiting for processes to exit. May 10 00:44:23.541261 systemd[1]: session-22.scope: Deactivated successfully. May 10 00:44:23.542001 systemd-logind[1342]: Removed session 22. May 10 00:44:23.570638 sshd[3845]: Accepted publickey for core from 139.178.89.65 port 56626 ssh2: RSA SHA256:XIiY2M1wm02XXDRKVwO0PLWdmJcZoEVRw8Nio1wxBSQ May 10 00:44:23.571877 sshd[3845]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 10 00:44:23.575541 systemd[1]: Started session-23.scope. May 10 00:44:23.576077 systemd-logind[1342]: New session 23 of user core. May 10 00:44:25.270323 env[1375]: time="2025-05-10T00:44:25.270289640Z" level=info msg="StopContainer for \"adcc7525bb609b9139126018dae923fcdf79eff98c995541fada4214f7b5b834\" with timeout 30 (s)" May 10 00:44:25.270821 env[1375]: time="2025-05-10T00:44:25.270793139Z" level=info msg="Stop container \"adcc7525bb609b9139126018dae923fcdf79eff98c995541fada4214f7b5b834\" with signal terminated" May 10 00:44:25.308367 systemd[1]: run-containerd-runc-k8s.io-84c2df00a2d9ae51142c7e672a69177305be8f0719f57135a0e1f6047515e011-runc.aNtJPn.mount: Deactivated successfully. May 10 00:44:25.318975 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-adcc7525bb609b9139126018dae923fcdf79eff98c995541fada4214f7b5b834-rootfs.mount: Deactivated successfully. May 10 00:44:25.326088 env[1375]: time="2025-05-10T00:44:25.326044350Z" level=info msg="shim disconnected" id=adcc7525bb609b9139126018dae923fcdf79eff98c995541fada4214f7b5b834 May 10 00:44:25.326088 env[1375]: time="2025-05-10T00:44:25.326084599Z" level=warning msg="cleaning up after shim disconnected" id=adcc7525bb609b9139126018dae923fcdf79eff98c995541fada4214f7b5b834 namespace=k8s.io May 10 00:44:25.326374 env[1375]: time="2025-05-10T00:44:25.326098751Z" level=info msg="cleaning up dead shim" May 10 00:44:25.334566 env[1375]: time="2025-05-10T00:44:25.334527837Z" level=warning msg="cleanup warnings time=\"2025-05-10T00:44:25Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3892 runtime=io.containerd.runc.v2\n" May 10 00:44:25.336163 env[1375]: time="2025-05-10T00:44:25.336133728Z" level=info msg="StopContainer for \"adcc7525bb609b9139126018dae923fcdf79eff98c995541fada4214f7b5b834\" returns successfully" May 10 00:44:25.336641 env[1375]: time="2025-05-10T00:44:25.336616917Z" level=info msg="StopPodSandbox for \"b8c6df627179679384c908bd3dd39337b6a0fd4438207707b3eb0517ddeaff99\"" May 10 00:44:25.336690 env[1375]: time="2025-05-10T00:44:25.336672642Z" level=info msg="Container to stop \"adcc7525bb609b9139126018dae923fcdf79eff98c995541fada4214f7b5b834\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 10 00:44:25.338519 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-b8c6df627179679384c908bd3dd39337b6a0fd4438207707b3eb0517ddeaff99-shm.mount: Deactivated successfully. May 10 00:44:25.347869 env[1375]: time="2025-05-10T00:44:25.347464526Z" level=error msg="failed to reload cni configuration after receiving fs change event(\"/etc/cni/net.d/05-cilium.conf\": REMOVE)" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" May 10 00:44:25.351636 env[1375]: time="2025-05-10T00:44:25.351607765Z" level=info msg="StopContainer for \"84c2df00a2d9ae51142c7e672a69177305be8f0719f57135a0e1f6047515e011\" with timeout 2 (s)" May 10 00:44:25.351888 env[1375]: time="2025-05-10T00:44:25.351869227Z" level=info msg="Stop container \"84c2df00a2d9ae51142c7e672a69177305be8f0719f57135a0e1f6047515e011\" with signal terminated" May 10 00:44:25.363950 systemd-networkd[1110]: lxc_health: Link DOWN May 10 00:44:25.364261 systemd-networkd[1110]: lxc_health: Lost carrier May 10 00:44:25.403715 env[1375]: time="2025-05-10T00:44:25.403106205Z" level=info msg="shim disconnected" id=b8c6df627179679384c908bd3dd39337b6a0fd4438207707b3eb0517ddeaff99 May 10 00:44:25.403715 env[1375]: time="2025-05-10T00:44:25.403635202Z" level=warning msg="cleaning up after shim disconnected" id=b8c6df627179679384c908bd3dd39337b6a0fd4438207707b3eb0517ddeaff99 namespace=k8s.io May 10 00:44:25.403715 env[1375]: time="2025-05-10T00:44:25.403645930Z" level=info msg="cleaning up dead shim" May 10 00:44:25.415096 env[1375]: time="2025-05-10T00:44:25.415055155Z" level=warning msg="cleanup warnings time=\"2025-05-10T00:44:25Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3936 runtime=io.containerd.runc.v2\n" May 10 00:44:25.417011 env[1375]: time="2025-05-10T00:44:25.416914404Z" level=info msg="TearDown network for sandbox \"b8c6df627179679384c908bd3dd39337b6a0fd4438207707b3eb0517ddeaff99\" successfully" May 10 00:44:25.417011 env[1375]: time="2025-05-10T00:44:25.416941952Z" level=info msg="StopPodSandbox for \"b8c6df627179679384c908bd3dd39337b6a0fd4438207707b3eb0517ddeaff99\" returns successfully" May 10 00:44:25.456329 env[1375]: time="2025-05-10T00:44:25.456290582Z" level=info msg="shim disconnected" id=84c2df00a2d9ae51142c7e672a69177305be8f0719f57135a0e1f6047515e011 May 10 00:44:25.456489 env[1375]: time="2025-05-10T00:44:25.456473076Z" level=warning msg="cleaning up after shim disconnected" id=84c2df00a2d9ae51142c7e672a69177305be8f0719f57135a0e1f6047515e011 namespace=k8s.io May 10 00:44:25.456542 env[1375]: time="2025-05-10T00:44:25.456532516Z" level=info msg="cleaning up dead shim" May 10 00:44:25.463116 env[1375]: time="2025-05-10T00:44:25.463081041Z" level=warning msg="cleanup warnings time=\"2025-05-10T00:44:25Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3962 runtime=io.containerd.runc.v2\n" May 10 00:44:25.463926 env[1375]: time="2025-05-10T00:44:25.463873818Z" level=info msg="StopContainer for \"84c2df00a2d9ae51142c7e672a69177305be8f0719f57135a0e1f6047515e011\" returns successfully" May 10 00:44:25.464271 env[1375]: time="2025-05-10T00:44:25.464246553Z" level=info msg="StopPodSandbox for \"6f49f71072fad4a81d5be8c44e9308a0d235941528aafe83b406ca7d718cfc1e\"" May 10 00:44:25.464370 env[1375]: time="2025-05-10T00:44:25.464357844Z" level=info msg="Container to stop \"905227d6256131b22c1d3908dfb084bc598d8af705a928c6d2409fc309dfe62e\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 10 00:44:25.464429 env[1375]: time="2025-05-10T00:44:25.464415363Z" level=info msg="Container to stop \"6a757ac046ea2030a54db32acf4c62e0680307d74c4b09ea6ff15c1e84c74796\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 10 00:44:25.464489 env[1375]: time="2025-05-10T00:44:25.464471740Z" level=info msg="Container to stop \"8952d1d23ebbe81c5ea39e82c7d33900aaf2b73296c926c67b68908119759b6d\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 10 00:44:25.464549 env[1375]: time="2025-05-10T00:44:25.464532953Z" level=info msg="Container to stop \"3ff79049377529728cc46f9178bcc19976126f729934984404722c2f0b27b0b3\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 10 00:44:25.464614 env[1375]: time="2025-05-10T00:44:25.464599099Z" level=info msg="Container to stop \"84c2df00a2d9ae51142c7e672a69177305be8f0719f57135a0e1f6047515e011\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 10 00:44:25.511131 env[1375]: time="2025-05-10T00:44:25.511083748Z" level=info msg="shim disconnected" id=6f49f71072fad4a81d5be8c44e9308a0d235941528aafe83b406ca7d718cfc1e May 10 00:44:25.511131 env[1375]: time="2025-05-10T00:44:25.511127128Z" level=warning msg="cleaning up after shim disconnected" id=6f49f71072fad4a81d5be8c44e9308a0d235941528aafe83b406ca7d718cfc1e namespace=k8s.io May 10 00:44:25.511267 env[1375]: time="2025-05-10T00:44:25.511138844Z" level=info msg="cleaning up dead shim" May 10 00:44:25.516501 env[1375]: time="2025-05-10T00:44:25.516476010Z" level=warning msg="cleanup warnings time=\"2025-05-10T00:44:25Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3995 runtime=io.containerd.runc.v2\n" May 10 00:44:25.523185 env[1375]: time="2025-05-10T00:44:25.523129732Z" level=info msg="TearDown network for sandbox \"6f49f71072fad4a81d5be8c44e9308a0d235941528aafe83b406ca7d718cfc1e\" successfully" May 10 00:44:25.523185 env[1375]: time="2025-05-10T00:44:25.523149997Z" level=info msg="StopPodSandbox for \"6f49f71072fad4a81d5be8c44e9308a0d235941528aafe83b406ca7d718cfc1e\" returns successfully" May 10 00:44:25.524578 kubelet[2305]: I0510 00:44:25.524477 2305 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vf5t\" (UniqueName: \"kubernetes.io/projected/1fdaf146-04ca-4d78-a309-770e5f747397-kube-api-access-6vf5t\") pod \"1fdaf146-04ca-4d78-a309-770e5f747397\" (UID: \"1fdaf146-04ca-4d78-a309-770e5f747397\") " May 10 00:44:25.524578 kubelet[2305]: I0510 00:44:25.524521 2305 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/1fdaf146-04ca-4d78-a309-770e5f747397-cilium-config-path\") pod \"1fdaf146-04ca-4d78-a309-770e5f747397\" (UID: \"1fdaf146-04ca-4d78-a309-770e5f747397\") " May 10 00:44:25.603986 kubelet[2305]: I0510 00:44:25.600601 2305 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fdaf146-04ca-4d78-a309-770e5f747397-kube-api-access-6vf5t" (OuterVolumeSpecName: "kube-api-access-6vf5t") pod "1fdaf146-04ca-4d78-a309-770e5f747397" (UID: "1fdaf146-04ca-4d78-a309-770e5f747397"). InnerVolumeSpecName "kube-api-access-6vf5t". PluginName "kubernetes.io/projected", VolumeGidValue "" May 10 00:44:25.604150 kubelet[2305]: I0510 00:44:25.604135 2305 scope.go:117] "RemoveContainer" containerID="adcc7525bb609b9139126018dae923fcdf79eff98c995541fada4214f7b5b834" May 10 00:44:25.604822 kubelet[2305]: I0510 00:44:25.576097 2305 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1fdaf146-04ca-4d78-a309-770e5f747397-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "1fdaf146-04ca-4d78-a309-770e5f747397" (UID: "1fdaf146-04ca-4d78-a309-770e5f747397"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" May 10 00:44:25.605202 env[1375]: time="2025-05-10T00:44:25.605171888Z" level=info msg="RemoveContainer for \"adcc7525bb609b9139126018dae923fcdf79eff98c995541fada4214f7b5b834\"" May 10 00:44:25.631443 env[1375]: time="2025-05-10T00:44:25.631409349Z" level=info msg="RemoveContainer for \"adcc7525bb609b9139126018dae923fcdf79eff98c995541fada4214f7b5b834\" returns successfully" May 10 00:44:25.631648 kubelet[2305]: I0510 00:44:25.631625 2305 scope.go:117] "RemoveContainer" containerID="adcc7525bb609b9139126018dae923fcdf79eff98c995541fada4214f7b5b834" May 10 00:44:25.631858 env[1375]: time="2025-05-10T00:44:25.631792191Z" level=error msg="ContainerStatus for \"adcc7525bb609b9139126018dae923fcdf79eff98c995541fada4214f7b5b834\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"adcc7525bb609b9139126018dae923fcdf79eff98c995541fada4214f7b5b834\": not found" May 10 00:44:25.642516 kubelet[2305]: I0510 00:44:25.642488 2305 reconciler_common.go:289] "Volume detached for volume \"kube-api-access-6vf5t\" (UniqueName: \"kubernetes.io/projected/1fdaf146-04ca-4d78-a309-770e5f747397-kube-api-access-6vf5t\") on node \"localhost\" DevicePath \"\"" May 10 00:44:25.642625 kubelet[2305]: I0510 00:44:25.642613 2305 reconciler_common.go:289] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/1fdaf146-04ca-4d78-a309-770e5f747397-cilium-config-path\") on node \"localhost\" DevicePath \"\"" May 10 00:44:25.664596 kubelet[2305]: E0510 00:44:25.648594 2305 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"adcc7525bb609b9139126018dae923fcdf79eff98c995541fada4214f7b5b834\": not found" containerID="adcc7525bb609b9139126018dae923fcdf79eff98c995541fada4214f7b5b834" May 10 00:44:25.670108 kubelet[2305]: I0510 00:44:25.670039 2305 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"adcc7525bb609b9139126018dae923fcdf79eff98c995541fada4214f7b5b834"} err="failed to get container status \"adcc7525bb609b9139126018dae923fcdf79eff98c995541fada4214f7b5b834\": rpc error: code = NotFound desc = an error occurred when try to find container \"adcc7525bb609b9139126018dae923fcdf79eff98c995541fada4214f7b5b834\": not found" May 10 00:44:25.670163 kubelet[2305]: I0510 00:44:25.670109 2305 scope.go:117] "RemoveContainer" containerID="84c2df00a2d9ae51142c7e672a69177305be8f0719f57135a0e1f6047515e011" May 10 00:44:25.679254 env[1375]: time="2025-05-10T00:44:25.670915859Z" level=info msg="RemoveContainer for \"84c2df00a2d9ae51142c7e672a69177305be8f0719f57135a0e1f6047515e011\"" May 10 00:44:25.692874 env[1375]: time="2025-05-10T00:44:25.692851138Z" level=info msg="RemoveContainer for \"84c2df00a2d9ae51142c7e672a69177305be8f0719f57135a0e1f6047515e011\" returns successfully" May 10 00:44:25.692979 kubelet[2305]: I0510 00:44:25.692964 2305 scope.go:117] "RemoveContainer" containerID="905227d6256131b22c1d3908dfb084bc598d8af705a928c6d2409fc309dfe62e" May 10 00:44:25.693688 env[1375]: time="2025-05-10T00:44:25.693665725Z" level=info msg="RemoveContainer for \"905227d6256131b22c1d3908dfb084bc598d8af705a928c6d2409fc309dfe62e\"" May 10 00:44:25.719881 env[1375]: time="2025-05-10T00:44:25.719851197Z" level=info msg="RemoveContainer for \"905227d6256131b22c1d3908dfb084bc598d8af705a928c6d2409fc309dfe62e\" returns successfully" May 10 00:44:25.719971 kubelet[2305]: I0510 00:44:25.719956 2305 scope.go:117] "RemoveContainer" containerID="6a757ac046ea2030a54db32acf4c62e0680307d74c4b09ea6ff15c1e84c74796" May 10 00:44:25.720652 env[1375]: time="2025-05-10T00:44:25.720634108Z" level=info msg="RemoveContainer for \"6a757ac046ea2030a54db32acf4c62e0680307d74c4b09ea6ff15c1e84c74796\"" May 10 00:44:25.740159 env[1375]: time="2025-05-10T00:44:25.740137172Z" level=info msg="RemoveContainer for \"6a757ac046ea2030a54db32acf4c62e0680307d74c4b09ea6ff15c1e84c74796\" returns successfully" May 10 00:44:25.749716 env[1375]: time="2025-05-10T00:44:25.741328446Z" level=info msg="RemoveContainer for \"3ff79049377529728cc46f9178bcc19976126f729934984404722c2f0b27b0b3\"" May 10 00:44:25.749762 kubelet[2305]: I0510 00:44:25.740382 2305 scope.go:117] "RemoveContainer" containerID="3ff79049377529728cc46f9178bcc19976126f729934984404722c2f0b27b0b3" May 10 00:44:25.749762 kubelet[2305]: I0510 00:44:25.743770 2305 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-cilium-cgroup\") pod \"088fca65-9cfd-456a-ab25-9f5687c194bf\" (UID: \"088fca65-9cfd-456a-ab25-9f5687c194bf\") " May 10 00:44:25.749762 kubelet[2305]: I0510 00:44:25.743811 2305 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/088fca65-9cfd-456a-ab25-9f5687c194bf-clustermesh-secrets\") pod \"088fca65-9cfd-456a-ab25-9f5687c194bf\" (UID: \"088fca65-9cfd-456a-ab25-9f5687c194bf\") " May 10 00:44:25.749762 kubelet[2305]: I0510 00:44:25.743826 2305 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-host-proc-sys-net\") pod \"088fca65-9cfd-456a-ab25-9f5687c194bf\" (UID: \"088fca65-9cfd-456a-ab25-9f5687c194bf\") " May 10 00:44:25.749762 kubelet[2305]: I0510 00:44:25.743839 2305 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-hostproc\") pod \"088fca65-9cfd-456a-ab25-9f5687c194bf\" (UID: \"088fca65-9cfd-456a-ab25-9f5687c194bf\") " May 10 00:44:25.749762 kubelet[2305]: I0510 00:44:25.743853 2305 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/088fca65-9cfd-456a-ab25-9f5687c194bf-cilium-config-path\") pod \"088fca65-9cfd-456a-ab25-9f5687c194bf\" (UID: \"088fca65-9cfd-456a-ab25-9f5687c194bf\") " May 10 00:44:25.749762 kubelet[2305]: I0510 00:44:25.743879 2305 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-lib-modules\") pod \"088fca65-9cfd-456a-ab25-9f5687c194bf\" (UID: \"088fca65-9cfd-456a-ab25-9f5687c194bf\") " May 10 00:44:25.749953 kubelet[2305]: I0510 00:44:25.743894 2305 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-host-proc-sys-kernel\") pod \"088fca65-9cfd-456a-ab25-9f5687c194bf\" (UID: \"088fca65-9cfd-456a-ab25-9f5687c194bf\") " May 10 00:44:25.749953 kubelet[2305]: I0510 00:44:25.743905 2305 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-etc-cni-netd\") pod \"088fca65-9cfd-456a-ab25-9f5687c194bf\" (UID: \"088fca65-9cfd-456a-ab25-9f5687c194bf\") " May 10 00:44:25.749953 kubelet[2305]: I0510 00:44:25.743919 2305 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gzlf\" (UniqueName: \"kubernetes.io/projected/088fca65-9cfd-456a-ab25-9f5687c194bf-kube-api-access-6gzlf\") pod \"088fca65-9cfd-456a-ab25-9f5687c194bf\" (UID: \"088fca65-9cfd-456a-ab25-9f5687c194bf\") " May 10 00:44:25.749953 kubelet[2305]: I0510 00:44:25.743929 2305 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-cilium-run\") pod \"088fca65-9cfd-456a-ab25-9f5687c194bf\" (UID: \"088fca65-9cfd-456a-ab25-9f5687c194bf\") " May 10 00:44:25.749953 kubelet[2305]: I0510 00:44:25.743950 2305 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-cni-path\") pod \"088fca65-9cfd-456a-ab25-9f5687c194bf\" (UID: \"088fca65-9cfd-456a-ab25-9f5687c194bf\") " May 10 00:44:25.749953 kubelet[2305]: I0510 00:44:25.743963 2305 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-xtables-lock\") pod \"088fca65-9cfd-456a-ab25-9f5687c194bf\" (UID: \"088fca65-9cfd-456a-ab25-9f5687c194bf\") " May 10 00:44:25.750101 kubelet[2305]: I0510 00:44:25.743974 2305 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-bpf-maps\") pod \"088fca65-9cfd-456a-ab25-9f5687c194bf\" (UID: \"088fca65-9cfd-456a-ab25-9f5687c194bf\") " May 10 00:44:25.750101 kubelet[2305]: I0510 00:44:25.743986 2305 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/088fca65-9cfd-456a-ab25-9f5687c194bf-hubble-tls\") pod \"088fca65-9cfd-456a-ab25-9f5687c194bf\" (UID: \"088fca65-9cfd-456a-ab25-9f5687c194bf\") " May 10 00:44:25.753957 kubelet[2305]: I0510 00:44:25.753939 2305 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "088fca65-9cfd-456a-ab25-9f5687c194bf" (UID: "088fca65-9cfd-456a-ab25-9f5687c194bf"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 10 00:44:25.754042 kubelet[2305]: I0510 00:44:25.753942 2305 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "088fca65-9cfd-456a-ab25-9f5687c194bf" (UID: "088fca65-9cfd-456a-ab25-9f5687c194bf"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 10 00:44:25.754104 kubelet[2305]: I0510 00:44:25.753960 2305 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "088fca65-9cfd-456a-ab25-9f5687c194bf" (UID: "088fca65-9cfd-456a-ab25-9f5687c194bf"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 10 00:44:25.754284 kubelet[2305]: I0510 00:44:25.753970 2305 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "088fca65-9cfd-456a-ab25-9f5687c194bf" (UID: "088fca65-9cfd-456a-ab25-9f5687c194bf"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 10 00:44:25.754348 kubelet[2305]: I0510 00:44:25.754047 2305 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-cni-path" (OuterVolumeSpecName: "cni-path") pod "088fca65-9cfd-456a-ab25-9f5687c194bf" (UID: "088fca65-9cfd-456a-ab25-9f5687c194bf"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 10 00:44:25.754409 kubelet[2305]: I0510 00:44:25.754058 2305 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-hostproc" (OuterVolumeSpecName: "hostproc") pod "088fca65-9cfd-456a-ab25-9f5687c194bf" (UID: "088fca65-9cfd-456a-ab25-9f5687c194bf"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 10 00:44:25.754467 kubelet[2305]: I0510 00:44:25.754066 2305 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "088fca65-9cfd-456a-ab25-9f5687c194bf" (UID: "088fca65-9cfd-456a-ab25-9f5687c194bf"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 10 00:44:25.754523 kubelet[2305]: I0510 00:44:25.754074 2305 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "088fca65-9cfd-456a-ab25-9f5687c194bf" (UID: "088fca65-9cfd-456a-ab25-9f5687c194bf"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 10 00:44:25.764873 kubelet[2305]: I0510 00:44:25.754117 2305 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "088fca65-9cfd-456a-ab25-9f5687c194bf" (UID: "088fca65-9cfd-456a-ab25-9f5687c194bf"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 10 00:44:25.764873 kubelet[2305]: I0510 00:44:25.754124 2305 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "088fca65-9cfd-456a-ab25-9f5687c194bf" (UID: "088fca65-9cfd-456a-ab25-9f5687c194bf"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 10 00:44:25.764873 kubelet[2305]: I0510 00:44:25.763098 2305 scope.go:117] "RemoveContainer" containerID="8952d1d23ebbe81c5ea39e82c7d33900aaf2b73296c926c67b68908119759b6d" May 10 00:44:25.764989 env[1375]: time="2025-05-10T00:44:25.762964345Z" level=info msg="RemoveContainer for \"3ff79049377529728cc46f9178bcc19976126f729934984404722c2f0b27b0b3\" returns successfully" May 10 00:44:25.764989 env[1375]: time="2025-05-10T00:44:25.763893635Z" level=info msg="RemoveContainer for \"8952d1d23ebbe81c5ea39e82c7d33900aaf2b73296c926c67b68908119759b6d\"" May 10 00:44:25.781677 env[1375]: time="2025-05-10T00:44:25.780595516Z" level=info msg="RemoveContainer for \"8952d1d23ebbe81c5ea39e82c7d33900aaf2b73296c926c67b68908119759b6d\" returns successfully" May 10 00:44:25.781677 env[1375]: time="2025-05-10T00:44:25.780820872Z" level=error msg="ContainerStatus for \"84c2df00a2d9ae51142c7e672a69177305be8f0719f57135a0e1f6047515e011\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"84c2df00a2d9ae51142c7e672a69177305be8f0719f57135a0e1f6047515e011\": not found" May 10 00:44:25.781677 env[1375]: time="2025-05-10T00:44:25.781057031Z" level=error msg="ContainerStatus for \"905227d6256131b22c1d3908dfb084bc598d8af705a928c6d2409fc309dfe62e\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"905227d6256131b22c1d3908dfb084bc598d8af705a928c6d2409fc309dfe62e\": not found" May 10 00:44:25.781677 env[1375]: time="2025-05-10T00:44:25.781272358Z" level=error msg="ContainerStatus for \"6a757ac046ea2030a54db32acf4c62e0680307d74c4b09ea6ff15c1e84c74796\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"6a757ac046ea2030a54db32acf4c62e0680307d74c4b09ea6ff15c1e84c74796\": not found" May 10 00:44:25.781677 env[1375]: time="2025-05-10T00:44:25.781471409Z" level=error msg="ContainerStatus for \"3ff79049377529728cc46f9178bcc19976126f729934984404722c2f0b27b0b3\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"3ff79049377529728cc46f9178bcc19976126f729934984404722c2f0b27b0b3\": not found" May 10 00:44:25.781857 kubelet[2305]: I0510 00:44:25.780709 2305 scope.go:117] "RemoveContainer" containerID="84c2df00a2d9ae51142c7e672a69177305be8f0719f57135a0e1f6047515e011" May 10 00:44:25.781857 kubelet[2305]: E0510 00:44:25.780934 2305 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"84c2df00a2d9ae51142c7e672a69177305be8f0719f57135a0e1f6047515e011\": not found" containerID="84c2df00a2d9ae51142c7e672a69177305be8f0719f57135a0e1f6047515e011" May 10 00:44:25.781857 kubelet[2305]: I0510 00:44:25.780951 2305 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"84c2df00a2d9ae51142c7e672a69177305be8f0719f57135a0e1f6047515e011"} err="failed to get container status \"84c2df00a2d9ae51142c7e672a69177305be8f0719f57135a0e1f6047515e011\": rpc error: code = NotFound desc = an error occurred when try to find container \"84c2df00a2d9ae51142c7e672a69177305be8f0719f57135a0e1f6047515e011\": not found" May 10 00:44:25.781857 kubelet[2305]: I0510 00:44:25.780966 2305 scope.go:117] "RemoveContainer" containerID="905227d6256131b22c1d3908dfb084bc598d8af705a928c6d2409fc309dfe62e" May 10 00:44:25.781857 kubelet[2305]: E0510 00:44:25.781138 2305 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"905227d6256131b22c1d3908dfb084bc598d8af705a928c6d2409fc309dfe62e\": not found" containerID="905227d6256131b22c1d3908dfb084bc598d8af705a928c6d2409fc309dfe62e" May 10 00:44:25.781857 kubelet[2305]: I0510 00:44:25.781151 2305 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"905227d6256131b22c1d3908dfb084bc598d8af705a928c6d2409fc309dfe62e"} err="failed to get container status \"905227d6256131b22c1d3908dfb084bc598d8af705a928c6d2409fc309dfe62e\": rpc error: code = NotFound desc = an error occurred when try to find container \"905227d6256131b22c1d3908dfb084bc598d8af705a928c6d2409fc309dfe62e\": not found" May 10 00:44:25.781857 kubelet[2305]: I0510 00:44:25.781161 2305 scope.go:117] "RemoveContainer" containerID="6a757ac046ea2030a54db32acf4c62e0680307d74c4b09ea6ff15c1e84c74796" May 10 00:44:25.782079 env[1375]: time="2025-05-10T00:44:25.781715990Z" level=error msg="ContainerStatus for \"8952d1d23ebbe81c5ea39e82c7d33900aaf2b73296c926c67b68908119759b6d\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"8952d1d23ebbe81c5ea39e82c7d33900aaf2b73296c926c67b68908119759b6d\": not found" May 10 00:44:25.782111 kubelet[2305]: E0510 00:44:25.781351 2305 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"6a757ac046ea2030a54db32acf4c62e0680307d74c4b09ea6ff15c1e84c74796\": not found" containerID="6a757ac046ea2030a54db32acf4c62e0680307d74c4b09ea6ff15c1e84c74796" May 10 00:44:25.782111 kubelet[2305]: I0510 00:44:25.781363 2305 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"6a757ac046ea2030a54db32acf4c62e0680307d74c4b09ea6ff15c1e84c74796"} err="failed to get container status \"6a757ac046ea2030a54db32acf4c62e0680307d74c4b09ea6ff15c1e84c74796\": rpc error: code = NotFound desc = an error occurred when try to find container \"6a757ac046ea2030a54db32acf4c62e0680307d74c4b09ea6ff15c1e84c74796\": not found" May 10 00:44:25.782111 kubelet[2305]: I0510 00:44:25.781380 2305 scope.go:117] "RemoveContainer" containerID="3ff79049377529728cc46f9178bcc19976126f729934984404722c2f0b27b0b3" May 10 00:44:25.782111 kubelet[2305]: E0510 00:44:25.781554 2305 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"3ff79049377529728cc46f9178bcc19976126f729934984404722c2f0b27b0b3\": not found" containerID="3ff79049377529728cc46f9178bcc19976126f729934984404722c2f0b27b0b3" May 10 00:44:25.782111 kubelet[2305]: I0510 00:44:25.781567 2305 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"3ff79049377529728cc46f9178bcc19976126f729934984404722c2f0b27b0b3"} err="failed to get container status \"3ff79049377529728cc46f9178bcc19976126f729934984404722c2f0b27b0b3\": rpc error: code = NotFound desc = an error occurred when try to find container \"3ff79049377529728cc46f9178bcc19976126f729934984404722c2f0b27b0b3\": not found" May 10 00:44:25.782111 kubelet[2305]: I0510 00:44:25.781578 2305 scope.go:117] "RemoveContainer" containerID="8952d1d23ebbe81c5ea39e82c7d33900aaf2b73296c926c67b68908119759b6d" May 10 00:44:25.788027 kubelet[2305]: E0510 00:44:25.781874 2305 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"8952d1d23ebbe81c5ea39e82c7d33900aaf2b73296c926c67b68908119759b6d\": not found" containerID="8952d1d23ebbe81c5ea39e82c7d33900aaf2b73296c926c67b68908119759b6d" May 10 00:44:25.788027 kubelet[2305]: I0510 00:44:25.781890 2305 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"8952d1d23ebbe81c5ea39e82c7d33900aaf2b73296c926c67b68908119759b6d"} err="failed to get container status \"8952d1d23ebbe81c5ea39e82c7d33900aaf2b73296c926c67b68908119759b6d\": rpc error: code = NotFound desc = an error occurred when try to find container \"8952d1d23ebbe81c5ea39e82c7d33900aaf2b73296c926c67b68908119759b6d\": not found" May 10 00:44:25.788027 kubelet[2305]: I0510 00:44:25.784479 2305 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/088fca65-9cfd-456a-ab25-9f5687c194bf-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "088fca65-9cfd-456a-ab25-9f5687c194bf" (UID: "088fca65-9cfd-456a-ab25-9f5687c194bf"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" May 10 00:44:25.794513 kubelet[2305]: I0510 00:44:25.794485 2305 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/088fca65-9cfd-456a-ab25-9f5687c194bf-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "088fca65-9cfd-456a-ab25-9f5687c194bf" (UID: "088fca65-9cfd-456a-ab25-9f5687c194bf"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" May 10 00:44:25.796022 kubelet[2305]: I0510 00:44:25.796004 2305 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/088fca65-9cfd-456a-ab25-9f5687c194bf-kube-api-access-6gzlf" (OuterVolumeSpecName: "kube-api-access-6gzlf") pod "088fca65-9cfd-456a-ab25-9f5687c194bf" (UID: "088fca65-9cfd-456a-ab25-9f5687c194bf"). InnerVolumeSpecName "kube-api-access-6gzlf". PluginName "kubernetes.io/projected", VolumeGidValue "" May 10 00:44:25.799836 kubelet[2305]: I0510 00:44:25.799811 2305 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/088fca65-9cfd-456a-ab25-9f5687c194bf-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "088fca65-9cfd-456a-ab25-9f5687c194bf" (UID: "088fca65-9cfd-456a-ab25-9f5687c194bf"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" May 10 00:44:25.845070 kubelet[2305]: I0510 00:44:25.845043 2305 reconciler_common.go:289] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-host-proc-sys-kernel\") on node \"localhost\" DevicePath \"\"" May 10 00:44:25.845227 kubelet[2305]: I0510 00:44:25.845214 2305 reconciler_common.go:289] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-etc-cni-netd\") on node \"localhost\" DevicePath \"\"" May 10 00:44:25.845338 kubelet[2305]: I0510 00:44:25.845328 2305 reconciler_common.go:289] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/088fca65-9cfd-456a-ab25-9f5687c194bf-cilium-config-path\") on node \"localhost\" DevicePath \"\"" May 10 00:44:25.845444 kubelet[2305]: I0510 00:44:25.845410 2305 reconciler_common.go:289] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-lib-modules\") on node \"localhost\" DevicePath \"\"" May 10 00:44:25.845520 kubelet[2305]: I0510 00:44:25.845508 2305 reconciler_common.go:289] "Volume detached for volume \"kube-api-access-6gzlf\" (UniqueName: \"kubernetes.io/projected/088fca65-9cfd-456a-ab25-9f5687c194bf-kube-api-access-6gzlf\") on node \"localhost\" DevicePath \"\"" May 10 00:44:25.845588 kubelet[2305]: I0510 00:44:25.845579 2305 reconciler_common.go:289] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-cilium-run\") on node \"localhost\" DevicePath \"\"" May 10 00:44:25.845652 kubelet[2305]: I0510 00:44:25.845642 2305 reconciler_common.go:289] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-cni-path\") on node \"localhost\" DevicePath \"\"" May 10 00:44:25.845752 kubelet[2305]: I0510 00:44:25.845741 2305 reconciler_common.go:289] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-xtables-lock\") on node \"localhost\" DevicePath \"\"" May 10 00:44:25.845830 kubelet[2305]: I0510 00:44:25.845818 2305 reconciler_common.go:289] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-bpf-maps\") on node \"localhost\" DevicePath \"\"" May 10 00:44:25.845894 kubelet[2305]: I0510 00:44:25.845884 2305 reconciler_common.go:289] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/088fca65-9cfd-456a-ab25-9f5687c194bf-hubble-tls\") on node \"localhost\" DevicePath \"\"" May 10 00:44:25.845965 kubelet[2305]: I0510 00:44:25.845953 2305 reconciler_common.go:289] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-cilium-cgroup\") on node \"localhost\" DevicePath \"\"" May 10 00:44:25.846036 kubelet[2305]: I0510 00:44:25.846024 2305 reconciler_common.go:289] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/088fca65-9cfd-456a-ab25-9f5687c194bf-clustermesh-secrets\") on node \"localhost\" DevicePath \"\"" May 10 00:44:25.846105 kubelet[2305]: I0510 00:44:25.846095 2305 reconciler_common.go:289] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-hostproc\") on node \"localhost\" DevicePath \"\"" May 10 00:44:25.846195 kubelet[2305]: I0510 00:44:25.846183 2305 reconciler_common.go:289] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/088fca65-9cfd-456a-ab25-9f5687c194bf-host-proc-sys-net\") on node \"localhost\" DevicePath \"\"" May 10 00:44:26.300064 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-84c2df00a2d9ae51142c7e672a69177305be8f0719f57135a0e1f6047515e011-rootfs.mount: Deactivated successfully. May 10 00:44:26.300184 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-b8c6df627179679384c908bd3dd39337b6a0fd4438207707b3eb0517ddeaff99-rootfs.mount: Deactivated successfully. May 10 00:44:26.300268 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-6f49f71072fad4a81d5be8c44e9308a0d235941528aafe83b406ca7d718cfc1e-rootfs.mount: Deactivated successfully. May 10 00:44:26.300336 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-6f49f71072fad4a81d5be8c44e9308a0d235941528aafe83b406ca7d718cfc1e-shm.mount: Deactivated successfully. May 10 00:44:26.300412 systemd[1]: var-lib-kubelet-pods-1fdaf146\x2d04ca\x2d4d78\x2da309\x2d770e5f747397-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d6vf5t.mount: Deactivated successfully. May 10 00:44:26.300491 systemd[1]: var-lib-kubelet-pods-088fca65\x2d9cfd\x2d456a\x2dab25\x2d9f5687c194bf-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d6gzlf.mount: Deactivated successfully. May 10 00:44:26.300577 systemd[1]: var-lib-kubelet-pods-088fca65\x2d9cfd\x2d456a\x2dab25\x2d9f5687c194bf-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. May 10 00:44:26.300648 systemd[1]: var-lib-kubelet-pods-088fca65\x2d9cfd\x2d456a\x2dab25\x2d9f5687c194bf-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. May 10 00:44:27.226066 systemd[1]: Started sshd@22-139.178.70.106:22-139.178.89.65:47910.service. May 10 00:44:27.232914 sshd[3845]: pam_unix(sshd:session): session closed for user core May 10 00:44:27.234912 systemd[1]: sshd@21-139.178.70.106:22-139.178.89.65:56626.service: Deactivated successfully. May 10 00:44:27.235632 systemd[1]: session-23.scope: Deactivated successfully. May 10 00:44:27.237060 systemd-logind[1342]: Session 23 logged out. Waiting for processes to exit. May 10 00:44:27.237665 systemd-logind[1342]: Removed session 23. May 10 00:44:27.241725 env[1375]: time="2025-05-10T00:44:27.241568406Z" level=info msg="StopPodSandbox for \"6f49f71072fad4a81d5be8c44e9308a0d235941528aafe83b406ca7d718cfc1e\"" May 10 00:44:27.241725 env[1375]: time="2025-05-10T00:44:27.241636970Z" level=info msg="TearDown network for sandbox \"6f49f71072fad4a81d5be8c44e9308a0d235941528aafe83b406ca7d718cfc1e\" successfully" May 10 00:44:27.241725 env[1375]: time="2025-05-10T00:44:27.241664550Z" level=info msg="StopPodSandbox for \"6f49f71072fad4a81d5be8c44e9308a0d235941528aafe83b406ca7d718cfc1e\" returns successfully" May 10 00:44:27.242639 env[1375]: time="2025-05-10T00:44:27.242116489Z" level=info msg="RemovePodSandbox for \"6f49f71072fad4a81d5be8c44e9308a0d235941528aafe83b406ca7d718cfc1e\"" May 10 00:44:27.242639 env[1375]: time="2025-05-10T00:44:27.242137304Z" level=info msg="Forcibly stopping sandbox \"6f49f71072fad4a81d5be8c44e9308a0d235941528aafe83b406ca7d718cfc1e\"" May 10 00:44:27.242639 env[1375]: time="2025-05-10T00:44:27.242188258Z" level=info msg="TearDown network for sandbox \"6f49f71072fad4a81d5be8c44e9308a0d235941528aafe83b406ca7d718cfc1e\" successfully" May 10 00:44:27.244527 env[1375]: time="2025-05-10T00:44:27.244447725Z" level=info msg="RemovePodSandbox \"6f49f71072fad4a81d5be8c44e9308a0d235941528aafe83b406ca7d718cfc1e\" returns successfully" May 10 00:44:27.244875 env[1375]: time="2025-05-10T00:44:27.244748548Z" level=info msg="StopPodSandbox for \"b8c6df627179679384c908bd3dd39337b6a0fd4438207707b3eb0517ddeaff99\"" May 10 00:44:27.244875 env[1375]: time="2025-05-10T00:44:27.244803973Z" level=info msg="TearDown network for sandbox \"b8c6df627179679384c908bd3dd39337b6a0fd4438207707b3eb0517ddeaff99\" successfully" May 10 00:44:27.244875 env[1375]: time="2025-05-10T00:44:27.244829515Z" level=info msg="StopPodSandbox for \"b8c6df627179679384c908bd3dd39337b6a0fd4438207707b3eb0517ddeaff99\" returns successfully" May 10 00:44:27.245384 env[1375]: time="2025-05-10T00:44:27.245038234Z" level=info msg="RemovePodSandbox for \"b8c6df627179679384c908bd3dd39337b6a0fd4438207707b3eb0517ddeaff99\"" May 10 00:44:27.245432 env[1375]: time="2025-05-10T00:44:27.245385449Z" level=info msg="Forcibly stopping sandbox \"b8c6df627179679384c908bd3dd39337b6a0fd4438207707b3eb0517ddeaff99\"" May 10 00:44:27.245480 env[1375]: time="2025-05-10T00:44:27.245459104Z" level=info msg="TearDown network for sandbox \"b8c6df627179679384c908bd3dd39337b6a0fd4438207707b3eb0517ddeaff99\" successfully" May 10 00:44:27.250172 env[1375]: time="2025-05-10T00:44:27.250141940Z" level=info msg="RemovePodSandbox \"b8c6df627179679384c908bd3dd39337b6a0fd4438207707b3eb0517ddeaff99\" returns successfully" May 10 00:44:27.265325 kubelet[2305]: I0510 00:44:27.265297 2305 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="088fca65-9cfd-456a-ab25-9f5687c194bf" path="/var/lib/kubelet/pods/088fca65-9cfd-456a-ab25-9f5687c194bf/volumes" May 10 00:44:27.267153 kubelet[2305]: I0510 00:44:27.267133 2305 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fdaf146-04ca-4d78-a309-770e5f747397" path="/var/lib/kubelet/pods/1fdaf146-04ca-4d78-a309-770e5f747397/volumes" May 10 00:44:27.288518 sshd[4022]: Accepted publickey for core from 139.178.89.65 port 47910 ssh2: RSA SHA256:XIiY2M1wm02XXDRKVwO0PLWdmJcZoEVRw8Nio1wxBSQ May 10 00:44:27.289782 sshd[4022]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 10 00:44:27.297385 systemd[1]: Started session-24.scope. May 10 00:44:27.298087 systemd-logind[1342]: New session 24 of user core. May 10 00:44:27.333194 kubelet[2305]: E0510 00:44:27.333167 2305 kubelet.go:2900] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" May 10 00:44:27.872682 systemd[1]: Started sshd@23-139.178.70.106:22-139.178.89.65:47912.service. May 10 00:44:27.878637 sshd[4022]: pam_unix(sshd:session): session closed for user core May 10 00:44:27.882240 systemd[1]: sshd@22-139.178.70.106:22-139.178.89.65:47910.service: Deactivated successfully. May 10 00:44:27.882891 systemd[1]: session-24.scope: Deactivated successfully. May 10 00:44:27.884576 systemd-logind[1342]: Session 24 logged out. Waiting for processes to exit. May 10 00:44:27.885299 systemd-logind[1342]: Removed session 24. May 10 00:44:27.905890 kubelet[2305]: I0510 00:44:27.900381 2305 topology_manager.go:215] "Topology Admit Handler" podUID="79537fb0-5b06-4d58-bdc3-9a3409a62e1a" podNamespace="kube-system" podName="cilium-fmd9x" May 10 00:44:27.905890 kubelet[2305]: E0510 00:44:27.904175 2305 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="088fca65-9cfd-456a-ab25-9f5687c194bf" containerName="mount-bpf-fs" May 10 00:44:27.905890 kubelet[2305]: E0510 00:44:27.904215 2305 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="1fdaf146-04ca-4d78-a309-770e5f747397" containerName="cilium-operator" May 10 00:44:27.905890 kubelet[2305]: E0510 00:44:27.904223 2305 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="088fca65-9cfd-456a-ab25-9f5687c194bf" containerName="clean-cilium-state" May 10 00:44:27.905890 kubelet[2305]: E0510 00:44:27.904227 2305 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="088fca65-9cfd-456a-ab25-9f5687c194bf" containerName="cilium-agent" May 10 00:44:27.905890 kubelet[2305]: E0510 00:44:27.904231 2305 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="088fca65-9cfd-456a-ab25-9f5687c194bf" containerName="mount-cgroup" May 10 00:44:27.905890 kubelet[2305]: E0510 00:44:27.904234 2305 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="088fca65-9cfd-456a-ab25-9f5687c194bf" containerName="apply-sysctl-overwrites" May 10 00:44:27.905890 kubelet[2305]: I0510 00:44:27.904264 2305 memory_manager.go:354] "RemoveStaleState removing state" podUID="088fca65-9cfd-456a-ab25-9f5687c194bf" containerName="cilium-agent" May 10 00:44:27.905890 kubelet[2305]: I0510 00:44:27.904269 2305 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fdaf146-04ca-4d78-a309-770e5f747397" containerName="cilium-operator" May 10 00:44:27.933060 sshd[4036]: Accepted publickey for core from 139.178.89.65 port 47912 ssh2: RSA SHA256:XIiY2M1wm02XXDRKVwO0PLWdmJcZoEVRw8Nio1wxBSQ May 10 00:44:27.934204 sshd[4036]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 10 00:44:27.937276 systemd[1]: Started session-25.scope. May 10 00:44:27.937506 systemd-logind[1342]: New session 25 of user core. May 10 00:44:28.147038 kubelet[2305]: I0510 00:44:28.146629 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-cilium-run\") pod \"cilium-fmd9x\" (UID: \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\") " pod="kube-system/cilium-fmd9x" May 10 00:44:28.147038 kubelet[2305]: I0510 00:44:28.146662 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-lib-modules\") pod \"cilium-fmd9x\" (UID: \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\") " pod="kube-system/cilium-fmd9x" May 10 00:44:28.147038 kubelet[2305]: I0510 00:44:28.146723 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-hubble-tls\") pod \"cilium-fmd9x\" (UID: \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\") " pod="kube-system/cilium-fmd9x" May 10 00:44:28.147038 kubelet[2305]: I0510 00:44:28.146738 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-cilium-cgroup\") pod \"cilium-fmd9x\" (UID: \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\") " pod="kube-system/cilium-fmd9x" May 10 00:44:28.147038 kubelet[2305]: I0510 00:44:28.146750 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-host-proc-sys-net\") pod \"cilium-fmd9x\" (UID: \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\") " pod="kube-system/cilium-fmd9x" May 10 00:44:28.147038 kubelet[2305]: I0510 00:44:28.146760 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-host-proc-sys-kernel\") pod \"cilium-fmd9x\" (UID: \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\") " pod="kube-system/cilium-fmd9x" May 10 00:44:28.147246 kubelet[2305]: I0510 00:44:28.146771 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-hostproc\") pod \"cilium-fmd9x\" (UID: \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\") " pod="kube-system/cilium-fmd9x" May 10 00:44:28.147246 kubelet[2305]: I0510 00:44:28.146787 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-cni-path\") pod \"cilium-fmd9x\" (UID: \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\") " pod="kube-system/cilium-fmd9x" May 10 00:44:28.147246 kubelet[2305]: I0510 00:44:28.146801 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-clustermesh-secrets\") pod \"cilium-fmd9x\" (UID: \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\") " pod="kube-system/cilium-fmd9x" May 10 00:44:28.147246 kubelet[2305]: I0510 00:44:28.146811 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-cilium-config-path\") pod \"cilium-fmd9x\" (UID: \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\") " pod="kube-system/cilium-fmd9x" May 10 00:44:28.147246 kubelet[2305]: I0510 00:44:28.146824 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-cilium-ipsec-secrets\") pod \"cilium-fmd9x\" (UID: \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\") " pod="kube-system/cilium-fmd9x" May 10 00:44:28.147246 kubelet[2305]: I0510 00:44:28.146833 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-bpf-maps\") pod \"cilium-fmd9x\" (UID: \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\") " pod="kube-system/cilium-fmd9x" May 10 00:44:28.147368 kubelet[2305]: I0510 00:44:28.146842 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-xtables-lock\") pod \"cilium-fmd9x\" (UID: \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\") " pod="kube-system/cilium-fmd9x" May 10 00:44:28.147368 kubelet[2305]: I0510 00:44:28.146859 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-etc-cni-netd\") pod \"cilium-fmd9x\" (UID: \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\") " pod="kube-system/cilium-fmd9x" May 10 00:44:28.147368 kubelet[2305]: I0510 00:44:28.146871 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrvv4\" (UniqueName: \"kubernetes.io/projected/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-kube-api-access-mrvv4\") pod \"cilium-fmd9x\" (UID: \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\") " pod="kube-system/cilium-fmd9x" May 10 00:44:28.363496 env[1375]: time="2025-05-10T00:44:28.363442223Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-fmd9x,Uid:79537fb0-5b06-4d58-bdc3-9a3409a62e1a,Namespace:kube-system,Attempt:0,}" May 10 00:44:28.405913 env[1375]: time="2025-05-10T00:44:28.405812740Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 10 00:44:28.405913 env[1375]: time="2025-05-10T00:44:28.405847088Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 10 00:44:28.406082 env[1375]: time="2025-05-10T00:44:28.406053013Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 10 00:44:28.406851 env[1375]: time="2025-05-10T00:44:28.406243388Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/034f81679d0e4106f73f701ad825f880acc67fede9a153e3b92e18b442ca8bad pid=4060 runtime=io.containerd.runc.v2 May 10 00:44:28.438813 env[1375]: time="2025-05-10T00:44:28.438768331Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-fmd9x,Uid:79537fb0-5b06-4d58-bdc3-9a3409a62e1a,Namespace:kube-system,Attempt:0,} returns sandbox id \"034f81679d0e4106f73f701ad825f880acc67fede9a153e3b92e18b442ca8bad\"" May 10 00:44:28.440909 env[1375]: time="2025-05-10T00:44:28.440677686Z" level=info msg="CreateContainer within sandbox \"034f81679d0e4106f73f701ad825f880acc67fede9a153e3b92e18b442ca8bad\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" May 10 00:44:28.486458 env[1375]: time="2025-05-10T00:44:28.486425029Z" level=info msg="CreateContainer within sandbox \"034f81679d0e4106f73f701ad825f880acc67fede9a153e3b92e18b442ca8bad\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"b0b979ef7ab0e55652b063140781bfa7a1e56c9c0278d871751af1d23089f159\"" May 10 00:44:28.487852 env[1375]: time="2025-05-10T00:44:28.487831660Z" level=info msg="StartContainer for \"b0b979ef7ab0e55652b063140781bfa7a1e56c9c0278d871751af1d23089f159\"" May 10 00:44:28.539570 env[1375]: time="2025-05-10T00:44:28.539530962Z" level=info msg="StartContainer for \"b0b979ef7ab0e55652b063140781bfa7a1e56c9c0278d871751af1d23089f159\" returns successfully" May 10 00:44:28.572835 systemd[1]: Started sshd@24-139.178.70.106:22-139.178.89.65:47922.service. May 10 00:44:28.573829 sshd[4036]: pam_unix(sshd:session): session closed for user core May 10 00:44:28.582236 systemd[1]: sshd@23-139.178.70.106:22-139.178.89.65:47912.service: Deactivated successfully. May 10 00:44:28.582746 systemd[1]: session-25.scope: Deactivated successfully. May 10 00:44:28.585911 systemd-logind[1342]: Session 25 logged out. Waiting for processes to exit. May 10 00:44:28.586536 systemd-logind[1342]: Removed session 25. May 10 00:44:28.624841 sshd[4124]: Accepted publickey for core from 139.178.89.65 port 47922 ssh2: RSA SHA256:XIiY2M1wm02XXDRKVwO0PLWdmJcZoEVRw8Nio1wxBSQ May 10 00:44:28.625666 sshd[4124]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 10 00:44:28.628584 systemd[1]: Started session-26.scope. May 10 00:44:28.628921 systemd-logind[1342]: New session 26 of user core. May 10 00:44:28.741559 env[1375]: time="2025-05-10T00:44:28.741514054Z" level=info msg="shim disconnected" id=b0b979ef7ab0e55652b063140781bfa7a1e56c9c0278d871751af1d23089f159 May 10 00:44:28.741559 env[1375]: time="2025-05-10T00:44:28.741556743Z" level=warning msg="cleaning up after shim disconnected" id=b0b979ef7ab0e55652b063140781bfa7a1e56c9c0278d871751af1d23089f159 namespace=k8s.io May 10 00:44:28.741742 env[1375]: time="2025-05-10T00:44:28.741565744Z" level=info msg="cleaning up dead shim" May 10 00:44:28.747292 env[1375]: time="2025-05-10T00:44:28.747247914Z" level=warning msg="cleanup warnings time=\"2025-05-10T00:44:28Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4155 runtime=io.containerd.runc.v2\n" May 10 00:44:28.896972 kubelet[2305]: I0510 00:44:28.896941 2305 setters.go:580] "Node became not ready" node="localhost" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-05-10T00:44:28Z","lastTransitionTime":"2025-05-10T00:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} May 10 00:44:29.553790 env[1375]: time="2025-05-10T00:44:29.553741364Z" level=info msg="StopPodSandbox for \"034f81679d0e4106f73f701ad825f880acc67fede9a153e3b92e18b442ca8bad\"" May 10 00:44:29.553790 env[1375]: time="2025-05-10T00:44:29.553789788Z" level=info msg="Container to stop \"b0b979ef7ab0e55652b063140781bfa7a1e56c9c0278d871751af1d23089f159\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 10 00:44:29.555498 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-034f81679d0e4106f73f701ad825f880acc67fede9a153e3b92e18b442ca8bad-shm.mount: Deactivated successfully. May 10 00:44:29.574060 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-034f81679d0e4106f73f701ad825f880acc67fede9a153e3b92e18b442ca8bad-rootfs.mount: Deactivated successfully. May 10 00:44:29.580100 env[1375]: time="2025-05-10T00:44:29.580052243Z" level=info msg="shim disconnected" id=034f81679d0e4106f73f701ad825f880acc67fede9a153e3b92e18b442ca8bad May 10 00:44:29.580100 env[1375]: time="2025-05-10T00:44:29.580093904Z" level=warning msg="cleaning up after shim disconnected" id=034f81679d0e4106f73f701ad825f880acc67fede9a153e3b92e18b442ca8bad namespace=k8s.io May 10 00:44:29.580100 env[1375]: time="2025-05-10T00:44:29.580101052Z" level=info msg="cleaning up dead shim" May 10 00:44:29.585738 env[1375]: time="2025-05-10T00:44:29.585697118Z" level=warning msg="cleanup warnings time=\"2025-05-10T00:44:29Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4188 runtime=io.containerd.runc.v2\n" May 10 00:44:29.585936 env[1375]: time="2025-05-10T00:44:29.585912953Z" level=info msg="TearDown network for sandbox \"034f81679d0e4106f73f701ad825f880acc67fede9a153e3b92e18b442ca8bad\" successfully" May 10 00:44:29.585936 env[1375]: time="2025-05-10T00:44:29.585933495Z" level=info msg="StopPodSandbox for \"034f81679d0e4106f73f701ad825f880acc67fede9a153e3b92e18b442ca8bad\" returns successfully" May 10 00:44:29.756667 kubelet[2305]: I0510 00:44:29.756623 2305 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-clustermesh-secrets\") pod \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\" (UID: \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\") " May 10 00:44:29.756667 kubelet[2305]: I0510 00:44:29.756660 2305 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-cilium-run\") pod \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\" (UID: \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\") " May 10 00:44:29.756862 kubelet[2305]: I0510 00:44:29.756825 2305 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-host-proc-sys-kernel\") pod \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\" (UID: \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\") " May 10 00:44:29.756862 kubelet[2305]: I0510 00:44:29.756847 2305 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-etc-cni-netd\") pod \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\" (UID: \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\") " May 10 00:44:29.756926 kubelet[2305]: I0510 00:44:29.756866 2305 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrvv4\" (UniqueName: \"kubernetes.io/projected/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-kube-api-access-mrvv4\") pod \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\" (UID: \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\") " May 10 00:44:29.756926 kubelet[2305]: I0510 00:44:29.756882 2305 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-cilium-cgroup\") pod \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\" (UID: \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\") " May 10 00:44:29.756983 kubelet[2305]: I0510 00:44:29.756927 2305 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-cni-path\") pod \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\" (UID: \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\") " May 10 00:44:29.756983 kubelet[2305]: I0510 00:44:29.756940 2305 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-xtables-lock\") pod \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\" (UID: \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\") " May 10 00:44:29.756983 kubelet[2305]: I0510 00:44:29.756952 2305 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-bpf-maps\") pod \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\" (UID: \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\") " May 10 00:44:29.756983 kubelet[2305]: I0510 00:44:29.756964 2305 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-hubble-tls\") pod \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\" (UID: \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\") " May 10 00:44:29.757087 kubelet[2305]: I0510 00:44:29.756995 2305 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-hostproc\") pod \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\" (UID: \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\") " May 10 00:44:29.757087 kubelet[2305]: I0510 00:44:29.757011 2305 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-lib-modules\") pod \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\" (UID: \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\") " May 10 00:44:29.757087 kubelet[2305]: I0510 00:44:29.757025 2305 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-host-proc-sys-net\") pod \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\" (UID: \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\") " May 10 00:44:29.757087 kubelet[2305]: I0510 00:44:29.757041 2305 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-cilium-ipsec-secrets\") pod \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\" (UID: \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\") " May 10 00:44:29.757087 kubelet[2305]: I0510 00:44:29.757073 2305 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-cilium-config-path\") pod \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\" (UID: \"79537fb0-5b06-4d58-bdc3-9a3409a62e1a\") " May 10 00:44:29.760890 kubelet[2305]: I0510 00:44:29.757645 2305 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "79537fb0-5b06-4d58-bdc3-9a3409a62e1a" (UID: "79537fb0-5b06-4d58-bdc3-9a3409a62e1a"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 10 00:44:29.760890 kubelet[2305]: I0510 00:44:29.757673 2305 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "79537fb0-5b06-4d58-bdc3-9a3409a62e1a" (UID: "79537fb0-5b06-4d58-bdc3-9a3409a62e1a"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 10 00:44:29.760890 kubelet[2305]: I0510 00:44:29.757689 2305 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "79537fb0-5b06-4d58-bdc3-9a3409a62e1a" (UID: "79537fb0-5b06-4d58-bdc3-9a3409a62e1a"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 10 00:44:29.760890 kubelet[2305]: I0510 00:44:29.757745 2305 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "79537fb0-5b06-4d58-bdc3-9a3409a62e1a" (UID: "79537fb0-5b06-4d58-bdc3-9a3409a62e1a"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 10 00:44:29.760890 kubelet[2305]: I0510 00:44:29.757744 2305 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "79537fb0-5b06-4d58-bdc3-9a3409a62e1a" (UID: "79537fb0-5b06-4d58-bdc3-9a3409a62e1a"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 10 00:44:29.761066 kubelet[2305]: I0510 00:44:29.758225 2305 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "79537fb0-5b06-4d58-bdc3-9a3409a62e1a" (UID: "79537fb0-5b06-4d58-bdc3-9a3409a62e1a"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 10 00:44:29.761066 kubelet[2305]: I0510 00:44:29.758245 2305 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-cni-path" (OuterVolumeSpecName: "cni-path") pod "79537fb0-5b06-4d58-bdc3-9a3409a62e1a" (UID: "79537fb0-5b06-4d58-bdc3-9a3409a62e1a"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 10 00:44:29.761014 systemd[1]: var-lib-kubelet-pods-79537fb0\x2d5b06\x2d4d58\x2dbdc3\x2d9a3409a62e1a-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. May 10 00:44:29.761634 kubelet[2305]: I0510 00:44:29.761618 2305 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "79537fb0-5b06-4d58-bdc3-9a3409a62e1a" (UID: "79537fb0-5b06-4d58-bdc3-9a3409a62e1a"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" May 10 00:44:29.761742 kubelet[2305]: I0510 00:44:29.761729 2305 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "79537fb0-5b06-4d58-bdc3-9a3409a62e1a" (UID: "79537fb0-5b06-4d58-bdc3-9a3409a62e1a"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 10 00:44:29.761822 kubelet[2305]: I0510 00:44:29.761810 2305 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-hostproc" (OuterVolumeSpecName: "hostproc") pod "79537fb0-5b06-4d58-bdc3-9a3409a62e1a" (UID: "79537fb0-5b06-4d58-bdc3-9a3409a62e1a"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 10 00:44:29.761969 kubelet[2305]: I0510 00:44:29.761957 2305 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "79537fb0-5b06-4d58-bdc3-9a3409a62e1a" (UID: "79537fb0-5b06-4d58-bdc3-9a3409a62e1a"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 10 00:44:29.762669 kubelet[2305]: I0510 00:44:29.762647 2305 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "79537fb0-5b06-4d58-bdc3-9a3409a62e1a" (UID: "79537fb0-5b06-4d58-bdc3-9a3409a62e1a"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" May 10 00:44:29.765476 systemd[1]: var-lib-kubelet-pods-79537fb0\x2d5b06\x2d4d58\x2dbdc3\x2d9a3409a62e1a-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. May 10 00:44:29.765941 kubelet[2305]: I0510 00:44:29.765926 2305 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "79537fb0-5b06-4d58-bdc3-9a3409a62e1a" (UID: "79537fb0-5b06-4d58-bdc3-9a3409a62e1a"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" May 10 00:44:29.767331 kubelet[2305]: I0510 00:44:29.767313 2305 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-kube-api-access-mrvv4" (OuterVolumeSpecName: "kube-api-access-mrvv4") pod "79537fb0-5b06-4d58-bdc3-9a3409a62e1a" (UID: "79537fb0-5b06-4d58-bdc3-9a3409a62e1a"). InnerVolumeSpecName "kube-api-access-mrvv4". PluginName "kubernetes.io/projected", VolumeGidValue "" May 10 00:44:29.767944 kubelet[2305]: I0510 00:44:29.767930 2305 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-cilium-ipsec-secrets" (OuterVolumeSpecName: "cilium-ipsec-secrets") pod "79537fb0-5b06-4d58-bdc3-9a3409a62e1a" (UID: "79537fb0-5b06-4d58-bdc3-9a3409a62e1a"). InnerVolumeSpecName "cilium-ipsec-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" May 10 00:44:29.859258 kubelet[2305]: I0510 00:44:29.858043 2305 reconciler_common.go:289] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-xtables-lock\") on node \"localhost\" DevicePath \"\"" May 10 00:44:29.859258 kubelet[2305]: I0510 00:44:29.858076 2305 reconciler_common.go:289] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-hubble-tls\") on node \"localhost\" DevicePath \"\"" May 10 00:44:29.859258 kubelet[2305]: I0510 00:44:29.858083 2305 reconciler_common.go:289] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-bpf-maps\") on node \"localhost\" DevicePath \"\"" May 10 00:44:29.859258 kubelet[2305]: I0510 00:44:29.858089 2305 reconciler_common.go:289] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-hostproc\") on node \"localhost\" DevicePath \"\"" May 10 00:44:29.859258 kubelet[2305]: I0510 00:44:29.858094 2305 reconciler_common.go:289] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-lib-modules\") on node \"localhost\" DevicePath \"\"" May 10 00:44:29.859258 kubelet[2305]: I0510 00:44:29.858101 2305 reconciler_common.go:289] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-host-proc-sys-net\") on node \"localhost\" DevicePath \"\"" May 10 00:44:29.859258 kubelet[2305]: I0510 00:44:29.858110 2305 reconciler_common.go:289] "Volume detached for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-cilium-ipsec-secrets\") on node \"localhost\" DevicePath \"\"" May 10 00:44:29.859258 kubelet[2305]: I0510 00:44:29.858118 2305 reconciler_common.go:289] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-cilium-config-path\") on node \"localhost\" DevicePath \"\"" May 10 00:44:29.859631 kubelet[2305]: I0510 00:44:29.858126 2305 reconciler_common.go:289] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-clustermesh-secrets\") on node \"localhost\" DevicePath \"\"" May 10 00:44:29.859631 kubelet[2305]: I0510 00:44:29.858133 2305 reconciler_common.go:289] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-cilium-run\") on node \"localhost\" DevicePath \"\"" May 10 00:44:29.859631 kubelet[2305]: I0510 00:44:29.858142 2305 reconciler_common.go:289] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-host-proc-sys-kernel\") on node \"localhost\" DevicePath \"\"" May 10 00:44:29.859631 kubelet[2305]: I0510 00:44:29.858148 2305 reconciler_common.go:289] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-etc-cni-netd\") on node \"localhost\" DevicePath \"\"" May 10 00:44:29.859631 kubelet[2305]: I0510 00:44:29.858155 2305 reconciler_common.go:289] "Volume detached for volume \"kube-api-access-mrvv4\" (UniqueName: \"kubernetes.io/projected/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-kube-api-access-mrvv4\") on node \"localhost\" DevicePath \"\"" May 10 00:44:29.859631 kubelet[2305]: I0510 00:44:29.858162 2305 reconciler_common.go:289] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-cilium-cgroup\") on node \"localhost\" DevicePath \"\"" May 10 00:44:29.859631 kubelet[2305]: I0510 00:44:29.858168 2305 reconciler_common.go:289] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/79537fb0-5b06-4d58-bdc3-9a3409a62e1a-cni-path\") on node \"localhost\" DevicePath \"\"" May 10 00:44:30.264497 systemd[1]: var-lib-kubelet-pods-79537fb0\x2d5b06\x2d4d58\x2dbdc3\x2d9a3409a62e1a-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dmrvv4.mount: Deactivated successfully. May 10 00:44:30.264592 systemd[1]: var-lib-kubelet-pods-79537fb0\x2d5b06\x2d4d58\x2dbdc3\x2d9a3409a62e1a-volumes-kubernetes.io\x7esecret-cilium\x2dipsec\x2dsecrets.mount: Deactivated successfully. May 10 00:44:30.566151 kubelet[2305]: I0510 00:44:30.566014 2305 scope.go:117] "RemoveContainer" containerID="b0b979ef7ab0e55652b063140781bfa7a1e56c9c0278d871751af1d23089f159" May 10 00:44:30.569169 env[1375]: time="2025-05-10T00:44:30.569143343Z" level=info msg="RemoveContainer for \"b0b979ef7ab0e55652b063140781bfa7a1e56c9c0278d871751af1d23089f159\"" May 10 00:44:30.585871 env[1375]: time="2025-05-10T00:44:30.585781770Z" level=info msg="RemoveContainer for \"b0b979ef7ab0e55652b063140781bfa7a1e56c9c0278d871751af1d23089f159\" returns successfully" May 10 00:44:30.627976 kubelet[2305]: I0510 00:44:30.627954 2305 topology_manager.go:215] "Topology Admit Handler" podUID="46caebdb-21c8-4485-b6cb-281fa0c1a36f" podNamespace="kube-system" podName="cilium-l5lpj" May 10 00:44:30.628131 kubelet[2305]: E0510 00:44:30.628122 2305 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="79537fb0-5b06-4d58-bdc3-9a3409a62e1a" containerName="mount-cgroup" May 10 00:44:30.628197 kubelet[2305]: I0510 00:44:30.628190 2305 memory_manager.go:354] "RemoveStaleState removing state" podUID="79537fb0-5b06-4d58-bdc3-9a3409a62e1a" containerName="mount-cgroup" May 10 00:44:30.763412 kubelet[2305]: I0510 00:44:30.763378 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/46caebdb-21c8-4485-b6cb-281fa0c1a36f-cilium-config-path\") pod \"cilium-l5lpj\" (UID: \"46caebdb-21c8-4485-b6cb-281fa0c1a36f\") " pod="kube-system/cilium-l5lpj" May 10 00:44:30.763557 kubelet[2305]: I0510 00:44:30.763413 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/46caebdb-21c8-4485-b6cb-281fa0c1a36f-host-proc-sys-net\") pod \"cilium-l5lpj\" (UID: \"46caebdb-21c8-4485-b6cb-281fa0c1a36f\") " pod="kube-system/cilium-l5lpj" May 10 00:44:30.763557 kubelet[2305]: I0510 00:44:30.763441 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/46caebdb-21c8-4485-b6cb-281fa0c1a36f-bpf-maps\") pod \"cilium-l5lpj\" (UID: \"46caebdb-21c8-4485-b6cb-281fa0c1a36f\") " pod="kube-system/cilium-l5lpj" May 10 00:44:30.763557 kubelet[2305]: I0510 00:44:30.763452 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/46caebdb-21c8-4485-b6cb-281fa0c1a36f-xtables-lock\") pod \"cilium-l5lpj\" (UID: \"46caebdb-21c8-4485-b6cb-281fa0c1a36f\") " pod="kube-system/cilium-l5lpj" May 10 00:44:30.763557 kubelet[2305]: I0510 00:44:30.763466 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/46caebdb-21c8-4485-b6cb-281fa0c1a36f-cilium-cgroup\") pod \"cilium-l5lpj\" (UID: \"46caebdb-21c8-4485-b6cb-281fa0c1a36f\") " pod="kube-system/cilium-l5lpj" May 10 00:44:30.763557 kubelet[2305]: I0510 00:44:30.763477 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/46caebdb-21c8-4485-b6cb-281fa0c1a36f-etc-cni-netd\") pod \"cilium-l5lpj\" (UID: \"46caebdb-21c8-4485-b6cb-281fa0c1a36f\") " pod="kube-system/cilium-l5lpj" May 10 00:44:30.763557 kubelet[2305]: I0510 00:44:30.763488 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/46caebdb-21c8-4485-b6cb-281fa0c1a36f-host-proc-sys-kernel\") pod \"cilium-l5lpj\" (UID: \"46caebdb-21c8-4485-b6cb-281fa0c1a36f\") " pod="kube-system/cilium-l5lpj" May 10 00:44:30.763787 kubelet[2305]: I0510 00:44:30.763516 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpk48\" (UniqueName: \"kubernetes.io/projected/46caebdb-21c8-4485-b6cb-281fa0c1a36f-kube-api-access-cpk48\") pod \"cilium-l5lpj\" (UID: \"46caebdb-21c8-4485-b6cb-281fa0c1a36f\") " pod="kube-system/cilium-l5lpj" May 10 00:44:30.763787 kubelet[2305]: I0510 00:44:30.763534 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/46caebdb-21c8-4485-b6cb-281fa0c1a36f-hostproc\") pod \"cilium-l5lpj\" (UID: \"46caebdb-21c8-4485-b6cb-281fa0c1a36f\") " pod="kube-system/cilium-l5lpj" May 10 00:44:30.763787 kubelet[2305]: I0510 00:44:30.763545 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/46caebdb-21c8-4485-b6cb-281fa0c1a36f-cni-path\") pod \"cilium-l5lpj\" (UID: \"46caebdb-21c8-4485-b6cb-281fa0c1a36f\") " pod="kube-system/cilium-l5lpj" May 10 00:44:30.763787 kubelet[2305]: I0510 00:44:30.763556 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/46caebdb-21c8-4485-b6cb-281fa0c1a36f-hubble-tls\") pod \"cilium-l5lpj\" (UID: \"46caebdb-21c8-4485-b6cb-281fa0c1a36f\") " pod="kube-system/cilium-l5lpj" May 10 00:44:30.763787 kubelet[2305]: I0510 00:44:30.763570 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/46caebdb-21c8-4485-b6cb-281fa0c1a36f-lib-modules\") pod \"cilium-l5lpj\" (UID: \"46caebdb-21c8-4485-b6cb-281fa0c1a36f\") " pod="kube-system/cilium-l5lpj" May 10 00:44:30.763787 kubelet[2305]: I0510 00:44:30.763593 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/46caebdb-21c8-4485-b6cb-281fa0c1a36f-cilium-ipsec-secrets\") pod \"cilium-l5lpj\" (UID: \"46caebdb-21c8-4485-b6cb-281fa0c1a36f\") " pod="kube-system/cilium-l5lpj" May 10 00:44:30.763942 kubelet[2305]: I0510 00:44:30.763606 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/46caebdb-21c8-4485-b6cb-281fa0c1a36f-cilium-run\") pod \"cilium-l5lpj\" (UID: \"46caebdb-21c8-4485-b6cb-281fa0c1a36f\") " pod="kube-system/cilium-l5lpj" May 10 00:44:30.763942 kubelet[2305]: I0510 00:44:30.763617 2305 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/46caebdb-21c8-4485-b6cb-281fa0c1a36f-clustermesh-secrets\") pod \"cilium-l5lpj\" (UID: \"46caebdb-21c8-4485-b6cb-281fa0c1a36f\") " pod="kube-system/cilium-l5lpj" May 10 00:44:30.932806 env[1375]: time="2025-05-10T00:44:30.932729250Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-l5lpj,Uid:46caebdb-21c8-4485-b6cb-281fa0c1a36f,Namespace:kube-system,Attempt:0,}" May 10 00:44:31.016737 env[1375]: time="2025-05-10T00:44:31.016663253Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 10 00:44:31.016737 env[1375]: time="2025-05-10T00:44:31.016712272Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 10 00:44:31.016902 env[1375]: time="2025-05-10T00:44:31.016724778Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 10 00:44:31.017098 env[1375]: time="2025-05-10T00:44:31.017066158Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/9113a9a1385270243a3169d01b2eda72f8d3545716769e65b3f42c0592400802 pid=4217 runtime=io.containerd.runc.v2 May 10 00:44:31.044001 env[1375]: time="2025-05-10T00:44:31.043956782Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-l5lpj,Uid:46caebdb-21c8-4485-b6cb-281fa0c1a36f,Namespace:kube-system,Attempt:0,} returns sandbox id \"9113a9a1385270243a3169d01b2eda72f8d3545716769e65b3f42c0592400802\"" May 10 00:44:31.046482 env[1375]: time="2025-05-10T00:44:31.046454590Z" level=info msg="CreateContainer within sandbox \"9113a9a1385270243a3169d01b2eda72f8d3545716769e65b3f42c0592400802\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" May 10 00:44:31.133216 env[1375]: time="2025-05-10T00:44:31.133169360Z" level=info msg="CreateContainer within sandbox \"9113a9a1385270243a3169d01b2eda72f8d3545716769e65b3f42c0592400802\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"f4fa998a39f319ecdaede5f38352fd95d38f5ba9d3f84de53dca872369a7074d\"" May 10 00:44:31.134872 env[1375]: time="2025-05-10T00:44:31.134025857Z" level=info msg="StartContainer for \"f4fa998a39f319ecdaede5f38352fd95d38f5ba9d3f84de53dca872369a7074d\"" May 10 00:44:31.183988 env[1375]: time="2025-05-10T00:44:31.183909907Z" level=info msg="StartContainer for \"f4fa998a39f319ecdaede5f38352fd95d38f5ba9d3f84de53dca872369a7074d\" returns successfully" May 10 00:44:31.226812 env[1375]: time="2025-05-10T00:44:31.226763071Z" level=info msg="shim disconnected" id=f4fa998a39f319ecdaede5f38352fd95d38f5ba9d3f84de53dca872369a7074d May 10 00:44:31.226812 env[1375]: time="2025-05-10T00:44:31.226810110Z" level=warning msg="cleaning up after shim disconnected" id=f4fa998a39f319ecdaede5f38352fd95d38f5ba9d3f84de53dca872369a7074d namespace=k8s.io May 10 00:44:31.227011 env[1375]: time="2025-05-10T00:44:31.226821455Z" level=info msg="cleaning up dead shim" May 10 00:44:31.233013 env[1375]: time="2025-05-10T00:44:31.232973932Z" level=warning msg="cleanup warnings time=\"2025-05-10T00:44:31Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4301 runtime=io.containerd.runc.v2\n" May 10 00:44:31.261789 kubelet[2305]: I0510 00:44:31.261764 2305 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79537fb0-5b06-4d58-bdc3-9a3409a62e1a" path="/var/lib/kubelet/pods/79537fb0-5b06-4d58-bdc3-9a3409a62e1a/volumes" May 10 00:44:31.571808 env[1375]: time="2025-05-10T00:44:31.571777230Z" level=info msg="CreateContainer within sandbox \"9113a9a1385270243a3169d01b2eda72f8d3545716769e65b3f42c0592400802\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" May 10 00:44:31.581241 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount89817789.mount: Deactivated successfully. May 10 00:44:31.589321 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2821434914.mount: Deactivated successfully. May 10 00:44:31.592491 env[1375]: time="2025-05-10T00:44:31.592464105Z" level=info msg="CreateContainer within sandbox \"9113a9a1385270243a3169d01b2eda72f8d3545716769e65b3f42c0592400802\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"68f4071a22bedd2a1bd842cfbe878c3cc8df6e442c84a250ba6685c685fb4167\"" May 10 00:44:31.593897 env[1375]: time="2025-05-10T00:44:31.592919880Z" level=info msg="StartContainer for \"68f4071a22bedd2a1bd842cfbe878c3cc8df6e442c84a250ba6685c685fb4167\"" May 10 00:44:31.637767 env[1375]: time="2025-05-10T00:44:31.637737290Z" level=info msg="StartContainer for \"68f4071a22bedd2a1bd842cfbe878c3cc8df6e442c84a250ba6685c685fb4167\" returns successfully" May 10 00:44:31.808653 env[1375]: time="2025-05-10T00:44:31.808616621Z" level=info msg="shim disconnected" id=68f4071a22bedd2a1bd842cfbe878c3cc8df6e442c84a250ba6685c685fb4167 May 10 00:44:31.808861 env[1375]: time="2025-05-10T00:44:31.808845827Z" level=warning msg="cleaning up after shim disconnected" id=68f4071a22bedd2a1bd842cfbe878c3cc8df6e442c84a250ba6685c685fb4167 namespace=k8s.io May 10 00:44:31.808929 env[1375]: time="2025-05-10T00:44:31.808915617Z" level=info msg="cleaning up dead shim" May 10 00:44:31.815044 env[1375]: time="2025-05-10T00:44:31.815023526Z" level=warning msg="cleanup warnings time=\"2025-05-10T00:44:31Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4363 runtime=io.containerd.runc.v2\n" May 10 00:44:32.334018 kubelet[2305]: E0510 00:44:32.333985 2305 kubelet.go:2900] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" May 10 00:44:32.572653 env[1375]: time="2025-05-10T00:44:32.572626843Z" level=info msg="CreateContainer within sandbox \"9113a9a1385270243a3169d01b2eda72f8d3545716769e65b3f42c0592400802\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" May 10 00:44:32.613562 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1596840216.mount: Deactivated successfully. May 10 00:44:32.619284 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2137417756.mount: Deactivated successfully. May 10 00:44:32.657439 env[1375]: time="2025-05-10T00:44:32.657399363Z" level=info msg="CreateContainer within sandbox \"9113a9a1385270243a3169d01b2eda72f8d3545716769e65b3f42c0592400802\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"875f45d34fc9b5f2b9d5923c5b49b80e6f439aaec9d8d4ce0e4f3925aea9f4c8\"" May 10 00:44:32.658091 env[1375]: time="2025-05-10T00:44:32.658073309Z" level=info msg="StartContainer for \"875f45d34fc9b5f2b9d5923c5b49b80e6f439aaec9d8d4ce0e4f3925aea9f4c8\"" May 10 00:44:32.702686 env[1375]: time="2025-05-10T00:44:32.702662819Z" level=info msg="StartContainer for \"875f45d34fc9b5f2b9d5923c5b49b80e6f439aaec9d8d4ce0e4f3925aea9f4c8\" returns successfully" May 10 00:44:32.758677 env[1375]: time="2025-05-10T00:44:32.758643570Z" level=info msg="shim disconnected" id=875f45d34fc9b5f2b9d5923c5b49b80e6f439aaec9d8d4ce0e4f3925aea9f4c8 May 10 00:44:32.758677 env[1375]: time="2025-05-10T00:44:32.758674897Z" level=warning msg="cleaning up after shim disconnected" id=875f45d34fc9b5f2b9d5923c5b49b80e6f439aaec9d8d4ce0e4f3925aea9f4c8 namespace=k8s.io May 10 00:44:32.758677 env[1375]: time="2025-05-10T00:44:32.758681036Z" level=info msg="cleaning up dead shim" May 10 00:44:32.764191 env[1375]: time="2025-05-10T00:44:32.764162536Z" level=warning msg="cleanup warnings time=\"2025-05-10T00:44:32Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4421 runtime=io.containerd.runc.v2\n" May 10 00:44:33.576084 env[1375]: time="2025-05-10T00:44:33.576048185Z" level=info msg="CreateContainer within sandbox \"9113a9a1385270243a3169d01b2eda72f8d3545716769e65b3f42c0592400802\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" May 10 00:44:33.609303 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3513681418.mount: Deactivated successfully. May 10 00:44:33.617225 env[1375]: time="2025-05-10T00:44:33.617190274Z" level=info msg="CreateContainer within sandbox \"9113a9a1385270243a3169d01b2eda72f8d3545716769e65b3f42c0592400802\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"86b6dd4a8316741a7a5ae718d66ec63b54c829d5d4d67d2ff333927a4d2844d8\"" May 10 00:44:33.617730 env[1375]: time="2025-05-10T00:44:33.617696032Z" level=info msg="StartContainer for \"86b6dd4a8316741a7a5ae718d66ec63b54c829d5d4d67d2ff333927a4d2844d8\"" May 10 00:44:33.656907 env[1375]: time="2025-05-10T00:44:33.656869362Z" level=info msg="StartContainer for \"86b6dd4a8316741a7a5ae718d66ec63b54c829d5d4d67d2ff333927a4d2844d8\" returns successfully" May 10 00:44:33.670942 env[1375]: time="2025-05-10T00:44:33.670908610Z" level=info msg="shim disconnected" id=86b6dd4a8316741a7a5ae718d66ec63b54c829d5d4d67d2ff333927a4d2844d8 May 10 00:44:33.671076 env[1375]: time="2025-05-10T00:44:33.670978354Z" level=warning msg="cleaning up after shim disconnected" id=86b6dd4a8316741a7a5ae718d66ec63b54c829d5d4d67d2ff333927a4d2844d8 namespace=k8s.io May 10 00:44:33.671076 env[1375]: time="2025-05-10T00:44:33.670987396Z" level=info msg="cleaning up dead shim" May 10 00:44:33.675734 env[1375]: time="2025-05-10T00:44:33.675666979Z" level=warning msg="cleanup warnings time=\"2025-05-10T00:44:33Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4480 runtime=io.containerd.runc.v2\n" May 10 00:44:34.264859 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-86b6dd4a8316741a7a5ae718d66ec63b54c829d5d4d67d2ff333927a4d2844d8-rootfs.mount: Deactivated successfully. May 10 00:44:34.579979 env[1375]: time="2025-05-10T00:44:34.579539847Z" level=info msg="CreateContainer within sandbox \"9113a9a1385270243a3169d01b2eda72f8d3545716769e65b3f42c0592400802\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" May 10 00:44:34.647073 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1041791096.mount: Deactivated successfully. May 10 00:44:34.652492 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2806624268.mount: Deactivated successfully. May 10 00:44:34.693976 env[1375]: time="2025-05-10T00:44:34.693947293Z" level=info msg="CreateContainer within sandbox \"9113a9a1385270243a3169d01b2eda72f8d3545716769e65b3f42c0592400802\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"b4f8805a4fd6e66354ab55b538efd7ba07299793ab7a2fe45047d4674c2ddca8\"" May 10 00:44:34.695361 env[1375]: time="2025-05-10T00:44:34.694614833Z" level=info msg="StartContainer for \"b4f8805a4fd6e66354ab55b538efd7ba07299793ab7a2fe45047d4674c2ddca8\"" May 10 00:44:34.740044 env[1375]: time="2025-05-10T00:44:34.740010738Z" level=info msg="StartContainer for \"b4f8805a4fd6e66354ab55b538efd7ba07299793ab7a2fe45047d4674c2ddca8\" returns successfully" May 10 00:44:36.835720 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni)) May 10 00:44:37.190544 systemd[1]: run-containerd-runc-k8s.io-b4f8805a4fd6e66354ab55b538efd7ba07299793ab7a2fe45047d4674c2ddca8-runc.5XXeMj.mount: Deactivated successfully. May 10 00:44:39.546820 systemd-networkd[1110]: lxc_health: Link UP May 10 00:44:39.552661 systemd-networkd[1110]: lxc_health: Gained carrier May 10 00:44:39.552775 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc_health: link becomes ready May 10 00:44:40.952626 kubelet[2305]: I0510 00:44:40.952579 2305 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-l5lpj" podStartSLOduration=10.952565539 podStartE2EDuration="10.952565539s" podCreationTimestamp="2025-05-10 00:44:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-10 00:44:35.616531271 +0000 UTC m=+128.465146782" watchObservedRunningTime="2025-05-10 00:44:40.952565539 +0000 UTC m=+133.801181045" May 10 00:44:41.338855 systemd-networkd[1110]: lxc_health: Gained IPv6LL May 10 00:44:42.039949 systemd[1]: run-containerd-runc-k8s.io-b4f8805a4fd6e66354ab55b538efd7ba07299793ab7a2fe45047d4674c2ddca8-runc.dHOYOa.mount: Deactivated successfully. May 10 00:44:44.191827 systemd[1]: run-containerd-runc-k8s.io-b4f8805a4fd6e66354ab55b538efd7ba07299793ab7a2fe45047d4674c2ddca8-runc.AvMnxi.mount: Deactivated successfully. May 10 00:44:46.263353 systemd[1]: run-containerd-runc-k8s.io-b4f8805a4fd6e66354ab55b538efd7ba07299793ab7a2fe45047d4674c2ddca8-runc.TRiIxH.mount: Deactivated successfully. May 10 00:44:46.316805 sshd[4124]: pam_unix(sshd:session): session closed for user core May 10 00:44:46.326022 systemd[1]: sshd@24-139.178.70.106:22-139.178.89.65:47922.service: Deactivated successfully. May 10 00:44:46.326740 systemd[1]: session-26.scope: Deactivated successfully. May 10 00:44:46.326768 systemd-logind[1342]: Session 26 logged out. Waiting for processes to exit. May 10 00:44:46.327480 systemd-logind[1342]: Removed session 26.