Aug 13 00:25:21.163985 kernel: Booting Linux on physical CPU 0x0000120000 [0x413fd0c1] Aug 13 00:25:21.164008 kernel: Linux version 6.6.100-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT Tue Aug 12 22:21:53 -00 2025 Aug 13 00:25:21.164016 kernel: KASLR enabled Aug 13 00:25:21.164022 kernel: efi: EFI v2.7 by American Megatrends Aug 13 00:25:21.164028 kernel: efi: ACPI 2.0=0xec090000 SMBIOS 3.0=0xf0a1ff98 ESRT=0xea48e818 RNG=0xebf10018 MEMRESERVE=0xe4751f98 Aug 13 00:25:21.164033 kernel: random: crng init done Aug 13 00:25:21.164055 kernel: esrt: Reserving ESRT space from 0x00000000ea48e818 to 0x00000000ea48e878. Aug 13 00:25:21.164061 kernel: ACPI: Early table checksum verification disabled Aug 13 00:25:21.164069 kernel: ACPI: RSDP 0x00000000EC090000 000024 (v02 Ampere) Aug 13 00:25:21.164074 kernel: ACPI: XSDT 0x00000000EC080000 0000A4 (v01 Ampere Altra 00000000 AMI 01000013) Aug 13 00:25:21.164081 kernel: ACPI: FACP 0x00000000EC060000 000114 (v06 Ampere Altra 00000000 INTL 20190509) Aug 13 00:25:21.164087 kernel: ACPI: DSDT 0x00000000EC000000 019B57 (v02 Ampere Jade 00000001 INTL 20200717) Aug 13 00:25:21.164093 kernel: ACPI: DBG2 0x00000000EC070000 00005C (v00 Ampere Altra 00000000 INTL 20190509) Aug 13 00:25:21.164099 kernel: ACPI: GTDT 0x00000000EC050000 000110 (v03 Ampere Altra 00000000 INTL 20190509) Aug 13 00:25:21.164107 kernel: ACPI: SSDT 0x00000000EC040000 00002D (v02 Ampere Altra 00000001 INTL 20190509) Aug 13 00:25:21.164114 kernel: ACPI: FIDT 0x00000000EBFF0000 00009C (v01 ALASKA A M I 01072009 AMI 00010013) Aug 13 00:25:21.164120 kernel: ACPI: SPCR 0x00000000EBFE0000 000050 (v02 ALASKA A M I 01072009 AMI 0005000F) Aug 13 00:25:21.164126 kernel: ACPI: BGRT 0x00000000EBFD0000 000038 (v01 ALASKA A M I 01072009 AMI 00010013) Aug 13 00:25:21.164133 kernel: ACPI: MCFG 0x00000000EBFC0000 0000AC (v01 Ampere Altra 00000001 AMP. 01000013) Aug 13 00:25:21.164139 kernel: ACPI: IORT 0x00000000EBFB0000 000610 (v00 Ampere Altra 00000000 AMP. 01000013) Aug 13 00:25:21.164145 kernel: ACPI: PPTT 0x00000000EBF90000 006E60 (v02 Ampere Altra 00000000 AMP. 01000013) Aug 13 00:25:21.164152 kernel: ACPI: SLIT 0x00000000EBF80000 00002D (v01 Ampere Altra 00000000 AMP. 01000013) Aug 13 00:25:21.164158 kernel: ACPI: SRAT 0x00000000EBF70000 0006D0 (v03 Ampere Altra 00000000 AMP. 01000013) Aug 13 00:25:21.164166 kernel: ACPI: APIC 0x00000000EBFA0000 0019F4 (v05 Ampere Altra 00000003 AMI 01000013) Aug 13 00:25:21.164173 kernel: ACPI: PCCT 0x00000000EBF50000 000576 (v02 Ampere Altra 00000003 AMP. 01000013) Aug 13 00:25:21.164179 kernel: ACPI: WSMT 0x00000000EBF40000 000028 (v01 ALASKA A M I 01072009 AMI 00010013) Aug 13 00:25:21.164185 kernel: ACPI: FPDT 0x00000000EBF30000 000044 (v01 ALASKA A M I 01072009 AMI 01000013) Aug 13 00:25:21.164192 kernel: ACPI: SPCR: console: pl011,mmio32,0x100002600000,115200 Aug 13 00:25:21.164198 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x88300000-0x883fffff] Aug 13 00:25:21.164204 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x90000000-0xffffffff] Aug 13 00:25:21.164211 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x80000000000-0x8007fffffff] Aug 13 00:25:21.164217 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x80100000000-0x83fffffffff] Aug 13 00:25:21.164223 kernel: NUMA: NODE_DATA [mem 0x83fdffcd800-0x83fdffd2fff] Aug 13 00:25:21.164229 kernel: Zone ranges: Aug 13 00:25:21.164236 kernel: DMA [mem 0x0000000088300000-0x00000000ffffffff] Aug 13 00:25:21.164243 kernel: DMA32 empty Aug 13 00:25:21.164250 kernel: Normal [mem 0x0000000100000000-0x0000083fffffffff] Aug 13 00:25:21.164256 kernel: Movable zone start for each node Aug 13 00:25:21.164262 kernel: Early memory node ranges Aug 13 00:25:21.164268 kernel: node 0: [mem 0x0000000088300000-0x00000000883fffff] Aug 13 00:25:21.164277 kernel: node 0: [mem 0x0000000090000000-0x0000000091ffffff] Aug 13 00:25:21.164284 kernel: node 0: [mem 0x0000000092000000-0x0000000093ffffff] Aug 13 00:25:21.164292 kernel: node 0: [mem 0x0000000094000000-0x00000000eba47fff] Aug 13 00:25:21.164299 kernel: node 0: [mem 0x00000000eba48000-0x00000000ebedcfff] Aug 13 00:25:21.164305 kernel: node 0: [mem 0x00000000ebedd000-0x00000000ebeddfff] Aug 13 00:25:21.164312 kernel: node 0: [mem 0x00000000ebede000-0x00000000ebedffff] Aug 13 00:25:21.164319 kernel: node 0: [mem 0x00000000ebee0000-0x00000000ec0fffff] Aug 13 00:25:21.164325 kernel: node 0: [mem 0x00000000ec100000-0x00000000ec10ffff] Aug 13 00:25:21.164332 kernel: node 0: [mem 0x00000000ec110000-0x00000000ee54ffff] Aug 13 00:25:21.164338 kernel: node 0: [mem 0x00000000ee550000-0x00000000f765ffff] Aug 13 00:25:21.164345 kernel: node 0: [mem 0x00000000f7660000-0x00000000f784ffff] Aug 13 00:25:21.164352 kernel: node 0: [mem 0x00000000f7850000-0x00000000f7fdffff] Aug 13 00:25:21.164360 kernel: node 0: [mem 0x00000000f7fe0000-0x00000000ffc8efff] Aug 13 00:25:21.164367 kernel: node 0: [mem 0x00000000ffc8f000-0x00000000ffc8ffff] Aug 13 00:25:21.164373 kernel: node 0: [mem 0x00000000ffc90000-0x00000000ffffffff] Aug 13 00:25:21.164380 kernel: node 0: [mem 0x0000080000000000-0x000008007fffffff] Aug 13 00:25:21.164387 kernel: node 0: [mem 0x0000080100000000-0x0000083fffffffff] Aug 13 00:25:21.164393 kernel: Initmem setup node 0 [mem 0x0000000088300000-0x0000083fffffffff] Aug 13 00:25:21.164400 kernel: On node 0, zone DMA: 768 pages in unavailable ranges Aug 13 00:25:21.164407 kernel: On node 0, zone DMA: 31744 pages in unavailable ranges Aug 13 00:25:21.164413 kernel: psci: probing for conduit method from ACPI. Aug 13 00:25:21.164420 kernel: psci: PSCIv1.1 detected in firmware. Aug 13 00:25:21.164426 kernel: psci: Using standard PSCI v0.2 function IDs Aug 13 00:25:21.164434 kernel: psci: MIGRATE_INFO_TYPE not supported. Aug 13 00:25:21.164441 kernel: psci: SMC Calling Convention v1.2 Aug 13 00:25:21.164448 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x0 -> Node 0 Aug 13 00:25:21.164455 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x100 -> Node 0 Aug 13 00:25:21.164461 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x10000 -> Node 0 Aug 13 00:25:21.164468 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x10100 -> Node 0 Aug 13 00:25:21.164475 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x20000 -> Node 0 Aug 13 00:25:21.164481 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x20100 -> Node 0 Aug 13 00:25:21.164488 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x30000 -> Node 0 Aug 13 00:25:21.164494 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x30100 -> Node 0 Aug 13 00:25:21.164501 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x40000 -> Node 0 Aug 13 00:25:21.164507 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x40100 -> Node 0 Aug 13 00:25:21.164515 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x50000 -> Node 0 Aug 13 00:25:21.164522 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x50100 -> Node 0 Aug 13 00:25:21.164529 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x60000 -> Node 0 Aug 13 00:25:21.164535 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x60100 -> Node 0 Aug 13 00:25:21.164542 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x70000 -> Node 0 Aug 13 00:25:21.164548 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x70100 -> Node 0 Aug 13 00:25:21.164555 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x80000 -> Node 0 Aug 13 00:25:21.164562 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x80100 -> Node 0 Aug 13 00:25:21.164568 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x90000 -> Node 0 Aug 13 00:25:21.164575 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x90100 -> Node 0 Aug 13 00:25:21.164582 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0xa0000 -> Node 0 Aug 13 00:25:21.164588 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0xa0100 -> Node 0 Aug 13 00:25:21.164596 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0xb0000 -> Node 0 Aug 13 00:25:21.164603 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0xb0100 -> Node 0 Aug 13 00:25:21.164609 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0xc0000 -> Node 0 Aug 13 00:25:21.164616 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0xc0100 -> Node 0 Aug 13 00:25:21.164623 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0xd0000 -> Node 0 Aug 13 00:25:21.164629 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0xd0100 -> Node 0 Aug 13 00:25:21.164636 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0xe0000 -> Node 0 Aug 13 00:25:21.164642 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0xe0100 -> Node 0 Aug 13 00:25:21.164649 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0xf0000 -> Node 0 Aug 13 00:25:21.164655 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0xf0100 -> Node 0 Aug 13 00:25:21.164662 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x100000 -> Node 0 Aug 13 00:25:21.164669 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x100100 -> Node 0 Aug 13 00:25:21.164676 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x110000 -> Node 0 Aug 13 00:25:21.164683 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x110100 -> Node 0 Aug 13 00:25:21.164689 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x120000 -> Node 0 Aug 13 00:25:21.164696 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x120100 -> Node 0 Aug 13 00:25:21.164703 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x130000 -> Node 0 Aug 13 00:25:21.164709 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x130100 -> Node 0 Aug 13 00:25:21.164716 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x140000 -> Node 0 Aug 13 00:25:21.164722 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x140100 -> Node 0 Aug 13 00:25:21.164729 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x150000 -> Node 0 Aug 13 00:25:21.164735 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x150100 -> Node 0 Aug 13 00:25:21.164742 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x160000 -> Node 0 Aug 13 00:25:21.164750 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x160100 -> Node 0 Aug 13 00:25:21.164757 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x170000 -> Node 0 Aug 13 00:25:21.164763 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x170100 -> Node 0 Aug 13 00:25:21.164770 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x180000 -> Node 0 Aug 13 00:25:21.164777 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x180100 -> Node 0 Aug 13 00:25:21.164783 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x190000 -> Node 0 Aug 13 00:25:21.164790 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x190100 -> Node 0 Aug 13 00:25:21.164797 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1a0000 -> Node 0 Aug 13 00:25:21.164809 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1a0100 -> Node 0 Aug 13 00:25:21.164816 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1b0000 -> Node 0 Aug 13 00:25:21.164825 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1b0100 -> Node 0 Aug 13 00:25:21.164832 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1c0000 -> Node 0 Aug 13 00:25:21.164839 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1c0100 -> Node 0 Aug 13 00:25:21.164846 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1d0000 -> Node 0 Aug 13 00:25:21.164853 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1d0100 -> Node 0 Aug 13 00:25:21.164860 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1e0000 -> Node 0 Aug 13 00:25:21.164869 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1e0100 -> Node 0 Aug 13 00:25:21.164876 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1f0000 -> Node 0 Aug 13 00:25:21.164883 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1f0100 -> Node 0 Aug 13 00:25:21.164889 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x200000 -> Node 0 Aug 13 00:25:21.164897 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x200100 -> Node 0 Aug 13 00:25:21.164904 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x210000 -> Node 0 Aug 13 00:25:21.164911 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x210100 -> Node 0 Aug 13 00:25:21.164918 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x220000 -> Node 0 Aug 13 00:25:21.164925 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x220100 -> Node 0 Aug 13 00:25:21.164932 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x230000 -> Node 0 Aug 13 00:25:21.164939 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x230100 -> Node 0 Aug 13 00:25:21.164947 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x240000 -> Node 0 Aug 13 00:25:21.164955 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x240100 -> Node 0 Aug 13 00:25:21.164962 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x250000 -> Node 0 Aug 13 00:25:21.164969 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x250100 -> Node 0 Aug 13 00:25:21.164976 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x260000 -> Node 0 Aug 13 00:25:21.164983 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x260100 -> Node 0 Aug 13 00:25:21.164990 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x270000 -> Node 0 Aug 13 00:25:21.164997 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x270100 -> Node 0 Aug 13 00:25:21.165004 kernel: percpu: Embedded 31 pages/cpu s86696 r8192 d32088 u126976 Aug 13 00:25:21.165011 kernel: pcpu-alloc: s86696 r8192 d32088 u126976 alloc=31*4096 Aug 13 00:25:21.165019 kernel: pcpu-alloc: [0] 00 [0] 01 [0] 02 [0] 03 [0] 04 [0] 05 [0] 06 [0] 07 Aug 13 00:25:21.165026 kernel: pcpu-alloc: [0] 08 [0] 09 [0] 10 [0] 11 [0] 12 [0] 13 [0] 14 [0] 15 Aug 13 00:25:21.165034 kernel: pcpu-alloc: [0] 16 [0] 17 [0] 18 [0] 19 [0] 20 [0] 21 [0] 22 [0] 23 Aug 13 00:25:21.165073 kernel: pcpu-alloc: [0] 24 [0] 25 [0] 26 [0] 27 [0] 28 [0] 29 [0] 30 [0] 31 Aug 13 00:25:21.165080 kernel: pcpu-alloc: [0] 32 [0] 33 [0] 34 [0] 35 [0] 36 [0] 37 [0] 38 [0] 39 Aug 13 00:25:21.165088 kernel: pcpu-alloc: [0] 40 [0] 41 [0] 42 [0] 43 [0] 44 [0] 45 [0] 46 [0] 47 Aug 13 00:25:21.165095 kernel: pcpu-alloc: [0] 48 [0] 49 [0] 50 [0] 51 [0] 52 [0] 53 [0] 54 [0] 55 Aug 13 00:25:21.165102 kernel: pcpu-alloc: [0] 56 [0] 57 [0] 58 [0] 59 [0] 60 [0] 61 [0] 62 [0] 63 Aug 13 00:25:21.165109 kernel: pcpu-alloc: [0] 64 [0] 65 [0] 66 [0] 67 [0] 68 [0] 69 [0] 70 [0] 71 Aug 13 00:25:21.165116 kernel: pcpu-alloc: [0] 72 [0] 73 [0] 74 [0] 75 [0] 76 [0] 77 [0] 78 [0] 79 Aug 13 00:25:21.165123 kernel: Detected PIPT I-cache on CPU0 Aug 13 00:25:21.165130 kernel: CPU features: detected: GIC system register CPU interface Aug 13 00:25:21.165137 kernel: CPU features: detected: Virtualization Host Extensions Aug 13 00:25:21.165146 kernel: CPU features: detected: Hardware dirty bit management Aug 13 00:25:21.165153 kernel: CPU features: detected: Spectre-v4 Aug 13 00:25:21.165160 kernel: CPU features: detected: Spectre-BHB Aug 13 00:25:21.165168 kernel: CPU features: kernel page table isolation forced ON by KASLR Aug 13 00:25:21.165175 kernel: CPU features: detected: Kernel page table isolation (KPTI) Aug 13 00:25:21.165182 kernel: CPU features: detected: ARM erratum 1418040 Aug 13 00:25:21.165189 kernel: CPU features: detected: SSBS not fully self-synchronizing Aug 13 00:25:21.165196 kernel: alternatives: applying boot alternatives Aug 13 00:25:21.165205 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty0 console=ttyS1,115200n8 flatcar.first_boot=detected acpi=force flatcar.oem.id=packet flatcar.autologin verity.usrhash=2f9df6e9e6c671c457040a64675390bbff42294b08c628cd2dc472ed8120146a Aug 13 00:25:21.165212 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Aug 13 00:25:21.165221 kernel: printk: log_buf_len individual max cpu contribution: 4096 bytes Aug 13 00:25:21.165228 kernel: printk: log_buf_len total cpu_extra contributions: 323584 bytes Aug 13 00:25:21.165235 kernel: printk: log_buf_len min size: 262144 bytes Aug 13 00:25:21.165242 kernel: printk: log_buf_len: 1048576 bytes Aug 13 00:25:21.165249 kernel: printk: early log buf free: 250024(95%) Aug 13 00:25:21.165256 kernel: Dentry cache hash table entries: 16777216 (order: 15, 134217728 bytes, linear) Aug 13 00:25:21.165263 kernel: Inode-cache hash table entries: 8388608 (order: 14, 67108864 bytes, linear) Aug 13 00:25:21.165270 kernel: Fallback order for Node 0: 0 Aug 13 00:25:21.165277 kernel: Built 1 zonelists, mobility grouping on. Total pages: 65996028 Aug 13 00:25:21.165284 kernel: Policy zone: Normal Aug 13 00:25:21.165291 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Aug 13 00:25:21.165298 kernel: software IO TLB: area num 128. Aug 13 00:25:21.165307 kernel: software IO TLB: mapped [mem 0x00000000fbc8f000-0x00000000ffc8f000] (64MB) Aug 13 00:25:21.165314 kernel: Memory: 262922400K/268174336K available (10304K kernel code, 2186K rwdata, 8108K rodata, 39424K init, 897K bss, 5251936K reserved, 0K cma-reserved) Aug 13 00:25:21.165321 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=80, Nodes=1 Aug 13 00:25:21.165328 kernel: rcu: Preemptible hierarchical RCU implementation. Aug 13 00:25:21.165336 kernel: rcu: RCU event tracing is enabled. Aug 13 00:25:21.165343 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=80. Aug 13 00:25:21.165350 kernel: Trampoline variant of Tasks RCU enabled. Aug 13 00:25:21.165358 kernel: Tracing variant of Tasks RCU enabled. Aug 13 00:25:21.165365 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Aug 13 00:25:21.165372 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=80 Aug 13 00:25:21.165379 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Aug 13 00:25:21.165388 kernel: GICv3: GIC: Using split EOI/Deactivate mode Aug 13 00:25:21.165395 kernel: GICv3: 672 SPIs implemented Aug 13 00:25:21.165402 kernel: GICv3: 0 Extended SPIs implemented Aug 13 00:25:21.165409 kernel: Root IRQ handler: gic_handle_irq Aug 13 00:25:21.165416 kernel: GICv3: GICv3 features: 16 PPIs Aug 13 00:25:21.165423 kernel: GICv3: CPU0: found redistributor 120000 region 0:0x00001001005c0000 Aug 13 00:25:21.165430 kernel: SRAT: PXM 0 -> ITS 0 -> Node 0 Aug 13 00:25:21.165437 kernel: SRAT: PXM 0 -> ITS 1 -> Node 0 Aug 13 00:25:21.165444 kernel: SRAT: PXM 0 -> ITS 2 -> Node 0 Aug 13 00:25:21.165451 kernel: SRAT: PXM 0 -> ITS 3 -> Node 0 Aug 13 00:25:21.165458 kernel: SRAT: PXM 0 -> ITS 4 -> Node 0 Aug 13 00:25:21.165465 kernel: SRAT: PXM 0 -> ITS 5 -> Node 0 Aug 13 00:25:21.165472 kernel: SRAT: PXM 0 -> ITS 6 -> Node 0 Aug 13 00:25:21.165481 kernel: SRAT: PXM 0 -> ITS 7 -> Node 0 Aug 13 00:25:21.165488 kernel: ITS [mem 0x100100040000-0x10010005ffff] Aug 13 00:25:21.165495 kernel: ITS@0x0000100100040000: allocated 8192 Devices @80000270000 (indirect, esz 8, psz 64K, shr 1) Aug 13 00:25:21.165502 kernel: ITS@0x0000100100040000: allocated 32768 Interrupt Collections @80000280000 (flat, esz 2, psz 64K, shr 1) Aug 13 00:25:21.165510 kernel: ITS [mem 0x100100060000-0x10010007ffff] Aug 13 00:25:21.165517 kernel: ITS@0x0000100100060000: allocated 8192 Devices @800002a0000 (indirect, esz 8, psz 64K, shr 1) Aug 13 00:25:21.165524 kernel: ITS@0x0000100100060000: allocated 32768 Interrupt Collections @800002b0000 (flat, esz 2, psz 64K, shr 1) Aug 13 00:25:21.165532 kernel: ITS [mem 0x100100080000-0x10010009ffff] Aug 13 00:25:21.165539 kernel: ITS@0x0000100100080000: allocated 8192 Devices @800002d0000 (indirect, esz 8, psz 64K, shr 1) Aug 13 00:25:21.165546 kernel: ITS@0x0000100100080000: allocated 32768 Interrupt Collections @800002e0000 (flat, esz 2, psz 64K, shr 1) Aug 13 00:25:21.165553 kernel: ITS [mem 0x1001000a0000-0x1001000bffff] Aug 13 00:25:21.165562 kernel: ITS@0x00001001000a0000: allocated 8192 Devices @80000300000 (indirect, esz 8, psz 64K, shr 1) Aug 13 00:25:21.165569 kernel: ITS@0x00001001000a0000: allocated 32768 Interrupt Collections @80000310000 (flat, esz 2, psz 64K, shr 1) Aug 13 00:25:21.165576 kernel: ITS [mem 0x1001000c0000-0x1001000dffff] Aug 13 00:25:21.165583 kernel: ITS@0x00001001000c0000: allocated 8192 Devices @80000330000 (indirect, esz 8, psz 64K, shr 1) Aug 13 00:25:21.165590 kernel: ITS@0x00001001000c0000: allocated 32768 Interrupt Collections @80000340000 (flat, esz 2, psz 64K, shr 1) Aug 13 00:25:21.165597 kernel: ITS [mem 0x1001000e0000-0x1001000fffff] Aug 13 00:25:21.165604 kernel: ITS@0x00001001000e0000: allocated 8192 Devices @80000360000 (indirect, esz 8, psz 64K, shr 1) Aug 13 00:25:21.165612 kernel: ITS@0x00001001000e0000: allocated 32768 Interrupt Collections @80000370000 (flat, esz 2, psz 64K, shr 1) Aug 13 00:25:21.165619 kernel: ITS [mem 0x100100100000-0x10010011ffff] Aug 13 00:25:21.165626 kernel: ITS@0x0000100100100000: allocated 8192 Devices @80000390000 (indirect, esz 8, psz 64K, shr 1) Aug 13 00:25:21.165633 kernel: ITS@0x0000100100100000: allocated 32768 Interrupt Collections @800003a0000 (flat, esz 2, psz 64K, shr 1) Aug 13 00:25:21.165642 kernel: ITS [mem 0x100100120000-0x10010013ffff] Aug 13 00:25:21.165649 kernel: ITS@0x0000100100120000: allocated 8192 Devices @800003c0000 (indirect, esz 8, psz 64K, shr 1) Aug 13 00:25:21.165656 kernel: ITS@0x0000100100120000: allocated 32768 Interrupt Collections @800003d0000 (flat, esz 2, psz 64K, shr 1) Aug 13 00:25:21.165663 kernel: GICv3: using LPI property table @0x00000800003e0000 Aug 13 00:25:21.165670 kernel: GICv3: CPU0: using allocated LPI pending table @0x00000800003f0000 Aug 13 00:25:21.165678 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Aug 13 00:25:21.165685 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.165692 kernel: ACPI GTDT: found 1 memory-mapped timer block(s). Aug 13 00:25:21.165700 kernel: arch_timer: cp15 and mmio timer(s) running at 25.00MHz (phys/phys). Aug 13 00:25:21.165707 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns Aug 13 00:25:21.165714 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns Aug 13 00:25:21.165723 kernel: Console: colour dummy device 80x25 Aug 13 00:25:21.165730 kernel: printk: console [tty0] enabled Aug 13 00:25:21.165737 kernel: ACPI: Core revision 20230628 Aug 13 00:25:21.165745 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000) Aug 13 00:25:21.165752 kernel: pid_max: default: 81920 minimum: 640 Aug 13 00:25:21.165759 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Aug 13 00:25:21.165767 kernel: landlock: Up and running. Aug 13 00:25:21.165774 kernel: SELinux: Initializing. Aug 13 00:25:21.165781 kernel: Mount-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Aug 13 00:25:21.165790 kernel: Mountpoint-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Aug 13 00:25:21.165797 kernel: RCU Tasks: Setting shift to 7 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=80. Aug 13 00:25:21.165805 kernel: RCU Tasks Trace: Setting shift to 7 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=80. Aug 13 00:25:21.165812 kernel: rcu: Hierarchical SRCU implementation. Aug 13 00:25:21.165820 kernel: rcu: Max phase no-delay instances is 400. Aug 13 00:25:21.165827 kernel: Platform MSI: ITS@0x100100040000 domain created Aug 13 00:25:21.165834 kernel: Platform MSI: ITS@0x100100060000 domain created Aug 13 00:25:21.165841 kernel: Platform MSI: ITS@0x100100080000 domain created Aug 13 00:25:21.165849 kernel: Platform MSI: ITS@0x1001000a0000 domain created Aug 13 00:25:21.165857 kernel: Platform MSI: ITS@0x1001000c0000 domain created Aug 13 00:25:21.165864 kernel: Platform MSI: ITS@0x1001000e0000 domain created Aug 13 00:25:21.165872 kernel: Platform MSI: ITS@0x100100100000 domain created Aug 13 00:25:21.165879 kernel: Platform MSI: ITS@0x100100120000 domain created Aug 13 00:25:21.165886 kernel: PCI/MSI: ITS@0x100100040000 domain created Aug 13 00:25:21.165893 kernel: PCI/MSI: ITS@0x100100060000 domain created Aug 13 00:25:21.165901 kernel: PCI/MSI: ITS@0x100100080000 domain created Aug 13 00:25:21.165908 kernel: PCI/MSI: ITS@0x1001000a0000 domain created Aug 13 00:25:21.165915 kernel: PCI/MSI: ITS@0x1001000c0000 domain created Aug 13 00:25:21.165922 kernel: PCI/MSI: ITS@0x1001000e0000 domain created Aug 13 00:25:21.165931 kernel: PCI/MSI: ITS@0x100100100000 domain created Aug 13 00:25:21.165938 kernel: PCI/MSI: ITS@0x100100120000 domain created Aug 13 00:25:21.165946 kernel: Remapping and enabling EFI services. Aug 13 00:25:21.165953 kernel: smp: Bringing up secondary CPUs ... Aug 13 00:25:21.165960 kernel: Detected PIPT I-cache on CPU1 Aug 13 00:25:21.165967 kernel: GICv3: CPU1: found redistributor 1a0000 region 0:0x00001001007c0000 Aug 13 00:25:21.165975 kernel: GICv3: CPU1: using allocated LPI pending table @0x0000080000800000 Aug 13 00:25:21.165982 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.165989 kernel: CPU1: Booted secondary processor 0x00001a0000 [0x413fd0c1] Aug 13 00:25:21.165998 kernel: Detected PIPT I-cache on CPU2 Aug 13 00:25:21.166005 kernel: GICv3: CPU2: found redistributor 140000 region 0:0x0000100100640000 Aug 13 00:25:21.166012 kernel: GICv3: CPU2: using allocated LPI pending table @0x0000080000810000 Aug 13 00:25:21.166019 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166026 kernel: CPU2: Booted secondary processor 0x0000140000 [0x413fd0c1] Aug 13 00:25:21.166034 kernel: Detected PIPT I-cache on CPU3 Aug 13 00:25:21.166043 kernel: GICv3: CPU3: found redistributor 1c0000 region 0:0x0000100100840000 Aug 13 00:25:21.166050 kernel: GICv3: CPU3: using allocated LPI pending table @0x0000080000820000 Aug 13 00:25:21.166058 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166065 kernel: CPU3: Booted secondary processor 0x00001c0000 [0x413fd0c1] Aug 13 00:25:21.166074 kernel: Detected PIPT I-cache on CPU4 Aug 13 00:25:21.166081 kernel: GICv3: CPU4: found redistributor 100000 region 0:0x0000100100540000 Aug 13 00:25:21.166089 kernel: GICv3: CPU4: using allocated LPI pending table @0x0000080000830000 Aug 13 00:25:21.166096 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166103 kernel: CPU4: Booted secondary processor 0x0000100000 [0x413fd0c1] Aug 13 00:25:21.166110 kernel: Detected PIPT I-cache on CPU5 Aug 13 00:25:21.166117 kernel: GICv3: CPU5: found redistributor 180000 region 0:0x0000100100740000 Aug 13 00:25:21.166124 kernel: GICv3: CPU5: using allocated LPI pending table @0x0000080000840000 Aug 13 00:25:21.166132 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166140 kernel: CPU5: Booted secondary processor 0x0000180000 [0x413fd0c1] Aug 13 00:25:21.166148 kernel: Detected PIPT I-cache on CPU6 Aug 13 00:25:21.166155 kernel: GICv3: CPU6: found redistributor 160000 region 0:0x00001001006c0000 Aug 13 00:25:21.166162 kernel: GICv3: CPU6: using allocated LPI pending table @0x0000080000850000 Aug 13 00:25:21.166169 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166176 kernel: CPU6: Booted secondary processor 0x0000160000 [0x413fd0c1] Aug 13 00:25:21.166184 kernel: Detected PIPT I-cache on CPU7 Aug 13 00:25:21.166191 kernel: GICv3: CPU7: found redistributor 1e0000 region 0:0x00001001008c0000 Aug 13 00:25:21.166198 kernel: GICv3: CPU7: using allocated LPI pending table @0x0000080000860000 Aug 13 00:25:21.166207 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166215 kernel: CPU7: Booted secondary processor 0x00001e0000 [0x413fd0c1] Aug 13 00:25:21.166222 kernel: Detected PIPT I-cache on CPU8 Aug 13 00:25:21.166229 kernel: GICv3: CPU8: found redistributor a0000 region 0:0x00001001003c0000 Aug 13 00:25:21.166236 kernel: GICv3: CPU8: using allocated LPI pending table @0x0000080000870000 Aug 13 00:25:21.166244 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166251 kernel: CPU8: Booted secondary processor 0x00000a0000 [0x413fd0c1] Aug 13 00:25:21.166258 kernel: Detected PIPT I-cache on CPU9 Aug 13 00:25:21.166265 kernel: GICv3: CPU9: found redistributor 220000 region 0:0x00001001009c0000 Aug 13 00:25:21.166272 kernel: GICv3: CPU9: using allocated LPI pending table @0x0000080000880000 Aug 13 00:25:21.166281 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166288 kernel: CPU9: Booted secondary processor 0x0000220000 [0x413fd0c1] Aug 13 00:25:21.166295 kernel: Detected PIPT I-cache on CPU10 Aug 13 00:25:21.166303 kernel: GICv3: CPU10: found redistributor c0000 region 0:0x0000100100440000 Aug 13 00:25:21.166310 kernel: GICv3: CPU10: using allocated LPI pending table @0x0000080000890000 Aug 13 00:25:21.166317 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166324 kernel: CPU10: Booted secondary processor 0x00000c0000 [0x413fd0c1] Aug 13 00:25:21.166331 kernel: Detected PIPT I-cache on CPU11 Aug 13 00:25:21.166338 kernel: GICv3: CPU11: found redistributor 240000 region 0:0x0000100100a40000 Aug 13 00:25:21.166346 kernel: GICv3: CPU11: using allocated LPI pending table @0x00000800008a0000 Aug 13 00:25:21.166354 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166361 kernel: CPU11: Booted secondary processor 0x0000240000 [0x413fd0c1] Aug 13 00:25:21.166368 kernel: Detected PIPT I-cache on CPU12 Aug 13 00:25:21.166375 kernel: GICv3: CPU12: found redistributor 80000 region 0:0x0000100100340000 Aug 13 00:25:21.166383 kernel: GICv3: CPU12: using allocated LPI pending table @0x00000800008b0000 Aug 13 00:25:21.166390 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166397 kernel: CPU12: Booted secondary processor 0x0000080000 [0x413fd0c1] Aug 13 00:25:21.166404 kernel: Detected PIPT I-cache on CPU13 Aug 13 00:25:21.166411 kernel: GICv3: CPU13: found redistributor 200000 region 0:0x0000100100940000 Aug 13 00:25:21.166420 kernel: GICv3: CPU13: using allocated LPI pending table @0x00000800008c0000 Aug 13 00:25:21.166428 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166435 kernel: CPU13: Booted secondary processor 0x0000200000 [0x413fd0c1] Aug 13 00:25:21.166442 kernel: Detected PIPT I-cache on CPU14 Aug 13 00:25:21.166449 kernel: GICv3: CPU14: found redistributor e0000 region 0:0x00001001004c0000 Aug 13 00:25:21.166457 kernel: GICv3: CPU14: using allocated LPI pending table @0x00000800008d0000 Aug 13 00:25:21.166464 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166471 kernel: CPU14: Booted secondary processor 0x00000e0000 [0x413fd0c1] Aug 13 00:25:21.166478 kernel: Detected PIPT I-cache on CPU15 Aug 13 00:25:21.166487 kernel: GICv3: CPU15: found redistributor 260000 region 0:0x0000100100ac0000 Aug 13 00:25:21.166494 kernel: GICv3: CPU15: using allocated LPI pending table @0x00000800008e0000 Aug 13 00:25:21.166502 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166509 kernel: CPU15: Booted secondary processor 0x0000260000 [0x413fd0c1] Aug 13 00:25:21.166516 kernel: Detected PIPT I-cache on CPU16 Aug 13 00:25:21.166523 kernel: GICv3: CPU16: found redistributor 20000 region 0:0x00001001001c0000 Aug 13 00:25:21.166530 kernel: GICv3: CPU16: using allocated LPI pending table @0x00000800008f0000 Aug 13 00:25:21.166537 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166545 kernel: CPU16: Booted secondary processor 0x0000020000 [0x413fd0c1] Aug 13 00:25:21.166552 kernel: Detected PIPT I-cache on CPU17 Aug 13 00:25:21.166568 kernel: GICv3: CPU17: found redistributor 40000 region 0:0x0000100100240000 Aug 13 00:25:21.166577 kernel: GICv3: CPU17: using allocated LPI pending table @0x0000080000900000 Aug 13 00:25:21.166585 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166592 kernel: CPU17: Booted secondary processor 0x0000040000 [0x413fd0c1] Aug 13 00:25:21.166600 kernel: Detected PIPT I-cache on CPU18 Aug 13 00:25:21.166607 kernel: GICv3: CPU18: found redistributor 0 region 0:0x0000100100140000 Aug 13 00:25:21.166615 kernel: GICv3: CPU18: using allocated LPI pending table @0x0000080000910000 Aug 13 00:25:21.166622 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166630 kernel: CPU18: Booted secondary processor 0x0000000000 [0x413fd0c1] Aug 13 00:25:21.166639 kernel: Detected PIPT I-cache on CPU19 Aug 13 00:25:21.166646 kernel: GICv3: CPU19: found redistributor 60000 region 0:0x00001001002c0000 Aug 13 00:25:21.166654 kernel: GICv3: CPU19: using allocated LPI pending table @0x0000080000920000 Aug 13 00:25:21.166662 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166669 kernel: CPU19: Booted secondary processor 0x0000060000 [0x413fd0c1] Aug 13 00:25:21.166677 kernel: Detected PIPT I-cache on CPU20 Aug 13 00:25:21.166686 kernel: GICv3: CPU20: found redistributor 130000 region 0:0x0000100100600000 Aug 13 00:25:21.166695 kernel: GICv3: CPU20: using allocated LPI pending table @0x0000080000930000 Aug 13 00:25:21.166703 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166710 kernel: CPU20: Booted secondary processor 0x0000130000 [0x413fd0c1] Aug 13 00:25:21.166718 kernel: Detected PIPT I-cache on CPU21 Aug 13 00:25:21.166725 kernel: GICv3: CPU21: found redistributor 1b0000 region 0:0x0000100100800000 Aug 13 00:25:21.166734 kernel: GICv3: CPU21: using allocated LPI pending table @0x0000080000940000 Aug 13 00:25:21.166742 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166750 kernel: CPU21: Booted secondary processor 0x00001b0000 [0x413fd0c1] Aug 13 00:25:21.166757 kernel: Detected PIPT I-cache on CPU22 Aug 13 00:25:21.166766 kernel: GICv3: CPU22: found redistributor 150000 region 0:0x0000100100680000 Aug 13 00:25:21.166774 kernel: GICv3: CPU22: using allocated LPI pending table @0x0000080000950000 Aug 13 00:25:21.166781 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166789 kernel: CPU22: Booted secondary processor 0x0000150000 [0x413fd0c1] Aug 13 00:25:21.166797 kernel: Detected PIPT I-cache on CPU23 Aug 13 00:25:21.166804 kernel: GICv3: CPU23: found redistributor 1d0000 region 0:0x0000100100880000 Aug 13 00:25:21.166812 kernel: GICv3: CPU23: using allocated LPI pending table @0x0000080000960000 Aug 13 00:25:21.166820 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166827 kernel: CPU23: Booted secondary processor 0x00001d0000 [0x413fd0c1] Aug 13 00:25:21.166836 kernel: Detected PIPT I-cache on CPU24 Aug 13 00:25:21.166844 kernel: GICv3: CPU24: found redistributor 110000 region 0:0x0000100100580000 Aug 13 00:25:21.166851 kernel: GICv3: CPU24: using allocated LPI pending table @0x0000080000970000 Aug 13 00:25:21.166859 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166866 kernel: CPU24: Booted secondary processor 0x0000110000 [0x413fd0c1] Aug 13 00:25:21.166874 kernel: Detected PIPT I-cache on CPU25 Aug 13 00:25:21.166881 kernel: GICv3: CPU25: found redistributor 190000 region 0:0x0000100100780000 Aug 13 00:25:21.166889 kernel: GICv3: CPU25: using allocated LPI pending table @0x0000080000980000 Aug 13 00:25:21.166896 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166905 kernel: CPU25: Booted secondary processor 0x0000190000 [0x413fd0c1] Aug 13 00:25:21.166913 kernel: Detected PIPT I-cache on CPU26 Aug 13 00:25:21.166921 kernel: GICv3: CPU26: found redistributor 170000 region 0:0x0000100100700000 Aug 13 00:25:21.166928 kernel: GICv3: CPU26: using allocated LPI pending table @0x0000080000990000 Aug 13 00:25:21.166936 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166943 kernel: CPU26: Booted secondary processor 0x0000170000 [0x413fd0c1] Aug 13 00:25:21.166951 kernel: Detected PIPT I-cache on CPU27 Aug 13 00:25:21.166958 kernel: GICv3: CPU27: found redistributor 1f0000 region 0:0x0000100100900000 Aug 13 00:25:21.166966 kernel: GICv3: CPU27: using allocated LPI pending table @0x00000800009a0000 Aug 13 00:25:21.166974 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166983 kernel: CPU27: Booted secondary processor 0x00001f0000 [0x413fd0c1] Aug 13 00:25:21.166990 kernel: Detected PIPT I-cache on CPU28 Aug 13 00:25:21.166998 kernel: GICv3: CPU28: found redistributor b0000 region 0:0x0000100100400000 Aug 13 00:25:21.167006 kernel: GICv3: CPU28: using allocated LPI pending table @0x00000800009b0000 Aug 13 00:25:21.167013 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167021 kernel: CPU28: Booted secondary processor 0x00000b0000 [0x413fd0c1] Aug 13 00:25:21.167028 kernel: Detected PIPT I-cache on CPU29 Aug 13 00:25:21.167036 kernel: GICv3: CPU29: found redistributor 230000 region 0:0x0000100100a00000 Aug 13 00:25:21.167065 kernel: GICv3: CPU29: using allocated LPI pending table @0x00000800009c0000 Aug 13 00:25:21.167075 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167083 kernel: CPU29: Booted secondary processor 0x0000230000 [0x413fd0c1] Aug 13 00:25:21.167090 kernel: Detected PIPT I-cache on CPU30 Aug 13 00:25:21.167098 kernel: GICv3: CPU30: found redistributor d0000 region 0:0x0000100100480000 Aug 13 00:25:21.167105 kernel: GICv3: CPU30: using allocated LPI pending table @0x00000800009d0000 Aug 13 00:25:21.167113 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167120 kernel: CPU30: Booted secondary processor 0x00000d0000 [0x413fd0c1] Aug 13 00:25:21.167128 kernel: Detected PIPT I-cache on CPU31 Aug 13 00:25:21.167135 kernel: GICv3: CPU31: found redistributor 250000 region 0:0x0000100100a80000 Aug 13 00:25:21.167143 kernel: GICv3: CPU31: using allocated LPI pending table @0x00000800009e0000 Aug 13 00:25:21.167152 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167160 kernel: CPU31: Booted secondary processor 0x0000250000 [0x413fd0c1] Aug 13 00:25:21.167168 kernel: Detected PIPT I-cache on CPU32 Aug 13 00:25:21.167175 kernel: GICv3: CPU32: found redistributor 90000 region 0:0x0000100100380000 Aug 13 00:25:21.167183 kernel: GICv3: CPU32: using allocated LPI pending table @0x00000800009f0000 Aug 13 00:25:21.167191 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167198 kernel: CPU32: Booted secondary processor 0x0000090000 [0x413fd0c1] Aug 13 00:25:21.167206 kernel: Detected PIPT I-cache on CPU33 Aug 13 00:25:21.167213 kernel: GICv3: CPU33: found redistributor 210000 region 0:0x0000100100980000 Aug 13 00:25:21.167222 kernel: GICv3: CPU33: using allocated LPI pending table @0x0000080000a00000 Aug 13 00:25:21.167230 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167237 kernel: CPU33: Booted secondary processor 0x0000210000 [0x413fd0c1] Aug 13 00:25:21.167245 kernel: Detected PIPT I-cache on CPU34 Aug 13 00:25:21.167253 kernel: GICv3: CPU34: found redistributor f0000 region 0:0x0000100100500000 Aug 13 00:25:21.167260 kernel: GICv3: CPU34: using allocated LPI pending table @0x0000080000a10000 Aug 13 00:25:21.167268 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167277 kernel: CPU34: Booted secondary processor 0x00000f0000 [0x413fd0c1] Aug 13 00:25:21.167285 kernel: Detected PIPT I-cache on CPU35 Aug 13 00:25:21.167292 kernel: GICv3: CPU35: found redistributor 270000 region 0:0x0000100100b00000 Aug 13 00:25:21.167302 kernel: GICv3: CPU35: using allocated LPI pending table @0x0000080000a20000 Aug 13 00:25:21.167309 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167317 kernel: CPU35: Booted secondary processor 0x0000270000 [0x413fd0c1] Aug 13 00:25:21.167324 kernel: Detected PIPT I-cache on CPU36 Aug 13 00:25:21.167332 kernel: GICv3: CPU36: found redistributor 30000 region 0:0x0000100100200000 Aug 13 00:25:21.167340 kernel: GICv3: CPU36: using allocated LPI pending table @0x0000080000a30000 Aug 13 00:25:21.167347 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167355 kernel: CPU36: Booted secondary processor 0x0000030000 [0x413fd0c1] Aug 13 00:25:21.167362 kernel: Detected PIPT I-cache on CPU37 Aug 13 00:25:21.167371 kernel: GICv3: CPU37: found redistributor 50000 region 0:0x0000100100280000 Aug 13 00:25:21.167379 kernel: GICv3: CPU37: using allocated LPI pending table @0x0000080000a40000 Aug 13 00:25:21.167387 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167394 kernel: CPU37: Booted secondary processor 0x0000050000 [0x413fd0c1] Aug 13 00:25:21.167402 kernel: Detected PIPT I-cache on CPU38 Aug 13 00:25:21.167409 kernel: GICv3: CPU38: found redistributor 10000 region 0:0x0000100100180000 Aug 13 00:25:21.167417 kernel: GICv3: CPU38: using allocated LPI pending table @0x0000080000a50000 Aug 13 00:25:21.167425 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167432 kernel: CPU38: Booted secondary processor 0x0000010000 [0x413fd0c1] Aug 13 00:25:21.167440 kernel: Detected PIPT I-cache on CPU39 Aug 13 00:25:21.167449 kernel: GICv3: CPU39: found redistributor 70000 region 0:0x0000100100300000 Aug 13 00:25:21.167456 kernel: GICv3: CPU39: using allocated LPI pending table @0x0000080000a60000 Aug 13 00:25:21.167464 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167472 kernel: CPU39: Booted secondary processor 0x0000070000 [0x413fd0c1] Aug 13 00:25:21.167479 kernel: Detected PIPT I-cache on CPU40 Aug 13 00:25:21.167487 kernel: GICv3: CPU40: found redistributor 120100 region 0:0x00001001005e0000 Aug 13 00:25:21.167495 kernel: GICv3: CPU40: using allocated LPI pending table @0x0000080000a70000 Aug 13 00:25:21.167504 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167511 kernel: CPU40: Booted secondary processor 0x0000120100 [0x413fd0c1] Aug 13 00:25:21.167519 kernel: Detected PIPT I-cache on CPU41 Aug 13 00:25:21.167527 kernel: GICv3: CPU41: found redistributor 1a0100 region 0:0x00001001007e0000 Aug 13 00:25:21.167534 kernel: GICv3: CPU41: using allocated LPI pending table @0x0000080000a80000 Aug 13 00:25:21.167542 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167549 kernel: CPU41: Booted secondary processor 0x00001a0100 [0x413fd0c1] Aug 13 00:25:21.167557 kernel: Detected PIPT I-cache on CPU42 Aug 13 00:25:21.167565 kernel: GICv3: CPU42: found redistributor 140100 region 0:0x0000100100660000 Aug 13 00:25:21.167572 kernel: GICv3: CPU42: using allocated LPI pending table @0x0000080000a90000 Aug 13 00:25:21.167581 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167589 kernel: CPU42: Booted secondary processor 0x0000140100 [0x413fd0c1] Aug 13 00:25:21.167596 kernel: Detected PIPT I-cache on CPU43 Aug 13 00:25:21.167604 kernel: GICv3: CPU43: found redistributor 1c0100 region 0:0x0000100100860000 Aug 13 00:25:21.167612 kernel: GICv3: CPU43: using allocated LPI pending table @0x0000080000aa0000 Aug 13 00:25:21.167620 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167627 kernel: CPU43: Booted secondary processor 0x00001c0100 [0x413fd0c1] Aug 13 00:25:21.167635 kernel: Detected PIPT I-cache on CPU44 Aug 13 00:25:21.167642 kernel: GICv3: CPU44: found redistributor 100100 region 0:0x0000100100560000 Aug 13 00:25:21.167651 kernel: GICv3: CPU44: using allocated LPI pending table @0x0000080000ab0000 Aug 13 00:25:21.167659 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167667 kernel: CPU44: Booted secondary processor 0x0000100100 [0x413fd0c1] Aug 13 00:25:21.167674 kernel: Detected PIPT I-cache on CPU45 Aug 13 00:25:21.167682 kernel: GICv3: CPU45: found redistributor 180100 region 0:0x0000100100760000 Aug 13 00:25:21.167689 kernel: GICv3: CPU45: using allocated LPI pending table @0x0000080000ac0000 Aug 13 00:25:21.167697 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167705 kernel: CPU45: Booted secondary processor 0x0000180100 [0x413fd0c1] Aug 13 00:25:21.167712 kernel: Detected PIPT I-cache on CPU46 Aug 13 00:25:21.167720 kernel: GICv3: CPU46: found redistributor 160100 region 0:0x00001001006e0000 Aug 13 00:25:21.167729 kernel: GICv3: CPU46: using allocated LPI pending table @0x0000080000ad0000 Aug 13 00:25:21.167738 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167746 kernel: CPU46: Booted secondary processor 0x0000160100 [0x413fd0c1] Aug 13 00:25:21.167753 kernel: Detected PIPT I-cache on CPU47 Aug 13 00:25:21.167761 kernel: GICv3: CPU47: found redistributor 1e0100 region 0:0x00001001008e0000 Aug 13 00:25:21.167769 kernel: GICv3: CPU47: using allocated LPI pending table @0x0000080000ae0000 Aug 13 00:25:21.167776 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167784 kernel: CPU47: Booted secondary processor 0x00001e0100 [0x413fd0c1] Aug 13 00:25:21.167791 kernel: Detected PIPT I-cache on CPU48 Aug 13 00:25:21.167800 kernel: GICv3: CPU48: found redistributor a0100 region 0:0x00001001003e0000 Aug 13 00:25:21.167808 kernel: GICv3: CPU48: using allocated LPI pending table @0x0000080000af0000 Aug 13 00:25:21.167816 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167823 kernel: CPU48: Booted secondary processor 0x00000a0100 [0x413fd0c1] Aug 13 00:25:21.167831 kernel: Detected PIPT I-cache on CPU49 Aug 13 00:25:21.167839 kernel: GICv3: CPU49: found redistributor 220100 region 0:0x00001001009e0000 Aug 13 00:25:21.167846 kernel: GICv3: CPU49: using allocated LPI pending table @0x0000080000b00000 Aug 13 00:25:21.167854 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167861 kernel: CPU49: Booted secondary processor 0x0000220100 [0x413fd0c1] Aug 13 00:25:21.167869 kernel: Detected PIPT I-cache on CPU50 Aug 13 00:25:21.167878 kernel: GICv3: CPU50: found redistributor c0100 region 0:0x0000100100460000 Aug 13 00:25:21.167886 kernel: GICv3: CPU50: using allocated LPI pending table @0x0000080000b10000 Aug 13 00:25:21.167893 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167901 kernel: CPU50: Booted secondary processor 0x00000c0100 [0x413fd0c1] Aug 13 00:25:21.167908 kernel: Detected PIPT I-cache on CPU51 Aug 13 00:25:21.167916 kernel: GICv3: CPU51: found redistributor 240100 region 0:0x0000100100a60000 Aug 13 00:25:21.167923 kernel: GICv3: CPU51: using allocated LPI pending table @0x0000080000b20000 Aug 13 00:25:21.167931 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167938 kernel: CPU51: Booted secondary processor 0x0000240100 [0x413fd0c1] Aug 13 00:25:21.167947 kernel: Detected PIPT I-cache on CPU52 Aug 13 00:25:21.167955 kernel: GICv3: CPU52: found redistributor 80100 region 0:0x0000100100360000 Aug 13 00:25:21.167963 kernel: GICv3: CPU52: using allocated LPI pending table @0x0000080000b30000 Aug 13 00:25:21.167970 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167978 kernel: CPU52: Booted secondary processor 0x0000080100 [0x413fd0c1] Aug 13 00:25:21.167986 kernel: Detected PIPT I-cache on CPU53 Aug 13 00:25:21.167993 kernel: GICv3: CPU53: found redistributor 200100 region 0:0x0000100100960000 Aug 13 00:25:21.168002 kernel: GICv3: CPU53: using allocated LPI pending table @0x0000080000b40000 Aug 13 00:25:21.168010 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168017 kernel: CPU53: Booted secondary processor 0x0000200100 [0x413fd0c1] Aug 13 00:25:21.168026 kernel: Detected PIPT I-cache on CPU54 Aug 13 00:25:21.168034 kernel: GICv3: CPU54: found redistributor e0100 region 0:0x00001001004e0000 Aug 13 00:25:21.168044 kernel: GICv3: CPU54: using allocated LPI pending table @0x0000080000b50000 Aug 13 00:25:21.168052 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168059 kernel: CPU54: Booted secondary processor 0x00000e0100 [0x413fd0c1] Aug 13 00:25:21.168067 kernel: Detected PIPT I-cache on CPU55 Aug 13 00:25:21.168075 kernel: GICv3: CPU55: found redistributor 260100 region 0:0x0000100100ae0000 Aug 13 00:25:21.168082 kernel: GICv3: CPU55: using allocated LPI pending table @0x0000080000b60000 Aug 13 00:25:21.168090 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168100 kernel: CPU55: Booted secondary processor 0x0000260100 [0x413fd0c1] Aug 13 00:25:21.168107 kernel: Detected PIPT I-cache on CPU56 Aug 13 00:25:21.168115 kernel: GICv3: CPU56: found redistributor 20100 region 0:0x00001001001e0000 Aug 13 00:25:21.168122 kernel: GICv3: CPU56: using allocated LPI pending table @0x0000080000b70000 Aug 13 00:25:21.168130 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168138 kernel: CPU56: Booted secondary processor 0x0000020100 [0x413fd0c1] Aug 13 00:25:21.168145 kernel: Detected PIPT I-cache on CPU57 Aug 13 00:25:21.168153 kernel: GICv3: CPU57: found redistributor 40100 region 0:0x0000100100260000 Aug 13 00:25:21.168161 kernel: GICv3: CPU57: using allocated LPI pending table @0x0000080000b80000 Aug 13 00:25:21.168170 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168177 kernel: CPU57: Booted secondary processor 0x0000040100 [0x413fd0c1] Aug 13 00:25:21.168185 kernel: Detected PIPT I-cache on CPU58 Aug 13 00:25:21.168193 kernel: GICv3: CPU58: found redistributor 100 region 0:0x0000100100160000 Aug 13 00:25:21.168200 kernel: GICv3: CPU58: using allocated LPI pending table @0x0000080000b90000 Aug 13 00:25:21.168208 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168216 kernel: CPU58: Booted secondary processor 0x0000000100 [0x413fd0c1] Aug 13 00:25:21.168223 kernel: Detected PIPT I-cache on CPU59 Aug 13 00:25:21.168231 kernel: GICv3: CPU59: found redistributor 60100 region 0:0x00001001002e0000 Aug 13 00:25:21.168238 kernel: GICv3: CPU59: using allocated LPI pending table @0x0000080000ba0000 Aug 13 00:25:21.168247 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168255 kernel: CPU59: Booted secondary processor 0x0000060100 [0x413fd0c1] Aug 13 00:25:21.168263 kernel: Detected PIPT I-cache on CPU60 Aug 13 00:25:21.168270 kernel: GICv3: CPU60: found redistributor 130100 region 0:0x0000100100620000 Aug 13 00:25:21.168278 kernel: GICv3: CPU60: using allocated LPI pending table @0x0000080000bb0000 Aug 13 00:25:21.168286 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168293 kernel: CPU60: Booted secondary processor 0x0000130100 [0x413fd0c1] Aug 13 00:25:21.168301 kernel: Detected PIPT I-cache on CPU61 Aug 13 00:25:21.168308 kernel: GICv3: CPU61: found redistributor 1b0100 region 0:0x0000100100820000 Aug 13 00:25:21.168318 kernel: GICv3: CPU61: using allocated LPI pending table @0x0000080000bc0000 Aug 13 00:25:21.168325 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168333 kernel: CPU61: Booted secondary processor 0x00001b0100 [0x413fd0c1] Aug 13 00:25:21.168341 kernel: Detected PIPT I-cache on CPU62 Aug 13 00:25:21.168349 kernel: GICv3: CPU62: found redistributor 150100 region 0:0x00001001006a0000 Aug 13 00:25:21.168357 kernel: GICv3: CPU62: using allocated LPI pending table @0x0000080000bd0000 Aug 13 00:25:21.168365 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168372 kernel: CPU62: Booted secondary processor 0x0000150100 [0x413fd0c1] Aug 13 00:25:21.168380 kernel: Detected PIPT I-cache on CPU63 Aug 13 00:25:21.168387 kernel: GICv3: CPU63: found redistributor 1d0100 region 0:0x00001001008a0000 Aug 13 00:25:21.168397 kernel: GICv3: CPU63: using allocated LPI pending table @0x0000080000be0000 Aug 13 00:25:21.168404 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168412 kernel: CPU63: Booted secondary processor 0x00001d0100 [0x413fd0c1] Aug 13 00:25:21.168420 kernel: Detected PIPT I-cache on CPU64 Aug 13 00:25:21.168427 kernel: GICv3: CPU64: found redistributor 110100 region 0:0x00001001005a0000 Aug 13 00:25:21.168435 kernel: GICv3: CPU64: using allocated LPI pending table @0x0000080000bf0000 Aug 13 00:25:21.168443 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168450 kernel: CPU64: Booted secondary processor 0x0000110100 [0x413fd0c1] Aug 13 00:25:21.168458 kernel: Detected PIPT I-cache on CPU65 Aug 13 00:25:21.168467 kernel: GICv3: CPU65: found redistributor 190100 region 0:0x00001001007a0000 Aug 13 00:25:21.168474 kernel: GICv3: CPU65: using allocated LPI pending table @0x0000080000c00000 Aug 13 00:25:21.168482 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168490 kernel: CPU65: Booted secondary processor 0x0000190100 [0x413fd0c1] Aug 13 00:25:21.168497 kernel: Detected PIPT I-cache on CPU66 Aug 13 00:25:21.168505 kernel: GICv3: CPU66: found redistributor 170100 region 0:0x0000100100720000 Aug 13 00:25:21.168513 kernel: GICv3: CPU66: using allocated LPI pending table @0x0000080000c10000 Aug 13 00:25:21.168520 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168528 kernel: CPU66: Booted secondary processor 0x0000170100 [0x413fd0c1] Aug 13 00:25:21.168535 kernel: Detected PIPT I-cache on CPU67 Aug 13 00:25:21.168545 kernel: GICv3: CPU67: found redistributor 1f0100 region 0:0x0000100100920000 Aug 13 00:25:21.168552 kernel: GICv3: CPU67: using allocated LPI pending table @0x0000080000c20000 Aug 13 00:25:21.168560 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168568 kernel: CPU67: Booted secondary processor 0x00001f0100 [0x413fd0c1] Aug 13 00:25:21.168575 kernel: Detected PIPT I-cache on CPU68 Aug 13 00:25:21.168583 kernel: GICv3: CPU68: found redistributor b0100 region 0:0x0000100100420000 Aug 13 00:25:21.168591 kernel: GICv3: CPU68: using allocated LPI pending table @0x0000080000c30000 Aug 13 00:25:21.168598 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168606 kernel: CPU68: Booted secondary processor 0x00000b0100 [0x413fd0c1] Aug 13 00:25:21.168615 kernel: Detected PIPT I-cache on CPU69 Aug 13 00:25:21.168623 kernel: GICv3: CPU69: found redistributor 230100 region 0:0x0000100100a20000 Aug 13 00:25:21.168630 kernel: GICv3: CPU69: using allocated LPI pending table @0x0000080000c40000 Aug 13 00:25:21.168638 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168645 kernel: CPU69: Booted secondary processor 0x0000230100 [0x413fd0c1] Aug 13 00:25:21.168653 kernel: Detected PIPT I-cache on CPU70 Aug 13 00:25:21.168661 kernel: GICv3: CPU70: found redistributor d0100 region 0:0x00001001004a0000 Aug 13 00:25:21.168669 kernel: GICv3: CPU70: using allocated LPI pending table @0x0000080000c50000 Aug 13 00:25:21.168677 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168684 kernel: CPU70: Booted secondary processor 0x00000d0100 [0x413fd0c1] Aug 13 00:25:21.168693 kernel: Detected PIPT I-cache on CPU71 Aug 13 00:25:21.168701 kernel: GICv3: CPU71: found redistributor 250100 region 0:0x0000100100aa0000 Aug 13 00:25:21.168709 kernel: GICv3: CPU71: using allocated LPI pending table @0x0000080000c60000 Aug 13 00:25:21.168716 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168724 kernel: CPU71: Booted secondary processor 0x0000250100 [0x413fd0c1] Aug 13 00:25:21.168731 kernel: Detected PIPT I-cache on CPU72 Aug 13 00:25:21.168739 kernel: GICv3: CPU72: found redistributor 90100 region 0:0x00001001003a0000 Aug 13 00:25:21.168747 kernel: GICv3: CPU72: using allocated LPI pending table @0x0000080000c70000 Aug 13 00:25:21.168754 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168763 kernel: CPU72: Booted secondary processor 0x0000090100 [0x413fd0c1] Aug 13 00:25:21.168771 kernel: Detected PIPT I-cache on CPU73 Aug 13 00:25:21.168779 kernel: GICv3: CPU73: found redistributor 210100 region 0:0x00001001009a0000 Aug 13 00:25:21.168786 kernel: GICv3: CPU73: using allocated LPI pending table @0x0000080000c80000 Aug 13 00:25:21.168794 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168801 kernel: CPU73: Booted secondary processor 0x0000210100 [0x413fd0c1] Aug 13 00:25:21.168809 kernel: Detected PIPT I-cache on CPU74 Aug 13 00:25:21.168817 kernel: GICv3: CPU74: found redistributor f0100 region 0:0x0000100100520000 Aug 13 00:25:21.168824 kernel: GICv3: CPU74: using allocated LPI pending table @0x0000080000c90000 Aug 13 00:25:21.168833 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168841 kernel: CPU74: Booted secondary processor 0x00000f0100 [0x413fd0c1] Aug 13 00:25:21.168849 kernel: Detected PIPT I-cache on CPU75 Aug 13 00:25:21.168856 kernel: GICv3: CPU75: found redistributor 270100 region 0:0x0000100100b20000 Aug 13 00:25:21.168864 kernel: GICv3: CPU75: using allocated LPI pending table @0x0000080000ca0000 Aug 13 00:25:21.168872 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168879 kernel: CPU75: Booted secondary processor 0x0000270100 [0x413fd0c1] Aug 13 00:25:21.168887 kernel: Detected PIPT I-cache on CPU76 Aug 13 00:25:21.168894 kernel: GICv3: CPU76: found redistributor 30100 region 0:0x0000100100220000 Aug 13 00:25:21.168902 kernel: GICv3: CPU76: using allocated LPI pending table @0x0000080000cb0000 Aug 13 00:25:21.168911 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168919 kernel: CPU76: Booted secondary processor 0x0000030100 [0x413fd0c1] Aug 13 00:25:21.168926 kernel: Detected PIPT I-cache on CPU77 Aug 13 00:25:21.168934 kernel: GICv3: CPU77: found redistributor 50100 region 0:0x00001001002a0000 Aug 13 00:25:21.168942 kernel: GICv3: CPU77: using allocated LPI pending table @0x0000080000cc0000 Aug 13 00:25:21.168949 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168957 kernel: CPU77: Booted secondary processor 0x0000050100 [0x413fd0c1] Aug 13 00:25:21.168964 kernel: Detected PIPT I-cache on CPU78 Aug 13 00:25:21.168972 kernel: GICv3: CPU78: found redistributor 10100 region 0:0x00001001001a0000 Aug 13 00:25:21.168981 kernel: GICv3: CPU78: using allocated LPI pending table @0x0000080000cd0000 Aug 13 00:25:21.168989 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168996 kernel: CPU78: Booted secondary processor 0x0000010100 [0x413fd0c1] Aug 13 00:25:21.169004 kernel: Detected PIPT I-cache on CPU79 Aug 13 00:25:21.169011 kernel: GICv3: CPU79: found redistributor 70100 region 0:0x0000100100320000 Aug 13 00:25:21.169019 kernel: GICv3: CPU79: using allocated LPI pending table @0x0000080000ce0000 Aug 13 00:25:21.169027 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.169035 kernel: CPU79: Booted secondary processor 0x0000070100 [0x413fd0c1] Aug 13 00:25:21.169044 kernel: smp: Brought up 1 node, 80 CPUs Aug 13 00:25:21.169052 kernel: SMP: Total of 80 processors activated. Aug 13 00:25:21.169061 kernel: CPU features: detected: 32-bit EL0 Support Aug 13 00:25:21.169068 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence Aug 13 00:25:21.169076 kernel: CPU features: detected: Common not Private translations Aug 13 00:25:21.169084 kernel: CPU features: detected: CRC32 instructions Aug 13 00:25:21.169092 kernel: CPU features: detected: Enhanced Virtualization Traps Aug 13 00:25:21.169099 kernel: CPU features: detected: RCpc load-acquire (LDAPR) Aug 13 00:25:21.169107 kernel: CPU features: detected: LSE atomic instructions Aug 13 00:25:21.169115 kernel: CPU features: detected: Privileged Access Never Aug 13 00:25:21.169122 kernel: CPU features: detected: RAS Extension Support Aug 13 00:25:21.169131 kernel: CPU features: detected: Speculative Store Bypassing Safe (SSBS) Aug 13 00:25:21.169139 kernel: CPU: All CPU(s) started at EL2 Aug 13 00:25:21.169147 kernel: alternatives: applying system-wide alternatives Aug 13 00:25:21.169154 kernel: devtmpfs: initialized Aug 13 00:25:21.169162 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Aug 13 00:25:21.169170 kernel: futex hash table entries: 32768 (order: 9, 2097152 bytes, linear) Aug 13 00:25:21.169177 kernel: pinctrl core: initialized pinctrl subsystem Aug 13 00:25:21.169185 kernel: SMBIOS 3.4.0 present. Aug 13 00:25:21.169192 kernel: DMI: GIGABYTE R272-P30-JG/MP32-AR0-JG, BIOS F17a (SCP: 1.07.20210713) 07/22/2021 Aug 13 00:25:21.169201 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Aug 13 00:25:21.169209 kernel: DMA: preallocated 4096 KiB GFP_KERNEL pool for atomic allocations Aug 13 00:25:21.169217 kernel: DMA: preallocated 4096 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Aug 13 00:25:21.169225 kernel: DMA: preallocated 4096 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Aug 13 00:25:21.169232 kernel: audit: initializing netlink subsys (disabled) Aug 13 00:25:21.169240 kernel: audit: type=2000 audit(0.042:1): state=initialized audit_enabled=0 res=1 Aug 13 00:25:21.169248 kernel: thermal_sys: Registered thermal governor 'step_wise' Aug 13 00:25:21.169255 kernel: cpuidle: using governor menu Aug 13 00:25:21.169263 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Aug 13 00:25:21.169272 kernel: ASID allocator initialised with 32768 entries Aug 13 00:25:21.169280 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Aug 13 00:25:21.169288 kernel: Serial: AMBA PL011 UART driver Aug 13 00:25:21.169295 kernel: Modules: 2G module region forced by RANDOMIZE_MODULE_REGION_FULL Aug 13 00:25:21.169303 kernel: Modules: 0 pages in range for non-PLT usage Aug 13 00:25:21.169311 kernel: Modules: 509008 pages in range for PLT usage Aug 13 00:25:21.169318 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Aug 13 00:25:21.169326 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page Aug 13 00:25:21.169334 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages Aug 13 00:25:21.169343 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page Aug 13 00:25:21.169350 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Aug 13 00:25:21.169358 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page Aug 13 00:25:21.169366 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages Aug 13 00:25:21.169373 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page Aug 13 00:25:21.169381 kernel: ACPI: Added _OSI(Module Device) Aug 13 00:25:21.169388 kernel: ACPI: Added _OSI(Processor Device) Aug 13 00:25:21.169396 kernel: ACPI: Added _OSI(Processor Aggregator Device) Aug 13 00:25:21.169404 kernel: ACPI: 2 ACPI AML tables successfully acquired and loaded Aug 13 00:25:21.169413 kernel: ACPI: Interpreter enabled Aug 13 00:25:21.169421 kernel: ACPI: Using GIC for interrupt routing Aug 13 00:25:21.169428 kernel: ACPI: MCFG table detected, 8 entries Aug 13 00:25:21.169436 kernel: ACPI: IORT: SMMU-v3[33ffe0000000] Mapped to Proximity domain 0 Aug 13 00:25:21.169443 kernel: ACPI: IORT: SMMU-v3[37ffe0000000] Mapped to Proximity domain 0 Aug 13 00:25:21.169451 kernel: ACPI: IORT: SMMU-v3[3bffe0000000] Mapped to Proximity domain 0 Aug 13 00:25:21.169459 kernel: ACPI: IORT: SMMU-v3[3fffe0000000] Mapped to Proximity domain 0 Aug 13 00:25:21.169466 kernel: ACPI: IORT: SMMU-v3[23ffe0000000] Mapped to Proximity domain 0 Aug 13 00:25:21.169474 kernel: ACPI: IORT: SMMU-v3[27ffe0000000] Mapped to Proximity domain 0 Aug 13 00:25:21.169483 kernel: ACPI: IORT: SMMU-v3[2bffe0000000] Mapped to Proximity domain 0 Aug 13 00:25:21.169491 kernel: ACPI: IORT: SMMU-v3[2fffe0000000] Mapped to Proximity domain 0 Aug 13 00:25:21.169499 kernel: ARMH0011:00: ttyAMA0 at MMIO 0x100002600000 (irq = 19, base_baud = 0) is a SBSA Aug 13 00:25:21.169507 kernel: printk: console [ttyAMA0] enabled Aug 13 00:25:21.169515 kernel: ARMH0011:01: ttyAMA1 at MMIO 0x100002620000 (irq = 20, base_baud = 0) is a SBSA Aug 13 00:25:21.169522 kernel: ACPI: PCI Root Bridge [PCI1] (domain 000d [bus 00-ff]) Aug 13 00:25:21.169671 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Aug 13 00:25:21.169762 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug PME LTR] Aug 13 00:25:21.169831 kernel: acpi PNP0A08:00: _OSC: OS now controls [AER PCIeCapability] Aug 13 00:25:21.169893 kernel: acpi PNP0A08:00: MCFG quirk: ECAM at [mem 0x37fff0000000-0x37ffffffffff] for [bus 00-ff] with pci_32b_read_ops Aug 13 00:25:21.169956 kernel: acpi PNP0A08:00: ECAM area [mem 0x37fff0000000-0x37ffffffffff] reserved by PNP0C02:00 Aug 13 00:25:21.170018 kernel: acpi PNP0A08:00: ECAM at [mem 0x37fff0000000-0x37ffffffffff] for [bus 00-ff] Aug 13 00:25:21.170028 kernel: PCI host bridge to bus 000d:00 Aug 13 00:25:21.170103 kernel: pci_bus 000d:00: root bus resource [mem 0x50000000-0x5fffffff window] Aug 13 00:25:21.170164 kernel: pci_bus 000d:00: root bus resource [mem 0x340000000000-0x37ffdfffffff window] Aug 13 00:25:21.170222 kernel: pci_bus 000d:00: root bus resource [bus 00-ff] Aug 13 00:25:21.170300 kernel: pci 000d:00:00.0: [1def:e100] type 00 class 0x060000 Aug 13 00:25:21.170375 kernel: pci 000d:00:01.0: [1def:e101] type 01 class 0x060400 Aug 13 00:25:21.170441 kernel: pci 000d:00:01.0: enabling Extended Tags Aug 13 00:25:21.170507 kernel: pci 000d:00:01.0: supports D1 D2 Aug 13 00:25:21.170570 kernel: pci 000d:00:01.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.170648 kernel: pci 000d:00:02.0: [1def:e102] type 01 class 0x060400 Aug 13 00:25:21.170715 kernel: pci 000d:00:02.0: supports D1 D2 Aug 13 00:25:21.170779 kernel: pci 000d:00:02.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.170851 kernel: pci 000d:00:03.0: [1def:e103] type 01 class 0x060400 Aug 13 00:25:21.170917 kernel: pci 000d:00:03.0: supports D1 D2 Aug 13 00:25:21.170980 kernel: pci 000d:00:03.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.171055 kernel: pci 000d:00:04.0: [1def:e104] type 01 class 0x060400 Aug 13 00:25:21.171126 kernel: pci 000d:00:04.0: supports D1 D2 Aug 13 00:25:21.171191 kernel: pci 000d:00:04.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.171202 kernel: acpiphp: Slot [1] registered Aug 13 00:25:21.171209 kernel: acpiphp: Slot [2] registered Aug 13 00:25:21.171217 kernel: acpiphp: Slot [3] registered Aug 13 00:25:21.171225 kernel: acpiphp: Slot [4] registered Aug 13 00:25:21.171284 kernel: pci_bus 000d:00: on NUMA node 0 Aug 13 00:25:21.171353 kernel: pci 000d:00:01.0: bridge window [io 0x1000-0x0fff] to [bus 01] add_size 1000 Aug 13 00:25:21.171418 kernel: pci 000d:00:01.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 01] add_size 200000 add_align 100000 Aug 13 00:25:21.171483 kernel: pci 000d:00:01.0: bridge window [mem 0x00100000-0x000fffff] to [bus 01] add_size 200000 add_align 100000 Aug 13 00:25:21.171548 kernel: pci 000d:00:02.0: bridge window [io 0x1000-0x0fff] to [bus 02] add_size 1000 Aug 13 00:25:21.171612 kernel: pci 000d:00:02.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 02] add_size 200000 add_align 100000 Aug 13 00:25:21.171677 kernel: pci 000d:00:02.0: bridge window [mem 0x00100000-0x000fffff] to [bus 02] add_size 200000 add_align 100000 Aug 13 00:25:21.171742 kernel: pci 000d:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 03] add_size 1000 Aug 13 00:25:21.171813 kernel: pci 000d:00:03.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 03] add_size 200000 add_align 100000 Aug 13 00:25:21.171876 kernel: pci 000d:00:03.0: bridge window [mem 0x00100000-0x000fffff] to [bus 03] add_size 200000 add_align 100000 Aug 13 00:25:21.171942 kernel: pci 000d:00:04.0: bridge window [io 0x1000-0x0fff] to [bus 04] add_size 1000 Aug 13 00:25:21.172005 kernel: pci 000d:00:04.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 04] add_size 200000 add_align 100000 Aug 13 00:25:21.172073 kernel: pci 000d:00:04.0: bridge window [mem 0x00100000-0x000fffff] to [bus 04] add_size 200000 add_align 100000 Aug 13 00:25:21.172138 kernel: pci 000d:00:01.0: BAR 14: assigned [mem 0x50000000-0x501fffff] Aug 13 00:25:21.172204 kernel: pci 000d:00:01.0: BAR 15: assigned [mem 0x340000000000-0x3400001fffff 64bit pref] Aug 13 00:25:21.172272 kernel: pci 000d:00:02.0: BAR 14: assigned [mem 0x50200000-0x503fffff] Aug 13 00:25:21.172336 kernel: pci 000d:00:02.0: BAR 15: assigned [mem 0x340000200000-0x3400003fffff 64bit pref] Aug 13 00:25:21.172402 kernel: pci 000d:00:03.0: BAR 14: assigned [mem 0x50400000-0x505fffff] Aug 13 00:25:21.172467 kernel: pci 000d:00:03.0: BAR 15: assigned [mem 0x340000400000-0x3400005fffff 64bit pref] Aug 13 00:25:21.172532 kernel: pci 000d:00:04.0: BAR 14: assigned [mem 0x50600000-0x507fffff] Aug 13 00:25:21.172596 kernel: pci 000d:00:04.0: BAR 15: assigned [mem 0x340000600000-0x3400007fffff 64bit pref] Aug 13 00:25:21.172661 kernel: pci 000d:00:01.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.172724 kernel: pci 000d:00:01.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.172792 kernel: pci 000d:00:02.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.172856 kernel: pci 000d:00:02.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.172920 kernel: pci 000d:00:03.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.172984 kernel: pci 000d:00:03.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.173056 kernel: pci 000d:00:04.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.173120 kernel: pci 000d:00:04.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.173185 kernel: pci 000d:00:04.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.173255 kernel: pci 000d:00:04.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.173319 kernel: pci 000d:00:03.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.173383 kernel: pci 000d:00:03.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.173448 kernel: pci 000d:00:02.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.173513 kernel: pci 000d:00:02.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.173576 kernel: pci 000d:00:01.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.173641 kernel: pci 000d:00:01.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.173704 kernel: pci 000d:00:01.0: PCI bridge to [bus 01] Aug 13 00:25:21.173772 kernel: pci 000d:00:01.0: bridge window [mem 0x50000000-0x501fffff] Aug 13 00:25:21.173836 kernel: pci 000d:00:01.0: bridge window [mem 0x340000000000-0x3400001fffff 64bit pref] Aug 13 00:25:21.173901 kernel: pci 000d:00:02.0: PCI bridge to [bus 02] Aug 13 00:25:21.173966 kernel: pci 000d:00:02.0: bridge window [mem 0x50200000-0x503fffff] Aug 13 00:25:21.174031 kernel: pci 000d:00:02.0: bridge window [mem 0x340000200000-0x3400003fffff 64bit pref] Aug 13 00:25:21.174099 kernel: pci 000d:00:03.0: PCI bridge to [bus 03] Aug 13 00:25:21.174164 kernel: pci 000d:00:03.0: bridge window [mem 0x50400000-0x505fffff] Aug 13 00:25:21.174230 kernel: pci 000d:00:03.0: bridge window [mem 0x340000400000-0x3400005fffff 64bit pref] Aug 13 00:25:21.174295 kernel: pci 000d:00:04.0: PCI bridge to [bus 04] Aug 13 00:25:21.174358 kernel: pci 000d:00:04.0: bridge window [mem 0x50600000-0x507fffff] Aug 13 00:25:21.174423 kernel: pci 000d:00:04.0: bridge window [mem 0x340000600000-0x3400007fffff 64bit pref] Aug 13 00:25:21.174481 kernel: pci_bus 000d:00: resource 4 [mem 0x50000000-0x5fffffff window] Aug 13 00:25:21.174538 kernel: pci_bus 000d:00: resource 5 [mem 0x340000000000-0x37ffdfffffff window] Aug 13 00:25:21.174611 kernel: pci_bus 000d:01: resource 1 [mem 0x50000000-0x501fffff] Aug 13 00:25:21.174671 kernel: pci_bus 000d:01: resource 2 [mem 0x340000000000-0x3400001fffff 64bit pref] Aug 13 00:25:21.174739 kernel: pci_bus 000d:02: resource 1 [mem 0x50200000-0x503fffff] Aug 13 00:25:21.174799 kernel: pci_bus 000d:02: resource 2 [mem 0x340000200000-0x3400003fffff 64bit pref] Aug 13 00:25:21.174874 kernel: pci_bus 000d:03: resource 1 [mem 0x50400000-0x505fffff] Aug 13 00:25:21.174933 kernel: pci_bus 000d:03: resource 2 [mem 0x340000400000-0x3400005fffff 64bit pref] Aug 13 00:25:21.175004 kernel: pci_bus 000d:04: resource 1 [mem 0x50600000-0x507fffff] Aug 13 00:25:21.175066 kernel: pci_bus 000d:04: resource 2 [mem 0x340000600000-0x3400007fffff 64bit pref] Aug 13 00:25:21.175077 kernel: ACPI: PCI Root Bridge [PCI3] (domain 0000 [bus 00-ff]) Aug 13 00:25:21.175149 kernel: acpi PNP0A08:01: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Aug 13 00:25:21.175213 kernel: acpi PNP0A08:01: _OSC: platform does not support [PCIeHotplug PME LTR] Aug 13 00:25:21.175275 kernel: acpi PNP0A08:01: _OSC: OS now controls [AER PCIeCapability] Aug 13 00:25:21.175339 kernel: acpi PNP0A08:01: MCFG quirk: ECAM at [mem 0x3ffff0000000-0x3fffffffffff] for [bus 00-ff] with pci_32b_read_ops Aug 13 00:25:21.175401 kernel: acpi PNP0A08:01: ECAM area [mem 0x3ffff0000000-0x3fffffffffff] reserved by PNP0C02:00 Aug 13 00:25:21.175463 kernel: acpi PNP0A08:01: ECAM at [mem 0x3ffff0000000-0x3fffffffffff] for [bus 00-ff] Aug 13 00:25:21.175473 kernel: PCI host bridge to bus 0000:00 Aug 13 00:25:21.175537 kernel: pci_bus 0000:00: root bus resource [mem 0x70000000-0x7fffffff window] Aug 13 00:25:21.175597 kernel: pci_bus 0000:00: root bus resource [mem 0x3c0000000000-0x3fffdfffffff window] Aug 13 00:25:21.175653 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Aug 13 00:25:21.175728 kernel: pci 0000:00:00.0: [1def:e100] type 00 class 0x060000 Aug 13 00:25:21.175801 kernel: pci 0000:00:01.0: [1def:e101] type 01 class 0x060400 Aug 13 00:25:21.175867 kernel: pci 0000:00:01.0: enabling Extended Tags Aug 13 00:25:21.175931 kernel: pci 0000:00:01.0: supports D1 D2 Aug 13 00:25:21.175997 kernel: pci 0000:00:01.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.176071 kernel: pci 0000:00:02.0: [1def:e102] type 01 class 0x060400 Aug 13 00:25:21.176138 kernel: pci 0000:00:02.0: supports D1 D2 Aug 13 00:25:21.176206 kernel: pci 0000:00:02.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.176277 kernel: pci 0000:00:03.0: [1def:e103] type 01 class 0x060400 Aug 13 00:25:21.176342 kernel: pci 0000:00:03.0: supports D1 D2 Aug 13 00:25:21.176407 kernel: pci 0000:00:03.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.176480 kernel: pci 0000:00:04.0: [1def:e104] type 01 class 0x060400 Aug 13 00:25:21.176545 kernel: pci 0000:00:04.0: supports D1 D2 Aug 13 00:25:21.176610 kernel: pci 0000:00:04.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.176621 kernel: acpiphp: Slot [1-1] registered Aug 13 00:25:21.176629 kernel: acpiphp: Slot [2-1] registered Aug 13 00:25:21.176637 kernel: acpiphp: Slot [3-1] registered Aug 13 00:25:21.176645 kernel: acpiphp: Slot [4-1] registered Aug 13 00:25:21.176702 kernel: pci_bus 0000:00: on NUMA node 0 Aug 13 00:25:21.176767 kernel: pci 0000:00:01.0: bridge window [io 0x1000-0x0fff] to [bus 01] add_size 1000 Aug 13 00:25:21.176831 kernel: pci 0000:00:01.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 01] add_size 200000 add_align 100000 Aug 13 00:25:21.176896 kernel: pci 0000:00:01.0: bridge window [mem 0x00100000-0x000fffff] to [bus 01] add_size 200000 add_align 100000 Aug 13 00:25:21.176962 kernel: pci 0000:00:02.0: bridge window [io 0x1000-0x0fff] to [bus 02] add_size 1000 Aug 13 00:25:21.177027 kernel: pci 0000:00:02.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 02] add_size 200000 add_align 100000 Aug 13 00:25:21.177095 kernel: pci 0000:00:02.0: bridge window [mem 0x00100000-0x000fffff] to [bus 02] add_size 200000 add_align 100000 Aug 13 00:25:21.177160 kernel: pci 0000:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 03] add_size 1000 Aug 13 00:25:21.177223 kernel: pci 0000:00:03.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 03] add_size 200000 add_align 100000 Aug 13 00:25:21.177288 kernel: pci 0000:00:03.0: bridge window [mem 0x00100000-0x000fffff] to [bus 03] add_size 200000 add_align 100000 Aug 13 00:25:21.177352 kernel: pci 0000:00:04.0: bridge window [io 0x1000-0x0fff] to [bus 04] add_size 1000 Aug 13 00:25:21.177419 kernel: pci 0000:00:04.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 04] add_size 200000 add_align 100000 Aug 13 00:25:21.177483 kernel: pci 0000:00:04.0: bridge window [mem 0x00100000-0x000fffff] to [bus 04] add_size 200000 add_align 100000 Aug 13 00:25:21.177547 kernel: pci 0000:00:01.0: BAR 14: assigned [mem 0x70000000-0x701fffff] Aug 13 00:25:21.177610 kernel: pci 0000:00:01.0: BAR 15: assigned [mem 0x3c0000000000-0x3c00001fffff 64bit pref] Aug 13 00:25:21.177675 kernel: pci 0000:00:02.0: BAR 14: assigned [mem 0x70200000-0x703fffff] Aug 13 00:25:21.177740 kernel: pci 0000:00:02.0: BAR 15: assigned [mem 0x3c0000200000-0x3c00003fffff 64bit pref] Aug 13 00:25:21.177804 kernel: pci 0000:00:03.0: BAR 14: assigned [mem 0x70400000-0x705fffff] Aug 13 00:25:21.177871 kernel: pci 0000:00:03.0: BAR 15: assigned [mem 0x3c0000400000-0x3c00005fffff 64bit pref] Aug 13 00:25:21.177934 kernel: pci 0000:00:04.0: BAR 14: assigned [mem 0x70600000-0x707fffff] Aug 13 00:25:21.177999 kernel: pci 0000:00:04.0: BAR 15: assigned [mem 0x3c0000600000-0x3c00007fffff 64bit pref] Aug 13 00:25:21.178066 kernel: pci 0000:00:01.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.178130 kernel: pci 0000:00:01.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.178194 kernel: pci 0000:00:02.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.178259 kernel: pci 0000:00:02.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.178322 kernel: pci 0000:00:03.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.178390 kernel: pci 0000:00:03.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.178454 kernel: pci 0000:00:04.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.178518 kernel: pci 0000:00:04.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.178581 kernel: pci 0000:00:04.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.178645 kernel: pci 0000:00:04.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.178708 kernel: pci 0000:00:03.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.178771 kernel: pci 0000:00:03.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.178835 kernel: pci 0000:00:02.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.178898 kernel: pci 0000:00:02.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.178966 kernel: pci 0000:00:01.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.179029 kernel: pci 0000:00:01.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.179097 kernel: pci 0000:00:01.0: PCI bridge to [bus 01] Aug 13 00:25:21.179161 kernel: pci 0000:00:01.0: bridge window [mem 0x70000000-0x701fffff] Aug 13 00:25:21.179225 kernel: pci 0000:00:01.0: bridge window [mem 0x3c0000000000-0x3c00001fffff 64bit pref] Aug 13 00:25:21.179287 kernel: pci 0000:00:02.0: PCI bridge to [bus 02] Aug 13 00:25:21.179352 kernel: pci 0000:00:02.0: bridge window [mem 0x70200000-0x703fffff] Aug 13 00:25:21.179418 kernel: pci 0000:00:02.0: bridge window [mem 0x3c0000200000-0x3c00003fffff 64bit pref] Aug 13 00:25:21.179485 kernel: pci 0000:00:03.0: PCI bridge to [bus 03] Aug 13 00:25:21.179548 kernel: pci 0000:00:03.0: bridge window [mem 0x70400000-0x705fffff] Aug 13 00:25:21.179612 kernel: pci 0000:00:03.0: bridge window [mem 0x3c0000400000-0x3c00005fffff 64bit pref] Aug 13 00:25:21.179678 kernel: pci 0000:00:04.0: PCI bridge to [bus 04] Aug 13 00:25:21.179742 kernel: pci 0000:00:04.0: bridge window [mem 0x70600000-0x707fffff] Aug 13 00:25:21.179807 kernel: pci 0000:00:04.0: bridge window [mem 0x3c0000600000-0x3c00007fffff 64bit pref] Aug 13 00:25:21.179866 kernel: pci_bus 0000:00: resource 4 [mem 0x70000000-0x7fffffff window] Aug 13 00:25:21.179923 kernel: pci_bus 0000:00: resource 5 [mem 0x3c0000000000-0x3fffdfffffff window] Aug 13 00:25:21.179991 kernel: pci_bus 0000:01: resource 1 [mem 0x70000000-0x701fffff] Aug 13 00:25:21.180059 kernel: pci_bus 0000:01: resource 2 [mem 0x3c0000000000-0x3c00001fffff 64bit pref] Aug 13 00:25:21.180126 kernel: pci_bus 0000:02: resource 1 [mem 0x70200000-0x703fffff] Aug 13 00:25:21.180186 kernel: pci_bus 0000:02: resource 2 [mem 0x3c0000200000-0x3c00003fffff 64bit pref] Aug 13 00:25:21.180262 kernel: pci_bus 0000:03: resource 1 [mem 0x70400000-0x705fffff] Aug 13 00:25:21.180322 kernel: pci_bus 0000:03: resource 2 [mem 0x3c0000400000-0x3c00005fffff 64bit pref] Aug 13 00:25:21.180389 kernel: pci_bus 0000:04: resource 1 [mem 0x70600000-0x707fffff] Aug 13 00:25:21.180453 kernel: pci_bus 0000:04: resource 2 [mem 0x3c0000600000-0x3c00007fffff 64bit pref] Aug 13 00:25:21.180463 kernel: ACPI: PCI Root Bridge [PCI7] (domain 0005 [bus 00-ff]) Aug 13 00:25:21.180533 kernel: acpi PNP0A08:02: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Aug 13 00:25:21.180596 kernel: acpi PNP0A08:02: _OSC: platform does not support [PCIeHotplug PME LTR] Aug 13 00:25:21.180658 kernel: acpi PNP0A08:02: _OSC: OS now controls [AER PCIeCapability] Aug 13 00:25:21.180722 kernel: acpi PNP0A08:02: MCFG quirk: ECAM at [mem 0x2ffff0000000-0x2fffffffffff] for [bus 00-ff] with pci_32b_read_ops Aug 13 00:25:21.180784 kernel: acpi PNP0A08:02: ECAM area [mem 0x2ffff0000000-0x2fffffffffff] reserved by PNP0C02:00 Aug 13 00:25:21.180847 kernel: acpi PNP0A08:02: ECAM at [mem 0x2ffff0000000-0x2fffffffffff] for [bus 00-ff] Aug 13 00:25:21.180858 kernel: PCI host bridge to bus 0005:00 Aug 13 00:25:21.180923 kernel: pci_bus 0005:00: root bus resource [mem 0x30000000-0x3fffffff window] Aug 13 00:25:21.180981 kernel: pci_bus 0005:00: root bus resource [mem 0x2c0000000000-0x2fffdfffffff window] Aug 13 00:25:21.181041 kernel: pci_bus 0005:00: root bus resource [bus 00-ff] Aug 13 00:25:21.181113 kernel: pci 0005:00:00.0: [1def:e110] type 00 class 0x060000 Aug 13 00:25:21.181188 kernel: pci 0005:00:01.0: [1def:e111] type 01 class 0x060400 Aug 13 00:25:21.181252 kernel: pci 0005:00:01.0: supports D1 D2 Aug 13 00:25:21.181317 kernel: pci 0005:00:01.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.181386 kernel: pci 0005:00:03.0: [1def:e113] type 01 class 0x060400 Aug 13 00:25:21.181453 kernel: pci 0005:00:03.0: supports D1 D2 Aug 13 00:25:21.181517 kernel: pci 0005:00:03.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.181590 kernel: pci 0005:00:05.0: [1def:e115] type 01 class 0x060400 Aug 13 00:25:21.181659 kernel: pci 0005:00:05.0: supports D1 D2 Aug 13 00:25:21.181724 kernel: pci 0005:00:05.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.181796 kernel: pci 0005:00:07.0: [1def:e117] type 01 class 0x060400 Aug 13 00:25:21.181861 kernel: pci 0005:00:07.0: supports D1 D2 Aug 13 00:25:21.181927 kernel: pci 0005:00:07.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.181938 kernel: acpiphp: Slot [1-2] registered Aug 13 00:25:21.181946 kernel: acpiphp: Slot [2-2] registered Aug 13 00:25:21.182018 kernel: pci 0005:03:00.0: [144d:a808] type 00 class 0x010802 Aug 13 00:25:21.182093 kernel: pci 0005:03:00.0: reg 0x10: [mem 0x30110000-0x30113fff 64bit] Aug 13 00:25:21.182159 kernel: pci 0005:03:00.0: reg 0x30: [mem 0x30100000-0x3010ffff pref] Aug 13 00:25:21.182234 kernel: pci 0005:04:00.0: [144d:a808] type 00 class 0x010802 Aug 13 00:25:21.182301 kernel: pci 0005:04:00.0: reg 0x10: [mem 0x30010000-0x30013fff 64bit] Aug 13 00:25:21.182367 kernel: pci 0005:04:00.0: reg 0x30: [mem 0x30000000-0x3000ffff pref] Aug 13 00:25:21.182426 kernel: pci_bus 0005:00: on NUMA node 0 Aug 13 00:25:21.182491 kernel: pci 0005:00:01.0: bridge window [io 0x1000-0x0fff] to [bus 01] add_size 1000 Aug 13 00:25:21.182559 kernel: pci 0005:00:01.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 01] add_size 200000 add_align 100000 Aug 13 00:25:21.182623 kernel: pci 0005:00:01.0: bridge window [mem 0x00100000-0x000fffff] to [bus 01] add_size 200000 add_align 100000 Aug 13 00:25:21.182689 kernel: pci 0005:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 02] add_size 1000 Aug 13 00:25:21.182753 kernel: pci 0005:00:03.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 02] add_size 200000 add_align 100000 Aug 13 00:25:21.182839 kernel: pci 0005:00:03.0: bridge window [mem 0x00100000-0x000fffff] to [bus 02] add_size 200000 add_align 100000 Aug 13 00:25:21.182920 kernel: pci 0005:00:05.0: bridge window [io 0x1000-0x0fff] to [bus 03] add_size 1000 Aug 13 00:25:21.182988 kernel: pci 0005:00:05.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 03] add_size 200000 add_align 100000 Aug 13 00:25:21.183057 kernel: pci 0005:00:05.0: bridge window [mem 0x00100000-0x001fffff] to [bus 03] add_size 100000 add_align 100000 Aug 13 00:25:21.183139 kernel: pci 0005:00:07.0: bridge window [io 0x1000-0x0fff] to [bus 04] add_size 1000 Aug 13 00:25:21.183204 kernel: pci 0005:00:07.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 04] add_size 200000 add_align 100000 Aug 13 00:25:21.183268 kernel: pci 0005:00:07.0: bridge window [mem 0x00100000-0x001fffff] to [bus 04] add_size 100000 add_align 100000 Aug 13 00:25:21.183333 kernel: pci 0005:00:01.0: BAR 14: assigned [mem 0x30000000-0x301fffff] Aug 13 00:25:21.183399 kernel: pci 0005:00:01.0: BAR 15: assigned [mem 0x2c0000000000-0x2c00001fffff 64bit pref] Aug 13 00:25:21.183466 kernel: pci 0005:00:03.0: BAR 14: assigned [mem 0x30200000-0x303fffff] Aug 13 00:25:21.183532 kernel: pci 0005:00:03.0: BAR 15: assigned [mem 0x2c0000200000-0x2c00003fffff 64bit pref] Aug 13 00:25:21.183597 kernel: pci 0005:00:05.0: BAR 14: assigned [mem 0x30400000-0x305fffff] Aug 13 00:25:21.183664 kernel: pci 0005:00:05.0: BAR 15: assigned [mem 0x2c0000400000-0x2c00005fffff 64bit pref] Aug 13 00:25:21.183729 kernel: pci 0005:00:07.0: BAR 14: assigned [mem 0x30600000-0x307fffff] Aug 13 00:25:21.183792 kernel: pci 0005:00:07.0: BAR 15: assigned [mem 0x2c0000600000-0x2c00007fffff 64bit pref] Aug 13 00:25:21.183856 kernel: pci 0005:00:01.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.183920 kernel: pci 0005:00:01.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.183988 kernel: pci 0005:00:03.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.184055 kernel: pci 0005:00:03.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.184121 kernel: pci 0005:00:05.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.184185 kernel: pci 0005:00:05.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.184250 kernel: pci 0005:00:07.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.184313 kernel: pci 0005:00:07.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.184377 kernel: pci 0005:00:07.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.184440 kernel: pci 0005:00:07.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.184507 kernel: pci 0005:00:05.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.184573 kernel: pci 0005:00:05.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.184637 kernel: pci 0005:00:03.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.184702 kernel: pci 0005:00:03.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.184766 kernel: pci 0005:00:01.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.184832 kernel: pci 0005:00:01.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.184895 kernel: pci 0005:00:01.0: PCI bridge to [bus 01] Aug 13 00:25:21.184961 kernel: pci 0005:00:01.0: bridge window [mem 0x30000000-0x301fffff] Aug 13 00:25:21.185024 kernel: pci 0005:00:01.0: bridge window [mem 0x2c0000000000-0x2c00001fffff 64bit pref] Aug 13 00:25:21.185095 kernel: pci 0005:00:03.0: PCI bridge to [bus 02] Aug 13 00:25:21.185159 kernel: pci 0005:00:03.0: bridge window [mem 0x30200000-0x303fffff] Aug 13 00:25:21.185224 kernel: pci 0005:00:03.0: bridge window [mem 0x2c0000200000-0x2c00003fffff 64bit pref] Aug 13 00:25:21.185291 kernel: pci 0005:03:00.0: BAR 6: assigned [mem 0x30400000-0x3040ffff pref] Aug 13 00:25:21.185361 kernel: pci 0005:03:00.0: BAR 0: assigned [mem 0x30410000-0x30413fff 64bit] Aug 13 00:25:21.185424 kernel: pci 0005:00:05.0: PCI bridge to [bus 03] Aug 13 00:25:21.185491 kernel: pci 0005:00:05.0: bridge window [mem 0x30400000-0x305fffff] Aug 13 00:25:21.185558 kernel: pci 0005:00:05.0: bridge window [mem 0x2c0000400000-0x2c00005fffff 64bit pref] Aug 13 00:25:21.185626 kernel: pci 0005:04:00.0: BAR 6: assigned [mem 0x30600000-0x3060ffff pref] Aug 13 00:25:21.185692 kernel: pci 0005:04:00.0: BAR 0: assigned [mem 0x30610000-0x30613fff 64bit] Aug 13 00:25:21.185755 kernel: pci 0005:00:07.0: PCI bridge to [bus 04] Aug 13 00:25:21.185819 kernel: pci 0005:00:07.0: bridge window [mem 0x30600000-0x307fffff] Aug 13 00:25:21.185882 kernel: pci 0005:00:07.0: bridge window [mem 0x2c0000600000-0x2c00007fffff 64bit pref] Aug 13 00:25:21.185944 kernel: pci_bus 0005:00: resource 4 [mem 0x30000000-0x3fffffff window] Aug 13 00:25:21.186001 kernel: pci_bus 0005:00: resource 5 [mem 0x2c0000000000-0x2fffdfffffff window] Aug 13 00:25:21.186075 kernel: pci_bus 0005:01: resource 1 [mem 0x30000000-0x301fffff] Aug 13 00:25:21.186135 kernel: pci_bus 0005:01: resource 2 [mem 0x2c0000000000-0x2c00001fffff 64bit pref] Aug 13 00:25:21.186211 kernel: pci_bus 0005:02: resource 1 [mem 0x30200000-0x303fffff] Aug 13 00:25:21.186270 kernel: pci_bus 0005:02: resource 2 [mem 0x2c0000200000-0x2c00003fffff 64bit pref] Aug 13 00:25:21.186343 kernel: pci_bus 0005:03: resource 1 [mem 0x30400000-0x305fffff] Aug 13 00:25:21.186402 kernel: pci_bus 0005:03: resource 2 [mem 0x2c0000400000-0x2c00005fffff 64bit pref] Aug 13 00:25:21.186470 kernel: pci_bus 0005:04: resource 1 [mem 0x30600000-0x307fffff] Aug 13 00:25:21.186530 kernel: pci_bus 0005:04: resource 2 [mem 0x2c0000600000-0x2c00007fffff 64bit pref] Aug 13 00:25:21.186540 kernel: ACPI: PCI Root Bridge [PCI5] (domain 0003 [bus 00-ff]) Aug 13 00:25:21.186611 kernel: acpi PNP0A08:03: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Aug 13 00:25:21.186677 kernel: acpi PNP0A08:03: _OSC: platform does not support [PCIeHotplug PME LTR] Aug 13 00:25:21.186740 kernel: acpi PNP0A08:03: _OSC: OS now controls [AER PCIeCapability] Aug 13 00:25:21.186802 kernel: acpi PNP0A08:03: MCFG quirk: ECAM at [mem 0x27fff0000000-0x27ffffffffff] for [bus 00-ff] with pci_32b_read_ops Aug 13 00:25:21.186864 kernel: acpi PNP0A08:03: ECAM area [mem 0x27fff0000000-0x27ffffffffff] reserved by PNP0C02:00 Aug 13 00:25:21.186926 kernel: acpi PNP0A08:03: ECAM at [mem 0x27fff0000000-0x27ffffffffff] for [bus 00-ff] Aug 13 00:25:21.186936 kernel: PCI host bridge to bus 0003:00 Aug 13 00:25:21.186999 kernel: pci_bus 0003:00: root bus resource [mem 0x10000000-0x1fffffff window] Aug 13 00:25:21.187075 kernel: pci_bus 0003:00: root bus resource [mem 0x240000000000-0x27ffdfffffff window] Aug 13 00:25:21.187134 kernel: pci_bus 0003:00: root bus resource [bus 00-ff] Aug 13 00:25:21.187211 kernel: pci 0003:00:00.0: [1def:e110] type 00 class 0x060000 Aug 13 00:25:21.187283 kernel: pci 0003:00:01.0: [1def:e111] type 01 class 0x060400 Aug 13 00:25:21.187351 kernel: pci 0003:00:01.0: supports D1 D2 Aug 13 00:25:21.187421 kernel: pci 0003:00:01.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.187499 kernel: pci 0003:00:03.0: [1def:e113] type 01 class 0x060400 Aug 13 00:25:21.187571 kernel: pci 0003:00:03.0: supports D1 D2 Aug 13 00:25:21.187639 kernel: pci 0003:00:03.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.187713 kernel: pci 0003:00:05.0: [1def:e115] type 01 class 0x060400 Aug 13 00:25:21.187778 kernel: pci 0003:00:05.0: supports D1 D2 Aug 13 00:25:21.187843 kernel: pci 0003:00:05.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.187853 kernel: acpiphp: Slot [1-3] registered Aug 13 00:25:21.187862 kernel: acpiphp: Slot [2-3] registered Aug 13 00:25:21.187935 kernel: pci 0003:03:00.0: [8086:1521] type 00 class 0x020000 Aug 13 00:25:21.188003 kernel: pci 0003:03:00.0: reg 0x10: [mem 0x10020000-0x1003ffff] Aug 13 00:25:21.188074 kernel: pci 0003:03:00.0: reg 0x18: [io 0x0020-0x003f] Aug 13 00:25:21.188141 kernel: pci 0003:03:00.0: reg 0x1c: [mem 0x10044000-0x10047fff] Aug 13 00:25:21.188207 kernel: pci 0003:03:00.0: PME# supported from D0 D3hot D3cold Aug 13 00:25:21.188273 kernel: pci 0003:03:00.0: reg 0x184: [mem 0x240000060000-0x240000063fff 64bit pref] Aug 13 00:25:21.188340 kernel: pci 0003:03:00.0: VF(n) BAR0 space: [mem 0x240000060000-0x24000007ffff 64bit pref] (contains BAR0 for 8 VFs) Aug 13 00:25:21.188411 kernel: pci 0003:03:00.0: reg 0x190: [mem 0x240000040000-0x240000043fff 64bit pref] Aug 13 00:25:21.188477 kernel: pci 0003:03:00.0: VF(n) BAR3 space: [mem 0x240000040000-0x24000005ffff 64bit pref] (contains BAR3 for 8 VFs) Aug 13 00:25:21.188543 kernel: pci 0003:03:00.0: 8.000 Gb/s available PCIe bandwidth, limited by 5.0 GT/s PCIe x2 link at 0003:00:05.0 (capable of 16.000 Gb/s with 5.0 GT/s PCIe x4 link) Aug 13 00:25:21.188621 kernel: pci 0003:03:00.1: [8086:1521] type 00 class 0x020000 Aug 13 00:25:21.188687 kernel: pci 0003:03:00.1: reg 0x10: [mem 0x10000000-0x1001ffff] Aug 13 00:25:21.188755 kernel: pci 0003:03:00.1: reg 0x18: [io 0x0000-0x001f] Aug 13 00:25:21.188821 kernel: pci 0003:03:00.1: reg 0x1c: [mem 0x10040000-0x10043fff] Aug 13 00:25:21.188888 kernel: pci 0003:03:00.1: PME# supported from D0 D3hot D3cold Aug 13 00:25:21.188955 kernel: pci 0003:03:00.1: reg 0x184: [mem 0x240000020000-0x240000023fff 64bit pref] Aug 13 00:25:21.189021 kernel: pci 0003:03:00.1: VF(n) BAR0 space: [mem 0x240000020000-0x24000003ffff 64bit pref] (contains BAR0 for 8 VFs) Aug 13 00:25:21.189183 kernel: pci 0003:03:00.1: reg 0x190: [mem 0x240000000000-0x240000003fff 64bit pref] Aug 13 00:25:21.189250 kernel: pci 0003:03:00.1: VF(n) BAR3 space: [mem 0x240000000000-0x24000001ffff 64bit pref] (contains BAR3 for 8 VFs) Aug 13 00:25:21.189309 kernel: pci_bus 0003:00: on NUMA node 0 Aug 13 00:25:21.189374 kernel: pci 0003:00:01.0: bridge window [io 0x1000-0x0fff] to [bus 01] add_size 1000 Aug 13 00:25:21.189442 kernel: pci 0003:00:01.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 01] add_size 200000 add_align 100000 Aug 13 00:25:21.189504 kernel: pci 0003:00:01.0: bridge window [mem 0x00100000-0x000fffff] to [bus 01] add_size 200000 add_align 100000 Aug 13 00:25:21.189569 kernel: pci 0003:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 02] add_size 1000 Aug 13 00:25:21.189632 kernel: pci 0003:00:03.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 02] add_size 200000 add_align 100000 Aug 13 00:25:21.189696 kernel: pci 0003:00:03.0: bridge window [mem 0x00100000-0x000fffff] to [bus 02] add_size 200000 add_align 100000 Aug 13 00:25:21.189760 kernel: pci 0003:00:05.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 03-04] add_size 300000 add_align 100000 Aug 13 00:25:21.189823 kernel: pci 0003:00:05.0: bridge window [mem 0x00100000-0x001fffff] to [bus 03-04] add_size 100000 add_align 100000 Aug 13 00:25:21.189889 kernel: pci 0003:00:01.0: BAR 14: assigned [mem 0x10000000-0x101fffff] Aug 13 00:25:21.189952 kernel: pci 0003:00:01.0: BAR 15: assigned [mem 0x240000000000-0x2400001fffff 64bit pref] Aug 13 00:25:21.190015 kernel: pci 0003:00:03.0: BAR 14: assigned [mem 0x10200000-0x103fffff] Aug 13 00:25:21.190081 kernel: pci 0003:00:03.0: BAR 15: assigned [mem 0x240000200000-0x2400003fffff 64bit pref] Aug 13 00:25:21.190144 kernel: pci 0003:00:05.0: BAR 14: assigned [mem 0x10400000-0x105fffff] Aug 13 00:25:21.190220 kernel: pci 0003:00:05.0: BAR 15: assigned [mem 0x240000400000-0x2400006fffff 64bit pref] Aug 13 00:25:21.190284 kernel: pci 0003:00:01.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.190346 kernel: pci 0003:00:01.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.190412 kernel: pci 0003:00:03.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.190475 kernel: pci 0003:00:03.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.190539 kernel: pci 0003:00:05.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.190601 kernel: pci 0003:00:05.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.190665 kernel: pci 0003:00:05.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.190727 kernel: pci 0003:00:05.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.190789 kernel: pci 0003:00:03.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.190853 kernel: pci 0003:00:03.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.190918 kernel: pci 0003:00:01.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.190983 kernel: pci 0003:00:01.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.191049 kernel: pci 0003:00:01.0: PCI bridge to [bus 01] Aug 13 00:25:21.191113 kernel: pci 0003:00:01.0: bridge window [mem 0x10000000-0x101fffff] Aug 13 00:25:21.191177 kernel: pci 0003:00:01.0: bridge window [mem 0x240000000000-0x2400001fffff 64bit pref] Aug 13 00:25:21.191240 kernel: pci 0003:00:03.0: PCI bridge to [bus 02] Aug 13 00:25:21.191302 kernel: pci 0003:00:03.0: bridge window [mem 0x10200000-0x103fffff] Aug 13 00:25:21.191368 kernel: pci 0003:00:03.0: bridge window [mem 0x240000200000-0x2400003fffff 64bit pref] Aug 13 00:25:21.191434 kernel: pci 0003:03:00.0: BAR 0: assigned [mem 0x10400000-0x1041ffff] Aug 13 00:25:21.191500 kernel: pci 0003:03:00.1: BAR 0: assigned [mem 0x10420000-0x1043ffff] Aug 13 00:25:21.191567 kernel: pci 0003:03:00.0: BAR 3: assigned [mem 0x10440000-0x10443fff] Aug 13 00:25:21.191633 kernel: pci 0003:03:00.0: BAR 7: assigned [mem 0x240000400000-0x24000041ffff 64bit pref] Aug 13 00:25:21.191699 kernel: pci 0003:03:00.0: BAR 10: assigned [mem 0x240000420000-0x24000043ffff 64bit pref] Aug 13 00:25:21.191766 kernel: pci 0003:03:00.1: BAR 3: assigned [mem 0x10444000-0x10447fff] Aug 13 00:25:21.191832 kernel: pci 0003:03:00.1: BAR 7: assigned [mem 0x240000440000-0x24000045ffff 64bit pref] Aug 13 00:25:21.191896 kernel: pci 0003:03:00.1: BAR 10: assigned [mem 0x240000460000-0x24000047ffff 64bit pref] Aug 13 00:25:21.191962 kernel: pci 0003:03:00.0: BAR 2: no space for [io size 0x0020] Aug 13 00:25:21.192027 kernel: pci 0003:03:00.0: BAR 2: failed to assign [io size 0x0020] Aug 13 00:25:21.192096 kernel: pci 0003:03:00.1: BAR 2: no space for [io size 0x0020] Aug 13 00:25:21.192162 kernel: pci 0003:03:00.1: BAR 2: failed to assign [io size 0x0020] Aug 13 00:25:21.192228 kernel: pci 0003:03:00.0: BAR 2: no space for [io size 0x0020] Aug 13 00:25:21.192293 kernel: pci 0003:03:00.0: BAR 2: failed to assign [io size 0x0020] Aug 13 00:25:21.192358 kernel: pci 0003:03:00.1: BAR 2: no space for [io size 0x0020] Aug 13 00:25:21.192423 kernel: pci 0003:03:00.1: BAR 2: failed to assign [io size 0x0020] Aug 13 00:25:21.192487 kernel: pci 0003:00:05.0: PCI bridge to [bus 03-04] Aug 13 00:25:21.192551 kernel: pci 0003:00:05.0: bridge window [mem 0x10400000-0x105fffff] Aug 13 00:25:21.192613 kernel: pci 0003:00:05.0: bridge window [mem 0x240000400000-0x2400006fffff 64bit pref] Aug 13 00:25:21.192672 kernel: pci_bus 0003:00: Some PCI device resources are unassigned, try booting with pci=realloc Aug 13 00:25:21.192731 kernel: pci_bus 0003:00: resource 4 [mem 0x10000000-0x1fffffff window] Aug 13 00:25:21.192787 kernel: pci_bus 0003:00: resource 5 [mem 0x240000000000-0x27ffdfffffff window] Aug 13 00:25:21.192864 kernel: pci_bus 0003:01: resource 1 [mem 0x10000000-0x101fffff] Aug 13 00:25:21.192924 kernel: pci_bus 0003:01: resource 2 [mem 0x240000000000-0x2400001fffff 64bit pref] Aug 13 00:25:21.192990 kernel: pci_bus 0003:02: resource 1 [mem 0x10200000-0x103fffff] Aug 13 00:25:21.193053 kernel: pci_bus 0003:02: resource 2 [mem 0x240000200000-0x2400003fffff 64bit pref] Aug 13 00:25:21.193122 kernel: pci_bus 0003:03: resource 1 [mem 0x10400000-0x105fffff] Aug 13 00:25:21.193180 kernel: pci_bus 0003:03: resource 2 [mem 0x240000400000-0x2400006fffff 64bit pref] Aug 13 00:25:21.193191 kernel: ACPI: PCI Root Bridge [PCI0] (domain 000c [bus 00-ff]) Aug 13 00:25:21.193261 kernel: acpi PNP0A08:04: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Aug 13 00:25:21.193323 kernel: acpi PNP0A08:04: _OSC: platform does not support [PCIeHotplug PME LTR] Aug 13 00:25:21.193386 kernel: acpi PNP0A08:04: _OSC: OS now controls [AER PCIeCapability] Aug 13 00:25:21.193448 kernel: acpi PNP0A08:04: MCFG quirk: ECAM at [mem 0x33fff0000000-0x33ffffffffff] for [bus 00-ff] with pci_32b_read_ops Aug 13 00:25:21.193514 kernel: acpi PNP0A08:04: ECAM area [mem 0x33fff0000000-0x33ffffffffff] reserved by PNP0C02:00 Aug 13 00:25:21.193575 kernel: acpi PNP0A08:04: ECAM at [mem 0x33fff0000000-0x33ffffffffff] for [bus 00-ff] Aug 13 00:25:21.193585 kernel: PCI host bridge to bus 000c:00 Aug 13 00:25:21.193652 kernel: pci_bus 000c:00: root bus resource [mem 0x40000000-0x4fffffff window] Aug 13 00:25:21.193709 kernel: pci_bus 000c:00: root bus resource [mem 0x300000000000-0x33ffdfffffff window] Aug 13 00:25:21.193767 kernel: pci_bus 000c:00: root bus resource [bus 00-ff] Aug 13 00:25:21.193840 kernel: pci 000c:00:00.0: [1def:e100] type 00 class 0x060000 Aug 13 00:25:21.193917 kernel: pci 000c:00:01.0: [1def:e101] type 01 class 0x060400 Aug 13 00:25:21.193984 kernel: pci 000c:00:01.0: enabling Extended Tags Aug 13 00:25:21.194052 kernel: pci 000c:00:01.0: supports D1 D2 Aug 13 00:25:21.194119 kernel: pci 000c:00:01.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.194190 kernel: pci 000c:00:02.0: [1def:e102] type 01 class 0x060400 Aug 13 00:25:21.194256 kernel: pci 000c:00:02.0: supports D1 D2 Aug 13 00:25:21.194320 kernel: pci 000c:00:02.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.194395 kernel: pci 000c:00:03.0: [1def:e103] type 01 class 0x060400 Aug 13 00:25:21.194462 kernel: pci 000c:00:03.0: supports D1 D2 Aug 13 00:25:21.194530 kernel: pci 000c:00:03.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.194601 kernel: pci 000c:00:04.0: [1def:e104] type 01 class 0x060400 Aug 13 00:25:21.194667 kernel: pci 000c:00:04.0: supports D1 D2 Aug 13 00:25:21.194731 kernel: pci 000c:00:04.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.194741 kernel: acpiphp: Slot [1-4] registered Aug 13 00:25:21.194753 kernel: acpiphp: Slot [2-4] registered Aug 13 00:25:21.194761 kernel: acpiphp: Slot [3-2] registered Aug 13 00:25:21.194769 kernel: acpiphp: Slot [4-2] registered Aug 13 00:25:21.194829 kernel: pci_bus 000c:00: on NUMA node 0 Aug 13 00:25:21.194893 kernel: pci 000c:00:01.0: bridge window [io 0x1000-0x0fff] to [bus 01] add_size 1000 Aug 13 00:25:21.194958 kernel: pci 000c:00:01.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 01] add_size 200000 add_align 100000 Aug 13 00:25:21.195022 kernel: pci 000c:00:01.0: bridge window [mem 0x00100000-0x000fffff] to [bus 01] add_size 200000 add_align 100000 Aug 13 00:25:21.195277 kernel: pci 000c:00:02.0: bridge window [io 0x1000-0x0fff] to [bus 02] add_size 1000 Aug 13 00:25:21.195348 kernel: pci 000c:00:02.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 02] add_size 200000 add_align 100000 Aug 13 00:25:21.195411 kernel: pci 000c:00:02.0: bridge window [mem 0x00100000-0x000fffff] to [bus 02] add_size 200000 add_align 100000 Aug 13 00:25:21.195475 kernel: pci 000c:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 03] add_size 1000 Aug 13 00:25:21.195539 kernel: pci 000c:00:03.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 03] add_size 200000 add_align 100000 Aug 13 00:25:21.195602 kernel: pci 000c:00:03.0: bridge window [mem 0x00100000-0x000fffff] to [bus 03] add_size 200000 add_align 100000 Aug 13 00:25:21.195666 kernel: pci 000c:00:04.0: bridge window [io 0x1000-0x0fff] to [bus 04] add_size 1000 Aug 13 00:25:21.195728 kernel: pci 000c:00:04.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 04] add_size 200000 add_align 100000 Aug 13 00:25:21.195794 kernel: pci 000c:00:04.0: bridge window [mem 0x00100000-0x000fffff] to [bus 04] add_size 200000 add_align 100000 Aug 13 00:25:21.195858 kernel: pci 000c:00:01.0: BAR 14: assigned [mem 0x40000000-0x401fffff] Aug 13 00:25:21.195922 kernel: pci 000c:00:01.0: BAR 15: assigned [mem 0x300000000000-0x3000001fffff 64bit pref] Aug 13 00:25:21.195984 kernel: pci 000c:00:02.0: BAR 14: assigned [mem 0x40200000-0x403fffff] Aug 13 00:25:21.196052 kernel: pci 000c:00:02.0: BAR 15: assigned [mem 0x300000200000-0x3000003fffff 64bit pref] Aug 13 00:25:21.196116 kernel: pci 000c:00:03.0: BAR 14: assigned [mem 0x40400000-0x405fffff] Aug 13 00:25:21.196180 kernel: pci 000c:00:03.0: BAR 15: assigned [mem 0x300000400000-0x3000005fffff 64bit pref] Aug 13 00:25:21.196247 kernel: pci 000c:00:04.0: BAR 14: assigned [mem 0x40600000-0x407fffff] Aug 13 00:25:21.196309 kernel: pci 000c:00:04.0: BAR 15: assigned [mem 0x300000600000-0x3000007fffff 64bit pref] Aug 13 00:25:21.196373 kernel: pci 000c:00:01.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.196436 kernel: pci 000c:00:01.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.196499 kernel: pci 000c:00:02.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.196562 kernel: pci 000c:00:02.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.196625 kernel: pci 000c:00:03.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.196688 kernel: pci 000c:00:03.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.196753 kernel: pci 000c:00:04.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.196816 kernel: pci 000c:00:04.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.196878 kernel: pci 000c:00:04.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.196940 kernel: pci 000c:00:04.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.197003 kernel: pci 000c:00:03.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.197070 kernel: pci 000c:00:03.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.197133 kernel: pci 000c:00:02.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.197196 kernel: pci 000c:00:02.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.197259 kernel: pci 000c:00:01.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.197325 kernel: pci 000c:00:01.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.197387 kernel: pci 000c:00:01.0: PCI bridge to [bus 01] Aug 13 00:25:21.197451 kernel: pci 000c:00:01.0: bridge window [mem 0x40000000-0x401fffff] Aug 13 00:25:21.197514 kernel: pci 000c:00:01.0: bridge window [mem 0x300000000000-0x3000001fffff 64bit pref] Aug 13 00:25:21.197578 kernel: pci 000c:00:02.0: PCI bridge to [bus 02] Aug 13 00:25:21.197641 kernel: pci 000c:00:02.0: bridge window [mem 0x40200000-0x403fffff] Aug 13 00:25:21.197706 kernel: pci 000c:00:02.0: bridge window [mem 0x300000200000-0x3000003fffff 64bit pref] Aug 13 00:25:21.197771 kernel: pci 000c:00:03.0: PCI bridge to [bus 03] Aug 13 00:25:21.197834 kernel: pci 000c:00:03.0: bridge window [mem 0x40400000-0x405fffff] Aug 13 00:25:21.197898 kernel: pci 000c:00:03.0: bridge window [mem 0x300000400000-0x3000005fffff 64bit pref] Aug 13 00:25:21.197961 kernel: pci 000c:00:04.0: PCI bridge to [bus 04] Aug 13 00:25:21.198024 kernel: pci 000c:00:04.0: bridge window [mem 0x40600000-0x407fffff] Aug 13 00:25:21.198093 kernel: pci 000c:00:04.0: bridge window [mem 0x300000600000-0x3000007fffff 64bit pref] Aug 13 00:25:21.198155 kernel: pci_bus 000c:00: resource 4 [mem 0x40000000-0x4fffffff window] Aug 13 00:25:21.198211 kernel: pci_bus 000c:00: resource 5 [mem 0x300000000000-0x33ffdfffffff window] Aug 13 00:25:21.198279 kernel: pci_bus 000c:01: resource 1 [mem 0x40000000-0x401fffff] Aug 13 00:25:21.198338 kernel: pci_bus 000c:01: resource 2 [mem 0x300000000000-0x3000001fffff 64bit pref] Aug 13 00:25:21.198412 kernel: pci_bus 000c:02: resource 1 [mem 0x40200000-0x403fffff] Aug 13 00:25:21.198471 kernel: pci_bus 000c:02: resource 2 [mem 0x300000200000-0x3000003fffff 64bit pref] Aug 13 00:25:21.198541 kernel: pci_bus 000c:03: resource 1 [mem 0x40400000-0x405fffff] Aug 13 00:25:21.198599 kernel: pci_bus 000c:03: resource 2 [mem 0x300000400000-0x3000005fffff 64bit pref] Aug 13 00:25:21.198666 kernel: pci_bus 000c:04: resource 1 [mem 0x40600000-0x407fffff] Aug 13 00:25:21.198724 kernel: pci_bus 000c:04: resource 2 [mem 0x300000600000-0x3000007fffff 64bit pref] Aug 13 00:25:21.198734 kernel: ACPI: PCI Root Bridge [PCI4] (domain 0002 [bus 00-ff]) Aug 13 00:25:21.198804 kernel: acpi PNP0A08:05: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Aug 13 00:25:21.198866 kernel: acpi PNP0A08:05: _OSC: platform does not support [PCIeHotplug PME LTR] Aug 13 00:25:21.198930 kernel: acpi PNP0A08:05: _OSC: OS now controls [AER PCIeCapability] Aug 13 00:25:21.198990 kernel: acpi PNP0A08:05: MCFG quirk: ECAM at [mem 0x23fff0000000-0x23ffffffffff] for [bus 00-ff] with pci_32b_read_ops Aug 13 00:25:21.199055 kernel: acpi PNP0A08:05: ECAM area [mem 0x23fff0000000-0x23ffffffffff] reserved by PNP0C02:00 Aug 13 00:25:21.199116 kernel: acpi PNP0A08:05: ECAM at [mem 0x23fff0000000-0x23ffffffffff] for [bus 00-ff] Aug 13 00:25:21.199127 kernel: PCI host bridge to bus 0002:00 Aug 13 00:25:21.199189 kernel: pci_bus 0002:00: root bus resource [mem 0x00800000-0x0fffffff window] Aug 13 00:25:21.199246 kernel: pci_bus 0002:00: root bus resource [mem 0x200000000000-0x23ffdfffffff window] Aug 13 00:25:21.199304 kernel: pci_bus 0002:00: root bus resource [bus 00-ff] Aug 13 00:25:21.199376 kernel: pci 0002:00:00.0: [1def:e110] type 00 class 0x060000 Aug 13 00:25:21.199447 kernel: pci 0002:00:01.0: [1def:e111] type 01 class 0x060400 Aug 13 00:25:21.199511 kernel: pci 0002:00:01.0: supports D1 D2 Aug 13 00:25:21.199574 kernel: pci 0002:00:01.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.199645 kernel: pci 0002:00:03.0: [1def:e113] type 01 class 0x060400 Aug 13 00:25:21.199713 kernel: pci 0002:00:03.0: supports D1 D2 Aug 13 00:25:21.199776 kernel: pci 0002:00:03.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.199844 kernel: pci 0002:00:05.0: [1def:e115] type 01 class 0x060400 Aug 13 00:25:21.199909 kernel: pci 0002:00:05.0: supports D1 D2 Aug 13 00:25:21.199971 kernel: pci 0002:00:05.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.200045 kernel: pci 0002:00:07.0: [1def:e117] type 01 class 0x060400 Aug 13 00:25:21.200109 kernel: pci 0002:00:07.0: supports D1 D2 Aug 13 00:25:21.200175 kernel: pci 0002:00:07.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.200185 kernel: acpiphp: Slot [1-5] registered Aug 13 00:25:21.200194 kernel: acpiphp: Slot [2-5] registered Aug 13 00:25:21.200202 kernel: acpiphp: Slot [3-3] registered Aug 13 00:25:21.200210 kernel: acpiphp: Slot [4-3] registered Aug 13 00:25:21.200264 kernel: pci_bus 0002:00: on NUMA node 0 Aug 13 00:25:21.200327 kernel: pci 0002:00:01.0: bridge window [io 0x1000-0x0fff] to [bus 01] add_size 1000 Aug 13 00:25:21.200391 kernel: pci 0002:00:01.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 01] add_size 200000 add_align 100000 Aug 13 00:25:21.200455 kernel: pci 0002:00:01.0: bridge window [mem 0x00100000-0x000fffff] to [bus 01] add_size 200000 add_align 100000 Aug 13 00:25:21.200521 kernel: pci 0002:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 02] add_size 1000 Aug 13 00:25:21.200584 kernel: pci 0002:00:03.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 02] add_size 200000 add_align 100000 Aug 13 00:25:21.200647 kernel: pci 0002:00:03.0: bridge window [mem 0x00100000-0x000fffff] to [bus 02] add_size 200000 add_align 100000 Aug 13 00:25:21.200711 kernel: pci 0002:00:05.0: bridge window [io 0x1000-0x0fff] to [bus 03] add_size 1000 Aug 13 00:25:21.200778 kernel: pci 0002:00:05.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 03] add_size 200000 add_align 100000 Aug 13 00:25:21.200843 kernel: pci 0002:00:05.0: bridge window [mem 0x00100000-0x000fffff] to [bus 03] add_size 200000 add_align 100000 Aug 13 00:25:21.200908 kernel: pci 0002:00:07.0: bridge window [io 0x1000-0x0fff] to [bus 04] add_size 1000 Aug 13 00:25:21.200972 kernel: pci 0002:00:07.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 04] add_size 200000 add_align 100000 Aug 13 00:25:21.201035 kernel: pci 0002:00:07.0: bridge window [mem 0x00100000-0x000fffff] to [bus 04] add_size 200000 add_align 100000 Aug 13 00:25:21.201102 kernel: pci 0002:00:01.0: BAR 14: assigned [mem 0x00800000-0x009fffff] Aug 13 00:25:21.201166 kernel: pci 0002:00:01.0: BAR 15: assigned [mem 0x200000000000-0x2000001fffff 64bit pref] Aug 13 00:25:21.201232 kernel: pci 0002:00:03.0: BAR 14: assigned [mem 0x00a00000-0x00bfffff] Aug 13 00:25:21.201296 kernel: pci 0002:00:03.0: BAR 15: assigned [mem 0x200000200000-0x2000003fffff 64bit pref] Aug 13 00:25:21.201360 kernel: pci 0002:00:05.0: BAR 14: assigned [mem 0x00c00000-0x00dfffff] Aug 13 00:25:21.201423 kernel: pci 0002:00:05.0: BAR 15: assigned [mem 0x200000400000-0x2000005fffff 64bit pref] Aug 13 00:25:21.201487 kernel: pci 0002:00:07.0: BAR 14: assigned [mem 0x00e00000-0x00ffffff] Aug 13 00:25:21.201549 kernel: pci 0002:00:07.0: BAR 15: assigned [mem 0x200000600000-0x2000007fffff 64bit pref] Aug 13 00:25:21.201613 kernel: pci 0002:00:01.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.201678 kernel: pci 0002:00:01.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.201742 kernel: pci 0002:00:03.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.201805 kernel: pci 0002:00:03.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.201868 kernel: pci 0002:00:05.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.201931 kernel: pci 0002:00:05.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.201998 kernel: pci 0002:00:07.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.202068 kernel: pci 0002:00:07.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.202132 kernel: pci 0002:00:07.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.202195 kernel: pci 0002:00:07.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.202265 kernel: pci 0002:00:05.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.202329 kernel: pci 0002:00:05.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.202394 kernel: pci 0002:00:03.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.202461 kernel: pci 0002:00:03.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.202525 kernel: pci 0002:00:01.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.202590 kernel: pci 0002:00:01.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.202655 kernel: pci 0002:00:01.0: PCI bridge to [bus 01] Aug 13 00:25:21.202719 kernel: pci 0002:00:01.0: bridge window [mem 0x00800000-0x009fffff] Aug 13 00:25:21.202784 kernel: pci 0002:00:01.0: bridge window [mem 0x200000000000-0x2000001fffff 64bit pref] Aug 13 00:25:21.202848 kernel: pci 0002:00:03.0: PCI bridge to [bus 02] Aug 13 00:25:21.202911 kernel: pci 0002:00:03.0: bridge window [mem 0x00a00000-0x00bfffff] Aug 13 00:25:21.202976 kernel: pci 0002:00:03.0: bridge window [mem 0x200000200000-0x2000003fffff 64bit pref] Aug 13 00:25:21.203137 kernel: pci 0002:00:05.0: PCI bridge to [bus 03] Aug 13 00:25:21.203218 kernel: pci 0002:00:05.0: bridge window [mem 0x00c00000-0x00dfffff] Aug 13 00:25:21.203288 kernel: pci 0002:00:05.0: bridge window [mem 0x200000400000-0x2000005fffff 64bit pref] Aug 13 00:25:21.203353 kernel: pci 0002:00:07.0: PCI bridge to [bus 04] Aug 13 00:25:21.203415 kernel: pci 0002:00:07.0: bridge window [mem 0x00e00000-0x00ffffff] Aug 13 00:25:21.203478 kernel: pci 0002:00:07.0: bridge window [mem 0x200000600000-0x2000007fffff 64bit pref] Aug 13 00:25:21.203538 kernel: pci_bus 0002:00: resource 4 [mem 0x00800000-0x0fffffff window] Aug 13 00:25:21.203595 kernel: pci_bus 0002:00: resource 5 [mem 0x200000000000-0x23ffdfffffff window] Aug 13 00:25:21.203663 kernel: pci_bus 0002:01: resource 1 [mem 0x00800000-0x009fffff] Aug 13 00:25:21.203772 kernel: pci_bus 0002:01: resource 2 [mem 0x200000000000-0x2000001fffff 64bit pref] Aug 13 00:25:21.203844 kernel: pci_bus 0002:02: resource 1 [mem 0x00a00000-0x00bfffff] Aug 13 00:25:21.203904 kernel: pci_bus 0002:02: resource 2 [mem 0x200000200000-0x2000003fffff 64bit pref] Aug 13 00:25:21.203978 kernel: pci_bus 0002:03: resource 1 [mem 0x00c00000-0x00dfffff] Aug 13 00:25:21.204044 kernel: pci_bus 0002:03: resource 2 [mem 0x200000400000-0x2000005fffff 64bit pref] Aug 13 00:25:21.204110 kernel: pci_bus 0002:04: resource 1 [mem 0x00e00000-0x00ffffff] Aug 13 00:25:21.204172 kernel: pci_bus 0002:04: resource 2 [mem 0x200000600000-0x2000007fffff 64bit pref] Aug 13 00:25:21.204183 kernel: ACPI: PCI Root Bridge [PCI2] (domain 0001 [bus 00-ff]) Aug 13 00:25:21.204253 kernel: acpi PNP0A08:06: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Aug 13 00:25:21.204315 kernel: acpi PNP0A08:06: _OSC: platform does not support [PCIeHotplug PME LTR] Aug 13 00:25:21.204377 kernel: acpi PNP0A08:06: _OSC: OS now controls [AER PCIeCapability] Aug 13 00:25:21.204437 kernel: acpi PNP0A08:06: MCFG quirk: ECAM at [mem 0x3bfff0000000-0x3bffffffffff] for [bus 00-ff] with pci_32b_read_ops Aug 13 00:25:21.204500 kernel: acpi PNP0A08:06: ECAM area [mem 0x3bfff0000000-0x3bffffffffff] reserved by PNP0C02:00 Aug 13 00:25:21.204563 kernel: acpi PNP0A08:06: ECAM at [mem 0x3bfff0000000-0x3bffffffffff] for [bus 00-ff] Aug 13 00:25:21.204574 kernel: PCI host bridge to bus 0001:00 Aug 13 00:25:21.204636 kernel: pci_bus 0001:00: root bus resource [mem 0x60000000-0x6fffffff window] Aug 13 00:25:21.204694 kernel: pci_bus 0001:00: root bus resource [mem 0x380000000000-0x3bffdfffffff window] Aug 13 00:25:21.204749 kernel: pci_bus 0001:00: root bus resource [bus 00-ff] Aug 13 00:25:21.204820 kernel: pci 0001:00:00.0: [1def:e100] type 00 class 0x060000 Aug 13 00:25:21.204894 kernel: pci 0001:00:01.0: [1def:e101] type 01 class 0x060400 Aug 13 00:25:21.204959 kernel: pci 0001:00:01.0: enabling Extended Tags Aug 13 00:25:21.205022 kernel: pci 0001:00:01.0: supports D1 D2 Aug 13 00:25:21.205090 kernel: pci 0001:00:01.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.205160 kernel: pci 0001:00:02.0: [1def:e102] type 01 class 0x060400 Aug 13 00:25:21.205225 kernel: pci 0001:00:02.0: supports D1 D2 Aug 13 00:25:21.205291 kernel: pci 0001:00:02.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.205362 kernel: pci 0001:00:03.0: [1def:e103] type 01 class 0x060400 Aug 13 00:25:21.205426 kernel: pci 0001:00:03.0: supports D1 D2 Aug 13 00:25:21.205489 kernel: pci 0001:00:03.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.205559 kernel: pci 0001:00:04.0: [1def:e104] type 01 class 0x060400 Aug 13 00:25:21.205624 kernel: pci 0001:00:04.0: supports D1 D2 Aug 13 00:25:21.205687 kernel: pci 0001:00:04.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.205699 kernel: acpiphp: Slot [1-6] registered Aug 13 00:25:21.205773 kernel: pci 0001:01:00.0: [15b3:1015] type 00 class 0x020000 Aug 13 00:25:21.205840 kernel: pci 0001:01:00.0: reg 0x10: [mem 0x380002000000-0x380003ffffff 64bit pref] Aug 13 00:25:21.205905 kernel: pci 0001:01:00.0: reg 0x30: [mem 0x60100000-0x601fffff pref] Aug 13 00:25:21.205971 kernel: pci 0001:01:00.0: PME# supported from D3cold Aug 13 00:25:21.206035 kernel: pci 0001:01:00.0: reg 0x1a4: [mem 0x380004800000-0x3800048fffff 64bit pref] Aug 13 00:25:21.206112 kernel: pci 0001:01:00.0: VF(n) BAR0 space: [mem 0x380004800000-0x380004ffffff 64bit pref] (contains BAR0 for 8 VFs) Aug 13 00:25:21.206179 kernel: pci 0001:01:00.0: 31.504 Gb/s available PCIe bandwidth, limited by 8.0 GT/s PCIe x4 link at 0001:00:01.0 (capable of 63.008 Gb/s with 8.0 GT/s PCIe x8 link) Aug 13 00:25:21.206258 kernel: pci 0001:01:00.1: [15b3:1015] type 00 class 0x020000 Aug 13 00:25:21.206326 kernel: pci 0001:01:00.1: reg 0x10: [mem 0x380000000000-0x380001ffffff 64bit pref] Aug 13 00:25:21.206391 kernel: pci 0001:01:00.1: reg 0x30: [mem 0x60000000-0x600fffff pref] Aug 13 00:25:21.206457 kernel: pci 0001:01:00.1: PME# supported from D3cold Aug 13 00:25:21.206521 kernel: pci 0001:01:00.1: reg 0x1a4: [mem 0x380004000000-0x3800040fffff 64bit pref] Aug 13 00:25:21.206586 kernel: pci 0001:01:00.1: VF(n) BAR0 space: [mem 0x380004000000-0x3800047fffff 64bit pref] (contains BAR0 for 8 VFs) Aug 13 00:25:21.206599 kernel: acpiphp: Slot [2-6] registered Aug 13 00:25:21.206607 kernel: acpiphp: Slot [3-4] registered Aug 13 00:25:21.206616 kernel: acpiphp: Slot [4-4] registered Aug 13 00:25:21.206672 kernel: pci_bus 0001:00: on NUMA node 0 Aug 13 00:25:21.206736 kernel: pci 0001:00:01.0: bridge window [io 0x1000-0x0fff] to [bus 01] add_size 1000 Aug 13 00:25:21.206815 kernel: pci 0001:00:01.0: bridge window [mem 0x02000000-0x05ffffff 64bit pref] to [bus 01] add_size 2000000 add_align 2000000 Aug 13 00:25:21.206881 kernel: pci 0001:00:02.0: bridge window [io 0x1000-0x0fff] to [bus 02] add_size 1000 Aug 13 00:25:21.206944 kernel: pci 0001:00:02.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 02] add_size 200000 add_align 100000 Aug 13 00:25:21.207011 kernel: pci 0001:00:02.0: bridge window [mem 0x00100000-0x000fffff] to [bus 02] add_size 200000 add_align 100000 Aug 13 00:25:21.207200 kernel: pci 0001:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 03] add_size 1000 Aug 13 00:25:21.207269 kernel: pci 0001:00:03.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 03] add_size 200000 add_align 100000 Aug 13 00:25:21.207332 kernel: pci 0001:00:03.0: bridge window [mem 0x00100000-0x000fffff] to [bus 03] add_size 200000 add_align 100000 Aug 13 00:25:21.207397 kernel: pci 0001:00:04.0: bridge window [io 0x1000-0x0fff] to [bus 04] add_size 1000 Aug 13 00:25:21.207460 kernel: pci 0001:00:04.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 04] add_size 200000 add_align 100000 Aug 13 00:25:21.207523 kernel: pci 0001:00:04.0: bridge window [mem 0x00100000-0x000fffff] to [bus 04] add_size 200000 add_align 100000 Aug 13 00:25:21.207591 kernel: pci 0001:00:01.0: BAR 15: assigned [mem 0x380000000000-0x380005ffffff 64bit pref] Aug 13 00:25:21.207655 kernel: pci 0001:00:01.0: BAR 14: assigned [mem 0x60000000-0x601fffff] Aug 13 00:25:21.207720 kernel: pci 0001:00:02.0: BAR 14: assigned [mem 0x60200000-0x603fffff] Aug 13 00:25:21.207783 kernel: pci 0001:00:02.0: BAR 15: assigned [mem 0x380006000000-0x3800061fffff 64bit pref] Aug 13 00:25:21.207849 kernel: pci 0001:00:03.0: BAR 14: assigned [mem 0x60400000-0x605fffff] Aug 13 00:25:21.207912 kernel: pci 0001:00:03.0: BAR 15: assigned [mem 0x380006200000-0x3800063fffff 64bit pref] Aug 13 00:25:21.207975 kernel: pci 0001:00:04.0: BAR 14: assigned [mem 0x60600000-0x607fffff] Aug 13 00:25:21.208042 kernel: pci 0001:00:04.0: BAR 15: assigned [mem 0x380006400000-0x3800065fffff 64bit pref] Aug 13 00:25:21.208106 kernel: pci 0001:00:01.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.208169 kernel: pci 0001:00:01.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.208233 kernel: pci 0001:00:02.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.208295 kernel: pci 0001:00:02.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.208359 kernel: pci 0001:00:03.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.208421 kernel: pci 0001:00:03.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.208484 kernel: pci 0001:00:04.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.208550 kernel: pci 0001:00:04.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.208613 kernel: pci 0001:00:04.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.208676 kernel: pci 0001:00:04.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.208740 kernel: pci 0001:00:03.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.208801 kernel: pci 0001:00:03.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.208864 kernel: pci 0001:00:02.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.208927 kernel: pci 0001:00:02.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.208990 kernel: pci 0001:00:01.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.209058 kernel: pci 0001:00:01.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.209124 kernel: pci 0001:01:00.0: BAR 0: assigned [mem 0x380000000000-0x380001ffffff 64bit pref] Aug 13 00:25:21.209195 kernel: pci 0001:01:00.1: BAR 0: assigned [mem 0x380002000000-0x380003ffffff 64bit pref] Aug 13 00:25:21.209260 kernel: pci 0001:01:00.0: BAR 6: assigned [mem 0x60000000-0x600fffff pref] Aug 13 00:25:21.209325 kernel: pci 0001:01:00.0: BAR 7: assigned [mem 0x380004000000-0x3800047fffff 64bit pref] Aug 13 00:25:21.209389 kernel: pci 0001:01:00.1: BAR 6: assigned [mem 0x60100000-0x601fffff pref] Aug 13 00:25:21.209455 kernel: pci 0001:01:00.1: BAR 7: assigned [mem 0x380004800000-0x380004ffffff 64bit pref] Aug 13 00:25:21.209517 kernel: pci 0001:00:01.0: PCI bridge to [bus 01] Aug 13 00:25:21.209580 kernel: pci 0001:00:01.0: bridge window [mem 0x60000000-0x601fffff] Aug 13 00:25:21.209647 kernel: pci 0001:00:01.0: bridge window [mem 0x380000000000-0x380005ffffff 64bit pref] Aug 13 00:25:21.209711 kernel: pci 0001:00:02.0: PCI bridge to [bus 02] Aug 13 00:25:21.209773 kernel: pci 0001:00:02.0: bridge window [mem 0x60200000-0x603fffff] Aug 13 00:25:21.209836 kernel: pci 0001:00:02.0: bridge window [mem 0x380006000000-0x3800061fffff 64bit pref] Aug 13 00:25:21.209900 kernel: pci 0001:00:03.0: PCI bridge to [bus 03] Aug 13 00:25:21.209962 kernel: pci 0001:00:03.0: bridge window [mem 0x60400000-0x605fffff] Aug 13 00:25:21.210028 kernel: pci 0001:00:03.0: bridge window [mem 0x380006200000-0x3800063fffff 64bit pref] Aug 13 00:25:21.210095 kernel: pci 0001:00:04.0: PCI bridge to [bus 04] Aug 13 00:25:21.210159 kernel: pci 0001:00:04.0: bridge window [mem 0x60600000-0x607fffff] Aug 13 00:25:21.210222 kernel: pci 0001:00:04.0: bridge window [mem 0x380006400000-0x3800065fffff 64bit pref] Aug 13 00:25:21.210281 kernel: pci_bus 0001:00: resource 4 [mem 0x60000000-0x6fffffff window] Aug 13 00:25:21.210337 kernel: pci_bus 0001:00: resource 5 [mem 0x380000000000-0x3bffdfffffff window] Aug 13 00:25:21.210413 kernel: pci_bus 0001:01: resource 1 [mem 0x60000000-0x601fffff] Aug 13 00:25:21.210475 kernel: pci_bus 0001:01: resource 2 [mem 0x380000000000-0x380005ffffff 64bit pref] Aug 13 00:25:21.210542 kernel: pci_bus 0001:02: resource 1 [mem 0x60200000-0x603fffff] Aug 13 00:25:21.210600 kernel: pci_bus 0001:02: resource 2 [mem 0x380006000000-0x3800061fffff 64bit pref] Aug 13 00:25:21.210666 kernel: pci_bus 0001:03: resource 1 [mem 0x60400000-0x605fffff] Aug 13 00:25:21.210724 kernel: pci_bus 0001:03: resource 2 [mem 0x380006200000-0x3800063fffff 64bit pref] Aug 13 00:25:21.210790 kernel: pci_bus 0001:04: resource 1 [mem 0x60600000-0x607fffff] Aug 13 00:25:21.210852 kernel: pci_bus 0001:04: resource 2 [mem 0x380006400000-0x3800065fffff 64bit pref] Aug 13 00:25:21.210863 kernel: ACPI: PCI Root Bridge [PCI6] (domain 0004 [bus 00-ff]) Aug 13 00:25:21.210931 kernel: acpi PNP0A08:07: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Aug 13 00:25:21.210994 kernel: acpi PNP0A08:07: _OSC: platform does not support [PCIeHotplug PME LTR] Aug 13 00:25:21.211060 kernel: acpi PNP0A08:07: _OSC: OS now controls [AER PCIeCapability] Aug 13 00:25:21.211122 kernel: acpi PNP0A08:07: MCFG quirk: ECAM at [mem 0x2bfff0000000-0x2bffffffffff] for [bus 00-ff] with pci_32b_read_ops Aug 13 00:25:21.211182 kernel: acpi PNP0A08:07: ECAM area [mem 0x2bfff0000000-0x2bffffffffff] reserved by PNP0C02:00 Aug 13 00:25:21.211246 kernel: acpi PNP0A08:07: ECAM at [mem 0x2bfff0000000-0x2bffffffffff] for [bus 00-ff] Aug 13 00:25:21.211257 kernel: PCI host bridge to bus 0004:00 Aug 13 00:25:21.211319 kernel: pci_bus 0004:00: root bus resource [mem 0x20000000-0x2fffffff window] Aug 13 00:25:21.211376 kernel: pci_bus 0004:00: root bus resource [mem 0x280000000000-0x2bffdfffffff window] Aug 13 00:25:21.211432 kernel: pci_bus 0004:00: root bus resource [bus 00-ff] Aug 13 00:25:21.211502 kernel: pci 0004:00:00.0: [1def:e110] type 00 class 0x060000 Aug 13 00:25:21.211577 kernel: pci 0004:00:01.0: [1def:e111] type 01 class 0x060400 Aug 13 00:25:21.211643 kernel: pci 0004:00:01.0: supports D1 D2 Aug 13 00:25:21.211708 kernel: pci 0004:00:01.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.211778 kernel: pci 0004:00:03.0: [1def:e113] type 01 class 0x060400 Aug 13 00:25:21.211843 kernel: pci 0004:00:03.0: supports D1 D2 Aug 13 00:25:21.211906 kernel: pci 0004:00:03.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.211977 kernel: pci 0004:00:05.0: [1def:e115] type 01 class 0x060400 Aug 13 00:25:21.212048 kernel: pci 0004:00:05.0: supports D1 D2 Aug 13 00:25:21.212112 kernel: pci 0004:00:05.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.212184 kernel: pci 0004:01:00.0: [1a03:1150] type 01 class 0x060400 Aug 13 00:25:21.212251 kernel: pci 0004:01:00.0: enabling Extended Tags Aug 13 00:25:21.212315 kernel: pci 0004:01:00.0: supports D1 D2 Aug 13 00:25:21.212380 kernel: pci 0004:01:00.0: PME# supported from D0 D1 D2 D3hot D3cold Aug 13 00:25:21.212457 kernel: pci_bus 0004:02: extended config space not accessible Aug 13 00:25:21.212536 kernel: pci 0004:02:00.0: [1a03:2000] type 00 class 0x030000 Aug 13 00:25:21.212605 kernel: pci 0004:02:00.0: reg 0x10: [mem 0x20000000-0x21ffffff] Aug 13 00:25:21.212672 kernel: pci 0004:02:00.0: reg 0x14: [mem 0x22000000-0x2201ffff] Aug 13 00:25:21.212740 kernel: pci 0004:02:00.0: reg 0x18: [io 0x0000-0x007f] Aug 13 00:25:21.212807 kernel: pci 0004:02:00.0: BAR 0: assigned to efifb Aug 13 00:25:21.212874 kernel: pci 0004:02:00.0: supports D1 D2 Aug 13 00:25:21.212942 kernel: pci 0004:02:00.0: PME# supported from D0 D1 D2 D3hot D3cold Aug 13 00:25:21.213019 kernel: pci 0004:03:00.0: [1912:0014] type 00 class 0x0c0330 Aug 13 00:25:21.213090 kernel: pci 0004:03:00.0: reg 0x10: [mem 0x22200000-0x22201fff 64bit] Aug 13 00:25:21.213155 kernel: pci 0004:03:00.0: PME# supported from D0 D3hot D3cold Aug 13 00:25:21.213215 kernel: pci_bus 0004:00: on NUMA node 0 Aug 13 00:25:21.213280 kernel: pci 0004:00:01.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 01-02] add_size 200000 add_align 100000 Aug 13 00:25:21.213345 kernel: pci 0004:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 03] add_size 1000 Aug 13 00:25:21.213408 kernel: pci 0004:00:03.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 03] add_size 200000 add_align 100000 Aug 13 00:25:21.213472 kernel: pci 0004:00:03.0: bridge window [mem 0x00100000-0x001fffff] to [bus 03] add_size 100000 add_align 100000 Aug 13 00:25:21.213540 kernel: pci 0004:00:05.0: bridge window [io 0x1000-0x0fff] to [bus 04] add_size 1000 Aug 13 00:25:21.213604 kernel: pci 0004:00:05.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 04] add_size 200000 add_align 100000 Aug 13 00:25:21.213668 kernel: pci 0004:00:05.0: bridge window [mem 0x00100000-0x000fffff] to [bus 04] add_size 200000 add_align 100000 Aug 13 00:25:21.213731 kernel: pci 0004:00:01.0: BAR 14: assigned [mem 0x20000000-0x22ffffff] Aug 13 00:25:21.213795 kernel: pci 0004:00:01.0: BAR 15: assigned [mem 0x280000000000-0x2800001fffff 64bit pref] Aug 13 00:25:21.213858 kernel: pci 0004:00:03.0: BAR 14: assigned [mem 0x23000000-0x231fffff] Aug 13 00:25:21.213924 kernel: pci 0004:00:03.0: BAR 15: assigned [mem 0x280000200000-0x2800003fffff 64bit pref] Aug 13 00:25:21.213987 kernel: pci 0004:00:05.0: BAR 14: assigned [mem 0x23200000-0x233fffff] Aug 13 00:25:21.214055 kernel: pci 0004:00:05.0: BAR 15: assigned [mem 0x280000400000-0x2800005fffff 64bit pref] Aug 13 00:25:21.214119 kernel: pci 0004:00:01.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.214184 kernel: pci 0004:00:01.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.214247 kernel: pci 0004:00:03.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.214311 kernel: pci 0004:00:03.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.214374 kernel: pci 0004:00:05.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.214441 kernel: pci 0004:00:05.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.214506 kernel: pci 0004:00:01.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.214570 kernel: pci 0004:00:01.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.214634 kernel: pci 0004:00:05.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.214697 kernel: pci 0004:00:05.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.214761 kernel: pci 0004:00:03.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.214824 kernel: pci 0004:00:03.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.214891 kernel: pci 0004:01:00.0: BAR 14: assigned [mem 0x20000000-0x22ffffff] Aug 13 00:25:21.214956 kernel: pci 0004:01:00.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.215025 kernel: pci 0004:01:00.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.215097 kernel: pci 0004:02:00.0: BAR 0: assigned [mem 0x20000000-0x21ffffff] Aug 13 00:25:21.215166 kernel: pci 0004:02:00.0: BAR 1: assigned [mem 0x22000000-0x2201ffff] Aug 13 00:25:21.215235 kernel: pci 0004:02:00.0: BAR 2: no space for [io size 0x0080] Aug 13 00:25:21.215302 kernel: pci 0004:02:00.0: BAR 2: failed to assign [io size 0x0080] Aug 13 00:25:21.215367 kernel: pci 0004:01:00.0: PCI bridge to [bus 02] Aug 13 00:25:21.215432 kernel: pci 0004:01:00.0: bridge window [mem 0x20000000-0x22ffffff] Aug 13 00:25:21.215497 kernel: pci 0004:00:01.0: PCI bridge to [bus 01-02] Aug 13 00:25:21.215562 kernel: pci 0004:00:01.0: bridge window [mem 0x20000000-0x22ffffff] Aug 13 00:25:21.215625 kernel: pci 0004:00:01.0: bridge window [mem 0x280000000000-0x2800001fffff 64bit pref] Aug 13 00:25:21.215694 kernel: pci 0004:03:00.0: BAR 0: assigned [mem 0x23000000-0x23001fff 64bit] Aug 13 00:25:21.215758 kernel: pci 0004:00:03.0: PCI bridge to [bus 03] Aug 13 00:25:21.215822 kernel: pci 0004:00:03.0: bridge window [mem 0x23000000-0x231fffff] Aug 13 00:25:21.215885 kernel: pci 0004:00:03.0: bridge window [mem 0x280000200000-0x2800003fffff 64bit pref] Aug 13 00:25:21.215949 kernel: pci 0004:00:05.0: PCI bridge to [bus 04] Aug 13 00:25:21.216014 kernel: pci 0004:00:05.0: bridge window [mem 0x23200000-0x233fffff] Aug 13 00:25:21.216082 kernel: pci 0004:00:05.0: bridge window [mem 0x280000400000-0x2800005fffff 64bit pref] Aug 13 00:25:21.216141 kernel: pci_bus 0004:00: Some PCI device resources are unassigned, try booting with pci=realloc Aug 13 00:25:21.216201 kernel: pci_bus 0004:00: resource 4 [mem 0x20000000-0x2fffffff window] Aug 13 00:25:21.216258 kernel: pci_bus 0004:00: resource 5 [mem 0x280000000000-0x2bffdfffffff window] Aug 13 00:25:21.216328 kernel: pci_bus 0004:01: resource 1 [mem 0x20000000-0x22ffffff] Aug 13 00:25:21.216390 kernel: pci_bus 0004:01: resource 2 [mem 0x280000000000-0x2800001fffff 64bit pref] Aug 13 00:25:21.216458 kernel: pci_bus 0004:02: resource 1 [mem 0x20000000-0x22ffffff] Aug 13 00:25:21.216525 kernel: pci_bus 0004:03: resource 1 [mem 0x23000000-0x231fffff] Aug 13 00:25:21.216585 kernel: pci_bus 0004:03: resource 2 [mem 0x280000200000-0x2800003fffff 64bit pref] Aug 13 00:25:21.216652 kernel: pci_bus 0004:04: resource 1 [mem 0x23200000-0x233fffff] Aug 13 00:25:21.216713 kernel: pci_bus 0004:04: resource 2 [mem 0x280000400000-0x2800005fffff 64bit pref] Aug 13 00:25:21.216723 kernel: iommu: Default domain type: Translated Aug 13 00:25:21.216734 kernel: iommu: DMA domain TLB invalidation policy: strict mode Aug 13 00:25:21.216742 kernel: efivars: Registered efivars operations Aug 13 00:25:21.216811 kernel: pci 0004:02:00.0: vgaarb: setting as boot VGA device Aug 13 00:25:21.216880 kernel: pci 0004:02:00.0: vgaarb: bridge control possible Aug 13 00:25:21.216950 kernel: pci 0004:02:00.0: vgaarb: VGA device added: decodes=io+mem,owns=none,locks=none Aug 13 00:25:21.216961 kernel: vgaarb: loaded Aug 13 00:25:21.216970 kernel: clocksource: Switched to clocksource arch_sys_counter Aug 13 00:25:21.216980 kernel: VFS: Disk quotas dquot_6.6.0 Aug 13 00:25:21.216988 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Aug 13 00:25:21.216996 kernel: pnp: PnP ACPI init Aug 13 00:25:21.217068 kernel: system 00:00: [mem 0x3bfff0000000-0x3bffffffffff window] could not be reserved Aug 13 00:25:21.217130 kernel: system 00:00: [mem 0x3ffff0000000-0x3fffffffffff window] could not be reserved Aug 13 00:25:21.217189 kernel: system 00:00: [mem 0x23fff0000000-0x23ffffffffff window] could not be reserved Aug 13 00:25:21.217248 kernel: system 00:00: [mem 0x27fff0000000-0x27ffffffffff window] could not be reserved Aug 13 00:25:21.217305 kernel: system 00:00: [mem 0x2bfff0000000-0x2bffffffffff window] could not be reserved Aug 13 00:25:21.217369 kernel: system 00:00: [mem 0x2ffff0000000-0x2fffffffffff window] could not be reserved Aug 13 00:25:21.217428 kernel: system 00:00: [mem 0x33fff0000000-0x33ffffffffff window] could not be reserved Aug 13 00:25:21.217487 kernel: system 00:00: [mem 0x37fff0000000-0x37ffffffffff window] could not be reserved Aug 13 00:25:21.217497 kernel: pnp: PnP ACPI: found 1 devices Aug 13 00:25:21.217505 kernel: NET: Registered PF_INET protocol family Aug 13 00:25:21.217513 kernel: IP idents hash table entries: 262144 (order: 9, 2097152 bytes, linear) Aug 13 00:25:21.217521 kernel: tcp_listen_portaddr_hash hash table entries: 65536 (order: 8, 1048576 bytes, linear) Aug 13 00:25:21.217532 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Aug 13 00:25:21.217540 kernel: TCP established hash table entries: 524288 (order: 10, 4194304 bytes, linear) Aug 13 00:25:21.217548 kernel: TCP bind hash table entries: 65536 (order: 9, 2097152 bytes, linear) Aug 13 00:25:21.217556 kernel: TCP: Hash tables configured (established 524288 bind 65536) Aug 13 00:25:21.217565 kernel: UDP hash table entries: 65536 (order: 9, 2097152 bytes, linear) Aug 13 00:25:21.217573 kernel: UDP-Lite hash table entries: 65536 (order: 9, 2097152 bytes, linear) Aug 13 00:25:21.217581 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Aug 13 00:25:21.217650 kernel: pci 0001:01:00.0: CLS mismatch (64 != 32), using 64 bytes Aug 13 00:25:21.217661 kernel: kvm [1]: IPA Size Limit: 48 bits Aug 13 00:25:21.217671 kernel: kvm [1]: GICv3: no GICV resource entry Aug 13 00:25:21.217679 kernel: kvm [1]: disabling GICv2 emulation Aug 13 00:25:21.217687 kernel: kvm [1]: GIC system register CPU interface enabled Aug 13 00:25:21.217695 kernel: kvm [1]: vgic interrupt IRQ9 Aug 13 00:25:21.217703 kernel: kvm [1]: VHE mode initialized successfully Aug 13 00:25:21.217711 kernel: Initialise system trusted keyrings Aug 13 00:25:21.217719 kernel: workingset: timestamp_bits=39 max_order=26 bucket_order=0 Aug 13 00:25:21.217728 kernel: Key type asymmetric registered Aug 13 00:25:21.217735 kernel: Asymmetric key parser 'x509' registered Aug 13 00:25:21.217745 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Aug 13 00:25:21.217753 kernel: io scheduler mq-deadline registered Aug 13 00:25:21.217761 kernel: io scheduler kyber registered Aug 13 00:25:21.217771 kernel: io scheduler bfq registered Aug 13 00:25:21.217779 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0C:00/input/input0 Aug 13 00:25:21.217787 kernel: ACPI: button: Power Button [PWRB] Aug 13 00:25:21.217795 kernel: ACPI GTDT: found 1 SBSA generic Watchdog(s). Aug 13 00:25:21.217803 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Aug 13 00:25:21.217875 kernel: arm-smmu-v3 arm-smmu-v3.0.auto: option mask 0x0 Aug 13 00:25:21.217940 kernel: arm-smmu-v3 arm-smmu-v3.0.auto: IDR0.COHACC overridden by FW configuration (false) Aug 13 00:25:21.218001 kernel: arm-smmu-v3 arm-smmu-v3.0.auto: ias 48-bit, oas 48-bit (features 0x000c1eff) Aug 13 00:25:21.218067 kernel: arm-smmu-v3 arm-smmu-v3.0.auto: allocated 262144 entries for cmdq Aug 13 00:25:21.218126 kernel: arm-smmu-v3 arm-smmu-v3.0.auto: allocated 131072 entries for evtq Aug 13 00:25:21.218187 kernel: arm-smmu-v3 arm-smmu-v3.0.auto: allocated 262144 entries for priq Aug 13 00:25:21.218254 kernel: arm-smmu-v3 arm-smmu-v3.1.auto: option mask 0x0 Aug 13 00:25:21.218318 kernel: arm-smmu-v3 arm-smmu-v3.1.auto: IDR0.COHACC overridden by FW configuration (false) Aug 13 00:25:21.218378 kernel: arm-smmu-v3 arm-smmu-v3.1.auto: ias 48-bit, oas 48-bit (features 0x000c1eff) Aug 13 00:25:21.218438 kernel: arm-smmu-v3 arm-smmu-v3.1.auto: allocated 262144 entries for cmdq Aug 13 00:25:21.218496 kernel: arm-smmu-v3 arm-smmu-v3.1.auto: allocated 131072 entries for evtq Aug 13 00:25:21.218556 kernel: arm-smmu-v3 arm-smmu-v3.1.auto: allocated 262144 entries for priq Aug 13 00:25:21.218623 kernel: arm-smmu-v3 arm-smmu-v3.2.auto: option mask 0x0 Aug 13 00:25:21.218684 kernel: arm-smmu-v3 arm-smmu-v3.2.auto: IDR0.COHACC overridden by FW configuration (false) Aug 13 00:25:21.218747 kernel: arm-smmu-v3 arm-smmu-v3.2.auto: ias 48-bit, oas 48-bit (features 0x000c1eff) Aug 13 00:25:21.218806 kernel: arm-smmu-v3 arm-smmu-v3.2.auto: allocated 262144 entries for cmdq Aug 13 00:25:21.218867 kernel: arm-smmu-v3 arm-smmu-v3.2.auto: allocated 131072 entries for evtq Aug 13 00:25:21.218926 kernel: arm-smmu-v3 arm-smmu-v3.2.auto: allocated 262144 entries for priq Aug 13 00:25:21.218993 kernel: arm-smmu-v3 arm-smmu-v3.3.auto: option mask 0x0 Aug 13 00:25:21.219197 kernel: arm-smmu-v3 arm-smmu-v3.3.auto: IDR0.COHACC overridden by FW configuration (false) Aug 13 00:25:21.219265 kernel: arm-smmu-v3 arm-smmu-v3.3.auto: ias 48-bit, oas 48-bit (features 0x000c1eff) Aug 13 00:25:21.219324 kernel: arm-smmu-v3 arm-smmu-v3.3.auto: allocated 262144 entries for cmdq Aug 13 00:25:21.219382 kernel: arm-smmu-v3 arm-smmu-v3.3.auto: allocated 131072 entries for evtq Aug 13 00:25:21.219441 kernel: arm-smmu-v3 arm-smmu-v3.3.auto: allocated 262144 entries for priq Aug 13 00:25:21.219514 kernel: arm-smmu-v3 arm-smmu-v3.4.auto: option mask 0x0 Aug 13 00:25:21.219573 kernel: arm-smmu-v3 arm-smmu-v3.4.auto: IDR0.COHACC overridden by FW configuration (false) Aug 13 00:25:21.219632 kernel: arm-smmu-v3 arm-smmu-v3.4.auto: ias 48-bit, oas 48-bit (features 0x000c1eff) Aug 13 00:25:21.219692 kernel: arm-smmu-v3 arm-smmu-v3.4.auto: allocated 262144 entries for cmdq Aug 13 00:25:21.219751 kernel: arm-smmu-v3 arm-smmu-v3.4.auto: allocated 131072 entries for evtq Aug 13 00:25:21.219809 kernel: arm-smmu-v3 arm-smmu-v3.4.auto: allocated 262144 entries for priq Aug 13 00:25:21.219876 kernel: arm-smmu-v3 arm-smmu-v3.5.auto: option mask 0x0 Aug 13 00:25:21.219935 kernel: arm-smmu-v3 arm-smmu-v3.5.auto: IDR0.COHACC overridden by FW configuration (false) Aug 13 00:25:21.219993 kernel: arm-smmu-v3 arm-smmu-v3.5.auto: ias 48-bit, oas 48-bit (features 0x000c1eff) Aug 13 00:25:21.220055 kernel: arm-smmu-v3 arm-smmu-v3.5.auto: allocated 262144 entries for cmdq Aug 13 00:25:21.220117 kernel: arm-smmu-v3 arm-smmu-v3.5.auto: allocated 131072 entries for evtq Aug 13 00:25:21.220174 kernel: arm-smmu-v3 arm-smmu-v3.5.auto: allocated 262144 entries for priq Aug 13 00:25:21.220241 kernel: arm-smmu-v3 arm-smmu-v3.6.auto: option mask 0x0 Aug 13 00:25:21.220300 kernel: arm-smmu-v3 arm-smmu-v3.6.auto: IDR0.COHACC overridden by FW configuration (false) Aug 13 00:25:21.220359 kernel: arm-smmu-v3 arm-smmu-v3.6.auto: ias 48-bit, oas 48-bit (features 0x000c1eff) Aug 13 00:25:21.220418 kernel: arm-smmu-v3 arm-smmu-v3.6.auto: allocated 262144 entries for cmdq Aug 13 00:25:21.220478 kernel: arm-smmu-v3 arm-smmu-v3.6.auto: allocated 131072 entries for evtq Aug 13 00:25:21.220537 kernel: arm-smmu-v3 arm-smmu-v3.6.auto: allocated 262144 entries for priq Aug 13 00:25:21.220601 kernel: arm-smmu-v3 arm-smmu-v3.7.auto: option mask 0x0 Aug 13 00:25:21.220660 kernel: arm-smmu-v3 arm-smmu-v3.7.auto: IDR0.COHACC overridden by FW configuration (false) Aug 13 00:25:21.220718 kernel: arm-smmu-v3 arm-smmu-v3.7.auto: ias 48-bit, oas 48-bit (features 0x000c1eff) Aug 13 00:25:21.220777 kernel: arm-smmu-v3 arm-smmu-v3.7.auto: allocated 262144 entries for cmdq Aug 13 00:25:21.220838 kernel: arm-smmu-v3 arm-smmu-v3.7.auto: allocated 131072 entries for evtq Aug 13 00:25:21.220897 kernel: arm-smmu-v3 arm-smmu-v3.7.auto: allocated 262144 entries for priq Aug 13 00:25:21.220907 kernel: thunder_xcv, ver 1.0 Aug 13 00:25:21.220916 kernel: thunder_bgx, ver 1.0 Aug 13 00:25:21.220924 kernel: nicpf, ver 1.0 Aug 13 00:25:21.220932 kernel: nicvf, ver 1.0 Aug 13 00:25:21.220996 kernel: rtc-efi rtc-efi.0: registered as rtc0 Aug 13 00:25:21.221060 kernel: rtc-efi rtc-efi.0: setting system clock to 2025-08-13T00:25:19 UTC (1755044719) Aug 13 00:25:21.221073 kernel: efifb: probing for efifb Aug 13 00:25:21.221082 kernel: efifb: framebuffer at 0x20000000, using 1876k, total 1875k Aug 13 00:25:21.221092 kernel: efifb: mode is 800x600x32, linelength=3200, pages=1 Aug 13 00:25:21.221100 kernel: efifb: scrolling: redraw Aug 13 00:25:21.221108 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Aug 13 00:25:21.221116 kernel: Console: switching to colour frame buffer device 100x37 Aug 13 00:25:21.221124 kernel: fb0: EFI VGA frame buffer device Aug 13 00:25:21.221132 kernel: SMCCC: SOC_ID: ID = jep106:0a16:0001 Revision = 0x000000a1 Aug 13 00:25:21.221141 kernel: hid: raw HID events driver (C) Jiri Kosina Aug 13 00:25:21.221150 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 7 counters available Aug 13 00:25:21.221158 kernel: watchdog: Delayed init of the lockup detector failed: -19 Aug 13 00:25:21.221167 kernel: watchdog: Hard watchdog permanently disabled Aug 13 00:25:21.221175 kernel: NET: Registered PF_INET6 protocol family Aug 13 00:25:21.221183 kernel: Segment Routing with IPv6 Aug 13 00:25:21.221191 kernel: In-situ OAM (IOAM) with IPv6 Aug 13 00:25:21.221199 kernel: NET: Registered PF_PACKET protocol family Aug 13 00:25:21.221207 kernel: Key type dns_resolver registered Aug 13 00:25:21.221215 kernel: registered taskstats version 1 Aug 13 00:25:21.221225 kernel: Loading compiled-in X.509 certificates Aug 13 00:25:21.221233 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.100-flatcar: 7263800c6d21650660e2b030c1023dce09b1e8b6' Aug 13 00:25:21.221241 kernel: Key type .fscrypt registered Aug 13 00:25:21.221249 kernel: Key type fscrypt-provisioning registered Aug 13 00:25:21.221257 kernel: ima: No TPM chip found, activating TPM-bypass! Aug 13 00:25:21.221265 kernel: ima: Allocated hash algorithm: sha1 Aug 13 00:25:21.221273 kernel: ima: No architecture policies found Aug 13 00:25:21.221281 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Aug 13 00:25:21.221349 kernel: pcieport 000d:00:01.0: Adding to iommu group 0 Aug 13 00:25:21.221418 kernel: pcieport 000d:00:01.0: AER: enabled with IRQ 91 Aug 13 00:25:21.221483 kernel: pcieport 000d:00:02.0: Adding to iommu group 1 Aug 13 00:25:21.221548 kernel: pcieport 000d:00:02.0: AER: enabled with IRQ 91 Aug 13 00:25:21.221614 kernel: pcieport 000d:00:03.0: Adding to iommu group 2 Aug 13 00:25:21.221678 kernel: pcieport 000d:00:03.0: AER: enabled with IRQ 91 Aug 13 00:25:21.221744 kernel: pcieport 000d:00:04.0: Adding to iommu group 3 Aug 13 00:25:21.221807 kernel: pcieport 000d:00:04.0: AER: enabled with IRQ 91 Aug 13 00:25:21.221873 kernel: pcieport 0000:00:01.0: Adding to iommu group 4 Aug 13 00:25:21.221940 kernel: pcieport 0000:00:01.0: AER: enabled with IRQ 92 Aug 13 00:25:21.222006 kernel: pcieport 0000:00:02.0: Adding to iommu group 5 Aug 13 00:25:21.222080 kernel: pcieport 0000:00:02.0: AER: enabled with IRQ 92 Aug 13 00:25:21.222148 kernel: pcieport 0000:00:03.0: Adding to iommu group 6 Aug 13 00:25:21.222212 kernel: pcieport 0000:00:03.0: AER: enabled with IRQ 92 Aug 13 00:25:21.222278 kernel: pcieport 0000:00:04.0: Adding to iommu group 7 Aug 13 00:25:21.222342 kernel: pcieport 0000:00:04.0: AER: enabled with IRQ 92 Aug 13 00:25:21.222407 kernel: pcieport 0005:00:01.0: Adding to iommu group 8 Aug 13 00:25:21.222471 kernel: pcieport 0005:00:01.0: AER: enabled with IRQ 93 Aug 13 00:25:21.222540 kernel: pcieport 0005:00:03.0: Adding to iommu group 9 Aug 13 00:25:21.222604 kernel: pcieport 0005:00:03.0: AER: enabled with IRQ 93 Aug 13 00:25:21.222668 kernel: pcieport 0005:00:05.0: Adding to iommu group 10 Aug 13 00:25:21.222732 kernel: pcieport 0005:00:05.0: AER: enabled with IRQ 93 Aug 13 00:25:21.222797 kernel: pcieport 0005:00:07.0: Adding to iommu group 11 Aug 13 00:25:21.222861 kernel: pcieport 0005:00:07.0: AER: enabled with IRQ 93 Aug 13 00:25:21.222927 kernel: pcieport 0003:00:01.0: Adding to iommu group 12 Aug 13 00:25:21.222990 kernel: pcieport 0003:00:01.0: AER: enabled with IRQ 94 Aug 13 00:25:21.223061 kernel: pcieport 0003:00:03.0: Adding to iommu group 13 Aug 13 00:25:21.223126 kernel: pcieport 0003:00:03.0: AER: enabled with IRQ 94 Aug 13 00:25:21.223190 kernel: pcieport 0003:00:05.0: Adding to iommu group 14 Aug 13 00:25:21.223254 kernel: pcieport 0003:00:05.0: AER: enabled with IRQ 94 Aug 13 00:25:21.223320 kernel: pcieport 000c:00:01.0: Adding to iommu group 15 Aug 13 00:25:21.223384 kernel: pcieport 000c:00:01.0: AER: enabled with IRQ 95 Aug 13 00:25:21.223450 kernel: pcieport 000c:00:02.0: Adding to iommu group 16 Aug 13 00:25:21.223514 kernel: pcieport 000c:00:02.0: AER: enabled with IRQ 95 Aug 13 00:25:21.223581 kernel: pcieport 000c:00:03.0: Adding to iommu group 17 Aug 13 00:25:21.223646 kernel: pcieport 000c:00:03.0: AER: enabled with IRQ 95 Aug 13 00:25:21.223711 kernel: pcieport 000c:00:04.0: Adding to iommu group 18 Aug 13 00:25:21.223777 kernel: pcieport 000c:00:04.0: AER: enabled with IRQ 95 Aug 13 00:25:21.223843 kernel: pcieport 0002:00:01.0: Adding to iommu group 19 Aug 13 00:25:21.223907 kernel: pcieport 0002:00:01.0: AER: enabled with IRQ 96 Aug 13 00:25:21.223971 kernel: pcieport 0002:00:03.0: Adding to iommu group 20 Aug 13 00:25:21.224034 kernel: pcieport 0002:00:03.0: AER: enabled with IRQ 96 Aug 13 00:25:21.224106 kernel: pcieport 0002:00:05.0: Adding to iommu group 21 Aug 13 00:25:21.224169 kernel: pcieport 0002:00:05.0: AER: enabled with IRQ 96 Aug 13 00:25:21.224234 kernel: pcieport 0002:00:07.0: Adding to iommu group 22 Aug 13 00:25:21.224297 kernel: pcieport 0002:00:07.0: AER: enabled with IRQ 96 Aug 13 00:25:21.224362 kernel: pcieport 0001:00:01.0: Adding to iommu group 23 Aug 13 00:25:21.224425 kernel: pcieport 0001:00:01.0: AER: enabled with IRQ 97 Aug 13 00:25:21.224490 kernel: pcieport 0001:00:02.0: Adding to iommu group 24 Aug 13 00:25:21.224552 kernel: pcieport 0001:00:02.0: AER: enabled with IRQ 97 Aug 13 00:25:21.224616 kernel: pcieport 0001:00:03.0: Adding to iommu group 25 Aug 13 00:25:21.224682 kernel: pcieport 0001:00:03.0: AER: enabled with IRQ 97 Aug 13 00:25:21.224746 kernel: pcieport 0001:00:04.0: Adding to iommu group 26 Aug 13 00:25:21.224810 kernel: pcieport 0001:00:04.0: AER: enabled with IRQ 97 Aug 13 00:25:21.224874 kernel: pcieport 0004:00:01.0: Adding to iommu group 27 Aug 13 00:25:21.224938 kernel: pcieport 0004:00:01.0: AER: enabled with IRQ 98 Aug 13 00:25:21.225002 kernel: pcieport 0004:00:03.0: Adding to iommu group 28 Aug 13 00:25:21.225069 kernel: pcieport 0004:00:03.0: AER: enabled with IRQ 98 Aug 13 00:25:21.225136 kernel: pcieport 0004:00:05.0: Adding to iommu group 29 Aug 13 00:25:21.225200 kernel: pcieport 0004:00:05.0: AER: enabled with IRQ 98 Aug 13 00:25:21.225267 kernel: pcieport 0004:01:00.0: Adding to iommu group 30 Aug 13 00:25:21.225278 kernel: clk: Disabling unused clocks Aug 13 00:25:21.225286 kernel: Freeing unused kernel memory: 39424K Aug 13 00:25:21.225294 kernel: Run /init as init process Aug 13 00:25:21.225302 kernel: with arguments: Aug 13 00:25:21.225310 kernel: /init Aug 13 00:25:21.225318 kernel: with environment: Aug 13 00:25:21.225328 kernel: HOME=/ Aug 13 00:25:21.225336 kernel: TERM=linux Aug 13 00:25:21.225344 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Aug 13 00:25:21.225354 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Aug 13 00:25:21.225364 systemd[1]: Detected architecture arm64. Aug 13 00:25:21.225373 systemd[1]: Running in initrd. Aug 13 00:25:21.225381 systemd[1]: No hostname configured, using default hostname. Aug 13 00:25:21.225389 systemd[1]: Hostname set to . Aug 13 00:25:21.225399 systemd[1]: Initializing machine ID from random generator. Aug 13 00:25:21.225408 systemd[1]: Queued start job for default target initrd.target. Aug 13 00:25:21.225416 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Aug 13 00:25:21.225425 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Aug 13 00:25:21.225434 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Aug 13 00:25:21.225442 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Aug 13 00:25:21.225451 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Aug 13 00:25:21.225460 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Aug 13 00:25:21.225470 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Aug 13 00:25:21.225479 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Aug 13 00:25:21.225488 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Aug 13 00:25:21.225496 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Aug 13 00:25:21.225505 systemd[1]: Reached target paths.target - Path Units. Aug 13 00:25:21.225513 systemd[1]: Reached target slices.target - Slice Units. Aug 13 00:25:21.225523 systemd[1]: Reached target swap.target - Swaps. Aug 13 00:25:21.225531 systemd[1]: Reached target timers.target - Timer Units. Aug 13 00:25:21.225540 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Aug 13 00:25:21.225548 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Aug 13 00:25:21.225557 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Aug 13 00:25:21.225565 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Aug 13 00:25:21.225574 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Aug 13 00:25:21.225582 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Aug 13 00:25:21.225590 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Aug 13 00:25:21.225600 systemd[1]: Reached target sockets.target - Socket Units. Aug 13 00:25:21.225608 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Aug 13 00:25:21.225617 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Aug 13 00:25:21.225625 systemd[1]: Finished network-cleanup.service - Network Cleanup. Aug 13 00:25:21.225634 systemd[1]: Starting systemd-fsck-usr.service... Aug 13 00:25:21.225642 systemd[1]: Starting systemd-journald.service - Journal Service... Aug 13 00:25:21.225650 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Aug 13 00:25:21.225680 systemd-journald[898]: Collecting audit messages is disabled. Aug 13 00:25:21.225703 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 00:25:21.225711 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Aug 13 00:25:21.225720 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Aug 13 00:25:21.225728 kernel: Bridge firewalling registered Aug 13 00:25:21.225737 systemd-journald[898]: Journal started Aug 13 00:25:21.225758 systemd-journald[898]: Runtime Journal (/run/log/journal/04c65096637449eca2db76111d8b0bcc) is 8.0M, max 4.0G, 3.9G free. Aug 13 00:25:21.183537 systemd-modules-load[900]: Inserted module 'overlay' Aug 13 00:25:21.260203 systemd[1]: Started systemd-journald.service - Journal Service. Aug 13 00:25:21.207000 systemd-modules-load[900]: Inserted module 'br_netfilter' Aug 13 00:25:21.265857 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Aug 13 00:25:21.276666 systemd[1]: Finished systemd-fsck-usr.service. Aug 13 00:25:21.287484 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Aug 13 00:25:21.298221 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 00:25:21.328230 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Aug 13 00:25:21.334301 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Aug 13 00:25:21.365128 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Aug 13 00:25:21.371408 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Aug 13 00:25:21.387753 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Aug 13 00:25:21.404140 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Aug 13 00:25:21.420820 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Aug 13 00:25:21.432276 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Aug 13 00:25:21.458193 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Aug 13 00:25:21.468277 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Aug 13 00:25:21.480267 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Aug 13 00:25:21.505747 dracut-cmdline[942]: dracut-dracut-053 Aug 13 00:25:21.505747 dracut-cmdline[942]: Using kernel command line parameters: rd.driver.pre=btrfs BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty0 console=ttyS1,115200n8 flatcar.first_boot=detected acpi=force flatcar.oem.id=packet flatcar.autologin verity.usrhash=2f9df6e9e6c671c457040a64675390bbff42294b08c628cd2dc472ed8120146a Aug 13 00:25:21.494420 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Aug 13 00:25:21.506669 systemd-resolved[944]: Positive Trust Anchors: Aug 13 00:25:21.506678 systemd-resolved[944]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Aug 13 00:25:21.506710 systemd-resolved[944]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Aug 13 00:25:21.521751 systemd-resolved[944]: Defaulting to hostname 'linux'. Aug 13 00:25:21.523261 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Aug 13 00:25:21.557704 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Aug 13 00:25:21.662044 kernel: SCSI subsystem initialized Aug 13 00:25:21.677042 kernel: Loading iSCSI transport class v2.0-870. Aug 13 00:25:21.696046 kernel: iscsi: registered transport (tcp) Aug 13 00:25:21.723158 kernel: iscsi: registered transport (qla4xxx) Aug 13 00:25:21.723180 kernel: QLogic iSCSI HBA Driver Aug 13 00:25:21.767515 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Aug 13 00:25:21.788153 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Aug 13 00:25:21.833034 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Aug 13 00:25:21.833073 kernel: device-mapper: uevent: version 1.0.3 Aug 13 00:25:21.842653 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Aug 13 00:25:21.908049 kernel: raid6: neonx8 gen() 15845 MB/s Aug 13 00:25:21.933043 kernel: raid6: neonx4 gen() 15733 MB/s Aug 13 00:25:21.958047 kernel: raid6: neonx2 gen() 13376 MB/s Aug 13 00:25:21.983047 kernel: raid6: neonx1 gen() 10511 MB/s Aug 13 00:25:22.008047 kernel: raid6: int64x8 gen() 6984 MB/s Aug 13 00:25:22.033047 kernel: raid6: int64x4 gen() 7374 MB/s Aug 13 00:25:22.058046 kernel: raid6: int64x2 gen() 6155 MB/s Aug 13 00:25:22.086059 kernel: raid6: int64x1 gen() 5075 MB/s Aug 13 00:25:22.086081 kernel: raid6: using algorithm neonx8 gen() 15845 MB/s Aug 13 00:25:22.120469 kernel: raid6: .... xor() 11977 MB/s, rmw enabled Aug 13 00:25:22.120490 kernel: raid6: using neon recovery algorithm Aug 13 00:25:22.140048 kernel: xor: measuring software checksum speed Aug 13 00:25:22.151684 kernel: 8regs : 19119 MB/sec Aug 13 00:25:22.151705 kernel: 32regs : 19679 MB/sec Aug 13 00:25:22.167116 kernel: arm64_neon : 26726 MB/sec Aug 13 00:25:22.167137 kernel: xor: using function: arm64_neon (26726 MB/sec) Aug 13 00:25:22.228046 kernel: Btrfs loaded, zoned=no, fsverity=no Aug 13 00:25:22.239102 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Aug 13 00:25:22.266163 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Aug 13 00:25:22.279353 systemd-udevd[1140]: Using default interface naming scheme 'v255'. Aug 13 00:25:22.282421 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Aug 13 00:25:22.305186 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Aug 13 00:25:22.319341 dracut-pre-trigger[1150]: rd.md=0: removing MD RAID activation Aug 13 00:25:22.345804 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Aug 13 00:25:22.372203 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Aug 13 00:25:22.481804 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Aug 13 00:25:22.510165 kernel: pps_core: LinuxPPS API ver. 1 registered Aug 13 00:25:22.510190 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Aug 13 00:25:22.524216 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Aug 13 00:25:22.662658 kernel: ACPI: bus type USB registered Aug 13 00:25:22.662682 kernel: usbcore: registered new interface driver usbfs Aug 13 00:25:22.662707 kernel: usbcore: registered new interface driver hub Aug 13 00:25:22.662727 kernel: usbcore: registered new device driver usb Aug 13 00:25:22.662748 kernel: PTP clock support registered Aug 13 00:25:22.662769 kernel: xhci_hcd 0004:03:00.0: Adding to iommu group 31 Aug 13 00:25:22.662934 kernel: xhci_hcd 0004:03:00.0: xHCI Host Controller Aug 13 00:25:22.663020 kernel: xhci_hcd 0004:03:00.0: new USB bus registered, assigned bus number 1 Aug 13 00:25:22.663110 kernel: xhci_hcd 0004:03:00.0: Zeroing 64bit base registers, expecting fault Aug 13 00:25:22.663190 kernel: igb: Intel(R) Gigabit Ethernet Network Driver Aug 13 00:25:22.663203 kernel: mlx5_core 0001:01:00.0: Adding to iommu group 32 Aug 13 00:25:22.663291 kernel: igb: Copyright (c) 2007-2014 Intel Corporation. Aug 13 00:25:22.663301 kernel: igb 0003:03:00.0: Adding to iommu group 33 Aug 13 00:25:22.686046 kernel: nvme 0005:03:00.0: Adding to iommu group 34 Aug 13 00:25:22.686176 kernel: nvme 0005:04:00.0: Adding to iommu group 35 Aug 13 00:25:22.702468 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Aug 13 00:25:22.714068 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Aug 13 00:25:22.720139 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Aug 13 00:25:22.737408 systemd[1]: Reached target remote-fs.target - Remote File Systems. Aug 13 00:25:22.748963 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Aug 13 00:25:22.749116 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Aug 13 00:25:22.766363 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Aug 13 00:25:22.790245 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Aug 13 00:25:22.800932 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Aug 13 00:25:22.961755 kernel: xhci_hcd 0004:03:00.0: hcc params 0x014051cf hci version 0x100 quirks 0x0000001100000010 Aug 13 00:25:22.961980 kernel: xhci_hcd 0004:03:00.0: xHCI Host Controller Aug 13 00:25:22.962077 kernel: xhci_hcd 0004:03:00.0: new USB bus registered, assigned bus number 2 Aug 13 00:25:22.962158 kernel: xhci_hcd 0004:03:00.0: Host supports USB 3.0 SuperSpeed Aug 13 00:25:22.962235 kernel: hub 1-0:1.0: USB hub found Aug 13 00:25:22.962334 kernel: hub 1-0:1.0: 4 ports detected Aug 13 00:25:22.962413 kernel: mlx5_core 0001:01:00.0: firmware version: 14.31.1014 Aug 13 00:25:22.962499 kernel: usb usb2: We don't know the algorithms for LPM for this host, disabling LPM. Aug 13 00:25:22.962592 kernel: hub 2-0:1.0: USB hub found Aug 13 00:25:22.962680 kernel: mlx5_core 0001:01:00.0: 31.504 Gb/s available PCIe bandwidth, limited by 8.0 GT/s PCIe x4 link at 0001:00:01.0 (capable of 63.008 Gb/s with 8.0 GT/s PCIe x8 link) Aug 13 00:25:22.962762 kernel: hub 2-0:1.0: 4 ports detected Aug 13 00:25:22.801107 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 00:25:22.879227 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 00:25:22.975255 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 00:25:22.984643 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Aug 13 00:25:23.001622 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Aug 13 00:25:23.001727 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 00:25:23.058570 kernel: nvme nvme0: pci function 0005:04:00.0 Aug 13 00:25:23.058714 kernel: nvme nvme1: pci function 0005:03:00.0 Aug 13 00:25:23.058802 kernel: nvme nvme0: Shutdown timeout set to 8 seconds Aug 13 00:25:23.058871 kernel: nvme nvme1: Shutdown timeout set to 8 seconds Aug 13 00:25:23.054257 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 00:25:23.073331 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 00:25:23.100685 kernel: nvme nvme1: 32/0/0 default/read/poll queues Aug 13 00:25:23.100878 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Aug 13 00:25:23.103044 kernel: GPT:9289727 != 1875385007 Aug 13 00:25:23.103067 kernel: GPT:Alternate GPT header not at the end of the disk. Aug 13 00:25:23.103086 kernel: GPT:9289727 != 1875385007 Aug 13 00:25:23.103110 kernel: GPT: Use GNU Parted to correct GPT errors. Aug 13 00:25:23.103129 kernel: nvme1n1: p1 p2 p3 p4 p6 p7 p9 Aug 13 00:25:23.104044 kernel: nvme nvme0: 32/0/0 default/read/poll queues Aug 13 00:25:23.105042 kernel: igb 0003:03:00.0: added PHC on eth0 Aug 13 00:25:23.105149 kernel: igb 0003:03:00.0: Intel(R) Gigabit Ethernet Network Connection Aug 13 00:25:23.105237 kernel: igb 0003:03:00.0: eth0: (PCIe:5.0Gb/s:Width x2) 18:c0:4d:0c:6f:ac Aug 13 00:25:23.105328 kernel: igb 0003:03:00.0: eth0: PBA No: 106300-000 Aug 13 00:25:23.105408 kernel: igb 0003:03:00.0: Using MSI-X interrupts. 8 rx queue(s), 8 tx queue(s) Aug 13 00:25:23.105500 kernel: igb 0003:03:00.1: Adding to iommu group 36 Aug 13 00:25:23.162051 kernel: BTRFS: device fsid 03408483-5051-409a-aab4-4e6d5027e982 devid 1 transid 41 /dev/nvme1n1p3 scanned by (udev-worker) (1186) Aug 13 00:25:23.162077 kernel: igb 0003:03:00.1: added PHC on eth1 Aug 13 00:25:23.162243 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/nvme1n1p6 scanned by (udev-worker) (1195) Aug 13 00:25:23.171074 kernel: igb 0003:03:00.1: Intel(R) Gigabit Ethernet Network Connection Aug 13 00:25:23.289991 kernel: igb 0003:03:00.1: eth1: (PCIe:5.0Gb/s:Width x2) 18:c0:4d:0c:6f:ad Aug 13 00:25:23.301685 kernel: igb 0003:03:00.1: eth1: PBA No: 106300-000 Aug 13 00:25:23.311260 kernel: igb 0003:03:00.1: Using MSI-X interrupts. 8 rx queue(s), 8 tx queue(s) Aug 13 00:25:23.318434 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - SAMSUNG MZ1LB960HAJQ-00007 EFI-SYSTEM. Aug 13 00:25:23.369587 kernel: igb 0003:03:00.1 eno2: renamed from eth1 Aug 13 00:25:23.369706 kernel: mlx5_core 0001:01:00.0: Port module event: module 0, Cable plugged Aug 13 00:25:23.369795 kernel: igb 0003:03:00.0 eno1: renamed from eth0 Aug 13 00:25:23.369873 kernel: usb 1-3: new high-speed USB device number 2 using xhci_hcd Aug 13 00:25:23.378800 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - SAMSUNG MZ1LB960HAJQ-00007 ROOT. Aug 13 00:25:23.396606 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - SAMSUNG MZ1LB960HAJQ-00007 USR-A. Aug 13 00:25:23.411230 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - SAMSUNG MZ1LB960HAJQ-00007 USR-A. Aug 13 00:25:23.420510 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - SAMSUNG MZ1LB960HAJQ-00007 OEM. Aug 13 00:25:23.445185 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Aug 13 00:25:23.471365 kernel: nvme1n1: p1 p2 p3 p4 p6 p7 p9 Aug 13 00:25:23.451476 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Aug 13 00:25:23.507111 kernel: hub 1-3:1.0: USB hub found Aug 13 00:25:23.507265 kernel: hub 1-3:1.0: 4 ports detected Aug 13 00:25:23.507361 disk-uuid[1285]: Primary Header is updated. Aug 13 00:25:23.507361 disk-uuid[1285]: Secondary Entries is updated. Aug 13 00:25:23.507361 disk-uuid[1285]: Secondary Header is updated. Aug 13 00:25:23.524795 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Aug 13 00:25:23.595051 kernel: usb 2-3: new SuperSpeed USB device number 2 using xhci_hcd Aug 13 00:25:23.630052 kernel: hub 2-3:1.0: USB hub found Aug 13 00:25:23.630289 kernel: hub 2-3:1.0: 4 ports detected Aug 13 00:25:23.675048 kernel: mlx5_core 0001:01:00.0: MLX5E: StrdRq(0) RqSz(1024) StrdSz(256) RxCqeCmprss(0 basic) Aug 13 00:25:23.690042 kernel: mlx5_core 0001:01:00.1: Adding to iommu group 37 Aug 13 00:25:23.712868 kernel: mlx5_core 0001:01:00.1: firmware version: 14.31.1014 Aug 13 00:25:23.713031 kernel: mlx5_core 0001:01:00.1: 31.504 Gb/s available PCIe bandwidth, limited by 8.0 GT/s PCIe x4 link at 0001:00:01.0 (capable of 63.008 Gb/s with 8.0 GT/s PCIe x8 link) Aug 13 00:25:24.065724 kernel: mlx5_core 0001:01:00.1: Port module event: module 1, Cable plugged Aug 13 00:25:24.373050 kernel: mlx5_core 0001:01:00.1: MLX5E: StrdRq(0) RqSz(1024) StrdSz(256) RxCqeCmprss(0 basic) Aug 13 00:25:24.389043 kernel: mlx5_core 0001:01:00.0 enP1p1s0f0np0: renamed from eth0 Aug 13 00:25:24.407043 kernel: mlx5_core 0001:01:00.1 enP1p1s0f1np1: renamed from eth1 Aug 13 00:25:24.472011 disk-uuid[1286]: The operation has completed successfully. Aug 13 00:25:24.477454 kernel: nvme1n1: p1 p2 p3 p4 p6 p7 p9 Aug 13 00:25:24.496808 systemd[1]: disk-uuid.service: Deactivated successfully. Aug 13 00:25:24.496891 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Aug 13 00:25:24.530141 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Aug 13 00:25:24.540464 sh[1484]: Success Aug 13 00:25:24.559043 kernel: device-mapper: verity: sha256 using implementation "sha256-ce" Aug 13 00:25:24.591866 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Aug 13 00:25:24.612302 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Aug 13 00:25:24.622633 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Aug 13 00:25:24.717702 kernel: BTRFS info (device dm-0): first mount of filesystem 03408483-5051-409a-aab4-4e6d5027e982 Aug 13 00:25:24.717733 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm Aug 13 00:25:24.717753 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Aug 13 00:25:24.717773 kernel: BTRFS info (device dm-0): disabling log replay at mount time Aug 13 00:25:24.717792 kernel: BTRFS info (device dm-0): using free space tree Aug 13 00:25:24.717811 kernel: BTRFS info (device dm-0): enabling ssd optimizations Aug 13 00:25:24.719122 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Aug 13 00:25:24.729444 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Aug 13 00:25:24.744143 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Aug 13 00:25:24.750280 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Aug 13 00:25:24.863485 kernel: BTRFS info (device nvme1n1p6): first mount of filesystem dbce4b09-c4b8-4cc9-bd11-416717f60c7d Aug 13 00:25:24.863503 kernel: BTRFS info (device nvme1n1p6): using crc32c (crc32c-generic) checksum algorithm Aug 13 00:25:24.863513 kernel: BTRFS info (device nvme1n1p6): using free space tree Aug 13 00:25:24.863523 kernel: BTRFS info (device nvme1n1p6): enabling ssd optimizations Aug 13 00:25:24.863534 kernel: BTRFS info (device nvme1n1p6): auto enabling async discard Aug 13 00:25:24.863544 kernel: BTRFS info (device nvme1n1p6): last unmount of filesystem dbce4b09-c4b8-4cc9-bd11-416717f60c7d Aug 13 00:25:24.860305 systemd[1]: Finished ignition-setup.service - Ignition (setup). Aug 13 00:25:24.888232 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Aug 13 00:25:24.899007 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Aug 13 00:25:24.928204 systemd[1]: Starting systemd-networkd.service - Network Configuration... Aug 13 00:25:24.948207 systemd-networkd[1687]: lo: Link UP Aug 13 00:25:24.948212 systemd-networkd[1687]: lo: Gained carrier Aug 13 00:25:24.951802 systemd-networkd[1687]: Enumeration completed Aug 13 00:25:24.951888 systemd[1]: Started systemd-networkd.service - Network Configuration. Aug 13 00:25:24.952978 systemd-networkd[1687]: eno1: Configuring with /usr/lib/systemd/network/zz-default.network. Aug 13 00:25:24.959738 systemd[1]: Reached target network.target - Network. Aug 13 00:25:24.984580 ignition[1674]: Ignition 2.19.0 Aug 13 00:25:24.993975 unknown[1674]: fetched base config from "system" Aug 13 00:25:24.984586 ignition[1674]: Stage: fetch-offline Aug 13 00:25:24.993984 unknown[1674]: fetched user config from "system" Aug 13 00:25:24.984688 ignition[1674]: no configs at "/usr/lib/ignition/base.d" Aug 13 00:25:24.997134 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Aug 13 00:25:24.984696 ignition[1674]: no config dir at "/usr/lib/ignition/base.platform.d/packet" Aug 13 00:25:25.004652 systemd-networkd[1687]: eno2: Configuring with /usr/lib/systemd/network/zz-default.network. Aug 13 00:25:24.985044 ignition[1674]: parsed url from cmdline: "" Aug 13 00:25:25.007345 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Aug 13 00:25:24.985047 ignition[1674]: no config URL provided Aug 13 00:25:25.019206 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Aug 13 00:25:24.985051 ignition[1674]: reading system config file "/usr/lib/ignition/user.ign" Aug 13 00:25:25.057004 systemd-networkd[1687]: enP1p1s0f0np0: Configuring with /usr/lib/systemd/network/zz-default.network. Aug 13 00:25:24.985105 ignition[1674]: parsing config with SHA512: 4549cbe70e82d968e80ac52d6c85114c45051bf18f0ae7feabe644370be95ed1e38d2b867246b279ac8a182ffa114a88b569684cec4568d7a77cb44ae95e1a7c Aug 13 00:25:24.994457 ignition[1674]: fetch-offline: fetch-offline passed Aug 13 00:25:24.994461 ignition[1674]: POST message to Packet Timeline Aug 13 00:25:24.994466 ignition[1674]: POST Status error: resource requires networking Aug 13 00:25:24.994525 ignition[1674]: Ignition finished successfully Aug 13 00:25:25.042984 ignition[1711]: Ignition 2.19.0 Aug 13 00:25:25.042990 ignition[1711]: Stage: kargs Aug 13 00:25:25.043145 ignition[1711]: no configs at "/usr/lib/ignition/base.d" Aug 13 00:25:25.043154 ignition[1711]: no config dir at "/usr/lib/ignition/base.platform.d/packet" Aug 13 00:25:25.044073 ignition[1711]: kargs: kargs passed Aug 13 00:25:25.044077 ignition[1711]: POST message to Packet Timeline Aug 13 00:25:25.044089 ignition[1711]: GET https://metadata.packet.net/metadata: attempt #1 Aug 13 00:25:25.046665 ignition[1711]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:36229->[::1]:53: read: connection refused Aug 13 00:25:25.246793 ignition[1711]: GET https://metadata.packet.net/metadata: attempt #2 Aug 13 00:25:25.247253 ignition[1711]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:48311->[::1]:53: read: connection refused Aug 13 00:25:25.639053 kernel: mlx5_core 0001:01:00.0 enP1p1s0f0np0: Link up Aug 13 00:25:25.641771 systemd-networkd[1687]: enP1p1s0f1np1: Configuring with /usr/lib/systemd/network/zz-default.network. Aug 13 00:25:25.648011 ignition[1711]: GET https://metadata.packet.net/metadata: attempt #3 Aug 13 00:25:25.649909 ignition[1711]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:47865->[::1]:53: read: connection refused Aug 13 00:25:26.243053 kernel: mlx5_core 0001:01:00.1 enP1p1s0f1np1: Link up Aug 13 00:25:26.245657 systemd-networkd[1687]: eno1: Link UP Aug 13 00:25:26.245790 systemd-networkd[1687]: eno2: Link UP Aug 13 00:25:26.245911 systemd-networkd[1687]: enP1p1s0f0np0: Link UP Aug 13 00:25:26.246057 systemd-networkd[1687]: enP1p1s0f0np0: Gained carrier Aug 13 00:25:26.257177 systemd-networkd[1687]: enP1p1s0f1np1: Link UP Aug 13 00:25:26.286072 systemd-networkd[1687]: enP1p1s0f0np0: DHCPv4 address 147.75.53.22/30, gateway 147.75.53.21 acquired from 145.40.76.140 Aug 13 00:25:26.450052 ignition[1711]: GET https://metadata.packet.net/metadata: attempt #4 Aug 13 00:25:26.450575 ignition[1711]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:56469->[::1]:53: read: connection refused Aug 13 00:25:26.646252 systemd-networkd[1687]: enP1p1s0f1np1: Gained carrier Aug 13 00:25:27.526239 systemd-networkd[1687]: enP1p1s0f0np0: Gained IPv6LL Aug 13 00:25:28.051291 ignition[1711]: GET https://metadata.packet.net/metadata: attempt #5 Aug 13 00:25:28.051684 ignition[1711]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:44801->[::1]:53: read: connection refused Aug 13 00:25:28.422290 systemd-networkd[1687]: enP1p1s0f1np1: Gained IPv6LL Aug 13 00:25:31.254177 ignition[1711]: GET https://metadata.packet.net/metadata: attempt #6 Aug 13 00:25:32.553372 ignition[1711]: GET result: OK Aug 13 00:25:33.400212 ignition[1711]: Ignition finished successfully Aug 13 00:25:33.403318 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Aug 13 00:25:33.414160 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Aug 13 00:25:33.429647 ignition[1732]: Ignition 2.19.0 Aug 13 00:25:33.429653 ignition[1732]: Stage: disks Aug 13 00:25:33.429811 ignition[1732]: no configs at "/usr/lib/ignition/base.d" Aug 13 00:25:33.429821 ignition[1732]: no config dir at "/usr/lib/ignition/base.platform.d/packet" Aug 13 00:25:33.430800 ignition[1732]: disks: disks passed Aug 13 00:25:33.430804 ignition[1732]: POST message to Packet Timeline Aug 13 00:25:33.430818 ignition[1732]: GET https://metadata.packet.net/metadata: attempt #1 Aug 13 00:25:34.320034 ignition[1732]: GET result: OK Aug 13 00:25:34.794294 ignition[1732]: Ignition finished successfully Aug 13 00:25:34.796433 systemd[1]: Finished ignition-disks.service - Ignition (disks). Aug 13 00:25:34.802899 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Aug 13 00:25:34.810471 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Aug 13 00:25:34.818478 systemd[1]: Reached target local-fs.target - Local File Systems. Aug 13 00:25:34.827024 systemd[1]: Reached target sysinit.target - System Initialization. Aug 13 00:25:34.835899 systemd[1]: Reached target basic.target - Basic System. Aug 13 00:25:34.854188 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Aug 13 00:25:34.869589 systemd-fsck[1751]: ROOT: clean, 14/553520 files, 52654/553472 blocks Aug 13 00:25:34.873214 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Aug 13 00:25:34.891146 systemd[1]: Mounting sysroot.mount - /sysroot... Aug 13 00:25:34.956972 systemd[1]: Mounted sysroot.mount - /sysroot. Aug 13 00:25:34.962083 kernel: EXT4-fs (nvme1n1p9): mounted filesystem 128aec8b-f05d-48ed-8996-c9e8b21a7810 r/w with ordered data mode. Quota mode: none. Aug 13 00:25:34.967460 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Aug 13 00:25:34.989122 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Aug 13 00:25:35.082295 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/nvme1n1p6 scanned by mount (1762) Aug 13 00:25:35.082314 kernel: BTRFS info (device nvme1n1p6): first mount of filesystem dbce4b09-c4b8-4cc9-bd11-416717f60c7d Aug 13 00:25:35.082324 kernel: BTRFS info (device nvme1n1p6): using crc32c (crc32c-generic) checksum algorithm Aug 13 00:25:35.082334 kernel: BTRFS info (device nvme1n1p6): using free space tree Aug 13 00:25:35.082344 kernel: BTRFS info (device nvme1n1p6): enabling ssd optimizations Aug 13 00:25:35.082359 kernel: BTRFS info (device nvme1n1p6): auto enabling async discard Aug 13 00:25:34.995327 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Aug 13 00:25:35.088609 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Aug 13 00:25:35.099531 systemd[1]: Starting flatcar-static-network.service - Flatcar Static Network Agent... Aug 13 00:25:35.115403 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Aug 13 00:25:35.115431 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Aug 13 00:25:35.148055 coreos-metadata[1781]: Aug 13 00:25:35.146 INFO Fetching https://metadata.packet.net/metadata: Attempt #1 Aug 13 00:25:35.164854 coreos-metadata[1780]: Aug 13 00:25:35.146 INFO Fetching https://metadata.packet.net/metadata: Attempt #1 Aug 13 00:25:35.128888 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Aug 13 00:25:35.142626 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Aug 13 00:25:35.170222 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Aug 13 00:25:35.203734 initrd-setup-root[1798]: cut: /sysroot/etc/passwd: No such file or directory Aug 13 00:25:35.209920 initrd-setup-root[1805]: cut: /sysroot/etc/group: No such file or directory Aug 13 00:25:35.216353 initrd-setup-root[1812]: cut: /sysroot/etc/shadow: No such file or directory Aug 13 00:25:35.222753 initrd-setup-root[1819]: cut: /sysroot/etc/gshadow: No such file or directory Aug 13 00:25:35.293449 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Aug 13 00:25:35.321106 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Aug 13 00:25:35.352086 kernel: BTRFS info (device nvme1n1p6): last unmount of filesystem dbce4b09-c4b8-4cc9-bd11-416717f60c7d Aug 13 00:25:35.327560 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Aug 13 00:25:35.358418 systemd[1]: sysroot-oem.mount: Deactivated successfully. Aug 13 00:25:35.373825 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Aug 13 00:25:35.386281 ignition[1891]: INFO : Ignition 2.19.0 Aug 13 00:25:35.386281 ignition[1891]: INFO : Stage: mount Aug 13 00:25:35.397064 ignition[1891]: INFO : no configs at "/usr/lib/ignition/base.d" Aug 13 00:25:35.397064 ignition[1891]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/packet" Aug 13 00:25:35.397064 ignition[1891]: INFO : mount: mount passed Aug 13 00:25:35.397064 ignition[1891]: INFO : POST message to Packet Timeline Aug 13 00:25:35.397064 ignition[1891]: INFO : GET https://metadata.packet.net/metadata: attempt #1 Aug 13 00:25:35.983464 coreos-metadata[1781]: Aug 13 00:25:35.983 INFO Fetch successful Aug 13 00:25:36.033361 systemd[1]: flatcar-static-network.service: Deactivated successfully. Aug 13 00:25:36.033449 systemd[1]: Finished flatcar-static-network.service - Flatcar Static Network Agent. Aug 13 00:25:36.111306 ignition[1891]: INFO : GET result: OK Aug 13 00:25:36.456809 coreos-metadata[1780]: Aug 13 00:25:36.456 INFO Fetch successful Aug 13 00:25:36.502007 coreos-metadata[1780]: Aug 13 00:25:36.501 INFO wrote hostname ci-4081.3.5-a-588adb827d to /sysroot/etc/hostname Aug 13 00:25:36.505201 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Aug 13 00:25:39.546855 ignition[1891]: INFO : Ignition finished successfully Aug 13 00:25:39.549085 systemd[1]: Finished ignition-mount.service - Ignition (mount). Aug 13 00:25:39.565111 systemd[1]: Starting ignition-files.service - Ignition (files)... Aug 13 00:25:39.577332 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Aug 13 00:25:39.603046 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/nvme1n1p6 scanned by mount (1925) Aug 13 00:25:39.627151 kernel: BTRFS info (device nvme1n1p6): first mount of filesystem dbce4b09-c4b8-4cc9-bd11-416717f60c7d Aug 13 00:25:39.627175 kernel: BTRFS info (device nvme1n1p6): using crc32c (crc32c-generic) checksum algorithm Aug 13 00:25:39.640161 kernel: BTRFS info (device nvme1n1p6): using free space tree Aug 13 00:25:39.663065 kernel: BTRFS info (device nvme1n1p6): enabling ssd optimizations Aug 13 00:25:39.663087 kernel: BTRFS info (device nvme1n1p6): auto enabling async discard Aug 13 00:25:39.671243 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Aug 13 00:25:39.704442 ignition[1945]: INFO : Ignition 2.19.0 Aug 13 00:25:39.704442 ignition[1945]: INFO : Stage: files Aug 13 00:25:39.713942 ignition[1945]: INFO : no configs at "/usr/lib/ignition/base.d" Aug 13 00:25:39.713942 ignition[1945]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/packet" Aug 13 00:25:39.713942 ignition[1945]: DEBUG : files: compiled without relabeling support, skipping Aug 13 00:25:39.713942 ignition[1945]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Aug 13 00:25:39.713942 ignition[1945]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Aug 13 00:25:39.713942 ignition[1945]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Aug 13 00:25:39.713942 ignition[1945]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Aug 13 00:25:39.713942 ignition[1945]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Aug 13 00:25:39.713942 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-arm64.tar.gz" Aug 13 00:25:39.713942 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.3-linux-arm64.tar.gz: attempt #1 Aug 13 00:25:39.709979 unknown[1945]: wrote ssh authorized keys file for user: core Aug 13 00:25:39.806785 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Aug 13 00:25:39.973814 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-arm64.tar.gz" Aug 13 00:25:39.984327 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Aug 13 00:25:39.984327 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Aug 13 00:25:39.984327 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Aug 13 00:25:39.984327 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Aug 13 00:25:39.984327 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Aug 13 00:25:39.984327 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Aug 13 00:25:39.984327 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Aug 13 00:25:39.984327 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Aug 13 00:25:39.984327 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Aug 13 00:25:39.984327 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Aug 13 00:25:39.984327 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-arm64.raw" Aug 13 00:25:39.984327 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-arm64.raw" Aug 13 00:25:39.984327 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-arm64.raw" Aug 13 00:25:39.984327 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.33.0-arm64.raw: attempt #1 Aug 13 00:25:40.383402 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Aug 13 00:25:40.812326 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-arm64.raw" Aug 13 00:25:40.824960 ignition[1945]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Aug 13 00:25:40.824960 ignition[1945]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Aug 13 00:25:40.824960 ignition[1945]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Aug 13 00:25:40.824960 ignition[1945]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Aug 13 00:25:40.824960 ignition[1945]: INFO : files: op(d): [started] setting preset to enabled for "prepare-helm.service" Aug 13 00:25:40.824960 ignition[1945]: INFO : files: op(d): [finished] setting preset to enabled for "prepare-helm.service" Aug 13 00:25:40.824960 ignition[1945]: INFO : files: createResultFile: createFiles: op(e): [started] writing file "/sysroot/etc/.ignition-result.json" Aug 13 00:25:40.824960 ignition[1945]: INFO : files: createResultFile: createFiles: op(e): [finished] writing file "/sysroot/etc/.ignition-result.json" Aug 13 00:25:40.824960 ignition[1945]: INFO : files: files passed Aug 13 00:25:40.824960 ignition[1945]: INFO : POST message to Packet Timeline Aug 13 00:25:40.824960 ignition[1945]: INFO : GET https://metadata.packet.net/metadata: attempt #1 Aug 13 00:25:41.597597 ignition[1945]: INFO : GET result: OK Aug 13 00:25:42.176234 ignition[1945]: INFO : Ignition finished successfully Aug 13 00:25:42.178643 systemd[1]: Finished ignition-files.service - Ignition (files). Aug 13 00:25:42.198160 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Aug 13 00:25:42.210687 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Aug 13 00:25:42.229343 systemd[1]: ignition-quench.service: Deactivated successfully. Aug 13 00:25:42.229422 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Aug 13 00:25:42.247948 initrd-setup-root-after-ignition[1990]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Aug 13 00:25:42.247948 initrd-setup-root-after-ignition[1990]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Aug 13 00:25:42.242389 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Aug 13 00:25:42.296038 initrd-setup-root-after-ignition[1995]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Aug 13 00:25:42.255520 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Aug 13 00:25:42.284184 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Aug 13 00:25:42.329626 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Aug 13 00:25:42.331086 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Aug 13 00:25:42.339403 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Aug 13 00:25:42.349763 systemd[1]: Reached target initrd.target - Initrd Default Target. Aug 13 00:25:42.367799 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Aug 13 00:25:42.378135 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Aug 13 00:25:42.400901 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Aug 13 00:25:42.426149 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Aug 13 00:25:42.441139 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Aug 13 00:25:42.450601 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Aug 13 00:25:42.462394 systemd[1]: Stopped target timers.target - Timer Units. Aug 13 00:25:42.474160 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Aug 13 00:25:42.474263 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Aug 13 00:25:42.486094 systemd[1]: Stopped target initrd.target - Initrd Default Target. Aug 13 00:25:42.497550 systemd[1]: Stopped target basic.target - Basic System. Aug 13 00:25:42.509200 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Aug 13 00:25:42.520805 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Aug 13 00:25:42.532403 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Aug 13 00:25:42.543942 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Aug 13 00:25:42.555434 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Aug 13 00:25:42.566956 systemd[1]: Stopped target sysinit.target - System Initialization. Aug 13 00:25:42.578460 systemd[1]: Stopped target local-fs.target - Local File Systems. Aug 13 00:25:42.595563 systemd[1]: Stopped target swap.target - Swaps. Aug 13 00:25:42.607209 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Aug 13 00:25:42.607304 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Aug 13 00:25:42.619094 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Aug 13 00:25:42.630441 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Aug 13 00:25:42.641974 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Aug 13 00:25:42.646079 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Aug 13 00:25:42.653270 systemd[1]: dracut-initqueue.service: Deactivated successfully. Aug 13 00:25:42.653365 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Aug 13 00:25:42.664744 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Aug 13 00:25:42.664848 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Aug 13 00:25:42.676025 systemd[1]: Stopped target paths.target - Path Units. Aug 13 00:25:42.687223 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Aug 13 00:25:42.693071 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Aug 13 00:25:42.704380 systemd[1]: Stopped target slices.target - Slice Units. Aug 13 00:25:42.715983 systemd[1]: Stopped target sockets.target - Socket Units. Aug 13 00:25:42.727446 systemd[1]: iscsid.socket: Deactivated successfully. Aug 13 00:25:42.727546 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Aug 13 00:25:42.828438 ignition[2017]: INFO : Ignition 2.19.0 Aug 13 00:25:42.828438 ignition[2017]: INFO : Stage: umount Aug 13 00:25:42.828438 ignition[2017]: INFO : no configs at "/usr/lib/ignition/base.d" Aug 13 00:25:42.828438 ignition[2017]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/packet" Aug 13 00:25:42.828438 ignition[2017]: INFO : umount: umount passed Aug 13 00:25:42.828438 ignition[2017]: INFO : POST message to Packet Timeline Aug 13 00:25:42.828438 ignition[2017]: INFO : GET https://metadata.packet.net/metadata: attempt #1 Aug 13 00:25:42.739085 systemd[1]: iscsiuio.socket: Deactivated successfully. Aug 13 00:25:42.739170 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Aug 13 00:25:42.750814 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Aug 13 00:25:42.750902 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Aug 13 00:25:42.762536 systemd[1]: ignition-files.service: Deactivated successfully. Aug 13 00:25:42.762618 systemd[1]: Stopped ignition-files.service - Ignition (files). Aug 13 00:25:42.774252 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Aug 13 00:25:42.774333 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Aug 13 00:25:42.798168 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Aug 13 00:25:42.810084 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Aug 13 00:25:42.810185 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Aug 13 00:25:42.837234 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Aug 13 00:25:42.849008 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Aug 13 00:25:42.849120 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Aug 13 00:25:42.860605 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Aug 13 00:25:42.860689 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Aug 13 00:25:42.879589 systemd[1]: sysroot-boot.mount: Deactivated successfully. Aug 13 00:25:42.880389 systemd[1]: sysroot-boot.service: Deactivated successfully. Aug 13 00:25:42.880474 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Aug 13 00:25:42.891983 systemd[1]: initrd-cleanup.service: Deactivated successfully. Aug 13 00:25:42.892063 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Aug 13 00:25:44.109702 ignition[2017]: INFO : GET result: OK Aug 13 00:25:44.500732 ignition[2017]: INFO : Ignition finished successfully Aug 13 00:25:44.502997 systemd[1]: ignition-mount.service: Deactivated successfully. Aug 13 00:25:44.503209 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Aug 13 00:25:44.511019 systemd[1]: Stopped target network.target - Network. Aug 13 00:25:44.520358 systemd[1]: ignition-disks.service: Deactivated successfully. Aug 13 00:25:44.520419 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Aug 13 00:25:44.530127 systemd[1]: ignition-kargs.service: Deactivated successfully. Aug 13 00:25:44.530160 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Aug 13 00:25:44.539655 systemd[1]: ignition-setup.service: Deactivated successfully. Aug 13 00:25:44.539701 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Aug 13 00:25:44.549470 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Aug 13 00:25:44.549520 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Aug 13 00:25:44.559307 systemd[1]: initrd-setup-root.service: Deactivated successfully. Aug 13 00:25:44.559337 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Aug 13 00:25:44.569269 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Aug 13 00:25:44.577068 systemd-networkd[1687]: enP1p1s0f1np1: DHCPv6 lease lost Aug 13 00:25:44.578886 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Aug 13 00:25:44.585063 systemd-networkd[1687]: enP1p1s0f0np0: DHCPv6 lease lost Aug 13 00:25:44.590717 systemd[1]: systemd-resolved.service: Deactivated successfully. Aug 13 00:25:44.590897 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Aug 13 00:25:44.601026 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Aug 13 00:25:44.601112 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Aug 13 00:25:44.609040 systemd[1]: systemd-networkd.service: Deactivated successfully. Aug 13 00:25:44.609222 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Aug 13 00:25:44.619467 systemd[1]: systemd-networkd.socket: Deactivated successfully. Aug 13 00:25:44.619635 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Aug 13 00:25:44.640140 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Aug 13 00:25:44.648557 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Aug 13 00:25:44.648624 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Aug 13 00:25:44.658792 systemd[1]: systemd-sysctl.service: Deactivated successfully. Aug 13 00:25:44.658825 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Aug 13 00:25:44.669131 systemd[1]: systemd-modules-load.service: Deactivated successfully. Aug 13 00:25:44.669161 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Aug 13 00:25:44.679535 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Aug 13 00:25:44.702427 systemd[1]: systemd-udevd.service: Deactivated successfully. Aug 13 00:25:44.702569 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Aug 13 00:25:44.714285 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Aug 13 00:25:44.714439 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Aug 13 00:25:44.723757 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Aug 13 00:25:44.723824 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Aug 13 00:25:44.734546 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Aug 13 00:25:44.734584 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Aug 13 00:25:44.750802 systemd[1]: dracut-cmdline.service: Deactivated successfully. Aug 13 00:25:44.750868 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Aug 13 00:25:44.761999 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Aug 13 00:25:44.762049 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Aug 13 00:25:44.789198 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Aug 13 00:25:44.796356 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Aug 13 00:25:44.796418 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Aug 13 00:25:44.807652 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Aug 13 00:25:44.807700 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 00:25:44.819275 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Aug 13 00:25:44.819348 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Aug 13 00:25:45.349988 systemd[1]: network-cleanup.service: Deactivated successfully. Aug 13 00:25:45.350101 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Aug 13 00:25:45.361447 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Aug 13 00:25:45.384155 systemd[1]: Starting initrd-switch-root.service - Switch Root... Aug 13 00:25:45.393798 systemd[1]: Switching root. Aug 13 00:25:45.457680 systemd-journald[898]: Journal stopped Aug 13 00:25:21.163985 kernel: Booting Linux on physical CPU 0x0000120000 [0x413fd0c1] Aug 13 00:25:21.164008 kernel: Linux version 6.6.100-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT Tue Aug 12 22:21:53 -00 2025 Aug 13 00:25:21.164016 kernel: KASLR enabled Aug 13 00:25:21.164022 kernel: efi: EFI v2.7 by American Megatrends Aug 13 00:25:21.164028 kernel: efi: ACPI 2.0=0xec090000 SMBIOS 3.0=0xf0a1ff98 ESRT=0xea48e818 RNG=0xebf10018 MEMRESERVE=0xe4751f98 Aug 13 00:25:21.164033 kernel: random: crng init done Aug 13 00:25:21.164055 kernel: esrt: Reserving ESRT space from 0x00000000ea48e818 to 0x00000000ea48e878. Aug 13 00:25:21.164061 kernel: ACPI: Early table checksum verification disabled Aug 13 00:25:21.164069 kernel: ACPI: RSDP 0x00000000EC090000 000024 (v02 Ampere) Aug 13 00:25:21.164074 kernel: ACPI: XSDT 0x00000000EC080000 0000A4 (v01 Ampere Altra 00000000 AMI 01000013) Aug 13 00:25:21.164081 kernel: ACPI: FACP 0x00000000EC060000 000114 (v06 Ampere Altra 00000000 INTL 20190509) Aug 13 00:25:21.164087 kernel: ACPI: DSDT 0x00000000EC000000 019B57 (v02 Ampere Jade 00000001 INTL 20200717) Aug 13 00:25:21.164093 kernel: ACPI: DBG2 0x00000000EC070000 00005C (v00 Ampere Altra 00000000 INTL 20190509) Aug 13 00:25:21.164099 kernel: ACPI: GTDT 0x00000000EC050000 000110 (v03 Ampere Altra 00000000 INTL 20190509) Aug 13 00:25:21.164107 kernel: ACPI: SSDT 0x00000000EC040000 00002D (v02 Ampere Altra 00000001 INTL 20190509) Aug 13 00:25:21.164114 kernel: ACPI: FIDT 0x00000000EBFF0000 00009C (v01 ALASKA A M I 01072009 AMI 00010013) Aug 13 00:25:21.164120 kernel: ACPI: SPCR 0x00000000EBFE0000 000050 (v02 ALASKA A M I 01072009 AMI 0005000F) Aug 13 00:25:21.164126 kernel: ACPI: BGRT 0x00000000EBFD0000 000038 (v01 ALASKA A M I 01072009 AMI 00010013) Aug 13 00:25:21.164133 kernel: ACPI: MCFG 0x00000000EBFC0000 0000AC (v01 Ampere Altra 00000001 AMP. 01000013) Aug 13 00:25:21.164139 kernel: ACPI: IORT 0x00000000EBFB0000 000610 (v00 Ampere Altra 00000000 AMP. 01000013) Aug 13 00:25:21.164145 kernel: ACPI: PPTT 0x00000000EBF90000 006E60 (v02 Ampere Altra 00000000 AMP. 01000013) Aug 13 00:25:21.164152 kernel: ACPI: SLIT 0x00000000EBF80000 00002D (v01 Ampere Altra 00000000 AMP. 01000013) Aug 13 00:25:21.164158 kernel: ACPI: SRAT 0x00000000EBF70000 0006D0 (v03 Ampere Altra 00000000 AMP. 01000013) Aug 13 00:25:21.164166 kernel: ACPI: APIC 0x00000000EBFA0000 0019F4 (v05 Ampere Altra 00000003 AMI 01000013) Aug 13 00:25:21.164173 kernel: ACPI: PCCT 0x00000000EBF50000 000576 (v02 Ampere Altra 00000003 AMP. 01000013) Aug 13 00:25:21.164179 kernel: ACPI: WSMT 0x00000000EBF40000 000028 (v01 ALASKA A M I 01072009 AMI 00010013) Aug 13 00:25:21.164185 kernel: ACPI: FPDT 0x00000000EBF30000 000044 (v01 ALASKA A M I 01072009 AMI 01000013) Aug 13 00:25:21.164192 kernel: ACPI: SPCR: console: pl011,mmio32,0x100002600000,115200 Aug 13 00:25:21.164198 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x88300000-0x883fffff] Aug 13 00:25:21.164204 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x90000000-0xffffffff] Aug 13 00:25:21.164211 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x80000000000-0x8007fffffff] Aug 13 00:25:21.164217 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x80100000000-0x83fffffffff] Aug 13 00:25:21.164223 kernel: NUMA: NODE_DATA [mem 0x83fdffcd800-0x83fdffd2fff] Aug 13 00:25:21.164229 kernel: Zone ranges: Aug 13 00:25:21.164236 kernel: DMA [mem 0x0000000088300000-0x00000000ffffffff] Aug 13 00:25:21.164243 kernel: DMA32 empty Aug 13 00:25:21.164250 kernel: Normal [mem 0x0000000100000000-0x0000083fffffffff] Aug 13 00:25:21.164256 kernel: Movable zone start for each node Aug 13 00:25:21.164262 kernel: Early memory node ranges Aug 13 00:25:21.164268 kernel: node 0: [mem 0x0000000088300000-0x00000000883fffff] Aug 13 00:25:21.164277 kernel: node 0: [mem 0x0000000090000000-0x0000000091ffffff] Aug 13 00:25:21.164284 kernel: node 0: [mem 0x0000000092000000-0x0000000093ffffff] Aug 13 00:25:21.164292 kernel: node 0: [mem 0x0000000094000000-0x00000000eba47fff] Aug 13 00:25:21.164299 kernel: node 0: [mem 0x00000000eba48000-0x00000000ebedcfff] Aug 13 00:25:21.164305 kernel: node 0: [mem 0x00000000ebedd000-0x00000000ebeddfff] Aug 13 00:25:21.164312 kernel: node 0: [mem 0x00000000ebede000-0x00000000ebedffff] Aug 13 00:25:21.164319 kernel: node 0: [mem 0x00000000ebee0000-0x00000000ec0fffff] Aug 13 00:25:21.164325 kernel: node 0: [mem 0x00000000ec100000-0x00000000ec10ffff] Aug 13 00:25:21.164332 kernel: node 0: [mem 0x00000000ec110000-0x00000000ee54ffff] Aug 13 00:25:21.164338 kernel: node 0: [mem 0x00000000ee550000-0x00000000f765ffff] Aug 13 00:25:21.164345 kernel: node 0: [mem 0x00000000f7660000-0x00000000f784ffff] Aug 13 00:25:21.164352 kernel: node 0: [mem 0x00000000f7850000-0x00000000f7fdffff] Aug 13 00:25:21.164360 kernel: node 0: [mem 0x00000000f7fe0000-0x00000000ffc8efff] Aug 13 00:25:21.164367 kernel: node 0: [mem 0x00000000ffc8f000-0x00000000ffc8ffff] Aug 13 00:25:21.164373 kernel: node 0: [mem 0x00000000ffc90000-0x00000000ffffffff] Aug 13 00:25:21.164380 kernel: node 0: [mem 0x0000080000000000-0x000008007fffffff] Aug 13 00:25:21.164387 kernel: node 0: [mem 0x0000080100000000-0x0000083fffffffff] Aug 13 00:25:21.164393 kernel: Initmem setup node 0 [mem 0x0000000088300000-0x0000083fffffffff] Aug 13 00:25:21.164400 kernel: On node 0, zone DMA: 768 pages in unavailable ranges Aug 13 00:25:21.164407 kernel: On node 0, zone DMA: 31744 pages in unavailable ranges Aug 13 00:25:21.164413 kernel: psci: probing for conduit method from ACPI. Aug 13 00:25:21.164420 kernel: psci: PSCIv1.1 detected in firmware. Aug 13 00:25:21.164426 kernel: psci: Using standard PSCI v0.2 function IDs Aug 13 00:25:21.164434 kernel: psci: MIGRATE_INFO_TYPE not supported. Aug 13 00:25:21.164441 kernel: psci: SMC Calling Convention v1.2 Aug 13 00:25:21.164448 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x0 -> Node 0 Aug 13 00:25:21.164455 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x100 -> Node 0 Aug 13 00:25:21.164461 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x10000 -> Node 0 Aug 13 00:25:21.164468 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x10100 -> Node 0 Aug 13 00:25:21.164475 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x20000 -> Node 0 Aug 13 00:25:21.164481 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x20100 -> Node 0 Aug 13 00:25:21.164488 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x30000 -> Node 0 Aug 13 00:25:21.164494 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x30100 -> Node 0 Aug 13 00:25:21.164501 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x40000 -> Node 0 Aug 13 00:25:21.164507 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x40100 -> Node 0 Aug 13 00:25:21.164515 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x50000 -> Node 0 Aug 13 00:25:21.164522 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x50100 -> Node 0 Aug 13 00:25:21.164529 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x60000 -> Node 0 Aug 13 00:25:21.164535 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x60100 -> Node 0 Aug 13 00:25:21.164542 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x70000 -> Node 0 Aug 13 00:25:21.164548 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x70100 -> Node 0 Aug 13 00:25:21.164555 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x80000 -> Node 0 Aug 13 00:25:21.164562 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x80100 -> Node 0 Aug 13 00:25:21.164568 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x90000 -> Node 0 Aug 13 00:25:21.164575 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x90100 -> Node 0 Aug 13 00:25:21.164582 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0xa0000 -> Node 0 Aug 13 00:25:21.164588 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0xa0100 -> Node 0 Aug 13 00:25:21.164596 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0xb0000 -> Node 0 Aug 13 00:25:21.164603 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0xb0100 -> Node 0 Aug 13 00:25:21.164609 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0xc0000 -> Node 0 Aug 13 00:25:21.164616 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0xc0100 -> Node 0 Aug 13 00:25:21.164623 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0xd0000 -> Node 0 Aug 13 00:25:21.164629 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0xd0100 -> Node 0 Aug 13 00:25:21.164636 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0xe0000 -> Node 0 Aug 13 00:25:21.164642 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0xe0100 -> Node 0 Aug 13 00:25:21.164649 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0xf0000 -> Node 0 Aug 13 00:25:21.164655 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0xf0100 -> Node 0 Aug 13 00:25:21.164662 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x100000 -> Node 0 Aug 13 00:25:21.164669 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x100100 -> Node 0 Aug 13 00:25:21.164676 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x110000 -> Node 0 Aug 13 00:25:21.164683 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x110100 -> Node 0 Aug 13 00:25:21.164689 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x120000 -> Node 0 Aug 13 00:25:21.164696 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x120100 -> Node 0 Aug 13 00:25:21.164703 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x130000 -> Node 0 Aug 13 00:25:21.164709 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x130100 -> Node 0 Aug 13 00:25:21.164716 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x140000 -> Node 0 Aug 13 00:25:21.164722 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x140100 -> Node 0 Aug 13 00:25:21.164729 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x150000 -> Node 0 Aug 13 00:25:21.164735 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x150100 -> Node 0 Aug 13 00:25:21.164742 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x160000 -> Node 0 Aug 13 00:25:21.164750 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x160100 -> Node 0 Aug 13 00:25:21.164757 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x170000 -> Node 0 Aug 13 00:25:21.164763 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x170100 -> Node 0 Aug 13 00:25:21.164770 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x180000 -> Node 0 Aug 13 00:25:21.164777 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x180100 -> Node 0 Aug 13 00:25:21.164783 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x190000 -> Node 0 Aug 13 00:25:21.164790 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x190100 -> Node 0 Aug 13 00:25:21.164797 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1a0000 -> Node 0 Aug 13 00:25:21.164809 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1a0100 -> Node 0 Aug 13 00:25:21.164816 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1b0000 -> Node 0 Aug 13 00:25:21.164825 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1b0100 -> Node 0 Aug 13 00:25:21.164832 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1c0000 -> Node 0 Aug 13 00:25:21.164839 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1c0100 -> Node 0 Aug 13 00:25:21.164846 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1d0000 -> Node 0 Aug 13 00:25:21.164853 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1d0100 -> Node 0 Aug 13 00:25:21.164860 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1e0000 -> Node 0 Aug 13 00:25:21.164869 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1e0100 -> Node 0 Aug 13 00:25:21.164876 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1f0000 -> Node 0 Aug 13 00:25:21.164883 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1f0100 -> Node 0 Aug 13 00:25:21.164889 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x200000 -> Node 0 Aug 13 00:25:21.164897 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x200100 -> Node 0 Aug 13 00:25:21.164904 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x210000 -> Node 0 Aug 13 00:25:21.164911 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x210100 -> Node 0 Aug 13 00:25:21.164918 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x220000 -> Node 0 Aug 13 00:25:21.164925 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x220100 -> Node 0 Aug 13 00:25:21.164932 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x230000 -> Node 0 Aug 13 00:25:21.164939 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x230100 -> Node 0 Aug 13 00:25:21.164947 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x240000 -> Node 0 Aug 13 00:25:21.164955 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x240100 -> Node 0 Aug 13 00:25:21.164962 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x250000 -> Node 0 Aug 13 00:25:21.164969 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x250100 -> Node 0 Aug 13 00:25:21.164976 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x260000 -> Node 0 Aug 13 00:25:21.164983 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x260100 -> Node 0 Aug 13 00:25:21.164990 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x270000 -> Node 0 Aug 13 00:25:21.164997 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x270100 -> Node 0 Aug 13 00:25:21.165004 kernel: percpu: Embedded 31 pages/cpu s86696 r8192 d32088 u126976 Aug 13 00:25:21.165011 kernel: pcpu-alloc: s86696 r8192 d32088 u126976 alloc=31*4096 Aug 13 00:25:21.165019 kernel: pcpu-alloc: [0] 00 [0] 01 [0] 02 [0] 03 [0] 04 [0] 05 [0] 06 [0] 07 Aug 13 00:25:21.165026 kernel: pcpu-alloc: [0] 08 [0] 09 [0] 10 [0] 11 [0] 12 [0] 13 [0] 14 [0] 15 Aug 13 00:25:21.165034 kernel: pcpu-alloc: [0] 16 [0] 17 [0] 18 [0] 19 [0] 20 [0] 21 [0] 22 [0] 23 Aug 13 00:25:21.165073 kernel: pcpu-alloc: [0] 24 [0] 25 [0] 26 [0] 27 [0] 28 [0] 29 [0] 30 [0] 31 Aug 13 00:25:21.165080 kernel: pcpu-alloc: [0] 32 [0] 33 [0] 34 [0] 35 [0] 36 [0] 37 [0] 38 [0] 39 Aug 13 00:25:21.165088 kernel: pcpu-alloc: [0] 40 [0] 41 [0] 42 [0] 43 [0] 44 [0] 45 [0] 46 [0] 47 Aug 13 00:25:21.165095 kernel: pcpu-alloc: [0] 48 [0] 49 [0] 50 [0] 51 [0] 52 [0] 53 [0] 54 [0] 55 Aug 13 00:25:21.165102 kernel: pcpu-alloc: [0] 56 [0] 57 [0] 58 [0] 59 [0] 60 [0] 61 [0] 62 [0] 63 Aug 13 00:25:21.165109 kernel: pcpu-alloc: [0] 64 [0] 65 [0] 66 [0] 67 [0] 68 [0] 69 [0] 70 [0] 71 Aug 13 00:25:21.165116 kernel: pcpu-alloc: [0] 72 [0] 73 [0] 74 [0] 75 [0] 76 [0] 77 [0] 78 [0] 79 Aug 13 00:25:21.165123 kernel: Detected PIPT I-cache on CPU0 Aug 13 00:25:21.165130 kernel: CPU features: detected: GIC system register CPU interface Aug 13 00:25:21.165137 kernel: CPU features: detected: Virtualization Host Extensions Aug 13 00:25:21.165146 kernel: CPU features: detected: Hardware dirty bit management Aug 13 00:25:21.165153 kernel: CPU features: detected: Spectre-v4 Aug 13 00:25:21.165160 kernel: CPU features: detected: Spectre-BHB Aug 13 00:25:21.165168 kernel: CPU features: kernel page table isolation forced ON by KASLR Aug 13 00:25:21.165175 kernel: CPU features: detected: Kernel page table isolation (KPTI) Aug 13 00:25:21.165182 kernel: CPU features: detected: ARM erratum 1418040 Aug 13 00:25:21.165189 kernel: CPU features: detected: SSBS not fully self-synchronizing Aug 13 00:25:21.165196 kernel: alternatives: applying boot alternatives Aug 13 00:25:21.165205 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty0 console=ttyS1,115200n8 flatcar.first_boot=detected acpi=force flatcar.oem.id=packet flatcar.autologin verity.usrhash=2f9df6e9e6c671c457040a64675390bbff42294b08c628cd2dc472ed8120146a Aug 13 00:25:21.165212 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Aug 13 00:25:21.165221 kernel: printk: log_buf_len individual max cpu contribution: 4096 bytes Aug 13 00:25:21.165228 kernel: printk: log_buf_len total cpu_extra contributions: 323584 bytes Aug 13 00:25:21.165235 kernel: printk: log_buf_len min size: 262144 bytes Aug 13 00:25:21.165242 kernel: printk: log_buf_len: 1048576 bytes Aug 13 00:25:21.165249 kernel: printk: early log buf free: 250024(95%) Aug 13 00:25:21.165256 kernel: Dentry cache hash table entries: 16777216 (order: 15, 134217728 bytes, linear) Aug 13 00:25:21.165263 kernel: Inode-cache hash table entries: 8388608 (order: 14, 67108864 bytes, linear) Aug 13 00:25:21.165270 kernel: Fallback order for Node 0: 0 Aug 13 00:25:21.165277 kernel: Built 1 zonelists, mobility grouping on. Total pages: 65996028 Aug 13 00:25:21.165284 kernel: Policy zone: Normal Aug 13 00:25:21.165291 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Aug 13 00:25:21.165298 kernel: software IO TLB: area num 128. Aug 13 00:25:21.165307 kernel: software IO TLB: mapped [mem 0x00000000fbc8f000-0x00000000ffc8f000] (64MB) Aug 13 00:25:21.165314 kernel: Memory: 262922400K/268174336K available (10304K kernel code, 2186K rwdata, 8108K rodata, 39424K init, 897K bss, 5251936K reserved, 0K cma-reserved) Aug 13 00:25:21.165321 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=80, Nodes=1 Aug 13 00:25:21.165328 kernel: rcu: Preemptible hierarchical RCU implementation. Aug 13 00:25:21.165336 kernel: rcu: RCU event tracing is enabled. Aug 13 00:25:21.165343 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=80. Aug 13 00:25:21.165350 kernel: Trampoline variant of Tasks RCU enabled. Aug 13 00:25:21.165358 kernel: Tracing variant of Tasks RCU enabled. Aug 13 00:25:21.165365 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Aug 13 00:25:21.165372 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=80 Aug 13 00:25:21.165379 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Aug 13 00:25:21.165388 kernel: GICv3: GIC: Using split EOI/Deactivate mode Aug 13 00:25:21.165395 kernel: GICv3: 672 SPIs implemented Aug 13 00:25:21.165402 kernel: GICv3: 0 Extended SPIs implemented Aug 13 00:25:21.165409 kernel: Root IRQ handler: gic_handle_irq Aug 13 00:25:21.165416 kernel: GICv3: GICv3 features: 16 PPIs Aug 13 00:25:21.165423 kernel: GICv3: CPU0: found redistributor 120000 region 0:0x00001001005c0000 Aug 13 00:25:21.165430 kernel: SRAT: PXM 0 -> ITS 0 -> Node 0 Aug 13 00:25:21.165437 kernel: SRAT: PXM 0 -> ITS 1 -> Node 0 Aug 13 00:25:21.165444 kernel: SRAT: PXM 0 -> ITS 2 -> Node 0 Aug 13 00:25:21.165451 kernel: SRAT: PXM 0 -> ITS 3 -> Node 0 Aug 13 00:25:21.165458 kernel: SRAT: PXM 0 -> ITS 4 -> Node 0 Aug 13 00:25:21.165465 kernel: SRAT: PXM 0 -> ITS 5 -> Node 0 Aug 13 00:25:21.165472 kernel: SRAT: PXM 0 -> ITS 6 -> Node 0 Aug 13 00:25:21.165481 kernel: SRAT: PXM 0 -> ITS 7 -> Node 0 Aug 13 00:25:21.165488 kernel: ITS [mem 0x100100040000-0x10010005ffff] Aug 13 00:25:21.165495 kernel: ITS@0x0000100100040000: allocated 8192 Devices @80000270000 (indirect, esz 8, psz 64K, shr 1) Aug 13 00:25:21.165502 kernel: ITS@0x0000100100040000: allocated 32768 Interrupt Collections @80000280000 (flat, esz 2, psz 64K, shr 1) Aug 13 00:25:21.165510 kernel: ITS [mem 0x100100060000-0x10010007ffff] Aug 13 00:25:21.165517 kernel: ITS@0x0000100100060000: allocated 8192 Devices @800002a0000 (indirect, esz 8, psz 64K, shr 1) Aug 13 00:25:21.165524 kernel: ITS@0x0000100100060000: allocated 32768 Interrupt Collections @800002b0000 (flat, esz 2, psz 64K, shr 1) Aug 13 00:25:21.165532 kernel: ITS [mem 0x100100080000-0x10010009ffff] Aug 13 00:25:21.165539 kernel: ITS@0x0000100100080000: allocated 8192 Devices @800002d0000 (indirect, esz 8, psz 64K, shr 1) Aug 13 00:25:21.165546 kernel: ITS@0x0000100100080000: allocated 32768 Interrupt Collections @800002e0000 (flat, esz 2, psz 64K, shr 1) Aug 13 00:25:21.165553 kernel: ITS [mem 0x1001000a0000-0x1001000bffff] Aug 13 00:25:21.165562 kernel: ITS@0x00001001000a0000: allocated 8192 Devices @80000300000 (indirect, esz 8, psz 64K, shr 1) Aug 13 00:25:21.165569 kernel: ITS@0x00001001000a0000: allocated 32768 Interrupt Collections @80000310000 (flat, esz 2, psz 64K, shr 1) Aug 13 00:25:21.165576 kernel: ITS [mem 0x1001000c0000-0x1001000dffff] Aug 13 00:25:21.165583 kernel: ITS@0x00001001000c0000: allocated 8192 Devices @80000330000 (indirect, esz 8, psz 64K, shr 1) Aug 13 00:25:21.165590 kernel: ITS@0x00001001000c0000: allocated 32768 Interrupt Collections @80000340000 (flat, esz 2, psz 64K, shr 1) Aug 13 00:25:21.165597 kernel: ITS [mem 0x1001000e0000-0x1001000fffff] Aug 13 00:25:21.165604 kernel: ITS@0x00001001000e0000: allocated 8192 Devices @80000360000 (indirect, esz 8, psz 64K, shr 1) Aug 13 00:25:21.165612 kernel: ITS@0x00001001000e0000: allocated 32768 Interrupt Collections @80000370000 (flat, esz 2, psz 64K, shr 1) Aug 13 00:25:21.165619 kernel: ITS [mem 0x100100100000-0x10010011ffff] Aug 13 00:25:21.165626 kernel: ITS@0x0000100100100000: allocated 8192 Devices @80000390000 (indirect, esz 8, psz 64K, shr 1) Aug 13 00:25:21.165633 kernel: ITS@0x0000100100100000: allocated 32768 Interrupt Collections @800003a0000 (flat, esz 2, psz 64K, shr 1) Aug 13 00:25:21.165642 kernel: ITS [mem 0x100100120000-0x10010013ffff] Aug 13 00:25:21.165649 kernel: ITS@0x0000100100120000: allocated 8192 Devices @800003c0000 (indirect, esz 8, psz 64K, shr 1) Aug 13 00:25:21.165656 kernel: ITS@0x0000100100120000: allocated 32768 Interrupt Collections @800003d0000 (flat, esz 2, psz 64K, shr 1) Aug 13 00:25:21.165663 kernel: GICv3: using LPI property table @0x00000800003e0000 Aug 13 00:25:21.165670 kernel: GICv3: CPU0: using allocated LPI pending table @0x00000800003f0000 Aug 13 00:25:21.165678 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Aug 13 00:25:21.165685 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.165692 kernel: ACPI GTDT: found 1 memory-mapped timer block(s). Aug 13 00:25:21.165700 kernel: arch_timer: cp15 and mmio timer(s) running at 25.00MHz (phys/phys). Aug 13 00:25:21.165707 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns Aug 13 00:25:21.165714 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns Aug 13 00:25:21.165723 kernel: Console: colour dummy device 80x25 Aug 13 00:25:21.165730 kernel: printk: console [tty0] enabled Aug 13 00:25:21.165737 kernel: ACPI: Core revision 20230628 Aug 13 00:25:21.165745 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000) Aug 13 00:25:21.165752 kernel: pid_max: default: 81920 minimum: 640 Aug 13 00:25:21.165759 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Aug 13 00:25:21.165767 kernel: landlock: Up and running. Aug 13 00:25:21.165774 kernel: SELinux: Initializing. Aug 13 00:25:21.165781 kernel: Mount-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Aug 13 00:25:21.165790 kernel: Mountpoint-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Aug 13 00:25:21.165797 kernel: RCU Tasks: Setting shift to 7 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=80. Aug 13 00:25:21.165805 kernel: RCU Tasks Trace: Setting shift to 7 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=80. Aug 13 00:25:21.165812 kernel: rcu: Hierarchical SRCU implementation. Aug 13 00:25:21.165820 kernel: rcu: Max phase no-delay instances is 400. Aug 13 00:25:21.165827 kernel: Platform MSI: ITS@0x100100040000 domain created Aug 13 00:25:21.165834 kernel: Platform MSI: ITS@0x100100060000 domain created Aug 13 00:25:21.165841 kernel: Platform MSI: ITS@0x100100080000 domain created Aug 13 00:25:21.165849 kernel: Platform MSI: ITS@0x1001000a0000 domain created Aug 13 00:25:21.165857 kernel: Platform MSI: ITS@0x1001000c0000 domain created Aug 13 00:25:21.165864 kernel: Platform MSI: ITS@0x1001000e0000 domain created Aug 13 00:25:21.165872 kernel: Platform MSI: ITS@0x100100100000 domain created Aug 13 00:25:21.165879 kernel: Platform MSI: ITS@0x100100120000 domain created Aug 13 00:25:21.165886 kernel: PCI/MSI: ITS@0x100100040000 domain created Aug 13 00:25:21.165893 kernel: PCI/MSI: ITS@0x100100060000 domain created Aug 13 00:25:21.165901 kernel: PCI/MSI: ITS@0x100100080000 domain created Aug 13 00:25:21.165908 kernel: PCI/MSI: ITS@0x1001000a0000 domain created Aug 13 00:25:21.165915 kernel: PCI/MSI: ITS@0x1001000c0000 domain created Aug 13 00:25:21.165922 kernel: PCI/MSI: ITS@0x1001000e0000 domain created Aug 13 00:25:21.165931 kernel: PCI/MSI: ITS@0x100100100000 domain created Aug 13 00:25:21.165938 kernel: PCI/MSI: ITS@0x100100120000 domain created Aug 13 00:25:21.165946 kernel: Remapping and enabling EFI services. Aug 13 00:25:21.165953 kernel: smp: Bringing up secondary CPUs ... Aug 13 00:25:21.165960 kernel: Detected PIPT I-cache on CPU1 Aug 13 00:25:21.165967 kernel: GICv3: CPU1: found redistributor 1a0000 region 0:0x00001001007c0000 Aug 13 00:25:21.165975 kernel: GICv3: CPU1: using allocated LPI pending table @0x0000080000800000 Aug 13 00:25:21.165982 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.165989 kernel: CPU1: Booted secondary processor 0x00001a0000 [0x413fd0c1] Aug 13 00:25:21.165998 kernel: Detected PIPT I-cache on CPU2 Aug 13 00:25:21.166005 kernel: GICv3: CPU2: found redistributor 140000 region 0:0x0000100100640000 Aug 13 00:25:21.166012 kernel: GICv3: CPU2: using allocated LPI pending table @0x0000080000810000 Aug 13 00:25:21.166019 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166026 kernel: CPU2: Booted secondary processor 0x0000140000 [0x413fd0c1] Aug 13 00:25:21.166034 kernel: Detected PIPT I-cache on CPU3 Aug 13 00:25:21.166043 kernel: GICv3: CPU3: found redistributor 1c0000 region 0:0x0000100100840000 Aug 13 00:25:21.166050 kernel: GICv3: CPU3: using allocated LPI pending table @0x0000080000820000 Aug 13 00:25:21.166058 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166065 kernel: CPU3: Booted secondary processor 0x00001c0000 [0x413fd0c1] Aug 13 00:25:21.166074 kernel: Detected PIPT I-cache on CPU4 Aug 13 00:25:21.166081 kernel: GICv3: CPU4: found redistributor 100000 region 0:0x0000100100540000 Aug 13 00:25:21.166089 kernel: GICv3: CPU4: using allocated LPI pending table @0x0000080000830000 Aug 13 00:25:21.166096 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166103 kernel: CPU4: Booted secondary processor 0x0000100000 [0x413fd0c1] Aug 13 00:25:21.166110 kernel: Detected PIPT I-cache on CPU5 Aug 13 00:25:21.166117 kernel: GICv3: CPU5: found redistributor 180000 region 0:0x0000100100740000 Aug 13 00:25:21.166124 kernel: GICv3: CPU5: using allocated LPI pending table @0x0000080000840000 Aug 13 00:25:21.166132 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166140 kernel: CPU5: Booted secondary processor 0x0000180000 [0x413fd0c1] Aug 13 00:25:21.166148 kernel: Detected PIPT I-cache on CPU6 Aug 13 00:25:21.166155 kernel: GICv3: CPU6: found redistributor 160000 region 0:0x00001001006c0000 Aug 13 00:25:21.166162 kernel: GICv3: CPU6: using allocated LPI pending table @0x0000080000850000 Aug 13 00:25:21.166169 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166176 kernel: CPU6: Booted secondary processor 0x0000160000 [0x413fd0c1] Aug 13 00:25:21.166184 kernel: Detected PIPT I-cache on CPU7 Aug 13 00:25:21.166191 kernel: GICv3: CPU7: found redistributor 1e0000 region 0:0x00001001008c0000 Aug 13 00:25:21.166198 kernel: GICv3: CPU7: using allocated LPI pending table @0x0000080000860000 Aug 13 00:25:21.166207 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166215 kernel: CPU7: Booted secondary processor 0x00001e0000 [0x413fd0c1] Aug 13 00:25:21.166222 kernel: Detected PIPT I-cache on CPU8 Aug 13 00:25:21.166229 kernel: GICv3: CPU8: found redistributor a0000 region 0:0x00001001003c0000 Aug 13 00:25:21.166236 kernel: GICv3: CPU8: using allocated LPI pending table @0x0000080000870000 Aug 13 00:25:21.166244 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166251 kernel: CPU8: Booted secondary processor 0x00000a0000 [0x413fd0c1] Aug 13 00:25:21.166258 kernel: Detected PIPT I-cache on CPU9 Aug 13 00:25:21.166265 kernel: GICv3: CPU9: found redistributor 220000 region 0:0x00001001009c0000 Aug 13 00:25:21.166272 kernel: GICv3: CPU9: using allocated LPI pending table @0x0000080000880000 Aug 13 00:25:21.166281 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166288 kernel: CPU9: Booted secondary processor 0x0000220000 [0x413fd0c1] Aug 13 00:25:21.166295 kernel: Detected PIPT I-cache on CPU10 Aug 13 00:25:21.166303 kernel: GICv3: CPU10: found redistributor c0000 region 0:0x0000100100440000 Aug 13 00:25:21.166310 kernel: GICv3: CPU10: using allocated LPI pending table @0x0000080000890000 Aug 13 00:25:21.166317 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166324 kernel: CPU10: Booted secondary processor 0x00000c0000 [0x413fd0c1] Aug 13 00:25:21.166331 kernel: Detected PIPT I-cache on CPU11 Aug 13 00:25:21.166338 kernel: GICv3: CPU11: found redistributor 240000 region 0:0x0000100100a40000 Aug 13 00:25:21.166346 kernel: GICv3: CPU11: using allocated LPI pending table @0x00000800008a0000 Aug 13 00:25:21.166354 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166361 kernel: CPU11: Booted secondary processor 0x0000240000 [0x413fd0c1] Aug 13 00:25:21.166368 kernel: Detected PIPT I-cache on CPU12 Aug 13 00:25:21.166375 kernel: GICv3: CPU12: found redistributor 80000 region 0:0x0000100100340000 Aug 13 00:25:21.166383 kernel: GICv3: CPU12: using allocated LPI pending table @0x00000800008b0000 Aug 13 00:25:21.166390 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166397 kernel: CPU12: Booted secondary processor 0x0000080000 [0x413fd0c1] Aug 13 00:25:21.166404 kernel: Detected PIPT I-cache on CPU13 Aug 13 00:25:21.166411 kernel: GICv3: CPU13: found redistributor 200000 region 0:0x0000100100940000 Aug 13 00:25:21.166420 kernel: GICv3: CPU13: using allocated LPI pending table @0x00000800008c0000 Aug 13 00:25:21.166428 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166435 kernel: CPU13: Booted secondary processor 0x0000200000 [0x413fd0c1] Aug 13 00:25:21.166442 kernel: Detected PIPT I-cache on CPU14 Aug 13 00:25:21.166449 kernel: GICv3: CPU14: found redistributor e0000 region 0:0x00001001004c0000 Aug 13 00:25:21.166457 kernel: GICv3: CPU14: using allocated LPI pending table @0x00000800008d0000 Aug 13 00:25:21.166464 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166471 kernel: CPU14: Booted secondary processor 0x00000e0000 [0x413fd0c1] Aug 13 00:25:21.166478 kernel: Detected PIPT I-cache on CPU15 Aug 13 00:25:21.166487 kernel: GICv3: CPU15: found redistributor 260000 region 0:0x0000100100ac0000 Aug 13 00:25:21.166494 kernel: GICv3: CPU15: using allocated LPI pending table @0x00000800008e0000 Aug 13 00:25:21.166502 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166509 kernel: CPU15: Booted secondary processor 0x0000260000 [0x413fd0c1] Aug 13 00:25:21.166516 kernel: Detected PIPT I-cache on CPU16 Aug 13 00:25:21.166523 kernel: GICv3: CPU16: found redistributor 20000 region 0:0x00001001001c0000 Aug 13 00:25:21.166530 kernel: GICv3: CPU16: using allocated LPI pending table @0x00000800008f0000 Aug 13 00:25:21.166537 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166545 kernel: CPU16: Booted secondary processor 0x0000020000 [0x413fd0c1] Aug 13 00:25:21.166552 kernel: Detected PIPT I-cache on CPU17 Aug 13 00:25:21.166568 kernel: GICv3: CPU17: found redistributor 40000 region 0:0x0000100100240000 Aug 13 00:25:21.166577 kernel: GICv3: CPU17: using allocated LPI pending table @0x0000080000900000 Aug 13 00:25:21.166585 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166592 kernel: CPU17: Booted secondary processor 0x0000040000 [0x413fd0c1] Aug 13 00:25:21.166600 kernel: Detected PIPT I-cache on CPU18 Aug 13 00:25:21.166607 kernel: GICv3: CPU18: found redistributor 0 region 0:0x0000100100140000 Aug 13 00:25:21.166615 kernel: GICv3: CPU18: using allocated LPI pending table @0x0000080000910000 Aug 13 00:25:21.166622 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166630 kernel: CPU18: Booted secondary processor 0x0000000000 [0x413fd0c1] Aug 13 00:25:21.166639 kernel: Detected PIPT I-cache on CPU19 Aug 13 00:25:21.166646 kernel: GICv3: CPU19: found redistributor 60000 region 0:0x00001001002c0000 Aug 13 00:25:21.166654 kernel: GICv3: CPU19: using allocated LPI pending table @0x0000080000920000 Aug 13 00:25:21.166662 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166669 kernel: CPU19: Booted secondary processor 0x0000060000 [0x413fd0c1] Aug 13 00:25:21.166677 kernel: Detected PIPT I-cache on CPU20 Aug 13 00:25:21.166686 kernel: GICv3: CPU20: found redistributor 130000 region 0:0x0000100100600000 Aug 13 00:25:21.166695 kernel: GICv3: CPU20: using allocated LPI pending table @0x0000080000930000 Aug 13 00:25:21.166703 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166710 kernel: CPU20: Booted secondary processor 0x0000130000 [0x413fd0c1] Aug 13 00:25:21.166718 kernel: Detected PIPT I-cache on CPU21 Aug 13 00:25:21.166725 kernel: GICv3: CPU21: found redistributor 1b0000 region 0:0x0000100100800000 Aug 13 00:25:21.166734 kernel: GICv3: CPU21: using allocated LPI pending table @0x0000080000940000 Aug 13 00:25:21.166742 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166750 kernel: CPU21: Booted secondary processor 0x00001b0000 [0x413fd0c1] Aug 13 00:25:21.166757 kernel: Detected PIPT I-cache on CPU22 Aug 13 00:25:21.166766 kernel: GICv3: CPU22: found redistributor 150000 region 0:0x0000100100680000 Aug 13 00:25:21.166774 kernel: GICv3: CPU22: using allocated LPI pending table @0x0000080000950000 Aug 13 00:25:21.166781 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166789 kernel: CPU22: Booted secondary processor 0x0000150000 [0x413fd0c1] Aug 13 00:25:21.166797 kernel: Detected PIPT I-cache on CPU23 Aug 13 00:25:21.166804 kernel: GICv3: CPU23: found redistributor 1d0000 region 0:0x0000100100880000 Aug 13 00:25:21.166812 kernel: GICv3: CPU23: using allocated LPI pending table @0x0000080000960000 Aug 13 00:25:21.166820 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166827 kernel: CPU23: Booted secondary processor 0x00001d0000 [0x413fd0c1] Aug 13 00:25:21.166836 kernel: Detected PIPT I-cache on CPU24 Aug 13 00:25:21.166844 kernel: GICv3: CPU24: found redistributor 110000 region 0:0x0000100100580000 Aug 13 00:25:21.166851 kernel: GICv3: CPU24: using allocated LPI pending table @0x0000080000970000 Aug 13 00:25:21.166859 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166866 kernel: CPU24: Booted secondary processor 0x0000110000 [0x413fd0c1] Aug 13 00:25:21.166874 kernel: Detected PIPT I-cache on CPU25 Aug 13 00:25:21.166881 kernel: GICv3: CPU25: found redistributor 190000 region 0:0x0000100100780000 Aug 13 00:25:21.166889 kernel: GICv3: CPU25: using allocated LPI pending table @0x0000080000980000 Aug 13 00:25:21.166896 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166905 kernel: CPU25: Booted secondary processor 0x0000190000 [0x413fd0c1] Aug 13 00:25:21.166913 kernel: Detected PIPT I-cache on CPU26 Aug 13 00:25:21.166921 kernel: GICv3: CPU26: found redistributor 170000 region 0:0x0000100100700000 Aug 13 00:25:21.166928 kernel: GICv3: CPU26: using allocated LPI pending table @0x0000080000990000 Aug 13 00:25:21.166936 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166943 kernel: CPU26: Booted secondary processor 0x0000170000 [0x413fd0c1] Aug 13 00:25:21.166951 kernel: Detected PIPT I-cache on CPU27 Aug 13 00:25:21.166958 kernel: GICv3: CPU27: found redistributor 1f0000 region 0:0x0000100100900000 Aug 13 00:25:21.166966 kernel: GICv3: CPU27: using allocated LPI pending table @0x00000800009a0000 Aug 13 00:25:21.166974 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.166983 kernel: CPU27: Booted secondary processor 0x00001f0000 [0x413fd0c1] Aug 13 00:25:21.166990 kernel: Detected PIPT I-cache on CPU28 Aug 13 00:25:21.166998 kernel: GICv3: CPU28: found redistributor b0000 region 0:0x0000100100400000 Aug 13 00:25:21.167006 kernel: GICv3: CPU28: using allocated LPI pending table @0x00000800009b0000 Aug 13 00:25:21.167013 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167021 kernel: CPU28: Booted secondary processor 0x00000b0000 [0x413fd0c1] Aug 13 00:25:21.167028 kernel: Detected PIPT I-cache on CPU29 Aug 13 00:25:21.167036 kernel: GICv3: CPU29: found redistributor 230000 region 0:0x0000100100a00000 Aug 13 00:25:21.167065 kernel: GICv3: CPU29: using allocated LPI pending table @0x00000800009c0000 Aug 13 00:25:21.167075 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167083 kernel: CPU29: Booted secondary processor 0x0000230000 [0x413fd0c1] Aug 13 00:25:21.167090 kernel: Detected PIPT I-cache on CPU30 Aug 13 00:25:21.167098 kernel: GICv3: CPU30: found redistributor d0000 region 0:0x0000100100480000 Aug 13 00:25:21.167105 kernel: GICv3: CPU30: using allocated LPI pending table @0x00000800009d0000 Aug 13 00:25:21.167113 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167120 kernel: CPU30: Booted secondary processor 0x00000d0000 [0x413fd0c1] Aug 13 00:25:21.167128 kernel: Detected PIPT I-cache on CPU31 Aug 13 00:25:21.167135 kernel: GICv3: CPU31: found redistributor 250000 region 0:0x0000100100a80000 Aug 13 00:25:21.167143 kernel: GICv3: CPU31: using allocated LPI pending table @0x00000800009e0000 Aug 13 00:25:21.167152 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167160 kernel: CPU31: Booted secondary processor 0x0000250000 [0x413fd0c1] Aug 13 00:25:21.167168 kernel: Detected PIPT I-cache on CPU32 Aug 13 00:25:21.167175 kernel: GICv3: CPU32: found redistributor 90000 region 0:0x0000100100380000 Aug 13 00:25:21.167183 kernel: GICv3: CPU32: using allocated LPI pending table @0x00000800009f0000 Aug 13 00:25:21.167191 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167198 kernel: CPU32: Booted secondary processor 0x0000090000 [0x413fd0c1] Aug 13 00:25:21.167206 kernel: Detected PIPT I-cache on CPU33 Aug 13 00:25:21.167213 kernel: GICv3: CPU33: found redistributor 210000 region 0:0x0000100100980000 Aug 13 00:25:21.167222 kernel: GICv3: CPU33: using allocated LPI pending table @0x0000080000a00000 Aug 13 00:25:21.167230 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167237 kernel: CPU33: Booted secondary processor 0x0000210000 [0x413fd0c1] Aug 13 00:25:21.167245 kernel: Detected PIPT I-cache on CPU34 Aug 13 00:25:21.167253 kernel: GICv3: CPU34: found redistributor f0000 region 0:0x0000100100500000 Aug 13 00:25:21.167260 kernel: GICv3: CPU34: using allocated LPI pending table @0x0000080000a10000 Aug 13 00:25:21.167268 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167277 kernel: CPU34: Booted secondary processor 0x00000f0000 [0x413fd0c1] Aug 13 00:25:21.167285 kernel: Detected PIPT I-cache on CPU35 Aug 13 00:25:21.167292 kernel: GICv3: CPU35: found redistributor 270000 region 0:0x0000100100b00000 Aug 13 00:25:21.167302 kernel: GICv3: CPU35: using allocated LPI pending table @0x0000080000a20000 Aug 13 00:25:21.167309 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167317 kernel: CPU35: Booted secondary processor 0x0000270000 [0x413fd0c1] Aug 13 00:25:21.167324 kernel: Detected PIPT I-cache on CPU36 Aug 13 00:25:21.167332 kernel: GICv3: CPU36: found redistributor 30000 region 0:0x0000100100200000 Aug 13 00:25:21.167340 kernel: GICv3: CPU36: using allocated LPI pending table @0x0000080000a30000 Aug 13 00:25:21.167347 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167355 kernel: CPU36: Booted secondary processor 0x0000030000 [0x413fd0c1] Aug 13 00:25:21.167362 kernel: Detected PIPT I-cache on CPU37 Aug 13 00:25:21.167371 kernel: GICv3: CPU37: found redistributor 50000 region 0:0x0000100100280000 Aug 13 00:25:21.167379 kernel: GICv3: CPU37: using allocated LPI pending table @0x0000080000a40000 Aug 13 00:25:21.167387 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167394 kernel: CPU37: Booted secondary processor 0x0000050000 [0x413fd0c1] Aug 13 00:25:21.167402 kernel: Detected PIPT I-cache on CPU38 Aug 13 00:25:21.167409 kernel: GICv3: CPU38: found redistributor 10000 region 0:0x0000100100180000 Aug 13 00:25:21.167417 kernel: GICv3: CPU38: using allocated LPI pending table @0x0000080000a50000 Aug 13 00:25:21.167425 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167432 kernel: CPU38: Booted secondary processor 0x0000010000 [0x413fd0c1] Aug 13 00:25:21.167440 kernel: Detected PIPT I-cache on CPU39 Aug 13 00:25:21.167449 kernel: GICv3: CPU39: found redistributor 70000 region 0:0x0000100100300000 Aug 13 00:25:21.167456 kernel: GICv3: CPU39: using allocated LPI pending table @0x0000080000a60000 Aug 13 00:25:21.167464 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167472 kernel: CPU39: Booted secondary processor 0x0000070000 [0x413fd0c1] Aug 13 00:25:21.167479 kernel: Detected PIPT I-cache on CPU40 Aug 13 00:25:21.167487 kernel: GICv3: CPU40: found redistributor 120100 region 0:0x00001001005e0000 Aug 13 00:25:21.167495 kernel: GICv3: CPU40: using allocated LPI pending table @0x0000080000a70000 Aug 13 00:25:21.167504 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167511 kernel: CPU40: Booted secondary processor 0x0000120100 [0x413fd0c1] Aug 13 00:25:21.167519 kernel: Detected PIPT I-cache on CPU41 Aug 13 00:25:21.167527 kernel: GICv3: CPU41: found redistributor 1a0100 region 0:0x00001001007e0000 Aug 13 00:25:21.167534 kernel: GICv3: CPU41: using allocated LPI pending table @0x0000080000a80000 Aug 13 00:25:21.167542 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167549 kernel: CPU41: Booted secondary processor 0x00001a0100 [0x413fd0c1] Aug 13 00:25:21.167557 kernel: Detected PIPT I-cache on CPU42 Aug 13 00:25:21.167565 kernel: GICv3: CPU42: found redistributor 140100 region 0:0x0000100100660000 Aug 13 00:25:21.167572 kernel: GICv3: CPU42: using allocated LPI pending table @0x0000080000a90000 Aug 13 00:25:21.167581 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167589 kernel: CPU42: Booted secondary processor 0x0000140100 [0x413fd0c1] Aug 13 00:25:21.167596 kernel: Detected PIPT I-cache on CPU43 Aug 13 00:25:21.167604 kernel: GICv3: CPU43: found redistributor 1c0100 region 0:0x0000100100860000 Aug 13 00:25:21.167612 kernel: GICv3: CPU43: using allocated LPI pending table @0x0000080000aa0000 Aug 13 00:25:21.167620 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167627 kernel: CPU43: Booted secondary processor 0x00001c0100 [0x413fd0c1] Aug 13 00:25:21.167635 kernel: Detected PIPT I-cache on CPU44 Aug 13 00:25:21.167642 kernel: GICv3: CPU44: found redistributor 100100 region 0:0x0000100100560000 Aug 13 00:25:21.167651 kernel: GICv3: CPU44: using allocated LPI pending table @0x0000080000ab0000 Aug 13 00:25:21.167659 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167667 kernel: CPU44: Booted secondary processor 0x0000100100 [0x413fd0c1] Aug 13 00:25:21.167674 kernel: Detected PIPT I-cache on CPU45 Aug 13 00:25:21.167682 kernel: GICv3: CPU45: found redistributor 180100 region 0:0x0000100100760000 Aug 13 00:25:21.167689 kernel: GICv3: CPU45: using allocated LPI pending table @0x0000080000ac0000 Aug 13 00:25:21.167697 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167705 kernel: CPU45: Booted secondary processor 0x0000180100 [0x413fd0c1] Aug 13 00:25:21.167712 kernel: Detected PIPT I-cache on CPU46 Aug 13 00:25:21.167720 kernel: GICv3: CPU46: found redistributor 160100 region 0:0x00001001006e0000 Aug 13 00:25:21.167729 kernel: GICv3: CPU46: using allocated LPI pending table @0x0000080000ad0000 Aug 13 00:25:21.167738 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167746 kernel: CPU46: Booted secondary processor 0x0000160100 [0x413fd0c1] Aug 13 00:25:21.167753 kernel: Detected PIPT I-cache on CPU47 Aug 13 00:25:21.167761 kernel: GICv3: CPU47: found redistributor 1e0100 region 0:0x00001001008e0000 Aug 13 00:25:21.167769 kernel: GICv3: CPU47: using allocated LPI pending table @0x0000080000ae0000 Aug 13 00:25:21.167776 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167784 kernel: CPU47: Booted secondary processor 0x00001e0100 [0x413fd0c1] Aug 13 00:25:21.167791 kernel: Detected PIPT I-cache on CPU48 Aug 13 00:25:21.167800 kernel: GICv3: CPU48: found redistributor a0100 region 0:0x00001001003e0000 Aug 13 00:25:21.167808 kernel: GICv3: CPU48: using allocated LPI pending table @0x0000080000af0000 Aug 13 00:25:21.167816 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167823 kernel: CPU48: Booted secondary processor 0x00000a0100 [0x413fd0c1] Aug 13 00:25:21.167831 kernel: Detected PIPT I-cache on CPU49 Aug 13 00:25:21.167839 kernel: GICv3: CPU49: found redistributor 220100 region 0:0x00001001009e0000 Aug 13 00:25:21.167846 kernel: GICv3: CPU49: using allocated LPI pending table @0x0000080000b00000 Aug 13 00:25:21.167854 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167861 kernel: CPU49: Booted secondary processor 0x0000220100 [0x413fd0c1] Aug 13 00:25:21.167869 kernel: Detected PIPT I-cache on CPU50 Aug 13 00:25:21.167878 kernel: GICv3: CPU50: found redistributor c0100 region 0:0x0000100100460000 Aug 13 00:25:21.167886 kernel: GICv3: CPU50: using allocated LPI pending table @0x0000080000b10000 Aug 13 00:25:21.167893 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167901 kernel: CPU50: Booted secondary processor 0x00000c0100 [0x413fd0c1] Aug 13 00:25:21.167908 kernel: Detected PIPT I-cache on CPU51 Aug 13 00:25:21.167916 kernel: GICv3: CPU51: found redistributor 240100 region 0:0x0000100100a60000 Aug 13 00:25:21.167923 kernel: GICv3: CPU51: using allocated LPI pending table @0x0000080000b20000 Aug 13 00:25:21.167931 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167938 kernel: CPU51: Booted secondary processor 0x0000240100 [0x413fd0c1] Aug 13 00:25:21.167947 kernel: Detected PIPT I-cache on CPU52 Aug 13 00:25:21.167955 kernel: GICv3: CPU52: found redistributor 80100 region 0:0x0000100100360000 Aug 13 00:25:21.167963 kernel: GICv3: CPU52: using allocated LPI pending table @0x0000080000b30000 Aug 13 00:25:21.167970 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.167978 kernel: CPU52: Booted secondary processor 0x0000080100 [0x413fd0c1] Aug 13 00:25:21.167986 kernel: Detected PIPT I-cache on CPU53 Aug 13 00:25:21.167993 kernel: GICv3: CPU53: found redistributor 200100 region 0:0x0000100100960000 Aug 13 00:25:21.168002 kernel: GICv3: CPU53: using allocated LPI pending table @0x0000080000b40000 Aug 13 00:25:21.168010 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168017 kernel: CPU53: Booted secondary processor 0x0000200100 [0x413fd0c1] Aug 13 00:25:21.168026 kernel: Detected PIPT I-cache on CPU54 Aug 13 00:25:21.168034 kernel: GICv3: CPU54: found redistributor e0100 region 0:0x00001001004e0000 Aug 13 00:25:21.168044 kernel: GICv3: CPU54: using allocated LPI pending table @0x0000080000b50000 Aug 13 00:25:21.168052 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168059 kernel: CPU54: Booted secondary processor 0x00000e0100 [0x413fd0c1] Aug 13 00:25:21.168067 kernel: Detected PIPT I-cache on CPU55 Aug 13 00:25:21.168075 kernel: GICv3: CPU55: found redistributor 260100 region 0:0x0000100100ae0000 Aug 13 00:25:21.168082 kernel: GICv3: CPU55: using allocated LPI pending table @0x0000080000b60000 Aug 13 00:25:21.168090 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168100 kernel: CPU55: Booted secondary processor 0x0000260100 [0x413fd0c1] Aug 13 00:25:21.168107 kernel: Detected PIPT I-cache on CPU56 Aug 13 00:25:21.168115 kernel: GICv3: CPU56: found redistributor 20100 region 0:0x00001001001e0000 Aug 13 00:25:21.168122 kernel: GICv3: CPU56: using allocated LPI pending table @0x0000080000b70000 Aug 13 00:25:21.168130 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168138 kernel: CPU56: Booted secondary processor 0x0000020100 [0x413fd0c1] Aug 13 00:25:21.168145 kernel: Detected PIPT I-cache on CPU57 Aug 13 00:25:21.168153 kernel: GICv3: CPU57: found redistributor 40100 region 0:0x0000100100260000 Aug 13 00:25:21.168161 kernel: GICv3: CPU57: using allocated LPI pending table @0x0000080000b80000 Aug 13 00:25:21.168170 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168177 kernel: CPU57: Booted secondary processor 0x0000040100 [0x413fd0c1] Aug 13 00:25:21.168185 kernel: Detected PIPT I-cache on CPU58 Aug 13 00:25:21.168193 kernel: GICv3: CPU58: found redistributor 100 region 0:0x0000100100160000 Aug 13 00:25:21.168200 kernel: GICv3: CPU58: using allocated LPI pending table @0x0000080000b90000 Aug 13 00:25:21.168208 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168216 kernel: CPU58: Booted secondary processor 0x0000000100 [0x413fd0c1] Aug 13 00:25:21.168223 kernel: Detected PIPT I-cache on CPU59 Aug 13 00:25:21.168231 kernel: GICv3: CPU59: found redistributor 60100 region 0:0x00001001002e0000 Aug 13 00:25:21.168238 kernel: GICv3: CPU59: using allocated LPI pending table @0x0000080000ba0000 Aug 13 00:25:21.168247 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168255 kernel: CPU59: Booted secondary processor 0x0000060100 [0x413fd0c1] Aug 13 00:25:21.168263 kernel: Detected PIPT I-cache on CPU60 Aug 13 00:25:21.168270 kernel: GICv3: CPU60: found redistributor 130100 region 0:0x0000100100620000 Aug 13 00:25:21.168278 kernel: GICv3: CPU60: using allocated LPI pending table @0x0000080000bb0000 Aug 13 00:25:21.168286 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168293 kernel: CPU60: Booted secondary processor 0x0000130100 [0x413fd0c1] Aug 13 00:25:21.168301 kernel: Detected PIPT I-cache on CPU61 Aug 13 00:25:21.168308 kernel: GICv3: CPU61: found redistributor 1b0100 region 0:0x0000100100820000 Aug 13 00:25:21.168318 kernel: GICv3: CPU61: using allocated LPI pending table @0x0000080000bc0000 Aug 13 00:25:21.168325 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168333 kernel: CPU61: Booted secondary processor 0x00001b0100 [0x413fd0c1] Aug 13 00:25:21.168341 kernel: Detected PIPT I-cache on CPU62 Aug 13 00:25:21.168349 kernel: GICv3: CPU62: found redistributor 150100 region 0:0x00001001006a0000 Aug 13 00:25:21.168357 kernel: GICv3: CPU62: using allocated LPI pending table @0x0000080000bd0000 Aug 13 00:25:21.168365 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168372 kernel: CPU62: Booted secondary processor 0x0000150100 [0x413fd0c1] Aug 13 00:25:21.168380 kernel: Detected PIPT I-cache on CPU63 Aug 13 00:25:21.168387 kernel: GICv3: CPU63: found redistributor 1d0100 region 0:0x00001001008a0000 Aug 13 00:25:21.168397 kernel: GICv3: CPU63: using allocated LPI pending table @0x0000080000be0000 Aug 13 00:25:21.168404 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168412 kernel: CPU63: Booted secondary processor 0x00001d0100 [0x413fd0c1] Aug 13 00:25:21.168420 kernel: Detected PIPT I-cache on CPU64 Aug 13 00:25:21.168427 kernel: GICv3: CPU64: found redistributor 110100 region 0:0x00001001005a0000 Aug 13 00:25:21.168435 kernel: GICv3: CPU64: using allocated LPI pending table @0x0000080000bf0000 Aug 13 00:25:21.168443 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168450 kernel: CPU64: Booted secondary processor 0x0000110100 [0x413fd0c1] Aug 13 00:25:21.168458 kernel: Detected PIPT I-cache on CPU65 Aug 13 00:25:21.168467 kernel: GICv3: CPU65: found redistributor 190100 region 0:0x00001001007a0000 Aug 13 00:25:21.168474 kernel: GICv3: CPU65: using allocated LPI pending table @0x0000080000c00000 Aug 13 00:25:21.168482 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168490 kernel: CPU65: Booted secondary processor 0x0000190100 [0x413fd0c1] Aug 13 00:25:21.168497 kernel: Detected PIPT I-cache on CPU66 Aug 13 00:25:21.168505 kernel: GICv3: CPU66: found redistributor 170100 region 0:0x0000100100720000 Aug 13 00:25:21.168513 kernel: GICv3: CPU66: using allocated LPI pending table @0x0000080000c10000 Aug 13 00:25:21.168520 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168528 kernel: CPU66: Booted secondary processor 0x0000170100 [0x413fd0c1] Aug 13 00:25:21.168535 kernel: Detected PIPT I-cache on CPU67 Aug 13 00:25:21.168545 kernel: GICv3: CPU67: found redistributor 1f0100 region 0:0x0000100100920000 Aug 13 00:25:21.168552 kernel: GICv3: CPU67: using allocated LPI pending table @0x0000080000c20000 Aug 13 00:25:21.168560 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168568 kernel: CPU67: Booted secondary processor 0x00001f0100 [0x413fd0c1] Aug 13 00:25:21.168575 kernel: Detected PIPT I-cache on CPU68 Aug 13 00:25:21.168583 kernel: GICv3: CPU68: found redistributor b0100 region 0:0x0000100100420000 Aug 13 00:25:21.168591 kernel: GICv3: CPU68: using allocated LPI pending table @0x0000080000c30000 Aug 13 00:25:21.168598 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168606 kernel: CPU68: Booted secondary processor 0x00000b0100 [0x413fd0c1] Aug 13 00:25:21.168615 kernel: Detected PIPT I-cache on CPU69 Aug 13 00:25:21.168623 kernel: GICv3: CPU69: found redistributor 230100 region 0:0x0000100100a20000 Aug 13 00:25:21.168630 kernel: GICv3: CPU69: using allocated LPI pending table @0x0000080000c40000 Aug 13 00:25:21.168638 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168645 kernel: CPU69: Booted secondary processor 0x0000230100 [0x413fd0c1] Aug 13 00:25:21.168653 kernel: Detected PIPT I-cache on CPU70 Aug 13 00:25:21.168661 kernel: GICv3: CPU70: found redistributor d0100 region 0:0x00001001004a0000 Aug 13 00:25:21.168669 kernel: GICv3: CPU70: using allocated LPI pending table @0x0000080000c50000 Aug 13 00:25:21.168677 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168684 kernel: CPU70: Booted secondary processor 0x00000d0100 [0x413fd0c1] Aug 13 00:25:21.168693 kernel: Detected PIPT I-cache on CPU71 Aug 13 00:25:21.168701 kernel: GICv3: CPU71: found redistributor 250100 region 0:0x0000100100aa0000 Aug 13 00:25:21.168709 kernel: GICv3: CPU71: using allocated LPI pending table @0x0000080000c60000 Aug 13 00:25:21.168716 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168724 kernel: CPU71: Booted secondary processor 0x0000250100 [0x413fd0c1] Aug 13 00:25:21.168731 kernel: Detected PIPT I-cache on CPU72 Aug 13 00:25:21.168739 kernel: GICv3: CPU72: found redistributor 90100 region 0:0x00001001003a0000 Aug 13 00:25:21.168747 kernel: GICv3: CPU72: using allocated LPI pending table @0x0000080000c70000 Aug 13 00:25:21.168754 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168763 kernel: CPU72: Booted secondary processor 0x0000090100 [0x413fd0c1] Aug 13 00:25:21.168771 kernel: Detected PIPT I-cache on CPU73 Aug 13 00:25:21.168779 kernel: GICv3: CPU73: found redistributor 210100 region 0:0x00001001009a0000 Aug 13 00:25:21.168786 kernel: GICv3: CPU73: using allocated LPI pending table @0x0000080000c80000 Aug 13 00:25:21.168794 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168801 kernel: CPU73: Booted secondary processor 0x0000210100 [0x413fd0c1] Aug 13 00:25:21.168809 kernel: Detected PIPT I-cache on CPU74 Aug 13 00:25:21.168817 kernel: GICv3: CPU74: found redistributor f0100 region 0:0x0000100100520000 Aug 13 00:25:21.168824 kernel: GICv3: CPU74: using allocated LPI pending table @0x0000080000c90000 Aug 13 00:25:21.168833 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168841 kernel: CPU74: Booted secondary processor 0x00000f0100 [0x413fd0c1] Aug 13 00:25:21.168849 kernel: Detected PIPT I-cache on CPU75 Aug 13 00:25:21.168856 kernel: GICv3: CPU75: found redistributor 270100 region 0:0x0000100100b20000 Aug 13 00:25:21.168864 kernel: GICv3: CPU75: using allocated LPI pending table @0x0000080000ca0000 Aug 13 00:25:21.168872 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168879 kernel: CPU75: Booted secondary processor 0x0000270100 [0x413fd0c1] Aug 13 00:25:21.168887 kernel: Detected PIPT I-cache on CPU76 Aug 13 00:25:21.168894 kernel: GICv3: CPU76: found redistributor 30100 region 0:0x0000100100220000 Aug 13 00:25:21.168902 kernel: GICv3: CPU76: using allocated LPI pending table @0x0000080000cb0000 Aug 13 00:25:21.168911 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168919 kernel: CPU76: Booted secondary processor 0x0000030100 [0x413fd0c1] Aug 13 00:25:21.168926 kernel: Detected PIPT I-cache on CPU77 Aug 13 00:25:21.168934 kernel: GICv3: CPU77: found redistributor 50100 region 0:0x00001001002a0000 Aug 13 00:25:21.168942 kernel: GICv3: CPU77: using allocated LPI pending table @0x0000080000cc0000 Aug 13 00:25:21.168949 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168957 kernel: CPU77: Booted secondary processor 0x0000050100 [0x413fd0c1] Aug 13 00:25:21.168964 kernel: Detected PIPT I-cache on CPU78 Aug 13 00:25:21.168972 kernel: GICv3: CPU78: found redistributor 10100 region 0:0x00001001001a0000 Aug 13 00:25:21.168981 kernel: GICv3: CPU78: using allocated LPI pending table @0x0000080000cd0000 Aug 13 00:25:21.168989 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.168996 kernel: CPU78: Booted secondary processor 0x0000010100 [0x413fd0c1] Aug 13 00:25:21.169004 kernel: Detected PIPT I-cache on CPU79 Aug 13 00:25:21.169011 kernel: GICv3: CPU79: found redistributor 70100 region 0:0x0000100100320000 Aug 13 00:25:21.169019 kernel: GICv3: CPU79: using allocated LPI pending table @0x0000080000ce0000 Aug 13 00:25:21.169027 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 00:25:21.169035 kernel: CPU79: Booted secondary processor 0x0000070100 [0x413fd0c1] Aug 13 00:25:21.169044 kernel: smp: Brought up 1 node, 80 CPUs Aug 13 00:25:21.169052 kernel: SMP: Total of 80 processors activated. Aug 13 00:25:21.169061 kernel: CPU features: detected: 32-bit EL0 Support Aug 13 00:25:21.169068 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence Aug 13 00:25:21.169076 kernel: CPU features: detected: Common not Private translations Aug 13 00:25:21.169084 kernel: CPU features: detected: CRC32 instructions Aug 13 00:25:21.169092 kernel: CPU features: detected: Enhanced Virtualization Traps Aug 13 00:25:21.169099 kernel: CPU features: detected: RCpc load-acquire (LDAPR) Aug 13 00:25:21.169107 kernel: CPU features: detected: LSE atomic instructions Aug 13 00:25:21.169115 kernel: CPU features: detected: Privileged Access Never Aug 13 00:25:21.169122 kernel: CPU features: detected: RAS Extension Support Aug 13 00:25:21.169131 kernel: CPU features: detected: Speculative Store Bypassing Safe (SSBS) Aug 13 00:25:21.169139 kernel: CPU: All CPU(s) started at EL2 Aug 13 00:25:21.169147 kernel: alternatives: applying system-wide alternatives Aug 13 00:25:21.169154 kernel: devtmpfs: initialized Aug 13 00:25:21.169162 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Aug 13 00:25:21.169170 kernel: futex hash table entries: 32768 (order: 9, 2097152 bytes, linear) Aug 13 00:25:21.169177 kernel: pinctrl core: initialized pinctrl subsystem Aug 13 00:25:21.169185 kernel: SMBIOS 3.4.0 present. Aug 13 00:25:21.169192 kernel: DMI: GIGABYTE R272-P30-JG/MP32-AR0-JG, BIOS F17a (SCP: 1.07.20210713) 07/22/2021 Aug 13 00:25:21.169201 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Aug 13 00:25:21.169209 kernel: DMA: preallocated 4096 KiB GFP_KERNEL pool for atomic allocations Aug 13 00:25:21.169217 kernel: DMA: preallocated 4096 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Aug 13 00:25:21.169225 kernel: DMA: preallocated 4096 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Aug 13 00:25:21.169232 kernel: audit: initializing netlink subsys (disabled) Aug 13 00:25:21.169240 kernel: audit: type=2000 audit(0.042:1): state=initialized audit_enabled=0 res=1 Aug 13 00:25:21.169248 kernel: thermal_sys: Registered thermal governor 'step_wise' Aug 13 00:25:21.169255 kernel: cpuidle: using governor menu Aug 13 00:25:21.169263 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Aug 13 00:25:21.169272 kernel: ASID allocator initialised with 32768 entries Aug 13 00:25:21.169280 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Aug 13 00:25:21.169288 kernel: Serial: AMBA PL011 UART driver Aug 13 00:25:21.169295 kernel: Modules: 2G module region forced by RANDOMIZE_MODULE_REGION_FULL Aug 13 00:25:21.169303 kernel: Modules: 0 pages in range for non-PLT usage Aug 13 00:25:21.169311 kernel: Modules: 509008 pages in range for PLT usage Aug 13 00:25:21.169318 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Aug 13 00:25:21.169326 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page Aug 13 00:25:21.169334 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages Aug 13 00:25:21.169343 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page Aug 13 00:25:21.169350 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Aug 13 00:25:21.169358 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page Aug 13 00:25:21.169366 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages Aug 13 00:25:21.169373 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page Aug 13 00:25:21.169381 kernel: ACPI: Added _OSI(Module Device) Aug 13 00:25:21.169388 kernel: ACPI: Added _OSI(Processor Device) Aug 13 00:25:21.169396 kernel: ACPI: Added _OSI(Processor Aggregator Device) Aug 13 00:25:21.169404 kernel: ACPI: 2 ACPI AML tables successfully acquired and loaded Aug 13 00:25:21.169413 kernel: ACPI: Interpreter enabled Aug 13 00:25:21.169421 kernel: ACPI: Using GIC for interrupt routing Aug 13 00:25:21.169428 kernel: ACPI: MCFG table detected, 8 entries Aug 13 00:25:21.169436 kernel: ACPI: IORT: SMMU-v3[33ffe0000000] Mapped to Proximity domain 0 Aug 13 00:25:21.169443 kernel: ACPI: IORT: SMMU-v3[37ffe0000000] Mapped to Proximity domain 0 Aug 13 00:25:21.169451 kernel: ACPI: IORT: SMMU-v3[3bffe0000000] Mapped to Proximity domain 0 Aug 13 00:25:21.169459 kernel: ACPI: IORT: SMMU-v3[3fffe0000000] Mapped to Proximity domain 0 Aug 13 00:25:21.169466 kernel: ACPI: IORT: SMMU-v3[23ffe0000000] Mapped to Proximity domain 0 Aug 13 00:25:21.169474 kernel: ACPI: IORT: SMMU-v3[27ffe0000000] Mapped to Proximity domain 0 Aug 13 00:25:21.169483 kernel: ACPI: IORT: SMMU-v3[2bffe0000000] Mapped to Proximity domain 0 Aug 13 00:25:21.169491 kernel: ACPI: IORT: SMMU-v3[2fffe0000000] Mapped to Proximity domain 0 Aug 13 00:25:21.169499 kernel: ARMH0011:00: ttyAMA0 at MMIO 0x100002600000 (irq = 19, base_baud = 0) is a SBSA Aug 13 00:25:21.169507 kernel: printk: console [ttyAMA0] enabled Aug 13 00:25:21.169515 kernel: ARMH0011:01: ttyAMA1 at MMIO 0x100002620000 (irq = 20, base_baud = 0) is a SBSA Aug 13 00:25:21.169522 kernel: ACPI: PCI Root Bridge [PCI1] (domain 000d [bus 00-ff]) Aug 13 00:25:21.169671 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Aug 13 00:25:21.169762 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug PME LTR] Aug 13 00:25:21.169831 kernel: acpi PNP0A08:00: _OSC: OS now controls [AER PCIeCapability] Aug 13 00:25:21.169893 kernel: acpi PNP0A08:00: MCFG quirk: ECAM at [mem 0x37fff0000000-0x37ffffffffff] for [bus 00-ff] with pci_32b_read_ops Aug 13 00:25:21.169956 kernel: acpi PNP0A08:00: ECAM area [mem 0x37fff0000000-0x37ffffffffff] reserved by PNP0C02:00 Aug 13 00:25:21.170018 kernel: acpi PNP0A08:00: ECAM at [mem 0x37fff0000000-0x37ffffffffff] for [bus 00-ff] Aug 13 00:25:21.170028 kernel: PCI host bridge to bus 000d:00 Aug 13 00:25:21.170103 kernel: pci_bus 000d:00: root bus resource [mem 0x50000000-0x5fffffff window] Aug 13 00:25:21.170164 kernel: pci_bus 000d:00: root bus resource [mem 0x340000000000-0x37ffdfffffff window] Aug 13 00:25:21.170222 kernel: pci_bus 000d:00: root bus resource [bus 00-ff] Aug 13 00:25:21.170300 kernel: pci 000d:00:00.0: [1def:e100] type 00 class 0x060000 Aug 13 00:25:21.170375 kernel: pci 000d:00:01.0: [1def:e101] type 01 class 0x060400 Aug 13 00:25:21.170441 kernel: pci 000d:00:01.0: enabling Extended Tags Aug 13 00:25:21.170507 kernel: pci 000d:00:01.0: supports D1 D2 Aug 13 00:25:21.170570 kernel: pci 000d:00:01.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.170648 kernel: pci 000d:00:02.0: [1def:e102] type 01 class 0x060400 Aug 13 00:25:21.170715 kernel: pci 000d:00:02.0: supports D1 D2 Aug 13 00:25:21.170779 kernel: pci 000d:00:02.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.170851 kernel: pci 000d:00:03.0: [1def:e103] type 01 class 0x060400 Aug 13 00:25:21.170917 kernel: pci 000d:00:03.0: supports D1 D2 Aug 13 00:25:21.170980 kernel: pci 000d:00:03.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.171055 kernel: pci 000d:00:04.0: [1def:e104] type 01 class 0x060400 Aug 13 00:25:21.171126 kernel: pci 000d:00:04.0: supports D1 D2 Aug 13 00:25:21.171191 kernel: pci 000d:00:04.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.171202 kernel: acpiphp: Slot [1] registered Aug 13 00:25:21.171209 kernel: acpiphp: Slot [2] registered Aug 13 00:25:21.171217 kernel: acpiphp: Slot [3] registered Aug 13 00:25:21.171225 kernel: acpiphp: Slot [4] registered Aug 13 00:25:21.171284 kernel: pci_bus 000d:00: on NUMA node 0 Aug 13 00:25:21.171353 kernel: pci 000d:00:01.0: bridge window [io 0x1000-0x0fff] to [bus 01] add_size 1000 Aug 13 00:25:21.171418 kernel: pci 000d:00:01.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 01] add_size 200000 add_align 100000 Aug 13 00:25:21.171483 kernel: pci 000d:00:01.0: bridge window [mem 0x00100000-0x000fffff] to [bus 01] add_size 200000 add_align 100000 Aug 13 00:25:21.171548 kernel: pci 000d:00:02.0: bridge window [io 0x1000-0x0fff] to [bus 02] add_size 1000 Aug 13 00:25:21.171612 kernel: pci 000d:00:02.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 02] add_size 200000 add_align 100000 Aug 13 00:25:21.171677 kernel: pci 000d:00:02.0: bridge window [mem 0x00100000-0x000fffff] to [bus 02] add_size 200000 add_align 100000 Aug 13 00:25:21.171742 kernel: pci 000d:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 03] add_size 1000 Aug 13 00:25:21.171813 kernel: pci 000d:00:03.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 03] add_size 200000 add_align 100000 Aug 13 00:25:21.171876 kernel: pci 000d:00:03.0: bridge window [mem 0x00100000-0x000fffff] to [bus 03] add_size 200000 add_align 100000 Aug 13 00:25:21.171942 kernel: pci 000d:00:04.0: bridge window [io 0x1000-0x0fff] to [bus 04] add_size 1000 Aug 13 00:25:21.172005 kernel: pci 000d:00:04.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 04] add_size 200000 add_align 100000 Aug 13 00:25:21.172073 kernel: pci 000d:00:04.0: bridge window [mem 0x00100000-0x000fffff] to [bus 04] add_size 200000 add_align 100000 Aug 13 00:25:21.172138 kernel: pci 000d:00:01.0: BAR 14: assigned [mem 0x50000000-0x501fffff] Aug 13 00:25:21.172204 kernel: pci 000d:00:01.0: BAR 15: assigned [mem 0x340000000000-0x3400001fffff 64bit pref] Aug 13 00:25:21.172272 kernel: pci 000d:00:02.0: BAR 14: assigned [mem 0x50200000-0x503fffff] Aug 13 00:25:21.172336 kernel: pci 000d:00:02.0: BAR 15: assigned [mem 0x340000200000-0x3400003fffff 64bit pref] Aug 13 00:25:21.172402 kernel: pci 000d:00:03.0: BAR 14: assigned [mem 0x50400000-0x505fffff] Aug 13 00:25:21.172467 kernel: pci 000d:00:03.0: BAR 15: assigned [mem 0x340000400000-0x3400005fffff 64bit pref] Aug 13 00:25:21.172532 kernel: pci 000d:00:04.0: BAR 14: assigned [mem 0x50600000-0x507fffff] Aug 13 00:25:21.172596 kernel: pci 000d:00:04.0: BAR 15: assigned [mem 0x340000600000-0x3400007fffff 64bit pref] Aug 13 00:25:21.172661 kernel: pci 000d:00:01.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.172724 kernel: pci 000d:00:01.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.172792 kernel: pci 000d:00:02.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.172856 kernel: pci 000d:00:02.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.172920 kernel: pci 000d:00:03.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.172984 kernel: pci 000d:00:03.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.173056 kernel: pci 000d:00:04.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.173120 kernel: pci 000d:00:04.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.173185 kernel: pci 000d:00:04.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.173255 kernel: pci 000d:00:04.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.173319 kernel: pci 000d:00:03.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.173383 kernel: pci 000d:00:03.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.173448 kernel: pci 000d:00:02.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.173513 kernel: pci 000d:00:02.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.173576 kernel: pci 000d:00:01.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.173641 kernel: pci 000d:00:01.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.173704 kernel: pci 000d:00:01.0: PCI bridge to [bus 01] Aug 13 00:25:21.173772 kernel: pci 000d:00:01.0: bridge window [mem 0x50000000-0x501fffff] Aug 13 00:25:21.173836 kernel: pci 000d:00:01.0: bridge window [mem 0x340000000000-0x3400001fffff 64bit pref] Aug 13 00:25:21.173901 kernel: pci 000d:00:02.0: PCI bridge to [bus 02] Aug 13 00:25:21.173966 kernel: pci 000d:00:02.0: bridge window [mem 0x50200000-0x503fffff] Aug 13 00:25:21.174031 kernel: pci 000d:00:02.0: bridge window [mem 0x340000200000-0x3400003fffff 64bit pref] Aug 13 00:25:21.174099 kernel: pci 000d:00:03.0: PCI bridge to [bus 03] Aug 13 00:25:21.174164 kernel: pci 000d:00:03.0: bridge window [mem 0x50400000-0x505fffff] Aug 13 00:25:21.174230 kernel: pci 000d:00:03.0: bridge window [mem 0x340000400000-0x3400005fffff 64bit pref] Aug 13 00:25:21.174295 kernel: pci 000d:00:04.0: PCI bridge to [bus 04] Aug 13 00:25:21.174358 kernel: pci 000d:00:04.0: bridge window [mem 0x50600000-0x507fffff] Aug 13 00:25:21.174423 kernel: pci 000d:00:04.0: bridge window [mem 0x340000600000-0x3400007fffff 64bit pref] Aug 13 00:25:21.174481 kernel: pci_bus 000d:00: resource 4 [mem 0x50000000-0x5fffffff window] Aug 13 00:25:21.174538 kernel: pci_bus 000d:00: resource 5 [mem 0x340000000000-0x37ffdfffffff window] Aug 13 00:25:21.174611 kernel: pci_bus 000d:01: resource 1 [mem 0x50000000-0x501fffff] Aug 13 00:25:21.174671 kernel: pci_bus 000d:01: resource 2 [mem 0x340000000000-0x3400001fffff 64bit pref] Aug 13 00:25:21.174739 kernel: pci_bus 000d:02: resource 1 [mem 0x50200000-0x503fffff] Aug 13 00:25:21.174799 kernel: pci_bus 000d:02: resource 2 [mem 0x340000200000-0x3400003fffff 64bit pref] Aug 13 00:25:21.174874 kernel: pci_bus 000d:03: resource 1 [mem 0x50400000-0x505fffff] Aug 13 00:25:21.174933 kernel: pci_bus 000d:03: resource 2 [mem 0x340000400000-0x3400005fffff 64bit pref] Aug 13 00:25:21.175004 kernel: pci_bus 000d:04: resource 1 [mem 0x50600000-0x507fffff] Aug 13 00:25:21.175066 kernel: pci_bus 000d:04: resource 2 [mem 0x340000600000-0x3400007fffff 64bit pref] Aug 13 00:25:21.175077 kernel: ACPI: PCI Root Bridge [PCI3] (domain 0000 [bus 00-ff]) Aug 13 00:25:21.175149 kernel: acpi PNP0A08:01: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Aug 13 00:25:21.175213 kernel: acpi PNP0A08:01: _OSC: platform does not support [PCIeHotplug PME LTR] Aug 13 00:25:21.175275 kernel: acpi PNP0A08:01: _OSC: OS now controls [AER PCIeCapability] Aug 13 00:25:21.175339 kernel: acpi PNP0A08:01: MCFG quirk: ECAM at [mem 0x3ffff0000000-0x3fffffffffff] for [bus 00-ff] with pci_32b_read_ops Aug 13 00:25:21.175401 kernel: acpi PNP0A08:01: ECAM area [mem 0x3ffff0000000-0x3fffffffffff] reserved by PNP0C02:00 Aug 13 00:25:21.175463 kernel: acpi PNP0A08:01: ECAM at [mem 0x3ffff0000000-0x3fffffffffff] for [bus 00-ff] Aug 13 00:25:21.175473 kernel: PCI host bridge to bus 0000:00 Aug 13 00:25:21.175537 kernel: pci_bus 0000:00: root bus resource [mem 0x70000000-0x7fffffff window] Aug 13 00:25:21.175597 kernel: pci_bus 0000:00: root bus resource [mem 0x3c0000000000-0x3fffdfffffff window] Aug 13 00:25:21.175653 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Aug 13 00:25:21.175728 kernel: pci 0000:00:00.0: [1def:e100] type 00 class 0x060000 Aug 13 00:25:21.175801 kernel: pci 0000:00:01.0: [1def:e101] type 01 class 0x060400 Aug 13 00:25:21.175867 kernel: pci 0000:00:01.0: enabling Extended Tags Aug 13 00:25:21.175931 kernel: pci 0000:00:01.0: supports D1 D2 Aug 13 00:25:21.175997 kernel: pci 0000:00:01.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.176071 kernel: pci 0000:00:02.0: [1def:e102] type 01 class 0x060400 Aug 13 00:25:21.176138 kernel: pci 0000:00:02.0: supports D1 D2 Aug 13 00:25:21.176206 kernel: pci 0000:00:02.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.176277 kernel: pci 0000:00:03.0: [1def:e103] type 01 class 0x060400 Aug 13 00:25:21.176342 kernel: pci 0000:00:03.0: supports D1 D2 Aug 13 00:25:21.176407 kernel: pci 0000:00:03.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.176480 kernel: pci 0000:00:04.0: [1def:e104] type 01 class 0x060400 Aug 13 00:25:21.176545 kernel: pci 0000:00:04.0: supports D1 D2 Aug 13 00:25:21.176610 kernel: pci 0000:00:04.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.176621 kernel: acpiphp: Slot [1-1] registered Aug 13 00:25:21.176629 kernel: acpiphp: Slot [2-1] registered Aug 13 00:25:21.176637 kernel: acpiphp: Slot [3-1] registered Aug 13 00:25:21.176645 kernel: acpiphp: Slot [4-1] registered Aug 13 00:25:21.176702 kernel: pci_bus 0000:00: on NUMA node 0 Aug 13 00:25:21.176767 kernel: pci 0000:00:01.0: bridge window [io 0x1000-0x0fff] to [bus 01] add_size 1000 Aug 13 00:25:21.176831 kernel: pci 0000:00:01.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 01] add_size 200000 add_align 100000 Aug 13 00:25:21.176896 kernel: pci 0000:00:01.0: bridge window [mem 0x00100000-0x000fffff] to [bus 01] add_size 200000 add_align 100000 Aug 13 00:25:21.176962 kernel: pci 0000:00:02.0: bridge window [io 0x1000-0x0fff] to [bus 02] add_size 1000 Aug 13 00:25:21.177027 kernel: pci 0000:00:02.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 02] add_size 200000 add_align 100000 Aug 13 00:25:21.177095 kernel: pci 0000:00:02.0: bridge window [mem 0x00100000-0x000fffff] to [bus 02] add_size 200000 add_align 100000 Aug 13 00:25:21.177160 kernel: pci 0000:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 03] add_size 1000 Aug 13 00:25:21.177223 kernel: pci 0000:00:03.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 03] add_size 200000 add_align 100000 Aug 13 00:25:21.177288 kernel: pci 0000:00:03.0: bridge window [mem 0x00100000-0x000fffff] to [bus 03] add_size 200000 add_align 100000 Aug 13 00:25:21.177352 kernel: pci 0000:00:04.0: bridge window [io 0x1000-0x0fff] to [bus 04] add_size 1000 Aug 13 00:25:21.177419 kernel: pci 0000:00:04.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 04] add_size 200000 add_align 100000 Aug 13 00:25:21.177483 kernel: pci 0000:00:04.0: bridge window [mem 0x00100000-0x000fffff] to [bus 04] add_size 200000 add_align 100000 Aug 13 00:25:21.177547 kernel: pci 0000:00:01.0: BAR 14: assigned [mem 0x70000000-0x701fffff] Aug 13 00:25:21.177610 kernel: pci 0000:00:01.0: BAR 15: assigned [mem 0x3c0000000000-0x3c00001fffff 64bit pref] Aug 13 00:25:21.177675 kernel: pci 0000:00:02.0: BAR 14: assigned [mem 0x70200000-0x703fffff] Aug 13 00:25:21.177740 kernel: pci 0000:00:02.0: BAR 15: assigned [mem 0x3c0000200000-0x3c00003fffff 64bit pref] Aug 13 00:25:21.177804 kernel: pci 0000:00:03.0: BAR 14: assigned [mem 0x70400000-0x705fffff] Aug 13 00:25:21.177871 kernel: pci 0000:00:03.0: BAR 15: assigned [mem 0x3c0000400000-0x3c00005fffff 64bit pref] Aug 13 00:25:21.177934 kernel: pci 0000:00:04.0: BAR 14: assigned [mem 0x70600000-0x707fffff] Aug 13 00:25:21.177999 kernel: pci 0000:00:04.0: BAR 15: assigned [mem 0x3c0000600000-0x3c00007fffff 64bit pref] Aug 13 00:25:21.178066 kernel: pci 0000:00:01.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.178130 kernel: pci 0000:00:01.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.178194 kernel: pci 0000:00:02.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.178259 kernel: pci 0000:00:02.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.178322 kernel: pci 0000:00:03.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.178390 kernel: pci 0000:00:03.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.178454 kernel: pci 0000:00:04.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.178518 kernel: pci 0000:00:04.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.178581 kernel: pci 0000:00:04.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.178645 kernel: pci 0000:00:04.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.178708 kernel: pci 0000:00:03.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.178771 kernel: pci 0000:00:03.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.178835 kernel: pci 0000:00:02.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.178898 kernel: pci 0000:00:02.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.178966 kernel: pci 0000:00:01.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.179029 kernel: pci 0000:00:01.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.179097 kernel: pci 0000:00:01.0: PCI bridge to [bus 01] Aug 13 00:25:21.179161 kernel: pci 0000:00:01.0: bridge window [mem 0x70000000-0x701fffff] Aug 13 00:25:21.179225 kernel: pci 0000:00:01.0: bridge window [mem 0x3c0000000000-0x3c00001fffff 64bit pref] Aug 13 00:25:21.179287 kernel: pci 0000:00:02.0: PCI bridge to [bus 02] Aug 13 00:25:21.179352 kernel: pci 0000:00:02.0: bridge window [mem 0x70200000-0x703fffff] Aug 13 00:25:21.179418 kernel: pci 0000:00:02.0: bridge window [mem 0x3c0000200000-0x3c00003fffff 64bit pref] Aug 13 00:25:21.179485 kernel: pci 0000:00:03.0: PCI bridge to [bus 03] Aug 13 00:25:21.179548 kernel: pci 0000:00:03.0: bridge window [mem 0x70400000-0x705fffff] Aug 13 00:25:21.179612 kernel: pci 0000:00:03.0: bridge window [mem 0x3c0000400000-0x3c00005fffff 64bit pref] Aug 13 00:25:21.179678 kernel: pci 0000:00:04.0: PCI bridge to [bus 04] Aug 13 00:25:21.179742 kernel: pci 0000:00:04.0: bridge window [mem 0x70600000-0x707fffff] Aug 13 00:25:21.179807 kernel: pci 0000:00:04.0: bridge window [mem 0x3c0000600000-0x3c00007fffff 64bit pref] Aug 13 00:25:21.179866 kernel: pci_bus 0000:00: resource 4 [mem 0x70000000-0x7fffffff window] Aug 13 00:25:21.179923 kernel: pci_bus 0000:00: resource 5 [mem 0x3c0000000000-0x3fffdfffffff window] Aug 13 00:25:21.179991 kernel: pci_bus 0000:01: resource 1 [mem 0x70000000-0x701fffff] Aug 13 00:25:21.180059 kernel: pci_bus 0000:01: resource 2 [mem 0x3c0000000000-0x3c00001fffff 64bit pref] Aug 13 00:25:21.180126 kernel: pci_bus 0000:02: resource 1 [mem 0x70200000-0x703fffff] Aug 13 00:25:21.180186 kernel: pci_bus 0000:02: resource 2 [mem 0x3c0000200000-0x3c00003fffff 64bit pref] Aug 13 00:25:21.180262 kernel: pci_bus 0000:03: resource 1 [mem 0x70400000-0x705fffff] Aug 13 00:25:21.180322 kernel: pci_bus 0000:03: resource 2 [mem 0x3c0000400000-0x3c00005fffff 64bit pref] Aug 13 00:25:21.180389 kernel: pci_bus 0000:04: resource 1 [mem 0x70600000-0x707fffff] Aug 13 00:25:21.180453 kernel: pci_bus 0000:04: resource 2 [mem 0x3c0000600000-0x3c00007fffff 64bit pref] Aug 13 00:25:21.180463 kernel: ACPI: PCI Root Bridge [PCI7] (domain 0005 [bus 00-ff]) Aug 13 00:25:21.180533 kernel: acpi PNP0A08:02: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Aug 13 00:25:21.180596 kernel: acpi PNP0A08:02: _OSC: platform does not support [PCIeHotplug PME LTR] Aug 13 00:25:21.180658 kernel: acpi PNP0A08:02: _OSC: OS now controls [AER PCIeCapability] Aug 13 00:25:21.180722 kernel: acpi PNP0A08:02: MCFG quirk: ECAM at [mem 0x2ffff0000000-0x2fffffffffff] for [bus 00-ff] with pci_32b_read_ops Aug 13 00:25:21.180784 kernel: acpi PNP0A08:02: ECAM area [mem 0x2ffff0000000-0x2fffffffffff] reserved by PNP0C02:00 Aug 13 00:25:21.180847 kernel: acpi PNP0A08:02: ECAM at [mem 0x2ffff0000000-0x2fffffffffff] for [bus 00-ff] Aug 13 00:25:21.180858 kernel: PCI host bridge to bus 0005:00 Aug 13 00:25:21.180923 kernel: pci_bus 0005:00: root bus resource [mem 0x30000000-0x3fffffff window] Aug 13 00:25:21.180981 kernel: pci_bus 0005:00: root bus resource [mem 0x2c0000000000-0x2fffdfffffff window] Aug 13 00:25:21.181041 kernel: pci_bus 0005:00: root bus resource [bus 00-ff] Aug 13 00:25:21.181113 kernel: pci 0005:00:00.0: [1def:e110] type 00 class 0x060000 Aug 13 00:25:21.181188 kernel: pci 0005:00:01.0: [1def:e111] type 01 class 0x060400 Aug 13 00:25:21.181252 kernel: pci 0005:00:01.0: supports D1 D2 Aug 13 00:25:21.181317 kernel: pci 0005:00:01.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.181386 kernel: pci 0005:00:03.0: [1def:e113] type 01 class 0x060400 Aug 13 00:25:21.181453 kernel: pci 0005:00:03.0: supports D1 D2 Aug 13 00:25:21.181517 kernel: pci 0005:00:03.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.181590 kernel: pci 0005:00:05.0: [1def:e115] type 01 class 0x060400 Aug 13 00:25:21.181659 kernel: pci 0005:00:05.0: supports D1 D2 Aug 13 00:25:21.181724 kernel: pci 0005:00:05.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.181796 kernel: pci 0005:00:07.0: [1def:e117] type 01 class 0x060400 Aug 13 00:25:21.181861 kernel: pci 0005:00:07.0: supports D1 D2 Aug 13 00:25:21.181927 kernel: pci 0005:00:07.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.181938 kernel: acpiphp: Slot [1-2] registered Aug 13 00:25:21.181946 kernel: acpiphp: Slot [2-2] registered Aug 13 00:25:21.182018 kernel: pci 0005:03:00.0: [144d:a808] type 00 class 0x010802 Aug 13 00:25:21.182093 kernel: pci 0005:03:00.0: reg 0x10: [mem 0x30110000-0x30113fff 64bit] Aug 13 00:25:21.182159 kernel: pci 0005:03:00.0: reg 0x30: [mem 0x30100000-0x3010ffff pref] Aug 13 00:25:21.182234 kernel: pci 0005:04:00.0: [144d:a808] type 00 class 0x010802 Aug 13 00:25:21.182301 kernel: pci 0005:04:00.0: reg 0x10: [mem 0x30010000-0x30013fff 64bit] Aug 13 00:25:21.182367 kernel: pci 0005:04:00.0: reg 0x30: [mem 0x30000000-0x3000ffff pref] Aug 13 00:25:21.182426 kernel: pci_bus 0005:00: on NUMA node 0 Aug 13 00:25:21.182491 kernel: pci 0005:00:01.0: bridge window [io 0x1000-0x0fff] to [bus 01] add_size 1000 Aug 13 00:25:21.182559 kernel: pci 0005:00:01.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 01] add_size 200000 add_align 100000 Aug 13 00:25:21.182623 kernel: pci 0005:00:01.0: bridge window [mem 0x00100000-0x000fffff] to [bus 01] add_size 200000 add_align 100000 Aug 13 00:25:21.182689 kernel: pci 0005:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 02] add_size 1000 Aug 13 00:25:21.182753 kernel: pci 0005:00:03.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 02] add_size 200000 add_align 100000 Aug 13 00:25:21.182839 kernel: pci 0005:00:03.0: bridge window [mem 0x00100000-0x000fffff] to [bus 02] add_size 200000 add_align 100000 Aug 13 00:25:21.182920 kernel: pci 0005:00:05.0: bridge window [io 0x1000-0x0fff] to [bus 03] add_size 1000 Aug 13 00:25:21.182988 kernel: pci 0005:00:05.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 03] add_size 200000 add_align 100000 Aug 13 00:25:21.183057 kernel: pci 0005:00:05.0: bridge window [mem 0x00100000-0x001fffff] to [bus 03] add_size 100000 add_align 100000 Aug 13 00:25:21.183139 kernel: pci 0005:00:07.0: bridge window [io 0x1000-0x0fff] to [bus 04] add_size 1000 Aug 13 00:25:21.183204 kernel: pci 0005:00:07.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 04] add_size 200000 add_align 100000 Aug 13 00:25:21.183268 kernel: pci 0005:00:07.0: bridge window [mem 0x00100000-0x001fffff] to [bus 04] add_size 100000 add_align 100000 Aug 13 00:25:21.183333 kernel: pci 0005:00:01.0: BAR 14: assigned [mem 0x30000000-0x301fffff] Aug 13 00:25:21.183399 kernel: pci 0005:00:01.0: BAR 15: assigned [mem 0x2c0000000000-0x2c00001fffff 64bit pref] Aug 13 00:25:21.183466 kernel: pci 0005:00:03.0: BAR 14: assigned [mem 0x30200000-0x303fffff] Aug 13 00:25:21.183532 kernel: pci 0005:00:03.0: BAR 15: assigned [mem 0x2c0000200000-0x2c00003fffff 64bit pref] Aug 13 00:25:21.183597 kernel: pci 0005:00:05.0: BAR 14: assigned [mem 0x30400000-0x305fffff] Aug 13 00:25:21.183664 kernel: pci 0005:00:05.0: BAR 15: assigned [mem 0x2c0000400000-0x2c00005fffff 64bit pref] Aug 13 00:25:21.183729 kernel: pci 0005:00:07.0: BAR 14: assigned [mem 0x30600000-0x307fffff] Aug 13 00:25:21.183792 kernel: pci 0005:00:07.0: BAR 15: assigned [mem 0x2c0000600000-0x2c00007fffff 64bit pref] Aug 13 00:25:21.183856 kernel: pci 0005:00:01.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.183920 kernel: pci 0005:00:01.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.183988 kernel: pci 0005:00:03.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.184055 kernel: pci 0005:00:03.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.184121 kernel: pci 0005:00:05.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.184185 kernel: pci 0005:00:05.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.184250 kernel: pci 0005:00:07.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.184313 kernel: pci 0005:00:07.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.184377 kernel: pci 0005:00:07.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.184440 kernel: pci 0005:00:07.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.184507 kernel: pci 0005:00:05.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.184573 kernel: pci 0005:00:05.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.184637 kernel: pci 0005:00:03.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.184702 kernel: pci 0005:00:03.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.184766 kernel: pci 0005:00:01.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.184832 kernel: pci 0005:00:01.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.184895 kernel: pci 0005:00:01.0: PCI bridge to [bus 01] Aug 13 00:25:21.184961 kernel: pci 0005:00:01.0: bridge window [mem 0x30000000-0x301fffff] Aug 13 00:25:21.185024 kernel: pci 0005:00:01.0: bridge window [mem 0x2c0000000000-0x2c00001fffff 64bit pref] Aug 13 00:25:21.185095 kernel: pci 0005:00:03.0: PCI bridge to [bus 02] Aug 13 00:25:21.185159 kernel: pci 0005:00:03.0: bridge window [mem 0x30200000-0x303fffff] Aug 13 00:25:21.185224 kernel: pci 0005:00:03.0: bridge window [mem 0x2c0000200000-0x2c00003fffff 64bit pref] Aug 13 00:25:21.185291 kernel: pci 0005:03:00.0: BAR 6: assigned [mem 0x30400000-0x3040ffff pref] Aug 13 00:25:21.185361 kernel: pci 0005:03:00.0: BAR 0: assigned [mem 0x30410000-0x30413fff 64bit] Aug 13 00:25:21.185424 kernel: pci 0005:00:05.0: PCI bridge to [bus 03] Aug 13 00:25:21.185491 kernel: pci 0005:00:05.0: bridge window [mem 0x30400000-0x305fffff] Aug 13 00:25:21.185558 kernel: pci 0005:00:05.0: bridge window [mem 0x2c0000400000-0x2c00005fffff 64bit pref] Aug 13 00:25:21.185626 kernel: pci 0005:04:00.0: BAR 6: assigned [mem 0x30600000-0x3060ffff pref] Aug 13 00:25:21.185692 kernel: pci 0005:04:00.0: BAR 0: assigned [mem 0x30610000-0x30613fff 64bit] Aug 13 00:25:21.185755 kernel: pci 0005:00:07.0: PCI bridge to [bus 04] Aug 13 00:25:21.185819 kernel: pci 0005:00:07.0: bridge window [mem 0x30600000-0x307fffff] Aug 13 00:25:21.185882 kernel: pci 0005:00:07.0: bridge window [mem 0x2c0000600000-0x2c00007fffff 64bit pref] Aug 13 00:25:21.185944 kernel: pci_bus 0005:00: resource 4 [mem 0x30000000-0x3fffffff window] Aug 13 00:25:21.186001 kernel: pci_bus 0005:00: resource 5 [mem 0x2c0000000000-0x2fffdfffffff window] Aug 13 00:25:21.186075 kernel: pci_bus 0005:01: resource 1 [mem 0x30000000-0x301fffff] Aug 13 00:25:21.186135 kernel: pci_bus 0005:01: resource 2 [mem 0x2c0000000000-0x2c00001fffff 64bit pref] Aug 13 00:25:21.186211 kernel: pci_bus 0005:02: resource 1 [mem 0x30200000-0x303fffff] Aug 13 00:25:21.186270 kernel: pci_bus 0005:02: resource 2 [mem 0x2c0000200000-0x2c00003fffff 64bit pref] Aug 13 00:25:21.186343 kernel: pci_bus 0005:03: resource 1 [mem 0x30400000-0x305fffff] Aug 13 00:25:21.186402 kernel: pci_bus 0005:03: resource 2 [mem 0x2c0000400000-0x2c00005fffff 64bit pref] Aug 13 00:25:21.186470 kernel: pci_bus 0005:04: resource 1 [mem 0x30600000-0x307fffff] Aug 13 00:25:21.186530 kernel: pci_bus 0005:04: resource 2 [mem 0x2c0000600000-0x2c00007fffff 64bit pref] Aug 13 00:25:21.186540 kernel: ACPI: PCI Root Bridge [PCI5] (domain 0003 [bus 00-ff]) Aug 13 00:25:21.186611 kernel: acpi PNP0A08:03: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Aug 13 00:25:21.186677 kernel: acpi PNP0A08:03: _OSC: platform does not support [PCIeHotplug PME LTR] Aug 13 00:25:21.186740 kernel: acpi PNP0A08:03: _OSC: OS now controls [AER PCIeCapability] Aug 13 00:25:21.186802 kernel: acpi PNP0A08:03: MCFG quirk: ECAM at [mem 0x27fff0000000-0x27ffffffffff] for [bus 00-ff] with pci_32b_read_ops Aug 13 00:25:21.186864 kernel: acpi PNP0A08:03: ECAM area [mem 0x27fff0000000-0x27ffffffffff] reserved by PNP0C02:00 Aug 13 00:25:21.186926 kernel: acpi PNP0A08:03: ECAM at [mem 0x27fff0000000-0x27ffffffffff] for [bus 00-ff] Aug 13 00:25:21.186936 kernel: PCI host bridge to bus 0003:00 Aug 13 00:25:21.186999 kernel: pci_bus 0003:00: root bus resource [mem 0x10000000-0x1fffffff window] Aug 13 00:25:21.187075 kernel: pci_bus 0003:00: root bus resource [mem 0x240000000000-0x27ffdfffffff window] Aug 13 00:25:21.187134 kernel: pci_bus 0003:00: root bus resource [bus 00-ff] Aug 13 00:25:21.187211 kernel: pci 0003:00:00.0: [1def:e110] type 00 class 0x060000 Aug 13 00:25:21.187283 kernel: pci 0003:00:01.0: [1def:e111] type 01 class 0x060400 Aug 13 00:25:21.187351 kernel: pci 0003:00:01.0: supports D1 D2 Aug 13 00:25:21.187421 kernel: pci 0003:00:01.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.187499 kernel: pci 0003:00:03.0: [1def:e113] type 01 class 0x060400 Aug 13 00:25:21.187571 kernel: pci 0003:00:03.0: supports D1 D2 Aug 13 00:25:21.187639 kernel: pci 0003:00:03.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.187713 kernel: pci 0003:00:05.0: [1def:e115] type 01 class 0x060400 Aug 13 00:25:21.187778 kernel: pci 0003:00:05.0: supports D1 D2 Aug 13 00:25:21.187843 kernel: pci 0003:00:05.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.187853 kernel: acpiphp: Slot [1-3] registered Aug 13 00:25:21.187862 kernel: acpiphp: Slot [2-3] registered Aug 13 00:25:21.187935 kernel: pci 0003:03:00.0: [8086:1521] type 00 class 0x020000 Aug 13 00:25:21.188003 kernel: pci 0003:03:00.0: reg 0x10: [mem 0x10020000-0x1003ffff] Aug 13 00:25:21.188074 kernel: pci 0003:03:00.0: reg 0x18: [io 0x0020-0x003f] Aug 13 00:25:21.188141 kernel: pci 0003:03:00.0: reg 0x1c: [mem 0x10044000-0x10047fff] Aug 13 00:25:21.188207 kernel: pci 0003:03:00.0: PME# supported from D0 D3hot D3cold Aug 13 00:25:21.188273 kernel: pci 0003:03:00.0: reg 0x184: [mem 0x240000060000-0x240000063fff 64bit pref] Aug 13 00:25:21.188340 kernel: pci 0003:03:00.0: VF(n) BAR0 space: [mem 0x240000060000-0x24000007ffff 64bit pref] (contains BAR0 for 8 VFs) Aug 13 00:25:21.188411 kernel: pci 0003:03:00.0: reg 0x190: [mem 0x240000040000-0x240000043fff 64bit pref] Aug 13 00:25:21.188477 kernel: pci 0003:03:00.0: VF(n) BAR3 space: [mem 0x240000040000-0x24000005ffff 64bit pref] (contains BAR3 for 8 VFs) Aug 13 00:25:21.188543 kernel: pci 0003:03:00.0: 8.000 Gb/s available PCIe bandwidth, limited by 5.0 GT/s PCIe x2 link at 0003:00:05.0 (capable of 16.000 Gb/s with 5.0 GT/s PCIe x4 link) Aug 13 00:25:21.188621 kernel: pci 0003:03:00.1: [8086:1521] type 00 class 0x020000 Aug 13 00:25:21.188687 kernel: pci 0003:03:00.1: reg 0x10: [mem 0x10000000-0x1001ffff] Aug 13 00:25:21.188755 kernel: pci 0003:03:00.1: reg 0x18: [io 0x0000-0x001f] Aug 13 00:25:21.188821 kernel: pci 0003:03:00.1: reg 0x1c: [mem 0x10040000-0x10043fff] Aug 13 00:25:21.188888 kernel: pci 0003:03:00.1: PME# supported from D0 D3hot D3cold Aug 13 00:25:21.188955 kernel: pci 0003:03:00.1: reg 0x184: [mem 0x240000020000-0x240000023fff 64bit pref] Aug 13 00:25:21.189021 kernel: pci 0003:03:00.1: VF(n) BAR0 space: [mem 0x240000020000-0x24000003ffff 64bit pref] (contains BAR0 for 8 VFs) Aug 13 00:25:21.189183 kernel: pci 0003:03:00.1: reg 0x190: [mem 0x240000000000-0x240000003fff 64bit pref] Aug 13 00:25:21.189250 kernel: pci 0003:03:00.1: VF(n) BAR3 space: [mem 0x240000000000-0x24000001ffff 64bit pref] (contains BAR3 for 8 VFs) Aug 13 00:25:21.189309 kernel: pci_bus 0003:00: on NUMA node 0 Aug 13 00:25:21.189374 kernel: pci 0003:00:01.0: bridge window [io 0x1000-0x0fff] to [bus 01] add_size 1000 Aug 13 00:25:21.189442 kernel: pci 0003:00:01.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 01] add_size 200000 add_align 100000 Aug 13 00:25:21.189504 kernel: pci 0003:00:01.0: bridge window [mem 0x00100000-0x000fffff] to [bus 01] add_size 200000 add_align 100000 Aug 13 00:25:21.189569 kernel: pci 0003:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 02] add_size 1000 Aug 13 00:25:21.189632 kernel: pci 0003:00:03.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 02] add_size 200000 add_align 100000 Aug 13 00:25:21.189696 kernel: pci 0003:00:03.0: bridge window [mem 0x00100000-0x000fffff] to [bus 02] add_size 200000 add_align 100000 Aug 13 00:25:21.189760 kernel: pci 0003:00:05.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 03-04] add_size 300000 add_align 100000 Aug 13 00:25:21.189823 kernel: pci 0003:00:05.0: bridge window [mem 0x00100000-0x001fffff] to [bus 03-04] add_size 100000 add_align 100000 Aug 13 00:25:21.189889 kernel: pci 0003:00:01.0: BAR 14: assigned [mem 0x10000000-0x101fffff] Aug 13 00:25:21.189952 kernel: pci 0003:00:01.0: BAR 15: assigned [mem 0x240000000000-0x2400001fffff 64bit pref] Aug 13 00:25:21.190015 kernel: pci 0003:00:03.0: BAR 14: assigned [mem 0x10200000-0x103fffff] Aug 13 00:25:21.190081 kernel: pci 0003:00:03.0: BAR 15: assigned [mem 0x240000200000-0x2400003fffff 64bit pref] Aug 13 00:25:21.190144 kernel: pci 0003:00:05.0: BAR 14: assigned [mem 0x10400000-0x105fffff] Aug 13 00:25:21.190220 kernel: pci 0003:00:05.0: BAR 15: assigned [mem 0x240000400000-0x2400006fffff 64bit pref] Aug 13 00:25:21.190284 kernel: pci 0003:00:01.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.190346 kernel: pci 0003:00:01.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.190412 kernel: pci 0003:00:03.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.190475 kernel: pci 0003:00:03.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.190539 kernel: pci 0003:00:05.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.190601 kernel: pci 0003:00:05.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.190665 kernel: pci 0003:00:05.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.190727 kernel: pci 0003:00:05.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.190789 kernel: pci 0003:00:03.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.190853 kernel: pci 0003:00:03.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.190918 kernel: pci 0003:00:01.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.190983 kernel: pci 0003:00:01.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.191049 kernel: pci 0003:00:01.0: PCI bridge to [bus 01] Aug 13 00:25:21.191113 kernel: pci 0003:00:01.0: bridge window [mem 0x10000000-0x101fffff] Aug 13 00:25:21.191177 kernel: pci 0003:00:01.0: bridge window [mem 0x240000000000-0x2400001fffff 64bit pref] Aug 13 00:25:21.191240 kernel: pci 0003:00:03.0: PCI bridge to [bus 02] Aug 13 00:25:21.191302 kernel: pci 0003:00:03.0: bridge window [mem 0x10200000-0x103fffff] Aug 13 00:25:21.191368 kernel: pci 0003:00:03.0: bridge window [mem 0x240000200000-0x2400003fffff 64bit pref] Aug 13 00:25:21.191434 kernel: pci 0003:03:00.0: BAR 0: assigned [mem 0x10400000-0x1041ffff] Aug 13 00:25:21.191500 kernel: pci 0003:03:00.1: BAR 0: assigned [mem 0x10420000-0x1043ffff] Aug 13 00:25:21.191567 kernel: pci 0003:03:00.0: BAR 3: assigned [mem 0x10440000-0x10443fff] Aug 13 00:25:21.191633 kernel: pci 0003:03:00.0: BAR 7: assigned [mem 0x240000400000-0x24000041ffff 64bit pref] Aug 13 00:25:21.191699 kernel: pci 0003:03:00.0: BAR 10: assigned [mem 0x240000420000-0x24000043ffff 64bit pref] Aug 13 00:25:21.191766 kernel: pci 0003:03:00.1: BAR 3: assigned [mem 0x10444000-0x10447fff] Aug 13 00:25:21.191832 kernel: pci 0003:03:00.1: BAR 7: assigned [mem 0x240000440000-0x24000045ffff 64bit pref] Aug 13 00:25:21.191896 kernel: pci 0003:03:00.1: BAR 10: assigned [mem 0x240000460000-0x24000047ffff 64bit pref] Aug 13 00:25:21.191962 kernel: pci 0003:03:00.0: BAR 2: no space for [io size 0x0020] Aug 13 00:25:21.192027 kernel: pci 0003:03:00.0: BAR 2: failed to assign [io size 0x0020] Aug 13 00:25:21.192096 kernel: pci 0003:03:00.1: BAR 2: no space for [io size 0x0020] Aug 13 00:25:21.192162 kernel: pci 0003:03:00.1: BAR 2: failed to assign [io size 0x0020] Aug 13 00:25:21.192228 kernel: pci 0003:03:00.0: BAR 2: no space for [io size 0x0020] Aug 13 00:25:21.192293 kernel: pci 0003:03:00.0: BAR 2: failed to assign [io size 0x0020] Aug 13 00:25:21.192358 kernel: pci 0003:03:00.1: BAR 2: no space for [io size 0x0020] Aug 13 00:25:21.192423 kernel: pci 0003:03:00.1: BAR 2: failed to assign [io size 0x0020] Aug 13 00:25:21.192487 kernel: pci 0003:00:05.0: PCI bridge to [bus 03-04] Aug 13 00:25:21.192551 kernel: pci 0003:00:05.0: bridge window [mem 0x10400000-0x105fffff] Aug 13 00:25:21.192613 kernel: pci 0003:00:05.0: bridge window [mem 0x240000400000-0x2400006fffff 64bit pref] Aug 13 00:25:21.192672 kernel: pci_bus 0003:00: Some PCI device resources are unassigned, try booting with pci=realloc Aug 13 00:25:21.192731 kernel: pci_bus 0003:00: resource 4 [mem 0x10000000-0x1fffffff window] Aug 13 00:25:21.192787 kernel: pci_bus 0003:00: resource 5 [mem 0x240000000000-0x27ffdfffffff window] Aug 13 00:25:21.192864 kernel: pci_bus 0003:01: resource 1 [mem 0x10000000-0x101fffff] Aug 13 00:25:21.192924 kernel: pci_bus 0003:01: resource 2 [mem 0x240000000000-0x2400001fffff 64bit pref] Aug 13 00:25:21.192990 kernel: pci_bus 0003:02: resource 1 [mem 0x10200000-0x103fffff] Aug 13 00:25:21.193053 kernel: pci_bus 0003:02: resource 2 [mem 0x240000200000-0x2400003fffff 64bit pref] Aug 13 00:25:21.193122 kernel: pci_bus 0003:03: resource 1 [mem 0x10400000-0x105fffff] Aug 13 00:25:21.193180 kernel: pci_bus 0003:03: resource 2 [mem 0x240000400000-0x2400006fffff 64bit pref] Aug 13 00:25:21.193191 kernel: ACPI: PCI Root Bridge [PCI0] (domain 000c [bus 00-ff]) Aug 13 00:25:21.193261 kernel: acpi PNP0A08:04: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Aug 13 00:25:21.193323 kernel: acpi PNP0A08:04: _OSC: platform does not support [PCIeHotplug PME LTR] Aug 13 00:25:21.193386 kernel: acpi PNP0A08:04: _OSC: OS now controls [AER PCIeCapability] Aug 13 00:25:21.193448 kernel: acpi PNP0A08:04: MCFG quirk: ECAM at [mem 0x33fff0000000-0x33ffffffffff] for [bus 00-ff] with pci_32b_read_ops Aug 13 00:25:21.193514 kernel: acpi PNP0A08:04: ECAM area [mem 0x33fff0000000-0x33ffffffffff] reserved by PNP0C02:00 Aug 13 00:25:21.193575 kernel: acpi PNP0A08:04: ECAM at [mem 0x33fff0000000-0x33ffffffffff] for [bus 00-ff] Aug 13 00:25:21.193585 kernel: PCI host bridge to bus 000c:00 Aug 13 00:25:21.193652 kernel: pci_bus 000c:00: root bus resource [mem 0x40000000-0x4fffffff window] Aug 13 00:25:21.193709 kernel: pci_bus 000c:00: root bus resource [mem 0x300000000000-0x33ffdfffffff window] Aug 13 00:25:21.193767 kernel: pci_bus 000c:00: root bus resource [bus 00-ff] Aug 13 00:25:21.193840 kernel: pci 000c:00:00.0: [1def:e100] type 00 class 0x060000 Aug 13 00:25:21.193917 kernel: pci 000c:00:01.0: [1def:e101] type 01 class 0x060400 Aug 13 00:25:21.193984 kernel: pci 000c:00:01.0: enabling Extended Tags Aug 13 00:25:21.194052 kernel: pci 000c:00:01.0: supports D1 D2 Aug 13 00:25:21.194119 kernel: pci 000c:00:01.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.194190 kernel: pci 000c:00:02.0: [1def:e102] type 01 class 0x060400 Aug 13 00:25:21.194256 kernel: pci 000c:00:02.0: supports D1 D2 Aug 13 00:25:21.194320 kernel: pci 000c:00:02.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.194395 kernel: pci 000c:00:03.0: [1def:e103] type 01 class 0x060400 Aug 13 00:25:21.194462 kernel: pci 000c:00:03.0: supports D1 D2 Aug 13 00:25:21.194530 kernel: pci 000c:00:03.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.194601 kernel: pci 000c:00:04.0: [1def:e104] type 01 class 0x060400 Aug 13 00:25:21.194667 kernel: pci 000c:00:04.0: supports D1 D2 Aug 13 00:25:21.194731 kernel: pci 000c:00:04.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.194741 kernel: acpiphp: Slot [1-4] registered Aug 13 00:25:21.194753 kernel: acpiphp: Slot [2-4] registered Aug 13 00:25:21.194761 kernel: acpiphp: Slot [3-2] registered Aug 13 00:25:21.194769 kernel: acpiphp: Slot [4-2] registered Aug 13 00:25:21.194829 kernel: pci_bus 000c:00: on NUMA node 0 Aug 13 00:25:21.194893 kernel: pci 000c:00:01.0: bridge window [io 0x1000-0x0fff] to [bus 01] add_size 1000 Aug 13 00:25:21.194958 kernel: pci 000c:00:01.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 01] add_size 200000 add_align 100000 Aug 13 00:25:21.195022 kernel: pci 000c:00:01.0: bridge window [mem 0x00100000-0x000fffff] to [bus 01] add_size 200000 add_align 100000 Aug 13 00:25:21.195277 kernel: pci 000c:00:02.0: bridge window [io 0x1000-0x0fff] to [bus 02] add_size 1000 Aug 13 00:25:21.195348 kernel: pci 000c:00:02.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 02] add_size 200000 add_align 100000 Aug 13 00:25:21.195411 kernel: pci 000c:00:02.0: bridge window [mem 0x00100000-0x000fffff] to [bus 02] add_size 200000 add_align 100000 Aug 13 00:25:21.195475 kernel: pci 000c:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 03] add_size 1000 Aug 13 00:25:21.195539 kernel: pci 000c:00:03.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 03] add_size 200000 add_align 100000 Aug 13 00:25:21.195602 kernel: pci 000c:00:03.0: bridge window [mem 0x00100000-0x000fffff] to [bus 03] add_size 200000 add_align 100000 Aug 13 00:25:21.195666 kernel: pci 000c:00:04.0: bridge window [io 0x1000-0x0fff] to [bus 04] add_size 1000 Aug 13 00:25:21.195728 kernel: pci 000c:00:04.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 04] add_size 200000 add_align 100000 Aug 13 00:25:21.195794 kernel: pci 000c:00:04.0: bridge window [mem 0x00100000-0x000fffff] to [bus 04] add_size 200000 add_align 100000 Aug 13 00:25:21.195858 kernel: pci 000c:00:01.0: BAR 14: assigned [mem 0x40000000-0x401fffff] Aug 13 00:25:21.195922 kernel: pci 000c:00:01.0: BAR 15: assigned [mem 0x300000000000-0x3000001fffff 64bit pref] Aug 13 00:25:21.195984 kernel: pci 000c:00:02.0: BAR 14: assigned [mem 0x40200000-0x403fffff] Aug 13 00:25:21.196052 kernel: pci 000c:00:02.0: BAR 15: assigned [mem 0x300000200000-0x3000003fffff 64bit pref] Aug 13 00:25:21.196116 kernel: pci 000c:00:03.0: BAR 14: assigned [mem 0x40400000-0x405fffff] Aug 13 00:25:21.196180 kernel: pci 000c:00:03.0: BAR 15: assigned [mem 0x300000400000-0x3000005fffff 64bit pref] Aug 13 00:25:21.196247 kernel: pci 000c:00:04.0: BAR 14: assigned [mem 0x40600000-0x407fffff] Aug 13 00:25:21.196309 kernel: pci 000c:00:04.0: BAR 15: assigned [mem 0x300000600000-0x3000007fffff 64bit pref] Aug 13 00:25:21.196373 kernel: pci 000c:00:01.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.196436 kernel: pci 000c:00:01.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.196499 kernel: pci 000c:00:02.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.196562 kernel: pci 000c:00:02.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.196625 kernel: pci 000c:00:03.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.196688 kernel: pci 000c:00:03.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.196753 kernel: pci 000c:00:04.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.196816 kernel: pci 000c:00:04.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.196878 kernel: pci 000c:00:04.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.196940 kernel: pci 000c:00:04.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.197003 kernel: pci 000c:00:03.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.197070 kernel: pci 000c:00:03.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.197133 kernel: pci 000c:00:02.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.197196 kernel: pci 000c:00:02.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.197259 kernel: pci 000c:00:01.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.197325 kernel: pci 000c:00:01.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.197387 kernel: pci 000c:00:01.0: PCI bridge to [bus 01] Aug 13 00:25:21.197451 kernel: pci 000c:00:01.0: bridge window [mem 0x40000000-0x401fffff] Aug 13 00:25:21.197514 kernel: pci 000c:00:01.0: bridge window [mem 0x300000000000-0x3000001fffff 64bit pref] Aug 13 00:25:21.197578 kernel: pci 000c:00:02.0: PCI bridge to [bus 02] Aug 13 00:25:21.197641 kernel: pci 000c:00:02.0: bridge window [mem 0x40200000-0x403fffff] Aug 13 00:25:21.197706 kernel: pci 000c:00:02.0: bridge window [mem 0x300000200000-0x3000003fffff 64bit pref] Aug 13 00:25:21.197771 kernel: pci 000c:00:03.0: PCI bridge to [bus 03] Aug 13 00:25:21.197834 kernel: pci 000c:00:03.0: bridge window [mem 0x40400000-0x405fffff] Aug 13 00:25:21.197898 kernel: pci 000c:00:03.0: bridge window [mem 0x300000400000-0x3000005fffff 64bit pref] Aug 13 00:25:21.197961 kernel: pci 000c:00:04.0: PCI bridge to [bus 04] Aug 13 00:25:21.198024 kernel: pci 000c:00:04.0: bridge window [mem 0x40600000-0x407fffff] Aug 13 00:25:21.198093 kernel: pci 000c:00:04.0: bridge window [mem 0x300000600000-0x3000007fffff 64bit pref] Aug 13 00:25:21.198155 kernel: pci_bus 000c:00: resource 4 [mem 0x40000000-0x4fffffff window] Aug 13 00:25:21.198211 kernel: pci_bus 000c:00: resource 5 [mem 0x300000000000-0x33ffdfffffff window] Aug 13 00:25:21.198279 kernel: pci_bus 000c:01: resource 1 [mem 0x40000000-0x401fffff] Aug 13 00:25:21.198338 kernel: pci_bus 000c:01: resource 2 [mem 0x300000000000-0x3000001fffff 64bit pref] Aug 13 00:25:21.198412 kernel: pci_bus 000c:02: resource 1 [mem 0x40200000-0x403fffff] Aug 13 00:25:21.198471 kernel: pci_bus 000c:02: resource 2 [mem 0x300000200000-0x3000003fffff 64bit pref] Aug 13 00:25:21.198541 kernel: pci_bus 000c:03: resource 1 [mem 0x40400000-0x405fffff] Aug 13 00:25:21.198599 kernel: pci_bus 000c:03: resource 2 [mem 0x300000400000-0x3000005fffff 64bit pref] Aug 13 00:25:21.198666 kernel: pci_bus 000c:04: resource 1 [mem 0x40600000-0x407fffff] Aug 13 00:25:21.198724 kernel: pci_bus 000c:04: resource 2 [mem 0x300000600000-0x3000007fffff 64bit pref] Aug 13 00:25:21.198734 kernel: ACPI: PCI Root Bridge [PCI4] (domain 0002 [bus 00-ff]) Aug 13 00:25:21.198804 kernel: acpi PNP0A08:05: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Aug 13 00:25:21.198866 kernel: acpi PNP0A08:05: _OSC: platform does not support [PCIeHotplug PME LTR] Aug 13 00:25:21.198930 kernel: acpi PNP0A08:05: _OSC: OS now controls [AER PCIeCapability] Aug 13 00:25:21.198990 kernel: acpi PNP0A08:05: MCFG quirk: ECAM at [mem 0x23fff0000000-0x23ffffffffff] for [bus 00-ff] with pci_32b_read_ops Aug 13 00:25:21.199055 kernel: acpi PNP0A08:05: ECAM area [mem 0x23fff0000000-0x23ffffffffff] reserved by PNP0C02:00 Aug 13 00:25:21.199116 kernel: acpi PNP0A08:05: ECAM at [mem 0x23fff0000000-0x23ffffffffff] for [bus 00-ff] Aug 13 00:25:21.199127 kernel: PCI host bridge to bus 0002:00 Aug 13 00:25:21.199189 kernel: pci_bus 0002:00: root bus resource [mem 0x00800000-0x0fffffff window] Aug 13 00:25:21.199246 kernel: pci_bus 0002:00: root bus resource [mem 0x200000000000-0x23ffdfffffff window] Aug 13 00:25:21.199304 kernel: pci_bus 0002:00: root bus resource [bus 00-ff] Aug 13 00:25:21.199376 kernel: pci 0002:00:00.0: [1def:e110] type 00 class 0x060000 Aug 13 00:25:21.199447 kernel: pci 0002:00:01.0: [1def:e111] type 01 class 0x060400 Aug 13 00:25:21.199511 kernel: pci 0002:00:01.0: supports D1 D2 Aug 13 00:25:21.199574 kernel: pci 0002:00:01.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.199645 kernel: pci 0002:00:03.0: [1def:e113] type 01 class 0x060400 Aug 13 00:25:21.199713 kernel: pci 0002:00:03.0: supports D1 D2 Aug 13 00:25:21.199776 kernel: pci 0002:00:03.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.199844 kernel: pci 0002:00:05.0: [1def:e115] type 01 class 0x060400 Aug 13 00:25:21.199909 kernel: pci 0002:00:05.0: supports D1 D2 Aug 13 00:25:21.199971 kernel: pci 0002:00:05.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.200045 kernel: pci 0002:00:07.0: [1def:e117] type 01 class 0x060400 Aug 13 00:25:21.200109 kernel: pci 0002:00:07.0: supports D1 D2 Aug 13 00:25:21.200175 kernel: pci 0002:00:07.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.200185 kernel: acpiphp: Slot [1-5] registered Aug 13 00:25:21.200194 kernel: acpiphp: Slot [2-5] registered Aug 13 00:25:21.200202 kernel: acpiphp: Slot [3-3] registered Aug 13 00:25:21.200210 kernel: acpiphp: Slot [4-3] registered Aug 13 00:25:21.200264 kernel: pci_bus 0002:00: on NUMA node 0 Aug 13 00:25:21.200327 kernel: pci 0002:00:01.0: bridge window [io 0x1000-0x0fff] to [bus 01] add_size 1000 Aug 13 00:25:21.200391 kernel: pci 0002:00:01.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 01] add_size 200000 add_align 100000 Aug 13 00:25:21.200455 kernel: pci 0002:00:01.0: bridge window [mem 0x00100000-0x000fffff] to [bus 01] add_size 200000 add_align 100000 Aug 13 00:25:21.200521 kernel: pci 0002:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 02] add_size 1000 Aug 13 00:25:21.200584 kernel: pci 0002:00:03.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 02] add_size 200000 add_align 100000 Aug 13 00:25:21.200647 kernel: pci 0002:00:03.0: bridge window [mem 0x00100000-0x000fffff] to [bus 02] add_size 200000 add_align 100000 Aug 13 00:25:21.200711 kernel: pci 0002:00:05.0: bridge window [io 0x1000-0x0fff] to [bus 03] add_size 1000 Aug 13 00:25:21.200778 kernel: pci 0002:00:05.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 03] add_size 200000 add_align 100000 Aug 13 00:25:21.200843 kernel: pci 0002:00:05.0: bridge window [mem 0x00100000-0x000fffff] to [bus 03] add_size 200000 add_align 100000 Aug 13 00:25:21.200908 kernel: pci 0002:00:07.0: bridge window [io 0x1000-0x0fff] to [bus 04] add_size 1000 Aug 13 00:25:21.200972 kernel: pci 0002:00:07.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 04] add_size 200000 add_align 100000 Aug 13 00:25:21.201035 kernel: pci 0002:00:07.0: bridge window [mem 0x00100000-0x000fffff] to [bus 04] add_size 200000 add_align 100000 Aug 13 00:25:21.201102 kernel: pci 0002:00:01.0: BAR 14: assigned [mem 0x00800000-0x009fffff] Aug 13 00:25:21.201166 kernel: pci 0002:00:01.0: BAR 15: assigned [mem 0x200000000000-0x2000001fffff 64bit pref] Aug 13 00:25:21.201232 kernel: pci 0002:00:03.0: BAR 14: assigned [mem 0x00a00000-0x00bfffff] Aug 13 00:25:21.201296 kernel: pci 0002:00:03.0: BAR 15: assigned [mem 0x200000200000-0x2000003fffff 64bit pref] Aug 13 00:25:21.201360 kernel: pci 0002:00:05.0: BAR 14: assigned [mem 0x00c00000-0x00dfffff] Aug 13 00:25:21.201423 kernel: pci 0002:00:05.0: BAR 15: assigned [mem 0x200000400000-0x2000005fffff 64bit pref] Aug 13 00:25:21.201487 kernel: pci 0002:00:07.0: BAR 14: assigned [mem 0x00e00000-0x00ffffff] Aug 13 00:25:21.201549 kernel: pci 0002:00:07.0: BAR 15: assigned [mem 0x200000600000-0x2000007fffff 64bit pref] Aug 13 00:25:21.201613 kernel: pci 0002:00:01.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.201678 kernel: pci 0002:00:01.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.201742 kernel: pci 0002:00:03.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.201805 kernel: pci 0002:00:03.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.201868 kernel: pci 0002:00:05.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.201931 kernel: pci 0002:00:05.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.201998 kernel: pci 0002:00:07.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.202068 kernel: pci 0002:00:07.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.202132 kernel: pci 0002:00:07.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.202195 kernel: pci 0002:00:07.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.202265 kernel: pci 0002:00:05.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.202329 kernel: pci 0002:00:05.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.202394 kernel: pci 0002:00:03.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.202461 kernel: pci 0002:00:03.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.202525 kernel: pci 0002:00:01.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.202590 kernel: pci 0002:00:01.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.202655 kernel: pci 0002:00:01.0: PCI bridge to [bus 01] Aug 13 00:25:21.202719 kernel: pci 0002:00:01.0: bridge window [mem 0x00800000-0x009fffff] Aug 13 00:25:21.202784 kernel: pci 0002:00:01.0: bridge window [mem 0x200000000000-0x2000001fffff 64bit pref] Aug 13 00:25:21.202848 kernel: pci 0002:00:03.0: PCI bridge to [bus 02] Aug 13 00:25:21.202911 kernel: pci 0002:00:03.0: bridge window [mem 0x00a00000-0x00bfffff] Aug 13 00:25:21.202976 kernel: pci 0002:00:03.0: bridge window [mem 0x200000200000-0x2000003fffff 64bit pref] Aug 13 00:25:21.203137 kernel: pci 0002:00:05.0: PCI bridge to [bus 03] Aug 13 00:25:21.203218 kernel: pci 0002:00:05.0: bridge window [mem 0x00c00000-0x00dfffff] Aug 13 00:25:21.203288 kernel: pci 0002:00:05.0: bridge window [mem 0x200000400000-0x2000005fffff 64bit pref] Aug 13 00:25:21.203353 kernel: pci 0002:00:07.0: PCI bridge to [bus 04] Aug 13 00:25:21.203415 kernel: pci 0002:00:07.0: bridge window [mem 0x00e00000-0x00ffffff] Aug 13 00:25:21.203478 kernel: pci 0002:00:07.0: bridge window [mem 0x200000600000-0x2000007fffff 64bit pref] Aug 13 00:25:21.203538 kernel: pci_bus 0002:00: resource 4 [mem 0x00800000-0x0fffffff window] Aug 13 00:25:21.203595 kernel: pci_bus 0002:00: resource 5 [mem 0x200000000000-0x23ffdfffffff window] Aug 13 00:25:21.203663 kernel: pci_bus 0002:01: resource 1 [mem 0x00800000-0x009fffff] Aug 13 00:25:21.203772 kernel: pci_bus 0002:01: resource 2 [mem 0x200000000000-0x2000001fffff 64bit pref] Aug 13 00:25:21.203844 kernel: pci_bus 0002:02: resource 1 [mem 0x00a00000-0x00bfffff] Aug 13 00:25:21.203904 kernel: pci_bus 0002:02: resource 2 [mem 0x200000200000-0x2000003fffff 64bit pref] Aug 13 00:25:21.203978 kernel: pci_bus 0002:03: resource 1 [mem 0x00c00000-0x00dfffff] Aug 13 00:25:21.204044 kernel: pci_bus 0002:03: resource 2 [mem 0x200000400000-0x2000005fffff 64bit pref] Aug 13 00:25:21.204110 kernel: pci_bus 0002:04: resource 1 [mem 0x00e00000-0x00ffffff] Aug 13 00:25:21.204172 kernel: pci_bus 0002:04: resource 2 [mem 0x200000600000-0x2000007fffff 64bit pref] Aug 13 00:25:21.204183 kernel: ACPI: PCI Root Bridge [PCI2] (domain 0001 [bus 00-ff]) Aug 13 00:25:21.204253 kernel: acpi PNP0A08:06: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Aug 13 00:25:21.204315 kernel: acpi PNP0A08:06: _OSC: platform does not support [PCIeHotplug PME LTR] Aug 13 00:25:21.204377 kernel: acpi PNP0A08:06: _OSC: OS now controls [AER PCIeCapability] Aug 13 00:25:21.204437 kernel: acpi PNP0A08:06: MCFG quirk: ECAM at [mem 0x3bfff0000000-0x3bffffffffff] for [bus 00-ff] with pci_32b_read_ops Aug 13 00:25:21.204500 kernel: acpi PNP0A08:06: ECAM area [mem 0x3bfff0000000-0x3bffffffffff] reserved by PNP0C02:00 Aug 13 00:25:21.204563 kernel: acpi PNP0A08:06: ECAM at [mem 0x3bfff0000000-0x3bffffffffff] for [bus 00-ff] Aug 13 00:25:21.204574 kernel: PCI host bridge to bus 0001:00 Aug 13 00:25:21.204636 kernel: pci_bus 0001:00: root bus resource [mem 0x60000000-0x6fffffff window] Aug 13 00:25:21.204694 kernel: pci_bus 0001:00: root bus resource [mem 0x380000000000-0x3bffdfffffff window] Aug 13 00:25:21.204749 kernel: pci_bus 0001:00: root bus resource [bus 00-ff] Aug 13 00:25:21.204820 kernel: pci 0001:00:00.0: [1def:e100] type 00 class 0x060000 Aug 13 00:25:21.204894 kernel: pci 0001:00:01.0: [1def:e101] type 01 class 0x060400 Aug 13 00:25:21.204959 kernel: pci 0001:00:01.0: enabling Extended Tags Aug 13 00:25:21.205022 kernel: pci 0001:00:01.0: supports D1 D2 Aug 13 00:25:21.205090 kernel: pci 0001:00:01.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.205160 kernel: pci 0001:00:02.0: [1def:e102] type 01 class 0x060400 Aug 13 00:25:21.205225 kernel: pci 0001:00:02.0: supports D1 D2 Aug 13 00:25:21.205291 kernel: pci 0001:00:02.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.205362 kernel: pci 0001:00:03.0: [1def:e103] type 01 class 0x060400 Aug 13 00:25:21.205426 kernel: pci 0001:00:03.0: supports D1 D2 Aug 13 00:25:21.205489 kernel: pci 0001:00:03.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.205559 kernel: pci 0001:00:04.0: [1def:e104] type 01 class 0x060400 Aug 13 00:25:21.205624 kernel: pci 0001:00:04.0: supports D1 D2 Aug 13 00:25:21.205687 kernel: pci 0001:00:04.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.205699 kernel: acpiphp: Slot [1-6] registered Aug 13 00:25:21.205773 kernel: pci 0001:01:00.0: [15b3:1015] type 00 class 0x020000 Aug 13 00:25:21.205840 kernel: pci 0001:01:00.0: reg 0x10: [mem 0x380002000000-0x380003ffffff 64bit pref] Aug 13 00:25:21.205905 kernel: pci 0001:01:00.0: reg 0x30: [mem 0x60100000-0x601fffff pref] Aug 13 00:25:21.205971 kernel: pci 0001:01:00.0: PME# supported from D3cold Aug 13 00:25:21.206035 kernel: pci 0001:01:00.0: reg 0x1a4: [mem 0x380004800000-0x3800048fffff 64bit pref] Aug 13 00:25:21.206112 kernel: pci 0001:01:00.0: VF(n) BAR0 space: [mem 0x380004800000-0x380004ffffff 64bit pref] (contains BAR0 for 8 VFs) Aug 13 00:25:21.206179 kernel: pci 0001:01:00.0: 31.504 Gb/s available PCIe bandwidth, limited by 8.0 GT/s PCIe x4 link at 0001:00:01.0 (capable of 63.008 Gb/s with 8.0 GT/s PCIe x8 link) Aug 13 00:25:21.206258 kernel: pci 0001:01:00.1: [15b3:1015] type 00 class 0x020000 Aug 13 00:25:21.206326 kernel: pci 0001:01:00.1: reg 0x10: [mem 0x380000000000-0x380001ffffff 64bit pref] Aug 13 00:25:21.206391 kernel: pci 0001:01:00.1: reg 0x30: [mem 0x60000000-0x600fffff pref] Aug 13 00:25:21.206457 kernel: pci 0001:01:00.1: PME# supported from D3cold Aug 13 00:25:21.206521 kernel: pci 0001:01:00.1: reg 0x1a4: [mem 0x380004000000-0x3800040fffff 64bit pref] Aug 13 00:25:21.206586 kernel: pci 0001:01:00.1: VF(n) BAR0 space: [mem 0x380004000000-0x3800047fffff 64bit pref] (contains BAR0 for 8 VFs) Aug 13 00:25:21.206599 kernel: acpiphp: Slot [2-6] registered Aug 13 00:25:21.206607 kernel: acpiphp: Slot [3-4] registered Aug 13 00:25:21.206616 kernel: acpiphp: Slot [4-4] registered Aug 13 00:25:21.206672 kernel: pci_bus 0001:00: on NUMA node 0 Aug 13 00:25:21.206736 kernel: pci 0001:00:01.0: bridge window [io 0x1000-0x0fff] to [bus 01] add_size 1000 Aug 13 00:25:21.206815 kernel: pci 0001:00:01.0: bridge window [mem 0x02000000-0x05ffffff 64bit pref] to [bus 01] add_size 2000000 add_align 2000000 Aug 13 00:25:21.206881 kernel: pci 0001:00:02.0: bridge window [io 0x1000-0x0fff] to [bus 02] add_size 1000 Aug 13 00:25:21.206944 kernel: pci 0001:00:02.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 02] add_size 200000 add_align 100000 Aug 13 00:25:21.207011 kernel: pci 0001:00:02.0: bridge window [mem 0x00100000-0x000fffff] to [bus 02] add_size 200000 add_align 100000 Aug 13 00:25:21.207200 kernel: pci 0001:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 03] add_size 1000 Aug 13 00:25:21.207269 kernel: pci 0001:00:03.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 03] add_size 200000 add_align 100000 Aug 13 00:25:21.207332 kernel: pci 0001:00:03.0: bridge window [mem 0x00100000-0x000fffff] to [bus 03] add_size 200000 add_align 100000 Aug 13 00:25:21.207397 kernel: pci 0001:00:04.0: bridge window [io 0x1000-0x0fff] to [bus 04] add_size 1000 Aug 13 00:25:21.207460 kernel: pci 0001:00:04.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 04] add_size 200000 add_align 100000 Aug 13 00:25:21.207523 kernel: pci 0001:00:04.0: bridge window [mem 0x00100000-0x000fffff] to [bus 04] add_size 200000 add_align 100000 Aug 13 00:25:21.207591 kernel: pci 0001:00:01.0: BAR 15: assigned [mem 0x380000000000-0x380005ffffff 64bit pref] Aug 13 00:25:21.207655 kernel: pci 0001:00:01.0: BAR 14: assigned [mem 0x60000000-0x601fffff] Aug 13 00:25:21.207720 kernel: pci 0001:00:02.0: BAR 14: assigned [mem 0x60200000-0x603fffff] Aug 13 00:25:21.207783 kernel: pci 0001:00:02.0: BAR 15: assigned [mem 0x380006000000-0x3800061fffff 64bit pref] Aug 13 00:25:21.207849 kernel: pci 0001:00:03.0: BAR 14: assigned [mem 0x60400000-0x605fffff] Aug 13 00:25:21.207912 kernel: pci 0001:00:03.0: BAR 15: assigned [mem 0x380006200000-0x3800063fffff 64bit pref] Aug 13 00:25:21.207975 kernel: pci 0001:00:04.0: BAR 14: assigned [mem 0x60600000-0x607fffff] Aug 13 00:25:21.208042 kernel: pci 0001:00:04.0: BAR 15: assigned [mem 0x380006400000-0x3800065fffff 64bit pref] Aug 13 00:25:21.208106 kernel: pci 0001:00:01.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.208169 kernel: pci 0001:00:01.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.208233 kernel: pci 0001:00:02.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.208295 kernel: pci 0001:00:02.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.208359 kernel: pci 0001:00:03.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.208421 kernel: pci 0001:00:03.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.208484 kernel: pci 0001:00:04.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.208550 kernel: pci 0001:00:04.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.208613 kernel: pci 0001:00:04.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.208676 kernel: pci 0001:00:04.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.208740 kernel: pci 0001:00:03.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.208801 kernel: pci 0001:00:03.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.208864 kernel: pci 0001:00:02.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.208927 kernel: pci 0001:00:02.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.208990 kernel: pci 0001:00:01.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.209058 kernel: pci 0001:00:01.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.209124 kernel: pci 0001:01:00.0: BAR 0: assigned [mem 0x380000000000-0x380001ffffff 64bit pref] Aug 13 00:25:21.209195 kernel: pci 0001:01:00.1: BAR 0: assigned [mem 0x380002000000-0x380003ffffff 64bit pref] Aug 13 00:25:21.209260 kernel: pci 0001:01:00.0: BAR 6: assigned [mem 0x60000000-0x600fffff pref] Aug 13 00:25:21.209325 kernel: pci 0001:01:00.0: BAR 7: assigned [mem 0x380004000000-0x3800047fffff 64bit pref] Aug 13 00:25:21.209389 kernel: pci 0001:01:00.1: BAR 6: assigned [mem 0x60100000-0x601fffff pref] Aug 13 00:25:21.209455 kernel: pci 0001:01:00.1: BAR 7: assigned [mem 0x380004800000-0x380004ffffff 64bit pref] Aug 13 00:25:21.209517 kernel: pci 0001:00:01.0: PCI bridge to [bus 01] Aug 13 00:25:21.209580 kernel: pci 0001:00:01.0: bridge window [mem 0x60000000-0x601fffff] Aug 13 00:25:21.209647 kernel: pci 0001:00:01.0: bridge window [mem 0x380000000000-0x380005ffffff 64bit pref] Aug 13 00:25:21.209711 kernel: pci 0001:00:02.0: PCI bridge to [bus 02] Aug 13 00:25:21.209773 kernel: pci 0001:00:02.0: bridge window [mem 0x60200000-0x603fffff] Aug 13 00:25:21.209836 kernel: pci 0001:00:02.0: bridge window [mem 0x380006000000-0x3800061fffff 64bit pref] Aug 13 00:25:21.209900 kernel: pci 0001:00:03.0: PCI bridge to [bus 03] Aug 13 00:25:21.209962 kernel: pci 0001:00:03.0: bridge window [mem 0x60400000-0x605fffff] Aug 13 00:25:21.210028 kernel: pci 0001:00:03.0: bridge window [mem 0x380006200000-0x3800063fffff 64bit pref] Aug 13 00:25:21.210095 kernel: pci 0001:00:04.0: PCI bridge to [bus 04] Aug 13 00:25:21.210159 kernel: pci 0001:00:04.0: bridge window [mem 0x60600000-0x607fffff] Aug 13 00:25:21.210222 kernel: pci 0001:00:04.0: bridge window [mem 0x380006400000-0x3800065fffff 64bit pref] Aug 13 00:25:21.210281 kernel: pci_bus 0001:00: resource 4 [mem 0x60000000-0x6fffffff window] Aug 13 00:25:21.210337 kernel: pci_bus 0001:00: resource 5 [mem 0x380000000000-0x3bffdfffffff window] Aug 13 00:25:21.210413 kernel: pci_bus 0001:01: resource 1 [mem 0x60000000-0x601fffff] Aug 13 00:25:21.210475 kernel: pci_bus 0001:01: resource 2 [mem 0x380000000000-0x380005ffffff 64bit pref] Aug 13 00:25:21.210542 kernel: pci_bus 0001:02: resource 1 [mem 0x60200000-0x603fffff] Aug 13 00:25:21.210600 kernel: pci_bus 0001:02: resource 2 [mem 0x380006000000-0x3800061fffff 64bit pref] Aug 13 00:25:21.210666 kernel: pci_bus 0001:03: resource 1 [mem 0x60400000-0x605fffff] Aug 13 00:25:21.210724 kernel: pci_bus 0001:03: resource 2 [mem 0x380006200000-0x3800063fffff 64bit pref] Aug 13 00:25:21.210790 kernel: pci_bus 0001:04: resource 1 [mem 0x60600000-0x607fffff] Aug 13 00:25:21.210852 kernel: pci_bus 0001:04: resource 2 [mem 0x380006400000-0x3800065fffff 64bit pref] Aug 13 00:25:21.210863 kernel: ACPI: PCI Root Bridge [PCI6] (domain 0004 [bus 00-ff]) Aug 13 00:25:21.210931 kernel: acpi PNP0A08:07: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Aug 13 00:25:21.210994 kernel: acpi PNP0A08:07: _OSC: platform does not support [PCIeHotplug PME LTR] Aug 13 00:25:21.211060 kernel: acpi PNP0A08:07: _OSC: OS now controls [AER PCIeCapability] Aug 13 00:25:21.211122 kernel: acpi PNP0A08:07: MCFG quirk: ECAM at [mem 0x2bfff0000000-0x2bffffffffff] for [bus 00-ff] with pci_32b_read_ops Aug 13 00:25:21.211182 kernel: acpi PNP0A08:07: ECAM area [mem 0x2bfff0000000-0x2bffffffffff] reserved by PNP0C02:00 Aug 13 00:25:21.211246 kernel: acpi PNP0A08:07: ECAM at [mem 0x2bfff0000000-0x2bffffffffff] for [bus 00-ff] Aug 13 00:25:21.211257 kernel: PCI host bridge to bus 0004:00 Aug 13 00:25:21.211319 kernel: pci_bus 0004:00: root bus resource [mem 0x20000000-0x2fffffff window] Aug 13 00:25:21.211376 kernel: pci_bus 0004:00: root bus resource [mem 0x280000000000-0x2bffdfffffff window] Aug 13 00:25:21.211432 kernel: pci_bus 0004:00: root bus resource [bus 00-ff] Aug 13 00:25:21.211502 kernel: pci 0004:00:00.0: [1def:e110] type 00 class 0x060000 Aug 13 00:25:21.211577 kernel: pci 0004:00:01.0: [1def:e111] type 01 class 0x060400 Aug 13 00:25:21.211643 kernel: pci 0004:00:01.0: supports D1 D2 Aug 13 00:25:21.211708 kernel: pci 0004:00:01.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.211778 kernel: pci 0004:00:03.0: [1def:e113] type 01 class 0x060400 Aug 13 00:25:21.211843 kernel: pci 0004:00:03.0: supports D1 D2 Aug 13 00:25:21.211906 kernel: pci 0004:00:03.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.211977 kernel: pci 0004:00:05.0: [1def:e115] type 01 class 0x060400 Aug 13 00:25:21.212048 kernel: pci 0004:00:05.0: supports D1 D2 Aug 13 00:25:21.212112 kernel: pci 0004:00:05.0: PME# supported from D0 D1 D3hot Aug 13 00:25:21.212184 kernel: pci 0004:01:00.0: [1a03:1150] type 01 class 0x060400 Aug 13 00:25:21.212251 kernel: pci 0004:01:00.0: enabling Extended Tags Aug 13 00:25:21.212315 kernel: pci 0004:01:00.0: supports D1 D2 Aug 13 00:25:21.212380 kernel: pci 0004:01:00.0: PME# supported from D0 D1 D2 D3hot D3cold Aug 13 00:25:21.212457 kernel: pci_bus 0004:02: extended config space not accessible Aug 13 00:25:21.212536 kernel: pci 0004:02:00.0: [1a03:2000] type 00 class 0x030000 Aug 13 00:25:21.212605 kernel: pci 0004:02:00.0: reg 0x10: [mem 0x20000000-0x21ffffff] Aug 13 00:25:21.212672 kernel: pci 0004:02:00.0: reg 0x14: [mem 0x22000000-0x2201ffff] Aug 13 00:25:21.212740 kernel: pci 0004:02:00.0: reg 0x18: [io 0x0000-0x007f] Aug 13 00:25:21.212807 kernel: pci 0004:02:00.0: BAR 0: assigned to efifb Aug 13 00:25:21.212874 kernel: pci 0004:02:00.0: supports D1 D2 Aug 13 00:25:21.212942 kernel: pci 0004:02:00.0: PME# supported from D0 D1 D2 D3hot D3cold Aug 13 00:25:21.213019 kernel: pci 0004:03:00.0: [1912:0014] type 00 class 0x0c0330 Aug 13 00:25:21.213090 kernel: pci 0004:03:00.0: reg 0x10: [mem 0x22200000-0x22201fff 64bit] Aug 13 00:25:21.213155 kernel: pci 0004:03:00.0: PME# supported from D0 D3hot D3cold Aug 13 00:25:21.213215 kernel: pci_bus 0004:00: on NUMA node 0 Aug 13 00:25:21.213280 kernel: pci 0004:00:01.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 01-02] add_size 200000 add_align 100000 Aug 13 00:25:21.213345 kernel: pci 0004:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 03] add_size 1000 Aug 13 00:25:21.213408 kernel: pci 0004:00:03.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 03] add_size 200000 add_align 100000 Aug 13 00:25:21.213472 kernel: pci 0004:00:03.0: bridge window [mem 0x00100000-0x001fffff] to [bus 03] add_size 100000 add_align 100000 Aug 13 00:25:21.213540 kernel: pci 0004:00:05.0: bridge window [io 0x1000-0x0fff] to [bus 04] add_size 1000 Aug 13 00:25:21.213604 kernel: pci 0004:00:05.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 04] add_size 200000 add_align 100000 Aug 13 00:25:21.213668 kernel: pci 0004:00:05.0: bridge window [mem 0x00100000-0x000fffff] to [bus 04] add_size 200000 add_align 100000 Aug 13 00:25:21.213731 kernel: pci 0004:00:01.0: BAR 14: assigned [mem 0x20000000-0x22ffffff] Aug 13 00:25:21.213795 kernel: pci 0004:00:01.0: BAR 15: assigned [mem 0x280000000000-0x2800001fffff 64bit pref] Aug 13 00:25:21.213858 kernel: pci 0004:00:03.0: BAR 14: assigned [mem 0x23000000-0x231fffff] Aug 13 00:25:21.213924 kernel: pci 0004:00:03.0: BAR 15: assigned [mem 0x280000200000-0x2800003fffff 64bit pref] Aug 13 00:25:21.213987 kernel: pci 0004:00:05.0: BAR 14: assigned [mem 0x23200000-0x233fffff] Aug 13 00:25:21.214055 kernel: pci 0004:00:05.0: BAR 15: assigned [mem 0x280000400000-0x2800005fffff 64bit pref] Aug 13 00:25:21.214119 kernel: pci 0004:00:01.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.214184 kernel: pci 0004:00:01.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.214247 kernel: pci 0004:00:03.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.214311 kernel: pci 0004:00:03.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.214374 kernel: pci 0004:00:05.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.214441 kernel: pci 0004:00:05.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.214506 kernel: pci 0004:00:01.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.214570 kernel: pci 0004:00:01.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.214634 kernel: pci 0004:00:05.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.214697 kernel: pci 0004:00:05.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.214761 kernel: pci 0004:00:03.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.214824 kernel: pci 0004:00:03.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.214891 kernel: pci 0004:01:00.0: BAR 14: assigned [mem 0x20000000-0x22ffffff] Aug 13 00:25:21.214956 kernel: pci 0004:01:00.0: BAR 13: no space for [io size 0x1000] Aug 13 00:25:21.215025 kernel: pci 0004:01:00.0: BAR 13: failed to assign [io size 0x1000] Aug 13 00:25:21.215097 kernel: pci 0004:02:00.0: BAR 0: assigned [mem 0x20000000-0x21ffffff] Aug 13 00:25:21.215166 kernel: pci 0004:02:00.0: BAR 1: assigned [mem 0x22000000-0x2201ffff] Aug 13 00:25:21.215235 kernel: pci 0004:02:00.0: BAR 2: no space for [io size 0x0080] Aug 13 00:25:21.215302 kernel: pci 0004:02:00.0: BAR 2: failed to assign [io size 0x0080] Aug 13 00:25:21.215367 kernel: pci 0004:01:00.0: PCI bridge to [bus 02] Aug 13 00:25:21.215432 kernel: pci 0004:01:00.0: bridge window [mem 0x20000000-0x22ffffff] Aug 13 00:25:21.215497 kernel: pci 0004:00:01.0: PCI bridge to [bus 01-02] Aug 13 00:25:21.215562 kernel: pci 0004:00:01.0: bridge window [mem 0x20000000-0x22ffffff] Aug 13 00:25:21.215625 kernel: pci 0004:00:01.0: bridge window [mem 0x280000000000-0x2800001fffff 64bit pref] Aug 13 00:25:21.215694 kernel: pci 0004:03:00.0: BAR 0: assigned [mem 0x23000000-0x23001fff 64bit] Aug 13 00:25:21.215758 kernel: pci 0004:00:03.0: PCI bridge to [bus 03] Aug 13 00:25:21.215822 kernel: pci 0004:00:03.0: bridge window [mem 0x23000000-0x231fffff] Aug 13 00:25:21.215885 kernel: pci 0004:00:03.0: bridge window [mem 0x280000200000-0x2800003fffff 64bit pref] Aug 13 00:25:21.215949 kernel: pci 0004:00:05.0: PCI bridge to [bus 04] Aug 13 00:25:21.216014 kernel: pci 0004:00:05.0: bridge window [mem 0x23200000-0x233fffff] Aug 13 00:25:21.216082 kernel: pci 0004:00:05.0: bridge window [mem 0x280000400000-0x2800005fffff 64bit pref] Aug 13 00:25:21.216141 kernel: pci_bus 0004:00: Some PCI device resources are unassigned, try booting with pci=realloc Aug 13 00:25:21.216201 kernel: pci_bus 0004:00: resource 4 [mem 0x20000000-0x2fffffff window] Aug 13 00:25:21.216258 kernel: pci_bus 0004:00: resource 5 [mem 0x280000000000-0x2bffdfffffff window] Aug 13 00:25:21.216328 kernel: pci_bus 0004:01: resource 1 [mem 0x20000000-0x22ffffff] Aug 13 00:25:21.216390 kernel: pci_bus 0004:01: resource 2 [mem 0x280000000000-0x2800001fffff 64bit pref] Aug 13 00:25:21.216458 kernel: pci_bus 0004:02: resource 1 [mem 0x20000000-0x22ffffff] Aug 13 00:25:21.216525 kernel: pci_bus 0004:03: resource 1 [mem 0x23000000-0x231fffff] Aug 13 00:25:21.216585 kernel: pci_bus 0004:03: resource 2 [mem 0x280000200000-0x2800003fffff 64bit pref] Aug 13 00:25:21.216652 kernel: pci_bus 0004:04: resource 1 [mem 0x23200000-0x233fffff] Aug 13 00:25:21.216713 kernel: pci_bus 0004:04: resource 2 [mem 0x280000400000-0x2800005fffff 64bit pref] Aug 13 00:25:21.216723 kernel: iommu: Default domain type: Translated Aug 13 00:25:21.216734 kernel: iommu: DMA domain TLB invalidation policy: strict mode Aug 13 00:25:21.216742 kernel: efivars: Registered efivars operations Aug 13 00:25:21.216811 kernel: pci 0004:02:00.0: vgaarb: setting as boot VGA device Aug 13 00:25:21.216880 kernel: pci 0004:02:00.0: vgaarb: bridge control possible Aug 13 00:25:21.216950 kernel: pci 0004:02:00.0: vgaarb: VGA device added: decodes=io+mem,owns=none,locks=none Aug 13 00:25:21.216961 kernel: vgaarb: loaded Aug 13 00:25:21.216970 kernel: clocksource: Switched to clocksource arch_sys_counter Aug 13 00:25:21.216980 kernel: VFS: Disk quotas dquot_6.6.0 Aug 13 00:25:21.216988 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Aug 13 00:25:21.216996 kernel: pnp: PnP ACPI init Aug 13 00:25:21.217068 kernel: system 00:00: [mem 0x3bfff0000000-0x3bffffffffff window] could not be reserved Aug 13 00:25:21.217130 kernel: system 00:00: [mem 0x3ffff0000000-0x3fffffffffff window] could not be reserved Aug 13 00:25:21.217189 kernel: system 00:00: [mem 0x23fff0000000-0x23ffffffffff window] could not be reserved Aug 13 00:25:21.217248 kernel: system 00:00: [mem 0x27fff0000000-0x27ffffffffff window] could not be reserved Aug 13 00:25:21.217305 kernel: system 00:00: [mem 0x2bfff0000000-0x2bffffffffff window] could not be reserved Aug 13 00:25:21.217369 kernel: system 00:00: [mem 0x2ffff0000000-0x2fffffffffff window] could not be reserved Aug 13 00:25:21.217428 kernel: system 00:00: [mem 0x33fff0000000-0x33ffffffffff window] could not be reserved Aug 13 00:25:21.217487 kernel: system 00:00: [mem 0x37fff0000000-0x37ffffffffff window] could not be reserved Aug 13 00:25:21.217497 kernel: pnp: PnP ACPI: found 1 devices Aug 13 00:25:21.217505 kernel: NET: Registered PF_INET protocol family Aug 13 00:25:21.217513 kernel: IP idents hash table entries: 262144 (order: 9, 2097152 bytes, linear) Aug 13 00:25:21.217521 kernel: tcp_listen_portaddr_hash hash table entries: 65536 (order: 8, 1048576 bytes, linear) Aug 13 00:25:21.217532 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Aug 13 00:25:21.217540 kernel: TCP established hash table entries: 524288 (order: 10, 4194304 bytes, linear) Aug 13 00:25:21.217548 kernel: TCP bind hash table entries: 65536 (order: 9, 2097152 bytes, linear) Aug 13 00:25:21.217556 kernel: TCP: Hash tables configured (established 524288 bind 65536) Aug 13 00:25:21.217565 kernel: UDP hash table entries: 65536 (order: 9, 2097152 bytes, linear) Aug 13 00:25:21.217573 kernel: UDP-Lite hash table entries: 65536 (order: 9, 2097152 bytes, linear) Aug 13 00:25:21.217581 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Aug 13 00:25:21.217650 kernel: pci 0001:01:00.0: CLS mismatch (64 != 32), using 64 bytes Aug 13 00:25:21.217661 kernel: kvm [1]: IPA Size Limit: 48 bits Aug 13 00:25:21.217671 kernel: kvm [1]: GICv3: no GICV resource entry Aug 13 00:25:21.217679 kernel: kvm [1]: disabling GICv2 emulation Aug 13 00:25:21.217687 kernel: kvm [1]: GIC system register CPU interface enabled Aug 13 00:25:21.217695 kernel: kvm [1]: vgic interrupt IRQ9 Aug 13 00:25:21.217703 kernel: kvm [1]: VHE mode initialized successfully Aug 13 00:25:21.217711 kernel: Initialise system trusted keyrings Aug 13 00:25:21.217719 kernel: workingset: timestamp_bits=39 max_order=26 bucket_order=0 Aug 13 00:25:21.217728 kernel: Key type asymmetric registered Aug 13 00:25:21.217735 kernel: Asymmetric key parser 'x509' registered Aug 13 00:25:21.217745 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Aug 13 00:25:21.217753 kernel: io scheduler mq-deadline registered Aug 13 00:25:21.217761 kernel: io scheduler kyber registered Aug 13 00:25:21.217771 kernel: io scheduler bfq registered Aug 13 00:25:21.217779 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0C:00/input/input0 Aug 13 00:25:21.217787 kernel: ACPI: button: Power Button [PWRB] Aug 13 00:25:21.217795 kernel: ACPI GTDT: found 1 SBSA generic Watchdog(s). Aug 13 00:25:21.217803 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Aug 13 00:25:21.217875 kernel: arm-smmu-v3 arm-smmu-v3.0.auto: option mask 0x0 Aug 13 00:25:21.217940 kernel: arm-smmu-v3 arm-smmu-v3.0.auto: IDR0.COHACC overridden by FW configuration (false) Aug 13 00:25:21.218001 kernel: arm-smmu-v3 arm-smmu-v3.0.auto: ias 48-bit, oas 48-bit (features 0x000c1eff) Aug 13 00:25:21.218067 kernel: arm-smmu-v3 arm-smmu-v3.0.auto: allocated 262144 entries for cmdq Aug 13 00:25:21.218126 kernel: arm-smmu-v3 arm-smmu-v3.0.auto: allocated 131072 entries for evtq Aug 13 00:25:21.218187 kernel: arm-smmu-v3 arm-smmu-v3.0.auto: allocated 262144 entries for priq Aug 13 00:25:21.218254 kernel: arm-smmu-v3 arm-smmu-v3.1.auto: option mask 0x0 Aug 13 00:25:21.218318 kernel: arm-smmu-v3 arm-smmu-v3.1.auto: IDR0.COHACC overridden by FW configuration (false) Aug 13 00:25:21.218378 kernel: arm-smmu-v3 arm-smmu-v3.1.auto: ias 48-bit, oas 48-bit (features 0x000c1eff) Aug 13 00:25:21.218438 kernel: arm-smmu-v3 arm-smmu-v3.1.auto: allocated 262144 entries for cmdq Aug 13 00:25:21.218496 kernel: arm-smmu-v3 arm-smmu-v3.1.auto: allocated 131072 entries for evtq Aug 13 00:25:21.218556 kernel: arm-smmu-v3 arm-smmu-v3.1.auto: allocated 262144 entries for priq Aug 13 00:25:21.218623 kernel: arm-smmu-v3 arm-smmu-v3.2.auto: option mask 0x0 Aug 13 00:25:21.218684 kernel: arm-smmu-v3 arm-smmu-v3.2.auto: IDR0.COHACC overridden by FW configuration (false) Aug 13 00:25:21.218747 kernel: arm-smmu-v3 arm-smmu-v3.2.auto: ias 48-bit, oas 48-bit (features 0x000c1eff) Aug 13 00:25:21.218806 kernel: arm-smmu-v3 arm-smmu-v3.2.auto: allocated 262144 entries for cmdq Aug 13 00:25:21.218867 kernel: arm-smmu-v3 arm-smmu-v3.2.auto: allocated 131072 entries for evtq Aug 13 00:25:21.218926 kernel: arm-smmu-v3 arm-smmu-v3.2.auto: allocated 262144 entries for priq Aug 13 00:25:21.218993 kernel: arm-smmu-v3 arm-smmu-v3.3.auto: option mask 0x0 Aug 13 00:25:21.219197 kernel: arm-smmu-v3 arm-smmu-v3.3.auto: IDR0.COHACC overridden by FW configuration (false) Aug 13 00:25:21.219265 kernel: arm-smmu-v3 arm-smmu-v3.3.auto: ias 48-bit, oas 48-bit (features 0x000c1eff) Aug 13 00:25:21.219324 kernel: arm-smmu-v3 arm-smmu-v3.3.auto: allocated 262144 entries for cmdq Aug 13 00:25:21.219382 kernel: arm-smmu-v3 arm-smmu-v3.3.auto: allocated 131072 entries for evtq Aug 13 00:25:21.219441 kernel: arm-smmu-v3 arm-smmu-v3.3.auto: allocated 262144 entries for priq Aug 13 00:25:21.219514 kernel: arm-smmu-v3 arm-smmu-v3.4.auto: option mask 0x0 Aug 13 00:25:21.219573 kernel: arm-smmu-v3 arm-smmu-v3.4.auto: IDR0.COHACC overridden by FW configuration (false) Aug 13 00:25:21.219632 kernel: arm-smmu-v3 arm-smmu-v3.4.auto: ias 48-bit, oas 48-bit (features 0x000c1eff) Aug 13 00:25:21.219692 kernel: arm-smmu-v3 arm-smmu-v3.4.auto: allocated 262144 entries for cmdq Aug 13 00:25:21.219751 kernel: arm-smmu-v3 arm-smmu-v3.4.auto: allocated 131072 entries for evtq Aug 13 00:25:21.219809 kernel: arm-smmu-v3 arm-smmu-v3.4.auto: allocated 262144 entries for priq Aug 13 00:25:21.219876 kernel: arm-smmu-v3 arm-smmu-v3.5.auto: option mask 0x0 Aug 13 00:25:21.219935 kernel: arm-smmu-v3 arm-smmu-v3.5.auto: IDR0.COHACC overridden by FW configuration (false) Aug 13 00:25:21.219993 kernel: arm-smmu-v3 arm-smmu-v3.5.auto: ias 48-bit, oas 48-bit (features 0x000c1eff) Aug 13 00:25:21.220055 kernel: arm-smmu-v3 arm-smmu-v3.5.auto: allocated 262144 entries for cmdq Aug 13 00:25:21.220117 kernel: arm-smmu-v3 arm-smmu-v3.5.auto: allocated 131072 entries for evtq Aug 13 00:25:21.220174 kernel: arm-smmu-v3 arm-smmu-v3.5.auto: allocated 262144 entries for priq Aug 13 00:25:21.220241 kernel: arm-smmu-v3 arm-smmu-v3.6.auto: option mask 0x0 Aug 13 00:25:21.220300 kernel: arm-smmu-v3 arm-smmu-v3.6.auto: IDR0.COHACC overridden by FW configuration (false) Aug 13 00:25:21.220359 kernel: arm-smmu-v3 arm-smmu-v3.6.auto: ias 48-bit, oas 48-bit (features 0x000c1eff) Aug 13 00:25:21.220418 kernel: arm-smmu-v3 arm-smmu-v3.6.auto: allocated 262144 entries for cmdq Aug 13 00:25:21.220478 kernel: arm-smmu-v3 arm-smmu-v3.6.auto: allocated 131072 entries for evtq Aug 13 00:25:21.220537 kernel: arm-smmu-v3 arm-smmu-v3.6.auto: allocated 262144 entries for priq Aug 13 00:25:21.220601 kernel: arm-smmu-v3 arm-smmu-v3.7.auto: option mask 0x0 Aug 13 00:25:21.220660 kernel: arm-smmu-v3 arm-smmu-v3.7.auto: IDR0.COHACC overridden by FW configuration (false) Aug 13 00:25:21.220718 kernel: arm-smmu-v3 arm-smmu-v3.7.auto: ias 48-bit, oas 48-bit (features 0x000c1eff) Aug 13 00:25:21.220777 kernel: arm-smmu-v3 arm-smmu-v3.7.auto: allocated 262144 entries for cmdq Aug 13 00:25:21.220838 kernel: arm-smmu-v3 arm-smmu-v3.7.auto: allocated 131072 entries for evtq Aug 13 00:25:21.220897 kernel: arm-smmu-v3 arm-smmu-v3.7.auto: allocated 262144 entries for priq Aug 13 00:25:21.220907 kernel: thunder_xcv, ver 1.0 Aug 13 00:25:21.220916 kernel: thunder_bgx, ver 1.0 Aug 13 00:25:21.220924 kernel: nicpf, ver 1.0 Aug 13 00:25:21.220932 kernel: nicvf, ver 1.0 Aug 13 00:25:21.220996 kernel: rtc-efi rtc-efi.0: registered as rtc0 Aug 13 00:25:21.221060 kernel: rtc-efi rtc-efi.0: setting system clock to 2025-08-13T00:25:19 UTC (1755044719) Aug 13 00:25:21.221073 kernel: efifb: probing for efifb Aug 13 00:25:21.221082 kernel: efifb: framebuffer at 0x20000000, using 1876k, total 1875k Aug 13 00:25:21.221092 kernel: efifb: mode is 800x600x32, linelength=3200, pages=1 Aug 13 00:25:21.221100 kernel: efifb: scrolling: redraw Aug 13 00:25:21.221108 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Aug 13 00:25:21.221116 kernel: Console: switching to colour frame buffer device 100x37 Aug 13 00:25:21.221124 kernel: fb0: EFI VGA frame buffer device Aug 13 00:25:21.221132 kernel: SMCCC: SOC_ID: ID = jep106:0a16:0001 Revision = 0x000000a1 Aug 13 00:25:21.221141 kernel: hid: raw HID events driver (C) Jiri Kosina Aug 13 00:25:21.221150 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 7 counters available Aug 13 00:25:21.221158 kernel: watchdog: Delayed init of the lockup detector failed: -19 Aug 13 00:25:21.221167 kernel: watchdog: Hard watchdog permanently disabled Aug 13 00:25:21.221175 kernel: NET: Registered PF_INET6 protocol family Aug 13 00:25:21.221183 kernel: Segment Routing with IPv6 Aug 13 00:25:21.221191 kernel: In-situ OAM (IOAM) with IPv6 Aug 13 00:25:21.221199 kernel: NET: Registered PF_PACKET protocol family Aug 13 00:25:21.221207 kernel: Key type dns_resolver registered Aug 13 00:25:21.221215 kernel: registered taskstats version 1 Aug 13 00:25:21.221225 kernel: Loading compiled-in X.509 certificates Aug 13 00:25:21.221233 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.100-flatcar: 7263800c6d21650660e2b030c1023dce09b1e8b6' Aug 13 00:25:21.221241 kernel: Key type .fscrypt registered Aug 13 00:25:21.221249 kernel: Key type fscrypt-provisioning registered Aug 13 00:25:21.221257 kernel: ima: No TPM chip found, activating TPM-bypass! Aug 13 00:25:21.221265 kernel: ima: Allocated hash algorithm: sha1 Aug 13 00:25:21.221273 kernel: ima: No architecture policies found Aug 13 00:25:21.221281 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Aug 13 00:25:21.221349 kernel: pcieport 000d:00:01.0: Adding to iommu group 0 Aug 13 00:25:21.221418 kernel: pcieport 000d:00:01.0: AER: enabled with IRQ 91 Aug 13 00:25:21.221483 kernel: pcieport 000d:00:02.0: Adding to iommu group 1 Aug 13 00:25:21.221548 kernel: pcieport 000d:00:02.0: AER: enabled with IRQ 91 Aug 13 00:25:21.221614 kernel: pcieport 000d:00:03.0: Adding to iommu group 2 Aug 13 00:25:21.221678 kernel: pcieport 000d:00:03.0: AER: enabled with IRQ 91 Aug 13 00:25:21.221744 kernel: pcieport 000d:00:04.0: Adding to iommu group 3 Aug 13 00:25:21.221807 kernel: pcieport 000d:00:04.0: AER: enabled with IRQ 91 Aug 13 00:25:21.221873 kernel: pcieport 0000:00:01.0: Adding to iommu group 4 Aug 13 00:25:21.221940 kernel: pcieport 0000:00:01.0: AER: enabled with IRQ 92 Aug 13 00:25:21.222006 kernel: pcieport 0000:00:02.0: Adding to iommu group 5 Aug 13 00:25:21.222080 kernel: pcieport 0000:00:02.0: AER: enabled with IRQ 92 Aug 13 00:25:21.222148 kernel: pcieport 0000:00:03.0: Adding to iommu group 6 Aug 13 00:25:21.222212 kernel: pcieport 0000:00:03.0: AER: enabled with IRQ 92 Aug 13 00:25:21.222278 kernel: pcieport 0000:00:04.0: Adding to iommu group 7 Aug 13 00:25:21.222342 kernel: pcieport 0000:00:04.0: AER: enabled with IRQ 92 Aug 13 00:25:21.222407 kernel: pcieport 0005:00:01.0: Adding to iommu group 8 Aug 13 00:25:21.222471 kernel: pcieport 0005:00:01.0: AER: enabled with IRQ 93 Aug 13 00:25:21.222540 kernel: pcieport 0005:00:03.0: Adding to iommu group 9 Aug 13 00:25:21.222604 kernel: pcieport 0005:00:03.0: AER: enabled with IRQ 93 Aug 13 00:25:21.222668 kernel: pcieport 0005:00:05.0: Adding to iommu group 10 Aug 13 00:25:21.222732 kernel: pcieport 0005:00:05.0: AER: enabled with IRQ 93 Aug 13 00:25:21.222797 kernel: pcieport 0005:00:07.0: Adding to iommu group 11 Aug 13 00:25:21.222861 kernel: pcieport 0005:00:07.0: AER: enabled with IRQ 93 Aug 13 00:25:21.222927 kernel: pcieport 0003:00:01.0: Adding to iommu group 12 Aug 13 00:25:21.222990 kernel: pcieport 0003:00:01.0: AER: enabled with IRQ 94 Aug 13 00:25:21.223061 kernel: pcieport 0003:00:03.0: Adding to iommu group 13 Aug 13 00:25:21.223126 kernel: pcieport 0003:00:03.0: AER: enabled with IRQ 94 Aug 13 00:25:21.223190 kernel: pcieport 0003:00:05.0: Adding to iommu group 14 Aug 13 00:25:21.223254 kernel: pcieport 0003:00:05.0: AER: enabled with IRQ 94 Aug 13 00:25:21.223320 kernel: pcieport 000c:00:01.0: Adding to iommu group 15 Aug 13 00:25:21.223384 kernel: pcieport 000c:00:01.0: AER: enabled with IRQ 95 Aug 13 00:25:21.223450 kernel: pcieport 000c:00:02.0: Adding to iommu group 16 Aug 13 00:25:21.223514 kernel: pcieport 000c:00:02.0: AER: enabled with IRQ 95 Aug 13 00:25:21.223581 kernel: pcieport 000c:00:03.0: Adding to iommu group 17 Aug 13 00:25:21.223646 kernel: pcieport 000c:00:03.0: AER: enabled with IRQ 95 Aug 13 00:25:21.223711 kernel: pcieport 000c:00:04.0: Adding to iommu group 18 Aug 13 00:25:21.223777 kernel: pcieport 000c:00:04.0: AER: enabled with IRQ 95 Aug 13 00:25:21.223843 kernel: pcieport 0002:00:01.0: Adding to iommu group 19 Aug 13 00:25:21.223907 kernel: pcieport 0002:00:01.0: AER: enabled with IRQ 96 Aug 13 00:25:21.223971 kernel: pcieport 0002:00:03.0: Adding to iommu group 20 Aug 13 00:25:21.224034 kernel: pcieport 0002:00:03.0: AER: enabled with IRQ 96 Aug 13 00:25:21.224106 kernel: pcieport 0002:00:05.0: Adding to iommu group 21 Aug 13 00:25:21.224169 kernel: pcieport 0002:00:05.0: AER: enabled with IRQ 96 Aug 13 00:25:21.224234 kernel: pcieport 0002:00:07.0: Adding to iommu group 22 Aug 13 00:25:21.224297 kernel: pcieport 0002:00:07.0: AER: enabled with IRQ 96 Aug 13 00:25:21.224362 kernel: pcieport 0001:00:01.0: Adding to iommu group 23 Aug 13 00:25:21.224425 kernel: pcieport 0001:00:01.0: AER: enabled with IRQ 97 Aug 13 00:25:21.224490 kernel: pcieport 0001:00:02.0: Adding to iommu group 24 Aug 13 00:25:21.224552 kernel: pcieport 0001:00:02.0: AER: enabled with IRQ 97 Aug 13 00:25:21.224616 kernel: pcieport 0001:00:03.0: Adding to iommu group 25 Aug 13 00:25:21.224682 kernel: pcieport 0001:00:03.0: AER: enabled with IRQ 97 Aug 13 00:25:21.224746 kernel: pcieport 0001:00:04.0: Adding to iommu group 26 Aug 13 00:25:21.224810 kernel: pcieport 0001:00:04.0: AER: enabled with IRQ 97 Aug 13 00:25:21.224874 kernel: pcieport 0004:00:01.0: Adding to iommu group 27 Aug 13 00:25:21.224938 kernel: pcieport 0004:00:01.0: AER: enabled with IRQ 98 Aug 13 00:25:21.225002 kernel: pcieport 0004:00:03.0: Adding to iommu group 28 Aug 13 00:25:21.225069 kernel: pcieport 0004:00:03.0: AER: enabled with IRQ 98 Aug 13 00:25:21.225136 kernel: pcieport 0004:00:05.0: Adding to iommu group 29 Aug 13 00:25:21.225200 kernel: pcieport 0004:00:05.0: AER: enabled with IRQ 98 Aug 13 00:25:21.225267 kernel: pcieport 0004:01:00.0: Adding to iommu group 30 Aug 13 00:25:21.225278 kernel: clk: Disabling unused clocks Aug 13 00:25:21.225286 kernel: Freeing unused kernel memory: 39424K Aug 13 00:25:21.225294 kernel: Run /init as init process Aug 13 00:25:21.225302 kernel: with arguments: Aug 13 00:25:21.225310 kernel: /init Aug 13 00:25:21.225318 kernel: with environment: Aug 13 00:25:21.225328 kernel: HOME=/ Aug 13 00:25:21.225336 kernel: TERM=linux Aug 13 00:25:21.225344 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Aug 13 00:25:21.225354 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Aug 13 00:25:21.225364 systemd[1]: Detected architecture arm64. Aug 13 00:25:21.225373 systemd[1]: Running in initrd. Aug 13 00:25:21.225381 systemd[1]: No hostname configured, using default hostname. Aug 13 00:25:21.225389 systemd[1]: Hostname set to . Aug 13 00:25:21.225399 systemd[1]: Initializing machine ID from random generator. Aug 13 00:25:21.225408 systemd[1]: Queued start job for default target initrd.target. Aug 13 00:25:21.225416 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Aug 13 00:25:21.225425 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Aug 13 00:25:21.225434 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Aug 13 00:25:21.225442 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Aug 13 00:25:21.225451 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Aug 13 00:25:21.225460 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Aug 13 00:25:21.225470 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Aug 13 00:25:21.225479 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Aug 13 00:25:21.225488 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Aug 13 00:25:21.225496 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Aug 13 00:25:21.225505 systemd[1]: Reached target paths.target - Path Units. Aug 13 00:25:21.225513 systemd[1]: Reached target slices.target - Slice Units. Aug 13 00:25:21.225523 systemd[1]: Reached target swap.target - Swaps. Aug 13 00:25:21.225531 systemd[1]: Reached target timers.target - Timer Units. Aug 13 00:25:21.225540 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Aug 13 00:25:21.225548 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Aug 13 00:25:21.225557 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Aug 13 00:25:21.225565 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Aug 13 00:25:21.225574 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Aug 13 00:25:21.225582 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Aug 13 00:25:21.225590 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Aug 13 00:25:21.225600 systemd[1]: Reached target sockets.target - Socket Units. Aug 13 00:25:21.225608 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Aug 13 00:25:21.225617 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Aug 13 00:25:21.225625 systemd[1]: Finished network-cleanup.service - Network Cleanup. Aug 13 00:25:21.225634 systemd[1]: Starting systemd-fsck-usr.service... Aug 13 00:25:21.225642 systemd[1]: Starting systemd-journald.service - Journal Service... Aug 13 00:25:21.225650 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Aug 13 00:25:21.225680 systemd-journald[898]: Collecting audit messages is disabled. Aug 13 00:25:21.225703 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 00:25:21.225711 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Aug 13 00:25:21.225720 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Aug 13 00:25:21.225728 kernel: Bridge firewalling registered Aug 13 00:25:21.225737 systemd-journald[898]: Journal started Aug 13 00:25:21.225758 systemd-journald[898]: Runtime Journal (/run/log/journal/04c65096637449eca2db76111d8b0bcc) is 8.0M, max 4.0G, 3.9G free. Aug 13 00:25:21.183537 systemd-modules-load[900]: Inserted module 'overlay' Aug 13 00:25:21.260203 systemd[1]: Started systemd-journald.service - Journal Service. Aug 13 00:25:21.207000 systemd-modules-load[900]: Inserted module 'br_netfilter' Aug 13 00:25:21.265857 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Aug 13 00:25:21.276666 systemd[1]: Finished systemd-fsck-usr.service. Aug 13 00:25:21.287484 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Aug 13 00:25:21.298221 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 00:25:21.328230 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Aug 13 00:25:21.334301 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Aug 13 00:25:21.365128 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Aug 13 00:25:21.371408 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Aug 13 00:25:21.387753 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Aug 13 00:25:21.404140 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Aug 13 00:25:21.420820 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Aug 13 00:25:21.432276 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Aug 13 00:25:21.458193 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Aug 13 00:25:21.468277 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Aug 13 00:25:21.480267 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Aug 13 00:25:21.505747 dracut-cmdline[942]: dracut-dracut-053 Aug 13 00:25:21.505747 dracut-cmdline[942]: Using kernel command line parameters: rd.driver.pre=btrfs BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty0 console=ttyS1,115200n8 flatcar.first_boot=detected acpi=force flatcar.oem.id=packet flatcar.autologin verity.usrhash=2f9df6e9e6c671c457040a64675390bbff42294b08c628cd2dc472ed8120146a Aug 13 00:25:21.494420 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Aug 13 00:25:21.506669 systemd-resolved[944]: Positive Trust Anchors: Aug 13 00:25:21.506678 systemd-resolved[944]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Aug 13 00:25:21.506710 systemd-resolved[944]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Aug 13 00:25:21.521751 systemd-resolved[944]: Defaulting to hostname 'linux'. Aug 13 00:25:21.523261 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Aug 13 00:25:21.557704 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Aug 13 00:25:21.662044 kernel: SCSI subsystem initialized Aug 13 00:25:21.677042 kernel: Loading iSCSI transport class v2.0-870. Aug 13 00:25:21.696046 kernel: iscsi: registered transport (tcp) Aug 13 00:25:21.723158 kernel: iscsi: registered transport (qla4xxx) Aug 13 00:25:21.723180 kernel: QLogic iSCSI HBA Driver Aug 13 00:25:21.767515 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Aug 13 00:25:21.788153 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Aug 13 00:25:21.833034 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Aug 13 00:25:21.833073 kernel: device-mapper: uevent: version 1.0.3 Aug 13 00:25:21.842653 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Aug 13 00:25:21.908049 kernel: raid6: neonx8 gen() 15845 MB/s Aug 13 00:25:21.933043 kernel: raid6: neonx4 gen() 15733 MB/s Aug 13 00:25:21.958047 kernel: raid6: neonx2 gen() 13376 MB/s Aug 13 00:25:21.983047 kernel: raid6: neonx1 gen() 10511 MB/s Aug 13 00:25:22.008047 kernel: raid6: int64x8 gen() 6984 MB/s Aug 13 00:25:22.033047 kernel: raid6: int64x4 gen() 7374 MB/s Aug 13 00:25:22.058046 kernel: raid6: int64x2 gen() 6155 MB/s Aug 13 00:25:22.086059 kernel: raid6: int64x1 gen() 5075 MB/s Aug 13 00:25:22.086081 kernel: raid6: using algorithm neonx8 gen() 15845 MB/s Aug 13 00:25:22.120469 kernel: raid6: .... xor() 11977 MB/s, rmw enabled Aug 13 00:25:22.120490 kernel: raid6: using neon recovery algorithm Aug 13 00:25:22.140048 kernel: xor: measuring software checksum speed Aug 13 00:25:22.151684 kernel: 8regs : 19119 MB/sec Aug 13 00:25:22.151705 kernel: 32regs : 19679 MB/sec Aug 13 00:25:22.167116 kernel: arm64_neon : 26726 MB/sec Aug 13 00:25:22.167137 kernel: xor: using function: arm64_neon (26726 MB/sec) Aug 13 00:25:22.228046 kernel: Btrfs loaded, zoned=no, fsverity=no Aug 13 00:25:22.239102 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Aug 13 00:25:22.266163 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Aug 13 00:25:22.279353 systemd-udevd[1140]: Using default interface naming scheme 'v255'. Aug 13 00:25:22.282421 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Aug 13 00:25:22.305186 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Aug 13 00:25:22.319341 dracut-pre-trigger[1150]: rd.md=0: removing MD RAID activation Aug 13 00:25:22.345804 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Aug 13 00:25:22.372203 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Aug 13 00:25:22.481804 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Aug 13 00:25:22.510165 kernel: pps_core: LinuxPPS API ver. 1 registered Aug 13 00:25:22.510190 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Aug 13 00:25:22.524216 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Aug 13 00:25:22.662658 kernel: ACPI: bus type USB registered Aug 13 00:25:22.662682 kernel: usbcore: registered new interface driver usbfs Aug 13 00:25:22.662707 kernel: usbcore: registered new interface driver hub Aug 13 00:25:22.662727 kernel: usbcore: registered new device driver usb Aug 13 00:25:22.662748 kernel: PTP clock support registered Aug 13 00:25:22.662769 kernel: xhci_hcd 0004:03:00.0: Adding to iommu group 31 Aug 13 00:25:22.662934 kernel: xhci_hcd 0004:03:00.0: xHCI Host Controller Aug 13 00:25:22.663020 kernel: xhci_hcd 0004:03:00.0: new USB bus registered, assigned bus number 1 Aug 13 00:25:22.663110 kernel: xhci_hcd 0004:03:00.0: Zeroing 64bit base registers, expecting fault Aug 13 00:25:22.663190 kernel: igb: Intel(R) Gigabit Ethernet Network Driver Aug 13 00:25:22.663203 kernel: mlx5_core 0001:01:00.0: Adding to iommu group 32 Aug 13 00:25:22.663291 kernel: igb: Copyright (c) 2007-2014 Intel Corporation. Aug 13 00:25:22.663301 kernel: igb 0003:03:00.0: Adding to iommu group 33 Aug 13 00:25:22.686046 kernel: nvme 0005:03:00.0: Adding to iommu group 34 Aug 13 00:25:22.686176 kernel: nvme 0005:04:00.0: Adding to iommu group 35 Aug 13 00:25:22.702468 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Aug 13 00:25:22.714068 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Aug 13 00:25:22.720139 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Aug 13 00:25:22.737408 systemd[1]: Reached target remote-fs.target - Remote File Systems. Aug 13 00:25:22.748963 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Aug 13 00:25:22.749116 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Aug 13 00:25:22.766363 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Aug 13 00:25:22.790245 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Aug 13 00:25:22.800932 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Aug 13 00:25:22.961755 kernel: xhci_hcd 0004:03:00.0: hcc params 0x014051cf hci version 0x100 quirks 0x0000001100000010 Aug 13 00:25:22.961980 kernel: xhci_hcd 0004:03:00.0: xHCI Host Controller Aug 13 00:25:22.962077 kernel: xhci_hcd 0004:03:00.0: new USB bus registered, assigned bus number 2 Aug 13 00:25:22.962158 kernel: xhci_hcd 0004:03:00.0: Host supports USB 3.0 SuperSpeed Aug 13 00:25:22.962235 kernel: hub 1-0:1.0: USB hub found Aug 13 00:25:22.962334 kernel: hub 1-0:1.0: 4 ports detected Aug 13 00:25:22.962413 kernel: mlx5_core 0001:01:00.0: firmware version: 14.31.1014 Aug 13 00:25:22.962499 kernel: usb usb2: We don't know the algorithms for LPM for this host, disabling LPM. Aug 13 00:25:22.962592 kernel: hub 2-0:1.0: USB hub found Aug 13 00:25:22.962680 kernel: mlx5_core 0001:01:00.0: 31.504 Gb/s available PCIe bandwidth, limited by 8.0 GT/s PCIe x4 link at 0001:00:01.0 (capable of 63.008 Gb/s with 8.0 GT/s PCIe x8 link) Aug 13 00:25:22.962762 kernel: hub 2-0:1.0: 4 ports detected Aug 13 00:25:22.801107 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 00:25:22.879227 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 00:25:22.975255 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 00:25:22.984643 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Aug 13 00:25:23.001622 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Aug 13 00:25:23.001727 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 00:25:23.058570 kernel: nvme nvme0: pci function 0005:04:00.0 Aug 13 00:25:23.058714 kernel: nvme nvme1: pci function 0005:03:00.0 Aug 13 00:25:23.058802 kernel: nvme nvme0: Shutdown timeout set to 8 seconds Aug 13 00:25:23.058871 kernel: nvme nvme1: Shutdown timeout set to 8 seconds Aug 13 00:25:23.054257 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 00:25:23.073331 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 00:25:23.100685 kernel: nvme nvme1: 32/0/0 default/read/poll queues Aug 13 00:25:23.100878 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Aug 13 00:25:23.103044 kernel: GPT:9289727 != 1875385007 Aug 13 00:25:23.103067 kernel: GPT:Alternate GPT header not at the end of the disk. Aug 13 00:25:23.103086 kernel: GPT:9289727 != 1875385007 Aug 13 00:25:23.103110 kernel: GPT: Use GNU Parted to correct GPT errors. Aug 13 00:25:23.103129 kernel: nvme1n1: p1 p2 p3 p4 p6 p7 p9 Aug 13 00:25:23.104044 kernel: nvme nvme0: 32/0/0 default/read/poll queues Aug 13 00:25:23.105042 kernel: igb 0003:03:00.0: added PHC on eth0 Aug 13 00:25:23.105149 kernel: igb 0003:03:00.0: Intel(R) Gigabit Ethernet Network Connection Aug 13 00:25:23.105237 kernel: igb 0003:03:00.0: eth0: (PCIe:5.0Gb/s:Width x2) 18:c0:4d:0c:6f:ac Aug 13 00:25:23.105328 kernel: igb 0003:03:00.0: eth0: PBA No: 106300-000 Aug 13 00:25:23.105408 kernel: igb 0003:03:00.0: Using MSI-X interrupts. 8 rx queue(s), 8 tx queue(s) Aug 13 00:25:23.105500 kernel: igb 0003:03:00.1: Adding to iommu group 36 Aug 13 00:25:23.162051 kernel: BTRFS: device fsid 03408483-5051-409a-aab4-4e6d5027e982 devid 1 transid 41 /dev/nvme1n1p3 scanned by (udev-worker) (1186) Aug 13 00:25:23.162077 kernel: igb 0003:03:00.1: added PHC on eth1 Aug 13 00:25:23.162243 kernel: BTRFS: device label OEM devid 1 transid 11 /dev/nvme1n1p6 scanned by (udev-worker) (1195) Aug 13 00:25:23.171074 kernel: igb 0003:03:00.1: Intel(R) Gigabit Ethernet Network Connection Aug 13 00:25:23.289991 kernel: igb 0003:03:00.1: eth1: (PCIe:5.0Gb/s:Width x2) 18:c0:4d:0c:6f:ad Aug 13 00:25:23.301685 kernel: igb 0003:03:00.1: eth1: PBA No: 106300-000 Aug 13 00:25:23.311260 kernel: igb 0003:03:00.1: Using MSI-X interrupts. 8 rx queue(s), 8 tx queue(s) Aug 13 00:25:23.318434 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - SAMSUNG MZ1LB960HAJQ-00007 EFI-SYSTEM. Aug 13 00:25:23.369587 kernel: igb 0003:03:00.1 eno2: renamed from eth1 Aug 13 00:25:23.369706 kernel: mlx5_core 0001:01:00.0: Port module event: module 0, Cable plugged Aug 13 00:25:23.369795 kernel: igb 0003:03:00.0 eno1: renamed from eth0 Aug 13 00:25:23.369873 kernel: usb 1-3: new high-speed USB device number 2 using xhci_hcd Aug 13 00:25:23.378800 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - SAMSUNG MZ1LB960HAJQ-00007 ROOT. Aug 13 00:25:23.396606 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - SAMSUNG MZ1LB960HAJQ-00007 USR-A. Aug 13 00:25:23.411230 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - SAMSUNG MZ1LB960HAJQ-00007 USR-A. Aug 13 00:25:23.420510 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - SAMSUNG MZ1LB960HAJQ-00007 OEM. Aug 13 00:25:23.445185 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Aug 13 00:25:23.471365 kernel: nvme1n1: p1 p2 p3 p4 p6 p7 p9 Aug 13 00:25:23.451476 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Aug 13 00:25:23.507111 kernel: hub 1-3:1.0: USB hub found Aug 13 00:25:23.507265 kernel: hub 1-3:1.0: 4 ports detected Aug 13 00:25:23.507361 disk-uuid[1285]: Primary Header is updated. Aug 13 00:25:23.507361 disk-uuid[1285]: Secondary Entries is updated. Aug 13 00:25:23.507361 disk-uuid[1285]: Secondary Header is updated. Aug 13 00:25:23.524795 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Aug 13 00:25:23.595051 kernel: usb 2-3: new SuperSpeed USB device number 2 using xhci_hcd Aug 13 00:25:23.630052 kernel: hub 2-3:1.0: USB hub found Aug 13 00:25:23.630289 kernel: hub 2-3:1.0: 4 ports detected Aug 13 00:25:23.675048 kernel: mlx5_core 0001:01:00.0: MLX5E: StrdRq(0) RqSz(1024) StrdSz(256) RxCqeCmprss(0 basic) Aug 13 00:25:23.690042 kernel: mlx5_core 0001:01:00.1: Adding to iommu group 37 Aug 13 00:25:23.712868 kernel: mlx5_core 0001:01:00.1: firmware version: 14.31.1014 Aug 13 00:25:23.713031 kernel: mlx5_core 0001:01:00.1: 31.504 Gb/s available PCIe bandwidth, limited by 8.0 GT/s PCIe x4 link at 0001:00:01.0 (capable of 63.008 Gb/s with 8.0 GT/s PCIe x8 link) Aug 13 00:25:24.065724 kernel: mlx5_core 0001:01:00.1: Port module event: module 1, Cable plugged Aug 13 00:25:24.373050 kernel: mlx5_core 0001:01:00.1: MLX5E: StrdRq(0) RqSz(1024) StrdSz(256) RxCqeCmprss(0 basic) Aug 13 00:25:24.389043 kernel: mlx5_core 0001:01:00.0 enP1p1s0f0np0: renamed from eth0 Aug 13 00:25:24.407043 kernel: mlx5_core 0001:01:00.1 enP1p1s0f1np1: renamed from eth1 Aug 13 00:25:24.472011 disk-uuid[1286]: The operation has completed successfully. Aug 13 00:25:24.477454 kernel: nvme1n1: p1 p2 p3 p4 p6 p7 p9 Aug 13 00:25:24.496808 systemd[1]: disk-uuid.service: Deactivated successfully. Aug 13 00:25:24.496891 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Aug 13 00:25:24.530141 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Aug 13 00:25:24.540464 sh[1484]: Success Aug 13 00:25:24.559043 kernel: device-mapper: verity: sha256 using implementation "sha256-ce" Aug 13 00:25:24.591866 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Aug 13 00:25:24.612302 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Aug 13 00:25:24.622633 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Aug 13 00:25:24.717702 kernel: BTRFS info (device dm-0): first mount of filesystem 03408483-5051-409a-aab4-4e6d5027e982 Aug 13 00:25:24.717733 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm Aug 13 00:25:24.717753 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Aug 13 00:25:24.717773 kernel: BTRFS info (device dm-0): disabling log replay at mount time Aug 13 00:25:24.717792 kernel: BTRFS info (device dm-0): using free space tree Aug 13 00:25:24.717811 kernel: BTRFS info (device dm-0): enabling ssd optimizations Aug 13 00:25:24.719122 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Aug 13 00:25:24.729444 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Aug 13 00:25:24.744143 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Aug 13 00:25:24.750280 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Aug 13 00:25:24.863485 kernel: BTRFS info (device nvme1n1p6): first mount of filesystem dbce4b09-c4b8-4cc9-bd11-416717f60c7d Aug 13 00:25:24.863503 kernel: BTRFS info (device nvme1n1p6): using crc32c (crc32c-generic) checksum algorithm Aug 13 00:25:24.863513 kernel: BTRFS info (device nvme1n1p6): using free space tree Aug 13 00:25:24.863523 kernel: BTRFS info (device nvme1n1p6): enabling ssd optimizations Aug 13 00:25:24.863534 kernel: BTRFS info (device nvme1n1p6): auto enabling async discard Aug 13 00:25:24.863544 kernel: BTRFS info (device nvme1n1p6): last unmount of filesystem dbce4b09-c4b8-4cc9-bd11-416717f60c7d Aug 13 00:25:24.860305 systemd[1]: Finished ignition-setup.service - Ignition (setup). Aug 13 00:25:24.888232 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Aug 13 00:25:24.899007 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Aug 13 00:25:24.928204 systemd[1]: Starting systemd-networkd.service - Network Configuration... Aug 13 00:25:24.948207 systemd-networkd[1687]: lo: Link UP Aug 13 00:25:24.948212 systemd-networkd[1687]: lo: Gained carrier Aug 13 00:25:24.951802 systemd-networkd[1687]: Enumeration completed Aug 13 00:25:24.951888 systemd[1]: Started systemd-networkd.service - Network Configuration. Aug 13 00:25:24.952978 systemd-networkd[1687]: eno1: Configuring with /usr/lib/systemd/network/zz-default.network. Aug 13 00:25:24.959738 systemd[1]: Reached target network.target - Network. Aug 13 00:25:24.984580 ignition[1674]: Ignition 2.19.0 Aug 13 00:25:24.993975 unknown[1674]: fetched base config from "system" Aug 13 00:25:24.984586 ignition[1674]: Stage: fetch-offline Aug 13 00:25:24.993984 unknown[1674]: fetched user config from "system" Aug 13 00:25:24.984688 ignition[1674]: no configs at "/usr/lib/ignition/base.d" Aug 13 00:25:24.997134 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Aug 13 00:25:24.984696 ignition[1674]: no config dir at "/usr/lib/ignition/base.platform.d/packet" Aug 13 00:25:25.004652 systemd-networkd[1687]: eno2: Configuring with /usr/lib/systemd/network/zz-default.network. Aug 13 00:25:24.985044 ignition[1674]: parsed url from cmdline: "" Aug 13 00:25:25.007345 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Aug 13 00:25:24.985047 ignition[1674]: no config URL provided Aug 13 00:25:25.019206 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Aug 13 00:25:24.985051 ignition[1674]: reading system config file "/usr/lib/ignition/user.ign" Aug 13 00:25:25.057004 systemd-networkd[1687]: enP1p1s0f0np0: Configuring with /usr/lib/systemd/network/zz-default.network. Aug 13 00:25:24.985105 ignition[1674]: parsing config with SHA512: 4549cbe70e82d968e80ac52d6c85114c45051bf18f0ae7feabe644370be95ed1e38d2b867246b279ac8a182ffa114a88b569684cec4568d7a77cb44ae95e1a7c Aug 13 00:25:24.994457 ignition[1674]: fetch-offline: fetch-offline passed Aug 13 00:25:24.994461 ignition[1674]: POST message to Packet Timeline Aug 13 00:25:24.994466 ignition[1674]: POST Status error: resource requires networking Aug 13 00:25:24.994525 ignition[1674]: Ignition finished successfully Aug 13 00:25:25.042984 ignition[1711]: Ignition 2.19.0 Aug 13 00:25:25.042990 ignition[1711]: Stage: kargs Aug 13 00:25:25.043145 ignition[1711]: no configs at "/usr/lib/ignition/base.d" Aug 13 00:25:25.043154 ignition[1711]: no config dir at "/usr/lib/ignition/base.platform.d/packet" Aug 13 00:25:25.044073 ignition[1711]: kargs: kargs passed Aug 13 00:25:25.044077 ignition[1711]: POST message to Packet Timeline Aug 13 00:25:25.044089 ignition[1711]: GET https://metadata.packet.net/metadata: attempt #1 Aug 13 00:25:25.046665 ignition[1711]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:36229->[::1]:53: read: connection refused Aug 13 00:25:25.246793 ignition[1711]: GET https://metadata.packet.net/metadata: attempt #2 Aug 13 00:25:25.247253 ignition[1711]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:48311->[::1]:53: read: connection refused Aug 13 00:25:25.639053 kernel: mlx5_core 0001:01:00.0 enP1p1s0f0np0: Link up Aug 13 00:25:25.641771 systemd-networkd[1687]: enP1p1s0f1np1: Configuring with /usr/lib/systemd/network/zz-default.network. Aug 13 00:25:25.648011 ignition[1711]: GET https://metadata.packet.net/metadata: attempt #3 Aug 13 00:25:25.649909 ignition[1711]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:47865->[::1]:53: read: connection refused Aug 13 00:25:26.243053 kernel: mlx5_core 0001:01:00.1 enP1p1s0f1np1: Link up Aug 13 00:25:26.245657 systemd-networkd[1687]: eno1: Link UP Aug 13 00:25:26.245790 systemd-networkd[1687]: eno2: Link UP Aug 13 00:25:26.245911 systemd-networkd[1687]: enP1p1s0f0np0: Link UP Aug 13 00:25:26.246057 systemd-networkd[1687]: enP1p1s0f0np0: Gained carrier Aug 13 00:25:26.257177 systemd-networkd[1687]: enP1p1s0f1np1: Link UP Aug 13 00:25:26.286072 systemd-networkd[1687]: enP1p1s0f0np0: DHCPv4 address 147.75.53.22/30, gateway 147.75.53.21 acquired from 145.40.76.140 Aug 13 00:25:26.450052 ignition[1711]: GET https://metadata.packet.net/metadata: attempt #4 Aug 13 00:25:26.450575 ignition[1711]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:56469->[::1]:53: read: connection refused Aug 13 00:25:26.646252 systemd-networkd[1687]: enP1p1s0f1np1: Gained carrier Aug 13 00:25:27.526239 systemd-networkd[1687]: enP1p1s0f0np0: Gained IPv6LL Aug 13 00:25:28.051291 ignition[1711]: GET https://metadata.packet.net/metadata: attempt #5 Aug 13 00:25:28.051684 ignition[1711]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:44801->[::1]:53: read: connection refused Aug 13 00:25:28.422290 systemd-networkd[1687]: enP1p1s0f1np1: Gained IPv6LL Aug 13 00:25:31.254177 ignition[1711]: GET https://metadata.packet.net/metadata: attempt #6 Aug 13 00:25:32.553372 ignition[1711]: GET result: OK Aug 13 00:25:33.400212 ignition[1711]: Ignition finished successfully Aug 13 00:25:33.403318 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Aug 13 00:25:33.414160 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Aug 13 00:25:33.429647 ignition[1732]: Ignition 2.19.0 Aug 13 00:25:33.429653 ignition[1732]: Stage: disks Aug 13 00:25:33.429811 ignition[1732]: no configs at "/usr/lib/ignition/base.d" Aug 13 00:25:33.429821 ignition[1732]: no config dir at "/usr/lib/ignition/base.platform.d/packet" Aug 13 00:25:33.430800 ignition[1732]: disks: disks passed Aug 13 00:25:33.430804 ignition[1732]: POST message to Packet Timeline Aug 13 00:25:33.430818 ignition[1732]: GET https://metadata.packet.net/metadata: attempt #1 Aug 13 00:25:34.320034 ignition[1732]: GET result: OK Aug 13 00:25:34.794294 ignition[1732]: Ignition finished successfully Aug 13 00:25:34.796433 systemd[1]: Finished ignition-disks.service - Ignition (disks). Aug 13 00:25:34.802899 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Aug 13 00:25:34.810471 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Aug 13 00:25:34.818478 systemd[1]: Reached target local-fs.target - Local File Systems. Aug 13 00:25:34.827024 systemd[1]: Reached target sysinit.target - System Initialization. Aug 13 00:25:34.835899 systemd[1]: Reached target basic.target - Basic System. Aug 13 00:25:34.854188 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Aug 13 00:25:34.869589 systemd-fsck[1751]: ROOT: clean, 14/553520 files, 52654/553472 blocks Aug 13 00:25:34.873214 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Aug 13 00:25:34.891146 systemd[1]: Mounting sysroot.mount - /sysroot... Aug 13 00:25:34.956972 systemd[1]: Mounted sysroot.mount - /sysroot. Aug 13 00:25:34.962083 kernel: EXT4-fs (nvme1n1p9): mounted filesystem 128aec8b-f05d-48ed-8996-c9e8b21a7810 r/w with ordered data mode. Quota mode: none. Aug 13 00:25:34.967460 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Aug 13 00:25:34.989122 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Aug 13 00:25:35.082295 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/nvme1n1p6 scanned by mount (1762) Aug 13 00:25:35.082314 kernel: BTRFS info (device nvme1n1p6): first mount of filesystem dbce4b09-c4b8-4cc9-bd11-416717f60c7d Aug 13 00:25:35.082324 kernel: BTRFS info (device nvme1n1p6): using crc32c (crc32c-generic) checksum algorithm Aug 13 00:25:35.082334 kernel: BTRFS info (device nvme1n1p6): using free space tree Aug 13 00:25:35.082344 kernel: BTRFS info (device nvme1n1p6): enabling ssd optimizations Aug 13 00:25:35.082359 kernel: BTRFS info (device nvme1n1p6): auto enabling async discard Aug 13 00:25:34.995327 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Aug 13 00:25:35.088609 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Aug 13 00:25:35.099531 systemd[1]: Starting flatcar-static-network.service - Flatcar Static Network Agent... Aug 13 00:25:35.115403 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Aug 13 00:25:35.115431 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Aug 13 00:25:35.148055 coreos-metadata[1781]: Aug 13 00:25:35.146 INFO Fetching https://metadata.packet.net/metadata: Attempt #1 Aug 13 00:25:35.164854 coreos-metadata[1780]: Aug 13 00:25:35.146 INFO Fetching https://metadata.packet.net/metadata: Attempt #1 Aug 13 00:25:35.128888 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Aug 13 00:25:35.142626 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Aug 13 00:25:35.170222 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Aug 13 00:25:35.203734 initrd-setup-root[1798]: cut: /sysroot/etc/passwd: No such file or directory Aug 13 00:25:35.209920 initrd-setup-root[1805]: cut: /sysroot/etc/group: No such file or directory Aug 13 00:25:35.216353 initrd-setup-root[1812]: cut: /sysroot/etc/shadow: No such file or directory Aug 13 00:25:35.222753 initrd-setup-root[1819]: cut: /sysroot/etc/gshadow: No such file or directory Aug 13 00:25:35.293449 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Aug 13 00:25:35.321106 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Aug 13 00:25:35.352086 kernel: BTRFS info (device nvme1n1p6): last unmount of filesystem dbce4b09-c4b8-4cc9-bd11-416717f60c7d Aug 13 00:25:35.327560 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Aug 13 00:25:35.358418 systemd[1]: sysroot-oem.mount: Deactivated successfully. Aug 13 00:25:35.373825 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Aug 13 00:25:35.386281 ignition[1891]: INFO : Ignition 2.19.0 Aug 13 00:25:35.386281 ignition[1891]: INFO : Stage: mount Aug 13 00:25:35.397064 ignition[1891]: INFO : no configs at "/usr/lib/ignition/base.d" Aug 13 00:25:35.397064 ignition[1891]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/packet" Aug 13 00:25:35.397064 ignition[1891]: INFO : mount: mount passed Aug 13 00:25:35.397064 ignition[1891]: INFO : POST message to Packet Timeline Aug 13 00:25:35.397064 ignition[1891]: INFO : GET https://metadata.packet.net/metadata: attempt #1 Aug 13 00:25:35.983464 coreos-metadata[1781]: Aug 13 00:25:35.983 INFO Fetch successful Aug 13 00:25:36.033361 systemd[1]: flatcar-static-network.service: Deactivated successfully. Aug 13 00:25:36.033449 systemd[1]: Finished flatcar-static-network.service - Flatcar Static Network Agent. Aug 13 00:25:36.111306 ignition[1891]: INFO : GET result: OK Aug 13 00:25:36.456809 coreos-metadata[1780]: Aug 13 00:25:36.456 INFO Fetch successful Aug 13 00:25:36.502007 coreos-metadata[1780]: Aug 13 00:25:36.501 INFO wrote hostname ci-4081.3.5-a-588adb827d to /sysroot/etc/hostname Aug 13 00:25:36.505201 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Aug 13 00:25:39.546855 ignition[1891]: INFO : Ignition finished successfully Aug 13 00:25:39.549085 systemd[1]: Finished ignition-mount.service - Ignition (mount). Aug 13 00:25:39.565111 systemd[1]: Starting ignition-files.service - Ignition (files)... Aug 13 00:25:39.577332 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Aug 13 00:25:39.603046 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/nvme1n1p6 scanned by mount (1925) Aug 13 00:25:39.627151 kernel: BTRFS info (device nvme1n1p6): first mount of filesystem dbce4b09-c4b8-4cc9-bd11-416717f60c7d Aug 13 00:25:39.627175 kernel: BTRFS info (device nvme1n1p6): using crc32c (crc32c-generic) checksum algorithm Aug 13 00:25:39.640161 kernel: BTRFS info (device nvme1n1p6): using free space tree Aug 13 00:25:39.663065 kernel: BTRFS info (device nvme1n1p6): enabling ssd optimizations Aug 13 00:25:39.663087 kernel: BTRFS info (device nvme1n1p6): auto enabling async discard Aug 13 00:25:39.671243 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Aug 13 00:25:39.704442 ignition[1945]: INFO : Ignition 2.19.0 Aug 13 00:25:39.704442 ignition[1945]: INFO : Stage: files Aug 13 00:25:39.713942 ignition[1945]: INFO : no configs at "/usr/lib/ignition/base.d" Aug 13 00:25:39.713942 ignition[1945]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/packet" Aug 13 00:25:39.713942 ignition[1945]: DEBUG : files: compiled without relabeling support, skipping Aug 13 00:25:39.713942 ignition[1945]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Aug 13 00:25:39.713942 ignition[1945]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Aug 13 00:25:39.713942 ignition[1945]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Aug 13 00:25:39.713942 ignition[1945]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Aug 13 00:25:39.713942 ignition[1945]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Aug 13 00:25:39.713942 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-arm64.tar.gz" Aug 13 00:25:39.713942 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.3-linux-arm64.tar.gz: attempt #1 Aug 13 00:25:39.709979 unknown[1945]: wrote ssh authorized keys file for user: core Aug 13 00:25:39.806785 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Aug 13 00:25:39.973814 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-arm64.tar.gz" Aug 13 00:25:39.984327 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Aug 13 00:25:39.984327 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Aug 13 00:25:39.984327 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Aug 13 00:25:39.984327 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Aug 13 00:25:39.984327 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Aug 13 00:25:39.984327 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Aug 13 00:25:39.984327 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Aug 13 00:25:39.984327 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Aug 13 00:25:39.984327 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Aug 13 00:25:39.984327 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Aug 13 00:25:39.984327 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-arm64.raw" Aug 13 00:25:39.984327 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.0-arm64.raw" Aug 13 00:25:39.984327 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-arm64.raw" Aug 13 00:25:39.984327 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://extensions.flatcar.org/extensions/kubernetes-v1.33.0-arm64.raw: attempt #1 Aug 13 00:25:40.383402 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Aug 13 00:25:40.812326 ignition[1945]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.0-arm64.raw" Aug 13 00:25:40.824960 ignition[1945]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Aug 13 00:25:40.824960 ignition[1945]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Aug 13 00:25:40.824960 ignition[1945]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Aug 13 00:25:40.824960 ignition[1945]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Aug 13 00:25:40.824960 ignition[1945]: INFO : files: op(d): [started] setting preset to enabled for "prepare-helm.service" Aug 13 00:25:40.824960 ignition[1945]: INFO : files: op(d): [finished] setting preset to enabled for "prepare-helm.service" Aug 13 00:25:40.824960 ignition[1945]: INFO : files: createResultFile: createFiles: op(e): [started] writing file "/sysroot/etc/.ignition-result.json" Aug 13 00:25:40.824960 ignition[1945]: INFO : files: createResultFile: createFiles: op(e): [finished] writing file "/sysroot/etc/.ignition-result.json" Aug 13 00:25:40.824960 ignition[1945]: INFO : files: files passed Aug 13 00:25:40.824960 ignition[1945]: INFO : POST message to Packet Timeline Aug 13 00:25:40.824960 ignition[1945]: INFO : GET https://metadata.packet.net/metadata: attempt #1 Aug 13 00:25:41.597597 ignition[1945]: INFO : GET result: OK Aug 13 00:25:42.176234 ignition[1945]: INFO : Ignition finished successfully Aug 13 00:25:42.178643 systemd[1]: Finished ignition-files.service - Ignition (files). Aug 13 00:25:42.198160 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Aug 13 00:25:42.210687 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Aug 13 00:25:42.229343 systemd[1]: ignition-quench.service: Deactivated successfully. Aug 13 00:25:42.229422 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Aug 13 00:25:42.247948 initrd-setup-root-after-ignition[1990]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Aug 13 00:25:42.247948 initrd-setup-root-after-ignition[1990]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Aug 13 00:25:42.242389 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Aug 13 00:25:42.296038 initrd-setup-root-after-ignition[1995]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Aug 13 00:25:42.255520 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Aug 13 00:25:42.284184 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Aug 13 00:25:42.329626 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Aug 13 00:25:42.331086 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Aug 13 00:25:42.339403 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Aug 13 00:25:42.349763 systemd[1]: Reached target initrd.target - Initrd Default Target. Aug 13 00:25:42.367799 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Aug 13 00:25:42.378135 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Aug 13 00:25:42.400901 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Aug 13 00:25:42.426149 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Aug 13 00:25:42.441139 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Aug 13 00:25:42.450601 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Aug 13 00:25:42.462394 systemd[1]: Stopped target timers.target - Timer Units. Aug 13 00:25:42.474160 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Aug 13 00:25:42.474263 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Aug 13 00:25:42.486094 systemd[1]: Stopped target initrd.target - Initrd Default Target. Aug 13 00:25:42.497550 systemd[1]: Stopped target basic.target - Basic System. Aug 13 00:25:42.509200 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Aug 13 00:25:42.520805 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Aug 13 00:25:42.532403 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Aug 13 00:25:42.543942 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Aug 13 00:25:42.555434 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Aug 13 00:25:42.566956 systemd[1]: Stopped target sysinit.target - System Initialization. Aug 13 00:25:42.578460 systemd[1]: Stopped target local-fs.target - Local File Systems. Aug 13 00:25:42.595563 systemd[1]: Stopped target swap.target - Swaps. Aug 13 00:25:42.607209 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Aug 13 00:25:42.607304 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Aug 13 00:25:42.619094 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Aug 13 00:25:42.630441 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Aug 13 00:25:42.641974 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Aug 13 00:25:42.646079 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Aug 13 00:25:42.653270 systemd[1]: dracut-initqueue.service: Deactivated successfully. Aug 13 00:25:42.653365 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Aug 13 00:25:42.664744 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Aug 13 00:25:42.664848 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Aug 13 00:25:42.676025 systemd[1]: Stopped target paths.target - Path Units. Aug 13 00:25:42.687223 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Aug 13 00:25:42.693071 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Aug 13 00:25:42.704380 systemd[1]: Stopped target slices.target - Slice Units. Aug 13 00:25:42.715983 systemd[1]: Stopped target sockets.target - Socket Units. Aug 13 00:25:42.727446 systemd[1]: iscsid.socket: Deactivated successfully. Aug 13 00:25:42.727546 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Aug 13 00:25:42.828438 ignition[2017]: INFO : Ignition 2.19.0 Aug 13 00:25:42.828438 ignition[2017]: INFO : Stage: umount Aug 13 00:25:42.828438 ignition[2017]: INFO : no configs at "/usr/lib/ignition/base.d" Aug 13 00:25:42.828438 ignition[2017]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/packet" Aug 13 00:25:42.828438 ignition[2017]: INFO : umount: umount passed Aug 13 00:25:42.828438 ignition[2017]: INFO : POST message to Packet Timeline Aug 13 00:25:42.828438 ignition[2017]: INFO : GET https://metadata.packet.net/metadata: attempt #1 Aug 13 00:25:42.739085 systemd[1]: iscsiuio.socket: Deactivated successfully. Aug 13 00:25:42.739170 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Aug 13 00:25:42.750814 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Aug 13 00:25:42.750902 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Aug 13 00:25:42.762536 systemd[1]: ignition-files.service: Deactivated successfully. Aug 13 00:25:42.762618 systemd[1]: Stopped ignition-files.service - Ignition (files). Aug 13 00:25:42.774252 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Aug 13 00:25:42.774333 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Aug 13 00:25:42.798168 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Aug 13 00:25:42.810084 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Aug 13 00:25:42.810185 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Aug 13 00:25:42.837234 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Aug 13 00:25:42.849008 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Aug 13 00:25:42.849120 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Aug 13 00:25:42.860605 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Aug 13 00:25:42.860689 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Aug 13 00:25:42.879589 systemd[1]: sysroot-boot.mount: Deactivated successfully. Aug 13 00:25:42.880389 systemd[1]: sysroot-boot.service: Deactivated successfully. Aug 13 00:25:42.880474 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Aug 13 00:25:42.891983 systemd[1]: initrd-cleanup.service: Deactivated successfully. Aug 13 00:25:42.892063 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Aug 13 00:25:44.109702 ignition[2017]: INFO : GET result: OK Aug 13 00:25:44.500732 ignition[2017]: INFO : Ignition finished successfully Aug 13 00:25:44.502997 systemd[1]: ignition-mount.service: Deactivated successfully. Aug 13 00:25:44.503209 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Aug 13 00:25:44.511019 systemd[1]: Stopped target network.target - Network. Aug 13 00:25:44.520358 systemd[1]: ignition-disks.service: Deactivated successfully. Aug 13 00:25:44.520419 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Aug 13 00:25:44.530127 systemd[1]: ignition-kargs.service: Deactivated successfully. Aug 13 00:25:44.530160 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Aug 13 00:25:44.539655 systemd[1]: ignition-setup.service: Deactivated successfully. Aug 13 00:25:44.539701 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Aug 13 00:25:44.549470 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Aug 13 00:25:44.549520 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Aug 13 00:25:44.559307 systemd[1]: initrd-setup-root.service: Deactivated successfully. Aug 13 00:25:44.559337 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Aug 13 00:25:44.569269 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Aug 13 00:25:44.577068 systemd-networkd[1687]: enP1p1s0f1np1: DHCPv6 lease lost Aug 13 00:25:44.578886 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Aug 13 00:25:44.585063 systemd-networkd[1687]: enP1p1s0f0np0: DHCPv6 lease lost Aug 13 00:25:44.590717 systemd[1]: systemd-resolved.service: Deactivated successfully. Aug 13 00:25:44.590897 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Aug 13 00:25:44.601026 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Aug 13 00:25:44.601112 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Aug 13 00:25:44.609040 systemd[1]: systemd-networkd.service: Deactivated successfully. Aug 13 00:25:44.609222 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Aug 13 00:25:44.619467 systemd[1]: systemd-networkd.socket: Deactivated successfully. Aug 13 00:25:44.619635 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Aug 13 00:25:44.640140 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Aug 13 00:25:44.648557 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Aug 13 00:25:44.648624 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Aug 13 00:25:44.658792 systemd[1]: systemd-sysctl.service: Deactivated successfully. Aug 13 00:25:44.658825 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Aug 13 00:25:44.669131 systemd[1]: systemd-modules-load.service: Deactivated successfully. Aug 13 00:25:44.669161 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Aug 13 00:25:44.679535 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Aug 13 00:25:44.702427 systemd[1]: systemd-udevd.service: Deactivated successfully. Aug 13 00:25:44.702569 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Aug 13 00:25:44.714285 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Aug 13 00:25:44.714439 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Aug 13 00:25:44.723757 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Aug 13 00:25:44.723824 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Aug 13 00:25:44.734546 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Aug 13 00:25:44.734584 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Aug 13 00:25:44.750802 systemd[1]: dracut-cmdline.service: Deactivated successfully. Aug 13 00:25:44.750868 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Aug 13 00:25:44.761999 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Aug 13 00:25:44.762049 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Aug 13 00:25:44.789198 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Aug 13 00:25:44.796356 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Aug 13 00:25:44.796418 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Aug 13 00:25:44.807652 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Aug 13 00:25:44.807700 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 00:25:44.819275 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Aug 13 00:25:44.819348 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Aug 13 00:25:45.349988 systemd[1]: network-cleanup.service: Deactivated successfully. Aug 13 00:25:45.350101 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Aug 13 00:25:45.361447 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Aug 13 00:25:45.384155 systemd[1]: Starting initrd-switch-root.service - Switch Root... Aug 13 00:25:45.393798 systemd[1]: Switching root. Aug 13 00:25:45.457680 systemd-journald[898]: Journal stopped Aug 13 00:25:47.476636 systemd-journald[898]: Received SIGTERM from PID 1 (systemd). Aug 13 00:25:47.476667 kernel: SELinux: policy capability network_peer_controls=1 Aug 13 00:25:47.476679 kernel: SELinux: policy capability open_perms=1 Aug 13 00:25:47.476687 kernel: SELinux: policy capability extended_socket_class=1 Aug 13 00:25:47.476694 kernel: SELinux: policy capability always_check_network=0 Aug 13 00:25:47.476702 kernel: SELinux: policy capability cgroup_seclabel=1 Aug 13 00:25:47.476711 kernel: SELinux: policy capability nnp_nosuid_transition=1 Aug 13 00:25:47.476721 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Aug 13 00:25:47.476734 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Aug 13 00:25:47.476742 kernel: audit: type=1403 audit(1755044745.684:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Aug 13 00:25:47.476752 systemd[1]: Successfully loaded SELinux policy in 116.231ms. Aug 13 00:25:47.476761 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 9.588ms. Aug 13 00:25:47.476771 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Aug 13 00:25:47.476780 systemd[1]: Detected architecture arm64. Aug 13 00:25:47.476792 systemd[1]: Detected first boot. Aug 13 00:25:47.476801 systemd[1]: Hostname set to . Aug 13 00:25:47.476810 systemd[1]: Initializing machine ID from random generator. Aug 13 00:25:47.476819 zram_generator::config[2079]: No configuration found. Aug 13 00:25:47.476831 systemd[1]: Populated /etc with preset unit settings. Aug 13 00:25:47.476840 systemd[1]: initrd-switch-root.service: Deactivated successfully. Aug 13 00:25:47.476849 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Aug 13 00:25:47.476858 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Aug 13 00:25:47.476867 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Aug 13 00:25:47.476877 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Aug 13 00:25:47.476886 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Aug 13 00:25:47.476895 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Aug 13 00:25:47.476906 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Aug 13 00:25:47.476916 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Aug 13 00:25:47.476925 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Aug 13 00:25:47.476934 systemd[1]: Created slice user.slice - User and Session Slice. Aug 13 00:25:47.476943 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Aug 13 00:25:47.476952 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Aug 13 00:25:47.476961 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Aug 13 00:25:47.476972 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Aug 13 00:25:47.476981 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Aug 13 00:25:47.476991 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Aug 13 00:25:47.477000 systemd[1]: Expecting device dev-ttyAMA0.device - /dev/ttyAMA0... Aug 13 00:25:47.477009 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Aug 13 00:25:47.477018 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Aug 13 00:25:47.477027 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Aug 13 00:25:47.477042 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Aug 13 00:25:47.477052 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Aug 13 00:25:47.477063 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Aug 13 00:25:47.477073 systemd[1]: Reached target remote-fs.target - Remote File Systems. Aug 13 00:25:47.477082 systemd[1]: Reached target slices.target - Slice Units. Aug 13 00:25:47.477091 systemd[1]: Reached target swap.target - Swaps. Aug 13 00:25:47.477100 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Aug 13 00:25:47.477110 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Aug 13 00:25:47.477119 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Aug 13 00:25:47.477130 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Aug 13 00:25:47.477141 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Aug 13 00:25:47.477151 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Aug 13 00:25:47.477161 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Aug 13 00:25:47.477171 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Aug 13 00:25:47.477182 systemd[1]: Mounting media.mount - External Media Directory... Aug 13 00:25:47.477191 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Aug 13 00:25:47.477201 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Aug 13 00:25:47.477210 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Aug 13 00:25:47.477220 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Aug 13 00:25:47.477229 systemd[1]: Reached target machines.target - Containers. Aug 13 00:25:47.477239 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Aug 13 00:25:47.477248 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Aug 13 00:25:47.477259 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Aug 13 00:25:47.477269 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Aug 13 00:25:47.477278 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Aug 13 00:25:47.477288 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Aug 13 00:25:47.477297 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Aug 13 00:25:47.477307 kernel: ACPI: bus type drm_connector registered Aug 13 00:25:47.477315 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Aug 13 00:25:47.477324 kernel: fuse: init (API version 7.39) Aug 13 00:25:47.477333 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Aug 13 00:25:47.477345 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Aug 13 00:25:47.477354 kernel: loop: module loaded Aug 13 00:25:47.477363 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Aug 13 00:25:47.477373 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Aug 13 00:25:47.477382 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Aug 13 00:25:47.477391 systemd[1]: Stopped systemd-fsck-usr.service. Aug 13 00:25:47.477401 systemd[1]: Starting systemd-journald.service - Journal Service... Aug 13 00:25:47.477410 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Aug 13 00:25:47.477438 systemd-journald[2180]: Collecting audit messages is disabled. Aug 13 00:25:47.477457 systemd-journald[2180]: Journal started Aug 13 00:25:47.477478 systemd-journald[2180]: Runtime Journal (/run/log/journal/a0781b455bf64ee59a2be93153e9d231) is 8.0M, max 4.0G, 3.9G free. Aug 13 00:25:46.195562 systemd[1]: Queued start job for default target multi-user.target. Aug 13 00:25:46.213429 systemd[1]: Unnecessary job was removed for dev-nvme1n1p6.device - /dev/nvme1n1p6. Aug 13 00:25:46.213757 systemd[1]: systemd-journald.service: Deactivated successfully. Aug 13 00:25:46.214029 systemd[1]: systemd-journald.service: Consumed 3.533s CPU time. Aug 13 00:25:47.501054 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Aug 13 00:25:47.528052 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Aug 13 00:25:47.549054 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Aug 13 00:25:47.572051 systemd[1]: verity-setup.service: Deactivated successfully. Aug 13 00:25:47.572094 systemd[1]: Stopped verity-setup.service. Aug 13 00:25:47.597052 systemd[1]: Started systemd-journald.service - Journal Service. Aug 13 00:25:47.602376 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Aug 13 00:25:47.607834 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Aug 13 00:25:47.613197 systemd[1]: Mounted media.mount - External Media Directory. Aug 13 00:25:47.618479 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Aug 13 00:25:47.623781 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Aug 13 00:25:47.629022 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Aug 13 00:25:47.636067 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Aug 13 00:25:47.641707 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Aug 13 00:25:47.647238 systemd[1]: modprobe@configfs.service: Deactivated successfully. Aug 13 00:25:47.649072 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Aug 13 00:25:47.654496 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Aug 13 00:25:47.654621 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Aug 13 00:25:47.661923 systemd[1]: modprobe@drm.service: Deactivated successfully. Aug 13 00:25:47.664069 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Aug 13 00:25:47.669270 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Aug 13 00:25:47.669417 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Aug 13 00:25:47.674675 systemd[1]: modprobe@fuse.service: Deactivated successfully. Aug 13 00:25:47.674828 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Aug 13 00:25:47.680022 systemd[1]: modprobe@loop.service: Deactivated successfully. Aug 13 00:25:47.681158 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Aug 13 00:25:47.686271 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Aug 13 00:25:47.691147 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Aug 13 00:25:47.697454 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Aug 13 00:25:47.702473 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Aug 13 00:25:47.717664 systemd[1]: Reached target network-pre.target - Preparation for Network. Aug 13 00:25:47.733200 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Aug 13 00:25:47.739200 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Aug 13 00:25:47.744058 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Aug 13 00:25:47.744089 systemd[1]: Reached target local-fs.target - Local File Systems. Aug 13 00:25:47.749647 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Aug 13 00:25:47.770206 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Aug 13 00:25:47.775980 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Aug 13 00:25:47.780775 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Aug 13 00:25:47.782099 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Aug 13 00:25:47.787808 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Aug 13 00:25:47.792616 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Aug 13 00:25:47.793673 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Aug 13 00:25:47.798680 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Aug 13 00:25:47.799579 systemd-journald[2180]: Time spent on flushing to /var/log/journal/a0781b455bf64ee59a2be93153e9d231 is 25.878ms for 2352 entries. Aug 13 00:25:47.799579 systemd-journald[2180]: System Journal (/var/log/journal/a0781b455bf64ee59a2be93153e9d231) is 8.0M, max 195.6M, 187.6M free. Aug 13 00:25:47.842723 systemd-journald[2180]: Received client request to flush runtime journal. Aug 13 00:25:47.842773 kernel: loop0: detected capacity change from 0 to 114328 Aug 13 00:25:47.799794 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Aug 13 00:25:47.817892 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Aug 13 00:25:47.823671 systemd[1]: Starting systemd-sysusers.service - Create System Users... Aug 13 00:25:47.829417 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Aug 13 00:25:47.845890 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Aug 13 00:25:47.856044 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Aug 13 00:25:47.859878 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Aug 13 00:25:47.866068 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Aug 13 00:25:47.870813 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Aug 13 00:25:47.876624 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Aug 13 00:25:47.881295 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Aug 13 00:25:47.887412 systemd[1]: Finished systemd-sysusers.service - Create System Users. Aug 13 00:25:47.897682 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Aug 13 00:25:47.913047 kernel: loop1: detected capacity change from 0 to 211168 Aug 13 00:25:47.922298 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Aug 13 00:25:47.928539 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Aug 13 00:25:47.934208 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Aug 13 00:25:47.934954 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Aug 13 00:25:47.942629 udevadm[2219]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Aug 13 00:25:47.950433 systemd-tmpfiles[2243]: ACLs are not supported, ignoring. Aug 13 00:25:47.950445 systemd-tmpfiles[2243]: ACLs are not supported, ignoring. Aug 13 00:25:47.954381 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Aug 13 00:25:47.965046 kernel: loop2: detected capacity change from 0 to 8 Aug 13 00:25:48.015049 kernel: loop3: detected capacity change from 0 to 114432 Aug 13 00:25:48.029000 ldconfig[2207]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Aug 13 00:25:48.030648 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Aug 13 00:25:48.073051 kernel: loop4: detected capacity change from 0 to 114328 Aug 13 00:25:48.089050 kernel: loop5: detected capacity change from 0 to 211168 Aug 13 00:25:48.106050 kernel: loop6: detected capacity change from 0 to 8 Aug 13 00:25:48.118046 kernel: loop7: detected capacity change from 0 to 114432 Aug 13 00:25:48.121591 (sd-merge)[2254]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-packet'. Aug 13 00:25:48.122013 (sd-merge)[2254]: Merged extensions into '/usr'. Aug 13 00:25:48.124908 systemd[1]: Reloading requested from client PID 2216 ('systemd-sysext') (unit systemd-sysext.service)... Aug 13 00:25:48.124922 systemd[1]: Reloading... Aug 13 00:25:48.174048 zram_generator::config[2279]: No configuration found. Aug 13 00:25:48.267413 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Aug 13 00:25:48.315864 systemd[1]: Reloading finished in 190 ms. Aug 13 00:25:48.344668 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Aug 13 00:25:48.349980 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Aug 13 00:25:48.372245 systemd[1]: Starting ensure-sysext.service... Aug 13 00:25:48.378342 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Aug 13 00:25:48.384804 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Aug 13 00:25:48.391618 systemd[1]: Reloading requested from client PID 2332 ('systemctl') (unit ensure-sysext.service)... Aug 13 00:25:48.391633 systemd[1]: Reloading... Aug 13 00:25:48.398437 systemd-tmpfiles[2334]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Aug 13 00:25:48.398685 systemd-tmpfiles[2334]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Aug 13 00:25:48.399303 systemd-tmpfiles[2334]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Aug 13 00:25:48.399509 systemd-tmpfiles[2334]: ACLs are not supported, ignoring. Aug 13 00:25:48.399551 systemd-tmpfiles[2334]: ACLs are not supported, ignoring. Aug 13 00:25:48.402145 systemd-tmpfiles[2334]: Detected autofs mount point /boot during canonicalization of boot. Aug 13 00:25:48.402150 systemd-tmpfiles[2334]: Skipping /boot Aug 13 00:25:48.408925 systemd-tmpfiles[2334]: Detected autofs mount point /boot during canonicalization of boot. Aug 13 00:25:48.408932 systemd-tmpfiles[2334]: Skipping /boot Aug 13 00:25:48.410710 systemd-udevd[2335]: Using default interface naming scheme 'v255'. Aug 13 00:25:48.437045 zram_generator::config[2363]: No configuration found. Aug 13 00:25:48.470051 kernel: BTRFS warning: duplicate device /dev/nvme1n1p3 devid 1 generation 41 scanned by (udev-worker) (2369) Aug 13 00:25:48.494052 kernel: IPMI message handler: version 39.2 Aug 13 00:25:48.505064 kernel: ipmi device interface Aug 13 00:25:48.516053 kernel: ipmi_ssif: IPMI SSIF Interface driver Aug 13 00:25:48.516135 kernel: ipmi_si: IPMI System Interface driver Aug 13 00:25:48.529607 kernel: ipmi_si: Unable to find any System Interface(s) Aug 13 00:25:48.552905 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Aug 13 00:25:48.616410 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - SAMSUNG MZ1LB960HAJQ-00007 OEM. Aug 13 00:25:48.620967 systemd[1]: Condition check resulted in dev-ttyAMA0.device - /dev/ttyAMA0 being skipped. Aug 13 00:25:48.621251 systemd[1]: Reloading finished in 229 ms. Aug 13 00:25:48.640664 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Aug 13 00:25:48.657379 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Aug 13 00:25:48.677880 systemd[1]: Finished ensure-sysext.service. Aug 13 00:25:48.682724 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Aug 13 00:25:48.713175 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Aug 13 00:25:48.719374 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Aug 13 00:25:48.724632 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Aug 13 00:25:48.725737 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Aug 13 00:25:48.731732 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Aug 13 00:25:48.737730 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Aug 13 00:25:48.743555 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Aug 13 00:25:48.744287 lvm[2555]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Aug 13 00:25:48.749548 augenrules[2574]: No rules Aug 13 00:25:48.749546 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Aug 13 00:25:48.754488 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Aug 13 00:25:48.755395 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Aug 13 00:25:48.761408 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Aug 13 00:25:48.768050 systemd[1]: Starting systemd-networkd.service - Network Configuration... Aug 13 00:25:48.774821 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Aug 13 00:25:48.781115 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Aug 13 00:25:48.786721 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Aug 13 00:25:48.792345 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 00:25:48.797753 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Aug 13 00:25:48.802688 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Aug 13 00:25:48.807676 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Aug 13 00:25:48.812544 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Aug 13 00:25:48.812672 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Aug 13 00:25:48.817656 systemd[1]: modprobe@drm.service: Deactivated successfully. Aug 13 00:25:48.817781 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Aug 13 00:25:48.823268 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Aug 13 00:25:48.823424 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Aug 13 00:25:48.828207 systemd[1]: modprobe@loop.service: Deactivated successfully. Aug 13 00:25:48.828325 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Aug 13 00:25:48.833131 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Aug 13 00:25:48.838345 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Aug 13 00:25:48.845453 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 00:25:48.857647 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Aug 13 00:25:48.863260 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Aug 13 00:25:48.880435 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Aug 13 00:25:48.884738 lvm[2613]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Aug 13 00:25:48.884966 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Aug 13 00:25:48.885034 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Aug 13 00:25:48.886207 systemd[1]: Starting systemd-update-done.service - Update is Completed... Aug 13 00:25:48.892723 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Aug 13 00:25:48.897591 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Aug 13 00:25:48.900217 systemd[1]: Finished systemd-update-done.service - Update is Completed. Aug 13 00:25:48.926193 systemd[1]: Started systemd-userdbd.service - User Database Manager. Aug 13 00:25:48.931496 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Aug 13 00:25:48.978906 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Aug 13 00:25:48.983577 systemd-resolved[2589]: Positive Trust Anchors: Aug 13 00:25:48.983589 systemd-resolved[2589]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Aug 13 00:25:48.983621 systemd-resolved[2589]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Aug 13 00:25:48.983943 systemd[1]: Reached target time-set.target - System Time Set. Aug 13 00:25:48.987270 systemd-resolved[2589]: Using system hostname 'ci-4081.3.5-a-588adb827d'. Aug 13 00:25:48.988861 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Aug 13 00:25:48.992387 systemd-networkd[2588]: lo: Link UP Aug 13 00:25:48.992394 systemd-networkd[2588]: lo: Gained carrier Aug 13 00:25:48.994326 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Aug 13 00:25:48.996072 systemd-networkd[2588]: bond0: netdev ready Aug 13 00:25:48.998743 systemd[1]: Reached target sysinit.target - System Initialization. Aug 13 00:25:49.003109 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Aug 13 00:25:49.005446 systemd-networkd[2588]: Enumeration completed Aug 13 00:25:49.007429 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Aug 13 00:25:49.011970 systemd[1]: Started logrotate.timer - Daily rotation of log files. Aug 13 00:25:49.016407 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Aug 13 00:25:49.019642 systemd-networkd[2588]: enP1p1s0f0np0: Configuring with /etc/systemd/network/10-0c:42:a1:5a:96:f8.network. Aug 13 00:25:49.020810 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Aug 13 00:25:49.025251 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Aug 13 00:25:49.025272 systemd[1]: Reached target paths.target - Path Units. Aug 13 00:25:49.029664 systemd[1]: Reached target timers.target - Timer Units. Aug 13 00:25:49.034640 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Aug 13 00:25:49.040434 systemd[1]: Starting docker.socket - Docker Socket for the API... Aug 13 00:25:49.048032 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Aug 13 00:25:49.052978 systemd[1]: Started systemd-networkd.service - Network Configuration. Aug 13 00:25:49.057632 systemd[1]: Listening on docker.socket - Docker Socket for the API. Aug 13 00:25:49.062208 systemd[1]: Reached target network.target - Network. Aug 13 00:25:49.066634 systemd[1]: Reached target sockets.target - Socket Units. Aug 13 00:25:49.070929 systemd[1]: Reached target basic.target - Basic System. Aug 13 00:25:49.075432 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Aug 13 00:25:49.075452 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Aug 13 00:25:49.086140 systemd[1]: Starting containerd.service - containerd container runtime... Aug 13 00:25:49.091718 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Aug 13 00:25:49.097364 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Aug 13 00:25:49.102970 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Aug 13 00:25:49.108676 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Aug 13 00:25:49.113180 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Aug 13 00:25:49.113715 coreos-metadata[2642]: Aug 13 00:25:49.113 INFO Fetching https://metadata.packet.net/metadata: Attempt #1 Aug 13 00:25:49.113915 jq[2646]: false Aug 13 00:25:49.114340 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Aug 13 00:25:49.116102 coreos-metadata[2642]: Aug 13 00:25:49.116 INFO Failed to fetch: error sending request for url (https://metadata.packet.net/metadata) Aug 13 00:25:49.119659 dbus-daemon[2643]: [system] SELinux support is enabled Aug 13 00:25:49.119873 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Aug 13 00:25:49.125511 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Aug 13 00:25:49.128514 extend-filesystems[2647]: Found loop4 Aug 13 00:25:49.134844 extend-filesystems[2647]: Found loop5 Aug 13 00:25:49.134844 extend-filesystems[2647]: Found loop6 Aug 13 00:25:49.134844 extend-filesystems[2647]: Found loop7 Aug 13 00:25:49.134844 extend-filesystems[2647]: Found nvme1n1 Aug 13 00:25:49.134844 extend-filesystems[2647]: Found nvme1n1p1 Aug 13 00:25:49.134844 extend-filesystems[2647]: Found nvme1n1p2 Aug 13 00:25:49.134844 extend-filesystems[2647]: Found nvme1n1p3 Aug 13 00:25:49.134844 extend-filesystems[2647]: Found usr Aug 13 00:25:49.134844 extend-filesystems[2647]: Found nvme1n1p4 Aug 13 00:25:49.134844 extend-filesystems[2647]: Found nvme1n1p6 Aug 13 00:25:49.134844 extend-filesystems[2647]: Found nvme1n1p7 Aug 13 00:25:49.134844 extend-filesystems[2647]: Found nvme1n1p9 Aug 13 00:25:49.134844 extend-filesystems[2647]: Checking size of /dev/nvme1n1p9 Aug 13 00:25:49.266222 kernel: EXT4-fs (nvme1n1p9): resizing filesystem from 553472 to 233815889 blocks Aug 13 00:25:49.266248 kernel: BTRFS warning: duplicate device /dev/nvme1n1p3 devid 1 generation 41 scanned by (udev-worker) (2412) Aug 13 00:25:49.131365 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Aug 13 00:25:49.266319 extend-filesystems[2647]: Resized partition /dev/nvme1n1p9 Aug 13 00:25:49.260299 dbus-daemon[2643]: [system] Successfully activated service 'org.freedesktop.systemd1' Aug 13 00:25:49.143627 systemd[1]: Starting systemd-logind.service - User Login Management... Aug 13 00:25:49.271019 extend-filesystems[2667]: resize2fs 1.47.1 (20-May-2024) Aug 13 00:25:49.149503 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Aug 13 00:25:49.190172 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Aug 13 00:25:49.190861 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Aug 13 00:25:49.191597 systemd[1]: Starting update-engine.service - Update Engine... Aug 13 00:25:49.280861 update_engine[2678]: I20250813 00:25:49.241050 2678 main.cc:92] Flatcar Update Engine starting Aug 13 00:25:49.280861 update_engine[2678]: I20250813 00:25:49.247049 2678 update_check_scheduler.cc:74] Next update check in 2m45s Aug 13 00:25:49.198186 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Aug 13 00:25:49.281198 jq[2679]: true Aug 13 00:25:49.206457 systemd[1]: Started dbus.service - D-Bus System Message Bus. Aug 13 00:25:49.219560 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Aug 13 00:25:49.281463 tar[2683]: linux-arm64/LICENSE Aug 13 00:25:49.281463 tar[2683]: linux-arm64/helm Aug 13 00:25:49.221072 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Aug 13 00:25:49.221426 systemd[1]: motdgen.service: Deactivated successfully. Aug 13 00:25:49.281833 jq[2684]: true Aug 13 00:25:49.221610 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Aug 13 00:25:49.225819 systemd-logind[2668]: Watching system buttons on /dev/input/event0 (Power Button) Aug 13 00:25:49.229137 systemd-logind[2668]: New seat seat0. Aug 13 00:25:49.230137 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Aug 13 00:25:49.230325 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Aug 13 00:25:49.244638 systemd[1]: Started systemd-logind.service - User Login Management. Aug 13 00:25:49.248874 (ntainerd)[2685]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Aug 13 00:25:49.270054 systemd[1]: Started update-engine.service - Update Engine. Aug 13 00:25:49.276878 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Aug 13 00:25:49.277256 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Aug 13 00:25:49.285234 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Aug 13 00:25:49.285339 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Aug 13 00:25:49.295060 bash[2708]: Updated "/home/core/.ssh/authorized_keys" Aug 13 00:25:49.307248 systemd[1]: Started locksmithd.service - Cluster reboot manager. Aug 13 00:25:49.314708 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Aug 13 00:25:49.324762 systemd[1]: Starting sshkeys.service... Aug 13 00:25:49.335358 locksmithd[2709]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Aug 13 00:25:49.338105 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Aug 13 00:25:49.344283 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Aug 13 00:25:49.364547 coreos-metadata[2726]: Aug 13 00:25:49.364 INFO Fetching https://metadata.packet.net/metadata: Attempt #1 Aug 13 00:25:49.365662 coreos-metadata[2726]: Aug 13 00:25:49.365 INFO Failed to fetch: error sending request for url (https://metadata.packet.net/metadata) Aug 13 00:25:49.391374 containerd[2685]: time="2025-08-13T00:25:49.391253880Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Aug 13 00:25:49.414024 containerd[2685]: time="2025-08-13T00:25:49.413981280Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Aug 13 00:25:49.415229 containerd[2685]: time="2025-08-13T00:25:49.415200200Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.100-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Aug 13 00:25:49.415272 containerd[2685]: time="2025-08-13T00:25:49.415226800Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Aug 13 00:25:49.415272 containerd[2685]: time="2025-08-13T00:25:49.415241040Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Aug 13 00:25:49.415401 containerd[2685]: time="2025-08-13T00:25:49.415385440Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Aug 13 00:25:49.415423 containerd[2685]: time="2025-08-13T00:25:49.415407080Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Aug 13 00:25:49.415479 containerd[2685]: time="2025-08-13T00:25:49.415464640Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Aug 13 00:25:49.415499 containerd[2685]: time="2025-08-13T00:25:49.415479080Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Aug 13 00:25:49.415644 containerd[2685]: time="2025-08-13T00:25:49.415627160Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Aug 13 00:25:49.415665 containerd[2685]: time="2025-08-13T00:25:49.415643840Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Aug 13 00:25:49.415665 containerd[2685]: time="2025-08-13T00:25:49.415657440Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Aug 13 00:25:49.415701 containerd[2685]: time="2025-08-13T00:25:49.415668440Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Aug 13 00:25:49.415755 containerd[2685]: time="2025-08-13T00:25:49.415740920Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Aug 13 00:25:49.415983 containerd[2685]: time="2025-08-13T00:25:49.415965680Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Aug 13 00:25:49.416090 containerd[2685]: time="2025-08-13T00:25:49.416076640Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Aug 13 00:25:49.416109 containerd[2685]: time="2025-08-13T00:25:49.416091680Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Aug 13 00:25:49.416182 containerd[2685]: time="2025-08-13T00:25:49.416171720Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Aug 13 00:25:49.416222 containerd[2685]: time="2025-08-13T00:25:49.416212800Z" level=info msg="metadata content store policy set" policy=shared Aug 13 00:25:49.423252 containerd[2685]: time="2025-08-13T00:25:49.423226880Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Aug 13 00:25:49.423320 containerd[2685]: time="2025-08-13T00:25:49.423274920Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Aug 13 00:25:49.423320 containerd[2685]: time="2025-08-13T00:25:49.423289240Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Aug 13 00:25:49.423320 containerd[2685]: time="2025-08-13T00:25:49.423303600Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Aug 13 00:25:49.423320 containerd[2685]: time="2025-08-13T00:25:49.423316400Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Aug 13 00:25:49.423463 containerd[2685]: time="2025-08-13T00:25:49.423449560Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Aug 13 00:25:49.423664 containerd[2685]: time="2025-08-13T00:25:49.423649040Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Aug 13 00:25:49.423770 containerd[2685]: time="2025-08-13T00:25:49.423755880Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Aug 13 00:25:49.423791 containerd[2685]: time="2025-08-13T00:25:49.423772960Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Aug 13 00:25:49.423809 containerd[2685]: time="2025-08-13T00:25:49.423786320Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Aug 13 00:25:49.423809 containerd[2685]: time="2025-08-13T00:25:49.423802840Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Aug 13 00:25:49.423841 containerd[2685]: time="2025-08-13T00:25:49.423816480Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Aug 13 00:25:49.423841 containerd[2685]: time="2025-08-13T00:25:49.423828800Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Aug 13 00:25:49.423873 containerd[2685]: time="2025-08-13T00:25:49.423842080Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Aug 13 00:25:49.423873 containerd[2685]: time="2025-08-13T00:25:49.423857360Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Aug 13 00:25:49.423910 containerd[2685]: time="2025-08-13T00:25:49.423889600Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Aug 13 00:25:49.423910 containerd[2685]: time="2025-08-13T00:25:49.423902720Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Aug 13 00:25:49.423942 containerd[2685]: time="2025-08-13T00:25:49.423914120Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Aug 13 00:25:49.423942 containerd[2685]: time="2025-08-13T00:25:49.423933120Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Aug 13 00:25:49.423974 containerd[2685]: time="2025-08-13T00:25:49.423946680Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Aug 13 00:25:49.423974 containerd[2685]: time="2025-08-13T00:25:49.423959000Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Aug 13 00:25:49.423974 containerd[2685]: time="2025-08-13T00:25:49.423971160Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Aug 13 00:25:49.424030 containerd[2685]: time="2025-08-13T00:25:49.423983120Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Aug 13 00:25:49.424030 containerd[2685]: time="2025-08-13T00:25:49.423996400Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Aug 13 00:25:49.424030 containerd[2685]: time="2025-08-13T00:25:49.424009480Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Aug 13 00:25:49.424030 containerd[2685]: time="2025-08-13T00:25:49.424025880Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Aug 13 00:25:49.424103 containerd[2685]: time="2025-08-13T00:25:49.424045600Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Aug 13 00:25:49.424103 containerd[2685]: time="2025-08-13T00:25:49.424060320Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Aug 13 00:25:49.424103 containerd[2685]: time="2025-08-13T00:25:49.424073520Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Aug 13 00:25:49.424103 containerd[2685]: time="2025-08-13T00:25:49.424084760Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Aug 13 00:25:49.424166 containerd[2685]: time="2025-08-13T00:25:49.424107600Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Aug 13 00:25:49.424166 containerd[2685]: time="2025-08-13T00:25:49.424124320Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Aug 13 00:25:49.424166 containerd[2685]: time="2025-08-13T00:25:49.424144320Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Aug 13 00:25:49.424166 containerd[2685]: time="2025-08-13T00:25:49.424156480Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Aug 13 00:25:49.424233 containerd[2685]: time="2025-08-13T00:25:49.424168240Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Aug 13 00:25:49.424286 containerd[2685]: time="2025-08-13T00:25:49.424276520Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Aug 13 00:25:49.424305 containerd[2685]: time="2025-08-13T00:25:49.424294440Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Aug 13 00:25:49.424327 containerd[2685]: time="2025-08-13T00:25:49.424305640Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Aug 13 00:25:49.424327 containerd[2685]: time="2025-08-13T00:25:49.424318040Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Aug 13 00:25:49.424368 containerd[2685]: time="2025-08-13T00:25:49.424327560Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Aug 13 00:25:49.424368 containerd[2685]: time="2025-08-13T00:25:49.424342320Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Aug 13 00:25:49.424368 containerd[2685]: time="2025-08-13T00:25:49.424353240Z" level=info msg="NRI interface is disabled by configuration." Aug 13 00:25:49.424368 containerd[2685]: time="2025-08-13T00:25:49.424365840Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Aug 13 00:25:49.424735 containerd[2685]: time="2025-08-13T00:25:49.424688280Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Aug 13 00:25:49.424830 containerd[2685]: time="2025-08-13T00:25:49.424744040Z" level=info msg="Connect containerd service" Aug 13 00:25:49.424830 containerd[2685]: time="2025-08-13T00:25:49.424768720Z" level=info msg="using legacy CRI server" Aug 13 00:25:49.424830 containerd[2685]: time="2025-08-13T00:25:49.424774840Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Aug 13 00:25:49.424887 containerd[2685]: time="2025-08-13T00:25:49.424856880Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Aug 13 00:25:49.426181 containerd[2685]: time="2025-08-13T00:25:49.426154760Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Aug 13 00:25:49.426403 containerd[2685]: time="2025-08-13T00:25:49.426365560Z" level=info msg="Start subscribing containerd event" Aug 13 00:25:49.426432 containerd[2685]: time="2025-08-13T00:25:49.426423320Z" level=info msg="Start recovering state" Aug 13 00:25:49.426514 containerd[2685]: time="2025-08-13T00:25:49.426504640Z" level=info msg="Start event monitor" Aug 13 00:25:49.426533 containerd[2685]: time="2025-08-13T00:25:49.426519560Z" level=info msg="Start snapshots syncer" Aug 13 00:25:49.426533 containerd[2685]: time="2025-08-13T00:25:49.426529560Z" level=info msg="Start cni network conf syncer for default" Aug 13 00:25:49.426566 containerd[2685]: time="2025-08-13T00:25:49.426536880Z" level=info msg="Start streaming server" Aug 13 00:25:49.426630 containerd[2685]: time="2025-08-13T00:25:49.426615000Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Aug 13 00:25:49.426666 containerd[2685]: time="2025-08-13T00:25:49.426657040Z" level=info msg=serving... address=/run/containerd/containerd.sock Aug 13 00:25:49.426716 containerd[2685]: time="2025-08-13T00:25:49.426706800Z" level=info msg="containerd successfully booted in 0.036251s" Aug 13 00:25:49.426765 systemd[1]: Started containerd.service - containerd container runtime. Aug 13 00:25:49.555517 sshd_keygen[2672]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Aug 13 00:25:49.574757 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Aug 13 00:25:49.592174 tar[2683]: linux-arm64/README.md Aug 13 00:25:49.592315 systemd[1]: Starting issuegen.service - Generate /run/issue... Aug 13 00:25:49.603808 systemd[1]: issuegen.service: Deactivated successfully. Aug 13 00:25:49.603996 systemd[1]: Finished issuegen.service - Generate /run/issue. Aug 13 00:25:49.610033 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Aug 13 00:25:49.617469 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Aug 13 00:25:49.630684 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Aug 13 00:25:49.636996 systemd[1]: Started getty@tty1.service - Getty on tty1. Aug 13 00:25:49.643220 systemd[1]: Started serial-getty@ttyAMA0.service - Serial Getty on ttyAMA0. Aug 13 00:25:49.648239 systemd[1]: Reached target getty.target - Login Prompts. Aug 13 00:25:49.674052 kernel: EXT4-fs (nvme1n1p9): resized filesystem to 233815889 Aug 13 00:25:49.689553 extend-filesystems[2667]: Filesystem at /dev/nvme1n1p9 is mounted on /; on-line resizing required Aug 13 00:25:49.689553 extend-filesystems[2667]: old_desc_blocks = 1, new_desc_blocks = 112 Aug 13 00:25:49.689553 extend-filesystems[2667]: The filesystem on /dev/nvme1n1p9 is now 233815889 (4k) blocks long. Aug 13 00:25:49.716615 extend-filesystems[2647]: Resized filesystem in /dev/nvme1n1p9 Aug 13 00:25:49.716615 extend-filesystems[2647]: Found nvme0n1 Aug 13 00:25:49.692155 systemd[1]: extend-filesystems.service: Deactivated successfully. Aug 13 00:25:49.692456 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Aug 13 00:25:50.116249 coreos-metadata[2642]: Aug 13 00:25:50.116 INFO Fetching https://metadata.packet.net/metadata: Attempt #2 Aug 13 00:25:50.116681 coreos-metadata[2642]: Aug 13 00:25:50.116 INFO Failed to fetch: error sending request for url (https://metadata.packet.net/metadata) Aug 13 00:25:50.336053 kernel: mlx5_core 0001:01:00.0 enP1p1s0f0np0: Link up Aug 13 00:25:50.352054 kernel: bond0: (slave enP1p1s0f0np0): Enslaving as a backup interface with an up link Aug 13 00:25:50.356449 systemd-networkd[2588]: enP1p1s0f1np1: Configuring with /etc/systemd/network/10-0c:42:a1:5a:96:f9.network. Aug 13 00:25:50.365791 coreos-metadata[2726]: Aug 13 00:25:50.365 INFO Fetching https://metadata.packet.net/metadata: Attempt #2 Aug 13 00:25:50.366137 coreos-metadata[2726]: Aug 13 00:25:50.366 INFO Failed to fetch: error sending request for url (https://metadata.packet.net/metadata) Aug 13 00:25:50.960054 kernel: mlx5_core 0001:01:00.1 enP1p1s0f1np1: Link up Aug 13 00:25:50.976952 systemd-networkd[2588]: bond0: Configuring with /etc/systemd/network/05-bond0.network. Aug 13 00:25:50.977046 kernel: bond0: (slave enP1p1s0f1np1): Enslaving as a backup interface with an up link Aug 13 00:25:50.978251 systemd-networkd[2588]: enP1p1s0f0np0: Link UP Aug 13 00:25:50.978576 systemd-networkd[2588]: enP1p1s0f0np0: Gained carrier Aug 13 00:25:50.997052 kernel: bond0: Warning: No 802.3ad response from the link partner for any adapters in the bond Aug 13 00:25:51.007511 systemd-networkd[2588]: enP1p1s0f1np1: Reconfiguring with /etc/systemd/network/10-0c:42:a1:5a:96:f8.network. Aug 13 00:25:51.007832 systemd-networkd[2588]: enP1p1s0f1np1: Link UP Aug 13 00:25:51.008119 systemd-networkd[2588]: enP1p1s0f1np1: Gained carrier Aug 13 00:25:51.018254 systemd-networkd[2588]: bond0: Link UP Aug 13 00:25:51.018557 systemd-networkd[2588]: bond0: Gained carrier Aug 13 00:25:51.018750 systemd-timesyncd[2591]: Network configuration changed, trying to establish connection. Aug 13 00:25:51.019311 systemd-timesyncd[2591]: Network configuration changed, trying to establish connection. Aug 13 00:25:51.019644 systemd-timesyncd[2591]: Network configuration changed, trying to establish connection. Aug 13 00:25:51.019782 systemd-timesyncd[2591]: Network configuration changed, trying to establish connection. Aug 13 00:25:51.097318 kernel: bond0: (slave enP1p1s0f0np0): link status definitely up, 25000 Mbps full duplex Aug 13 00:25:51.097347 kernel: bond0: active interface up! Aug 13 00:25:51.221049 kernel: bond0: (slave enP1p1s0f1np1): link status definitely up, 25000 Mbps full duplex Aug 13 00:25:52.103305 systemd-timesyncd[2591]: Network configuration changed, trying to establish connection. Aug 13 00:25:52.116777 coreos-metadata[2642]: Aug 13 00:25:52.116 INFO Fetching https://metadata.packet.net/metadata: Attempt #3 Aug 13 00:25:52.358076 systemd-networkd[2588]: bond0: Gained IPv6LL Aug 13 00:25:52.358267 systemd-timesyncd[2591]: Network configuration changed, trying to establish connection. Aug 13 00:25:52.360185 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Aug 13 00:25:52.365913 systemd[1]: Reached target network-online.target - Network is Online. Aug 13 00:25:52.366216 coreos-metadata[2726]: Aug 13 00:25:52.366 INFO Fetching https://metadata.packet.net/metadata: Attempt #3 Aug 13 00:25:52.388314 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 00:25:52.394959 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Aug 13 00:25:52.416484 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Aug 13 00:25:53.071087 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 00:25:53.077092 (kubelet)[2786]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Aug 13 00:25:53.492308 kubelet[2786]: E0813 00:25:53.492214 2786 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Aug 13 00:25:53.495126 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Aug 13 00:25:53.495273 systemd[1]: kubelet.service: Failed with result 'exit-code'. Aug 13 00:25:54.309443 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Aug 13 00:25:54.324352 systemd[1]: Started sshd@0-147.75.53.22:22-147.75.109.163:48590.service - OpenSSH per-connection server daemon (147.75.109.163:48590). Aug 13 00:25:54.482944 kernel: mlx5_core 0001:01:00.0: lag map: port 1:1 port 2:2 Aug 13 00:25:54.483250 kernel: mlx5_core 0001:01:00.0: shared_fdb:0 mode:queue_affinity Aug 13 00:25:54.608446 sshd[2808]: Accepted publickey for core from 147.75.109.163 port 48590 ssh2: RSA SHA256:u7vn2r+eDvbQsVz87adSciVbwQCBuSV2wy8SWfSsl1k Aug 13 00:25:54.610267 sshd[2808]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:25:54.618085 systemd-logind[2668]: New session 1 of user core. Aug 13 00:25:54.619532 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Aug 13 00:25:54.638268 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Aug 13 00:25:54.648569 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Aug 13 00:25:54.658109 systemd[1]: Starting user@500.service - User Manager for UID 500... Aug 13 00:25:54.666503 (systemd)[2815]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Aug 13 00:25:54.681928 login[2765]: pam_unix(login:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:25:54.684852 systemd-logind[2668]: New session 2 of user core. Aug 13 00:25:54.688484 login[2764]: pam_unix(login:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:25:54.691232 systemd-logind[2668]: New session 3 of user core. Aug 13 00:25:54.762737 systemd[2815]: Queued start job for default target default.target. Aug 13 00:25:54.774122 systemd[2815]: Created slice app.slice - User Application Slice. Aug 13 00:25:54.774150 systemd[2815]: Reached target paths.target - Paths. Aug 13 00:25:54.774162 systemd[2815]: Reached target timers.target - Timers. Aug 13 00:25:54.775448 systemd[2815]: Starting dbus.socket - D-Bus User Message Bus Socket... Aug 13 00:25:54.784767 systemd[2815]: Listening on dbus.socket - D-Bus User Message Bus Socket. Aug 13 00:25:54.784821 systemd[2815]: Reached target sockets.target - Sockets. Aug 13 00:25:54.784835 systemd[2815]: Reached target basic.target - Basic System. Aug 13 00:25:54.784876 systemd[2815]: Reached target default.target - Main User Target. Aug 13 00:25:54.784901 systemd[2815]: Startup finished in 113ms. Aug 13 00:25:54.785310 systemd[1]: Started user@500.service - User Manager for UID 500. Aug 13 00:25:54.786878 systemd[1]: Started session-1.scope - Session 1 of User core. Aug 13 00:25:54.787777 systemd[1]: Started session-2.scope - Session 2 of User core. Aug 13 00:25:54.788738 systemd[1]: Started session-3.scope - Session 3 of User core. Aug 13 00:25:55.012685 systemd[1]: Started sshd@1-147.75.53.22:22-147.75.109.163:48604.service - OpenSSH per-connection server daemon (147.75.109.163:48604). Aug 13 00:25:55.123380 coreos-metadata[2642]: Aug 13 00:25:55.123 INFO Fetch successful Aug 13 00:25:55.140636 coreos-metadata[2726]: Aug 13 00:25:55.140 INFO Fetch successful Aug 13 00:25:55.189331 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Aug 13 00:25:55.191339 systemd[1]: Starting packet-phone-home.service - Report Success to Packet... Aug 13 00:25:55.194951 unknown[2726]: wrote ssh authorized keys file for user: core Aug 13 00:25:55.212096 update-ssh-keys[2866]: Updated "/home/core/.ssh/authorized_keys" Aug 13 00:25:55.213275 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Aug 13 00:25:55.214784 systemd[1]: Finished sshkeys.service. Aug 13 00:25:55.312619 sshd[2855]: Accepted publickey for core from 147.75.109.163 port 48604 ssh2: RSA SHA256:u7vn2r+eDvbQsVz87adSciVbwQCBuSV2wy8SWfSsl1k Aug 13 00:25:55.313920 sshd[2855]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:25:55.317146 systemd-logind[2668]: New session 4 of user core. Aug 13 00:25:55.326207 systemd[1]: Started session-4.scope - Session 4 of User core. Aug 13 00:25:55.530473 sshd[2855]: pam_unix(sshd:session): session closed for user core Aug 13 00:25:55.533292 systemd[1]: sshd@1-147.75.53.22:22-147.75.109.163:48604.service: Deactivated successfully. Aug 13 00:25:55.535505 systemd[1]: session-4.scope: Deactivated successfully. Aug 13 00:25:55.536132 systemd-logind[2668]: Session 4 logged out. Waiting for processes to exit. Aug 13 00:25:55.536713 systemd-logind[2668]: Removed session 4. Aug 13 00:25:55.585267 systemd[1]: Started sshd@2-147.75.53.22:22-147.75.109.163:48612.service - OpenSSH per-connection server daemon (147.75.109.163:48612). Aug 13 00:25:55.887572 sshd[2877]: Accepted publickey for core from 147.75.109.163 port 48612 ssh2: RSA SHA256:u7vn2r+eDvbQsVz87adSciVbwQCBuSV2wy8SWfSsl1k Aug 13 00:25:55.888620 sshd[2877]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:25:55.891402 systemd-logind[2668]: New session 5 of user core. Aug 13 00:25:55.904147 systemd[1]: Started session-5.scope - Session 5 of User core. Aug 13 00:25:56.103345 sshd[2877]: pam_unix(sshd:session): session closed for user core Aug 13 00:25:56.105941 systemd[1]: sshd@2-147.75.53.22:22-147.75.109.163:48612.service: Deactivated successfully. Aug 13 00:25:56.108259 systemd[1]: session-5.scope: Deactivated successfully. Aug 13 00:25:56.108718 systemd-logind[2668]: Session 5 logged out. Waiting for processes to exit. Aug 13 00:25:56.109235 systemd-logind[2668]: Removed session 5. Aug 13 00:25:57.642001 systemd[1]: Finished packet-phone-home.service - Report Success to Packet. Aug 13 00:25:57.642478 systemd[1]: Reached target multi-user.target - Multi-User System. Aug 13 00:25:57.643121 systemd[1]: Startup finished in 3.234s (kernel) + 25.205s (initrd) + 12.075s (userspace) = 40.516s. Aug 13 00:26:03.562914 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Aug 13 00:26:03.572175 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 00:26:03.682440 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 00:26:03.686287 (kubelet)[2893]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Aug 13 00:26:03.727340 kubelet[2893]: E0813 00:26:03.727308 2893 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Aug 13 00:26:03.730323 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Aug 13 00:26:03.730458 systemd[1]: kubelet.service: Failed with result 'exit-code'. Aug 13 00:26:06.151296 systemd[1]: Started sshd@3-147.75.53.22:22-147.75.109.163:36258.service - OpenSSH per-connection server daemon (147.75.109.163:36258). Aug 13 00:26:06.429954 sshd[2917]: Accepted publickey for core from 147.75.109.163 port 36258 ssh2: RSA SHA256:u7vn2r+eDvbQsVz87adSciVbwQCBuSV2wy8SWfSsl1k Aug 13 00:26:06.431009 sshd[2917]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:26:06.434019 systemd-logind[2668]: New session 6 of user core. Aug 13 00:26:06.448202 systemd[1]: Started session-6.scope - Session 6 of User core. Aug 13 00:26:06.635109 sshd[2917]: pam_unix(sshd:session): session closed for user core Aug 13 00:26:06.638549 systemd[1]: sshd@3-147.75.53.22:22-147.75.109.163:36258.service: Deactivated successfully. Aug 13 00:26:06.640303 systemd[1]: session-6.scope: Deactivated successfully. Aug 13 00:26:06.641491 systemd-logind[2668]: Session 6 logged out. Waiting for processes to exit. Aug 13 00:26:06.642042 systemd-logind[2668]: Removed session 6. Aug 13 00:26:06.682291 systemd[1]: Started sshd@4-147.75.53.22:22-147.75.109.163:36212.service - OpenSSH per-connection server daemon (147.75.109.163:36212). Aug 13 00:26:06.956886 sshd[2924]: Accepted publickey for core from 147.75.109.163 port 36212 ssh2: RSA SHA256:u7vn2r+eDvbQsVz87adSciVbwQCBuSV2wy8SWfSsl1k Aug 13 00:26:06.957923 sshd[2924]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:26:06.960543 systemd-logind[2668]: New session 7 of user core. Aug 13 00:26:06.970152 systemd[1]: Started session-7.scope - Session 7 of User core. Aug 13 00:26:07.156250 sshd[2924]: pam_unix(sshd:session): session closed for user core Aug 13 00:26:07.158761 systemd[1]: sshd@4-147.75.53.22:22-147.75.109.163:36212.service: Deactivated successfully. Aug 13 00:26:07.160176 systemd[1]: session-7.scope: Deactivated successfully. Aug 13 00:26:07.160616 systemd-logind[2668]: Session 7 logged out. Waiting for processes to exit. Aug 13 00:26:07.161162 systemd-logind[2668]: Removed session 7. Aug 13 00:26:07.212165 systemd[1]: Started sshd@5-147.75.53.22:22-147.75.109.163:36220.service - OpenSSH per-connection server daemon (147.75.109.163:36220). Aug 13 00:26:07.514791 sshd[2931]: Accepted publickey for core from 147.75.109.163 port 36220 ssh2: RSA SHA256:u7vn2r+eDvbQsVz87adSciVbwQCBuSV2wy8SWfSsl1k Aug 13 00:26:07.515805 sshd[2931]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:26:07.518612 systemd-logind[2668]: New session 8 of user core. Aug 13 00:26:07.528149 systemd[1]: Started session-8.scope - Session 8 of User core. Aug 13 00:26:07.733996 sshd[2931]: pam_unix(sshd:session): session closed for user core Aug 13 00:26:07.737570 systemd[1]: sshd@5-147.75.53.22:22-147.75.109.163:36220.service: Deactivated successfully. Aug 13 00:26:07.739293 systemd[1]: session-8.scope: Deactivated successfully. Aug 13 00:26:07.740491 systemd-logind[2668]: Session 8 logged out. Waiting for processes to exit. Aug 13 00:26:07.741056 systemd-logind[2668]: Removed session 8. Aug 13 00:26:07.788206 systemd[1]: Started sshd@6-147.75.53.22:22-147.75.109.163:36234.service - OpenSSH per-connection server daemon (147.75.109.163:36234). Aug 13 00:26:08.092713 sshd[2938]: Accepted publickey for core from 147.75.109.163 port 36234 ssh2: RSA SHA256:u7vn2r+eDvbQsVz87adSciVbwQCBuSV2wy8SWfSsl1k Aug 13 00:26:08.093773 sshd[2938]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:26:08.096482 systemd-logind[2668]: New session 9 of user core. Aug 13 00:26:08.105144 systemd[1]: Started session-9.scope - Session 9 of User core. Aug 13 00:26:08.275220 sudo[2941]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Aug 13 00:26:08.275484 sudo[2941]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Aug 13 00:26:08.287785 sudo[2941]: pam_unix(sudo:session): session closed for user root Aug 13 00:26:08.331866 sshd[2938]: pam_unix(sshd:session): session closed for user core Aug 13 00:26:08.334786 systemd[1]: sshd@6-147.75.53.22:22-147.75.109.163:36234.service: Deactivated successfully. Aug 13 00:26:08.336263 systemd[1]: session-9.scope: Deactivated successfully. Aug 13 00:26:08.336751 systemd-logind[2668]: Session 9 logged out. Waiting for processes to exit. Aug 13 00:26:08.337345 systemd-logind[2668]: Removed session 9. Aug 13 00:26:08.381425 systemd[1]: Started sshd@7-147.75.53.22:22-147.75.109.163:36242.service - OpenSSH per-connection server daemon (147.75.109.163:36242). Aug 13 00:26:08.657144 sshd[2946]: Accepted publickey for core from 147.75.109.163 port 36242 ssh2: RSA SHA256:u7vn2r+eDvbQsVz87adSciVbwQCBuSV2wy8SWfSsl1k Aug 13 00:26:08.658271 sshd[2946]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:26:08.660874 systemd-logind[2668]: New session 10 of user core. Aug 13 00:26:08.672149 systemd[1]: Started session-10.scope - Session 10 of User core. Aug 13 00:26:08.821496 sudo[2950]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Aug 13 00:26:08.821755 sudo[2950]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Aug 13 00:26:08.824263 sudo[2950]: pam_unix(sudo:session): session closed for user root Aug 13 00:26:08.828596 sudo[2949]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Aug 13 00:26:08.828853 sudo[2949]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Aug 13 00:26:08.849313 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Aug 13 00:26:08.850415 auditctl[2953]: No rules Aug 13 00:26:08.851228 systemd[1]: audit-rules.service: Deactivated successfully. Aug 13 00:26:08.852122 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Aug 13 00:26:08.853863 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Aug 13 00:26:08.876438 augenrules[2971]: No rules Aug 13 00:26:08.877548 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Aug 13 00:26:08.878383 sudo[2949]: pam_unix(sudo:session): session closed for user root Aug 13 00:26:08.917339 sshd[2946]: pam_unix(sshd:session): session closed for user core Aug 13 00:26:08.920114 systemd[1]: sshd@7-147.75.53.22:22-147.75.109.163:36242.service: Deactivated successfully. Aug 13 00:26:08.922563 systemd[1]: session-10.scope: Deactivated successfully. Aug 13 00:26:08.923034 systemd-logind[2668]: Session 10 logged out. Waiting for processes to exit. Aug 13 00:26:08.923572 systemd-logind[2668]: Removed session 10. Aug 13 00:26:08.970286 systemd[1]: Started sshd@8-147.75.53.22:22-147.75.109.163:36252.service - OpenSSH per-connection server daemon (147.75.109.163:36252). Aug 13 00:26:09.279636 sshd[2979]: Accepted publickey for core from 147.75.109.163 port 36252 ssh2: RSA SHA256:u7vn2r+eDvbQsVz87adSciVbwQCBuSV2wy8SWfSsl1k Aug 13 00:26:09.280651 sshd[2979]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:26:09.283376 systemd-logind[2668]: New session 11 of user core. Aug 13 00:26:09.292154 systemd[1]: Started session-11.scope - Session 11 of User core. Aug 13 00:26:09.456448 sudo[2982]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Aug 13 00:26:09.456719 sudo[2982]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Aug 13 00:26:09.732256 systemd[1]: Starting docker.service - Docker Application Container Engine... Aug 13 00:26:09.732377 (dockerd)[3011]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Aug 13 00:26:09.944400 dockerd[3011]: time="2025-08-13T00:26:09.944355600Z" level=info msg="Starting up" Aug 13 00:26:10.004230 dockerd[3011]: time="2025-08-13T00:26:10.004146960Z" level=info msg="Loading containers: start." Aug 13 00:26:10.087048 kernel: Initializing XFRM netlink socket Aug 13 00:26:10.110559 systemd-timesyncd[2591]: Network configuration changed, trying to establish connection. Aug 13 00:26:10.152988 systemd-networkd[2588]: docker0: Link UP Aug 13 00:26:10.170173 dockerd[3011]: time="2025-08-13T00:26:10.170144280Z" level=info msg="Loading containers: done." Aug 13 00:26:10.178999 dockerd[3011]: time="2025-08-13T00:26:10.178961080Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Aug 13 00:26:10.179139 dockerd[3011]: time="2025-08-13T00:26:10.179046320Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Aug 13 00:26:10.179172 dockerd[3011]: time="2025-08-13T00:26:10.179140760Z" level=info msg="Daemon has completed initialization" Aug 13 00:26:10.199804 dockerd[3011]: time="2025-08-13T00:26:10.199686960Z" level=info msg="API listen on /run/docker.sock" Aug 13 00:26:10.199802 systemd[1]: Started docker.service - Docker Application Container Engine. Aug 13 00:26:11.882961 systemd-resolved[2589]: Clock change detected. Flushing caches. Aug 13 00:26:11.883143 systemd-timesyncd[2591]: Contacted time server [2604:2dc0:202:300::2459]:123 (2.flatcar.pool.ntp.org). Aug 13 00:26:11.883197 systemd-timesyncd[2591]: Initial clock synchronization to Wed 2025-08-13 00:26:11.882912 UTC. Aug 13 00:26:12.129571 containerd[2685]: time="2025-08-13T00:26:12.129538582Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.3\"" Aug 13 00:26:12.450825 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck4237627834-merged.mount: Deactivated successfully. Aug 13 00:26:12.695002 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1986538650.mount: Deactivated successfully. Aug 13 00:26:13.394636 containerd[2685]: time="2025-08-13T00:26:13.394591142Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.33.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:26:13.394951 containerd[2685]: time="2025-08-13T00:26:13.394616302Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.33.3: active requests=0, bytes read=27352096" Aug 13 00:26:13.395643 containerd[2685]: time="2025-08-13T00:26:13.395621702Z" level=info msg="ImageCreate event name:\"sha256:c0425f3fe3fbf33c17a14d49c43d4fd0b60b2254511902d5b2c29e53ca684fc9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:26:13.398479 containerd[2685]: time="2025-08-13T00:26:13.398452302Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:125a8b488def5ea24e2de5682ab1abf063163aae4d89ce21811a45f3ecf23816\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:26:13.399636 containerd[2685]: time="2025-08-13T00:26:13.399610302Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.33.3\" with image id \"sha256:c0425f3fe3fbf33c17a14d49c43d4fd0b60b2254511902d5b2c29e53ca684fc9\", repo tag \"registry.k8s.io/kube-apiserver:v1.33.3\", repo digest \"registry.k8s.io/kube-apiserver@sha256:125a8b488def5ea24e2de5682ab1abf063163aae4d89ce21811a45f3ecf23816\", size \"27348894\" in 1.27003028s" Aug 13 00:26:13.399668 containerd[2685]: time="2025-08-13T00:26:13.399645982Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.3\" returns image reference \"sha256:c0425f3fe3fbf33c17a14d49c43d4fd0b60b2254511902d5b2c29e53ca684fc9\"" Aug 13 00:26:13.400854 containerd[2685]: time="2025-08-13T00:26:13.400831942Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.3\"" Aug 13 00:26:14.491600 containerd[2685]: time="2025-08-13T00:26:14.491564102Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.33.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:26:14.491883 containerd[2685]: time="2025-08-13T00:26:14.491617102Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.33.3: active requests=0, bytes read=23537848" Aug 13 00:26:14.492650 containerd[2685]: time="2025-08-13T00:26:14.492631342Z" level=info msg="ImageCreate event name:\"sha256:ef439b94d49d41d1b377c316fb053adb88bf6b26ec7e63aaf3deba953b7c766f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:26:14.495515 containerd[2685]: time="2025-08-13T00:26:14.495492182Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:96091626e37c5d5920ee6c3203b783cc01a08f287ec0713aeb7809bb62ccea90\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:26:14.496628 containerd[2685]: time="2025-08-13T00:26:14.496597622Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.33.3\" with image id \"sha256:ef439b94d49d41d1b377c316fb053adb88bf6b26ec7e63aaf3deba953b7c766f\", repo tag \"registry.k8s.io/kube-controller-manager:v1.33.3\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:96091626e37c5d5920ee6c3203b783cc01a08f287ec0713aeb7809bb62ccea90\", size \"25092764\" in 1.09572732s" Aug 13 00:26:14.496661 containerd[2685]: time="2025-08-13T00:26:14.496636302Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.3\" returns image reference \"sha256:ef439b94d49d41d1b377c316fb053adb88bf6b26ec7e63aaf3deba953b7c766f\"" Aug 13 00:26:14.497009 containerd[2685]: time="2025-08-13T00:26:14.496989182Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.3\"" Aug 13 00:26:15.269032 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Aug 13 00:26:15.278377 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 00:26:15.388145 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 00:26:15.391937 (kubelet)[3293]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Aug 13 00:26:15.422747 kubelet[3293]: E0813 00:26:15.422714 3293 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Aug 13 00:26:15.425184 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Aug 13 00:26:15.425396 systemd[1]: kubelet.service: Failed with result 'exit-code'. Aug 13 00:26:15.434406 containerd[2685]: time="2025-08-13T00:26:15.434374502Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.33.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:26:15.434458 containerd[2685]: time="2025-08-13T00:26:15.434431262Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.33.3: active requests=0, bytes read=18293526" Aug 13 00:26:15.435622 containerd[2685]: time="2025-08-13T00:26:15.435597982Z" level=info msg="ImageCreate event name:\"sha256:c03972dff86ba78247043f2b6171ce436ab9323da7833b18924c3d8e29ea37a5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:26:15.438434 containerd[2685]: time="2025-08-13T00:26:15.438409542Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:f3a2ffdd7483168205236f7762e9a1933f17dd733bc0188b52bddab9c0762868\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:26:15.439536 containerd[2685]: time="2025-08-13T00:26:15.439517422Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.33.3\" with image id \"sha256:c03972dff86ba78247043f2b6171ce436ab9323da7833b18924c3d8e29ea37a5\", repo tag \"registry.k8s.io/kube-scheduler:v1.33.3\", repo digest \"registry.k8s.io/kube-scheduler@sha256:f3a2ffdd7483168205236f7762e9a1933f17dd733bc0188b52bddab9c0762868\", size \"19848460\" in 942.49816ms" Aug 13 00:26:15.439564 containerd[2685]: time="2025-08-13T00:26:15.439542542Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.3\" returns image reference \"sha256:c03972dff86ba78247043f2b6171ce436ab9323da7833b18924c3d8e29ea37a5\"" Aug 13 00:26:15.439914 containerd[2685]: time="2025-08-13T00:26:15.439889422Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.3\"" Aug 13 00:26:16.321355 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2988995929.mount: Deactivated successfully. Aug 13 00:26:16.514527 containerd[2685]: time="2025-08-13T00:26:16.514489942Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.33.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:26:16.514821 containerd[2685]: time="2025-08-13T00:26:16.514572742Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.33.3: active requests=0, bytes read=28199474" Aug 13 00:26:16.515262 containerd[2685]: time="2025-08-13T00:26:16.515241062Z" level=info msg="ImageCreate event name:\"sha256:738e99dbd7325e2cdd650d83d59a79c7ecb005ab0d5bf029fc15c54ee9359306\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:26:16.517038 containerd[2685]: time="2025-08-13T00:26:16.517015102Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:c69929cfba9e38305eb1e20ca859aeb90e0d2a7326eab9bb1e8298882fe626cd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:26:16.517751 containerd[2685]: time="2025-08-13T00:26:16.517721262Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.33.3\" with image id \"sha256:738e99dbd7325e2cdd650d83d59a79c7ecb005ab0d5bf029fc15c54ee9359306\", repo tag \"registry.k8s.io/kube-proxy:v1.33.3\", repo digest \"registry.k8s.io/kube-proxy@sha256:c69929cfba9e38305eb1e20ca859aeb90e0d2a7326eab9bb1e8298882fe626cd\", size \"28198491\" in 1.07779592s" Aug 13 00:26:16.517777 containerd[2685]: time="2025-08-13T00:26:16.517759102Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.3\" returns image reference \"sha256:738e99dbd7325e2cdd650d83d59a79c7ecb005ab0d5bf029fc15c54ee9359306\"" Aug 13 00:26:16.518087 containerd[2685]: time="2025-08-13T00:26:16.518073102Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\"" Aug 13 00:26:17.047385 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3923691462.mount: Deactivated successfully. Aug 13 00:26:17.748004 containerd[2685]: time="2025-08-13T00:26:17.747958822Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.12.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:26:17.748428 containerd[2685]: time="2025-08-13T00:26:17.747965582Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.12.0: active requests=0, bytes read=19152119" Aug 13 00:26:17.749209 containerd[2685]: time="2025-08-13T00:26:17.749185262Z" level=info msg="ImageCreate event name:\"sha256:f72407be9e08c3a1b29a88318cbfee87b9f2da489f84015a5090b1e386e4dbc1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:26:17.752075 containerd[2685]: time="2025-08-13T00:26:17.752053542Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:26:17.753368 containerd[2685]: time="2025-08-13T00:26:17.753322982Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.12.0\" with image id \"sha256:f72407be9e08c3a1b29a88318cbfee87b9f2da489f84015a5090b1e386e4dbc1\", repo tag \"registry.k8s.io/coredns/coredns:v1.12.0\", repo digest \"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\", size \"19148915\" in 1.2352166s" Aug 13 00:26:17.753399 containerd[2685]: time="2025-08-13T00:26:17.753379862Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\" returns image reference \"sha256:f72407be9e08c3a1b29a88318cbfee87b9f2da489f84015a5090b1e386e4dbc1\"" Aug 13 00:26:17.753793 containerd[2685]: time="2025-08-13T00:26:17.753775262Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Aug 13 00:26:18.311177 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3641688760.mount: Deactivated successfully. Aug 13 00:26:18.311638 containerd[2685]: time="2025-08-13T00:26:18.311591982Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:26:18.311711 containerd[2685]: time="2025-08-13T00:26:18.311681142Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=268705" Aug 13 00:26:18.312444 containerd[2685]: time="2025-08-13T00:26:18.312417662Z" level=info msg="ImageCreate event name:\"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:26:18.314582 containerd[2685]: time="2025-08-13T00:26:18.314534182Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:26:18.315346 containerd[2685]: time="2025-08-13T00:26:18.315311182Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"267933\" in 561.50636ms" Aug 13 00:26:18.315549 containerd[2685]: time="2025-08-13T00:26:18.315455942Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\"" Aug 13 00:26:18.315780 containerd[2685]: time="2025-08-13T00:26:18.315753822Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.21-0\"" Aug 13 00:26:18.769799 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2035557499.mount: Deactivated successfully. Aug 13 00:26:20.320490 containerd[2685]: time="2025-08-13T00:26:20.320409182Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.21-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:26:20.320859 containerd[2685]: time="2025-08-13T00:26:20.320492022Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.21-0: active requests=0, bytes read=69334601" Aug 13 00:26:20.321718 containerd[2685]: time="2025-08-13T00:26:20.321664702Z" level=info msg="ImageCreate event name:\"sha256:31747a36ce712f0bf61b50a0c06e99768522025e7b8daedd6dc63d1ae84837b5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:26:20.324821 containerd[2685]: time="2025-08-13T00:26:20.324771542Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:d58c035df557080a27387d687092e3fc2b64c6d0e3162dc51453a115f847d121\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:26:20.326137 containerd[2685]: time="2025-08-13T00:26:20.326105462Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.21-0\" with image id \"sha256:31747a36ce712f0bf61b50a0c06e99768522025e7b8daedd6dc63d1ae84837b5\", repo tag \"registry.k8s.io/etcd:3.5.21-0\", repo digest \"registry.k8s.io/etcd@sha256:d58c035df557080a27387d687092e3fc2b64c6d0e3162dc51453a115f847d121\", size \"70026017\" in 2.01023288s" Aug 13 00:26:20.326265 containerd[2685]: time="2025-08-13T00:26:20.326246942Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.21-0\" returns image reference \"sha256:31747a36ce712f0bf61b50a0c06e99768522025e7b8daedd6dc63d1ae84837b5\"" Aug 13 00:26:24.619376 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 00:26:24.633515 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 00:26:24.650947 systemd[1]: Reloading requested from client PID 3511 ('systemctl') (unit session-11.scope)... Aug 13 00:26:24.650960 systemd[1]: Reloading... Aug 13 00:26:24.719247 zram_generator::config[3553]: No configuration found. Aug 13 00:26:24.811373 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Aug 13 00:26:24.884165 systemd[1]: Reloading finished in 232 ms. Aug 13 00:26:24.936243 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 00:26:24.939289 systemd[1]: kubelet.service: Deactivated successfully. Aug 13 00:26:24.939490 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 00:26:24.941061 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 00:26:25.047015 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 00:26:25.050859 (kubelet)[3617]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Aug 13 00:26:25.093640 kubelet[3617]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Aug 13 00:26:25.093640 kubelet[3617]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Aug 13 00:26:25.093640 kubelet[3617]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Aug 13 00:26:25.093873 kubelet[3617]: I0813 00:26:25.093686 3617 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Aug 13 00:26:27.255962 kubelet[3617]: I0813 00:26:27.255927 3617 server.go:530] "Kubelet version" kubeletVersion="v1.33.0" Aug 13 00:26:27.255962 kubelet[3617]: I0813 00:26:27.255952 3617 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Aug 13 00:26:27.256378 kubelet[3617]: I0813 00:26:27.256139 3617 server.go:956] "Client rotation is on, will bootstrap in background" Aug 13 00:26:27.275624 kubelet[3617]: E0813 00:26:27.275598 3617 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://147.75.53.22:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 147.75.53.22:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Aug 13 00:26:27.276533 kubelet[3617]: I0813 00:26:27.276511 3617 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Aug 13 00:26:27.284906 kubelet[3617]: E0813 00:26:27.284872 3617 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Aug 13 00:26:27.284958 kubelet[3617]: I0813 00:26:27.284906 3617 server.go:1423] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Aug 13 00:26:27.306376 kubelet[3617]: I0813 00:26:27.306350 3617 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Aug 13 00:26:27.307384 kubelet[3617]: I0813 00:26:27.307343 3617 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Aug 13 00:26:27.307530 kubelet[3617]: I0813 00:26:27.307387 3617 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4081.3.5-a-588adb827d","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Aug 13 00:26:27.307648 kubelet[3617]: I0813 00:26:27.307600 3617 topology_manager.go:138] "Creating topology manager with none policy" Aug 13 00:26:27.307648 kubelet[3617]: I0813 00:26:27.307609 3617 container_manager_linux.go:303] "Creating device plugin manager" Aug 13 00:26:27.307844 kubelet[3617]: I0813 00:26:27.307834 3617 state_mem.go:36] "Initialized new in-memory state store" Aug 13 00:26:27.310722 kubelet[3617]: I0813 00:26:27.310706 3617 kubelet.go:480] "Attempting to sync node with API server" Aug 13 00:26:27.310742 kubelet[3617]: I0813 00:26:27.310730 3617 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Aug 13 00:26:27.310766 kubelet[3617]: I0813 00:26:27.310757 3617 kubelet.go:386] "Adding apiserver pod source" Aug 13 00:26:27.311851 kubelet[3617]: I0813 00:26:27.311835 3617 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Aug 13 00:26:27.312836 kubelet[3617]: I0813 00:26:27.312819 3617 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Aug 13 00:26:27.313481 kubelet[3617]: E0813 00:26:27.313458 3617 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://147.75.53.22:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081.3.5-a-588adb827d&limit=500&resourceVersion=0\": dial tcp 147.75.53.22:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Aug 13 00:26:27.313506 kubelet[3617]: I0813 00:26:27.313476 3617 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Aug 13 00:26:27.313600 kubelet[3617]: W0813 00:26:27.313592 3617 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Aug 13 00:26:27.314230 kubelet[3617]: E0813 00:26:27.314213 3617 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://147.75.53.22:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 147.75.53.22:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Aug 13 00:26:27.315732 kubelet[3617]: I0813 00:26:27.315721 3617 watchdog_linux.go:99] "Systemd watchdog is not enabled" Aug 13 00:26:27.315768 kubelet[3617]: I0813 00:26:27.315763 3617 server.go:1289] "Started kubelet" Aug 13 00:26:27.315840 kubelet[3617]: I0813 00:26:27.315798 3617 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Aug 13 00:26:27.315862 kubelet[3617]: I0813 00:26:27.315824 3617 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Aug 13 00:26:27.316130 kubelet[3617]: I0813 00:26:27.316119 3617 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Aug 13 00:26:27.317369 kubelet[3617]: I0813 00:26:27.317353 3617 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Aug 13 00:26:27.317395 kubelet[3617]: I0813 00:26:27.317361 3617 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Aug 13 00:26:27.317466 kubelet[3617]: I0813 00:26:27.317457 3617 volume_manager.go:297] "Starting Kubelet Volume Manager" Aug 13 00:26:27.317492 kubelet[3617]: I0813 00:26:27.317477 3617 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Aug 13 00:26:27.317492 kubelet[3617]: E0813 00:26:27.317477 3617 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4081.3.5-a-588adb827d\" not found" Aug 13 00:26:27.317573 kubelet[3617]: I0813 00:26:27.317556 3617 reconciler.go:26] "Reconciler: start to sync state" Aug 13 00:26:27.317599 kubelet[3617]: E0813 00:26:27.317564 3617 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Aug 13 00:26:27.317673 kubelet[3617]: E0813 00:26:27.317646 3617 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://147.75.53.22:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.3.5-a-588adb827d?timeout=10s\": dial tcp 147.75.53.22:6443: connect: connection refused" interval="200ms" Aug 13 00:26:27.317781 kubelet[3617]: I0813 00:26:27.317771 3617 factory.go:223] Registration of the systemd container factory successfully Aug 13 00:26:27.317846 kubelet[3617]: E0813 00:26:27.317802 3617 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://147.75.53.22:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 147.75.53.22:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Aug 13 00:26:27.317869 kubelet[3617]: I0813 00:26:27.317854 3617 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Aug 13 00:26:27.317938 kubelet[3617]: I0813 00:26:27.317859 3617 server.go:317] "Adding debug handlers to kubelet server" Aug 13 00:26:27.321041 kubelet[3617]: I0813 00:26:27.321012 3617 factory.go:223] Registration of the containerd container factory successfully Aug 13 00:26:27.321337 kubelet[3617]: E0813 00:26:27.318344 3617 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://147.75.53.22:6443/api/v1/namespaces/default/events\": dial tcp 147.75.53.22:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4081.3.5-a-588adb827d.185b2bef78daf246 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4081.3.5-a-588adb827d,UID:ci-4081.3.5-a-588adb827d,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4081.3.5-a-588adb827d,},FirstTimestamp:2025-08-13 00:26:27.315733062 +0000 UTC m=+2.261851761,LastTimestamp:2025-08-13 00:26:27.315733062 +0000 UTC m=+2.261851761,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4081.3.5-a-588adb827d,}" Aug 13 00:26:27.334158 kubelet[3617]: I0813 00:26:27.334140 3617 cpu_manager.go:221] "Starting CPU manager" policy="none" Aug 13 00:26:27.334158 kubelet[3617]: I0813 00:26:27.334155 3617 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Aug 13 00:26:27.334215 kubelet[3617]: I0813 00:26:27.334172 3617 state_mem.go:36] "Initialized new in-memory state store" Aug 13 00:26:27.334995 kubelet[3617]: I0813 00:26:27.334983 3617 policy_none.go:49] "None policy: Start" Aug 13 00:26:27.335027 kubelet[3617]: I0813 00:26:27.335001 3617 memory_manager.go:186] "Starting memorymanager" policy="None" Aug 13 00:26:27.335027 kubelet[3617]: I0813 00:26:27.335010 3617 state_mem.go:35] "Initializing new in-memory state store" Aug 13 00:26:27.335099 kubelet[3617]: I0813 00:26:27.335074 3617 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Aug 13 00:26:27.336095 kubelet[3617]: I0813 00:26:27.336084 3617 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Aug 13 00:26:27.336117 kubelet[3617]: I0813 00:26:27.336100 3617 status_manager.go:230] "Starting to sync pod status with apiserver" Aug 13 00:26:27.336117 kubelet[3617]: I0813 00:26:27.336114 3617 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Aug 13 00:26:27.336162 kubelet[3617]: I0813 00:26:27.336122 3617 kubelet.go:2436] "Starting kubelet main sync loop" Aug 13 00:26:27.336183 kubelet[3617]: E0813 00:26:27.336156 3617 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Aug 13 00:26:27.336524 kubelet[3617]: E0813 00:26:27.336507 3617 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://147.75.53.22:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 147.75.53.22:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Aug 13 00:26:27.338403 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Aug 13 00:26:27.354444 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Aug 13 00:26:27.356980 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Aug 13 00:26:27.366893 kubelet[3617]: E0813 00:26:27.366874 3617 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Aug 13 00:26:27.367067 kubelet[3617]: I0813 00:26:27.367055 3617 eviction_manager.go:189] "Eviction manager: starting control loop" Aug 13 00:26:27.367098 kubelet[3617]: I0813 00:26:27.367068 3617 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Aug 13 00:26:27.367217 kubelet[3617]: I0813 00:26:27.367202 3617 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Aug 13 00:26:27.367722 kubelet[3617]: E0813 00:26:27.367704 3617 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Aug 13 00:26:27.367757 kubelet[3617]: E0813 00:26:27.367749 3617 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4081.3.5-a-588adb827d\" not found" Aug 13 00:26:27.443158 systemd[1]: Created slice kubepods-burstable-poda946c50864af429b16b98607f47b8809.slice - libcontainer container kubepods-burstable-poda946c50864af429b16b98607f47b8809.slice. Aug 13 00:26:27.469127 kubelet[3617]: I0813 00:26:27.469098 3617 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081.3.5-a-588adb827d" Aug 13 00:26:27.469485 kubelet[3617]: E0813 00:26:27.469462 3617 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://147.75.53.22:6443/api/v1/nodes\": dial tcp 147.75.53.22:6443: connect: connection refused" node="ci-4081.3.5-a-588adb827d" Aug 13 00:26:27.472488 kubelet[3617]: E0813 00:26:27.472466 3617 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081.3.5-a-588adb827d\" not found" node="ci-4081.3.5-a-588adb827d" Aug 13 00:26:27.474696 systemd[1]: Created slice kubepods-burstable-podfb1dc440882620436e9aa04eece6f027.slice - libcontainer container kubepods-burstable-podfb1dc440882620436e9aa04eece6f027.slice. Aug 13 00:26:27.485226 kubelet[3617]: E0813 00:26:27.485207 3617 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081.3.5-a-588adb827d\" not found" node="ci-4081.3.5-a-588adb827d" Aug 13 00:26:27.487392 systemd[1]: Created slice kubepods-burstable-pod264c31b494a704168777c30aacde43f7.slice - libcontainer container kubepods-burstable-pod264c31b494a704168777c30aacde43f7.slice. Aug 13 00:26:27.488541 kubelet[3617]: E0813 00:26:27.488525 3617 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081.3.5-a-588adb827d\" not found" node="ci-4081.3.5-a-588adb827d" Aug 13 00:26:27.518000 kubelet[3617]: E0813 00:26:27.517944 3617 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://147.75.53.22:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.3.5-a-588adb827d?timeout=10s\": dial tcp 147.75.53.22:6443: connect: connection refused" interval="400ms" Aug 13 00:26:27.519013 kubelet[3617]: I0813 00:26:27.518996 3617 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/a946c50864af429b16b98607f47b8809-ca-certs\") pod \"kube-apiserver-ci-4081.3.5-a-588adb827d\" (UID: \"a946c50864af429b16b98607f47b8809\") " pod="kube-system/kube-apiserver-ci-4081.3.5-a-588adb827d" Aug 13 00:26:27.519058 kubelet[3617]: I0813 00:26:27.519023 3617 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/a946c50864af429b16b98607f47b8809-k8s-certs\") pod \"kube-apiserver-ci-4081.3.5-a-588adb827d\" (UID: \"a946c50864af429b16b98607f47b8809\") " pod="kube-system/kube-apiserver-ci-4081.3.5-a-588adb827d" Aug 13 00:26:27.519058 kubelet[3617]: I0813 00:26:27.519045 3617 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/a946c50864af429b16b98607f47b8809-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4081.3.5-a-588adb827d\" (UID: \"a946c50864af429b16b98607f47b8809\") " pod="kube-system/kube-apiserver-ci-4081.3.5-a-588adb827d" Aug 13 00:26:27.519101 kubelet[3617]: I0813 00:26:27.519063 3617 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/fb1dc440882620436e9aa04eece6f027-flexvolume-dir\") pod \"kube-controller-manager-ci-4081.3.5-a-588adb827d\" (UID: \"fb1dc440882620436e9aa04eece6f027\") " pod="kube-system/kube-controller-manager-ci-4081.3.5-a-588adb827d" Aug 13 00:26:27.519101 kubelet[3617]: I0813 00:26:27.519081 3617 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/fb1dc440882620436e9aa04eece6f027-kubeconfig\") pod \"kube-controller-manager-ci-4081.3.5-a-588adb827d\" (UID: \"fb1dc440882620436e9aa04eece6f027\") " pod="kube-system/kube-controller-manager-ci-4081.3.5-a-588adb827d" Aug 13 00:26:27.519101 kubelet[3617]: I0813 00:26:27.519097 3617 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/fb1dc440882620436e9aa04eece6f027-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4081.3.5-a-588adb827d\" (UID: \"fb1dc440882620436e9aa04eece6f027\") " pod="kube-system/kube-controller-manager-ci-4081.3.5-a-588adb827d" Aug 13 00:26:27.519177 kubelet[3617]: I0813 00:26:27.519152 3617 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/264c31b494a704168777c30aacde43f7-kubeconfig\") pod \"kube-scheduler-ci-4081.3.5-a-588adb827d\" (UID: \"264c31b494a704168777c30aacde43f7\") " pod="kube-system/kube-scheduler-ci-4081.3.5-a-588adb827d" Aug 13 00:26:27.519212 kubelet[3617]: I0813 00:26:27.519196 3617 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/fb1dc440882620436e9aa04eece6f027-ca-certs\") pod \"kube-controller-manager-ci-4081.3.5-a-588adb827d\" (UID: \"fb1dc440882620436e9aa04eece6f027\") " pod="kube-system/kube-controller-manager-ci-4081.3.5-a-588adb827d" Aug 13 00:26:27.519259 kubelet[3617]: I0813 00:26:27.519245 3617 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/fb1dc440882620436e9aa04eece6f027-k8s-certs\") pod \"kube-controller-manager-ci-4081.3.5-a-588adb827d\" (UID: \"fb1dc440882620436e9aa04eece6f027\") " pod="kube-system/kube-controller-manager-ci-4081.3.5-a-588adb827d" Aug 13 00:26:27.670773 kubelet[3617]: I0813 00:26:27.670753 3617 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081.3.5-a-588adb827d" Aug 13 00:26:27.671030 kubelet[3617]: E0813 00:26:27.671004 3617 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://147.75.53.22:6443/api/v1/nodes\": dial tcp 147.75.53.22:6443: connect: connection refused" node="ci-4081.3.5-a-588adb827d" Aug 13 00:26:27.774065 containerd[2685]: time="2025-08-13T00:26:27.774024822Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4081.3.5-a-588adb827d,Uid:a946c50864af429b16b98607f47b8809,Namespace:kube-system,Attempt:0,}" Aug 13 00:26:27.786554 containerd[2685]: time="2025-08-13T00:26:27.786530702Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4081.3.5-a-588adb827d,Uid:fb1dc440882620436e9aa04eece6f027,Namespace:kube-system,Attempt:0,}" Aug 13 00:26:27.790049 containerd[2685]: time="2025-08-13T00:26:27.790016702Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4081.3.5-a-588adb827d,Uid:264c31b494a704168777c30aacde43f7,Namespace:kube-system,Attempt:0,}" Aug 13 00:26:27.919272 kubelet[3617]: E0813 00:26:27.919246 3617 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://147.75.53.22:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081.3.5-a-588adb827d?timeout=10s\": dial tcp 147.75.53.22:6443: connect: connection refused" interval="800ms" Aug 13 00:26:28.073045 kubelet[3617]: I0813 00:26:28.072999 3617 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081.3.5-a-588adb827d" Aug 13 00:26:28.073281 kubelet[3617]: E0813 00:26:28.073257 3617 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://147.75.53.22:6443/api/v1/nodes\": dial tcp 147.75.53.22:6443: connect: connection refused" node="ci-4081.3.5-a-588adb827d" Aug 13 00:26:28.307117 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3805000460.mount: Deactivated successfully. Aug 13 00:26:28.307719 containerd[2685]: time="2025-08-13T00:26:28.307689502Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Aug 13 00:26:28.308267 containerd[2685]: time="2025-08-13T00:26:28.308248862Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Aug 13 00:26:28.308417 containerd[2685]: time="2025-08-13T00:26:28.308390462Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=269175" Aug 13 00:26:28.308502 containerd[2685]: time="2025-08-13T00:26:28.308489502Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Aug 13 00:26:28.308537 containerd[2685]: time="2025-08-13T00:26:28.308516222Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Aug 13 00:26:28.309287 containerd[2685]: time="2025-08-13T00:26:28.309269662Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Aug 13 00:26:28.312554 containerd[2685]: time="2025-08-13T00:26:28.312531342Z" level=info msg="ImageCreate event name:\"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Aug 13 00:26:28.313344 containerd[2685]: time="2025-08-13T00:26:28.313303022Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 526.71588ms" Aug 13 00:26:28.314908 containerd[2685]: time="2025-08-13T00:26:28.314882662Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Aug 13 00:26:28.315668 containerd[2685]: time="2025-08-13T00:26:28.315647782Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 541.54516ms" Aug 13 00:26:28.316252 containerd[2685]: time="2025-08-13T00:26:28.316227262Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 526.1588ms" Aug 13 00:26:28.428115 containerd[2685]: time="2025-08-13T00:26:28.427984582Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 00:26:28.428318 containerd[2685]: time="2025-08-13T00:26:28.427950782Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 00:26:28.428351 containerd[2685]: time="2025-08-13T00:26:28.428321742Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 00:26:28.428351 containerd[2685]: time="2025-08-13T00:26:28.428335422Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:26:28.428401 containerd[2685]: time="2025-08-13T00:26:28.428377542Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 00:26:28.428421 containerd[2685]: time="2025-08-13T00:26:28.428398582Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:26:28.428449 containerd[2685]: time="2025-08-13T00:26:28.428368502Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 00:26:28.428486 containerd[2685]: time="2025-08-13T00:26:28.428467622Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 00:26:28.428507 containerd[2685]: time="2025-08-13T00:26:28.428480902Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:26:28.428938 containerd[2685]: time="2025-08-13T00:26:28.428917342Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:26:28.428938 containerd[2685]: time="2025-08-13T00:26:28.428927822Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:26:28.428994 containerd[2685]: time="2025-08-13T00:26:28.428939822Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:26:28.450358 systemd[1]: Started cri-containerd-32b6b1d85d34f5251fa3d1213ddef384b27fd8a7b14237853197dc5ba42de01a.scope - libcontainer container 32b6b1d85d34f5251fa3d1213ddef384b27fd8a7b14237853197dc5ba42de01a. Aug 13 00:26:28.451757 systemd[1]: Started cri-containerd-a079c3818acccab6cc5294374ff65f2ff5748058b65ccc2abb2c1d9628389cdc.scope - libcontainer container a079c3818acccab6cc5294374ff65f2ff5748058b65ccc2abb2c1d9628389cdc. Aug 13 00:26:28.453180 systemd[1]: Started cri-containerd-e2d5faefe68d7cf46434965481e3899f7f730d00a397098f94096149f495cd41.scope - libcontainer container e2d5faefe68d7cf46434965481e3899f7f730d00a397098f94096149f495cd41. Aug 13 00:26:28.474034 containerd[2685]: time="2025-08-13T00:26:28.474002862Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4081.3.5-a-588adb827d,Uid:fb1dc440882620436e9aa04eece6f027,Namespace:kube-system,Attempt:0,} returns sandbox id \"32b6b1d85d34f5251fa3d1213ddef384b27fd8a7b14237853197dc5ba42de01a\"" Aug 13 00:26:28.475378 containerd[2685]: time="2025-08-13T00:26:28.475350142Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4081.3.5-a-588adb827d,Uid:264c31b494a704168777c30aacde43f7,Namespace:kube-system,Attempt:0,} returns sandbox id \"a079c3818acccab6cc5294374ff65f2ff5748058b65ccc2abb2c1d9628389cdc\"" Aug 13 00:26:28.476339 containerd[2685]: time="2025-08-13T00:26:28.476316862Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4081.3.5-a-588adb827d,Uid:a946c50864af429b16b98607f47b8809,Namespace:kube-system,Attempt:0,} returns sandbox id \"e2d5faefe68d7cf46434965481e3899f7f730d00a397098f94096149f495cd41\"" Aug 13 00:26:28.478158 containerd[2685]: time="2025-08-13T00:26:28.478138422Z" level=info msg="CreateContainer within sandbox \"32b6b1d85d34f5251fa3d1213ddef384b27fd8a7b14237853197dc5ba42de01a\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Aug 13 00:26:28.478444 containerd[2685]: time="2025-08-13T00:26:28.478426462Z" level=info msg="CreateContainer within sandbox \"e2d5faefe68d7cf46434965481e3899f7f730d00a397098f94096149f495cd41\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Aug 13 00:26:28.478820 containerd[2685]: time="2025-08-13T00:26:28.478802102Z" level=info msg="CreateContainer within sandbox \"a079c3818acccab6cc5294374ff65f2ff5748058b65ccc2abb2c1d9628389cdc\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Aug 13 00:26:28.484164 containerd[2685]: time="2025-08-13T00:26:28.484137382Z" level=info msg="CreateContainer within sandbox \"32b6b1d85d34f5251fa3d1213ddef384b27fd8a7b14237853197dc5ba42de01a\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"3a8ec1e433c22f5eb260f990696a52ff9d3b2a738df1496c6c33c4930b5c6319\"" Aug 13 00:26:28.484674 containerd[2685]: time="2025-08-13T00:26:28.484655622Z" level=info msg="StartContainer for \"3a8ec1e433c22f5eb260f990696a52ff9d3b2a738df1496c6c33c4930b5c6319\"" Aug 13 00:26:28.484995 containerd[2685]: time="2025-08-13T00:26:28.484974342Z" level=info msg="CreateContainer within sandbox \"e2d5faefe68d7cf46434965481e3899f7f730d00a397098f94096149f495cd41\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"f8fb3986b6c483d56d16ed5554afec55e4327d59c93069f3ad64f9c3d3acf551\"" Aug 13 00:26:28.485220 containerd[2685]: time="2025-08-13T00:26:28.485197702Z" level=info msg="CreateContainer within sandbox \"a079c3818acccab6cc5294374ff65f2ff5748058b65ccc2abb2c1d9628389cdc\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"2c4f2b4a7074807e6934c71f2617973b179cc44aa7472e816094dbb9fcd73f1d\"" Aug 13 00:26:28.485305 containerd[2685]: time="2025-08-13T00:26:28.485285022Z" level=info msg="StartContainer for \"f8fb3986b6c483d56d16ed5554afec55e4327d59c93069f3ad64f9c3d3acf551\"" Aug 13 00:26:28.485449 containerd[2685]: time="2025-08-13T00:26:28.485432142Z" level=info msg="StartContainer for \"2c4f2b4a7074807e6934c71f2617973b179cc44aa7472e816094dbb9fcd73f1d\"" Aug 13 00:26:28.510353 systemd[1]: Started cri-containerd-2c4f2b4a7074807e6934c71f2617973b179cc44aa7472e816094dbb9fcd73f1d.scope - libcontainer container 2c4f2b4a7074807e6934c71f2617973b179cc44aa7472e816094dbb9fcd73f1d. Aug 13 00:26:28.511484 systemd[1]: Started cri-containerd-3a8ec1e433c22f5eb260f990696a52ff9d3b2a738df1496c6c33c4930b5c6319.scope - libcontainer container 3a8ec1e433c22f5eb260f990696a52ff9d3b2a738df1496c6c33c4930b5c6319. Aug 13 00:26:28.512595 systemd[1]: Started cri-containerd-f8fb3986b6c483d56d16ed5554afec55e4327d59c93069f3ad64f9c3d3acf551.scope - libcontainer container f8fb3986b6c483d56d16ed5554afec55e4327d59c93069f3ad64f9c3d3acf551. Aug 13 00:26:28.531640 kubelet[3617]: E0813 00:26:28.531611 3617 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://147.75.53.22:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 147.75.53.22:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Aug 13 00:26:28.534377 containerd[2685]: time="2025-08-13T00:26:28.534290622Z" level=info msg="StartContainer for \"2c4f2b4a7074807e6934c71f2617973b179cc44aa7472e816094dbb9fcd73f1d\" returns successfully" Aug 13 00:26:28.535376 containerd[2685]: time="2025-08-13T00:26:28.535333622Z" level=info msg="StartContainer for \"3a8ec1e433c22f5eb260f990696a52ff9d3b2a738df1496c6c33c4930b5c6319\" returns successfully" Aug 13 00:26:28.536672 containerd[2685]: time="2025-08-13T00:26:28.536644302Z" level=info msg="StartContainer for \"f8fb3986b6c483d56d16ed5554afec55e4327d59c93069f3ad64f9c3d3acf551\" returns successfully" Aug 13 00:26:28.875296 kubelet[3617]: I0813 00:26:28.875275 3617 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081.3.5-a-588adb827d" Aug 13 00:26:29.341890 kubelet[3617]: E0813 00:26:29.341867 3617 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081.3.5-a-588adb827d\" not found" node="ci-4081.3.5-a-588adb827d" Aug 13 00:26:29.342222 kubelet[3617]: E0813 00:26:29.342208 3617 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081.3.5-a-588adb827d\" not found" node="ci-4081.3.5-a-588adb827d" Aug 13 00:26:29.343825 kubelet[3617]: E0813 00:26:29.343809 3617 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081.3.5-a-588adb827d\" not found" node="ci-4081.3.5-a-588adb827d" Aug 13 00:26:29.731826 kubelet[3617]: E0813 00:26:29.731729 3617 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-4081.3.5-a-588adb827d\" not found" node="ci-4081.3.5-a-588adb827d" Aug 13 00:26:29.835057 kubelet[3617]: I0813 00:26:29.834178 3617 kubelet_node_status.go:78] "Successfully registered node" node="ci-4081.3.5-a-588adb827d" Aug 13 00:26:29.918174 kubelet[3617]: I0813 00:26:29.918146 3617 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081.3.5-a-588adb827d" Aug 13 00:26:29.922211 kubelet[3617]: E0813 00:26:29.922186 3617 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4081.3.5-a-588adb827d\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4081.3.5-a-588adb827d" Aug 13 00:26:29.922211 kubelet[3617]: I0813 00:26:29.922211 3617 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4081.3.5-a-588adb827d" Aug 13 00:26:29.923560 kubelet[3617]: E0813 00:26:29.923544 3617 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4081.3.5-a-588adb827d\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-ci-4081.3.5-a-588adb827d" Aug 13 00:26:29.923595 kubelet[3617]: I0813 00:26:29.923561 3617 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081.3.5-a-588adb827d" Aug 13 00:26:29.924747 kubelet[3617]: E0813 00:26:29.924731 3617 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4081.3.5-a-588adb827d\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-4081.3.5-a-588adb827d" Aug 13 00:26:30.314297 kubelet[3617]: I0813 00:26:30.314271 3617 apiserver.go:52] "Watching apiserver" Aug 13 00:26:30.318436 kubelet[3617]: I0813 00:26:30.318416 3617 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Aug 13 00:26:30.343718 kubelet[3617]: I0813 00:26:30.343698 3617 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081.3.5-a-588adb827d" Aug 13 00:26:30.343810 kubelet[3617]: I0813 00:26:30.343790 3617 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081.3.5-a-588adb827d" Aug 13 00:26:30.345150 kubelet[3617]: E0813 00:26:30.345134 3617 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4081.3.5-a-588adb827d\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4081.3.5-a-588adb827d" Aug 13 00:26:30.345150 kubelet[3617]: E0813 00:26:30.345144 3617 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4081.3.5-a-588adb827d\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-4081.3.5-a-588adb827d" Aug 13 00:26:31.344759 kubelet[3617]: I0813 00:26:31.344725 3617 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081.3.5-a-588adb827d" Aug 13 00:26:31.347615 kubelet[3617]: I0813 00:26:31.347600 3617 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Aug 13 00:26:31.486314 systemd[1]: Reloading requested from client PID 4044 ('systemctl') (unit session-11.scope)... Aug 13 00:26:31.486324 systemd[1]: Reloading... Aug 13 00:26:31.556256 zram_generator::config[4088]: No configuration found. Aug 13 00:26:31.646614 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Aug 13 00:26:31.730042 systemd[1]: Reloading finished in 243 ms. Aug 13 00:26:31.765153 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 00:26:31.787107 systemd[1]: kubelet.service: Deactivated successfully. Aug 13 00:26:31.787402 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 00:26:31.787454 systemd[1]: kubelet.service: Consumed 2.660s CPU time, 153.4M memory peak, 0B memory swap peak. Aug 13 00:26:31.798602 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 00:26:31.907997 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 00:26:31.911926 (kubelet)[4148]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Aug 13 00:26:31.941801 kubelet[4148]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Aug 13 00:26:31.941801 kubelet[4148]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Aug 13 00:26:31.941801 kubelet[4148]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Aug 13 00:26:31.942096 kubelet[4148]: I0813 00:26:31.941838 4148 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Aug 13 00:26:31.946675 kubelet[4148]: I0813 00:26:31.946655 4148 server.go:530] "Kubelet version" kubeletVersion="v1.33.0" Aug 13 00:26:31.946675 kubelet[4148]: I0813 00:26:31.946674 4148 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Aug 13 00:26:31.946864 kubelet[4148]: I0813 00:26:31.946856 4148 server.go:956] "Client rotation is on, will bootstrap in background" Aug 13 00:26:31.947983 kubelet[4148]: I0813 00:26:31.947972 4148 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Aug 13 00:26:31.949938 kubelet[4148]: I0813 00:26:31.949920 4148 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Aug 13 00:26:31.952116 kubelet[4148]: E0813 00:26:31.952096 4148 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Aug 13 00:26:31.952172 kubelet[4148]: I0813 00:26:31.952118 4148 server.go:1423] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Aug 13 00:26:31.970494 kubelet[4148]: I0813 00:26:31.970469 4148 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Aug 13 00:26:31.970679 kubelet[4148]: I0813 00:26:31.970653 4148 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Aug 13 00:26:31.970818 kubelet[4148]: I0813 00:26:31.970678 4148 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4081.3.5-a-588adb827d","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Aug 13 00:26:31.970888 kubelet[4148]: I0813 00:26:31.970827 4148 topology_manager.go:138] "Creating topology manager with none policy" Aug 13 00:26:31.970888 kubelet[4148]: I0813 00:26:31.970836 4148 container_manager_linux.go:303] "Creating device plugin manager" Aug 13 00:26:31.970934 kubelet[4148]: I0813 00:26:31.970894 4148 state_mem.go:36] "Initialized new in-memory state store" Aug 13 00:26:31.971199 kubelet[4148]: I0813 00:26:31.971188 4148 kubelet.go:480] "Attempting to sync node with API server" Aug 13 00:26:31.971547 kubelet[4148]: I0813 00:26:31.971204 4148 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Aug 13 00:26:31.971547 kubelet[4148]: I0813 00:26:31.971227 4148 kubelet.go:386] "Adding apiserver pod source" Aug 13 00:26:31.971547 kubelet[4148]: I0813 00:26:31.971250 4148 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Aug 13 00:26:31.972166 kubelet[4148]: I0813 00:26:31.972144 4148 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Aug 13 00:26:31.972709 kubelet[4148]: I0813 00:26:31.972694 4148 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Aug 13 00:26:31.975590 kubelet[4148]: I0813 00:26:31.975573 4148 watchdog_linux.go:99] "Systemd watchdog is not enabled" Aug 13 00:26:31.975617 kubelet[4148]: I0813 00:26:31.975612 4148 server.go:1289] "Started kubelet" Aug 13 00:26:31.975725 kubelet[4148]: I0813 00:26:31.975666 4148 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Aug 13 00:26:31.975828 kubelet[4148]: I0813 00:26:31.975663 4148 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Aug 13 00:26:31.976043 kubelet[4148]: I0813 00:26:31.976031 4148 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Aug 13 00:26:31.976601 kubelet[4148]: I0813 00:26:31.976588 4148 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Aug 13 00:26:31.976653 kubelet[4148]: I0813 00:26:31.976637 4148 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Aug 13 00:26:31.976706 kubelet[4148]: E0813 00:26:31.976688 4148 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4081.3.5-a-588adb827d\" not found" Aug 13 00:26:31.976733 kubelet[4148]: I0813 00:26:31.976716 4148 volume_manager.go:297] "Starting Kubelet Volume Manager" Aug 13 00:26:31.976755 kubelet[4148]: I0813 00:26:31.976737 4148 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Aug 13 00:26:31.976829 kubelet[4148]: I0813 00:26:31.976817 4148 reconciler.go:26] "Reconciler: start to sync state" Aug 13 00:26:31.976969 kubelet[4148]: E0813 00:26:31.976956 4148 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Aug 13 00:26:31.977152 kubelet[4148]: I0813 00:26:31.977140 4148 factory.go:223] Registration of the systemd container factory successfully Aug 13 00:26:31.977254 kubelet[4148]: I0813 00:26:31.977240 4148 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Aug 13 00:26:31.977605 kubelet[4148]: I0813 00:26:31.977591 4148 server.go:317] "Adding debug handlers to kubelet server" Aug 13 00:26:31.977892 kubelet[4148]: I0813 00:26:31.977880 4148 factory.go:223] Registration of the containerd container factory successfully Aug 13 00:26:31.986397 kubelet[4148]: I0813 00:26:31.986364 4148 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Aug 13 00:26:31.987415 kubelet[4148]: I0813 00:26:31.987403 4148 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Aug 13 00:26:31.987440 kubelet[4148]: I0813 00:26:31.987419 4148 status_manager.go:230] "Starting to sync pod status with apiserver" Aug 13 00:26:31.987440 kubelet[4148]: I0813 00:26:31.987434 4148 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Aug 13 00:26:31.987440 kubelet[4148]: I0813 00:26:31.987441 4148 kubelet.go:2436] "Starting kubelet main sync loop" Aug 13 00:26:31.987497 kubelet[4148]: E0813 00:26:31.987478 4148 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Aug 13 00:26:32.007044 kubelet[4148]: I0813 00:26:32.007025 4148 cpu_manager.go:221] "Starting CPU manager" policy="none" Aug 13 00:26:32.007071 kubelet[4148]: I0813 00:26:32.007043 4148 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Aug 13 00:26:32.007071 kubelet[4148]: I0813 00:26:32.007065 4148 state_mem.go:36] "Initialized new in-memory state store" Aug 13 00:26:32.007207 kubelet[4148]: I0813 00:26:32.007194 4148 state_mem.go:88] "Updated default CPUSet" cpuSet="" Aug 13 00:26:32.007241 kubelet[4148]: I0813 00:26:32.007205 4148 state_mem.go:96] "Updated CPUSet assignments" assignments={} Aug 13 00:26:32.007241 kubelet[4148]: I0813 00:26:32.007224 4148 policy_none.go:49] "None policy: Start" Aug 13 00:26:32.007241 kubelet[4148]: I0813 00:26:32.007240 4148 memory_manager.go:186] "Starting memorymanager" policy="None" Aug 13 00:26:32.007300 kubelet[4148]: I0813 00:26:32.007250 4148 state_mem.go:35] "Initializing new in-memory state store" Aug 13 00:26:32.007340 kubelet[4148]: I0813 00:26:32.007330 4148 state_mem.go:75] "Updated machine memory state" Aug 13 00:26:32.010154 kubelet[4148]: E0813 00:26:32.010135 4148 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Aug 13 00:26:32.010305 kubelet[4148]: I0813 00:26:32.010293 4148 eviction_manager.go:189] "Eviction manager: starting control loop" Aug 13 00:26:32.010332 kubelet[4148]: I0813 00:26:32.010305 4148 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Aug 13 00:26:32.010452 kubelet[4148]: I0813 00:26:32.010438 4148 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Aug 13 00:26:32.010922 kubelet[4148]: E0813 00:26:32.010907 4148 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Aug 13 00:26:32.088156 kubelet[4148]: I0813 00:26:32.088133 4148 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081.3.5-a-588adb827d" Aug 13 00:26:32.088247 kubelet[4148]: I0813 00:26:32.088220 4148 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081.3.5-a-588adb827d" Aug 13 00:26:32.088348 kubelet[4148]: I0813 00:26:32.088329 4148 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4081.3.5-a-588adb827d" Aug 13 00:26:32.103610 kubelet[4148]: I0813 00:26:32.103589 4148 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Aug 13 00:26:32.103775 kubelet[4148]: I0813 00:26:32.103753 4148 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Aug 13 00:26:32.104187 kubelet[4148]: I0813 00:26:32.104173 4148 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Aug 13 00:26:32.104239 kubelet[4148]: E0813 00:26:32.104215 4148 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4081.3.5-a-588adb827d\" already exists" pod="kube-system/kube-apiserver-ci-4081.3.5-a-588adb827d" Aug 13 00:26:32.113233 kubelet[4148]: I0813 00:26:32.113205 4148 kubelet_node_status.go:75] "Attempting to register node" node="ci-4081.3.5-a-588adb827d" Aug 13 00:26:32.116734 kubelet[4148]: I0813 00:26:32.116713 4148 kubelet_node_status.go:124] "Node was previously registered" node="ci-4081.3.5-a-588adb827d" Aug 13 00:26:32.116815 kubelet[4148]: I0813 00:26:32.116773 4148 kubelet_node_status.go:78] "Successfully registered node" node="ci-4081.3.5-a-588adb827d" Aug 13 00:26:32.178403 kubelet[4148]: I0813 00:26:32.178330 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/a946c50864af429b16b98607f47b8809-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4081.3.5-a-588adb827d\" (UID: \"a946c50864af429b16b98607f47b8809\") " pod="kube-system/kube-apiserver-ci-4081.3.5-a-588adb827d" Aug 13 00:26:32.178403 kubelet[4148]: I0813 00:26:32.178375 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/fb1dc440882620436e9aa04eece6f027-kubeconfig\") pod \"kube-controller-manager-ci-4081.3.5-a-588adb827d\" (UID: \"fb1dc440882620436e9aa04eece6f027\") " pod="kube-system/kube-controller-manager-ci-4081.3.5-a-588adb827d" Aug 13 00:26:32.178497 kubelet[4148]: I0813 00:26:32.178409 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/fb1dc440882620436e9aa04eece6f027-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4081.3.5-a-588adb827d\" (UID: \"fb1dc440882620436e9aa04eece6f027\") " pod="kube-system/kube-controller-manager-ci-4081.3.5-a-588adb827d" Aug 13 00:26:32.178497 kubelet[4148]: I0813 00:26:32.178439 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/fb1dc440882620436e9aa04eece6f027-ca-certs\") pod \"kube-controller-manager-ci-4081.3.5-a-588adb827d\" (UID: \"fb1dc440882620436e9aa04eece6f027\") " pod="kube-system/kube-controller-manager-ci-4081.3.5-a-588adb827d" Aug 13 00:26:32.178585 kubelet[4148]: I0813 00:26:32.178511 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/fb1dc440882620436e9aa04eece6f027-flexvolume-dir\") pod \"kube-controller-manager-ci-4081.3.5-a-588adb827d\" (UID: \"fb1dc440882620436e9aa04eece6f027\") " pod="kube-system/kube-controller-manager-ci-4081.3.5-a-588adb827d" Aug 13 00:26:32.178585 kubelet[4148]: I0813 00:26:32.178555 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/fb1dc440882620436e9aa04eece6f027-k8s-certs\") pod \"kube-controller-manager-ci-4081.3.5-a-588adb827d\" (UID: \"fb1dc440882620436e9aa04eece6f027\") " pod="kube-system/kube-controller-manager-ci-4081.3.5-a-588adb827d" Aug 13 00:26:32.178657 kubelet[4148]: I0813 00:26:32.178586 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/264c31b494a704168777c30aacde43f7-kubeconfig\") pod \"kube-scheduler-ci-4081.3.5-a-588adb827d\" (UID: \"264c31b494a704168777c30aacde43f7\") " pod="kube-system/kube-scheduler-ci-4081.3.5-a-588adb827d" Aug 13 00:26:32.178657 kubelet[4148]: I0813 00:26:32.178614 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/a946c50864af429b16b98607f47b8809-ca-certs\") pod \"kube-apiserver-ci-4081.3.5-a-588adb827d\" (UID: \"a946c50864af429b16b98607f47b8809\") " pod="kube-system/kube-apiserver-ci-4081.3.5-a-588adb827d" Aug 13 00:26:32.178657 kubelet[4148]: I0813 00:26:32.178639 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/a946c50864af429b16b98607f47b8809-k8s-certs\") pod \"kube-apiserver-ci-4081.3.5-a-588adb827d\" (UID: \"a946c50864af429b16b98607f47b8809\") " pod="kube-system/kube-apiserver-ci-4081.3.5-a-588adb827d" Aug 13 00:26:32.971827 kubelet[4148]: I0813 00:26:32.971795 4148 apiserver.go:52] "Watching apiserver" Aug 13 00:26:32.977175 kubelet[4148]: I0813 00:26:32.977158 4148 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Aug 13 00:26:32.992432 kubelet[4148]: I0813 00:26:32.992409 4148 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081.3.5-a-588adb827d" Aug 13 00:26:32.992573 kubelet[4148]: I0813 00:26:32.992553 4148 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081.3.5-a-588adb827d" Aug 13 00:26:32.992639 kubelet[4148]: I0813 00:26:32.992626 4148 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4081.3.5-a-588adb827d" Aug 13 00:26:32.995287 kubelet[4148]: I0813 00:26:32.995273 4148 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Aug 13 00:26:32.995322 kubelet[4148]: I0813 00:26:32.995285 4148 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Aug 13 00:26:32.995322 kubelet[4148]: E0813 00:26:32.995309 4148 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4081.3.5-a-588adb827d\" already exists" pod="kube-system/kube-scheduler-ci-4081.3.5-a-588adb827d" Aug 13 00:26:32.995368 kubelet[4148]: E0813 00:26:32.995331 4148 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4081.3.5-a-588adb827d\" already exists" pod="kube-system/kube-apiserver-ci-4081.3.5-a-588adb827d" Aug 13 00:26:32.995368 kubelet[4148]: I0813 00:26:32.995352 4148 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Aug 13 00:26:32.995407 kubelet[4148]: E0813 00:26:32.995388 4148 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4081.3.5-a-588adb827d\" already exists" pod="kube-system/kube-controller-manager-ci-4081.3.5-a-588adb827d" Aug 13 00:26:33.005100 kubelet[4148]: I0813 00:26:33.005055 4148 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4081.3.5-a-588adb827d" podStartSLOduration=2.005031182 podStartE2EDuration="2.005031182s" podCreationTimestamp="2025-08-13 00:26:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 00:26:33.004991542 +0000 UTC m=+1.089948681" watchObservedRunningTime="2025-08-13 00:26:33.005031182 +0000 UTC m=+1.089988321" Aug 13 00:26:33.015382 kubelet[4148]: I0813 00:26:33.015346 4148 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4081.3.5-a-588adb827d" podStartSLOduration=1.015333182 podStartE2EDuration="1.015333182s" podCreationTimestamp="2025-08-13 00:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 00:26:33.015317102 +0000 UTC m=+1.100274241" watchObservedRunningTime="2025-08-13 00:26:33.015333182 +0000 UTC m=+1.100290321" Aug 13 00:26:33.015437 kubelet[4148]: I0813 00:26:33.015403 4148 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4081.3.5-a-588adb827d" podStartSLOduration=1.015400302 podStartE2EDuration="1.015400302s" podCreationTimestamp="2025-08-13 00:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 00:26:33.009924942 +0000 UTC m=+1.094882121" watchObservedRunningTime="2025-08-13 00:26:33.015400302 +0000 UTC m=+1.100357441" Aug 13 00:26:35.587279 update_engine[2678]: I20250813 00:26:35.587194 2678 update_attempter.cc:509] Updating boot flags... Aug 13 00:26:35.619242 kernel: BTRFS warning: duplicate device /dev/nvme1n1p3 devid 1 generation 41 scanned by (udev-worker) (4301) Aug 13 00:26:35.647239 kernel: BTRFS warning: duplicate device /dev/nvme1n1p3 devid 1 generation 41 scanned by (udev-worker) (4302) Aug 13 00:26:35.671244 kernel: BTRFS warning: duplicate device /dev/nvme1n1p3 devid 1 generation 41 scanned by (udev-worker) (4302) Aug 13 00:26:38.955032 kubelet[4148]: I0813 00:26:38.954992 4148 kuberuntime_manager.go:1746] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Aug 13 00:26:38.955451 kubelet[4148]: I0813 00:26:38.955423 4148 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Aug 13 00:26:38.955482 containerd[2685]: time="2025-08-13T00:26:38.955277862Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Aug 13 00:26:39.902713 systemd[1]: Created slice kubepods-besteffort-podfb58bdfc_7fb5_43d2_b158_781027294dfa.slice - libcontainer container kubepods-besteffort-podfb58bdfc_7fb5_43d2_b158_781027294dfa.slice. Aug 13 00:26:39.937194 kubelet[4148]: I0813 00:26:39.937157 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/fb58bdfc-7fb5-43d2-b158-781027294dfa-kube-proxy\") pod \"kube-proxy-vvghc\" (UID: \"fb58bdfc-7fb5-43d2-b158-781027294dfa\") " pod="kube-system/kube-proxy-vvghc" Aug 13 00:26:39.937194 kubelet[4148]: I0813 00:26:39.937193 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/fb58bdfc-7fb5-43d2-b158-781027294dfa-lib-modules\") pod \"kube-proxy-vvghc\" (UID: \"fb58bdfc-7fb5-43d2-b158-781027294dfa\") " pod="kube-system/kube-proxy-vvghc" Aug 13 00:26:39.937346 kubelet[4148]: I0813 00:26:39.937212 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/fb58bdfc-7fb5-43d2-b158-781027294dfa-xtables-lock\") pod \"kube-proxy-vvghc\" (UID: \"fb58bdfc-7fb5-43d2-b158-781027294dfa\") " pod="kube-system/kube-proxy-vvghc" Aug 13 00:26:39.937346 kubelet[4148]: I0813 00:26:39.937232 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7txr\" (UniqueName: \"kubernetes.io/projected/fb58bdfc-7fb5-43d2-b158-781027294dfa-kube-api-access-b7txr\") pod \"kube-proxy-vvghc\" (UID: \"fb58bdfc-7fb5-43d2-b158-781027294dfa\") " pod="kube-system/kube-proxy-vvghc" Aug 13 00:26:40.044299 kubelet[4148]: E0813 00:26:40.044267 4148 projected.go:289] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found Aug 13 00:26:40.044613 kubelet[4148]: E0813 00:26:40.044299 4148 projected.go:194] Error preparing data for projected volume kube-api-access-b7txr for pod kube-system/kube-proxy-vvghc: configmap "kube-root-ca.crt" not found Aug 13 00:26:40.044613 kubelet[4148]: E0813 00:26:40.044381 4148 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/fb58bdfc-7fb5-43d2-b158-781027294dfa-kube-api-access-b7txr podName:fb58bdfc-7fb5-43d2-b158-781027294dfa nodeName:}" failed. No retries permitted until 2025-08-13 00:26:40.544352902 +0000 UTC m=+8.629310041 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-b7txr" (UniqueName: "kubernetes.io/projected/fb58bdfc-7fb5-43d2-b158-781027294dfa-kube-api-access-b7txr") pod "kube-proxy-vvghc" (UID: "fb58bdfc-7fb5-43d2-b158-781027294dfa") : configmap "kube-root-ca.crt" not found Aug 13 00:26:40.160867 systemd[1]: Created slice kubepods-besteffort-podbe57f0c9_f9e6_43c1_ac71_2fed80aa0281.slice - libcontainer container kubepods-besteffort-podbe57f0c9_f9e6_43c1_ac71_2fed80aa0281.slice. Aug 13 00:26:40.239267 kubelet[4148]: I0813 00:26:40.239235 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/be57f0c9-f9e6-43c1-ac71-2fed80aa0281-var-lib-calico\") pod \"tigera-operator-747864d56d-n95tq\" (UID: \"be57f0c9-f9e6-43c1-ac71-2fed80aa0281\") " pod="tigera-operator/tigera-operator-747864d56d-n95tq" Aug 13 00:26:40.239267 kubelet[4148]: I0813 00:26:40.239266 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c22hn\" (UniqueName: \"kubernetes.io/projected/be57f0c9-f9e6-43c1-ac71-2fed80aa0281-kube-api-access-c22hn\") pod \"tigera-operator-747864d56d-n95tq\" (UID: \"be57f0c9-f9e6-43c1-ac71-2fed80aa0281\") " pod="tigera-operator/tigera-operator-747864d56d-n95tq" Aug 13 00:26:40.463356 containerd[2685]: time="2025-08-13T00:26:40.463283742Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-747864d56d-n95tq,Uid:be57f0c9-f9e6-43c1-ac71-2fed80aa0281,Namespace:tigera-operator,Attempt:0,}" Aug 13 00:26:40.481754 containerd[2685]: time="2025-08-13T00:26:40.481687302Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 00:26:40.481862 containerd[2685]: time="2025-08-13T00:26:40.481741222Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 00:26:40.481862 containerd[2685]: time="2025-08-13T00:26:40.481752982Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:26:40.481862 containerd[2685]: time="2025-08-13T00:26:40.481830782Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:26:40.507342 systemd[1]: Started cri-containerd-dedc33c3e062eddfda1e77b17017e51df434826b2abc7eb453699a855e594202.scope - libcontainer container dedc33c3e062eddfda1e77b17017e51df434826b2abc7eb453699a855e594202. Aug 13 00:26:40.530093 containerd[2685]: time="2025-08-13T00:26:40.530027902Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-747864d56d-n95tq,Uid:be57f0c9-f9e6-43c1-ac71-2fed80aa0281,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"dedc33c3e062eddfda1e77b17017e51df434826b2abc7eb453699a855e594202\"" Aug 13 00:26:40.531265 containerd[2685]: time="2025-08-13T00:26:40.531244582Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.3\"" Aug 13 00:26:40.817676 containerd[2685]: time="2025-08-13T00:26:40.817638462Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-vvghc,Uid:fb58bdfc-7fb5-43d2-b158-781027294dfa,Namespace:kube-system,Attempt:0,}" Aug 13 00:26:40.829992 containerd[2685]: time="2025-08-13T00:26:40.829926302Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 00:26:40.830033 containerd[2685]: time="2025-08-13T00:26:40.829987822Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 00:26:40.830033 containerd[2685]: time="2025-08-13T00:26:40.829999902Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:26:40.830102 containerd[2685]: time="2025-08-13T00:26:40.830082502Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:26:40.850351 systemd[1]: Started cri-containerd-b8a26057de0ca5a1d99f2fc538232dd5b74f94f6189d8d94b330f099800b3265.scope - libcontainer container b8a26057de0ca5a1d99f2fc538232dd5b74f94f6189d8d94b330f099800b3265. Aug 13 00:26:40.865253 containerd[2685]: time="2025-08-13T00:26:40.865215462Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-vvghc,Uid:fb58bdfc-7fb5-43d2-b158-781027294dfa,Namespace:kube-system,Attempt:0,} returns sandbox id \"b8a26057de0ca5a1d99f2fc538232dd5b74f94f6189d8d94b330f099800b3265\"" Aug 13 00:26:40.867615 containerd[2685]: time="2025-08-13T00:26:40.867584742Z" level=info msg="CreateContainer within sandbox \"b8a26057de0ca5a1d99f2fc538232dd5b74f94f6189d8d94b330f099800b3265\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Aug 13 00:26:40.876318 containerd[2685]: time="2025-08-13T00:26:40.876280662Z" level=info msg="CreateContainer within sandbox \"b8a26057de0ca5a1d99f2fc538232dd5b74f94f6189d8d94b330f099800b3265\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"4b310e99493a06b3c860e6f3348371aafd4d57e8d0a500c27790c3d6c41a5a14\"" Aug 13 00:26:40.876774 containerd[2685]: time="2025-08-13T00:26:40.876750742Z" level=info msg="StartContainer for \"4b310e99493a06b3c860e6f3348371aafd4d57e8d0a500c27790c3d6c41a5a14\"" Aug 13 00:26:40.905348 systemd[1]: Started cri-containerd-4b310e99493a06b3c860e6f3348371aafd4d57e8d0a500c27790c3d6c41a5a14.scope - libcontainer container 4b310e99493a06b3c860e6f3348371aafd4d57e8d0a500c27790c3d6c41a5a14. Aug 13 00:26:40.924412 containerd[2685]: time="2025-08-13T00:26:40.924382582Z" level=info msg="StartContainer for \"4b310e99493a06b3c860e6f3348371aafd4d57e8d0a500c27790c3d6c41a5a14\" returns successfully" Aug 13 00:26:41.668770 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1482493809.mount: Deactivated successfully. Aug 13 00:26:42.087677 containerd[2685]: time="2025-08-13T00:26:42.087628622Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.38.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:26:42.087677 containerd[2685]: time="2025-08-13T00:26:42.087657862Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.38.3: active requests=0, bytes read=22150610" Aug 13 00:26:42.088474 containerd[2685]: time="2025-08-13T00:26:42.088454022Z" level=info msg="ImageCreate event name:\"sha256:7f8a5b1dba618e907d5f7804e42b3bd7cd5766bc3b0a66da25ff2c687e356bb0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:26:42.090411 containerd[2685]: time="2025-08-13T00:26:42.090392142Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:dbf1bad0def7b5955dc8e4aeee96e23ead0bc5822f6872518e685cd0ed484121\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:26:42.091173 containerd[2685]: time="2025-08-13T00:26:42.091146262Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.38.3\" with image id \"sha256:7f8a5b1dba618e907d5f7804e42b3bd7cd5766bc3b0a66da25ff2c687e356bb0\", repo tag \"quay.io/tigera/operator:v1.38.3\", repo digest \"quay.io/tigera/operator@sha256:dbf1bad0def7b5955dc8e4aeee96e23ead0bc5822f6872518e685cd0ed484121\", size \"22146605\" in 1.559873s" Aug 13 00:26:42.091245 containerd[2685]: time="2025-08-13T00:26:42.091177982Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.38.3\" returns image reference \"sha256:7f8a5b1dba618e907d5f7804e42b3bd7cd5766bc3b0a66da25ff2c687e356bb0\"" Aug 13 00:26:42.093118 containerd[2685]: time="2025-08-13T00:26:42.093097422Z" level=info msg="CreateContainer within sandbox \"dedc33c3e062eddfda1e77b17017e51df434826b2abc7eb453699a855e594202\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Aug 13 00:26:42.099751 containerd[2685]: time="2025-08-13T00:26:42.099720982Z" level=info msg="CreateContainer within sandbox \"dedc33c3e062eddfda1e77b17017e51df434826b2abc7eb453699a855e594202\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"a86888d3e41b93c13d4b5269f05b3973c907620b954b9d175e62fab2ef38be86\"" Aug 13 00:26:42.100061 containerd[2685]: time="2025-08-13T00:26:42.100043542Z" level=info msg="StartContainer for \"a86888d3e41b93c13d4b5269f05b3973c907620b954b9d175e62fab2ef38be86\"" Aug 13 00:26:42.126337 systemd[1]: Started cri-containerd-a86888d3e41b93c13d4b5269f05b3973c907620b954b9d175e62fab2ef38be86.scope - libcontainer container a86888d3e41b93c13d4b5269f05b3973c907620b954b9d175e62fab2ef38be86. Aug 13 00:26:42.142558 containerd[2685]: time="2025-08-13T00:26:42.142525422Z" level=info msg="StartContainer for \"a86888d3e41b93c13d4b5269f05b3973c907620b954b9d175e62fab2ef38be86\" returns successfully" Aug 13 00:26:43.012223 kubelet[4148]: I0813 00:26:43.012174 4148 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-vvghc" podStartSLOduration=4.012159942 podStartE2EDuration="4.012159942s" podCreationTimestamp="2025-08-13 00:26:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 00:26:41.008993062 +0000 UTC m=+9.093950201" watchObservedRunningTime="2025-08-13 00:26:43.012159942 +0000 UTC m=+11.097117081" Aug 13 00:26:45.169157 kubelet[4148]: I0813 00:26:45.169106 4148 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-747864d56d-n95tq" podStartSLOduration=3.608351717 podStartE2EDuration="5.169092637s" podCreationTimestamp="2025-08-13 00:26:40 +0000 UTC" firstStartedPulling="2025-08-13 00:26:40.530998422 +0000 UTC m=+8.615955521" lastFinishedPulling="2025-08-13 00:26:42.091739302 +0000 UTC m=+10.176696441" observedRunningTime="2025-08-13 00:26:43.012103422 +0000 UTC m=+11.097060561" watchObservedRunningTime="2025-08-13 00:26:45.169092637 +0000 UTC m=+13.254049776" Aug 13 00:26:47.049554 sudo[2982]: pam_unix(sudo:session): session closed for user root Aug 13 00:26:47.095218 sshd[2979]: pam_unix(sshd:session): session closed for user core Aug 13 00:26:47.098469 systemd[1]: sshd@8-147.75.53.22:22-147.75.109.163:36252.service: Deactivated successfully. Aug 13 00:26:47.101687 systemd[1]: session-11.scope: Deactivated successfully. Aug 13 00:26:47.101845 systemd[1]: session-11.scope: Consumed 6.573s CPU time, 177.0M memory peak, 0B memory swap peak. Aug 13 00:26:47.102279 systemd-logind[2668]: Session 11 logged out. Waiting for processes to exit. Aug 13 00:26:47.102928 systemd-logind[2668]: Removed session 11. Aug 13 00:26:51.792501 systemd[1]: Created slice kubepods-besteffort-podf9284738_32e9_433d_9dba_08c7422cc95c.slice - libcontainer container kubepods-besteffort-podf9284738_32e9_433d_9dba_08c7422cc95c.slice. Aug 13 00:26:51.816508 kubelet[4148]: I0813 00:26:51.816472 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f9284738-32e9-433d-9dba-08c7422cc95c-tigera-ca-bundle\") pod \"calico-typha-549fc4fbbb-5dnvr\" (UID: \"f9284738-32e9-433d-9dba-08c7422cc95c\") " pod="calico-system/calico-typha-549fc4fbbb-5dnvr" Aug 13 00:26:51.816836 kubelet[4148]: I0813 00:26:51.816533 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/f9284738-32e9-433d-9dba-08c7422cc95c-typha-certs\") pod \"calico-typha-549fc4fbbb-5dnvr\" (UID: \"f9284738-32e9-433d-9dba-08c7422cc95c\") " pod="calico-system/calico-typha-549fc4fbbb-5dnvr" Aug 13 00:26:51.816836 kubelet[4148]: I0813 00:26:51.816555 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4v2cr\" (UniqueName: \"kubernetes.io/projected/f9284738-32e9-433d-9dba-08c7422cc95c-kube-api-access-4v2cr\") pod \"calico-typha-549fc4fbbb-5dnvr\" (UID: \"f9284738-32e9-433d-9dba-08c7422cc95c\") " pod="calico-system/calico-typha-549fc4fbbb-5dnvr" Aug 13 00:26:52.036811 systemd[1]: Created slice kubepods-besteffort-podf39c691f_8596_40b5_ada0_166ce1261c2d.slice - libcontainer container kubepods-besteffort-podf39c691f_8596_40b5_ada0_166ce1261c2d.slice. Aug 13 00:26:52.094703 containerd[2685]: time="2025-08-13T00:26:52.094624673Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-549fc4fbbb-5dnvr,Uid:f9284738-32e9-433d-9dba-08c7422cc95c,Namespace:calico-system,Attempt:0,}" Aug 13 00:26:52.107002 containerd[2685]: time="2025-08-13T00:26:52.106942969Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 00:26:52.107002 containerd[2685]: time="2025-08-13T00:26:52.106995169Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 00:26:52.107069 containerd[2685]: time="2025-08-13T00:26:52.107006929Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:26:52.107096 containerd[2685]: time="2025-08-13T00:26:52.107080929Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:26:52.118311 kubelet[4148]: I0813 00:26:52.118284 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/f39c691f-8596-40b5-ada0-166ce1261c2d-cni-log-dir\") pod \"calico-node-rjjzv\" (UID: \"f39c691f-8596-40b5-ada0-166ce1261c2d\") " pod="calico-system/calico-node-rjjzv" Aug 13 00:26:52.118372 kubelet[4148]: I0813 00:26:52.118319 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/f39c691f-8596-40b5-ada0-166ce1261c2d-flexvol-driver-host\") pod \"calico-node-rjjzv\" (UID: \"f39c691f-8596-40b5-ada0-166ce1261c2d\") " pod="calico-system/calico-node-rjjzv" Aug 13 00:26:52.118372 kubelet[4148]: I0813 00:26:52.118338 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/f39c691f-8596-40b5-ada0-166ce1261c2d-xtables-lock\") pod \"calico-node-rjjzv\" (UID: \"f39c691f-8596-40b5-ada0-166ce1261c2d\") " pod="calico-system/calico-node-rjjzv" Aug 13 00:26:52.118372 kubelet[4148]: I0813 00:26:52.118366 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f39c691f-8596-40b5-ada0-166ce1261c2d-tigera-ca-bundle\") pod \"calico-node-rjjzv\" (UID: \"f39c691f-8596-40b5-ada0-166ce1261c2d\") " pod="calico-system/calico-node-rjjzv" Aug 13 00:26:52.118460 kubelet[4148]: I0813 00:26:52.118385 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f39c691f-8596-40b5-ada0-166ce1261c2d-lib-modules\") pod \"calico-node-rjjzv\" (UID: \"f39c691f-8596-40b5-ada0-166ce1261c2d\") " pod="calico-system/calico-node-rjjzv" Aug 13 00:26:52.118460 kubelet[4148]: I0813 00:26:52.118399 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/f39c691f-8596-40b5-ada0-166ce1261c2d-policysync\") pod \"calico-node-rjjzv\" (UID: \"f39c691f-8596-40b5-ada0-166ce1261c2d\") " pod="calico-system/calico-node-rjjzv" Aug 13 00:26:52.118460 kubelet[4148]: I0813 00:26:52.118414 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/f39c691f-8596-40b5-ada0-166ce1261c2d-var-lib-calico\") pod \"calico-node-rjjzv\" (UID: \"f39c691f-8596-40b5-ada0-166ce1261c2d\") " pod="calico-system/calico-node-rjjzv" Aug 13 00:26:52.118530 kubelet[4148]: I0813 00:26:52.118457 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/f39c691f-8596-40b5-ada0-166ce1261c2d-var-run-calico\") pod \"calico-node-rjjzv\" (UID: \"f39c691f-8596-40b5-ada0-166ce1261c2d\") " pod="calico-system/calico-node-rjjzv" Aug 13 00:26:52.118530 kubelet[4148]: I0813 00:26:52.118491 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/f39c691f-8596-40b5-ada0-166ce1261c2d-node-certs\") pod \"calico-node-rjjzv\" (UID: \"f39c691f-8596-40b5-ada0-166ce1261c2d\") " pod="calico-system/calico-node-rjjzv" Aug 13 00:26:52.118530 kubelet[4148]: I0813 00:26:52.118510 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nb9wp\" (UniqueName: \"kubernetes.io/projected/f39c691f-8596-40b5-ada0-166ce1261c2d-kube-api-access-nb9wp\") pod \"calico-node-rjjzv\" (UID: \"f39c691f-8596-40b5-ada0-166ce1261c2d\") " pod="calico-system/calico-node-rjjzv" Aug 13 00:26:52.118597 kubelet[4148]: I0813 00:26:52.118531 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/f39c691f-8596-40b5-ada0-166ce1261c2d-cni-bin-dir\") pod \"calico-node-rjjzv\" (UID: \"f39c691f-8596-40b5-ada0-166ce1261c2d\") " pod="calico-system/calico-node-rjjzv" Aug 13 00:26:52.118597 kubelet[4148]: I0813 00:26:52.118547 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/f39c691f-8596-40b5-ada0-166ce1261c2d-cni-net-dir\") pod \"calico-node-rjjzv\" (UID: \"f39c691f-8596-40b5-ada0-166ce1261c2d\") " pod="calico-system/calico-node-rjjzv" Aug 13 00:26:52.132354 systemd[1]: Started cri-containerd-c8b4cff7f90816e4a1390629f4e12719078f8314bfbcfd18f4aa8b2bbbb431b1.scope - libcontainer container c8b4cff7f90816e4a1390629f4e12719078f8314bfbcfd18f4aa8b2bbbb431b1. Aug 13 00:26:52.155538 containerd[2685]: time="2025-08-13T00:26:52.155501711Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-549fc4fbbb-5dnvr,Uid:f9284738-32e9-433d-9dba-08c7422cc95c,Namespace:calico-system,Attempt:0,} returns sandbox id \"c8b4cff7f90816e4a1390629f4e12719078f8314bfbcfd18f4aa8b2bbbb431b1\"" Aug 13 00:26:52.156471 containerd[2685]: time="2025-08-13T00:26:52.156449112Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.2\"" Aug 13 00:26:52.220280 kubelet[4148]: E0813 00:26:52.220256 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.220280 kubelet[4148]: W0813 00:26:52.220276 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.220398 kubelet[4148]: E0813 00:26:52.220293 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.220526 kubelet[4148]: E0813 00:26:52.220517 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.220526 kubelet[4148]: W0813 00:26:52.220525 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.220597 kubelet[4148]: E0813 00:26:52.220533 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.220747 kubelet[4148]: E0813 00:26:52.220739 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.220747 kubelet[4148]: W0813 00:26:52.220746 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.220800 kubelet[4148]: E0813 00:26:52.220755 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.220980 kubelet[4148]: E0813 00:26:52.220968 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.220980 kubelet[4148]: W0813 00:26:52.220976 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.221074 kubelet[4148]: E0813 00:26:52.220984 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.221288 kubelet[4148]: E0813 00:26:52.221265 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.221288 kubelet[4148]: W0813 00:26:52.221282 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.221369 kubelet[4148]: E0813 00:26:52.221295 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.221454 kubelet[4148]: E0813 00:26:52.221445 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.221454 kubelet[4148]: W0813 00:26:52.221453 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.221534 kubelet[4148]: E0813 00:26:52.221461 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.221615 kubelet[4148]: E0813 00:26:52.221599 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.221615 kubelet[4148]: W0813 00:26:52.221607 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.221615 kubelet[4148]: E0813 00:26:52.221614 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.221820 kubelet[4148]: E0813 00:26:52.221813 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.221820 kubelet[4148]: W0813 00:26:52.221820 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.221875 kubelet[4148]: E0813 00:26:52.221827 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.222093 kubelet[4148]: E0813 00:26:52.222057 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.222093 kubelet[4148]: W0813 00:26:52.222065 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.222093 kubelet[4148]: E0813 00:26:52.222073 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.222441 kubelet[4148]: E0813 00:26:52.222268 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.222441 kubelet[4148]: W0813 00:26:52.222275 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.222441 kubelet[4148]: E0813 00:26:52.222284 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.222550 kubelet[4148]: E0813 00:26:52.222483 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.222550 kubelet[4148]: W0813 00:26:52.222491 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.222550 kubelet[4148]: E0813 00:26:52.222498 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.222843 kubelet[4148]: E0813 00:26:52.222825 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.222921 kubelet[4148]: W0813 00:26:52.222854 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.222921 kubelet[4148]: E0813 00:26:52.222869 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.223120 kubelet[4148]: E0813 00:26:52.223102 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.223152 kubelet[4148]: W0813 00:26:52.223120 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.223152 kubelet[4148]: E0813 00:26:52.223135 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.223366 kubelet[4148]: E0813 00:26:52.223353 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.223366 kubelet[4148]: W0813 00:26:52.223362 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.223434 kubelet[4148]: E0813 00:26:52.223371 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.223566 kubelet[4148]: E0813 00:26:52.223557 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.223566 kubelet[4148]: W0813 00:26:52.223565 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.223648 kubelet[4148]: E0813 00:26:52.223573 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.223786 kubelet[4148]: E0813 00:26:52.223778 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.223817 kubelet[4148]: W0813 00:26:52.223786 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.223817 kubelet[4148]: E0813 00:26:52.223794 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.224003 kubelet[4148]: E0813 00:26:52.223994 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.224003 kubelet[4148]: W0813 00:26:52.224003 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.224058 kubelet[4148]: E0813 00:26:52.224010 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.224212 kubelet[4148]: E0813 00:26:52.224203 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.224212 kubelet[4148]: W0813 00:26:52.224212 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.224296 kubelet[4148]: E0813 00:26:52.224219 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.224460 kubelet[4148]: E0813 00:26:52.224451 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.224497 kubelet[4148]: W0813 00:26:52.224460 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.224497 kubelet[4148]: E0813 00:26:52.224469 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.224675 kubelet[4148]: E0813 00:26:52.224665 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.224675 kubelet[4148]: W0813 00:26:52.224674 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.224740 kubelet[4148]: E0813 00:26:52.224682 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.224914 kubelet[4148]: E0813 00:26:52.224904 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.224914 kubelet[4148]: W0813 00:26:52.224913 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.224959 kubelet[4148]: E0813 00:26:52.224921 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.225109 kubelet[4148]: E0813 00:26:52.225101 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.225135 kubelet[4148]: W0813 00:26:52.225109 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.225135 kubelet[4148]: E0813 00:26:52.225116 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.225299 kubelet[4148]: E0813 00:26:52.225290 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.225299 kubelet[4148]: W0813 00:26:52.225298 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.225356 kubelet[4148]: E0813 00:26:52.225306 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.225555 kubelet[4148]: E0813 00:26:52.225546 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.225555 kubelet[4148]: W0813 00:26:52.225554 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.225633 kubelet[4148]: E0813 00:26:52.225562 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.229470 kubelet[4148]: E0813 00:26:52.229453 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.229470 kubelet[4148]: W0813 00:26:52.229466 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.229570 kubelet[4148]: E0813 00:26:52.229478 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.279136 kubelet[4148]: E0813 00:26:52.279081 4148 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-4wkp8" podUID="bb51369a-5ac7-4ac1-b819-64d9c4e312d0" Aug 13 00:26:52.307454 kubelet[4148]: E0813 00:26:52.307433 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.307454 kubelet[4148]: W0813 00:26:52.307449 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.307568 kubelet[4148]: E0813 00:26:52.307466 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.307676 kubelet[4148]: E0813 00:26:52.307667 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.307707 kubelet[4148]: W0813 00:26:52.307674 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.307732 kubelet[4148]: E0813 00:26:52.307709 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.307866 kubelet[4148]: E0813 00:26:52.307858 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.307890 kubelet[4148]: W0813 00:26:52.307866 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.307890 kubelet[4148]: E0813 00:26:52.307873 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.308025 kubelet[4148]: E0813 00:26:52.308017 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.308048 kubelet[4148]: W0813 00:26:52.308025 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.308048 kubelet[4148]: E0813 00:26:52.308032 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.308189 kubelet[4148]: E0813 00:26:52.308181 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.308215 kubelet[4148]: W0813 00:26:52.308189 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.308215 kubelet[4148]: E0813 00:26:52.308196 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.308338 kubelet[4148]: E0813 00:26:52.308330 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.308361 kubelet[4148]: W0813 00:26:52.308337 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.308361 kubelet[4148]: E0813 00:26:52.308344 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.308476 kubelet[4148]: E0813 00:26:52.308469 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.308503 kubelet[4148]: W0813 00:26:52.308476 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.308503 kubelet[4148]: E0813 00:26:52.308483 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.308681 kubelet[4148]: E0813 00:26:52.308674 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.308704 kubelet[4148]: W0813 00:26:52.308681 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.308704 kubelet[4148]: E0813 00:26:52.308687 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.308838 kubelet[4148]: E0813 00:26:52.308830 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.308863 kubelet[4148]: W0813 00:26:52.308837 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.308863 kubelet[4148]: E0813 00:26:52.308844 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.308987 kubelet[4148]: E0813 00:26:52.308980 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.309010 kubelet[4148]: W0813 00:26:52.308987 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.309010 kubelet[4148]: E0813 00:26:52.308995 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.309151 kubelet[4148]: E0813 00:26:52.309144 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.309174 kubelet[4148]: W0813 00:26:52.309151 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.309174 kubelet[4148]: E0813 00:26:52.309158 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.309307 kubelet[4148]: E0813 00:26:52.309300 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.309307 kubelet[4148]: W0813 00:26:52.309307 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.309366 kubelet[4148]: E0813 00:26:52.309315 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.309468 kubelet[4148]: E0813 00:26:52.309460 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.309494 kubelet[4148]: W0813 00:26:52.309468 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.309494 kubelet[4148]: E0813 00:26:52.309475 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.309622 kubelet[4148]: E0813 00:26:52.309614 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.309643 kubelet[4148]: W0813 00:26:52.309622 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.309643 kubelet[4148]: E0813 00:26:52.309629 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.309766 kubelet[4148]: E0813 00:26:52.309759 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.309792 kubelet[4148]: W0813 00:26:52.309766 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.309792 kubelet[4148]: E0813 00:26:52.309772 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.309919 kubelet[4148]: E0813 00:26:52.309911 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.309944 kubelet[4148]: W0813 00:26:52.309918 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.309944 kubelet[4148]: E0813 00:26:52.309926 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.310100 kubelet[4148]: E0813 00:26:52.310092 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.310165 kubelet[4148]: W0813 00:26:52.310100 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.310165 kubelet[4148]: E0813 00:26:52.310107 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.310263 kubelet[4148]: E0813 00:26:52.310246 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.310263 kubelet[4148]: W0813 00:26:52.310254 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.310263 kubelet[4148]: E0813 00:26:52.310260 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.310400 kubelet[4148]: E0813 00:26:52.310392 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.310400 kubelet[4148]: W0813 00:26:52.310399 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.310445 kubelet[4148]: E0813 00:26:52.310406 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.310564 kubelet[4148]: E0813 00:26:52.310557 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.310590 kubelet[4148]: W0813 00:26:52.310564 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.310590 kubelet[4148]: E0813 00:26:52.310571 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.323905 kubelet[4148]: E0813 00:26:52.323871 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.323905 kubelet[4148]: W0813 00:26:52.323885 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.323905 kubelet[4148]: E0813 00:26:52.323899 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.324029 kubelet[4148]: I0813 00:26:52.323923 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bb51369a-5ac7-4ac1-b819-64d9c4e312d0-kubelet-dir\") pod \"csi-node-driver-4wkp8\" (UID: \"bb51369a-5ac7-4ac1-b819-64d9c4e312d0\") " pod="calico-system/csi-node-driver-4wkp8" Aug 13 00:26:52.324125 kubelet[4148]: E0813 00:26:52.324114 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.324152 kubelet[4148]: W0813 00:26:52.324124 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.324152 kubelet[4148]: E0813 00:26:52.324132 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.324198 kubelet[4148]: I0813 00:26:52.324150 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/bb51369a-5ac7-4ac1-b819-64d9c4e312d0-varrun\") pod \"csi-node-driver-4wkp8\" (UID: \"bb51369a-5ac7-4ac1-b819-64d9c4e312d0\") " pod="calico-system/csi-node-driver-4wkp8" Aug 13 00:26:52.324361 kubelet[4148]: E0813 00:26:52.324347 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.324396 kubelet[4148]: W0813 00:26:52.324361 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.324396 kubelet[4148]: E0813 00:26:52.324373 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.324545 kubelet[4148]: E0813 00:26:52.324535 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.324545 kubelet[4148]: W0813 00:26:52.324543 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.324596 kubelet[4148]: E0813 00:26:52.324552 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.324718 kubelet[4148]: E0813 00:26:52.324709 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.324718 kubelet[4148]: W0813 00:26:52.324717 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.324775 kubelet[4148]: E0813 00:26:52.324725 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.324775 kubelet[4148]: I0813 00:26:52.324746 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/bb51369a-5ac7-4ac1-b819-64d9c4e312d0-socket-dir\") pod \"csi-node-driver-4wkp8\" (UID: \"bb51369a-5ac7-4ac1-b819-64d9c4e312d0\") " pod="calico-system/csi-node-driver-4wkp8" Aug 13 00:26:52.324922 kubelet[4148]: E0813 00:26:52.324912 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.324951 kubelet[4148]: W0813 00:26:52.324922 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.324951 kubelet[4148]: E0813 00:26:52.324932 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.325074 kubelet[4148]: E0813 00:26:52.325066 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.325101 kubelet[4148]: W0813 00:26:52.325073 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.325101 kubelet[4148]: E0813 00:26:52.325080 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.325304 kubelet[4148]: E0813 00:26:52.325290 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.325304 kubelet[4148]: W0813 00:26:52.325298 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.325357 kubelet[4148]: E0813 00:26:52.325307 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.325357 kubelet[4148]: I0813 00:26:52.325325 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/bb51369a-5ac7-4ac1-b819-64d9c4e312d0-registration-dir\") pod \"csi-node-driver-4wkp8\" (UID: \"bb51369a-5ac7-4ac1-b819-64d9c4e312d0\") " pod="calico-system/csi-node-driver-4wkp8" Aug 13 00:26:52.325499 kubelet[4148]: E0813 00:26:52.325487 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.325527 kubelet[4148]: W0813 00:26:52.325498 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.325527 kubelet[4148]: E0813 00:26:52.325509 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.325655 kubelet[4148]: E0813 00:26:52.325647 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.325681 kubelet[4148]: W0813 00:26:52.325655 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.325681 kubelet[4148]: E0813 00:26:52.325662 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.325819 kubelet[4148]: E0813 00:26:52.325811 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.325819 kubelet[4148]: W0813 00:26:52.325818 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.325871 kubelet[4148]: E0813 00:26:52.325826 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.325871 kubelet[4148]: I0813 00:26:52.325849 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6crr\" (UniqueName: \"kubernetes.io/projected/bb51369a-5ac7-4ac1-b819-64d9c4e312d0-kube-api-access-g6crr\") pod \"csi-node-driver-4wkp8\" (UID: \"bb51369a-5ac7-4ac1-b819-64d9c4e312d0\") " pod="calico-system/csi-node-driver-4wkp8" Aug 13 00:26:52.326069 kubelet[4148]: E0813 00:26:52.326051 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.326099 kubelet[4148]: W0813 00:26:52.326070 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.326099 kubelet[4148]: E0813 00:26:52.326083 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.326242 kubelet[4148]: E0813 00:26:52.326219 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.326242 kubelet[4148]: W0813 00:26:52.326227 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.326242 kubelet[4148]: E0813 00:26:52.326240 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.326420 kubelet[4148]: E0813 00:26:52.326410 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.326455 kubelet[4148]: W0813 00:26:52.326420 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.326455 kubelet[4148]: E0813 00:26:52.326429 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.326581 kubelet[4148]: E0813 00:26:52.326570 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.326581 kubelet[4148]: W0813 00:26:52.326577 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.326627 kubelet[4148]: E0813 00:26:52.326586 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.339117 containerd[2685]: time="2025-08-13T00:26:52.339086384Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-rjjzv,Uid:f39c691f-8596-40b5-ada0-166ce1261c2d,Namespace:calico-system,Attempt:0,}" Aug 13 00:26:52.351624 containerd[2685]: time="2025-08-13T00:26:52.351524040Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 00:26:52.351624 containerd[2685]: time="2025-08-13T00:26:52.351580240Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 00:26:52.351624 containerd[2685]: time="2025-08-13T00:26:52.351592880Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:26:52.352072 containerd[2685]: time="2025-08-13T00:26:52.352046840Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:26:52.379409 systemd[1]: Started cri-containerd-49068c488f9a318d05f6e0e80d5c70f277ccc8bbe32e3e6b64b9cf6d495d3342.scope - libcontainer container 49068c488f9a318d05f6e0e80d5c70f277ccc8bbe32e3e6b64b9cf6d495d3342. Aug 13 00:26:52.395009 containerd[2685]: time="2025-08-13T00:26:52.394969335Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-rjjzv,Uid:f39c691f-8596-40b5-ada0-166ce1261c2d,Namespace:calico-system,Attempt:0,} returns sandbox id \"49068c488f9a318d05f6e0e80d5c70f277ccc8bbe32e3e6b64b9cf6d495d3342\"" Aug 13 00:26:52.426388 kubelet[4148]: E0813 00:26:52.426358 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.426388 kubelet[4148]: W0813 00:26:52.426377 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.426492 kubelet[4148]: E0813 00:26:52.426396 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.426623 kubelet[4148]: E0813 00:26:52.426613 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.426623 kubelet[4148]: W0813 00:26:52.426622 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.426679 kubelet[4148]: E0813 00:26:52.426630 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.426881 kubelet[4148]: E0813 00:26:52.426865 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.426902 kubelet[4148]: W0813 00:26:52.426881 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.426902 kubelet[4148]: E0813 00:26:52.426896 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.427086 kubelet[4148]: E0813 00:26:52.427078 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.427109 kubelet[4148]: W0813 00:26:52.427086 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.427109 kubelet[4148]: E0813 00:26:52.427094 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.427285 kubelet[4148]: E0813 00:26:52.427277 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.427307 kubelet[4148]: W0813 00:26:52.427284 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.427307 kubelet[4148]: E0813 00:26:52.427292 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.427513 kubelet[4148]: E0813 00:26:52.427504 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.427513 kubelet[4148]: W0813 00:26:52.427513 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.427566 kubelet[4148]: E0813 00:26:52.427521 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.427755 kubelet[4148]: E0813 00:26:52.427746 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.427781 kubelet[4148]: W0813 00:26:52.427757 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.427781 kubelet[4148]: E0813 00:26:52.427765 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.427936 kubelet[4148]: E0813 00:26:52.427927 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.427960 kubelet[4148]: W0813 00:26:52.427936 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.427960 kubelet[4148]: E0813 00:26:52.427943 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.428147 kubelet[4148]: E0813 00:26:52.428138 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.428147 kubelet[4148]: W0813 00:26:52.428146 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.428197 kubelet[4148]: E0813 00:26:52.428153 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.428377 kubelet[4148]: E0813 00:26:52.428367 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.428377 kubelet[4148]: W0813 00:26:52.428376 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.428429 kubelet[4148]: E0813 00:26:52.428384 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.428597 kubelet[4148]: E0813 00:26:52.428588 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.428597 kubelet[4148]: W0813 00:26:52.428596 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.428640 kubelet[4148]: E0813 00:26:52.428604 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.428794 kubelet[4148]: E0813 00:26:52.428786 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.428819 kubelet[4148]: W0813 00:26:52.428794 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.428819 kubelet[4148]: E0813 00:26:52.428801 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.429131 kubelet[4148]: E0813 00:26:52.429118 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.429153 kubelet[4148]: W0813 00:26:52.429132 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.429153 kubelet[4148]: E0813 00:26:52.429145 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.429394 kubelet[4148]: E0813 00:26:52.429386 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.429415 kubelet[4148]: W0813 00:26:52.429394 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.429415 kubelet[4148]: E0813 00:26:52.429401 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.429553 kubelet[4148]: E0813 00:26:52.429546 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.429574 kubelet[4148]: W0813 00:26:52.429553 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.429574 kubelet[4148]: E0813 00:26:52.429560 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.429748 kubelet[4148]: E0813 00:26:52.429738 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.429748 kubelet[4148]: W0813 00:26:52.429748 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.429803 kubelet[4148]: E0813 00:26:52.429756 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.429991 kubelet[4148]: E0813 00:26:52.429981 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.430021 kubelet[4148]: W0813 00:26:52.429992 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.430021 kubelet[4148]: E0813 00:26:52.430001 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.430180 kubelet[4148]: E0813 00:26:52.430171 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.430180 kubelet[4148]: W0813 00:26:52.430180 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.430243 kubelet[4148]: E0813 00:26:52.430187 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.430441 kubelet[4148]: E0813 00:26:52.430432 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.430441 kubelet[4148]: W0813 00:26:52.430440 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.430504 kubelet[4148]: E0813 00:26:52.430448 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.430736 kubelet[4148]: E0813 00:26:52.430721 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.430758 kubelet[4148]: W0813 00:26:52.430736 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.430758 kubelet[4148]: E0813 00:26:52.430749 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.431045 kubelet[4148]: E0813 00:26:52.431036 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.431069 kubelet[4148]: W0813 00:26:52.431045 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.431069 kubelet[4148]: E0813 00:26:52.431053 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.431307 kubelet[4148]: E0813 00:26:52.431299 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.431335 kubelet[4148]: W0813 00:26:52.431309 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.431335 kubelet[4148]: E0813 00:26:52.431317 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.431577 kubelet[4148]: E0813 00:26:52.431568 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.431600 kubelet[4148]: W0813 00:26:52.431576 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.431600 kubelet[4148]: E0813 00:26:52.431584 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.431824 kubelet[4148]: E0813 00:26:52.431816 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.431846 kubelet[4148]: W0813 00:26:52.431825 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.431846 kubelet[4148]: E0813 00:26:52.431833 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.432068 kubelet[4148]: E0813 00:26:52.432060 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.432089 kubelet[4148]: W0813 00:26:52.432068 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.432089 kubelet[4148]: E0813 00:26:52.432075 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:52.440373 kubelet[4148]: E0813 00:26:52.440352 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:52.440373 kubelet[4148]: W0813 00:26:52.440366 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:52.440487 kubelet[4148]: E0813 00:26:52.440378 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:53.206259 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3611799663.mount: Deactivated successfully. Aug 13 00:26:53.607098 containerd[2685]: time="2025-08-13T00:26:53.607051267Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:26:53.607421 containerd[2685]: time="2025-08-13T00:26:53.607062747Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.30.2: active requests=0, bytes read=33087207" Aug 13 00:26:53.607835 containerd[2685]: time="2025-08-13T00:26:53.607814708Z" level=info msg="ImageCreate event name:\"sha256:bd819526ff844d29b60cd75e846a1f55306016ff269d881d52a9b6c7b2eef0b2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:26:53.609528 containerd[2685]: time="2025-08-13T00:26:53.609506990Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:da29d745efe5eb7d25f765d3aa439f3fe60710a458efe39c285e58b02bd961af\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:26:53.610179 containerd[2685]: time="2025-08-13T00:26:53.610156111Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.30.2\" with image id \"sha256:bd819526ff844d29b60cd75e846a1f55306016ff269d881d52a9b6c7b2eef0b2\", repo tag \"ghcr.io/flatcar/calico/typha:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:da29d745efe5eb7d25f765d3aa439f3fe60710a458efe39c285e58b02bd961af\", size \"33087061\" in 1.453677159s" Aug 13 00:26:53.610202 containerd[2685]: time="2025-08-13T00:26:53.610186511Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.30.2\" returns image reference \"sha256:bd819526ff844d29b60cd75e846a1f55306016ff269d881d52a9b6c7b2eef0b2\"" Aug 13 00:26:53.610842 containerd[2685]: time="2025-08-13T00:26:53.610818352Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\"" Aug 13 00:26:53.616052 containerd[2685]: time="2025-08-13T00:26:53.615975158Z" level=info msg="CreateContainer within sandbox \"c8b4cff7f90816e4a1390629f4e12719078f8314bfbcfd18f4aa8b2bbbb431b1\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Aug 13 00:26:53.621086 containerd[2685]: time="2025-08-13T00:26:53.621056124Z" level=info msg="CreateContainer within sandbox \"c8b4cff7f90816e4a1390629f4e12719078f8314bfbcfd18f4aa8b2bbbb431b1\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"e8562dde9084b61ba499be80856b937575af79ed0d55fa0146d4357ba6bb0d5a\"" Aug 13 00:26:53.621438 containerd[2685]: time="2025-08-13T00:26:53.621419804Z" level=info msg="StartContainer for \"e8562dde9084b61ba499be80856b937575af79ed0d55fa0146d4357ba6bb0d5a\"" Aug 13 00:26:53.652404 systemd[1]: Started cri-containerd-e8562dde9084b61ba499be80856b937575af79ed0d55fa0146d4357ba6bb0d5a.scope - libcontainer container e8562dde9084b61ba499be80856b937575af79ed0d55fa0146d4357ba6bb0d5a. Aug 13 00:26:53.676970 containerd[2685]: time="2025-08-13T00:26:53.676933791Z" level=info msg="StartContainer for \"e8562dde9084b61ba499be80856b937575af79ed0d55fa0146d4357ba6bb0d5a\" returns successfully" Aug 13 00:26:53.989221 kubelet[4148]: E0813 00:26:53.988206 4148 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-4wkp8" podUID="bb51369a-5ac7-4ac1-b819-64d9c4e312d0" Aug 13 00:26:54.029279 kubelet[4148]: I0813 00:26:54.029223 4148 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-549fc4fbbb-5dnvr" podStartSLOduration=1.574724848 podStartE2EDuration="3.029208768s" podCreationTimestamp="2025-08-13 00:26:51 +0000 UTC" firstStartedPulling="2025-08-13 00:26:52.156240592 +0000 UTC m=+20.241197731" lastFinishedPulling="2025-08-13 00:26:53.610724472 +0000 UTC m=+21.695681651" observedRunningTime="2025-08-13 00:26:54.029108688 +0000 UTC m=+22.114065867" watchObservedRunningTime="2025-08-13 00:26:54.029208768 +0000 UTC m=+22.114165907" Aug 13 00:26:54.122537 kubelet[4148]: E0813 00:26:54.122440 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:54.122537 kubelet[4148]: W0813 00:26:54.122460 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:54.122537 kubelet[4148]: E0813 00:26:54.122476 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:54.122753 kubelet[4148]: E0813 00:26:54.122744 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:54.122792 kubelet[4148]: W0813 00:26:54.122752 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:54.122818 kubelet[4148]: E0813 00:26:54.122794 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:54.123064 kubelet[4148]: E0813 00:26:54.123055 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:54.123091 kubelet[4148]: W0813 00:26:54.123064 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:54.123091 kubelet[4148]: E0813 00:26:54.123072 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:54.123344 kubelet[4148]: E0813 00:26:54.123336 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:54.123344 kubelet[4148]: W0813 00:26:54.123344 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:54.123391 kubelet[4148]: E0813 00:26:54.123351 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:54.123509 kubelet[4148]: E0813 00:26:54.123500 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:54.123536 kubelet[4148]: W0813 00:26:54.123509 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:54.123536 kubelet[4148]: E0813 00:26:54.123517 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:54.123774 kubelet[4148]: E0813 00:26:54.123766 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:54.123800 kubelet[4148]: W0813 00:26:54.123776 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:54.123800 kubelet[4148]: E0813 00:26:54.123784 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:54.124036 kubelet[4148]: E0813 00:26:54.124028 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:54.124062 kubelet[4148]: W0813 00:26:54.124035 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:54.124062 kubelet[4148]: E0813 00:26:54.124043 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:54.124256 kubelet[4148]: E0813 00:26:54.124248 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:54.124256 kubelet[4148]: W0813 00:26:54.124255 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:54.124310 kubelet[4148]: E0813 00:26:54.124263 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:54.124483 kubelet[4148]: E0813 00:26:54.124474 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:54.124483 kubelet[4148]: W0813 00:26:54.124482 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:54.124536 kubelet[4148]: E0813 00:26:54.124490 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:54.124751 kubelet[4148]: E0813 00:26:54.124743 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:54.124751 kubelet[4148]: W0813 00:26:54.124751 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:54.124801 kubelet[4148]: E0813 00:26:54.124758 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:54.125020 kubelet[4148]: E0813 00:26:54.125011 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:54.125020 kubelet[4148]: W0813 00:26:54.125020 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:54.125071 kubelet[4148]: E0813 00:26:54.125027 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:54.125296 kubelet[4148]: E0813 00:26:54.125287 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:54.125296 kubelet[4148]: W0813 00:26:54.125296 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:54.125343 kubelet[4148]: E0813 00:26:54.125303 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:54.125535 kubelet[4148]: E0813 00:26:54.125526 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:54.125535 kubelet[4148]: W0813 00:26:54.125534 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:54.125587 kubelet[4148]: E0813 00:26:54.125542 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:54.125748 kubelet[4148]: E0813 00:26:54.125740 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:54.125748 kubelet[4148]: W0813 00:26:54.125747 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:54.125795 kubelet[4148]: E0813 00:26:54.125754 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:54.126019 kubelet[4148]: E0813 00:26:54.126011 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:54.126019 kubelet[4148]: W0813 00:26:54.126019 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:54.126063 kubelet[4148]: E0813 00:26:54.126026 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:54.142431 kubelet[4148]: E0813 00:26:54.142348 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:54.142431 kubelet[4148]: W0813 00:26:54.142364 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:54.142431 kubelet[4148]: E0813 00:26:54.142377 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:54.142683 kubelet[4148]: E0813 00:26:54.142668 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:54.142683 kubelet[4148]: W0813 00:26:54.142678 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:54.142683 kubelet[4148]: E0813 00:26:54.142685 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:54.142984 kubelet[4148]: E0813 00:26:54.142974 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:54.142984 kubelet[4148]: W0813 00:26:54.142982 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:54.143038 kubelet[4148]: E0813 00:26:54.142992 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:54.143254 kubelet[4148]: E0813 00:26:54.143244 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:54.143254 kubelet[4148]: W0813 00:26:54.143252 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:54.143303 kubelet[4148]: E0813 00:26:54.143262 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:54.143423 kubelet[4148]: E0813 00:26:54.143414 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:54.143423 kubelet[4148]: W0813 00:26:54.143422 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:54.143479 kubelet[4148]: E0813 00:26:54.143430 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:54.143585 kubelet[4148]: E0813 00:26:54.143577 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:54.143585 kubelet[4148]: W0813 00:26:54.143585 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:54.143640 kubelet[4148]: E0813 00:26:54.143593 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:54.143753 kubelet[4148]: E0813 00:26:54.143744 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:54.143753 kubelet[4148]: W0813 00:26:54.143752 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:54.143803 kubelet[4148]: E0813 00:26:54.143759 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:54.144028 kubelet[4148]: E0813 00:26:54.144019 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:54.144055 kubelet[4148]: W0813 00:26:54.144027 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:54.144055 kubelet[4148]: E0813 00:26:54.144035 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:54.144256 kubelet[4148]: E0813 00:26:54.144247 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:54.144256 kubelet[4148]: W0813 00:26:54.144254 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:54.144313 kubelet[4148]: E0813 00:26:54.144261 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:54.144412 kubelet[4148]: E0813 00:26:54.144403 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:54.144412 kubelet[4148]: W0813 00:26:54.144411 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:54.144464 kubelet[4148]: E0813 00:26:54.144418 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:54.144648 kubelet[4148]: E0813 00:26:54.144640 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:54.144648 kubelet[4148]: W0813 00:26:54.144648 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:54.144703 kubelet[4148]: E0813 00:26:54.144655 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:54.145227 kubelet[4148]: E0813 00:26:54.144986 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:54.145227 kubelet[4148]: W0813 00:26:54.145004 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:54.145227 kubelet[4148]: E0813 00:26:54.145018 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:54.145227 kubelet[4148]: E0813 00:26:54.145245 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:54.145359 kubelet[4148]: W0813 00:26:54.145254 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:54.145359 kubelet[4148]: E0813 00:26:54.145263 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:54.145418 kubelet[4148]: E0813 00:26:54.145405 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:54.145418 kubelet[4148]: W0813 00:26:54.145412 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:54.145460 kubelet[4148]: E0813 00:26:54.145418 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:54.145650 kubelet[4148]: E0813 00:26:54.145641 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:54.145650 kubelet[4148]: W0813 00:26:54.145649 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:54.145708 kubelet[4148]: E0813 00:26:54.145656 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:54.146057 kubelet[4148]: E0813 00:26:54.145932 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:54.146057 kubelet[4148]: W0813 00:26:54.145946 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:54.146057 kubelet[4148]: E0813 00:26:54.145959 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:54.146211 kubelet[4148]: E0813 00:26:54.146200 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:54.146297 kubelet[4148]: W0813 00:26:54.146272 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:54.146345 kubelet[4148]: E0813 00:26:54.146306 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:54.146614 kubelet[4148]: E0813 00:26:54.146604 4148 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Aug 13 00:26:54.146635 kubelet[4148]: W0813 00:26:54.146614 4148 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Aug 13 00:26:54.146635 kubelet[4148]: E0813 00:26:54.146623 4148 plugins.go:703] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Aug 13 00:26:54.560552 containerd[2685]: time="2025-08-13T00:26:54.560510682Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:26:54.560682 containerd[2685]: time="2025-08-13T00:26:54.560523482Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2: active requests=0, bytes read=4266981" Aug 13 00:26:54.561234 containerd[2685]: time="2025-08-13T00:26:54.561212363Z" level=info msg="ImageCreate event name:\"sha256:53f638101e3d73f7dd5e42dc42fb3d94ae1978e8958677222c3de6ec1d8c3d4f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:26:54.563024 containerd[2685]: time="2025-08-13T00:26:54.563001605Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:972be127eaecd7d1a2d5393b8d14f1ae8f88550bee83e0519e9590c7e15eb41b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:26:54.563714 containerd[2685]: time="2025-08-13T00:26:54.563688685Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\" with image id \"sha256:53f638101e3d73f7dd5e42dc42fb3d94ae1978e8958677222c3de6ec1d8c3d4f\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:972be127eaecd7d1a2d5393b8d14f1ae8f88550bee83e0519e9590c7e15eb41b\", size \"5636182\" in 952.838573ms" Aug 13 00:26:54.563737 containerd[2685]: time="2025-08-13T00:26:54.563721245Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.30.2\" returns image reference \"sha256:53f638101e3d73f7dd5e42dc42fb3d94ae1978e8958677222c3de6ec1d8c3d4f\"" Aug 13 00:26:54.565733 containerd[2685]: time="2025-08-13T00:26:54.565713448Z" level=info msg="CreateContainer within sandbox \"49068c488f9a318d05f6e0e80d5c70f277ccc8bbe32e3e6b64b9cf6d495d3342\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Aug 13 00:26:54.572723 containerd[2685]: time="2025-08-13T00:26:54.572693655Z" level=info msg="CreateContainer within sandbox \"49068c488f9a318d05f6e0e80d5c70f277ccc8bbe32e3e6b64b9cf6d495d3342\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"a55e1df55be36c51d0f8a3c806066b4d87be6dc879ca71c604ea6c34a67f7ca3\"" Aug 13 00:26:54.573113 containerd[2685]: time="2025-08-13T00:26:54.573086656Z" level=info msg="StartContainer for \"a55e1df55be36c51d0f8a3c806066b4d87be6dc879ca71c604ea6c34a67f7ca3\"" Aug 13 00:26:54.606429 systemd[1]: Started cri-containerd-a55e1df55be36c51d0f8a3c806066b4d87be6dc879ca71c604ea6c34a67f7ca3.scope - libcontainer container a55e1df55be36c51d0f8a3c806066b4d87be6dc879ca71c604ea6c34a67f7ca3. Aug 13 00:26:54.624761 containerd[2685]: time="2025-08-13T00:26:54.624731753Z" level=info msg="StartContainer for \"a55e1df55be36c51d0f8a3c806066b4d87be6dc879ca71c604ea6c34a67f7ca3\" returns successfully" Aug 13 00:26:54.636590 systemd[1]: cri-containerd-a55e1df55be36c51d0f8a3c806066b4d87be6dc879ca71c604ea6c34a67f7ca3.scope: Deactivated successfully. Aug 13 00:26:54.791305 containerd[2685]: time="2025-08-13T00:26:54.791251379Z" level=info msg="shim disconnected" id=a55e1df55be36c51d0f8a3c806066b4d87be6dc879ca71c604ea6c34a67f7ca3 namespace=k8s.io Aug 13 00:26:54.791305 containerd[2685]: time="2025-08-13T00:26:54.791304340Z" level=warning msg="cleaning up after shim disconnected" id=a55e1df55be36c51d0f8a3c806066b4d87be6dc879ca71c604ea6c34a67f7ca3 namespace=k8s.io Aug 13 00:26:54.791444 containerd[2685]: time="2025-08-13T00:26:54.791312660Z" level=info msg="cleaning up dead shim" namespace=k8s.io Aug 13 00:26:54.921266 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a55e1df55be36c51d0f8a3c806066b4d87be6dc879ca71c604ea6c34a67f7ca3-rootfs.mount: Deactivated successfully. Aug 13 00:26:55.024474 kubelet[4148]: I0813 00:26:55.024445 4148 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Aug 13 00:26:55.025351 containerd[2685]: time="2025-08-13T00:26:55.025328199Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.2\"" Aug 13 00:26:55.988041 kubelet[4148]: E0813 00:26:55.988006 4148 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-4wkp8" podUID="bb51369a-5ac7-4ac1-b819-64d9c4e312d0" Aug 13 00:26:56.391039 kubelet[4148]: I0813 00:26:56.391009 4148 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Aug 13 00:26:57.059520 containerd[2685]: time="2025-08-13T00:26:57.059477857Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:26:57.059912 containerd[2685]: time="2025-08-13T00:26:57.059502657Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.30.2: active requests=0, bytes read=65888320" Aug 13 00:26:57.060198 containerd[2685]: time="2025-08-13T00:26:57.060177217Z" level=info msg="ImageCreate event name:\"sha256:f6e344d58b3c5524e767c7d1dd4cb29c85ce820b0f3005a603532b6a22db5588\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:26:57.062051 containerd[2685]: time="2025-08-13T00:26:57.062026659Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:50686775cc60acb78bd92a66fa2d84e1700b2d8e43a718fbadbf35e59baefb4d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:26:57.062831 containerd[2685]: time="2025-08-13T00:26:57.062807300Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.30.2\" with image id \"sha256:f6e344d58b3c5524e767c7d1dd4cb29c85ce820b0f3005a603532b6a22db5588\", repo tag \"ghcr.io/flatcar/calico/cni:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:50686775cc60acb78bd92a66fa2d84e1700b2d8e43a718fbadbf35e59baefb4d\", size \"67257561\" in 2.037445781s" Aug 13 00:26:57.062858 containerd[2685]: time="2025-08-13T00:26:57.062837900Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.30.2\" returns image reference \"sha256:f6e344d58b3c5524e767c7d1dd4cb29c85ce820b0f3005a603532b6a22db5588\"" Aug 13 00:26:57.076207 containerd[2685]: time="2025-08-13T00:26:57.076178992Z" level=info msg="CreateContainer within sandbox \"49068c488f9a318d05f6e0e80d5c70f277ccc8bbe32e3e6b64b9cf6d495d3342\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Aug 13 00:26:57.082182 containerd[2685]: time="2025-08-13T00:26:57.082150437Z" level=info msg="CreateContainer within sandbox \"49068c488f9a318d05f6e0e80d5c70f277ccc8bbe32e3e6b64b9cf6d495d3342\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"a0d12baf896e694d49ea3639f57c44ef8f625f1451d9d138874eacfb5898082a\"" Aug 13 00:26:57.082520 containerd[2685]: time="2025-08-13T00:26:57.082500278Z" level=info msg="StartContainer for \"a0d12baf896e694d49ea3639f57c44ef8f625f1451d9d138874eacfb5898082a\"" Aug 13 00:26:57.112356 systemd[1]: Started cri-containerd-a0d12baf896e694d49ea3639f57c44ef8f625f1451d9d138874eacfb5898082a.scope - libcontainer container a0d12baf896e694d49ea3639f57c44ef8f625f1451d9d138874eacfb5898082a. Aug 13 00:26:57.131267 containerd[2685]: time="2025-08-13T00:26:57.131226683Z" level=info msg="StartContainer for \"a0d12baf896e694d49ea3639f57c44ef8f625f1451d9d138874eacfb5898082a\" returns successfully" Aug 13 00:26:57.505609 containerd[2685]: time="2025-08-13T00:26:57.505531187Z" level=error msg="failed to reload cni configuration after receiving fs change event(WRITE \"/etc/cni/net.d/calico-kubeconfig\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Aug 13 00:26:57.507202 systemd[1]: cri-containerd-a0d12baf896e694d49ea3639f57c44ef8f625f1451d9d138874eacfb5898082a.scope: Deactivated successfully. Aug 13 00:26:57.558690 kubelet[4148]: I0813 00:26:57.558662 4148 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Aug 13 00:26:57.628541 systemd[1]: Created slice kubepods-burstable-pod75d28fd6_77d6_4c58_8897_8ce0729a65ea.slice - libcontainer container kubepods-burstable-pod75d28fd6_77d6_4c58_8897_8ce0729a65ea.slice. Aug 13 00:26:57.639665 containerd[2685]: time="2025-08-13T00:26:57.639610591Z" level=info msg="shim disconnected" id=a0d12baf896e694d49ea3639f57c44ef8f625f1451d9d138874eacfb5898082a namespace=k8s.io Aug 13 00:26:57.639665 containerd[2685]: time="2025-08-13T00:26:57.639663751Z" level=warning msg="cleaning up after shim disconnected" id=a0d12baf896e694d49ea3639f57c44ef8f625f1451d9d138874eacfb5898082a namespace=k8s.io Aug 13 00:26:57.639746 containerd[2685]: time="2025-08-13T00:26:57.639671831Z" level=info msg="cleaning up dead shim" namespace=k8s.io Aug 13 00:26:57.642678 systemd[1]: Created slice kubepods-burstable-pod085d2185_0930_470d_b50d_68703ac546fb.slice - libcontainer container kubepods-burstable-pod085d2185_0930_470d_b50d_68703ac546fb.slice. Aug 13 00:26:57.646516 systemd[1]: Created slice kubepods-besteffort-pod925e0d95_7160_42b1_a3c8_6b7795ca67bc.slice - libcontainer container kubepods-besteffort-pod925e0d95_7160_42b1_a3c8_6b7795ca67bc.slice. Aug 13 00:26:57.650255 systemd[1]: Created slice kubepods-besteffort-pod3dc1e9ad_8b04_42ba_98fc_165c43766984.slice - libcontainer container kubepods-besteffort-pod3dc1e9ad_8b04_42ba_98fc_165c43766984.slice. Aug 13 00:26:57.654113 systemd[1]: Created slice kubepods-besteffort-pode7e9845f_f3a1_4760_b84c_8f704f46a611.slice - libcontainer container kubepods-besteffort-pode7e9845f_f3a1_4760_b84c_8f704f46a611.slice. Aug 13 00:26:57.657754 systemd[1]: Created slice kubepods-besteffort-podb5da64a2_733d_4c0b_b26c_80dcab04992c.slice - libcontainer container kubepods-besteffort-podb5da64a2_733d_4c0b_b26c_80dcab04992c.slice. Aug 13 00:26:57.661486 systemd[1]: Created slice kubepods-besteffort-pod0ced7fba_ab7f_4d92_941c_7095a5aea251.slice - libcontainer container kubepods-besteffort-pod0ced7fba_ab7f_4d92_941c_7095a5aea251.slice. Aug 13 00:26:57.666797 systemd[1]: Created slice kubepods-besteffort-pod517c2f94_3455_4677_a3ec_3a8771258060.slice - libcontainer container kubepods-besteffort-pod517c2f94_3455_4677_a3ec_3a8771258060.slice. Aug 13 00:26:57.670135 kubelet[4148]: I0813 00:26:57.670109 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e7e9845f-f3a1-4760-b84c-8f704f46a611-whisker-ca-bundle\") pod \"whisker-67db74799b-hr8ld\" (UID: \"e7e9845f-f3a1-4760-b84c-8f704f46a611\") " pod="calico-system/whisker-67db74799b-hr8ld" Aug 13 00:26:57.670189 kubelet[4148]: I0813 00:26:57.670145 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gm4s4\" (UniqueName: \"kubernetes.io/projected/e7e9845f-f3a1-4760-b84c-8f704f46a611-kube-api-access-gm4s4\") pod \"whisker-67db74799b-hr8ld\" (UID: \"e7e9845f-f3a1-4760-b84c-8f704f46a611\") " pod="calico-system/whisker-67db74799b-hr8ld" Aug 13 00:26:57.670189 kubelet[4148]: I0813 00:26:57.670164 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/517c2f94-3455-4677-a3ec-3a8771258060-calico-apiserver-certs\") pod \"calico-apiserver-74488cd4c9-h5l86\" (UID: \"517c2f94-3455-4677-a3ec-3a8771258060\") " pod="calico-apiserver/calico-apiserver-74488cd4c9-h5l86" Aug 13 00:26:57.670295 kubelet[4148]: I0813 00:26:57.670265 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/925e0d95-7160-42b1-a3c8-6b7795ca67bc-tigera-ca-bundle\") pod \"calico-kube-controllers-f5cbb957c-wms9w\" (UID: \"925e0d95-7160-42b1-a3c8-6b7795ca67bc\") " pod="calico-system/calico-kube-controllers-f5cbb957c-wms9w" Aug 13 00:26:57.670353 kubelet[4148]: I0813 00:26:57.670336 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b5da64a2-733d-4c0b-b26c-80dcab04992c-goldmane-ca-bundle\") pod \"goldmane-768f4c5c69-nm6vx\" (UID: \"b5da64a2-733d-4c0b-b26c-80dcab04992c\") " pod="calico-system/goldmane-768f4c5c69-nm6vx" Aug 13 00:26:57.670407 kubelet[4148]: I0813 00:26:57.670385 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbmgz\" (UniqueName: \"kubernetes.io/projected/517c2f94-3455-4677-a3ec-3a8771258060-kube-api-access-pbmgz\") pod \"calico-apiserver-74488cd4c9-h5l86\" (UID: \"517c2f94-3455-4677-a3ec-3a8771258060\") " pod="calico-apiserver/calico-apiserver-74488cd4c9-h5l86" Aug 13 00:26:57.670474 kubelet[4148]: I0813 00:26:57.670418 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/75d28fd6-77d6-4c58-8897-8ce0729a65ea-config-volume\") pod \"coredns-674b8bbfcf-v7mtb\" (UID: \"75d28fd6-77d6-4c58-8897-8ce0729a65ea\") " pod="kube-system/coredns-674b8bbfcf-v7mtb" Aug 13 00:26:57.670474 kubelet[4148]: I0813 00:26:57.670437 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ff4j\" (UniqueName: \"kubernetes.io/projected/3dc1e9ad-8b04-42ba-98fc-165c43766984-kube-api-access-2ff4j\") pod \"calico-apiserver-54c5fc9686-lcpkt\" (UID: \"3dc1e9ad-8b04-42ba-98fc-165c43766984\") " pod="calico-apiserver/calico-apiserver-54c5fc9686-lcpkt" Aug 13 00:26:57.670520 kubelet[4148]: I0813 00:26:57.670487 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5da64a2-733d-4c0b-b26c-80dcab04992c-config\") pod \"goldmane-768f4c5c69-nm6vx\" (UID: \"b5da64a2-733d-4c0b-b26c-80dcab04992c\") " pod="calico-system/goldmane-768f4c5c69-nm6vx" Aug 13 00:26:57.670520 kubelet[4148]: I0813 00:26:57.670506 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"goldmane-key-pair\" (UniqueName: \"kubernetes.io/secret/b5da64a2-733d-4c0b-b26c-80dcab04992c-goldmane-key-pair\") pod \"goldmane-768f4c5c69-nm6vx\" (UID: \"b5da64a2-733d-4c0b-b26c-80dcab04992c\") " pod="calico-system/goldmane-768f4c5c69-nm6vx" Aug 13 00:26:57.670570 kubelet[4148]: I0813 00:26:57.670533 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/085d2185-0930-470d-b50d-68703ac546fb-config-volume\") pod \"coredns-674b8bbfcf-ftxjx\" (UID: \"085d2185-0930-470d-b50d-68703ac546fb\") " pod="kube-system/coredns-674b8bbfcf-ftxjx" Aug 13 00:26:57.670570 kubelet[4148]: I0813 00:26:57.670550 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4zz6\" (UniqueName: \"kubernetes.io/projected/085d2185-0930-470d-b50d-68703ac546fb-kube-api-access-n4zz6\") pod \"coredns-674b8bbfcf-ftxjx\" (UID: \"085d2185-0930-470d-b50d-68703ac546fb\") " pod="kube-system/coredns-674b8bbfcf-ftxjx" Aug 13 00:26:57.670570 kubelet[4148]: I0813 00:26:57.670564 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/e7e9845f-f3a1-4760-b84c-8f704f46a611-whisker-backend-key-pair\") pod \"whisker-67db74799b-hr8ld\" (UID: \"e7e9845f-f3a1-4760-b84c-8f704f46a611\") " pod="calico-system/whisker-67db74799b-hr8ld" Aug 13 00:26:57.670633 kubelet[4148]: I0813 00:26:57.670581 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6m4wd\" (UniqueName: \"kubernetes.io/projected/b5da64a2-733d-4c0b-b26c-80dcab04992c-kube-api-access-6m4wd\") pod \"goldmane-768f4c5c69-nm6vx\" (UID: \"b5da64a2-733d-4c0b-b26c-80dcab04992c\") " pod="calico-system/goldmane-768f4c5c69-nm6vx" Aug 13 00:26:57.670633 kubelet[4148]: I0813 00:26:57.670619 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fpl6\" (UniqueName: \"kubernetes.io/projected/75d28fd6-77d6-4c58-8897-8ce0729a65ea-kube-api-access-5fpl6\") pod \"coredns-674b8bbfcf-v7mtb\" (UID: \"75d28fd6-77d6-4c58-8897-8ce0729a65ea\") " pod="kube-system/coredns-674b8bbfcf-v7mtb" Aug 13 00:26:57.670682 kubelet[4148]: I0813 00:26:57.670668 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rl58n\" (UniqueName: \"kubernetes.io/projected/925e0d95-7160-42b1-a3c8-6b7795ca67bc-kube-api-access-rl58n\") pod \"calico-kube-controllers-f5cbb957c-wms9w\" (UID: \"925e0d95-7160-42b1-a3c8-6b7795ca67bc\") " pod="calico-system/calico-kube-controllers-f5cbb957c-wms9w" Aug 13 00:26:57.670709 kubelet[4148]: I0813 00:26:57.670692 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/0ced7fba-ab7f-4d92-941c-7095a5aea251-calico-apiserver-certs\") pod \"calico-apiserver-74488cd4c9-4dnl6\" (UID: \"0ced7fba-ab7f-4d92-941c-7095a5aea251\") " pod="calico-apiserver/calico-apiserver-74488cd4c9-4dnl6" Aug 13 00:26:57.670739 kubelet[4148]: I0813 00:26:57.670722 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvxdb\" (UniqueName: \"kubernetes.io/projected/0ced7fba-ab7f-4d92-941c-7095a5aea251-kube-api-access-wvxdb\") pod \"calico-apiserver-74488cd4c9-4dnl6\" (UID: \"0ced7fba-ab7f-4d92-941c-7095a5aea251\") " pod="calico-apiserver/calico-apiserver-74488cd4c9-4dnl6" Aug 13 00:26:57.670763 kubelet[4148]: I0813 00:26:57.670749 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/3dc1e9ad-8b04-42ba-98fc-165c43766984-calico-apiserver-certs\") pod \"calico-apiserver-54c5fc9686-lcpkt\" (UID: \"3dc1e9ad-8b04-42ba-98fc-165c43766984\") " pod="calico-apiserver/calico-apiserver-54c5fc9686-lcpkt" Aug 13 00:26:57.931415 containerd[2685]: time="2025-08-13T00:26:57.931384339Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-v7mtb,Uid:75d28fd6-77d6-4c58-8897-8ce0729a65ea,Namespace:kube-system,Attempt:0,}" Aug 13 00:26:57.944961 containerd[2685]: time="2025-08-13T00:26:57.944926392Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-ftxjx,Uid:085d2185-0930-470d-b50d-68703ac546fb,Namespace:kube-system,Attempt:0,}" Aug 13 00:26:57.949491 containerd[2685]: time="2025-08-13T00:26:57.949456236Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-f5cbb957c-wms9w,Uid:925e0d95-7160-42b1-a3c8-6b7795ca67bc,Namespace:calico-system,Attempt:0,}" Aug 13 00:26:57.953042 containerd[2685]: time="2025-08-13T00:26:57.952960039Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-54c5fc9686-lcpkt,Uid:3dc1e9ad-8b04-42ba-98fc-165c43766984,Namespace:calico-apiserver,Attempt:0,}" Aug 13 00:26:57.956423 containerd[2685]: time="2025-08-13T00:26:57.956399242Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-67db74799b-hr8ld,Uid:e7e9845f-f3a1-4760-b84c-8f704f46a611,Namespace:calico-system,Attempt:0,}" Aug 13 00:26:57.960062 containerd[2685]: time="2025-08-13T00:26:57.960041605Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-768f4c5c69-nm6vx,Uid:b5da64a2-733d-4c0b-b26c-80dcab04992c,Namespace:calico-system,Attempt:0,}" Aug 13 00:26:57.965591 containerd[2685]: time="2025-08-13T00:26:57.965570051Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-74488cd4c9-4dnl6,Uid:0ced7fba-ab7f-4d92-941c-7095a5aea251,Namespace:calico-apiserver,Attempt:0,}" Aug 13 00:26:57.969391 containerd[2685]: time="2025-08-13T00:26:57.969361094Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-74488cd4c9-h5l86,Uid:517c2f94-3455-4677-a3ec-3a8771258060,Namespace:calico-apiserver,Attempt:0,}" Aug 13 00:26:57.992187 containerd[2685]: time="2025-08-13T00:26:57.992137195Z" level=error msg="Failed to destroy network for sandbox \"e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:57.992534 containerd[2685]: time="2025-08-13T00:26:57.992508475Z" level=error msg="encountered an error cleaning up failed sandbox \"e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:57.992579 containerd[2685]: time="2025-08-13T00:26:57.992559755Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-v7mtb,Uid:75d28fd6-77d6-4c58-8897-8ce0729a65ea,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:57.992641 containerd[2685]: time="2025-08-13T00:26:57.992611755Z" level=error msg="Failed to destroy network for sandbox \"d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:57.992793 kubelet[4148]: E0813 00:26:57.992766 4148 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:57.992834 kubelet[4148]: E0813 00:26:57.992820 4148 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-v7mtb" Aug 13 00:26:57.992857 kubelet[4148]: E0813 00:26:57.992840 4148 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-v7mtb" Aug 13 00:26:57.992907 kubelet[4148]: E0813 00:26:57.992888 4148 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-v7mtb_kube-system(75d28fd6-77d6-4c58-8897-8ce0729a65ea)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-v7mtb_kube-system(75d28fd6-77d6-4c58-8897-8ce0729a65ea)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-v7mtb" podUID="75d28fd6-77d6-4c58-8897-8ce0729a65ea" Aug 13 00:26:57.992968 containerd[2685]: time="2025-08-13T00:26:57.992931716Z" level=error msg="encountered an error cleaning up failed sandbox \"d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:57.992991 containerd[2685]: time="2025-08-13T00:26:57.992977756Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-ftxjx,Uid:085d2185-0930-470d-b50d-68703ac546fb,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:57.993009 systemd[1]: Created slice kubepods-besteffort-podbb51369a_5ac7_4ac1_b819_64d9c4e312d0.slice - libcontainer container kubepods-besteffort-podbb51369a_5ac7_4ac1_b819_64d9c4e312d0.slice. Aug 13 00:26:57.993386 kubelet[4148]: E0813 00:26:57.993354 4148 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:57.993423 kubelet[4148]: E0813 00:26:57.993405 4148 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-ftxjx" Aug 13 00:26:57.993450 kubelet[4148]: E0813 00:26:57.993422 4148 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-674b8bbfcf-ftxjx" Aug 13 00:26:57.993490 kubelet[4148]: E0813 00:26:57.993468 4148 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-674b8bbfcf-ftxjx_kube-system(085d2185-0930-470d-b50d-68703ac546fb)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-674b8bbfcf-ftxjx_kube-system(085d2185-0930-470d-b50d-68703ac546fb)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-ftxjx" podUID="085d2185-0930-470d-b50d-68703ac546fb" Aug 13 00:26:57.995046 containerd[2685]: time="2025-08-13T00:26:57.995020478Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-4wkp8,Uid:bb51369a-5ac7-4ac1-b819-64d9c4e312d0,Namespace:calico-system,Attempt:0,}" Aug 13 00:26:58.006339 containerd[2685]: time="2025-08-13T00:26:58.006291288Z" level=error msg="Failed to destroy network for sandbox \"fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:58.006908 containerd[2685]: time="2025-08-13T00:26:58.006878168Z" level=error msg="encountered an error cleaning up failed sandbox \"fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:58.006964 containerd[2685]: time="2025-08-13T00:26:58.006933008Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-f5cbb957c-wms9w,Uid:925e0d95-7160-42b1-a3c8-6b7795ca67bc,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:58.007170 kubelet[4148]: E0813 00:26:58.007124 4148 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:58.007207 kubelet[4148]: E0813 00:26:58.007192 4148 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-f5cbb957c-wms9w" Aug 13 00:26:58.007247 kubelet[4148]: E0813 00:26:58.007211 4148 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-f5cbb957c-wms9w" Aug 13 00:26:58.007310 kubelet[4148]: E0813 00:26:58.007289 4148 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-f5cbb957c-wms9w_calico-system(925e0d95-7160-42b1-a3c8-6b7795ca67bc)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-f5cbb957c-wms9w_calico-system(925e0d95-7160-42b1-a3c8-6b7795ca67bc)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-f5cbb957c-wms9w" podUID="925e0d95-7160-42b1-a3c8-6b7795ca67bc" Aug 13 00:26:58.009319 containerd[2685]: time="2025-08-13T00:26:58.009284010Z" level=error msg="Failed to destroy network for sandbox \"7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:58.009550 containerd[2685]: time="2025-08-13T00:26:58.009525211Z" level=error msg="Failed to destroy network for sandbox \"74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:58.009628 containerd[2685]: time="2025-08-13T00:26:58.009605411Z" level=error msg="encountered an error cleaning up failed sandbox \"7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:58.009672 containerd[2685]: time="2025-08-13T00:26:58.009651491Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-54c5fc9686-lcpkt,Uid:3dc1e9ad-8b04-42ba-98fc-165c43766984,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:58.009850 kubelet[4148]: E0813 00:26:58.009811 4148 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:58.010210 kubelet[4148]: E0813 00:26:58.009870 4148 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-54c5fc9686-lcpkt" Aug 13 00:26:58.010210 kubelet[4148]: E0813 00:26:58.009892 4148 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-54c5fc9686-lcpkt" Aug 13 00:26:58.010210 kubelet[4148]: E0813 00:26:58.009943 4148 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-54c5fc9686-lcpkt_calico-apiserver(3dc1e9ad-8b04-42ba-98fc-165c43766984)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-54c5fc9686-lcpkt_calico-apiserver(3dc1e9ad-8b04-42ba-98fc-165c43766984)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-54c5fc9686-lcpkt" podUID="3dc1e9ad-8b04-42ba-98fc-165c43766984" Aug 13 00:26:58.010366 containerd[2685]: time="2025-08-13T00:26:58.009853811Z" level=error msg="encountered an error cleaning up failed sandbox \"74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:58.010366 containerd[2685]: time="2025-08-13T00:26:58.009893691Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-67db74799b-hr8ld,Uid:e7e9845f-f3a1-4760-b84c-8f704f46a611,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:58.010426 kubelet[4148]: E0813 00:26:58.010003 4148 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:58.010426 kubelet[4148]: E0813 00:26:58.010041 4148 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-67db74799b-hr8ld" Aug 13 00:26:58.010426 kubelet[4148]: E0813 00:26:58.010056 4148 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/whisker-67db74799b-hr8ld" Aug 13 00:26:58.010504 containerd[2685]: time="2025-08-13T00:26:58.010394091Z" level=error msg="Failed to destroy network for sandbox \"2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:58.010525 kubelet[4148]: E0813 00:26:58.010101 4148 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"whisker-67db74799b-hr8ld_calico-system(e7e9845f-f3a1-4760-b84c-8f704f46a611)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"whisker-67db74799b-hr8ld_calico-system(e7e9845f-f3a1-4760-b84c-8f704f46a611)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-67db74799b-hr8ld" podUID="e7e9845f-f3a1-4760-b84c-8f704f46a611" Aug 13 00:26:58.010722 containerd[2685]: time="2025-08-13T00:26:58.010691892Z" level=error msg="encountered an error cleaning up failed sandbox \"2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:58.010754 containerd[2685]: time="2025-08-13T00:26:58.010739652Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-74488cd4c9-4dnl6,Uid:0ced7fba-ab7f-4d92-941c-7095a5aea251,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:58.010866 kubelet[4148]: E0813 00:26:58.010843 4148 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:58.010891 kubelet[4148]: E0813 00:26:58.010880 4148 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-74488cd4c9-4dnl6" Aug 13 00:26:58.010917 kubelet[4148]: E0813 00:26:58.010898 4148 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-74488cd4c9-4dnl6" Aug 13 00:26:58.010949 kubelet[4148]: E0813 00:26:58.010932 4148 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-74488cd4c9-4dnl6_calico-apiserver(0ced7fba-ab7f-4d92-941c-7095a5aea251)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-74488cd4c9-4dnl6_calico-apiserver(0ced7fba-ab7f-4d92-941c-7095a5aea251)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-74488cd4c9-4dnl6" podUID="0ced7fba-ab7f-4d92-941c-7095a5aea251" Aug 13 00:26:58.012544 containerd[2685]: time="2025-08-13T00:26:58.012517773Z" level=error msg="Failed to destroy network for sandbox \"3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:58.012869 containerd[2685]: time="2025-08-13T00:26:58.012846253Z" level=error msg="encountered an error cleaning up failed sandbox \"3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:58.012905 containerd[2685]: time="2025-08-13T00:26:58.012888733Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-768f4c5c69-nm6vx,Uid:b5da64a2-733d-4c0b-b26c-80dcab04992c,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:58.013011 kubelet[4148]: E0813 00:26:58.012998 4148 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:58.013040 kubelet[4148]: E0813 00:26:58.013018 4148 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-768f4c5c69-nm6vx" Aug 13 00:26:58.013040 kubelet[4148]: E0813 00:26:58.013030 4148 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/goldmane-768f4c5c69-nm6vx" Aug 13 00:26:58.013084 kubelet[4148]: E0813 00:26:58.013060 4148 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"goldmane-768f4c5c69-nm6vx_calico-system(b5da64a2-733d-4c0b-b26c-80dcab04992c)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"goldmane-768f4c5c69-nm6vx_calico-system(b5da64a2-733d-4c0b-b26c-80dcab04992c)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-768f4c5c69-nm6vx" podUID="b5da64a2-733d-4c0b-b26c-80dcab04992c" Aug 13 00:26:58.015150 containerd[2685]: time="2025-08-13T00:26:58.015123375Z" level=error msg="Failed to destroy network for sandbox \"96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:58.015502 containerd[2685]: time="2025-08-13T00:26:58.015471376Z" level=error msg="encountered an error cleaning up failed sandbox \"96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:58.015544 containerd[2685]: time="2025-08-13T00:26:58.015521576Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-74488cd4c9-h5l86,Uid:517c2f94-3455-4677-a3ec-3a8771258060,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:58.015676 kubelet[4148]: E0813 00:26:58.015651 4148 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:58.015704 kubelet[4148]: E0813 00:26:58.015691 4148 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-74488cd4c9-h5l86" Aug 13 00:26:58.015728 kubelet[4148]: E0813 00:26:58.015707 4148 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-74488cd4c9-h5l86" Aug 13 00:26:58.015762 kubelet[4148]: E0813 00:26:58.015743 4148 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-74488cd4c9-h5l86_calico-apiserver(517c2f94-3455-4677-a3ec-3a8771258060)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-74488cd4c9-h5l86_calico-apiserver(517c2f94-3455-4677-a3ec-3a8771258060)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-74488cd4c9-h5l86" podUID="517c2f94-3455-4677-a3ec-3a8771258060" Aug 13 00:26:58.029221 kubelet[4148]: I0813 00:26:58.029206 4148 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" Aug 13 00:26:58.029764 containerd[2685]: time="2025-08-13T00:26:58.029746268Z" level=info msg="StopPodSandbox for \"e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f\"" Aug 13 00:26:58.029844 kubelet[4148]: I0813 00:26:58.029828 4148 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" Aug 13 00:26:58.029898 containerd[2685]: time="2025-08-13T00:26:58.029885708Z" level=info msg="Ensure that sandbox e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f in task-service has been cleanup successfully" Aug 13 00:26:58.030264 containerd[2685]: time="2025-08-13T00:26:58.030244628Z" level=info msg="StopPodSandbox for \"fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02\"" Aug 13 00:26:58.030398 containerd[2685]: time="2025-08-13T00:26:58.030385469Z" level=info msg="Ensure that sandbox fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02 in task-service has been cleanup successfully" Aug 13 00:26:58.030790 kubelet[4148]: I0813 00:26:58.030778 4148 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" Aug 13 00:26:58.031215 containerd[2685]: time="2025-08-13T00:26:58.031194869Z" level=info msg="StopPodSandbox for \"d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3\"" Aug 13 00:26:58.031376 containerd[2685]: time="2025-08-13T00:26:58.031362509Z" level=info msg="Ensure that sandbox d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3 in task-service has been cleanup successfully" Aug 13 00:26:58.033146 containerd[2685]: time="2025-08-13T00:26:58.033127551Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.2\"" Aug 13 00:26:58.033277 kubelet[4148]: I0813 00:26:58.033261 4148 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" Aug 13 00:26:58.033674 containerd[2685]: time="2025-08-13T00:26:58.033654151Z" level=info msg="StopPodSandbox for \"2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8\"" Aug 13 00:26:58.033811 containerd[2685]: time="2025-08-13T00:26:58.033797832Z" level=info msg="Ensure that sandbox 2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8 in task-service has been cleanup successfully" Aug 13 00:26:58.033968 kubelet[4148]: I0813 00:26:58.033953 4148 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" Aug 13 00:26:58.034419 containerd[2685]: time="2025-08-13T00:26:58.034394632Z" level=info msg="StopPodSandbox for \"3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd\"" Aug 13 00:26:58.034551 containerd[2685]: time="2025-08-13T00:26:58.034535152Z" level=info msg="Ensure that sandbox 3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd in task-service has been cleanup successfully" Aug 13 00:26:58.035015 kubelet[4148]: I0813 00:26:58.035001 4148 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" Aug 13 00:26:58.035480 containerd[2685]: time="2025-08-13T00:26:58.035458433Z" level=info msg="StopPodSandbox for \"96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce\"" Aug 13 00:26:58.035783 containerd[2685]: time="2025-08-13T00:26:58.035764873Z" level=info msg="Ensure that sandbox 96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce in task-service has been cleanup successfully" Aug 13 00:26:58.035954 kubelet[4148]: I0813 00:26:58.035943 4148 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" Aug 13 00:26:58.036411 containerd[2685]: time="2025-08-13T00:26:58.036390754Z" level=info msg="StopPodSandbox for \"74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef\"" Aug 13 00:26:58.036544 containerd[2685]: time="2025-08-13T00:26:58.036530154Z" level=info msg="Ensure that sandbox 74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef in task-service has been cleanup successfully" Aug 13 00:26:58.036704 kubelet[4148]: I0813 00:26:58.036690 4148 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" Aug 13 00:26:58.037112 containerd[2685]: time="2025-08-13T00:26:58.037086794Z" level=info msg="StopPodSandbox for \"7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530\"" Aug 13 00:26:58.037259 containerd[2685]: time="2025-08-13T00:26:58.037244794Z" level=info msg="Ensure that sandbox 7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530 in task-service has been cleanup successfully" Aug 13 00:26:58.039002 containerd[2685]: time="2025-08-13T00:26:58.038973356Z" level=error msg="Failed to destroy network for sandbox \"4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:58.039369 containerd[2685]: time="2025-08-13T00:26:58.039345036Z" level=error msg="encountered an error cleaning up failed sandbox \"4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:58.039415 containerd[2685]: time="2025-08-13T00:26:58.039396196Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-4wkp8,Uid:bb51369a-5ac7-4ac1-b819-64d9c4e312d0,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:58.039572 kubelet[4148]: E0813 00:26:58.039547 4148 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:58.039606 kubelet[4148]: E0813 00:26:58.039590 4148 kuberuntime_sandbox.go:70] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-4wkp8" Aug 13 00:26:58.039631 kubelet[4148]: E0813 00:26:58.039613 4148 kuberuntime_manager.go:1252] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-4wkp8" Aug 13 00:26:58.039906 kubelet[4148]: E0813 00:26:58.039652 4148 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-4wkp8_calico-system(bb51369a-5ac7-4ac1-b819-64d9c4e312d0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-4wkp8_calico-system(bb51369a-5ac7-4ac1-b819-64d9c4e312d0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-4wkp8" podUID="bb51369a-5ac7-4ac1-b819-64d9c4e312d0" Aug 13 00:26:58.051645 containerd[2685]: time="2025-08-13T00:26:58.051602607Z" level=error msg="StopPodSandbox for \"e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f\" failed" error="failed to destroy network for sandbox \"e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:58.051825 kubelet[4148]: E0813 00:26:58.051789 4148 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" Aug 13 00:26:58.051880 kubelet[4148]: E0813 00:26:58.051848 4148 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f"} Aug 13 00:26:58.051918 kubelet[4148]: E0813 00:26:58.051899 4148 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"75d28fd6-77d6-4c58-8897-8ce0729a65ea\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Aug 13 00:26:58.051970 kubelet[4148]: E0813 00:26:58.051928 4148 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"75d28fd6-77d6-4c58-8897-8ce0729a65ea\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-v7mtb" podUID="75d28fd6-77d6-4c58-8897-8ce0729a65ea" Aug 13 00:26:58.052247 containerd[2685]: time="2025-08-13T00:26:58.052214327Z" level=error msg="StopPodSandbox for \"fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02\" failed" error="failed to destroy network for sandbox \"fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:58.052358 kubelet[4148]: E0813 00:26:58.052333 4148 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" Aug 13 00:26:58.052386 kubelet[4148]: E0813 00:26:58.052368 4148 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02"} Aug 13 00:26:58.052408 kubelet[4148]: E0813 00:26:58.052395 4148 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"925e0d95-7160-42b1-a3c8-6b7795ca67bc\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Aug 13 00:26:58.052443 kubelet[4148]: E0813 00:26:58.052413 4148 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"925e0d95-7160-42b1-a3c8-6b7795ca67bc\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-f5cbb957c-wms9w" podUID="925e0d95-7160-42b1-a3c8-6b7795ca67bc" Aug 13 00:26:58.052548 containerd[2685]: time="2025-08-13T00:26:58.052511128Z" level=error msg="StopPodSandbox for \"d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3\" failed" error="failed to destroy network for sandbox \"d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:58.052661 kubelet[4148]: E0813 00:26:58.052646 4148 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" Aug 13 00:26:58.052685 kubelet[4148]: E0813 00:26:58.052665 4148 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3"} Aug 13 00:26:58.052705 kubelet[4148]: E0813 00:26:58.052683 4148 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"085d2185-0930-470d-b50d-68703ac546fb\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Aug 13 00:26:58.052739 kubelet[4148]: E0813 00:26:58.052699 4148 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"085d2185-0930-470d-b50d-68703ac546fb\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-674b8bbfcf-ftxjx" podUID="085d2185-0930-470d-b50d-68703ac546fb" Aug 13 00:26:58.055696 containerd[2685]: time="2025-08-13T00:26:58.055665730Z" level=error msg="StopPodSandbox for \"2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8\" failed" error="failed to destroy network for sandbox \"2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:58.055852 kubelet[4148]: E0813 00:26:58.055820 4148 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" Aug 13 00:26:58.055882 kubelet[4148]: E0813 00:26:58.055868 4148 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8"} Aug 13 00:26:58.055905 kubelet[4148]: E0813 00:26:58.055896 4148 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"0ced7fba-ab7f-4d92-941c-7095a5aea251\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Aug 13 00:26:58.055939 kubelet[4148]: E0813 00:26:58.055917 4148 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"0ced7fba-ab7f-4d92-941c-7095a5aea251\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-74488cd4c9-4dnl6" podUID="0ced7fba-ab7f-4d92-941c-7095a5aea251" Aug 13 00:26:58.055983 containerd[2685]: time="2025-08-13T00:26:58.055951931Z" level=error msg="StopPodSandbox for \"3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd\" failed" error="failed to destroy network for sandbox \"3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:58.056119 kubelet[4148]: E0813 00:26:58.056103 4148 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" Aug 13 00:26:58.056142 kubelet[4148]: E0813 00:26:58.056122 4148 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd"} Aug 13 00:26:58.056166 kubelet[4148]: E0813 00:26:58.056139 4148 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"b5da64a2-733d-4c0b-b26c-80dcab04992c\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Aug 13 00:26:58.056166 kubelet[4148]: E0813 00:26:58.056155 4148 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"b5da64a2-733d-4c0b-b26c-80dcab04992c\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/goldmane-768f4c5c69-nm6vx" podUID="b5da64a2-733d-4c0b-b26c-80dcab04992c" Aug 13 00:26:58.057990 containerd[2685]: time="2025-08-13T00:26:58.057959652Z" level=error msg="StopPodSandbox for \"7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530\" failed" error="failed to destroy network for sandbox \"7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:58.058104 kubelet[4148]: E0813 00:26:58.058081 4148 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" Aug 13 00:26:58.058130 kubelet[4148]: E0813 00:26:58.058113 4148 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530"} Aug 13 00:26:58.058151 kubelet[4148]: E0813 00:26:58.058137 4148 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"3dc1e9ad-8b04-42ba-98fc-165c43766984\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Aug 13 00:26:58.058193 kubelet[4148]: E0813 00:26:58.058154 4148 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"3dc1e9ad-8b04-42ba-98fc-165c43766984\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-54c5fc9686-lcpkt" podUID="3dc1e9ad-8b04-42ba-98fc-165c43766984" Aug 13 00:26:58.058491 containerd[2685]: time="2025-08-13T00:26:58.058416293Z" level=error msg="StopPodSandbox for \"74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef\" failed" error="failed to destroy network for sandbox \"74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:58.058565 kubelet[4148]: E0813 00:26:58.058546 4148 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" Aug 13 00:26:58.058593 kubelet[4148]: E0813 00:26:58.058567 4148 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef"} Aug 13 00:26:58.058614 kubelet[4148]: E0813 00:26:58.058588 4148 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"e7e9845f-f3a1-4760-b84c-8f704f46a611\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Aug 13 00:26:58.058614 kubelet[4148]: E0813 00:26:58.058604 4148 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"e7e9845f-f3a1-4760-b84c-8f704f46a611\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/whisker-67db74799b-hr8ld" podUID="e7e9845f-f3a1-4760-b84c-8f704f46a611" Aug 13 00:26:58.059157 containerd[2685]: time="2025-08-13T00:26:58.059125533Z" level=error msg="StopPodSandbox for \"96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce\" failed" error="failed to destroy network for sandbox \"96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:58.059260 kubelet[4148]: E0813 00:26:58.059236 4148 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" Aug 13 00:26:58.059297 kubelet[4148]: E0813 00:26:58.059267 4148 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce"} Aug 13 00:26:58.059297 kubelet[4148]: E0813 00:26:58.059289 4148 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"517c2f94-3455-4677-a3ec-3a8771258060\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Aug 13 00:26:58.059356 kubelet[4148]: E0813 00:26:58.059306 4148 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"517c2f94-3455-4677-a3ec-3a8771258060\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-74488cd4c9-h5l86" podUID="517c2f94-3455-4677-a3ec-3a8771258060" Aug 13 00:26:58.095254 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a0d12baf896e694d49ea3639f57c44ef8f625f1451d9d138874eacfb5898082a-rootfs.mount: Deactivated successfully. Aug 13 00:26:59.038552 kubelet[4148]: I0813 00:26:59.038525 4148 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" Aug 13 00:26:59.039062 containerd[2685]: time="2025-08-13T00:26:59.039025777Z" level=info msg="StopPodSandbox for \"4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652\"" Aug 13 00:26:59.039255 containerd[2685]: time="2025-08-13T00:26:59.039196457Z" level=info msg="Ensure that sandbox 4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652 in task-service has been cleanup successfully" Aug 13 00:26:59.059820 containerd[2685]: time="2025-08-13T00:26:59.059781194Z" level=error msg="StopPodSandbox for \"4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652\" failed" error="failed to destroy network for sandbox \"4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Aug 13 00:26:59.060001 kubelet[4148]: E0813 00:26:59.059968 4148 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" Aug 13 00:26:59.060043 kubelet[4148]: E0813 00:26:59.060015 4148 kuberuntime_manager.go:1586] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652"} Aug 13 00:26:59.060065 kubelet[4148]: E0813 00:26:59.060047 4148 kuberuntime_manager.go:1161] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"bb51369a-5ac7-4ac1-b819-64d9c4e312d0\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Aug 13 00:26:59.060120 kubelet[4148]: E0813 00:26:59.060065 4148 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"bb51369a-5ac7-4ac1-b819-64d9c4e312d0\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-4wkp8" podUID="bb51369a-5ac7-4ac1-b819-64d9c4e312d0" Aug 13 00:27:01.684449 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3945305811.mount: Deactivated successfully. Aug 13 00:27:01.706348 containerd[2685]: time="2025-08-13T00:27:01.706299775Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:27:01.706634 containerd[2685]: time="2025-08-13T00:27:01.706328935Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.30.2: active requests=0, bytes read=152544909" Aug 13 00:27:01.707046 containerd[2685]: time="2025-08-13T00:27:01.707024295Z" level=info msg="ImageCreate event name:\"sha256:1c6ddca599ddd18c061e797a7830b0aea985f8b023c5e43d815a9ed1088893a9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:27:01.708709 containerd[2685]: time="2025-08-13T00:27:01.708683417Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:e94d49349cc361ef2216d27dda4a097278984d778279f66e79b0616c827c6760\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:27:01.709328 containerd[2685]: time="2025-08-13T00:27:01.709303457Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.30.2\" with image id \"sha256:1c6ddca599ddd18c061e797a7830b0aea985f8b023c5e43d815a9ed1088893a9\", repo tag \"ghcr.io/flatcar/calico/node:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/node@sha256:e94d49349cc361ef2216d27dda4a097278984d778279f66e79b0616c827c6760\", size \"152544771\" in 3.676146906s" Aug 13 00:27:01.709371 containerd[2685]: time="2025-08-13T00:27:01.709334377Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.30.2\" returns image reference \"sha256:1c6ddca599ddd18c061e797a7830b0aea985f8b023c5e43d815a9ed1088893a9\"" Aug 13 00:27:01.715276 containerd[2685]: time="2025-08-13T00:27:01.715249061Z" level=info msg="CreateContainer within sandbox \"49068c488f9a318d05f6e0e80d5c70f277ccc8bbe32e3e6b64b9cf6d495d3342\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Aug 13 00:27:01.723140 containerd[2685]: time="2025-08-13T00:27:01.723104427Z" level=info msg="CreateContainer within sandbox \"49068c488f9a318d05f6e0e80d5c70f277ccc8bbe32e3e6b64b9cf6d495d3342\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"6d83c1741b266606fed85026ec22a1fa9963e22f6417b8762cbfd7a6a660448c\"" Aug 13 00:27:01.723485 containerd[2685]: time="2025-08-13T00:27:01.723456187Z" level=info msg="StartContainer for \"6d83c1741b266606fed85026ec22a1fa9963e22f6417b8762cbfd7a6a660448c\"" Aug 13 00:27:01.760402 systemd[1]: Started cri-containerd-6d83c1741b266606fed85026ec22a1fa9963e22f6417b8762cbfd7a6a660448c.scope - libcontainer container 6d83c1741b266606fed85026ec22a1fa9963e22f6417b8762cbfd7a6a660448c. Aug 13 00:27:01.781624 containerd[2685]: time="2025-08-13T00:27:01.781591708Z" level=info msg="StartContainer for \"6d83c1741b266606fed85026ec22a1fa9963e22f6417b8762cbfd7a6a660448c\" returns successfully" Aug 13 00:27:01.913222 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Aug 13 00:27:01.913298 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Aug 13 00:27:01.975461 containerd[2685]: time="2025-08-13T00:27:01.975366526Z" level=info msg="StopPodSandbox for \"74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef\"" Aug 13 00:27:02.055424 kubelet[4148]: I0813 00:27:02.055369 4148 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-rjjzv" podStartSLOduration=0.74125134 podStartE2EDuration="10.055352981s" podCreationTimestamp="2025-08-13 00:26:52 +0000 UTC" firstStartedPulling="2025-08-13 00:26:52.395810616 +0000 UTC m=+20.480767755" lastFinishedPulling="2025-08-13 00:27:01.709912257 +0000 UTC m=+29.794869396" observedRunningTime="2025-08-13 00:27:02.055169781 +0000 UTC m=+30.140126960" watchObservedRunningTime="2025-08-13 00:27:02.055352981 +0000 UTC m=+30.140310120" Aug 13 00:27:02.062366 containerd[2685]: 2025-08-13 00:27:02.016 [INFO][6245] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" Aug 13 00:27:02.062366 containerd[2685]: 2025-08-13 00:27:02.016 [INFO][6245] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" iface="eth0" netns="/var/run/netns/cni-f6cc869b-8f8a-d124-8124-d90577a6bbb1" Aug 13 00:27:02.062366 containerd[2685]: 2025-08-13 00:27:02.016 [INFO][6245] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" iface="eth0" netns="/var/run/netns/cni-f6cc869b-8f8a-d124-8124-d90577a6bbb1" Aug 13 00:27:02.062366 containerd[2685]: 2025-08-13 00:27:02.016 [INFO][6245] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" iface="eth0" netns="/var/run/netns/cni-f6cc869b-8f8a-d124-8124-d90577a6bbb1" Aug 13 00:27:02.062366 containerd[2685]: 2025-08-13 00:27:02.016 [INFO][6245] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" Aug 13 00:27:02.062366 containerd[2685]: 2025-08-13 00:27:02.016 [INFO][6245] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" Aug 13 00:27:02.062366 containerd[2685]: 2025-08-13 00:27:02.050 [INFO][6276] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" HandleID="k8s-pod-network.74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" Workload="ci--4081.3.5--a--588adb827d-k8s-whisker--67db74799b--hr8ld-eth0" Aug 13 00:27:02.062366 containerd[2685]: 2025-08-13 00:27:02.050 [INFO][6276] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:02.062366 containerd[2685]: 2025-08-13 00:27:02.050 [INFO][6276] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:02.062366 containerd[2685]: 2025-08-13 00:27:02.058 [WARNING][6276] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" HandleID="k8s-pod-network.74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" Workload="ci--4081.3.5--a--588adb827d-k8s-whisker--67db74799b--hr8ld-eth0" Aug 13 00:27:02.062366 containerd[2685]: 2025-08-13 00:27:02.058 [INFO][6276] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" HandleID="k8s-pod-network.74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" Workload="ci--4081.3.5--a--588adb827d-k8s-whisker--67db74799b--hr8ld-eth0" Aug 13 00:27:02.062366 containerd[2685]: 2025-08-13 00:27:02.059 [INFO][6276] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:02.062366 containerd[2685]: 2025-08-13 00:27:02.060 [INFO][6245] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" Aug 13 00:27:02.062706 containerd[2685]: time="2025-08-13T00:27:02.062469865Z" level=info msg="TearDown network for sandbox \"74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef\" successfully" Aug 13 00:27:02.062706 containerd[2685]: time="2025-08-13T00:27:02.062493505Z" level=info msg="StopPodSandbox for \"74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef\" returns successfully" Aug 13 00:27:02.096120 kubelet[4148]: I0813 00:27:02.096093 4148 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e7e9845f-f3a1-4760-b84c-8f704f46a611-whisker-ca-bundle\") pod \"e7e9845f-f3a1-4760-b84c-8f704f46a611\" (UID: \"e7e9845f-f3a1-4760-b84c-8f704f46a611\") " Aug 13 00:27:02.096281 kubelet[4148]: I0813 00:27:02.096267 4148 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gm4s4\" (UniqueName: \"kubernetes.io/projected/e7e9845f-f3a1-4760-b84c-8f704f46a611-kube-api-access-gm4s4\") pod \"e7e9845f-f3a1-4760-b84c-8f704f46a611\" (UID: \"e7e9845f-f3a1-4760-b84c-8f704f46a611\") " Aug 13 00:27:02.096745 kubelet[4148]: I0813 00:27:02.096375 4148 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/e7e9845f-f3a1-4760-b84c-8f704f46a611-whisker-backend-key-pair\") pod \"e7e9845f-f3a1-4760-b84c-8f704f46a611\" (UID: \"e7e9845f-f3a1-4760-b84c-8f704f46a611\") " Aug 13 00:27:02.096745 kubelet[4148]: I0813 00:27:02.096535 4148 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e9845f-f3a1-4760-b84c-8f704f46a611-whisker-ca-bundle" (OuterVolumeSpecName: "whisker-ca-bundle") pod "e7e9845f-f3a1-4760-b84c-8f704f46a611" (UID: "e7e9845f-f3a1-4760-b84c-8f704f46a611"). InnerVolumeSpecName "whisker-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Aug 13 00:27:02.098603 kubelet[4148]: I0813 00:27:02.098550 4148 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e9845f-f3a1-4760-b84c-8f704f46a611-kube-api-access-gm4s4" (OuterVolumeSpecName: "kube-api-access-gm4s4") pod "e7e9845f-f3a1-4760-b84c-8f704f46a611" (UID: "e7e9845f-f3a1-4760-b84c-8f704f46a611"). InnerVolumeSpecName "kube-api-access-gm4s4". PluginName "kubernetes.io/projected", VolumeGIDValue "" Aug 13 00:27:02.098675 kubelet[4148]: I0813 00:27:02.098656 4148 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e9845f-f3a1-4760-b84c-8f704f46a611-whisker-backend-key-pair" (OuterVolumeSpecName: "whisker-backend-key-pair") pod "e7e9845f-f3a1-4760-b84c-8f704f46a611" (UID: "e7e9845f-f3a1-4760-b84c-8f704f46a611"). InnerVolumeSpecName "whisker-backend-key-pair". PluginName "kubernetes.io/secret", VolumeGIDValue "" Aug 13 00:27:02.197034 kubelet[4148]: I0813 00:27:02.196980 4148 reconciler_common.go:299] "Volume detached for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e7e9845f-f3a1-4760-b84c-8f704f46a611-whisker-ca-bundle\") on node \"ci-4081.3.5-a-588adb827d\" DevicePath \"\"" Aug 13 00:27:02.197034 kubelet[4148]: I0813 00:27:02.197003 4148 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-gm4s4\" (UniqueName: \"kubernetes.io/projected/e7e9845f-f3a1-4760-b84c-8f704f46a611-kube-api-access-gm4s4\") on node \"ci-4081.3.5-a-588adb827d\" DevicePath \"\"" Aug 13 00:27:02.197034 kubelet[4148]: I0813 00:27:02.197013 4148 reconciler_common.go:299] "Volume detached for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/e7e9845f-f3a1-4760-b84c-8f704f46a611-whisker-backend-key-pair\") on node \"ci-4081.3.5-a-588adb827d\" DevicePath \"\"" Aug 13 00:27:02.685465 systemd[1]: run-netns-cni\x2df6cc869b\x2d8f8a\x2dd124\x2d8124\x2dd90577a6bbb1.mount: Deactivated successfully. Aug 13 00:27:02.685552 systemd[1]: var-lib-kubelet-pods-e7e9845f\x2df3a1\x2d4760\x2db84c\x2d8f704f46a611-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dgm4s4.mount: Deactivated successfully. Aug 13 00:27:02.685606 systemd[1]: var-lib-kubelet-pods-e7e9845f\x2df3a1\x2d4760\x2db84c\x2d8f704f46a611-volumes-kubernetes.io\x7esecret-whisker\x2dbackend\x2dkey\x2dpair.mount: Deactivated successfully. Aug 13 00:27:03.047052 kubelet[4148]: I0813 00:27:03.047027 4148 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Aug 13 00:27:03.050557 systemd[1]: Removed slice kubepods-besteffort-pode7e9845f_f3a1_4760_b84c_8f704f46a611.slice - libcontainer container kubepods-besteffort-pode7e9845f_f3a1_4760_b84c_8f704f46a611.slice. Aug 13 00:27:03.077975 systemd[1]: Created slice kubepods-besteffort-pod63f96dc6_e216_4be0_b317_bc4a0d37d6d4.slice - libcontainer container kubepods-besteffort-pod63f96dc6_e216_4be0_b317_bc4a0d37d6d4.slice. Aug 13 00:27:03.102487 kubelet[4148]: I0813 00:27:03.102447 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-backend-key-pair\" (UniqueName: \"kubernetes.io/secret/63f96dc6-e216-4be0-b317-bc4a0d37d6d4-whisker-backend-key-pair\") pod \"whisker-858bfd4c6-f72fd\" (UID: \"63f96dc6-e216-4be0-b317-bc4a0d37d6d4\") " pod="calico-system/whisker-858bfd4c6-f72fd" Aug 13 00:27:03.102487 kubelet[4148]: I0813 00:27:03.102489 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whisker-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/63f96dc6-e216-4be0-b317-bc4a0d37d6d4-whisker-ca-bundle\") pod \"whisker-858bfd4c6-f72fd\" (UID: \"63f96dc6-e216-4be0-b317-bc4a0d37d6d4\") " pod="calico-system/whisker-858bfd4c6-f72fd" Aug 13 00:27:03.102838 kubelet[4148]: I0813 00:27:03.102508 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8p8q\" (UniqueName: \"kubernetes.io/projected/63f96dc6-e216-4be0-b317-bc4a0d37d6d4-kube-api-access-f8p8q\") pod \"whisker-858bfd4c6-f72fd\" (UID: \"63f96dc6-e216-4be0-b317-bc4a0d37d6d4\") " pod="calico-system/whisker-858bfd4c6-f72fd" Aug 13 00:27:03.187295 kernel: bpftool[6492]: memfd_create() called without MFD_EXEC or MFD_NOEXEC_SEAL set Aug 13 00:27:03.344036 systemd-networkd[2588]: vxlan.calico: Link UP Aug 13 00:27:03.344041 systemd-networkd[2588]: vxlan.calico: Gained carrier Aug 13 00:27:03.380294 containerd[2685]: time="2025-08-13T00:27:03.380258208Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-858bfd4c6-f72fd,Uid:63f96dc6-e216-4be0-b317-bc4a0d37d6d4,Namespace:calico-system,Attempt:0,}" Aug 13 00:27:03.462657 systemd-networkd[2588]: caliaafa66a7136: Link UP Aug 13 00:27:03.463211 systemd-networkd[2588]: caliaafa66a7136: Gained carrier Aug 13 00:27:03.472075 containerd[2685]: 2025-08-13 00:27:03.414 [INFO][6593] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.5--a--588adb827d-k8s-whisker--858bfd4c6--f72fd-eth0 whisker-858bfd4c6- calico-system 63f96dc6-e216-4be0-b317-bc4a0d37d6d4 899 0 2025-08-13 00:27:03 +0000 UTC map[app.kubernetes.io/name:whisker k8s-app:whisker pod-template-hash:858bfd4c6 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:whisker] map[] [] [] []} {k8s ci-4081.3.5-a-588adb827d whisker-858bfd4c6-f72fd eth0 whisker [] [] [kns.calico-system ksa.calico-system.whisker] caliaafa66a7136 [] [] }} ContainerID="5719b814c2924890821526ae7a84c13e364e5f09440cdcf45cab5a2512849955" Namespace="calico-system" Pod="whisker-858bfd4c6-f72fd" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-whisker--858bfd4c6--f72fd-" Aug 13 00:27:03.472075 containerd[2685]: 2025-08-13 00:27:03.414 [INFO][6593] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="5719b814c2924890821526ae7a84c13e364e5f09440cdcf45cab5a2512849955" Namespace="calico-system" Pod="whisker-858bfd4c6-f72fd" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-whisker--858bfd4c6--f72fd-eth0" Aug 13 00:27:03.472075 containerd[2685]: 2025-08-13 00:27:03.434 [INFO][6715] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="5719b814c2924890821526ae7a84c13e364e5f09440cdcf45cab5a2512849955" HandleID="k8s-pod-network.5719b814c2924890821526ae7a84c13e364e5f09440cdcf45cab5a2512849955" Workload="ci--4081.3.5--a--588adb827d-k8s-whisker--858bfd4c6--f72fd-eth0" Aug 13 00:27:03.472075 containerd[2685]: 2025-08-13 00:27:03.434 [INFO][6715] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="5719b814c2924890821526ae7a84c13e364e5f09440cdcf45cab5a2512849955" HandleID="k8s-pod-network.5719b814c2924890821526ae7a84c13e364e5f09440cdcf45cab5a2512849955" Workload="ci--4081.3.5--a--588adb827d-k8s-whisker--858bfd4c6--f72fd-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x400053a800), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081.3.5-a-588adb827d", "pod":"whisker-858bfd4c6-f72fd", "timestamp":"2025-08-13 00:27:03.434048482 +0000 UTC"}, Hostname:"ci-4081.3.5-a-588adb827d", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 00:27:03.472075 containerd[2685]: 2025-08-13 00:27:03.434 [INFO][6715] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:03.472075 containerd[2685]: 2025-08-13 00:27:03.434 [INFO][6715] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:03.472075 containerd[2685]: 2025-08-13 00:27:03.434 [INFO][6715] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.5-a-588adb827d' Aug 13 00:27:03.472075 containerd[2685]: 2025-08-13 00:27:03.442 [INFO][6715] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.5719b814c2924890821526ae7a84c13e364e5f09440cdcf45cab5a2512849955" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:03.472075 containerd[2685]: 2025-08-13 00:27:03.445 [INFO][6715] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:03.472075 containerd[2685]: 2025-08-13 00:27:03.448 [INFO][6715] ipam/ipam.go 511: Trying affinity for 192.168.85.64/26 host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:03.472075 containerd[2685]: 2025-08-13 00:27:03.449 [INFO][6715] ipam/ipam.go 158: Attempting to load block cidr=192.168.85.64/26 host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:03.472075 containerd[2685]: 2025-08-13 00:27:03.451 [INFO][6715] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.85.64/26 host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:03.472075 containerd[2685]: 2025-08-13 00:27:03.451 [INFO][6715] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.85.64/26 handle="k8s-pod-network.5719b814c2924890821526ae7a84c13e364e5f09440cdcf45cab5a2512849955" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:03.472075 containerd[2685]: 2025-08-13 00:27:03.452 [INFO][6715] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.5719b814c2924890821526ae7a84c13e364e5f09440cdcf45cab5a2512849955 Aug 13 00:27:03.472075 containerd[2685]: 2025-08-13 00:27:03.456 [INFO][6715] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.85.64/26 handle="k8s-pod-network.5719b814c2924890821526ae7a84c13e364e5f09440cdcf45cab5a2512849955" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:03.472075 containerd[2685]: 2025-08-13 00:27:03.459 [INFO][6715] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.85.65/26] block=192.168.85.64/26 handle="k8s-pod-network.5719b814c2924890821526ae7a84c13e364e5f09440cdcf45cab5a2512849955" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:03.472075 containerd[2685]: 2025-08-13 00:27:03.459 [INFO][6715] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.85.65/26] handle="k8s-pod-network.5719b814c2924890821526ae7a84c13e364e5f09440cdcf45cab5a2512849955" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:03.472075 containerd[2685]: 2025-08-13 00:27:03.459 [INFO][6715] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:03.472075 containerd[2685]: 2025-08-13 00:27:03.459 [INFO][6715] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.85.65/26] IPv6=[] ContainerID="5719b814c2924890821526ae7a84c13e364e5f09440cdcf45cab5a2512849955" HandleID="k8s-pod-network.5719b814c2924890821526ae7a84c13e364e5f09440cdcf45cab5a2512849955" Workload="ci--4081.3.5--a--588adb827d-k8s-whisker--858bfd4c6--f72fd-eth0" Aug 13 00:27:03.472576 containerd[2685]: 2025-08-13 00:27:03.461 [INFO][6593] cni-plugin/k8s.go 418: Populated endpoint ContainerID="5719b814c2924890821526ae7a84c13e364e5f09440cdcf45cab5a2512849955" Namespace="calico-system" Pod="whisker-858bfd4c6-f72fd" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-whisker--858bfd4c6--f72fd-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--588adb827d-k8s-whisker--858bfd4c6--f72fd-eth0", GenerateName:"whisker-858bfd4c6-", Namespace:"calico-system", SelfLink:"", UID:"63f96dc6-e216-4be0-b317-bc4a0d37d6d4", ResourceVersion:"899", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 27, 3, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"858bfd4c6", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-588adb827d", ContainerID:"", Pod:"whisker-858bfd4c6-f72fd", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.85.65/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"caliaafa66a7136", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:27:03.472576 containerd[2685]: 2025-08-13 00:27:03.461 [INFO][6593] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.85.65/32] ContainerID="5719b814c2924890821526ae7a84c13e364e5f09440cdcf45cab5a2512849955" Namespace="calico-system" Pod="whisker-858bfd4c6-f72fd" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-whisker--858bfd4c6--f72fd-eth0" Aug 13 00:27:03.472576 containerd[2685]: 2025-08-13 00:27:03.461 [INFO][6593] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to caliaafa66a7136 ContainerID="5719b814c2924890821526ae7a84c13e364e5f09440cdcf45cab5a2512849955" Namespace="calico-system" Pod="whisker-858bfd4c6-f72fd" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-whisker--858bfd4c6--f72fd-eth0" Aug 13 00:27:03.472576 containerd[2685]: 2025-08-13 00:27:03.463 [INFO][6593] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="5719b814c2924890821526ae7a84c13e364e5f09440cdcf45cab5a2512849955" Namespace="calico-system" Pod="whisker-858bfd4c6-f72fd" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-whisker--858bfd4c6--f72fd-eth0" Aug 13 00:27:03.472576 containerd[2685]: 2025-08-13 00:27:03.463 [INFO][6593] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="5719b814c2924890821526ae7a84c13e364e5f09440cdcf45cab5a2512849955" Namespace="calico-system" Pod="whisker-858bfd4c6-f72fd" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-whisker--858bfd4c6--f72fd-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--588adb827d-k8s-whisker--858bfd4c6--f72fd-eth0", GenerateName:"whisker-858bfd4c6-", Namespace:"calico-system", SelfLink:"", UID:"63f96dc6-e216-4be0-b317-bc4a0d37d6d4", ResourceVersion:"899", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 27, 3, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"whisker", "k8s-app":"whisker", "pod-template-hash":"858bfd4c6", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"whisker"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-588adb827d", ContainerID:"5719b814c2924890821526ae7a84c13e364e5f09440cdcf45cab5a2512849955", Pod:"whisker-858bfd4c6-f72fd", Endpoint:"eth0", ServiceAccountName:"whisker", IPNetworks:[]string{"192.168.85.65/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.whisker"}, InterfaceName:"caliaafa66a7136", MAC:"22:6e:40:ff:34:2a", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:27:03.472576 containerd[2685]: 2025-08-13 00:27:03.469 [INFO][6593] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="5719b814c2924890821526ae7a84c13e364e5f09440cdcf45cab5a2512849955" Namespace="calico-system" Pod="whisker-858bfd4c6-f72fd" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-whisker--858bfd4c6--f72fd-eth0" Aug 13 00:27:03.484333 containerd[2685]: time="2025-08-13T00:27:03.484274993Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 00:27:03.484333 containerd[2685]: time="2025-08-13T00:27:03.484327913Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 00:27:03.484395 containerd[2685]: time="2025-08-13T00:27:03.484339193Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:27:03.484428 containerd[2685]: time="2025-08-13T00:27:03.484411673Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:27:03.508411 systemd[1]: Started cri-containerd-5719b814c2924890821526ae7a84c13e364e5f09440cdcf45cab5a2512849955.scope - libcontainer container 5719b814c2924890821526ae7a84c13e364e5f09440cdcf45cab5a2512849955. Aug 13 00:27:03.531716 containerd[2685]: time="2025-08-13T00:27:03.531683583Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:whisker-858bfd4c6-f72fd,Uid:63f96dc6-e216-4be0-b317-bc4a0d37d6d4,Namespace:calico-system,Attempt:0,} returns sandbox id \"5719b814c2924890821526ae7a84c13e364e5f09440cdcf45cab5a2512849955\"" Aug 13 00:27:03.532779 containerd[2685]: time="2025-08-13T00:27:03.532756023Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.2\"" Aug 13 00:27:03.990059 kubelet[4148]: I0813 00:27:03.990024 4148 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e9845f-f3a1-4760-b84c-8f704f46a611" path="/var/lib/kubelet/pods/e7e9845f-f3a1-4760-b84c-8f704f46a611/volumes" Aug 13 00:27:04.796583 containerd[2685]: time="2025-08-13T00:27:04.796542582Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:27:04.796983 containerd[2685]: time="2025-08-13T00:27:04.796601142Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker:v3.30.2: active requests=0, bytes read=4605614" Aug 13 00:27:04.798890 containerd[2685]: time="2025-08-13T00:27:04.798837903Z" level=info msg="ImageCreate event name:\"sha256:309942601a9ca6c4e92bcd09162824fef1c137a5c5d92fbbb45be0f29bfd1817\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:27:04.811967 containerd[2685]: time="2025-08-13T00:27:04.811915911Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker@sha256:31346d4524252a3b0d2a1d289c4985b8402b498b5ce82a12e682096ab7446678\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:27:04.812676 containerd[2685]: time="2025-08-13T00:27:04.812578911Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker:v3.30.2\" with image id \"sha256:309942601a9ca6c4e92bcd09162824fef1c137a5c5d92fbbb45be0f29bfd1817\", repo tag \"ghcr.io/flatcar/calico/whisker:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/whisker@sha256:31346d4524252a3b0d2a1d289c4985b8402b498b5ce82a12e682096ab7446678\", size \"5974847\" in 1.279794648s" Aug 13 00:27:04.812676 containerd[2685]: time="2025-08-13T00:27:04.812613991Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker:v3.30.2\" returns image reference \"sha256:309942601a9ca6c4e92bcd09162824fef1c137a5c5d92fbbb45be0f29bfd1817\"" Aug 13 00:27:04.814778 containerd[2685]: time="2025-08-13T00:27:04.814661753Z" level=info msg="CreateContainer within sandbox \"5719b814c2924890821526ae7a84c13e364e5f09440cdcf45cab5a2512849955\" for container &ContainerMetadata{Name:whisker,Attempt:0,}" Aug 13 00:27:04.819292 containerd[2685]: time="2025-08-13T00:27:04.819218235Z" level=info msg="CreateContainer within sandbox \"5719b814c2924890821526ae7a84c13e364e5f09440cdcf45cab5a2512849955\" for &ContainerMetadata{Name:whisker,Attempt:0,} returns container id \"274a24a248b6234e9a7e1be51e32ffc4a933c042b537ae48f9df37408bd9fda4\"" Aug 13 00:27:04.819539 containerd[2685]: time="2025-08-13T00:27:04.819522995Z" level=info msg="StartContainer for \"274a24a248b6234e9a7e1be51e32ffc4a933c042b537ae48f9df37408bd9fda4\"" Aug 13 00:27:04.852340 systemd[1]: Started cri-containerd-274a24a248b6234e9a7e1be51e32ffc4a933c042b537ae48f9df37408bd9fda4.scope - libcontainer container 274a24a248b6234e9a7e1be51e32ffc4a933c042b537ae48f9df37408bd9fda4. Aug 13 00:27:04.854321 systemd-networkd[2588]: vxlan.calico: Gained IPv6LL Aug 13 00:27:04.876454 containerd[2685]: time="2025-08-13T00:27:04.876416349Z" level=info msg="StartContainer for \"274a24a248b6234e9a7e1be51e32ffc4a933c042b537ae48f9df37408bd9fda4\" returns successfully" Aug 13 00:27:04.877177 containerd[2685]: time="2025-08-13T00:27:04.877163269Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\"" Aug 13 00:27:05.238342 systemd-networkd[2588]: caliaafa66a7136: Gained IPv6LL Aug 13 00:27:06.410051 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1994295052.mount: Deactivated successfully. Aug 13 00:27:06.412745 containerd[2685]: time="2025-08-13T00:27:06.412711103Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:27:06.412954 containerd[2685]: time="2025-08-13T00:27:06.412770383Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/whisker-backend:v3.30.2: active requests=0, bytes read=30814581" Aug 13 00:27:06.413528 containerd[2685]: time="2025-08-13T00:27:06.413508943Z" level=info msg="ImageCreate event name:\"sha256:8763d908c0cd23d0e87bc61ce1ba8371b86449688baf955e5eeff7f7d7e101c4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:27:06.415418 containerd[2685]: time="2025-08-13T00:27:06.415393104Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/whisker-backend@sha256:fbf7f21f5aba95930803ad7e7dea8b083220854eae72c2a7c51681c09c5614b5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:27:06.416165 containerd[2685]: time="2025-08-13T00:27:06.416141905Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\" with image id \"sha256:8763d908c0cd23d0e87bc61ce1ba8371b86449688baf955e5eeff7f7d7e101c4\", repo tag \"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/whisker-backend@sha256:fbf7f21f5aba95930803ad7e7dea8b083220854eae72c2a7c51681c09c5614b5\", size \"30814411\" in 1.538954876s" Aug 13 00:27:06.416191 containerd[2685]: time="2025-08-13T00:27:06.416172785Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/whisker-backend:v3.30.2\" returns image reference \"sha256:8763d908c0cd23d0e87bc61ce1ba8371b86449688baf955e5eeff7f7d7e101c4\"" Aug 13 00:27:06.418199 containerd[2685]: time="2025-08-13T00:27:06.418175986Z" level=info msg="CreateContainer within sandbox \"5719b814c2924890821526ae7a84c13e364e5f09440cdcf45cab5a2512849955\" for container &ContainerMetadata{Name:whisker-backend,Attempt:0,}" Aug 13 00:27:06.423120 containerd[2685]: time="2025-08-13T00:27:06.423094028Z" level=info msg="CreateContainer within sandbox \"5719b814c2924890821526ae7a84c13e364e5f09440cdcf45cab5a2512849955\" for &ContainerMetadata{Name:whisker-backend,Attempt:0,} returns container id \"8e0a85e173eff683a79a5488247747815af6d70dc04ac9b66237f69fdf936bc6\"" Aug 13 00:27:06.423487 containerd[2685]: time="2025-08-13T00:27:06.423462989Z" level=info msg="StartContainer for \"8e0a85e173eff683a79a5488247747815af6d70dc04ac9b66237f69fdf936bc6\"" Aug 13 00:27:06.457342 systemd[1]: Started cri-containerd-8e0a85e173eff683a79a5488247747815af6d70dc04ac9b66237f69fdf936bc6.scope - libcontainer container 8e0a85e173eff683a79a5488247747815af6d70dc04ac9b66237f69fdf936bc6. Aug 13 00:27:06.481708 containerd[2685]: time="2025-08-13T00:27:06.481677099Z" level=info msg="StartContainer for \"8e0a85e173eff683a79a5488247747815af6d70dc04ac9b66237f69fdf936bc6\" returns successfully" Aug 13 00:27:07.062542 kubelet[4148]: I0813 00:27:07.062480 4148 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/whisker-858bfd4c6-f72fd" podStartSLOduration=1.178288714 podStartE2EDuration="4.062463876s" podCreationTimestamp="2025-08-13 00:27:03 +0000 UTC" firstStartedPulling="2025-08-13 00:27:03.532553543 +0000 UTC m=+31.617510682" lastFinishedPulling="2025-08-13 00:27:06.416728705 +0000 UTC m=+34.501685844" observedRunningTime="2025-08-13 00:27:07.062210355 +0000 UTC m=+35.147167454" watchObservedRunningTime="2025-08-13 00:27:07.062463876 +0000 UTC m=+35.147421015" Aug 13 00:27:09.990758 containerd[2685]: time="2025-08-13T00:27:09.990713721Z" level=info msg="StopPodSandbox for \"96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce\"" Aug 13 00:27:09.991292 containerd[2685]: time="2025-08-13T00:27:09.990720801Z" level=info msg="StopPodSandbox for \"4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652\"" Aug 13 00:27:10.061877 containerd[2685]: 2025-08-13 00:27:10.032 [INFO][7050] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" Aug 13 00:27:10.061877 containerd[2685]: 2025-08-13 00:27:10.032 [INFO][7050] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" iface="eth0" netns="/var/run/netns/cni-781d23f2-6a4b-82db-4669-4dc0a2025392" Aug 13 00:27:10.061877 containerd[2685]: 2025-08-13 00:27:10.032 [INFO][7050] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" iface="eth0" netns="/var/run/netns/cni-781d23f2-6a4b-82db-4669-4dc0a2025392" Aug 13 00:27:10.061877 containerd[2685]: 2025-08-13 00:27:10.032 [INFO][7050] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" iface="eth0" netns="/var/run/netns/cni-781d23f2-6a4b-82db-4669-4dc0a2025392" Aug 13 00:27:10.061877 containerd[2685]: 2025-08-13 00:27:10.032 [INFO][7050] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" Aug 13 00:27:10.061877 containerd[2685]: 2025-08-13 00:27:10.032 [INFO][7050] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" Aug 13 00:27:10.061877 containerd[2685]: 2025-08-13 00:27:10.050 [INFO][7087] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" HandleID="k8s-pod-network.96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--h5l86-eth0" Aug 13 00:27:10.061877 containerd[2685]: 2025-08-13 00:27:10.050 [INFO][7087] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:10.061877 containerd[2685]: 2025-08-13 00:27:10.050 [INFO][7087] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:10.061877 containerd[2685]: 2025-08-13 00:27:10.057 [WARNING][7087] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" HandleID="k8s-pod-network.96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--h5l86-eth0" Aug 13 00:27:10.061877 containerd[2685]: 2025-08-13 00:27:10.057 [INFO][7087] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" HandleID="k8s-pod-network.96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--h5l86-eth0" Aug 13 00:27:10.061877 containerd[2685]: 2025-08-13 00:27:10.059 [INFO][7087] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:10.061877 containerd[2685]: 2025-08-13 00:27:10.060 [INFO][7050] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" Aug 13 00:27:10.062201 containerd[2685]: time="2025-08-13T00:27:10.062021150Z" level=info msg="TearDown network for sandbox \"96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce\" successfully" Aug 13 00:27:10.062201 containerd[2685]: time="2025-08-13T00:27:10.062048030Z" level=info msg="StopPodSandbox for \"96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce\" returns successfully" Aug 13 00:27:10.062529 containerd[2685]: time="2025-08-13T00:27:10.062504070Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-74488cd4c9-h5l86,Uid:517c2f94-3455-4677-a3ec-3a8771258060,Namespace:calico-apiserver,Attempt:1,}" Aug 13 00:27:10.063991 systemd[1]: run-netns-cni\x2d781d23f2\x2d6a4b\x2d82db\x2d4669\x2d4dc0a2025392.mount: Deactivated successfully. Aug 13 00:27:10.070969 containerd[2685]: 2025-08-13 00:27:10.032 [INFO][7051] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" Aug 13 00:27:10.070969 containerd[2685]: 2025-08-13 00:27:10.032 [INFO][7051] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" iface="eth0" netns="/var/run/netns/cni-60a8c7c8-9438-b2af-c81f-dd2ec335739d" Aug 13 00:27:10.070969 containerd[2685]: 2025-08-13 00:27:10.032 [INFO][7051] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" iface="eth0" netns="/var/run/netns/cni-60a8c7c8-9438-b2af-c81f-dd2ec335739d" Aug 13 00:27:10.070969 containerd[2685]: 2025-08-13 00:27:10.032 [INFO][7051] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" iface="eth0" netns="/var/run/netns/cni-60a8c7c8-9438-b2af-c81f-dd2ec335739d" Aug 13 00:27:10.070969 containerd[2685]: 2025-08-13 00:27:10.032 [INFO][7051] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" Aug 13 00:27:10.070969 containerd[2685]: 2025-08-13 00:27:10.032 [INFO][7051] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" Aug 13 00:27:10.070969 containerd[2685]: 2025-08-13 00:27:10.050 [INFO][7088] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" HandleID="k8s-pod-network.4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" Workload="ci--4081.3.5--a--588adb827d-k8s-csi--node--driver--4wkp8-eth0" Aug 13 00:27:10.070969 containerd[2685]: 2025-08-13 00:27:10.050 [INFO][7088] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:10.070969 containerd[2685]: 2025-08-13 00:27:10.059 [INFO][7088] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:10.070969 containerd[2685]: 2025-08-13 00:27:10.066 [WARNING][7088] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" HandleID="k8s-pod-network.4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" Workload="ci--4081.3.5--a--588adb827d-k8s-csi--node--driver--4wkp8-eth0" Aug 13 00:27:10.070969 containerd[2685]: 2025-08-13 00:27:10.066 [INFO][7088] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" HandleID="k8s-pod-network.4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" Workload="ci--4081.3.5--a--588adb827d-k8s-csi--node--driver--4wkp8-eth0" Aug 13 00:27:10.070969 containerd[2685]: 2025-08-13 00:27:10.067 [INFO][7088] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:10.070969 containerd[2685]: 2025-08-13 00:27:10.069 [INFO][7051] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" Aug 13 00:27:10.071252 containerd[2685]: time="2025-08-13T00:27:10.071104993Z" level=info msg="TearDown network for sandbox \"4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652\" successfully" Aug 13 00:27:10.071252 containerd[2685]: time="2025-08-13T00:27:10.071126073Z" level=info msg="StopPodSandbox for \"4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652\" returns successfully" Aug 13 00:27:10.071573 containerd[2685]: time="2025-08-13T00:27:10.071551754Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-4wkp8,Uid:bb51369a-5ac7-4ac1-b819-64d9c4e312d0,Namespace:calico-system,Attempt:1,}" Aug 13 00:27:10.072776 systemd[1]: run-netns-cni\x2d60a8c7c8\x2d9438\x2db2af\x2dc81f\x2ddd2ec335739d.mount: Deactivated successfully. Aug 13 00:27:10.148291 systemd-networkd[2588]: cali3c1d63c3857: Link UP Aug 13 00:27:10.148507 systemd-networkd[2588]: cali3c1d63c3857: Gained carrier Aug 13 00:27:10.156589 containerd[2685]: 2025-08-13 00:27:10.099 [INFO][7129] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--h5l86-eth0 calico-apiserver-74488cd4c9- calico-apiserver 517c2f94-3455-4677-a3ec-3a8771258060 932 0 2025-08-13 00:26:47 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:74488cd4c9 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4081.3.5-a-588adb827d calico-apiserver-74488cd4c9-h5l86 eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali3c1d63c3857 [] [] }} ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" Namespace="calico-apiserver" Pod="calico-apiserver-74488cd4c9-h5l86" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--h5l86-" Aug 13 00:27:10.156589 containerd[2685]: 2025-08-13 00:27:10.099 [INFO][7129] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" Namespace="calico-apiserver" Pod="calico-apiserver-74488cd4c9-h5l86" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--h5l86-eth0" Aug 13 00:27:10.156589 containerd[2685]: 2025-08-13 00:27:10.119 [INFO][7183] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" HandleID="k8s-pod-network.6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--h5l86-eth0" Aug 13 00:27:10.156589 containerd[2685]: 2025-08-13 00:27:10.119 [INFO][7183] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" HandleID="k8s-pod-network.6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--h5l86-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000363530), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4081.3.5-a-588adb827d", "pod":"calico-apiserver-74488cd4c9-h5l86", "timestamp":"2025-08-13 00:27:10.119812973 +0000 UTC"}, Hostname:"ci-4081.3.5-a-588adb827d", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 00:27:10.156589 containerd[2685]: 2025-08-13 00:27:10.120 [INFO][7183] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:10.156589 containerd[2685]: 2025-08-13 00:27:10.120 [INFO][7183] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:10.156589 containerd[2685]: 2025-08-13 00:27:10.120 [INFO][7183] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.5-a-588adb827d' Aug 13 00:27:10.156589 containerd[2685]: 2025-08-13 00:27:10.128 [INFO][7183] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:10.156589 containerd[2685]: 2025-08-13 00:27:10.131 [INFO][7183] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:10.156589 containerd[2685]: 2025-08-13 00:27:10.134 [INFO][7183] ipam/ipam.go 511: Trying affinity for 192.168.85.64/26 host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:10.156589 containerd[2685]: 2025-08-13 00:27:10.135 [INFO][7183] ipam/ipam.go 158: Attempting to load block cidr=192.168.85.64/26 host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:10.156589 containerd[2685]: 2025-08-13 00:27:10.137 [INFO][7183] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.85.64/26 host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:10.156589 containerd[2685]: 2025-08-13 00:27:10.137 [INFO][7183] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.85.64/26 handle="k8s-pod-network.6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:10.156589 containerd[2685]: 2025-08-13 00:27:10.138 [INFO][7183] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881 Aug 13 00:27:10.156589 containerd[2685]: 2025-08-13 00:27:10.140 [INFO][7183] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.85.64/26 handle="k8s-pod-network.6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:10.156589 containerd[2685]: 2025-08-13 00:27:10.144 [INFO][7183] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.85.66/26] block=192.168.85.64/26 handle="k8s-pod-network.6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:10.156589 containerd[2685]: 2025-08-13 00:27:10.144 [INFO][7183] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.85.66/26] handle="k8s-pod-network.6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:10.156589 containerd[2685]: 2025-08-13 00:27:10.144 [INFO][7183] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:10.156589 containerd[2685]: 2025-08-13 00:27:10.144 [INFO][7183] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.85.66/26] IPv6=[] ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" HandleID="k8s-pod-network.6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--h5l86-eth0" Aug 13 00:27:10.157063 containerd[2685]: 2025-08-13 00:27:10.146 [INFO][7129] cni-plugin/k8s.go 418: Populated endpoint ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" Namespace="calico-apiserver" Pod="calico-apiserver-74488cd4c9-h5l86" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--h5l86-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--h5l86-eth0", GenerateName:"calico-apiserver-74488cd4c9-", Namespace:"calico-apiserver", SelfLink:"", UID:"517c2f94-3455-4677-a3ec-3a8771258060", ResourceVersion:"932", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 26, 47, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"74488cd4c9", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-588adb827d", ContainerID:"", Pod:"calico-apiserver-74488cd4c9-h5l86", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.85.66/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali3c1d63c3857", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:27:10.157063 containerd[2685]: 2025-08-13 00:27:10.146 [INFO][7129] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.85.66/32] ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" Namespace="calico-apiserver" Pod="calico-apiserver-74488cd4c9-h5l86" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--h5l86-eth0" Aug 13 00:27:10.157063 containerd[2685]: 2025-08-13 00:27:10.146 [INFO][7129] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali3c1d63c3857 ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" Namespace="calico-apiserver" Pod="calico-apiserver-74488cd4c9-h5l86" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--h5l86-eth0" Aug 13 00:27:10.157063 containerd[2685]: 2025-08-13 00:27:10.148 [INFO][7129] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" Namespace="calico-apiserver" Pod="calico-apiserver-74488cd4c9-h5l86" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--h5l86-eth0" Aug 13 00:27:10.157063 containerd[2685]: 2025-08-13 00:27:10.148 [INFO][7129] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" Namespace="calico-apiserver" Pod="calico-apiserver-74488cd4c9-h5l86" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--h5l86-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--h5l86-eth0", GenerateName:"calico-apiserver-74488cd4c9-", Namespace:"calico-apiserver", SelfLink:"", UID:"517c2f94-3455-4677-a3ec-3a8771258060", ResourceVersion:"932", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 26, 47, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"74488cd4c9", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-588adb827d", ContainerID:"6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881", Pod:"calico-apiserver-74488cd4c9-h5l86", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.85.66/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali3c1d63c3857", MAC:"5e:2a:0b:f5:b9:4a", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:27:10.157063 containerd[2685]: 2025-08-13 00:27:10.154 [INFO][7129] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" Namespace="calico-apiserver" Pod="calico-apiserver-74488cd4c9-h5l86" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--h5l86-eth0" Aug 13 00:27:10.169405 containerd[2685]: time="2025-08-13T00:27:10.169064432Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 00:27:10.169405 containerd[2685]: time="2025-08-13T00:27:10.169385072Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 00:27:10.169405 containerd[2685]: time="2025-08-13T00:27:10.169397752Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:27:10.169542 containerd[2685]: time="2025-08-13T00:27:10.169473472Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:27:10.195348 systemd[1]: Started cri-containerd-6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881.scope - libcontainer container 6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881. Aug 13 00:27:10.218506 containerd[2685]: time="2025-08-13T00:27:10.218475452Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-74488cd4c9-h5l86,Uid:517c2f94-3455-4677-a3ec-3a8771258060,Namespace:calico-apiserver,Attempt:1,} returns sandbox id \"6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881\"" Aug 13 00:27:10.219559 containerd[2685]: time="2025-08-13T00:27:10.219541212Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.2\"" Aug 13 00:27:10.248829 systemd-networkd[2588]: calie22dd168197: Link UP Aug 13 00:27:10.249034 systemd-networkd[2588]: calie22dd168197: Gained carrier Aug 13 00:27:10.256748 containerd[2685]: 2025-08-13 00:27:10.101 [INFO][7146] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.5--a--588adb827d-k8s-csi--node--driver--4wkp8-eth0 csi-node-driver- calico-system bb51369a-5ac7-4ac1-b819-64d9c4e312d0 933 0 2025-08-13 00:26:52 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:8967bcb6f k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s ci-4081.3.5-a-588adb827d csi-node-driver-4wkp8 eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] calie22dd168197 [] [] }} ContainerID="28c2e0ac93ee78b239c852a9b7b6e26ab8f23ab9c93620b45125367839cf7097" Namespace="calico-system" Pod="csi-node-driver-4wkp8" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-csi--node--driver--4wkp8-" Aug 13 00:27:10.256748 containerd[2685]: 2025-08-13 00:27:10.102 [INFO][7146] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="28c2e0ac93ee78b239c852a9b7b6e26ab8f23ab9c93620b45125367839cf7097" Namespace="calico-system" Pod="csi-node-driver-4wkp8" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-csi--node--driver--4wkp8-eth0" Aug 13 00:27:10.256748 containerd[2685]: 2025-08-13 00:27:10.122 [INFO][7189] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="28c2e0ac93ee78b239c852a9b7b6e26ab8f23ab9c93620b45125367839cf7097" HandleID="k8s-pod-network.28c2e0ac93ee78b239c852a9b7b6e26ab8f23ab9c93620b45125367839cf7097" Workload="ci--4081.3.5--a--588adb827d-k8s-csi--node--driver--4wkp8-eth0" Aug 13 00:27:10.256748 containerd[2685]: 2025-08-13 00:27:10.122 [INFO][7189] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="28c2e0ac93ee78b239c852a9b7b6e26ab8f23ab9c93620b45125367839cf7097" HandleID="k8s-pod-network.28c2e0ac93ee78b239c852a9b7b6e26ab8f23ab9c93620b45125367839cf7097" Workload="ci--4081.3.5--a--588adb827d-k8s-csi--node--driver--4wkp8-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40003e1b80), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081.3.5-a-588adb827d", "pod":"csi-node-driver-4wkp8", "timestamp":"2025-08-13 00:27:10.122160934 +0000 UTC"}, Hostname:"ci-4081.3.5-a-588adb827d", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 00:27:10.256748 containerd[2685]: 2025-08-13 00:27:10.122 [INFO][7189] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:10.256748 containerd[2685]: 2025-08-13 00:27:10.144 [INFO][7189] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:10.256748 containerd[2685]: 2025-08-13 00:27:10.144 [INFO][7189] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.5-a-588adb827d' Aug 13 00:27:10.256748 containerd[2685]: 2025-08-13 00:27:10.229 [INFO][7189] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.28c2e0ac93ee78b239c852a9b7b6e26ab8f23ab9c93620b45125367839cf7097" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:10.256748 containerd[2685]: 2025-08-13 00:27:10.232 [INFO][7189] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:10.256748 containerd[2685]: 2025-08-13 00:27:10.235 [INFO][7189] ipam/ipam.go 511: Trying affinity for 192.168.85.64/26 host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:10.256748 containerd[2685]: 2025-08-13 00:27:10.237 [INFO][7189] ipam/ipam.go 158: Attempting to load block cidr=192.168.85.64/26 host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:10.256748 containerd[2685]: 2025-08-13 00:27:10.238 [INFO][7189] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.85.64/26 host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:10.256748 containerd[2685]: 2025-08-13 00:27:10.238 [INFO][7189] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.85.64/26 handle="k8s-pod-network.28c2e0ac93ee78b239c852a9b7b6e26ab8f23ab9c93620b45125367839cf7097" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:10.256748 containerd[2685]: 2025-08-13 00:27:10.239 [INFO][7189] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.28c2e0ac93ee78b239c852a9b7b6e26ab8f23ab9c93620b45125367839cf7097 Aug 13 00:27:10.256748 containerd[2685]: 2025-08-13 00:27:10.242 [INFO][7189] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.85.64/26 handle="k8s-pod-network.28c2e0ac93ee78b239c852a9b7b6e26ab8f23ab9c93620b45125367839cf7097" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:10.256748 containerd[2685]: 2025-08-13 00:27:10.245 [INFO][7189] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.85.67/26] block=192.168.85.64/26 handle="k8s-pod-network.28c2e0ac93ee78b239c852a9b7b6e26ab8f23ab9c93620b45125367839cf7097" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:10.256748 containerd[2685]: 2025-08-13 00:27:10.245 [INFO][7189] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.85.67/26] handle="k8s-pod-network.28c2e0ac93ee78b239c852a9b7b6e26ab8f23ab9c93620b45125367839cf7097" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:10.256748 containerd[2685]: 2025-08-13 00:27:10.245 [INFO][7189] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:10.256748 containerd[2685]: 2025-08-13 00:27:10.245 [INFO][7189] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.85.67/26] IPv6=[] ContainerID="28c2e0ac93ee78b239c852a9b7b6e26ab8f23ab9c93620b45125367839cf7097" HandleID="k8s-pod-network.28c2e0ac93ee78b239c852a9b7b6e26ab8f23ab9c93620b45125367839cf7097" Workload="ci--4081.3.5--a--588adb827d-k8s-csi--node--driver--4wkp8-eth0" Aug 13 00:27:10.257356 containerd[2685]: 2025-08-13 00:27:10.247 [INFO][7146] cni-plugin/k8s.go 418: Populated endpoint ContainerID="28c2e0ac93ee78b239c852a9b7b6e26ab8f23ab9c93620b45125367839cf7097" Namespace="calico-system" Pod="csi-node-driver-4wkp8" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-csi--node--driver--4wkp8-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--588adb827d-k8s-csi--node--driver--4wkp8-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"bb51369a-5ac7-4ac1-b819-64d9c4e312d0", ResourceVersion:"933", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 26, 52, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"8967bcb6f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-588adb827d", ContainerID:"", Pod:"csi-node-driver-4wkp8", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.85.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calie22dd168197", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:27:10.257356 containerd[2685]: 2025-08-13 00:27:10.247 [INFO][7146] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.85.67/32] ContainerID="28c2e0ac93ee78b239c852a9b7b6e26ab8f23ab9c93620b45125367839cf7097" Namespace="calico-system" Pod="csi-node-driver-4wkp8" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-csi--node--driver--4wkp8-eth0" Aug 13 00:27:10.257356 containerd[2685]: 2025-08-13 00:27:10.247 [INFO][7146] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calie22dd168197 ContainerID="28c2e0ac93ee78b239c852a9b7b6e26ab8f23ab9c93620b45125367839cf7097" Namespace="calico-system" Pod="csi-node-driver-4wkp8" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-csi--node--driver--4wkp8-eth0" Aug 13 00:27:10.257356 containerd[2685]: 2025-08-13 00:27:10.249 [INFO][7146] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="28c2e0ac93ee78b239c852a9b7b6e26ab8f23ab9c93620b45125367839cf7097" Namespace="calico-system" Pod="csi-node-driver-4wkp8" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-csi--node--driver--4wkp8-eth0" Aug 13 00:27:10.257356 containerd[2685]: 2025-08-13 00:27:10.249 [INFO][7146] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="28c2e0ac93ee78b239c852a9b7b6e26ab8f23ab9c93620b45125367839cf7097" Namespace="calico-system" Pod="csi-node-driver-4wkp8" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-csi--node--driver--4wkp8-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--588adb827d-k8s-csi--node--driver--4wkp8-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"bb51369a-5ac7-4ac1-b819-64d9c4e312d0", ResourceVersion:"933", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 26, 52, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"8967bcb6f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-588adb827d", ContainerID:"28c2e0ac93ee78b239c852a9b7b6e26ab8f23ab9c93620b45125367839cf7097", Pod:"csi-node-driver-4wkp8", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.85.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calie22dd168197", MAC:"62:4f:36:ab:a5:7a", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:27:10.257356 containerd[2685]: 2025-08-13 00:27:10.255 [INFO][7146] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="28c2e0ac93ee78b239c852a9b7b6e26ab8f23ab9c93620b45125367839cf7097" Namespace="calico-system" Pod="csi-node-driver-4wkp8" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-csi--node--driver--4wkp8-eth0" Aug 13 00:27:10.268698 containerd[2685]: time="2025-08-13T00:27:10.268357752Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 00:27:10.268730 containerd[2685]: time="2025-08-13T00:27:10.268691552Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 00:27:10.268730 containerd[2685]: time="2025-08-13T00:27:10.268704872Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:27:10.268806 containerd[2685]: time="2025-08-13T00:27:10.268786872Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:27:10.299404 systemd[1]: Started cri-containerd-28c2e0ac93ee78b239c852a9b7b6e26ab8f23ab9c93620b45125367839cf7097.scope - libcontainer container 28c2e0ac93ee78b239c852a9b7b6e26ab8f23ab9c93620b45125367839cf7097. Aug 13 00:27:10.315906 containerd[2685]: time="2025-08-13T00:27:10.315876891Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-4wkp8,Uid:bb51369a-5ac7-4ac1-b819-64d9c4e312d0,Namespace:calico-system,Attempt:1,} returns sandbox id \"28c2e0ac93ee78b239c852a9b7b6e26ab8f23ab9c93620b45125367839cf7097\"" Aug 13 00:27:11.254335 systemd-networkd[2588]: cali3c1d63c3857: Gained IPv6LL Aug 13 00:27:11.510304 systemd-networkd[2588]: calie22dd168197: Gained IPv6LL Aug 13 00:27:11.989104 containerd[2685]: time="2025-08-13T00:27:11.989057532Z" level=info msg="StopPodSandbox for \"2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8\"" Aug 13 00:27:11.999088 containerd[2685]: time="2025-08-13T00:27:11.999056775Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:27:11.999136 containerd[2685]: time="2025-08-13T00:27:11.999110455Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.30.2: active requests=0, bytes read=44517149" Aug 13 00:27:11.999839 containerd[2685]: time="2025-08-13T00:27:11.999816376Z" level=info msg="ImageCreate event name:\"sha256:3371ea1b18040228ef58c964e49b96f4291def748753dfbc0aef87a55f906b8f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:27:12.001683 containerd[2685]: time="2025-08-13T00:27:12.001659856Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:ec6b10660962e7caad70c47755049fad68f9fc2f7064e8bc7cb862583e02cc2b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:27:12.002389 containerd[2685]: time="2025-08-13T00:27:12.002364577Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.30.2\" with image id \"sha256:3371ea1b18040228ef58c964e49b96f4291def748753dfbc0aef87a55f906b8f\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:ec6b10660962e7caad70c47755049fad68f9fc2f7064e8bc7cb862583e02cc2b\", size \"45886406\" in 1.782793365s" Aug 13 00:27:12.002415 containerd[2685]: time="2025-08-13T00:27:12.002395377Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.30.2\" returns image reference \"sha256:3371ea1b18040228ef58c964e49b96f4291def748753dfbc0aef87a55f906b8f\"" Aug 13 00:27:12.003200 containerd[2685]: time="2025-08-13T00:27:12.003180977Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.2\"" Aug 13 00:27:12.004476 containerd[2685]: time="2025-08-13T00:27:12.004454577Z" level=info msg="CreateContainer within sandbox \"6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Aug 13 00:27:12.022646 containerd[2685]: time="2025-08-13T00:27:12.022614944Z" level=info msg="CreateContainer within sandbox \"6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"05d1a463839387260db0c040a26786524b1847fa73ef06c63092a8f10b27a4bd\"" Aug 13 00:27:12.022942 containerd[2685]: time="2025-08-13T00:27:12.022917984Z" level=info msg="StartContainer for \"05d1a463839387260db0c040a26786524b1847fa73ef06c63092a8f10b27a4bd\"" Aug 13 00:27:12.054371 containerd[2685]: 2025-08-13 00:27:12.025 [INFO][7356] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" Aug 13 00:27:12.054371 containerd[2685]: 2025-08-13 00:27:12.025 [INFO][7356] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" iface="eth0" netns="/var/run/netns/cni-250e5fee-017e-ba86-9ef5-6c9881800851" Aug 13 00:27:12.054371 containerd[2685]: 2025-08-13 00:27:12.025 [INFO][7356] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" iface="eth0" netns="/var/run/netns/cni-250e5fee-017e-ba86-9ef5-6c9881800851" Aug 13 00:27:12.054371 containerd[2685]: 2025-08-13 00:27:12.026 [INFO][7356] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" iface="eth0" netns="/var/run/netns/cni-250e5fee-017e-ba86-9ef5-6c9881800851" Aug 13 00:27:12.054371 containerd[2685]: 2025-08-13 00:27:12.026 [INFO][7356] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" Aug 13 00:27:12.054371 containerd[2685]: 2025-08-13 00:27:12.026 [INFO][7356] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" Aug 13 00:27:12.054371 containerd[2685]: 2025-08-13 00:27:12.042 [INFO][7383] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" HandleID="k8s-pod-network.2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--4dnl6-eth0" Aug 13 00:27:12.054371 containerd[2685]: 2025-08-13 00:27:12.043 [INFO][7383] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:12.054371 containerd[2685]: 2025-08-13 00:27:12.043 [INFO][7383] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:12.054371 containerd[2685]: 2025-08-13 00:27:12.050 [WARNING][7383] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" HandleID="k8s-pod-network.2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--4dnl6-eth0" Aug 13 00:27:12.054371 containerd[2685]: 2025-08-13 00:27:12.050 [INFO][7383] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" HandleID="k8s-pod-network.2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--4dnl6-eth0" Aug 13 00:27:12.054371 containerd[2685]: 2025-08-13 00:27:12.051 [INFO][7383] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:12.054371 containerd[2685]: 2025-08-13 00:27:12.053 [INFO][7356] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" Aug 13 00:27:12.054823 containerd[2685]: time="2025-08-13T00:27:12.054526715Z" level=info msg="TearDown network for sandbox \"2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8\" successfully" Aug 13 00:27:12.054823 containerd[2685]: time="2025-08-13T00:27:12.054547515Z" level=info msg="StopPodSandbox for \"2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8\" returns successfully" Aug 13 00:27:12.054962 containerd[2685]: time="2025-08-13T00:27:12.054934795Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-74488cd4c9-4dnl6,Uid:0ced7fba-ab7f-4d92-941c-7095a5aea251,Namespace:calico-apiserver,Attempt:1,}" Aug 13 00:27:12.061341 systemd[1]: Started cri-containerd-05d1a463839387260db0c040a26786524b1847fa73ef06c63092a8f10b27a4bd.scope - libcontainer container 05d1a463839387260db0c040a26786524b1847fa73ef06c63092a8f10b27a4bd. Aug 13 00:27:12.065212 systemd[1]: run-netns-cni\x2d250e5fee\x2d017e\x2dba86\x2d9ef5\x2d6c9881800851.mount: Deactivated successfully. Aug 13 00:27:12.085220 containerd[2685]: time="2025-08-13T00:27:12.085191526Z" level=info msg="StartContainer for \"05d1a463839387260db0c040a26786524b1847fa73ef06c63092a8f10b27a4bd\" returns successfully" Aug 13 00:27:12.134373 systemd-networkd[2588]: cali650bf866543: Link UP Aug 13 00:27:12.134608 systemd-networkd[2588]: cali650bf866543: Gained carrier Aug 13 00:27:12.141859 containerd[2685]: 2025-08-13 00:27:12.087 [INFO][7415] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--4dnl6-eth0 calico-apiserver-74488cd4c9- calico-apiserver 0ced7fba-ab7f-4d92-941c-7095a5aea251 950 0 2025-08-13 00:26:47 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:74488cd4c9 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4081.3.5-a-588adb827d calico-apiserver-74488cd4c9-4dnl6 eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali650bf866543 [] [] }} ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" Namespace="calico-apiserver" Pod="calico-apiserver-74488cd4c9-4dnl6" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--4dnl6-" Aug 13 00:27:12.141859 containerd[2685]: 2025-08-13 00:27:12.087 [INFO][7415] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" Namespace="calico-apiserver" Pod="calico-apiserver-74488cd4c9-4dnl6" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--4dnl6-eth0" Aug 13 00:27:12.141859 containerd[2685]: 2025-08-13 00:27:12.107 [INFO][7462] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" HandleID="k8s-pod-network.812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--4dnl6-eth0" Aug 13 00:27:12.141859 containerd[2685]: 2025-08-13 00:27:12.107 [INFO][7462] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" HandleID="k8s-pod-network.812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--4dnl6-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000736290), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4081.3.5-a-588adb827d", "pod":"calico-apiserver-74488cd4c9-4dnl6", "timestamp":"2025-08-13 00:27:12.107531093 +0000 UTC"}, Hostname:"ci-4081.3.5-a-588adb827d", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 00:27:12.141859 containerd[2685]: 2025-08-13 00:27:12.107 [INFO][7462] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:12.141859 containerd[2685]: 2025-08-13 00:27:12.107 [INFO][7462] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:12.141859 containerd[2685]: 2025-08-13 00:27:12.107 [INFO][7462] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.5-a-588adb827d' Aug 13 00:27:12.141859 containerd[2685]: 2025-08-13 00:27:12.116 [INFO][7462] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:12.141859 containerd[2685]: 2025-08-13 00:27:12.119 [INFO][7462] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:12.141859 containerd[2685]: 2025-08-13 00:27:12.122 [INFO][7462] ipam/ipam.go 511: Trying affinity for 192.168.85.64/26 host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:12.141859 containerd[2685]: 2025-08-13 00:27:12.123 [INFO][7462] ipam/ipam.go 158: Attempting to load block cidr=192.168.85.64/26 host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:12.141859 containerd[2685]: 2025-08-13 00:27:12.124 [INFO][7462] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.85.64/26 host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:12.141859 containerd[2685]: 2025-08-13 00:27:12.124 [INFO][7462] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.85.64/26 handle="k8s-pod-network.812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:12.141859 containerd[2685]: 2025-08-13 00:27:12.125 [INFO][7462] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6 Aug 13 00:27:12.141859 containerd[2685]: 2025-08-13 00:27:12.127 [INFO][7462] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.85.64/26 handle="k8s-pod-network.812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:12.141859 containerd[2685]: 2025-08-13 00:27:12.131 [INFO][7462] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.85.68/26] block=192.168.85.64/26 handle="k8s-pod-network.812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:12.141859 containerd[2685]: 2025-08-13 00:27:12.131 [INFO][7462] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.85.68/26] handle="k8s-pod-network.812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:12.141859 containerd[2685]: 2025-08-13 00:27:12.131 [INFO][7462] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:12.141859 containerd[2685]: 2025-08-13 00:27:12.131 [INFO][7462] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.85.68/26] IPv6=[] ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" HandleID="k8s-pod-network.812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--4dnl6-eth0" Aug 13 00:27:12.142313 containerd[2685]: 2025-08-13 00:27:12.132 [INFO][7415] cni-plugin/k8s.go 418: Populated endpoint ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" Namespace="calico-apiserver" Pod="calico-apiserver-74488cd4c9-4dnl6" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--4dnl6-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--4dnl6-eth0", GenerateName:"calico-apiserver-74488cd4c9-", Namespace:"calico-apiserver", SelfLink:"", UID:"0ced7fba-ab7f-4d92-941c-7095a5aea251", ResourceVersion:"950", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 26, 47, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"74488cd4c9", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-588adb827d", ContainerID:"", Pod:"calico-apiserver-74488cd4c9-4dnl6", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.85.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali650bf866543", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:27:12.142313 containerd[2685]: 2025-08-13 00:27:12.133 [INFO][7415] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.85.68/32] ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" Namespace="calico-apiserver" Pod="calico-apiserver-74488cd4c9-4dnl6" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--4dnl6-eth0" Aug 13 00:27:12.142313 containerd[2685]: 2025-08-13 00:27:12.133 [INFO][7415] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali650bf866543 ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" Namespace="calico-apiserver" Pod="calico-apiserver-74488cd4c9-4dnl6" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--4dnl6-eth0" Aug 13 00:27:12.142313 containerd[2685]: 2025-08-13 00:27:12.134 [INFO][7415] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" Namespace="calico-apiserver" Pod="calico-apiserver-74488cd4c9-4dnl6" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--4dnl6-eth0" Aug 13 00:27:12.142313 containerd[2685]: 2025-08-13 00:27:12.134 [INFO][7415] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" Namespace="calico-apiserver" Pod="calico-apiserver-74488cd4c9-4dnl6" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--4dnl6-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--4dnl6-eth0", GenerateName:"calico-apiserver-74488cd4c9-", Namespace:"calico-apiserver", SelfLink:"", UID:"0ced7fba-ab7f-4d92-941c-7095a5aea251", ResourceVersion:"950", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 26, 47, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"74488cd4c9", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-588adb827d", ContainerID:"812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6", Pod:"calico-apiserver-74488cd4c9-4dnl6", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.85.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali650bf866543", MAC:"06:b8:2f:b4:a5:bd", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:27:12.142313 containerd[2685]: 2025-08-13 00:27:12.140 [INFO][7415] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" Namespace="calico-apiserver" Pod="calico-apiserver-74488cd4c9-4dnl6" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--4dnl6-eth0" Aug 13 00:27:12.158068 containerd[2685]: time="2025-08-13T00:27:12.157955111Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 00:27:12.158092 containerd[2685]: time="2025-08-13T00:27:12.158063391Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 00:27:12.158092 containerd[2685]: time="2025-08-13T00:27:12.158076071Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:27:12.158169 containerd[2685]: time="2025-08-13T00:27:12.158149831Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:27:12.185406 systemd[1]: Started cri-containerd-812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6.scope - libcontainer container 812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6. Aug 13 00:27:12.208747 containerd[2685]: time="2025-08-13T00:27:12.208714889Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-74488cd4c9-4dnl6,Uid:0ced7fba-ab7f-4d92-941c-7095a5aea251,Namespace:calico-apiserver,Attempt:1,} returns sandbox id \"812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6\"" Aug 13 00:27:12.211175 containerd[2685]: time="2025-08-13T00:27:12.211145930Z" level=info msg="CreateContainer within sandbox \"812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Aug 13 00:27:12.215859 containerd[2685]: time="2025-08-13T00:27:12.215830571Z" level=info msg="CreateContainer within sandbox \"812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"7d1057fb03c5afd1c6c40fbde2f435dd8914ced368b563fcf13dc6e7b062cd11\"" Aug 13 00:27:12.216195 containerd[2685]: time="2025-08-13T00:27:12.216174571Z" level=info msg="StartContainer for \"7d1057fb03c5afd1c6c40fbde2f435dd8914ced368b563fcf13dc6e7b062cd11\"" Aug 13 00:27:12.247418 systemd[1]: Started cri-containerd-7d1057fb03c5afd1c6c40fbde2f435dd8914ced368b563fcf13dc6e7b062cd11.scope - libcontainer container 7d1057fb03c5afd1c6c40fbde2f435dd8914ced368b563fcf13dc6e7b062cd11. Aug 13 00:27:12.276427 containerd[2685]: time="2025-08-13T00:27:12.276398192Z" level=info msg="StartContainer for \"7d1057fb03c5afd1c6c40fbde2f435dd8914ced368b563fcf13dc6e7b062cd11\" returns successfully" Aug 13 00:27:12.989326 containerd[2685]: time="2025-08-13T00:27:12.989021441Z" level=info msg="StopPodSandbox for \"3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd\"" Aug 13 00:27:12.989326 containerd[2685]: time="2025-08-13T00:27:12.989059042Z" level=info msg="StopPodSandbox for \"e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f\"" Aug 13 00:27:12.989326 containerd[2685]: time="2025-08-13T00:27:12.989115482Z" level=info msg="StopPodSandbox for \"d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3\"" Aug 13 00:27:12.989326 containerd[2685]: time="2025-08-13T00:27:12.989275402Z" level=info msg="StopPodSandbox for \"fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02\"" Aug 13 00:27:13.061409 containerd[2685]: 2025-08-13 00:27:13.028 [INFO][7681] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" Aug 13 00:27:13.061409 containerd[2685]: 2025-08-13 00:27:13.028 [INFO][7681] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" iface="eth0" netns="/var/run/netns/cni-9002d6f5-fa6f-7448-ad2c-4fcf204577f7" Aug 13 00:27:13.061409 containerd[2685]: 2025-08-13 00:27:13.029 [INFO][7681] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" iface="eth0" netns="/var/run/netns/cni-9002d6f5-fa6f-7448-ad2c-4fcf204577f7" Aug 13 00:27:13.061409 containerd[2685]: 2025-08-13 00:27:13.029 [INFO][7681] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" iface="eth0" netns="/var/run/netns/cni-9002d6f5-fa6f-7448-ad2c-4fcf204577f7" Aug 13 00:27:13.061409 containerd[2685]: 2025-08-13 00:27:13.029 [INFO][7681] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" Aug 13 00:27:13.061409 containerd[2685]: 2025-08-13 00:27:13.029 [INFO][7681] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" Aug 13 00:27:13.061409 containerd[2685]: 2025-08-13 00:27:13.047 [INFO][7764] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" HandleID="k8s-pod-network.3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" Workload="ci--4081.3.5--a--588adb827d-k8s-goldmane--768f4c5c69--nm6vx-eth0" Aug 13 00:27:13.061409 containerd[2685]: 2025-08-13 00:27:13.048 [INFO][7764] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:13.061409 containerd[2685]: 2025-08-13 00:27:13.048 [INFO][7764] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:13.061409 containerd[2685]: 2025-08-13 00:27:13.055 [WARNING][7764] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" HandleID="k8s-pod-network.3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" Workload="ci--4081.3.5--a--588adb827d-k8s-goldmane--768f4c5c69--nm6vx-eth0" Aug 13 00:27:13.061409 containerd[2685]: 2025-08-13 00:27:13.055 [INFO][7764] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" HandleID="k8s-pod-network.3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" Workload="ci--4081.3.5--a--588adb827d-k8s-goldmane--768f4c5c69--nm6vx-eth0" Aug 13 00:27:13.061409 containerd[2685]: 2025-08-13 00:27:13.056 [INFO][7764] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:13.061409 containerd[2685]: 2025-08-13 00:27:13.058 [INFO][7681] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" Aug 13 00:27:13.061841 containerd[2685]: time="2025-08-13T00:27:13.061810106Z" level=info msg="TearDown network for sandbox \"3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd\" successfully" Aug 13 00:27:13.061896 containerd[2685]: time="2025-08-13T00:27:13.061884586Z" level=info msg="StopPodSandbox for \"3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd\" returns successfully" Aug 13 00:27:13.062667 containerd[2685]: time="2025-08-13T00:27:13.062634266Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-768f4c5c69-nm6vx,Uid:b5da64a2-733d-4c0b-b26c-80dcab04992c,Namespace:calico-system,Attempt:1,}" Aug 13 00:27:13.065041 systemd[1]: run-netns-cni\x2d9002d6f5\x2dfa6f\x2d7448\x2dad2c\x2d4fcf204577f7.mount: Deactivated successfully. Aug 13 00:27:13.074905 containerd[2685]: 2025-08-13 00:27:13.035 [INFO][7682] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" Aug 13 00:27:13.074905 containerd[2685]: 2025-08-13 00:27:13.035 [INFO][7682] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" iface="eth0" netns="/var/run/netns/cni-4277d4de-8591-827e-7ab5-e91518bfe27d" Aug 13 00:27:13.074905 containerd[2685]: 2025-08-13 00:27:13.035 [INFO][7682] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" iface="eth0" netns="/var/run/netns/cni-4277d4de-8591-827e-7ab5-e91518bfe27d" Aug 13 00:27:13.074905 containerd[2685]: 2025-08-13 00:27:13.036 [INFO][7682] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" iface="eth0" netns="/var/run/netns/cni-4277d4de-8591-827e-7ab5-e91518bfe27d" Aug 13 00:27:13.074905 containerd[2685]: 2025-08-13 00:27:13.036 [INFO][7682] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" Aug 13 00:27:13.074905 containerd[2685]: 2025-08-13 00:27:13.036 [INFO][7682] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" Aug 13 00:27:13.074905 containerd[2685]: 2025-08-13 00:27:13.054 [INFO][7770] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" HandleID="k8s-pod-network.fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--kube--controllers--f5cbb957c--wms9w-eth0" Aug 13 00:27:13.074905 containerd[2685]: 2025-08-13 00:27:13.054 [INFO][7770] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:13.074905 containerd[2685]: 2025-08-13 00:27:13.056 [INFO][7770] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:13.074905 containerd[2685]: 2025-08-13 00:27:13.064 [WARNING][7770] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" HandleID="k8s-pod-network.fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--kube--controllers--f5cbb957c--wms9w-eth0" Aug 13 00:27:13.074905 containerd[2685]: 2025-08-13 00:27:13.064 [INFO][7770] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" HandleID="k8s-pod-network.fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--kube--controllers--f5cbb957c--wms9w-eth0" Aug 13 00:27:13.074905 containerd[2685]: 2025-08-13 00:27:13.066 [INFO][7770] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:13.074905 containerd[2685]: 2025-08-13 00:27:13.073 [INFO][7682] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" Aug 13 00:27:13.075404 containerd[2685]: time="2025-08-13T00:27:13.075379350Z" level=info msg="TearDown network for sandbox \"fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02\" successfully" Aug 13 00:27:13.075478 containerd[2685]: time="2025-08-13T00:27:13.075465150Z" level=info msg="StopPodSandbox for \"fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02\" returns successfully" Aug 13 00:27:13.076223 containerd[2685]: time="2025-08-13T00:27:13.075923830Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-f5cbb957c-wms9w,Uid:925e0d95-7160-42b1-a3c8-6b7795ca67bc,Namespace:calico-system,Attempt:1,}" Aug 13 00:27:13.077201 systemd[1]: run-netns-cni\x2d4277d4de\x2d8591\x2d827e\x2d7ab5\x2de91518bfe27d.mount: Deactivated successfully. Aug 13 00:27:13.078533 containerd[2685]: 2025-08-13 00:27:13.037 [INFO][7688] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" Aug 13 00:27:13.078533 containerd[2685]: 2025-08-13 00:27:13.038 [INFO][7688] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" iface="eth0" netns="/var/run/netns/cni-4b2c6991-41fc-d1a2-5a73-7ec5ca37d349" Aug 13 00:27:13.078533 containerd[2685]: 2025-08-13 00:27:13.038 [INFO][7688] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" iface="eth0" netns="/var/run/netns/cni-4b2c6991-41fc-d1a2-5a73-7ec5ca37d349" Aug 13 00:27:13.078533 containerd[2685]: 2025-08-13 00:27:13.038 [INFO][7688] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" iface="eth0" netns="/var/run/netns/cni-4b2c6991-41fc-d1a2-5a73-7ec5ca37d349" Aug 13 00:27:13.078533 containerd[2685]: 2025-08-13 00:27:13.038 [INFO][7688] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" Aug 13 00:27:13.078533 containerd[2685]: 2025-08-13 00:27:13.038 [INFO][7688] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" Aug 13 00:27:13.078533 containerd[2685]: 2025-08-13 00:27:13.055 [INFO][7782] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" HandleID="k8s-pod-network.e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" Workload="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--v7mtb-eth0" Aug 13 00:27:13.078533 containerd[2685]: 2025-08-13 00:27:13.055 [INFO][7782] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:13.078533 containerd[2685]: 2025-08-13 00:27:13.066 [INFO][7782] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:13.078533 containerd[2685]: 2025-08-13 00:27:13.073 [WARNING][7782] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" HandleID="k8s-pod-network.e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" Workload="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--v7mtb-eth0" Aug 13 00:27:13.078533 containerd[2685]: 2025-08-13 00:27:13.073 [INFO][7782] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" HandleID="k8s-pod-network.e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" Workload="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--v7mtb-eth0" Aug 13 00:27:13.078533 containerd[2685]: 2025-08-13 00:27:13.074 [INFO][7782] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:13.078533 containerd[2685]: 2025-08-13 00:27:13.075 [INFO][7688] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" Aug 13 00:27:13.078834 containerd[2685]: time="2025-08-13T00:27:13.078698391Z" level=info msg="TearDown network for sandbox \"e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f\" successfully" Aug 13 00:27:13.078834 containerd[2685]: time="2025-08-13T00:27:13.078720751Z" level=info msg="StopPodSandbox for \"e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f\" returns successfully" Aug 13 00:27:13.079158 containerd[2685]: time="2025-08-13T00:27:13.079129991Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-v7mtb,Uid:75d28fd6-77d6-4c58-8897-8ce0729a65ea,Namespace:kube-system,Attempt:1,}" Aug 13 00:27:13.081399 systemd[1]: run-netns-cni\x2d4b2c6991\x2d41fc\x2dd1a2\x2d5a73\x2d7ec5ca37d349.mount: Deactivated successfully. Aug 13 00:27:13.087046 kubelet[4148]: I0813 00:27:13.087000 4148 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-74488cd4c9-4dnl6" podStartSLOduration=26.086981514 podStartE2EDuration="26.086981514s" podCreationTimestamp="2025-08-13 00:26:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 00:27:13.086314114 +0000 UTC m=+41.171271253" watchObservedRunningTime="2025-08-13 00:27:13.086981514 +0000 UTC m=+41.171938613" Aug 13 00:27:13.093111 containerd[2685]: 2025-08-13 00:27:13.038 [INFO][7687] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" Aug 13 00:27:13.093111 containerd[2685]: 2025-08-13 00:27:13.038 [INFO][7687] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" iface="eth0" netns="/var/run/netns/cni-50519924-17e8-677e-9bef-2e4d7828d6ae" Aug 13 00:27:13.093111 containerd[2685]: 2025-08-13 00:27:13.038 [INFO][7687] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" iface="eth0" netns="/var/run/netns/cni-50519924-17e8-677e-9bef-2e4d7828d6ae" Aug 13 00:27:13.093111 containerd[2685]: 2025-08-13 00:27:13.039 [INFO][7687] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" iface="eth0" netns="/var/run/netns/cni-50519924-17e8-677e-9bef-2e4d7828d6ae" Aug 13 00:27:13.093111 containerd[2685]: 2025-08-13 00:27:13.039 [INFO][7687] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" Aug 13 00:27:13.093111 containerd[2685]: 2025-08-13 00:27:13.039 [INFO][7687] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" Aug 13 00:27:13.093111 containerd[2685]: 2025-08-13 00:27:13.055 [INFO][7784] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" HandleID="k8s-pod-network.d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" Workload="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--ftxjx-eth0" Aug 13 00:27:13.093111 containerd[2685]: 2025-08-13 00:27:13.055 [INFO][7784] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:13.093111 containerd[2685]: 2025-08-13 00:27:13.074 [INFO][7784] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:13.093111 containerd[2685]: 2025-08-13 00:27:13.086 [WARNING][7784] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" HandleID="k8s-pod-network.d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" Workload="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--ftxjx-eth0" Aug 13 00:27:13.093111 containerd[2685]: 2025-08-13 00:27:13.086 [INFO][7784] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" HandleID="k8s-pod-network.d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" Workload="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--ftxjx-eth0" Aug 13 00:27:13.093111 containerd[2685]: 2025-08-13 00:27:13.090 [INFO][7784] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:13.093111 containerd[2685]: 2025-08-13 00:27:13.091 [INFO][7687] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" Aug 13 00:27:13.093438 containerd[2685]: time="2025-08-13T00:27:13.093261876Z" level=info msg="TearDown network for sandbox \"d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3\" successfully" Aug 13 00:27:13.093438 containerd[2685]: time="2025-08-13T00:27:13.093287036Z" level=info msg="StopPodSandbox for \"d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3\" returns successfully" Aug 13 00:27:13.094014 containerd[2685]: time="2025-08-13T00:27:13.093993316Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-ftxjx,Uid:085d2185-0930-470d-b50d-68703ac546fb,Namespace:kube-system,Attempt:1,}" Aug 13 00:27:13.129206 containerd[2685]: time="2025-08-13T00:27:13.129170288Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:27:13.129323 containerd[2685]: time="2025-08-13T00:27:13.129234048Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.30.2: active requests=0, bytes read=8225702" Aug 13 00:27:13.129999 containerd[2685]: time="2025-08-13T00:27:13.129979528Z" level=info msg="ImageCreate event name:\"sha256:14ecfabbdbebd1f5a36708f8b11a95a43baddd6a935d7d78c89a9c333849fcd2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:27:13.131843 containerd[2685]: time="2025-08-13T00:27:13.131822849Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:e570128aa8067a2f06b96d3cc98afa2e0a4b9790b435ee36ca051c8e72aeb8d0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:27:13.132576 containerd[2685]: time="2025-08-13T00:27:13.132543849Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.30.2\" with image id \"sha256:14ecfabbdbebd1f5a36708f8b11a95a43baddd6a935d7d78c89a9c333849fcd2\", repo tag \"ghcr.io/flatcar/calico/csi:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:e570128aa8067a2f06b96d3cc98afa2e0a4b9790b435ee36ca051c8e72aeb8d0\", size \"9594943\" in 1.129331312s" Aug 13 00:27:13.132611 containerd[2685]: time="2025-08-13T00:27:13.132580609Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.30.2\" returns image reference \"sha256:14ecfabbdbebd1f5a36708f8b11a95a43baddd6a935d7d78c89a9c333849fcd2\"" Aug 13 00:27:13.134701 containerd[2685]: time="2025-08-13T00:27:13.134675610Z" level=info msg="CreateContainer within sandbox \"28c2e0ac93ee78b239c852a9b7b6e26ab8f23ab9c93620b45125367839cf7097\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Aug 13 00:27:13.143561 containerd[2685]: time="2025-08-13T00:27:13.143530492Z" level=info msg="CreateContainer within sandbox \"28c2e0ac93ee78b239c852a9b7b6e26ab8f23ab9c93620b45125367839cf7097\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"e3c9f256fa1183a6b323163f1781eb77ec64c4338cf17fb68dc4c1f29f8344bf\"" Aug 13 00:27:13.143926 containerd[2685]: time="2025-08-13T00:27:13.143901893Z" level=info msg="StartContainer for \"e3c9f256fa1183a6b323163f1781eb77ec64c4338cf17fb68dc4c1f29f8344bf\"" Aug 13 00:27:13.155346 systemd-networkd[2588]: calicdc52df372f: Link UP Aug 13 00:27:13.155567 systemd-networkd[2588]: calicdc52df372f: Gained carrier Aug 13 00:27:13.162874 kubelet[4148]: I0813 00:27:13.162829 4148 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-74488cd4c9-h5l86" podStartSLOduration=24.379076654 podStartE2EDuration="26.162815459s" podCreationTimestamp="2025-08-13 00:26:47 +0000 UTC" firstStartedPulling="2025-08-13 00:27:10.219340892 +0000 UTC m=+38.304298031" lastFinishedPulling="2025-08-13 00:27:12.003079697 +0000 UTC m=+40.088036836" observedRunningTime="2025-08-13 00:27:13.095085037 +0000 UTC m=+41.180042176" watchObservedRunningTime="2025-08-13 00:27:13.162815459 +0000 UTC m=+41.247772598" Aug 13 00:27:13.163733 containerd[2685]: 2025-08-13 00:27:13.106 [INFO][7847] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.5--a--588adb827d-k8s-goldmane--768f4c5c69--nm6vx-eth0 goldmane-768f4c5c69- calico-system b5da64a2-733d-4c0b-b26c-80dcab04992c 964 0 2025-08-13 00:26:51 +0000 UTC map[app.kubernetes.io/name:goldmane k8s-app:goldmane pod-template-hash:768f4c5c69 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:goldmane] map[] [] [] []} {k8s ci-4081.3.5-a-588adb827d goldmane-768f4c5c69-nm6vx eth0 goldmane [] [] [kns.calico-system ksa.calico-system.goldmane] calicdc52df372f [] [] }} ContainerID="62cb1f87cf16165f451f6c892ad1adfd06ebd722336b883f76a6498daa505f2e" Namespace="calico-system" Pod="goldmane-768f4c5c69-nm6vx" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-goldmane--768f4c5c69--nm6vx-" Aug 13 00:27:13.163733 containerd[2685]: 2025-08-13 00:27:13.106 [INFO][7847] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="62cb1f87cf16165f451f6c892ad1adfd06ebd722336b883f76a6498daa505f2e" Namespace="calico-system" Pod="goldmane-768f4c5c69-nm6vx" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-goldmane--768f4c5c69--nm6vx-eth0" Aug 13 00:27:13.163733 containerd[2685]: 2025-08-13 00:27:13.127 [INFO][7943] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="62cb1f87cf16165f451f6c892ad1adfd06ebd722336b883f76a6498daa505f2e" HandleID="k8s-pod-network.62cb1f87cf16165f451f6c892ad1adfd06ebd722336b883f76a6498daa505f2e" Workload="ci--4081.3.5--a--588adb827d-k8s-goldmane--768f4c5c69--nm6vx-eth0" Aug 13 00:27:13.163733 containerd[2685]: 2025-08-13 00:27:13.127 [INFO][7943] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="62cb1f87cf16165f451f6c892ad1adfd06ebd722336b883f76a6498daa505f2e" HandleID="k8s-pod-network.62cb1f87cf16165f451f6c892ad1adfd06ebd722336b883f76a6498daa505f2e" Workload="ci--4081.3.5--a--588adb827d-k8s-goldmane--768f4c5c69--nm6vx-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000710c00), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081.3.5-a-588adb827d", "pod":"goldmane-768f4c5c69-nm6vx", "timestamp":"2025-08-13 00:27:13.127585007 +0000 UTC"}, Hostname:"ci-4081.3.5-a-588adb827d", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 00:27:13.163733 containerd[2685]: 2025-08-13 00:27:13.127 [INFO][7943] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:13.163733 containerd[2685]: 2025-08-13 00:27:13.127 [INFO][7943] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:13.163733 containerd[2685]: 2025-08-13 00:27:13.127 [INFO][7943] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.5-a-588adb827d' Aug 13 00:27:13.163733 containerd[2685]: 2025-08-13 00:27:13.135 [INFO][7943] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.62cb1f87cf16165f451f6c892ad1adfd06ebd722336b883f76a6498daa505f2e" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:13.163733 containerd[2685]: 2025-08-13 00:27:13.138 [INFO][7943] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:13.163733 containerd[2685]: 2025-08-13 00:27:13.141 [INFO][7943] ipam/ipam.go 511: Trying affinity for 192.168.85.64/26 host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:13.163733 containerd[2685]: 2025-08-13 00:27:13.143 [INFO][7943] ipam/ipam.go 158: Attempting to load block cidr=192.168.85.64/26 host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:13.163733 containerd[2685]: 2025-08-13 00:27:13.144 [INFO][7943] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.85.64/26 host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:13.163733 containerd[2685]: 2025-08-13 00:27:13.144 [INFO][7943] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.85.64/26 handle="k8s-pod-network.62cb1f87cf16165f451f6c892ad1adfd06ebd722336b883f76a6498daa505f2e" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:13.163733 containerd[2685]: 2025-08-13 00:27:13.146 [INFO][7943] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.62cb1f87cf16165f451f6c892ad1adfd06ebd722336b883f76a6498daa505f2e Aug 13 00:27:13.163733 containerd[2685]: 2025-08-13 00:27:13.148 [INFO][7943] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.85.64/26 handle="k8s-pod-network.62cb1f87cf16165f451f6c892ad1adfd06ebd722336b883f76a6498daa505f2e" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:13.163733 containerd[2685]: 2025-08-13 00:27:13.152 [INFO][7943] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.85.69/26] block=192.168.85.64/26 handle="k8s-pod-network.62cb1f87cf16165f451f6c892ad1adfd06ebd722336b883f76a6498daa505f2e" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:13.163733 containerd[2685]: 2025-08-13 00:27:13.152 [INFO][7943] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.85.69/26] handle="k8s-pod-network.62cb1f87cf16165f451f6c892ad1adfd06ebd722336b883f76a6498daa505f2e" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:13.163733 containerd[2685]: 2025-08-13 00:27:13.152 [INFO][7943] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:13.163733 containerd[2685]: 2025-08-13 00:27:13.152 [INFO][7943] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.85.69/26] IPv6=[] ContainerID="62cb1f87cf16165f451f6c892ad1adfd06ebd722336b883f76a6498daa505f2e" HandleID="k8s-pod-network.62cb1f87cf16165f451f6c892ad1adfd06ebd722336b883f76a6498daa505f2e" Workload="ci--4081.3.5--a--588adb827d-k8s-goldmane--768f4c5c69--nm6vx-eth0" Aug 13 00:27:13.164132 containerd[2685]: 2025-08-13 00:27:13.153 [INFO][7847] cni-plugin/k8s.go 418: Populated endpoint ContainerID="62cb1f87cf16165f451f6c892ad1adfd06ebd722336b883f76a6498daa505f2e" Namespace="calico-system" Pod="goldmane-768f4c5c69-nm6vx" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-goldmane--768f4c5c69--nm6vx-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--588adb827d-k8s-goldmane--768f4c5c69--nm6vx-eth0", GenerateName:"goldmane-768f4c5c69-", Namespace:"calico-system", SelfLink:"", UID:"b5da64a2-733d-4c0b-b26c-80dcab04992c", ResourceVersion:"964", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 26, 51, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"768f4c5c69", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-588adb827d", ContainerID:"", Pod:"goldmane-768f4c5c69-nm6vx", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.85.69/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"calicdc52df372f", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:27:13.164132 containerd[2685]: 2025-08-13 00:27:13.154 [INFO][7847] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.85.69/32] ContainerID="62cb1f87cf16165f451f6c892ad1adfd06ebd722336b883f76a6498daa505f2e" Namespace="calico-system" Pod="goldmane-768f4c5c69-nm6vx" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-goldmane--768f4c5c69--nm6vx-eth0" Aug 13 00:27:13.164132 containerd[2685]: 2025-08-13 00:27:13.154 [INFO][7847] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calicdc52df372f ContainerID="62cb1f87cf16165f451f6c892ad1adfd06ebd722336b883f76a6498daa505f2e" Namespace="calico-system" Pod="goldmane-768f4c5c69-nm6vx" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-goldmane--768f4c5c69--nm6vx-eth0" Aug 13 00:27:13.164132 containerd[2685]: 2025-08-13 00:27:13.155 [INFO][7847] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="62cb1f87cf16165f451f6c892ad1adfd06ebd722336b883f76a6498daa505f2e" Namespace="calico-system" Pod="goldmane-768f4c5c69-nm6vx" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-goldmane--768f4c5c69--nm6vx-eth0" Aug 13 00:27:13.164132 containerd[2685]: 2025-08-13 00:27:13.156 [INFO][7847] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="62cb1f87cf16165f451f6c892ad1adfd06ebd722336b883f76a6498daa505f2e" Namespace="calico-system" Pod="goldmane-768f4c5c69-nm6vx" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-goldmane--768f4c5c69--nm6vx-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--588adb827d-k8s-goldmane--768f4c5c69--nm6vx-eth0", GenerateName:"goldmane-768f4c5c69-", Namespace:"calico-system", SelfLink:"", UID:"b5da64a2-733d-4c0b-b26c-80dcab04992c", ResourceVersion:"964", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 26, 51, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"768f4c5c69", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-588adb827d", ContainerID:"62cb1f87cf16165f451f6c892ad1adfd06ebd722336b883f76a6498daa505f2e", Pod:"goldmane-768f4c5c69-nm6vx", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.85.69/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"calicdc52df372f", MAC:"8e:c4:09:ff:bd:5d", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:27:13.164132 containerd[2685]: 2025-08-13 00:27:13.162 [INFO][7847] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="62cb1f87cf16165f451f6c892ad1adfd06ebd722336b883f76a6498daa505f2e" Namespace="calico-system" Pod="goldmane-768f4c5c69-nm6vx" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-goldmane--768f4c5c69--nm6vx-eth0" Aug 13 00:27:13.176404 systemd[1]: Started cri-containerd-e3c9f256fa1183a6b323163f1781eb77ec64c4338cf17fb68dc4c1f29f8344bf.scope - libcontainer container e3c9f256fa1183a6b323163f1781eb77ec64c4338cf17fb68dc4c1f29f8344bf. Aug 13 00:27:13.183954 containerd[2685]: time="2025-08-13T00:27:13.183875226Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 00:27:13.183954 containerd[2685]: time="2025-08-13T00:27:13.183931026Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 00:27:13.183954 containerd[2685]: time="2025-08-13T00:27:13.183942306Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:27:13.184043 containerd[2685]: time="2025-08-13T00:27:13.184013946Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:27:13.193891 systemd[1]: Started cri-containerd-62cb1f87cf16165f451f6c892ad1adfd06ebd722336b883f76a6498daa505f2e.scope - libcontainer container 62cb1f87cf16165f451f6c892ad1adfd06ebd722336b883f76a6498daa505f2e. Aug 13 00:27:13.196779 containerd[2685]: time="2025-08-13T00:27:13.196748070Z" level=info msg="StartContainer for \"e3c9f256fa1183a6b323163f1781eb77ec64c4338cf17fb68dc4c1f29f8344bf\" returns successfully" Aug 13 00:27:13.197547 containerd[2685]: time="2025-08-13T00:27:13.197521070Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\"" Aug 13 00:27:13.217022 containerd[2685]: time="2025-08-13T00:27:13.216990516Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:goldmane-768f4c5c69-nm6vx,Uid:b5da64a2-733d-4c0b-b26c-80dcab04992c,Namespace:calico-system,Attempt:1,} returns sandbox id \"62cb1f87cf16165f451f6c892ad1adfd06ebd722336b883f76a6498daa505f2e\"" Aug 13 00:27:13.255970 systemd-networkd[2588]: calib4046b1b4a7: Link UP Aug 13 00:27:13.256273 systemd-networkd[2588]: calib4046b1b4a7: Gained carrier Aug 13 00:27:13.263757 containerd[2685]: 2025-08-13 00:27:13.110 [INFO][7868] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.5--a--588adb827d-k8s-calico--kube--controllers--f5cbb957c--wms9w-eth0 calico-kube-controllers-f5cbb957c- calico-system 925e0d95-7160-42b1-a3c8-6b7795ca67bc 965 0 2025-08-13 00:26:52 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:f5cbb957c projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s ci-4081.3.5-a-588adb827d calico-kube-controllers-f5cbb957c-wms9w eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] calib4046b1b4a7 [] [] }} ContainerID="7abc47012f7c901b3d8b53e8ae84366e84722e4aaed0067a97a0c021d17effd0" Namespace="calico-system" Pod="calico-kube-controllers-f5cbb957c-wms9w" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--kube--controllers--f5cbb957c--wms9w-" Aug 13 00:27:13.263757 containerd[2685]: 2025-08-13 00:27:13.110 [INFO][7868] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="7abc47012f7c901b3d8b53e8ae84366e84722e4aaed0067a97a0c021d17effd0" Namespace="calico-system" Pod="calico-kube-controllers-f5cbb957c-wms9w" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--kube--controllers--f5cbb957c--wms9w-eth0" Aug 13 00:27:13.263757 containerd[2685]: 2025-08-13 00:27:13.132 [INFO][7949] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="7abc47012f7c901b3d8b53e8ae84366e84722e4aaed0067a97a0c021d17effd0" HandleID="k8s-pod-network.7abc47012f7c901b3d8b53e8ae84366e84722e4aaed0067a97a0c021d17effd0" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--kube--controllers--f5cbb957c--wms9w-eth0" Aug 13 00:27:13.263757 containerd[2685]: 2025-08-13 00:27:13.132 [INFO][7949] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="7abc47012f7c901b3d8b53e8ae84366e84722e4aaed0067a97a0c021d17effd0" HandleID="k8s-pod-network.7abc47012f7c901b3d8b53e8ae84366e84722e4aaed0067a97a0c021d17effd0" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--kube--controllers--f5cbb957c--wms9w-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000791eb0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4081.3.5-a-588adb827d", "pod":"calico-kube-controllers-f5cbb957c-wms9w", "timestamp":"2025-08-13 00:27:13.132669369 +0000 UTC"}, Hostname:"ci-4081.3.5-a-588adb827d", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 00:27:13.263757 containerd[2685]: 2025-08-13 00:27:13.132 [INFO][7949] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:13.263757 containerd[2685]: 2025-08-13 00:27:13.152 [INFO][7949] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:13.263757 containerd[2685]: 2025-08-13 00:27:13.152 [INFO][7949] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.5-a-588adb827d' Aug 13 00:27:13.263757 containerd[2685]: 2025-08-13 00:27:13.236 [INFO][7949] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.7abc47012f7c901b3d8b53e8ae84366e84722e4aaed0067a97a0c021d17effd0" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:13.263757 containerd[2685]: 2025-08-13 00:27:13.239 [INFO][7949] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:13.263757 containerd[2685]: 2025-08-13 00:27:13.242 [INFO][7949] ipam/ipam.go 511: Trying affinity for 192.168.85.64/26 host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:13.263757 containerd[2685]: 2025-08-13 00:27:13.243 [INFO][7949] ipam/ipam.go 158: Attempting to load block cidr=192.168.85.64/26 host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:13.263757 containerd[2685]: 2025-08-13 00:27:13.245 [INFO][7949] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.85.64/26 host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:13.263757 containerd[2685]: 2025-08-13 00:27:13.245 [INFO][7949] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.85.64/26 handle="k8s-pod-network.7abc47012f7c901b3d8b53e8ae84366e84722e4aaed0067a97a0c021d17effd0" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:13.263757 containerd[2685]: 2025-08-13 00:27:13.246 [INFO][7949] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.7abc47012f7c901b3d8b53e8ae84366e84722e4aaed0067a97a0c021d17effd0 Aug 13 00:27:13.263757 containerd[2685]: 2025-08-13 00:27:13.248 [INFO][7949] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.85.64/26 handle="k8s-pod-network.7abc47012f7c901b3d8b53e8ae84366e84722e4aaed0067a97a0c021d17effd0" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:13.263757 containerd[2685]: 2025-08-13 00:27:13.252 [INFO][7949] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.85.70/26] block=192.168.85.64/26 handle="k8s-pod-network.7abc47012f7c901b3d8b53e8ae84366e84722e4aaed0067a97a0c021d17effd0" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:13.263757 containerd[2685]: 2025-08-13 00:27:13.252 [INFO][7949] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.85.70/26] handle="k8s-pod-network.7abc47012f7c901b3d8b53e8ae84366e84722e4aaed0067a97a0c021d17effd0" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:13.263757 containerd[2685]: 2025-08-13 00:27:13.252 [INFO][7949] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:13.263757 containerd[2685]: 2025-08-13 00:27:13.252 [INFO][7949] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.85.70/26] IPv6=[] ContainerID="7abc47012f7c901b3d8b53e8ae84366e84722e4aaed0067a97a0c021d17effd0" HandleID="k8s-pod-network.7abc47012f7c901b3d8b53e8ae84366e84722e4aaed0067a97a0c021d17effd0" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--kube--controllers--f5cbb957c--wms9w-eth0" Aug 13 00:27:13.264206 containerd[2685]: 2025-08-13 00:27:13.253 [INFO][7868] cni-plugin/k8s.go 418: Populated endpoint ContainerID="7abc47012f7c901b3d8b53e8ae84366e84722e4aaed0067a97a0c021d17effd0" Namespace="calico-system" Pod="calico-kube-controllers-f5cbb957c-wms9w" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--kube--controllers--f5cbb957c--wms9w-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--588adb827d-k8s-calico--kube--controllers--f5cbb957c--wms9w-eth0", GenerateName:"calico-kube-controllers-f5cbb957c-", Namespace:"calico-system", SelfLink:"", UID:"925e0d95-7160-42b1-a3c8-6b7795ca67bc", ResourceVersion:"965", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 26, 52, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"f5cbb957c", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-588adb827d", ContainerID:"", Pod:"calico-kube-controllers-f5cbb957c-wms9w", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.85.70/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calib4046b1b4a7", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:27:13.264206 containerd[2685]: 2025-08-13 00:27:13.254 [INFO][7868] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.85.70/32] ContainerID="7abc47012f7c901b3d8b53e8ae84366e84722e4aaed0067a97a0c021d17effd0" Namespace="calico-system" Pod="calico-kube-controllers-f5cbb957c-wms9w" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--kube--controllers--f5cbb957c--wms9w-eth0" Aug 13 00:27:13.264206 containerd[2685]: 2025-08-13 00:27:13.254 [INFO][7868] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calib4046b1b4a7 ContainerID="7abc47012f7c901b3d8b53e8ae84366e84722e4aaed0067a97a0c021d17effd0" Namespace="calico-system" Pod="calico-kube-controllers-f5cbb957c-wms9w" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--kube--controllers--f5cbb957c--wms9w-eth0" Aug 13 00:27:13.264206 containerd[2685]: 2025-08-13 00:27:13.256 [INFO][7868] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="7abc47012f7c901b3d8b53e8ae84366e84722e4aaed0067a97a0c021d17effd0" Namespace="calico-system" Pod="calico-kube-controllers-f5cbb957c-wms9w" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--kube--controllers--f5cbb957c--wms9w-eth0" Aug 13 00:27:13.264206 containerd[2685]: 2025-08-13 00:27:13.256 [INFO][7868] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="7abc47012f7c901b3d8b53e8ae84366e84722e4aaed0067a97a0c021d17effd0" Namespace="calico-system" Pod="calico-kube-controllers-f5cbb957c-wms9w" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--kube--controllers--f5cbb957c--wms9w-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--588adb827d-k8s-calico--kube--controllers--f5cbb957c--wms9w-eth0", GenerateName:"calico-kube-controllers-f5cbb957c-", Namespace:"calico-system", SelfLink:"", UID:"925e0d95-7160-42b1-a3c8-6b7795ca67bc", ResourceVersion:"965", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 26, 52, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"f5cbb957c", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-588adb827d", ContainerID:"7abc47012f7c901b3d8b53e8ae84366e84722e4aaed0067a97a0c021d17effd0", Pod:"calico-kube-controllers-f5cbb957c-wms9w", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.85.70/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calib4046b1b4a7", MAC:"32:50:fa:00:02:91", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:27:13.264206 containerd[2685]: 2025-08-13 00:27:13.262 [INFO][7868] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="7abc47012f7c901b3d8b53e8ae84366e84722e4aaed0067a97a0c021d17effd0" Namespace="calico-system" Pod="calico-kube-controllers-f5cbb957c-wms9w" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--kube--controllers--f5cbb957c--wms9w-eth0" Aug 13 00:27:13.276368 containerd[2685]: time="2025-08-13T00:27:13.276039016Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 00:27:13.276398 containerd[2685]: time="2025-08-13T00:27:13.276363576Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 00:27:13.276398 containerd[2685]: time="2025-08-13T00:27:13.276376776Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:27:13.276465 containerd[2685]: time="2025-08-13T00:27:13.276450816Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:27:13.300362 systemd[1]: Started cri-containerd-7abc47012f7c901b3d8b53e8ae84366e84722e4aaed0067a97a0c021d17effd0.scope - libcontainer container 7abc47012f7c901b3d8b53e8ae84366e84722e4aaed0067a97a0c021d17effd0. Aug 13 00:27:13.323995 containerd[2685]: time="2025-08-13T00:27:13.323967032Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-f5cbb957c-wms9w,Uid:925e0d95-7160-42b1-a3c8-6b7795ca67bc,Namespace:calico-system,Attempt:1,} returns sandbox id \"7abc47012f7c901b3d8b53e8ae84366e84722e4aaed0067a97a0c021d17effd0\"" Aug 13 00:27:13.369543 systemd-networkd[2588]: calid5091b26cdd: Link UP Aug 13 00:27:13.369917 systemd-networkd[2588]: calid5091b26cdd: Gained carrier Aug 13 00:27:13.378658 containerd[2685]: 2025-08-13 00:27:13.114 [INFO][7879] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--v7mtb-eth0 coredns-674b8bbfcf- kube-system 75d28fd6-77d6-4c58-8897-8ce0729a65ea 966 0 2025-08-13 00:26:40 +0000 UTC map[k8s-app:kube-dns pod-template-hash:674b8bbfcf projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4081.3.5-a-588adb827d coredns-674b8bbfcf-v7mtb eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] calid5091b26cdd [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="cd71e691f7ceac8c9846ac32c539f554ff5f9ea8ed0badd72a0ca75474176af9" Namespace="kube-system" Pod="coredns-674b8bbfcf-v7mtb" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--v7mtb-" Aug 13 00:27:13.378658 containerd[2685]: 2025-08-13 00:27:13.115 [INFO][7879] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="cd71e691f7ceac8c9846ac32c539f554ff5f9ea8ed0badd72a0ca75474176af9" Namespace="kube-system" Pod="coredns-674b8bbfcf-v7mtb" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--v7mtb-eth0" Aug 13 00:27:13.378658 containerd[2685]: 2025-08-13 00:27:13.135 [INFO][7967] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="cd71e691f7ceac8c9846ac32c539f554ff5f9ea8ed0badd72a0ca75474176af9" HandleID="k8s-pod-network.cd71e691f7ceac8c9846ac32c539f554ff5f9ea8ed0badd72a0ca75474176af9" Workload="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--v7mtb-eth0" Aug 13 00:27:13.378658 containerd[2685]: 2025-08-13 00:27:13.135 [INFO][7967] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="cd71e691f7ceac8c9846ac32c539f554ff5f9ea8ed0badd72a0ca75474176af9" HandleID="k8s-pod-network.cd71e691f7ceac8c9846ac32c539f554ff5f9ea8ed0badd72a0ca75474176af9" Workload="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--v7mtb-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000519c80), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4081.3.5-a-588adb827d", "pod":"coredns-674b8bbfcf-v7mtb", "timestamp":"2025-08-13 00:27:13.13549917 +0000 UTC"}, Hostname:"ci-4081.3.5-a-588adb827d", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 00:27:13.378658 containerd[2685]: 2025-08-13 00:27:13.135 [INFO][7967] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:13.378658 containerd[2685]: 2025-08-13 00:27:13.252 [INFO][7967] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:13.378658 containerd[2685]: 2025-08-13 00:27:13.252 [INFO][7967] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.5-a-588adb827d' Aug 13 00:27:13.378658 containerd[2685]: 2025-08-13 00:27:13.336 [INFO][7967] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.cd71e691f7ceac8c9846ac32c539f554ff5f9ea8ed0badd72a0ca75474176af9" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:13.378658 containerd[2685]: 2025-08-13 00:27:13.339 [INFO][7967] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:13.378658 containerd[2685]: 2025-08-13 00:27:13.342 [INFO][7967] ipam/ipam.go 511: Trying affinity for 192.168.85.64/26 host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:13.378658 containerd[2685]: 2025-08-13 00:27:13.344 [INFO][7967] ipam/ipam.go 158: Attempting to load block cidr=192.168.85.64/26 host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:13.378658 containerd[2685]: 2025-08-13 00:27:13.359 [INFO][7967] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.85.64/26 host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:13.378658 containerd[2685]: 2025-08-13 00:27:13.359 [INFO][7967] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.85.64/26 handle="k8s-pod-network.cd71e691f7ceac8c9846ac32c539f554ff5f9ea8ed0badd72a0ca75474176af9" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:13.378658 containerd[2685]: 2025-08-13 00:27:13.360 [INFO][7967] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.cd71e691f7ceac8c9846ac32c539f554ff5f9ea8ed0badd72a0ca75474176af9 Aug 13 00:27:13.378658 containerd[2685]: 2025-08-13 00:27:13.363 [INFO][7967] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.85.64/26 handle="k8s-pod-network.cd71e691f7ceac8c9846ac32c539f554ff5f9ea8ed0badd72a0ca75474176af9" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:13.378658 containerd[2685]: 2025-08-13 00:27:13.366 [INFO][7967] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.85.71/26] block=192.168.85.64/26 handle="k8s-pod-network.cd71e691f7ceac8c9846ac32c539f554ff5f9ea8ed0badd72a0ca75474176af9" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:13.378658 containerd[2685]: 2025-08-13 00:27:13.366 [INFO][7967] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.85.71/26] handle="k8s-pod-network.cd71e691f7ceac8c9846ac32c539f554ff5f9ea8ed0badd72a0ca75474176af9" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:13.378658 containerd[2685]: 2025-08-13 00:27:13.366 [INFO][7967] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:13.378658 containerd[2685]: 2025-08-13 00:27:13.366 [INFO][7967] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.85.71/26] IPv6=[] ContainerID="cd71e691f7ceac8c9846ac32c539f554ff5f9ea8ed0badd72a0ca75474176af9" HandleID="k8s-pod-network.cd71e691f7ceac8c9846ac32c539f554ff5f9ea8ed0badd72a0ca75474176af9" Workload="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--v7mtb-eth0" Aug 13 00:27:13.379085 containerd[2685]: 2025-08-13 00:27:13.368 [INFO][7879] cni-plugin/k8s.go 418: Populated endpoint ContainerID="cd71e691f7ceac8c9846ac32c539f554ff5f9ea8ed0badd72a0ca75474176af9" Namespace="kube-system" Pod="coredns-674b8bbfcf-v7mtb" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--v7mtb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--v7mtb-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"75d28fd6-77d6-4c58-8897-8ce0729a65ea", ResourceVersion:"966", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 26, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-588adb827d", ContainerID:"", Pod:"coredns-674b8bbfcf-v7mtb", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.85.71/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calid5091b26cdd", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:27:13.379085 containerd[2685]: 2025-08-13 00:27:13.368 [INFO][7879] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.85.71/32] ContainerID="cd71e691f7ceac8c9846ac32c539f554ff5f9ea8ed0badd72a0ca75474176af9" Namespace="kube-system" Pod="coredns-674b8bbfcf-v7mtb" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--v7mtb-eth0" Aug 13 00:27:13.379085 containerd[2685]: 2025-08-13 00:27:13.368 [INFO][7879] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calid5091b26cdd ContainerID="cd71e691f7ceac8c9846ac32c539f554ff5f9ea8ed0badd72a0ca75474176af9" Namespace="kube-system" Pod="coredns-674b8bbfcf-v7mtb" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--v7mtb-eth0" Aug 13 00:27:13.379085 containerd[2685]: 2025-08-13 00:27:13.369 [INFO][7879] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="cd71e691f7ceac8c9846ac32c539f554ff5f9ea8ed0badd72a0ca75474176af9" Namespace="kube-system" Pod="coredns-674b8bbfcf-v7mtb" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--v7mtb-eth0" Aug 13 00:27:13.379085 containerd[2685]: 2025-08-13 00:27:13.369 [INFO][7879] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="cd71e691f7ceac8c9846ac32c539f554ff5f9ea8ed0badd72a0ca75474176af9" Namespace="kube-system" Pod="coredns-674b8bbfcf-v7mtb" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--v7mtb-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--v7mtb-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"75d28fd6-77d6-4c58-8897-8ce0729a65ea", ResourceVersion:"966", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 26, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-588adb827d", ContainerID:"cd71e691f7ceac8c9846ac32c539f554ff5f9ea8ed0badd72a0ca75474176af9", Pod:"coredns-674b8bbfcf-v7mtb", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.85.71/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calid5091b26cdd", MAC:"3e:f8:32:42:23:68", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:27:13.379085 containerd[2685]: 2025-08-13 00:27:13.377 [INFO][7879] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="cd71e691f7ceac8c9846ac32c539f554ff5f9ea8ed0badd72a0ca75474176af9" Namespace="kube-system" Pod="coredns-674b8bbfcf-v7mtb" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--v7mtb-eth0" Aug 13 00:27:13.391661 containerd[2685]: time="2025-08-13T00:27:13.391544254Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 00:27:13.391691 containerd[2685]: time="2025-08-13T00:27:13.391656094Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 00:27:13.391691 containerd[2685]: time="2025-08-13T00:27:13.391669054Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:27:13.391769 containerd[2685]: time="2025-08-13T00:27:13.391752814Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:27:13.415358 systemd[1]: Started cri-containerd-cd71e691f7ceac8c9846ac32c539f554ff5f9ea8ed0badd72a0ca75474176af9.scope - libcontainer container cd71e691f7ceac8c9846ac32c539f554ff5f9ea8ed0badd72a0ca75474176af9. Aug 13 00:27:13.439045 containerd[2685]: time="2025-08-13T00:27:13.439017589Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-v7mtb,Uid:75d28fd6-77d6-4c58-8897-8ce0729a65ea,Namespace:kube-system,Attempt:1,} returns sandbox id \"cd71e691f7ceac8c9846ac32c539f554ff5f9ea8ed0badd72a0ca75474176af9\"" Aug 13 00:27:13.441483 containerd[2685]: time="2025-08-13T00:27:13.441461110Z" level=info msg="CreateContainer within sandbox \"cd71e691f7ceac8c9846ac32c539f554ff5f9ea8ed0badd72a0ca75474176af9\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Aug 13 00:27:13.449525 containerd[2685]: time="2025-08-13T00:27:13.449496313Z" level=info msg="CreateContainer within sandbox \"cd71e691f7ceac8c9846ac32c539f554ff5f9ea8ed0badd72a0ca75474176af9\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"33c8bbaacf83f0ac95d19f9e13dfbb72a10f0426e8bd3fd6f66887c9f7cba9b1\"" Aug 13 00:27:13.449906 containerd[2685]: time="2025-08-13T00:27:13.449881993Z" level=info msg="StartContainer for \"33c8bbaacf83f0ac95d19f9e13dfbb72a10f0426e8bd3fd6f66887c9f7cba9b1\"" Aug 13 00:27:13.457350 systemd-networkd[2588]: calic1d119fb847: Link UP Aug 13 00:27:13.457546 systemd-networkd[2588]: calic1d119fb847: Gained carrier Aug 13 00:27:13.466064 containerd[2685]: 2025-08-13 00:27:13.126 [INFO][7923] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--ftxjx-eth0 coredns-674b8bbfcf- kube-system 085d2185-0930-470d-b50d-68703ac546fb 967 0 2025-08-13 00:26:40 +0000 UTC map[k8s-app:kube-dns pod-template-hash:674b8bbfcf projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4081.3.5-a-588adb827d coredns-674b8bbfcf-ftxjx eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] calic1d119fb847 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] [] }} ContainerID="3afc2e5aedc5d5722bff856b27c1c16aad8ea7dc957c30eaf468179a695d337a" Namespace="kube-system" Pod="coredns-674b8bbfcf-ftxjx" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--ftxjx-" Aug 13 00:27:13.466064 containerd[2685]: 2025-08-13 00:27:13.127 [INFO][7923] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="3afc2e5aedc5d5722bff856b27c1c16aad8ea7dc957c30eaf468179a695d337a" Namespace="kube-system" Pod="coredns-674b8bbfcf-ftxjx" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--ftxjx-eth0" Aug 13 00:27:13.466064 containerd[2685]: 2025-08-13 00:27:13.147 [INFO][8001] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="3afc2e5aedc5d5722bff856b27c1c16aad8ea7dc957c30eaf468179a695d337a" HandleID="k8s-pod-network.3afc2e5aedc5d5722bff856b27c1c16aad8ea7dc957c30eaf468179a695d337a" Workload="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--ftxjx-eth0" Aug 13 00:27:13.466064 containerd[2685]: 2025-08-13 00:27:13.148 [INFO][8001] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="3afc2e5aedc5d5722bff856b27c1c16aad8ea7dc957c30eaf468179a695d337a" HandleID="k8s-pod-network.3afc2e5aedc5d5722bff856b27c1c16aad8ea7dc957c30eaf468179a695d337a" Workload="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--ftxjx-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40001b6ce0), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4081.3.5-a-588adb827d", "pod":"coredns-674b8bbfcf-ftxjx", "timestamp":"2025-08-13 00:27:13.147513414 +0000 UTC"}, Hostname:"ci-4081.3.5-a-588adb827d", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 00:27:13.466064 containerd[2685]: 2025-08-13 00:27:13.148 [INFO][8001] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:13.466064 containerd[2685]: 2025-08-13 00:27:13.366 [INFO][8001] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:13.466064 containerd[2685]: 2025-08-13 00:27:13.366 [INFO][8001] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.5-a-588adb827d' Aug 13 00:27:13.466064 containerd[2685]: 2025-08-13 00:27:13.436 [INFO][8001] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.3afc2e5aedc5d5722bff856b27c1c16aad8ea7dc957c30eaf468179a695d337a" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:13.466064 containerd[2685]: 2025-08-13 00:27:13.440 [INFO][8001] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:13.466064 containerd[2685]: 2025-08-13 00:27:13.443 [INFO][8001] ipam/ipam.go 511: Trying affinity for 192.168.85.64/26 host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:13.466064 containerd[2685]: 2025-08-13 00:27:13.444 [INFO][8001] ipam/ipam.go 158: Attempting to load block cidr=192.168.85.64/26 host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:13.466064 containerd[2685]: 2025-08-13 00:27:13.446 [INFO][8001] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.85.64/26 host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:13.466064 containerd[2685]: 2025-08-13 00:27:13.446 [INFO][8001] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.85.64/26 handle="k8s-pod-network.3afc2e5aedc5d5722bff856b27c1c16aad8ea7dc957c30eaf468179a695d337a" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:13.466064 containerd[2685]: 2025-08-13 00:27:13.447 [INFO][8001] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.3afc2e5aedc5d5722bff856b27c1c16aad8ea7dc957c30eaf468179a695d337a Aug 13 00:27:13.466064 containerd[2685]: 2025-08-13 00:27:13.450 [INFO][8001] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.85.64/26 handle="k8s-pod-network.3afc2e5aedc5d5722bff856b27c1c16aad8ea7dc957c30eaf468179a695d337a" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:13.466064 containerd[2685]: 2025-08-13 00:27:13.454 [INFO][8001] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.85.72/26] block=192.168.85.64/26 handle="k8s-pod-network.3afc2e5aedc5d5722bff856b27c1c16aad8ea7dc957c30eaf468179a695d337a" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:13.466064 containerd[2685]: 2025-08-13 00:27:13.454 [INFO][8001] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.85.72/26] handle="k8s-pod-network.3afc2e5aedc5d5722bff856b27c1c16aad8ea7dc957c30eaf468179a695d337a" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:13.466064 containerd[2685]: 2025-08-13 00:27:13.454 [INFO][8001] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:13.466064 containerd[2685]: 2025-08-13 00:27:13.454 [INFO][8001] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.85.72/26] IPv6=[] ContainerID="3afc2e5aedc5d5722bff856b27c1c16aad8ea7dc957c30eaf468179a695d337a" HandleID="k8s-pod-network.3afc2e5aedc5d5722bff856b27c1c16aad8ea7dc957c30eaf468179a695d337a" Workload="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--ftxjx-eth0" Aug 13 00:27:13.466520 containerd[2685]: 2025-08-13 00:27:13.455 [INFO][7923] cni-plugin/k8s.go 418: Populated endpoint ContainerID="3afc2e5aedc5d5722bff856b27c1c16aad8ea7dc957c30eaf468179a695d337a" Namespace="kube-system" Pod="coredns-674b8bbfcf-ftxjx" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--ftxjx-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--ftxjx-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"085d2185-0930-470d-b50d-68703ac546fb", ResourceVersion:"967", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 26, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-588adb827d", ContainerID:"", Pod:"coredns-674b8bbfcf-ftxjx", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.85.72/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calic1d119fb847", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:27:13.466520 containerd[2685]: 2025-08-13 00:27:13.455 [INFO][7923] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.85.72/32] ContainerID="3afc2e5aedc5d5722bff856b27c1c16aad8ea7dc957c30eaf468179a695d337a" Namespace="kube-system" Pod="coredns-674b8bbfcf-ftxjx" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--ftxjx-eth0" Aug 13 00:27:13.466520 containerd[2685]: 2025-08-13 00:27:13.455 [INFO][7923] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calic1d119fb847 ContainerID="3afc2e5aedc5d5722bff856b27c1c16aad8ea7dc957c30eaf468179a695d337a" Namespace="kube-system" Pod="coredns-674b8bbfcf-ftxjx" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--ftxjx-eth0" Aug 13 00:27:13.466520 containerd[2685]: 2025-08-13 00:27:13.457 [INFO][7923] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="3afc2e5aedc5d5722bff856b27c1c16aad8ea7dc957c30eaf468179a695d337a" Namespace="kube-system" Pod="coredns-674b8bbfcf-ftxjx" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--ftxjx-eth0" Aug 13 00:27:13.466520 containerd[2685]: 2025-08-13 00:27:13.457 [INFO][7923] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="3afc2e5aedc5d5722bff856b27c1c16aad8ea7dc957c30eaf468179a695d337a" Namespace="kube-system" Pod="coredns-674b8bbfcf-ftxjx" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--ftxjx-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--ftxjx-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"085d2185-0930-470d-b50d-68703ac546fb", ResourceVersion:"967", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 26, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-588adb827d", ContainerID:"3afc2e5aedc5d5722bff856b27c1c16aad8ea7dc957c30eaf468179a695d337a", Pod:"coredns-674b8bbfcf-ftxjx", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.85.72/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calic1d119fb847", MAC:"06:91:ba:48:cf:4b", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:27:13.466520 containerd[2685]: 2025-08-13 00:27:13.464 [INFO][7923] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="3afc2e5aedc5d5722bff856b27c1c16aad8ea7dc957c30eaf468179a695d337a" Namespace="kube-system" Pod="coredns-674b8bbfcf-ftxjx" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--ftxjx-eth0" Aug 13 00:27:13.477364 systemd[1]: Started cri-containerd-33c8bbaacf83f0ac95d19f9e13dfbb72a10f0426e8bd3fd6f66887c9f7cba9b1.scope - libcontainer container 33c8bbaacf83f0ac95d19f9e13dfbb72a10f0426e8bd3fd6f66887c9f7cba9b1. Aug 13 00:27:13.478680 containerd[2685]: time="2025-08-13T00:27:13.478601762Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 00:27:13.478680 containerd[2685]: time="2025-08-13T00:27:13.478658202Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 00:27:13.478680 containerd[2685]: time="2025-08-13T00:27:13.478669562Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:27:13.478774 containerd[2685]: time="2025-08-13T00:27:13.478753642Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:27:13.488286 systemd[1]: Started cri-containerd-3afc2e5aedc5d5722bff856b27c1c16aad8ea7dc957c30eaf468179a695d337a.scope - libcontainer container 3afc2e5aedc5d5722bff856b27c1c16aad8ea7dc957c30eaf468179a695d337a. Aug 13 00:27:13.495485 containerd[2685]: time="2025-08-13T00:27:13.495453088Z" level=info msg="StartContainer for \"33c8bbaacf83f0ac95d19f9e13dfbb72a10f0426e8bd3fd6f66887c9f7cba9b1\" returns successfully" Aug 13 00:27:13.511983 containerd[2685]: time="2025-08-13T00:27:13.511909253Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-ftxjx,Uid:085d2185-0930-470d-b50d-68703ac546fb,Namespace:kube-system,Attempt:1,} returns sandbox id \"3afc2e5aedc5d5722bff856b27c1c16aad8ea7dc957c30eaf468179a695d337a\"" Aug 13 00:27:13.514417 containerd[2685]: time="2025-08-13T00:27:13.514394534Z" level=info msg="CreateContainer within sandbox \"3afc2e5aedc5d5722bff856b27c1c16aad8ea7dc957c30eaf468179a695d337a\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Aug 13 00:27:13.520257 containerd[2685]: time="2025-08-13T00:27:13.520219176Z" level=info msg="CreateContainer within sandbox \"3afc2e5aedc5d5722bff856b27c1c16aad8ea7dc957c30eaf468179a695d337a\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"7db74fe8433280f1a189b1d643fdbf222a5175615ff940e32b425a9e844d9273\"" Aug 13 00:27:13.520575 containerd[2685]: time="2025-08-13T00:27:13.520558776Z" level=info msg="StartContainer for \"7db74fe8433280f1a189b1d643fdbf222a5175615ff940e32b425a9e844d9273\"" Aug 13 00:27:13.552481 systemd[1]: Started cri-containerd-7db74fe8433280f1a189b1d643fdbf222a5175615ff940e32b425a9e844d9273.scope - libcontainer container 7db74fe8433280f1a189b1d643fdbf222a5175615ff940e32b425a9e844d9273. Aug 13 00:27:13.558318 systemd-networkd[2588]: cali650bf866543: Gained IPv6LL Aug 13 00:27:13.570369 containerd[2685]: time="2025-08-13T00:27:13.570338032Z" level=info msg="StartContainer for \"7db74fe8433280f1a189b1d643fdbf222a5175615ff940e32b425a9e844d9273\" returns successfully" Aug 13 00:27:13.988966 containerd[2685]: time="2025-08-13T00:27:13.988933369Z" level=info msg="StopPodSandbox for \"7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530\"" Aug 13 00:27:14.053683 containerd[2685]: 2025-08-13 00:27:14.024 [INFO][8446] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" Aug 13 00:27:14.053683 containerd[2685]: 2025-08-13 00:27:14.024 [INFO][8446] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" iface="eth0" netns="/var/run/netns/cni-0a41a3bd-dd3a-e983-464b-f873d06c1806" Aug 13 00:27:14.053683 containerd[2685]: 2025-08-13 00:27:14.024 [INFO][8446] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" iface="eth0" netns="/var/run/netns/cni-0a41a3bd-dd3a-e983-464b-f873d06c1806" Aug 13 00:27:14.053683 containerd[2685]: 2025-08-13 00:27:14.024 [INFO][8446] cni-plugin/dataplane_linux.go 597: Workload's veth was already gone. Nothing to do. ContainerID="7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" iface="eth0" netns="/var/run/netns/cni-0a41a3bd-dd3a-e983-464b-f873d06c1806" Aug 13 00:27:14.053683 containerd[2685]: 2025-08-13 00:27:14.024 [INFO][8446] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" Aug 13 00:27:14.053683 containerd[2685]: 2025-08-13 00:27:14.024 [INFO][8446] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" Aug 13 00:27:14.053683 containerd[2685]: 2025-08-13 00:27:14.042 [INFO][8470] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" HandleID="k8s-pod-network.7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--lcpkt-eth0" Aug 13 00:27:14.053683 containerd[2685]: 2025-08-13 00:27:14.042 [INFO][8470] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:14.053683 containerd[2685]: 2025-08-13 00:27:14.042 [INFO][8470] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:14.053683 containerd[2685]: 2025-08-13 00:27:14.049 [WARNING][8470] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" HandleID="k8s-pod-network.7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--lcpkt-eth0" Aug 13 00:27:14.053683 containerd[2685]: 2025-08-13 00:27:14.050 [INFO][8470] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" HandleID="k8s-pod-network.7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--lcpkt-eth0" Aug 13 00:27:14.053683 containerd[2685]: 2025-08-13 00:27:14.051 [INFO][8470] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:14.053683 containerd[2685]: 2025-08-13 00:27:14.052 [INFO][8446] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" Aug 13 00:27:14.054200 containerd[2685]: time="2025-08-13T00:27:14.053850590Z" level=info msg="TearDown network for sandbox \"7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530\" successfully" Aug 13 00:27:14.054200 containerd[2685]: time="2025-08-13T00:27:14.053875430Z" level=info msg="StopPodSandbox for \"7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530\" returns successfully" Aug 13 00:27:14.054462 containerd[2685]: time="2025-08-13T00:27:14.054437390Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-54c5fc9686-lcpkt,Uid:3dc1e9ad-8b04-42ba-98fc-165c43766984,Namespace:calico-apiserver,Attempt:1,}" Aug 13 00:27:14.071240 systemd[1]: run-netns-cni\x2d0a41a3bd\x2ddd3a\x2de983\x2d464b\x2df873d06c1806.mount: Deactivated successfully. Aug 13 00:27:14.071323 systemd[1]: run-netns-cni\x2d50519924\x2d17e8\x2d677e\x2d9bef\x2d2e4d7828d6ae.mount: Deactivated successfully. Aug 13 00:27:14.079905 kubelet[4148]: I0813 00:27:14.079882 4148 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Aug 13 00:27:14.079905 kubelet[4148]: I0813 00:27:14.079891 4148 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Aug 13 00:27:14.085452 kubelet[4148]: I0813 00:27:14.085411 4148 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-ftxjx" podStartSLOduration=34.085397239 podStartE2EDuration="34.085397239s" podCreationTimestamp="2025-08-13 00:26:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 00:27:14.085160719 +0000 UTC m=+42.170117858" watchObservedRunningTime="2025-08-13 00:27:14.085397239 +0000 UTC m=+42.170354378" Aug 13 00:27:14.099388 kubelet[4148]: I0813 00:27:14.099343 4148 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-v7mtb" podStartSLOduration=34.099327124 podStartE2EDuration="34.099327124s" podCreationTimestamp="2025-08-13 00:26:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 00:27:14.098920204 +0000 UTC m=+42.183877343" watchObservedRunningTime="2025-08-13 00:27:14.099327124 +0000 UTC m=+42.184284263" Aug 13 00:27:14.165309 systemd-networkd[2588]: cali4364a52d32c: Link UP Aug 13 00:27:14.165529 systemd-networkd[2588]: cali4364a52d32c: Gained carrier Aug 13 00:27:14.183665 containerd[2685]: 2025-08-13 00:27:14.103 [INFO][8491] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--lcpkt-eth0 calico-apiserver-54c5fc9686- calico-apiserver 3dc1e9ad-8b04-42ba-98fc-165c43766984 1001 0 2025-08-13 00:26:48 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:54c5fc9686 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4081.3.5-a-588adb827d calico-apiserver-54c5fc9686-lcpkt eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali4364a52d32c [] [] }} ContainerID="10b7183e37555429f20aaf37a6f7d52ae00153679751bd35876ef5c03f417df3" Namespace="calico-apiserver" Pod="calico-apiserver-54c5fc9686-lcpkt" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--lcpkt-" Aug 13 00:27:14.183665 containerd[2685]: 2025-08-13 00:27:14.103 [INFO][8491] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="10b7183e37555429f20aaf37a6f7d52ae00153679751bd35876ef5c03f417df3" Namespace="calico-apiserver" Pod="calico-apiserver-54c5fc9686-lcpkt" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--lcpkt-eth0" Aug 13 00:27:14.183665 containerd[2685]: 2025-08-13 00:27:14.132 [INFO][8518] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="10b7183e37555429f20aaf37a6f7d52ae00153679751bd35876ef5c03f417df3" HandleID="k8s-pod-network.10b7183e37555429f20aaf37a6f7d52ae00153679751bd35876ef5c03f417df3" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--lcpkt-eth0" Aug 13 00:27:14.183665 containerd[2685]: 2025-08-13 00:27:14.132 [INFO][8518] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="10b7183e37555429f20aaf37a6f7d52ae00153679751bd35876ef5c03f417df3" HandleID="k8s-pod-network.10b7183e37555429f20aaf37a6f7d52ae00153679751bd35876ef5c03f417df3" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--lcpkt-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40001b6b10), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4081.3.5-a-588adb827d", "pod":"calico-apiserver-54c5fc9686-lcpkt", "timestamp":"2025-08-13 00:27:14.132495374 +0000 UTC"}, Hostname:"ci-4081.3.5-a-588adb827d", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 00:27:14.183665 containerd[2685]: 2025-08-13 00:27:14.132 [INFO][8518] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:14.183665 containerd[2685]: 2025-08-13 00:27:14.132 [INFO][8518] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:14.183665 containerd[2685]: 2025-08-13 00:27:14.132 [INFO][8518] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.5-a-588adb827d' Aug 13 00:27:14.183665 containerd[2685]: 2025-08-13 00:27:14.144 [INFO][8518] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.10b7183e37555429f20aaf37a6f7d52ae00153679751bd35876ef5c03f417df3" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:14.183665 containerd[2685]: 2025-08-13 00:27:14.148 [INFO][8518] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:14.183665 containerd[2685]: 2025-08-13 00:27:14.150 [INFO][8518] ipam/ipam.go 511: Trying affinity for 192.168.85.64/26 host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:14.183665 containerd[2685]: 2025-08-13 00:27:14.152 [INFO][8518] ipam/ipam.go 158: Attempting to load block cidr=192.168.85.64/26 host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:14.183665 containerd[2685]: 2025-08-13 00:27:14.154 [INFO][8518] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.85.64/26 host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:14.183665 containerd[2685]: 2025-08-13 00:27:14.154 [INFO][8518] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.85.64/26 handle="k8s-pod-network.10b7183e37555429f20aaf37a6f7d52ae00153679751bd35876ef5c03f417df3" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:14.183665 containerd[2685]: 2025-08-13 00:27:14.155 [INFO][8518] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.10b7183e37555429f20aaf37a6f7d52ae00153679751bd35876ef5c03f417df3 Aug 13 00:27:14.183665 containerd[2685]: 2025-08-13 00:27:14.157 [INFO][8518] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.85.64/26 handle="k8s-pod-network.10b7183e37555429f20aaf37a6f7d52ae00153679751bd35876ef5c03f417df3" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:14.183665 containerd[2685]: 2025-08-13 00:27:14.161 [INFO][8518] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.85.73/26] block=192.168.85.64/26 handle="k8s-pod-network.10b7183e37555429f20aaf37a6f7d52ae00153679751bd35876ef5c03f417df3" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:14.183665 containerd[2685]: 2025-08-13 00:27:14.161 [INFO][8518] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.85.73/26] handle="k8s-pod-network.10b7183e37555429f20aaf37a6f7d52ae00153679751bd35876ef5c03f417df3" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:14.183665 containerd[2685]: 2025-08-13 00:27:14.161 [INFO][8518] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:14.183665 containerd[2685]: 2025-08-13 00:27:14.161 [INFO][8518] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.85.73/26] IPv6=[] ContainerID="10b7183e37555429f20aaf37a6f7d52ae00153679751bd35876ef5c03f417df3" HandleID="k8s-pod-network.10b7183e37555429f20aaf37a6f7d52ae00153679751bd35876ef5c03f417df3" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--lcpkt-eth0" Aug 13 00:27:14.184115 containerd[2685]: 2025-08-13 00:27:14.163 [INFO][8491] cni-plugin/k8s.go 418: Populated endpoint ContainerID="10b7183e37555429f20aaf37a6f7d52ae00153679751bd35876ef5c03f417df3" Namespace="calico-apiserver" Pod="calico-apiserver-54c5fc9686-lcpkt" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--lcpkt-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--lcpkt-eth0", GenerateName:"calico-apiserver-54c5fc9686-", Namespace:"calico-apiserver", SelfLink:"", UID:"3dc1e9ad-8b04-42ba-98fc-165c43766984", ResourceVersion:"1001", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 26, 48, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"54c5fc9686", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-588adb827d", ContainerID:"", Pod:"calico-apiserver-54c5fc9686-lcpkt", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.85.73/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali4364a52d32c", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:27:14.184115 containerd[2685]: 2025-08-13 00:27:14.163 [INFO][8491] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.85.73/32] ContainerID="10b7183e37555429f20aaf37a6f7d52ae00153679751bd35876ef5c03f417df3" Namespace="calico-apiserver" Pod="calico-apiserver-54c5fc9686-lcpkt" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--lcpkt-eth0" Aug 13 00:27:14.184115 containerd[2685]: 2025-08-13 00:27:14.163 [INFO][8491] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali4364a52d32c ContainerID="10b7183e37555429f20aaf37a6f7d52ae00153679751bd35876ef5c03f417df3" Namespace="calico-apiserver" Pod="calico-apiserver-54c5fc9686-lcpkt" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--lcpkt-eth0" Aug 13 00:27:14.184115 containerd[2685]: 2025-08-13 00:27:14.165 [INFO][8491] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="10b7183e37555429f20aaf37a6f7d52ae00153679751bd35876ef5c03f417df3" Namespace="calico-apiserver" Pod="calico-apiserver-54c5fc9686-lcpkt" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--lcpkt-eth0" Aug 13 00:27:14.184115 containerd[2685]: 2025-08-13 00:27:14.165 [INFO][8491] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="10b7183e37555429f20aaf37a6f7d52ae00153679751bd35876ef5c03f417df3" Namespace="calico-apiserver" Pod="calico-apiserver-54c5fc9686-lcpkt" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--lcpkt-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--lcpkt-eth0", GenerateName:"calico-apiserver-54c5fc9686-", Namespace:"calico-apiserver", SelfLink:"", UID:"3dc1e9ad-8b04-42ba-98fc-165c43766984", ResourceVersion:"1001", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 26, 48, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"54c5fc9686", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-588adb827d", ContainerID:"10b7183e37555429f20aaf37a6f7d52ae00153679751bd35876ef5c03f417df3", Pod:"calico-apiserver-54c5fc9686-lcpkt", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.85.73/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali4364a52d32c", MAC:"76:5f:a0:fa:72:21", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:27:14.184115 containerd[2685]: 2025-08-13 00:27:14.182 [INFO][8491] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="10b7183e37555429f20aaf37a6f7d52ae00153679751bd35876ef5c03f417df3" Namespace="calico-apiserver" Pod="calico-apiserver-54c5fc9686-lcpkt" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--lcpkt-eth0" Aug 13 00:27:14.205699 containerd[2685]: time="2025-08-13T00:27:14.205629036Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 00:27:14.205699 containerd[2685]: time="2025-08-13T00:27:14.205691676Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 00:27:14.205831 containerd[2685]: time="2025-08-13T00:27:14.205703836Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:27:14.205831 containerd[2685]: time="2025-08-13T00:27:14.205788596Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:27:14.236370 systemd[1]: Started cri-containerd-10b7183e37555429f20aaf37a6f7d52ae00153679751bd35876ef5c03f417df3.scope - libcontainer container 10b7183e37555429f20aaf37a6f7d52ae00153679751bd35876ef5c03f417df3. Aug 13 00:27:14.260876 containerd[2685]: time="2025-08-13T00:27:14.260845613Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-54c5fc9686-lcpkt,Uid:3dc1e9ad-8b04-42ba-98fc-165c43766984,Namespace:calico-apiserver,Attempt:1,} returns sandbox id \"10b7183e37555429f20aaf37a6f7d52ae00153679751bd35876ef5c03f417df3\"" Aug 13 00:27:14.263119 containerd[2685]: time="2025-08-13T00:27:14.263094494Z" level=info msg="CreateContainer within sandbox \"10b7183e37555429f20aaf37a6f7d52ae00153679751bd35876ef5c03f417df3\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Aug 13 00:27:14.267781 containerd[2685]: time="2025-08-13T00:27:14.267751415Z" level=info msg="CreateContainer within sandbox \"10b7183e37555429f20aaf37a6f7d52ae00153679751bd35876ef5c03f417df3\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"843981f61aebda8bd40223feef37057b44d9eac0207849f6a3e1791b94fb5ca8\"" Aug 13 00:27:14.268114 containerd[2685]: time="2025-08-13T00:27:14.268093295Z" level=info msg="StartContainer for \"843981f61aebda8bd40223feef37057b44d9eac0207849f6a3e1791b94fb5ca8\"" Aug 13 00:27:14.302410 systemd[1]: Started cri-containerd-843981f61aebda8bd40223feef37057b44d9eac0207849f6a3e1791b94fb5ca8.scope - libcontainer container 843981f61aebda8bd40223feef37057b44d9eac0207849f6a3e1791b94fb5ca8. Aug 13 00:27:14.307309 kubelet[4148]: I0813 00:27:14.307283 4148 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Aug 13 00:27:14.324197 containerd[2685]: time="2025-08-13T00:27:14.324160473Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:27:14.324270 containerd[2685]: time="2025-08-13T00:27:14.324199073Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2: active requests=0, bytes read=13754366" Aug 13 00:27:14.324963 containerd[2685]: time="2025-08-13T00:27:14.324940033Z" level=info msg="ImageCreate event name:\"sha256:664ed31fb4687b0de23d6e6e116bc87b236790d7355871d3237c54452e02e27c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:27:14.326730 containerd[2685]: time="2025-08-13T00:27:14.326705914Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:8fec2de12dfa51bae89d941938a07af2598eb8bfcab55d0dded1d9c193d7b99f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:27:14.327478 containerd[2685]: time="2025-08-13T00:27:14.327451474Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\" with image id \"sha256:664ed31fb4687b0de23d6e6e116bc87b236790d7355871d3237c54452e02e27c\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:8fec2de12dfa51bae89d941938a07af2598eb8bfcab55d0dded1d9c193d7b99f\", size \"15123559\" in 1.129894964s" Aug 13 00:27:14.327510 containerd[2685]: time="2025-08-13T00:27:14.327482794Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.30.2\" returns image reference \"sha256:664ed31fb4687b0de23d6e6e116bc87b236790d7355871d3237c54452e02e27c\"" Aug 13 00:27:14.328172 containerd[2685]: time="2025-08-13T00:27:14.328154514Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.2\"" Aug 13 00:27:14.329756 containerd[2685]: time="2025-08-13T00:27:14.329729434Z" level=info msg="CreateContainer within sandbox \"28c2e0ac93ee78b239c852a9b7b6e26ab8f23ab9c93620b45125367839cf7097\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Aug 13 00:27:14.337407 containerd[2685]: time="2025-08-13T00:27:14.337371277Z" level=info msg="CreateContainer within sandbox \"28c2e0ac93ee78b239c852a9b7b6e26ab8f23ab9c93620b45125367839cf7097\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"1bebd9107f1a9ea637a75df3789e45c5bcfeefd4ecdaa6dc9177bcd383fb8c2f\"" Aug 13 00:27:14.337763 containerd[2685]: time="2025-08-13T00:27:14.337738797Z" level=info msg="StartContainer for \"1bebd9107f1a9ea637a75df3789e45c5bcfeefd4ecdaa6dc9177bcd383fb8c2f\"" Aug 13 00:27:14.340932 containerd[2685]: time="2025-08-13T00:27:14.340908358Z" level=info msg="StartContainer for \"843981f61aebda8bd40223feef37057b44d9eac0207849f6a3e1791b94fb5ca8\" returns successfully" Aug 13 00:27:14.365394 systemd[1]: Started cri-containerd-1bebd9107f1a9ea637a75df3789e45c5bcfeefd4ecdaa6dc9177bcd383fb8c2f.scope - libcontainer container 1bebd9107f1a9ea637a75df3789e45c5bcfeefd4ecdaa6dc9177bcd383fb8c2f. Aug 13 00:27:14.385032 containerd[2685]: time="2025-08-13T00:27:14.384996771Z" level=info msg="StartContainer for \"1bebd9107f1a9ea637a75df3789e45c5bcfeefd4ecdaa6dc9177bcd383fb8c2f\" returns successfully" Aug 13 00:27:14.646344 systemd-networkd[2588]: calic1d119fb847: Gained IPv6LL Aug 13 00:27:14.646622 systemd-networkd[2588]: calicdc52df372f: Gained IPv6LL Aug 13 00:27:15.030320 systemd-networkd[2588]: calib4046b1b4a7: Gained IPv6LL Aug 13 00:27:15.033783 kubelet[4148]: I0813 00:27:15.033756 4148 csi_plugin.go:106] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Aug 13 00:27:15.033783 kubelet[4148]: I0813 00:27:15.033787 4148 csi_plugin.go:119] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Aug 13 00:27:15.089684 kubelet[4148]: I0813 00:27:15.089637 4148 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-54c5fc9686-lcpkt" podStartSLOduration=27.089621186 podStartE2EDuration="27.089621186s" podCreationTimestamp="2025-08-13 00:26:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 00:27:15.089498746 +0000 UTC m=+43.174455885" watchObservedRunningTime="2025-08-13 00:27:15.089621186 +0000 UTC m=+43.174578285" Aug 13 00:27:15.096819 kubelet[4148]: I0813 00:27:15.096776 4148 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-4wkp8" podStartSLOduration=19.089996207 podStartE2EDuration="23.096754788s" podCreationTimestamp="2025-08-13 00:26:52 +0000 UTC" firstStartedPulling="2025-08-13 00:27:10.321309173 +0000 UTC m=+38.406266312" lastFinishedPulling="2025-08-13 00:27:14.328067754 +0000 UTC m=+42.413024893" observedRunningTime="2025-08-13 00:27:15.096626948 +0000 UTC m=+43.181584087" watchObservedRunningTime="2025-08-13 00:27:15.096754788 +0000 UTC m=+43.181711927" Aug 13 00:27:15.223336 systemd-networkd[2588]: calid5091b26cdd: Gained IPv6LL Aug 13 00:27:15.872028 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1167627311.mount: Deactivated successfully. Aug 13 00:27:16.079124 containerd[2685]: time="2025-08-13T00:27:16.079088510Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:27:16.079525 containerd[2685]: time="2025-08-13T00:27:16.079190830Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/goldmane:v3.30.2: active requests=0, bytes read=61838790" Aug 13 00:27:16.079946 containerd[2685]: time="2025-08-13T00:27:16.079925510Z" level=info msg="ImageCreate event name:\"sha256:1389d38feb576cfff09a57a2c028a53e51a72c658f295166960f770eaf07985f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:27:16.081979 containerd[2685]: time="2025-08-13T00:27:16.081952271Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/goldmane@sha256:a2b761fd93d824431ad93e59e8e670cdf00b478f4b532145297e1e67f2768305\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:27:16.082788 containerd[2685]: time="2025-08-13T00:27:16.082765951Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/goldmane:v3.30.2\" with image id \"sha256:1389d38feb576cfff09a57a2c028a53e51a72c658f295166960f770eaf07985f\", repo tag \"ghcr.io/flatcar/calico/goldmane:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/goldmane@sha256:a2b761fd93d824431ad93e59e8e670cdf00b478f4b532145297e1e67f2768305\", size \"61838636\" in 1.754583237s" Aug 13 00:27:16.082825 containerd[2685]: time="2025-08-13T00:27:16.082795511Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/goldmane:v3.30.2\" returns image reference \"sha256:1389d38feb576cfff09a57a2c028a53e51a72c658f295166960f770eaf07985f\"" Aug 13 00:27:16.083764 containerd[2685]: time="2025-08-13T00:27:16.083735431Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\"" Aug 13 00:27:16.084877 kubelet[4148]: I0813 00:27:16.084858 4148 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Aug 13 00:27:16.085168 containerd[2685]: time="2025-08-13T00:27:16.085052791Z" level=info msg="CreateContainer within sandbox \"62cb1f87cf16165f451f6c892ad1adfd06ebd722336b883f76a6498daa505f2e\" for container &ContainerMetadata{Name:goldmane,Attempt:0,}" Aug 13 00:27:16.092458 containerd[2685]: time="2025-08-13T00:27:16.092428553Z" level=info msg="CreateContainer within sandbox \"62cb1f87cf16165f451f6c892ad1adfd06ebd722336b883f76a6498daa505f2e\" for &ContainerMetadata{Name:goldmane,Attempt:0,} returns container id \"460b5755a01848c56c63fde6cadf9bb845eb093fb75e84e661e95af12604c0d5\"" Aug 13 00:27:16.092836 containerd[2685]: time="2025-08-13T00:27:16.092815434Z" level=info msg="StartContainer for \"460b5755a01848c56c63fde6cadf9bb845eb093fb75e84e661e95af12604c0d5\"" Aug 13 00:27:16.118329 systemd-networkd[2588]: cali4364a52d32c: Gained IPv6LL Aug 13 00:27:16.141428 systemd[1]: Started cri-containerd-460b5755a01848c56c63fde6cadf9bb845eb093fb75e84e661e95af12604c0d5.scope - libcontainer container 460b5755a01848c56c63fde6cadf9bb845eb093fb75e84e661e95af12604c0d5. Aug 13 00:27:16.168634 containerd[2685]: time="2025-08-13T00:27:16.168606054Z" level=info msg="StartContainer for \"460b5755a01848c56c63fde6cadf9bb845eb093fb75e84e661e95af12604c0d5\" returns successfully" Aug 13 00:27:17.095352 kubelet[4148]: I0813 00:27:17.095305 4148 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/goldmane-768f4c5c69-nm6vx" podStartSLOduration=23.229483269 podStartE2EDuration="26.095290263s" podCreationTimestamp="2025-08-13 00:26:51 +0000 UTC" firstStartedPulling="2025-08-13 00:27:13.217838957 +0000 UTC m=+41.302796096" lastFinishedPulling="2025-08-13 00:27:16.083645951 +0000 UTC m=+44.168603090" observedRunningTime="2025-08-13 00:27:17.095084903 +0000 UTC m=+45.180042042" watchObservedRunningTime="2025-08-13 00:27:17.095290263 +0000 UTC m=+45.180247402" Aug 13 00:27:18.085262 containerd[2685]: time="2025-08-13T00:27:18.085202712Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:27:18.085262 containerd[2685]: time="2025-08-13T00:27:18.085251192Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.30.2: active requests=0, bytes read=48128336" Aug 13 00:27:18.086125 containerd[2685]: time="2025-08-13T00:27:18.086105232Z" level=info msg="ImageCreate event name:\"sha256:ba9e7793995ca67a9b78aa06adda4e89cbd435b1e88ab1032ca665140517fa7a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:27:18.088003 containerd[2685]: time="2025-08-13T00:27:18.087975153Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:5d3ecdec3cbbe8f7009077102e35e8a2141161b59c548cf3f97829177677cbce\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 00:27:18.088750 containerd[2685]: time="2025-08-13T00:27:18.088725553Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\" with image id \"sha256:ba9e7793995ca67a9b78aa06adda4e89cbd435b1e88ab1032ca665140517fa7a\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:5d3ecdec3cbbe8f7009077102e35e8a2141161b59c548cf3f97829177677cbce\", size \"49497545\" in 2.004956762s" Aug 13 00:27:18.088779 containerd[2685]: time="2025-08-13T00:27:18.088756233Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.30.2\" returns image reference \"sha256:ba9e7793995ca67a9b78aa06adda4e89cbd435b1e88ab1032ca665140517fa7a\"" Aug 13 00:27:18.094642 containerd[2685]: time="2025-08-13T00:27:18.094616874Z" level=info msg="CreateContainer within sandbox \"7abc47012f7c901b3d8b53e8ae84366e84722e4aaed0067a97a0c021d17effd0\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Aug 13 00:27:18.099334 containerd[2685]: time="2025-08-13T00:27:18.099305475Z" level=info msg="CreateContainer within sandbox \"7abc47012f7c901b3d8b53e8ae84366e84722e4aaed0067a97a0c021d17effd0\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"f6685efcb3572069575aae7e636ec9a97c2b75b7165872aead50ac3df44d9517\"" Aug 13 00:27:18.099650 containerd[2685]: time="2025-08-13T00:27:18.099629195Z" level=info msg="StartContainer for \"f6685efcb3572069575aae7e636ec9a97c2b75b7165872aead50ac3df44d9517\"" Aug 13 00:27:18.121402 systemd[1]: Started cri-containerd-f6685efcb3572069575aae7e636ec9a97c2b75b7165872aead50ac3df44d9517.scope - libcontainer container f6685efcb3572069575aae7e636ec9a97c2b75b7165872aead50ac3df44d9517. Aug 13 00:27:18.145761 containerd[2685]: time="2025-08-13T00:27:18.145728966Z" level=info msg="StartContainer for \"f6685efcb3572069575aae7e636ec9a97c2b75b7165872aead50ac3df44d9517\" returns successfully" Aug 13 00:27:19.099103 kubelet[4148]: I0813 00:27:19.099002 4148 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-f5cbb957c-wms9w" podStartSLOduration=22.33452595 podStartE2EDuration="27.098987671s" podCreationTimestamp="2025-08-13 00:26:52 +0000 UTC" firstStartedPulling="2025-08-13 00:27:13.324837272 +0000 UTC m=+41.409794411" lastFinishedPulling="2025-08-13 00:27:18.089298993 +0000 UTC m=+46.174256132" observedRunningTime="2025-08-13 00:27:19.098657911 +0000 UTC m=+47.183615010" watchObservedRunningTime="2025-08-13 00:27:19.098987671 +0000 UTC m=+47.183944810" Aug 13 00:27:31.978631 containerd[2685]: time="2025-08-13T00:27:31.978586029Z" level=info msg="StopPodSandbox for \"7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530\"" Aug 13 00:27:32.037456 containerd[2685]: 2025-08-13 00:27:32.009 [WARNING][9106] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--lcpkt-eth0", GenerateName:"calico-apiserver-54c5fc9686-", Namespace:"calico-apiserver", SelfLink:"", UID:"3dc1e9ad-8b04-42ba-98fc-165c43766984", ResourceVersion:"1030", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 26, 48, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"54c5fc9686", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-588adb827d", ContainerID:"10b7183e37555429f20aaf37a6f7d52ae00153679751bd35876ef5c03f417df3", Pod:"calico-apiserver-54c5fc9686-lcpkt", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.85.73/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali4364a52d32c", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:27:32.037456 containerd[2685]: 2025-08-13 00:27:32.009 [INFO][9106] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" Aug 13 00:27:32.037456 containerd[2685]: 2025-08-13 00:27:32.009 [INFO][9106] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" iface="eth0" netns="" Aug 13 00:27:32.037456 containerd[2685]: 2025-08-13 00:27:32.009 [INFO][9106] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" Aug 13 00:27:32.037456 containerd[2685]: 2025-08-13 00:27:32.009 [INFO][9106] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" Aug 13 00:27:32.037456 containerd[2685]: 2025-08-13 00:27:32.026 [INFO][9131] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" HandleID="k8s-pod-network.7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--lcpkt-eth0" Aug 13 00:27:32.037456 containerd[2685]: 2025-08-13 00:27:32.026 [INFO][9131] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:32.037456 containerd[2685]: 2025-08-13 00:27:32.026 [INFO][9131] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:32.037456 containerd[2685]: 2025-08-13 00:27:32.034 [WARNING][9131] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" HandleID="k8s-pod-network.7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--lcpkt-eth0" Aug 13 00:27:32.037456 containerd[2685]: 2025-08-13 00:27:32.034 [INFO][9131] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" HandleID="k8s-pod-network.7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--lcpkt-eth0" Aug 13 00:27:32.037456 containerd[2685]: 2025-08-13 00:27:32.035 [INFO][9131] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:32.037456 containerd[2685]: 2025-08-13 00:27:32.036 [INFO][9106] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" Aug 13 00:27:32.037763 containerd[2685]: time="2025-08-13T00:27:32.037499034Z" level=info msg="TearDown network for sandbox \"7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530\" successfully" Aug 13 00:27:32.037763 containerd[2685]: time="2025-08-13T00:27:32.037526474Z" level=info msg="StopPodSandbox for \"7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530\" returns successfully" Aug 13 00:27:32.037950 containerd[2685]: time="2025-08-13T00:27:32.037922355Z" level=info msg="RemovePodSandbox for \"7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530\"" Aug 13 00:27:32.037976 containerd[2685]: time="2025-08-13T00:27:32.037957275Z" level=info msg="Forcibly stopping sandbox \"7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530\"" Aug 13 00:27:32.095936 containerd[2685]: 2025-08-13 00:27:32.067 [WARNING][9161] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--lcpkt-eth0", GenerateName:"calico-apiserver-54c5fc9686-", Namespace:"calico-apiserver", SelfLink:"", UID:"3dc1e9ad-8b04-42ba-98fc-165c43766984", ResourceVersion:"1030", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 26, 48, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"54c5fc9686", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-588adb827d", ContainerID:"10b7183e37555429f20aaf37a6f7d52ae00153679751bd35876ef5c03f417df3", Pod:"calico-apiserver-54c5fc9686-lcpkt", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.85.73/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali4364a52d32c", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:27:32.095936 containerd[2685]: 2025-08-13 00:27:32.067 [INFO][9161] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" Aug 13 00:27:32.095936 containerd[2685]: 2025-08-13 00:27:32.067 [INFO][9161] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" iface="eth0" netns="" Aug 13 00:27:32.095936 containerd[2685]: 2025-08-13 00:27:32.067 [INFO][9161] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" Aug 13 00:27:32.095936 containerd[2685]: 2025-08-13 00:27:32.067 [INFO][9161] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" Aug 13 00:27:32.095936 containerd[2685]: 2025-08-13 00:27:32.084 [INFO][9182] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" HandleID="k8s-pod-network.7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--lcpkt-eth0" Aug 13 00:27:32.095936 containerd[2685]: 2025-08-13 00:27:32.084 [INFO][9182] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:32.095936 containerd[2685]: 2025-08-13 00:27:32.085 [INFO][9182] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:32.095936 containerd[2685]: 2025-08-13 00:27:32.092 [WARNING][9182] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" HandleID="k8s-pod-network.7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--lcpkt-eth0" Aug 13 00:27:32.095936 containerd[2685]: 2025-08-13 00:27:32.092 [INFO][9182] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" HandleID="k8s-pod-network.7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--lcpkt-eth0" Aug 13 00:27:32.095936 containerd[2685]: 2025-08-13 00:27:32.093 [INFO][9182] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:32.095936 containerd[2685]: 2025-08-13 00:27:32.094 [INFO][9161] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530" Aug 13 00:27:32.096303 containerd[2685]: time="2025-08-13T00:27:32.095988080Z" level=info msg="TearDown network for sandbox \"7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530\" successfully" Aug 13 00:27:32.104235 containerd[2685]: time="2025-08-13T00:27:32.103817521Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Aug 13 00:27:32.104235 containerd[2685]: time="2025-08-13T00:27:32.103890961Z" level=info msg="RemovePodSandbox \"7fecfa0a84bbfb65ef6c386d34a4f64a009a081096e141173dc7c4527df48530\" returns successfully" Aug 13 00:27:32.108545 containerd[2685]: time="2025-08-13T00:27:32.107361081Z" level=info msg="StopPodSandbox for \"96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce\"" Aug 13 00:27:32.175142 containerd[2685]: 2025-08-13 00:27:32.145 [WARNING][9211] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--h5l86-eth0", GenerateName:"calico-apiserver-74488cd4c9-", Namespace:"calico-apiserver", SelfLink:"", UID:"517c2f94-3455-4677-a3ec-3a8771258060", ResourceVersion:"972", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 26, 47, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"74488cd4c9", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-588adb827d", ContainerID:"6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881", Pod:"calico-apiserver-74488cd4c9-h5l86", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.85.66/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali3c1d63c3857", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:27:32.175142 containerd[2685]: 2025-08-13 00:27:32.145 [INFO][9211] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" Aug 13 00:27:32.175142 containerd[2685]: 2025-08-13 00:27:32.145 [INFO][9211] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" iface="eth0" netns="" Aug 13 00:27:32.175142 containerd[2685]: 2025-08-13 00:27:32.145 [INFO][9211] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" Aug 13 00:27:32.175142 containerd[2685]: 2025-08-13 00:27:32.145 [INFO][9211] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" Aug 13 00:27:32.175142 containerd[2685]: 2025-08-13 00:27:32.163 [INFO][9231] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" HandleID="k8s-pod-network.96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--h5l86-eth0" Aug 13 00:27:32.175142 containerd[2685]: 2025-08-13 00:27:32.163 [INFO][9231] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:32.175142 containerd[2685]: 2025-08-13 00:27:32.163 [INFO][9231] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:32.175142 containerd[2685]: 2025-08-13 00:27:32.171 [WARNING][9231] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" HandleID="k8s-pod-network.96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--h5l86-eth0" Aug 13 00:27:32.175142 containerd[2685]: 2025-08-13 00:27:32.171 [INFO][9231] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" HandleID="k8s-pod-network.96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--h5l86-eth0" Aug 13 00:27:32.175142 containerd[2685]: 2025-08-13 00:27:32.172 [INFO][9231] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:32.175142 containerd[2685]: 2025-08-13 00:27:32.173 [INFO][9211] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" Aug 13 00:27:32.175551 containerd[2685]: time="2025-08-13T00:27:32.175174768Z" level=info msg="TearDown network for sandbox \"96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce\" successfully" Aug 13 00:27:32.175551 containerd[2685]: time="2025-08-13T00:27:32.175205088Z" level=info msg="StopPodSandbox for \"96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce\" returns successfully" Aug 13 00:27:32.175594 containerd[2685]: time="2025-08-13T00:27:32.175537408Z" level=info msg="RemovePodSandbox for \"96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce\"" Aug 13 00:27:32.175594 containerd[2685]: time="2025-08-13T00:27:32.175569328Z" level=info msg="Forcibly stopping sandbox \"96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce\"" Aug 13 00:27:32.234770 containerd[2685]: 2025-08-13 00:27:32.205 [WARNING][9261] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--h5l86-eth0", GenerateName:"calico-apiserver-74488cd4c9-", Namespace:"calico-apiserver", SelfLink:"", UID:"517c2f94-3455-4677-a3ec-3a8771258060", ResourceVersion:"972", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 26, 47, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"74488cd4c9", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-588adb827d", ContainerID:"6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881", Pod:"calico-apiserver-74488cd4c9-h5l86", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.85.66/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali3c1d63c3857", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:27:32.234770 containerd[2685]: 2025-08-13 00:27:32.205 [INFO][9261] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" Aug 13 00:27:32.234770 containerd[2685]: 2025-08-13 00:27:32.205 [INFO][9261] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" iface="eth0" netns="" Aug 13 00:27:32.234770 containerd[2685]: 2025-08-13 00:27:32.205 [INFO][9261] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" Aug 13 00:27:32.234770 containerd[2685]: 2025-08-13 00:27:32.205 [INFO][9261] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" Aug 13 00:27:32.234770 containerd[2685]: 2025-08-13 00:27:32.223 [INFO][9285] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" HandleID="k8s-pod-network.96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--h5l86-eth0" Aug 13 00:27:32.234770 containerd[2685]: 2025-08-13 00:27:32.223 [INFO][9285] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:32.234770 containerd[2685]: 2025-08-13 00:27:32.223 [INFO][9285] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:32.234770 containerd[2685]: 2025-08-13 00:27:32.231 [WARNING][9285] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" HandleID="k8s-pod-network.96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--h5l86-eth0" Aug 13 00:27:32.234770 containerd[2685]: 2025-08-13 00:27:32.231 [INFO][9285] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" HandleID="k8s-pod-network.96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--h5l86-eth0" Aug 13 00:27:32.234770 containerd[2685]: 2025-08-13 00:27:32.232 [INFO][9285] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:32.234770 containerd[2685]: 2025-08-13 00:27:32.233 [INFO][9261] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce" Aug 13 00:27:32.235065 containerd[2685]: time="2025-08-13T00:27:32.234765013Z" level=info msg="TearDown network for sandbox \"96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce\" successfully" Aug 13 00:27:32.236411 containerd[2685]: time="2025-08-13T00:27:32.236386454Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Aug 13 00:27:32.236459 containerd[2685]: time="2025-08-13T00:27:32.236447774Z" level=info msg="RemovePodSandbox \"96540c70b7d48970d56c07568397a982321a2ee1bac19b11d77a21446b0612ce\" returns successfully" Aug 13 00:27:32.236827 containerd[2685]: time="2025-08-13T00:27:32.236805974Z" level=info msg="StopPodSandbox for \"fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02\"" Aug 13 00:27:32.296190 containerd[2685]: 2025-08-13 00:27:32.266 [WARNING][9315] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--588adb827d-k8s-calico--kube--controllers--f5cbb957c--wms9w-eth0", GenerateName:"calico-kube-controllers-f5cbb957c-", Namespace:"calico-system", SelfLink:"", UID:"925e0d95-7160-42b1-a3c8-6b7795ca67bc", ResourceVersion:"1061", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 26, 52, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"f5cbb957c", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-588adb827d", ContainerID:"7abc47012f7c901b3d8b53e8ae84366e84722e4aaed0067a97a0c021d17effd0", Pod:"calico-kube-controllers-f5cbb957c-wms9w", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.85.70/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calib4046b1b4a7", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:27:32.296190 containerd[2685]: 2025-08-13 00:27:32.267 [INFO][9315] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" Aug 13 00:27:32.296190 containerd[2685]: 2025-08-13 00:27:32.267 [INFO][9315] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" iface="eth0" netns="" Aug 13 00:27:32.296190 containerd[2685]: 2025-08-13 00:27:32.267 [INFO][9315] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" Aug 13 00:27:32.296190 containerd[2685]: 2025-08-13 00:27:32.267 [INFO][9315] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" Aug 13 00:27:32.296190 containerd[2685]: 2025-08-13 00:27:32.284 [INFO][9341] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" HandleID="k8s-pod-network.fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--kube--controllers--f5cbb957c--wms9w-eth0" Aug 13 00:27:32.296190 containerd[2685]: 2025-08-13 00:27:32.285 [INFO][9341] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:32.296190 containerd[2685]: 2025-08-13 00:27:32.285 [INFO][9341] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:32.296190 containerd[2685]: 2025-08-13 00:27:32.292 [WARNING][9341] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" HandleID="k8s-pod-network.fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--kube--controllers--f5cbb957c--wms9w-eth0" Aug 13 00:27:32.296190 containerd[2685]: 2025-08-13 00:27:32.292 [INFO][9341] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" HandleID="k8s-pod-network.fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--kube--controllers--f5cbb957c--wms9w-eth0" Aug 13 00:27:32.296190 containerd[2685]: 2025-08-13 00:27:32.293 [INFO][9341] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:32.296190 containerd[2685]: 2025-08-13 00:27:32.294 [INFO][9315] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" Aug 13 00:27:32.296540 containerd[2685]: time="2025-08-13T00:27:32.296236299Z" level=info msg="TearDown network for sandbox \"fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02\" successfully" Aug 13 00:27:32.296540 containerd[2685]: time="2025-08-13T00:27:32.296262699Z" level=info msg="StopPodSandbox for \"fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02\" returns successfully" Aug 13 00:27:32.296639 containerd[2685]: time="2025-08-13T00:27:32.296616099Z" level=info msg="RemovePodSandbox for \"fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02\"" Aug 13 00:27:32.296664 containerd[2685]: time="2025-08-13T00:27:32.296648939Z" level=info msg="Forcibly stopping sandbox \"fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02\"" Aug 13 00:27:32.359819 containerd[2685]: 2025-08-13 00:27:32.330 [WARNING][9369] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--588adb827d-k8s-calico--kube--controllers--f5cbb957c--wms9w-eth0", GenerateName:"calico-kube-controllers-f5cbb957c-", Namespace:"calico-system", SelfLink:"", UID:"925e0d95-7160-42b1-a3c8-6b7795ca67bc", ResourceVersion:"1061", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 26, 52, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"f5cbb957c", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-588adb827d", ContainerID:"7abc47012f7c901b3d8b53e8ae84366e84722e4aaed0067a97a0c021d17effd0", Pod:"calico-kube-controllers-f5cbb957c-wms9w", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.85.70/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"calib4046b1b4a7", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:27:32.359819 containerd[2685]: 2025-08-13 00:27:32.330 [INFO][9369] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" Aug 13 00:27:32.359819 containerd[2685]: 2025-08-13 00:27:32.330 [INFO][9369] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" iface="eth0" netns="" Aug 13 00:27:32.359819 containerd[2685]: 2025-08-13 00:27:32.330 [INFO][9369] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" Aug 13 00:27:32.359819 containerd[2685]: 2025-08-13 00:27:32.330 [INFO][9369] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" Aug 13 00:27:32.359819 containerd[2685]: 2025-08-13 00:27:32.348 [INFO][9391] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" HandleID="k8s-pod-network.fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--kube--controllers--f5cbb957c--wms9w-eth0" Aug 13 00:27:32.359819 containerd[2685]: 2025-08-13 00:27:32.348 [INFO][9391] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:32.359819 containerd[2685]: 2025-08-13 00:27:32.348 [INFO][9391] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:32.359819 containerd[2685]: 2025-08-13 00:27:32.356 [WARNING][9391] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" HandleID="k8s-pod-network.fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--kube--controllers--f5cbb957c--wms9w-eth0" Aug 13 00:27:32.359819 containerd[2685]: 2025-08-13 00:27:32.356 [INFO][9391] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" HandleID="k8s-pod-network.fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--kube--controllers--f5cbb957c--wms9w-eth0" Aug 13 00:27:32.359819 containerd[2685]: 2025-08-13 00:27:32.357 [INFO][9391] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:32.359819 containerd[2685]: 2025-08-13 00:27:32.358 [INFO][9369] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02" Aug 13 00:27:32.360288 containerd[2685]: time="2025-08-13T00:27:32.359863425Z" level=info msg="TearDown network for sandbox \"fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02\" successfully" Aug 13 00:27:32.362181 containerd[2685]: time="2025-08-13T00:27:32.362151826Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Aug 13 00:27:32.362220 containerd[2685]: time="2025-08-13T00:27:32.362207186Z" level=info msg="RemovePodSandbox \"fc02ef2de6afbbc103bc97ca8c72b4c9be51f0639f89b265bf8bff00857d5f02\" returns successfully" Aug 13 00:27:32.362647 containerd[2685]: time="2025-08-13T00:27:32.362622306Z" level=info msg="StopPodSandbox for \"3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd\"" Aug 13 00:27:32.420740 containerd[2685]: 2025-08-13 00:27:32.392 [WARNING][9420] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--588adb827d-k8s-goldmane--768f4c5c69--nm6vx-eth0", GenerateName:"goldmane-768f4c5c69-", Namespace:"calico-system", SelfLink:"", UID:"b5da64a2-733d-4c0b-b26c-80dcab04992c", ResourceVersion:"1041", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 26, 51, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"768f4c5c69", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-588adb827d", ContainerID:"62cb1f87cf16165f451f6c892ad1adfd06ebd722336b883f76a6498daa505f2e", Pod:"goldmane-768f4c5c69-nm6vx", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.85.69/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"calicdc52df372f", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:27:32.420740 containerd[2685]: 2025-08-13 00:27:32.392 [INFO][9420] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" Aug 13 00:27:32.420740 containerd[2685]: 2025-08-13 00:27:32.392 [INFO][9420] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" iface="eth0" netns="" Aug 13 00:27:32.420740 containerd[2685]: 2025-08-13 00:27:32.392 [INFO][9420] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" Aug 13 00:27:32.420740 containerd[2685]: 2025-08-13 00:27:32.392 [INFO][9420] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" Aug 13 00:27:32.420740 containerd[2685]: 2025-08-13 00:27:32.409 [INFO][9443] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" HandleID="k8s-pod-network.3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" Workload="ci--4081.3.5--a--588adb827d-k8s-goldmane--768f4c5c69--nm6vx-eth0" Aug 13 00:27:32.420740 containerd[2685]: 2025-08-13 00:27:32.409 [INFO][9443] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:32.420740 containerd[2685]: 2025-08-13 00:27:32.409 [INFO][9443] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:32.420740 containerd[2685]: 2025-08-13 00:27:32.417 [WARNING][9443] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" HandleID="k8s-pod-network.3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" Workload="ci--4081.3.5--a--588adb827d-k8s-goldmane--768f4c5c69--nm6vx-eth0" Aug 13 00:27:32.420740 containerd[2685]: 2025-08-13 00:27:32.417 [INFO][9443] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" HandleID="k8s-pod-network.3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" Workload="ci--4081.3.5--a--588adb827d-k8s-goldmane--768f4c5c69--nm6vx-eth0" Aug 13 00:27:32.420740 containerd[2685]: 2025-08-13 00:27:32.418 [INFO][9443] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:32.420740 containerd[2685]: 2025-08-13 00:27:32.419 [INFO][9420] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" Aug 13 00:27:32.421007 containerd[2685]: time="2025-08-13T00:27:32.420786951Z" level=info msg="TearDown network for sandbox \"3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd\" successfully" Aug 13 00:27:32.421007 containerd[2685]: time="2025-08-13T00:27:32.420809271Z" level=info msg="StopPodSandbox for \"3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd\" returns successfully" Aug 13 00:27:32.421197 containerd[2685]: time="2025-08-13T00:27:32.421172431Z" level=info msg="RemovePodSandbox for \"3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd\"" Aug 13 00:27:32.421225 containerd[2685]: time="2025-08-13T00:27:32.421204431Z" level=info msg="Forcibly stopping sandbox \"3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd\"" Aug 13 00:27:32.479582 containerd[2685]: 2025-08-13 00:27:32.450 [WARNING][9471] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--588adb827d-k8s-goldmane--768f4c5c69--nm6vx-eth0", GenerateName:"goldmane-768f4c5c69-", Namespace:"calico-system", SelfLink:"", UID:"b5da64a2-733d-4c0b-b26c-80dcab04992c", ResourceVersion:"1041", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 26, 51, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"goldmane", "k8s-app":"goldmane", "pod-template-hash":"768f4c5c69", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"goldmane"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-588adb827d", ContainerID:"62cb1f87cf16165f451f6c892ad1adfd06ebd722336b883f76a6498daa505f2e", Pod:"goldmane-768f4c5c69-nm6vx", Endpoint:"eth0", ServiceAccountName:"goldmane", IPNetworks:[]string{"192.168.85.69/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.goldmane"}, InterfaceName:"calicdc52df372f", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:27:32.479582 containerd[2685]: 2025-08-13 00:27:32.450 [INFO][9471] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" Aug 13 00:27:32.479582 containerd[2685]: 2025-08-13 00:27:32.450 [INFO][9471] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" iface="eth0" netns="" Aug 13 00:27:32.479582 containerd[2685]: 2025-08-13 00:27:32.450 [INFO][9471] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" Aug 13 00:27:32.479582 containerd[2685]: 2025-08-13 00:27:32.451 [INFO][9471] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" Aug 13 00:27:32.479582 containerd[2685]: 2025-08-13 00:27:32.468 [INFO][9493] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" HandleID="k8s-pod-network.3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" Workload="ci--4081.3.5--a--588adb827d-k8s-goldmane--768f4c5c69--nm6vx-eth0" Aug 13 00:27:32.479582 containerd[2685]: 2025-08-13 00:27:32.468 [INFO][9493] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:32.479582 containerd[2685]: 2025-08-13 00:27:32.468 [INFO][9493] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:32.479582 containerd[2685]: 2025-08-13 00:27:32.475 [WARNING][9493] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" HandleID="k8s-pod-network.3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" Workload="ci--4081.3.5--a--588adb827d-k8s-goldmane--768f4c5c69--nm6vx-eth0" Aug 13 00:27:32.479582 containerd[2685]: 2025-08-13 00:27:32.475 [INFO][9493] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" HandleID="k8s-pod-network.3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" Workload="ci--4081.3.5--a--588adb827d-k8s-goldmane--768f4c5c69--nm6vx-eth0" Aug 13 00:27:32.479582 containerd[2685]: 2025-08-13 00:27:32.477 [INFO][9493] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:32.479582 containerd[2685]: 2025-08-13 00:27:32.478 [INFO][9471] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd" Aug 13 00:27:32.480027 containerd[2685]: time="2025-08-13T00:27:32.479613037Z" level=info msg="TearDown network for sandbox \"3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd\" successfully" Aug 13 00:27:32.481256 containerd[2685]: time="2025-08-13T00:27:32.481220317Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Aug 13 00:27:32.481301 containerd[2685]: time="2025-08-13T00:27:32.481287637Z" level=info msg="RemovePodSandbox \"3749b88c8d1c5679488634ec127f63e80baba0082f4872ca9d3bf6be0570dedd\" returns successfully" Aug 13 00:27:32.481631 containerd[2685]: time="2025-08-13T00:27:32.481608917Z" level=info msg="StopPodSandbox for \"d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3\"" Aug 13 00:27:32.540800 containerd[2685]: 2025-08-13 00:27:32.511 [WARNING][9523] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--ftxjx-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"085d2185-0930-470d-b50d-68703ac546fb", ResourceVersion:"1005", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 26, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-588adb827d", ContainerID:"3afc2e5aedc5d5722bff856b27c1c16aad8ea7dc957c30eaf468179a695d337a", Pod:"coredns-674b8bbfcf-ftxjx", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.85.72/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calic1d119fb847", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:27:32.540800 containerd[2685]: 2025-08-13 00:27:32.511 [INFO][9523] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" Aug 13 00:27:32.540800 containerd[2685]: 2025-08-13 00:27:32.511 [INFO][9523] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" iface="eth0" netns="" Aug 13 00:27:32.540800 containerd[2685]: 2025-08-13 00:27:32.511 [INFO][9523] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" Aug 13 00:27:32.540800 containerd[2685]: 2025-08-13 00:27:32.512 [INFO][9523] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" Aug 13 00:27:32.540800 containerd[2685]: 2025-08-13 00:27:32.529 [INFO][9544] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" HandleID="k8s-pod-network.d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" Workload="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--ftxjx-eth0" Aug 13 00:27:32.540800 containerd[2685]: 2025-08-13 00:27:32.529 [INFO][9544] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:32.540800 containerd[2685]: 2025-08-13 00:27:32.529 [INFO][9544] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:32.540800 containerd[2685]: 2025-08-13 00:27:32.537 [WARNING][9544] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" HandleID="k8s-pod-network.d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" Workload="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--ftxjx-eth0" Aug 13 00:27:32.540800 containerd[2685]: 2025-08-13 00:27:32.537 [INFO][9544] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" HandleID="k8s-pod-network.d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" Workload="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--ftxjx-eth0" Aug 13 00:27:32.540800 containerd[2685]: 2025-08-13 00:27:32.538 [INFO][9544] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:32.540800 containerd[2685]: 2025-08-13 00:27:32.539 [INFO][9523] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" Aug 13 00:27:32.541183 containerd[2685]: time="2025-08-13T00:27:32.540843723Z" level=info msg="TearDown network for sandbox \"d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3\" successfully" Aug 13 00:27:32.541183 containerd[2685]: time="2025-08-13T00:27:32.540872963Z" level=info msg="StopPodSandbox for \"d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3\" returns successfully" Aug 13 00:27:32.541244 containerd[2685]: time="2025-08-13T00:27:32.541189963Z" level=info msg="RemovePodSandbox for \"d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3\"" Aug 13 00:27:32.541244 containerd[2685]: time="2025-08-13T00:27:32.541215323Z" level=info msg="Forcibly stopping sandbox \"d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3\"" Aug 13 00:27:32.601182 containerd[2685]: 2025-08-13 00:27:32.571 [WARNING][9575] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--ftxjx-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"085d2185-0930-470d-b50d-68703ac546fb", ResourceVersion:"1005", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 26, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-588adb827d", ContainerID:"3afc2e5aedc5d5722bff856b27c1c16aad8ea7dc957c30eaf468179a695d337a", Pod:"coredns-674b8bbfcf-ftxjx", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.85.72/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calic1d119fb847", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:27:32.601182 containerd[2685]: 2025-08-13 00:27:32.572 [INFO][9575] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" Aug 13 00:27:32.601182 containerd[2685]: 2025-08-13 00:27:32.572 [INFO][9575] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" iface="eth0" netns="" Aug 13 00:27:32.601182 containerd[2685]: 2025-08-13 00:27:32.572 [INFO][9575] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" Aug 13 00:27:32.601182 containerd[2685]: 2025-08-13 00:27:32.572 [INFO][9575] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" Aug 13 00:27:32.601182 containerd[2685]: 2025-08-13 00:27:32.590 [INFO][9602] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" HandleID="k8s-pod-network.d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" Workload="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--ftxjx-eth0" Aug 13 00:27:32.601182 containerd[2685]: 2025-08-13 00:27:32.590 [INFO][9602] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:32.601182 containerd[2685]: 2025-08-13 00:27:32.590 [INFO][9602] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:32.601182 containerd[2685]: 2025-08-13 00:27:32.597 [WARNING][9602] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" HandleID="k8s-pod-network.d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" Workload="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--ftxjx-eth0" Aug 13 00:27:32.601182 containerd[2685]: 2025-08-13 00:27:32.597 [INFO][9602] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" HandleID="k8s-pod-network.d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" Workload="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--ftxjx-eth0" Aug 13 00:27:32.601182 containerd[2685]: 2025-08-13 00:27:32.598 [INFO][9602] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:32.601182 containerd[2685]: 2025-08-13 00:27:32.599 [INFO][9575] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3" Aug 13 00:27:32.601455 containerd[2685]: time="2025-08-13T00:27:32.601224849Z" level=info msg="TearDown network for sandbox \"d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3\" successfully" Aug 13 00:27:32.604641 containerd[2685]: time="2025-08-13T00:27:32.604606929Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Aug 13 00:27:32.604701 containerd[2685]: time="2025-08-13T00:27:32.604681569Z" level=info msg="RemovePodSandbox \"d77cee5f345584d563328bf828762cc47a8425317696b689f313f4f174412ac3\" returns successfully" Aug 13 00:27:32.605050 containerd[2685]: time="2025-08-13T00:27:32.605026489Z" level=info msg="StopPodSandbox for \"e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f\"" Aug 13 00:27:32.664099 containerd[2685]: 2025-08-13 00:27:32.635 [WARNING][9630] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--v7mtb-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"75d28fd6-77d6-4c58-8897-8ce0729a65ea", ResourceVersion:"1009", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 26, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-588adb827d", ContainerID:"cd71e691f7ceac8c9846ac32c539f554ff5f9ea8ed0badd72a0ca75474176af9", Pod:"coredns-674b8bbfcf-v7mtb", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.85.71/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calid5091b26cdd", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:27:32.664099 containerd[2685]: 2025-08-13 00:27:32.635 [INFO][9630] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" Aug 13 00:27:32.664099 containerd[2685]: 2025-08-13 00:27:32.635 [INFO][9630] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" iface="eth0" netns="" Aug 13 00:27:32.664099 containerd[2685]: 2025-08-13 00:27:32.635 [INFO][9630] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" Aug 13 00:27:32.664099 containerd[2685]: 2025-08-13 00:27:32.635 [INFO][9630] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" Aug 13 00:27:32.664099 containerd[2685]: 2025-08-13 00:27:32.653 [INFO][9652] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" HandleID="k8s-pod-network.e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" Workload="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--v7mtb-eth0" Aug 13 00:27:32.664099 containerd[2685]: 2025-08-13 00:27:32.653 [INFO][9652] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:32.664099 containerd[2685]: 2025-08-13 00:27:32.653 [INFO][9652] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:32.664099 containerd[2685]: 2025-08-13 00:27:32.660 [WARNING][9652] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" HandleID="k8s-pod-network.e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" Workload="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--v7mtb-eth0" Aug 13 00:27:32.664099 containerd[2685]: 2025-08-13 00:27:32.660 [INFO][9652] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" HandleID="k8s-pod-network.e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" Workload="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--v7mtb-eth0" Aug 13 00:27:32.664099 containerd[2685]: 2025-08-13 00:27:32.661 [INFO][9652] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:32.664099 containerd[2685]: 2025-08-13 00:27:32.662 [INFO][9630] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" Aug 13 00:27:32.664390 containerd[2685]: time="2025-08-13T00:27:32.664139015Z" level=info msg="TearDown network for sandbox \"e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f\" successfully" Aug 13 00:27:32.664390 containerd[2685]: time="2025-08-13T00:27:32.664161095Z" level=info msg="StopPodSandbox for \"e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f\" returns successfully" Aug 13 00:27:32.664537 containerd[2685]: time="2025-08-13T00:27:32.664513775Z" level=info msg="RemovePodSandbox for \"e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f\"" Aug 13 00:27:32.664564 containerd[2685]: time="2025-08-13T00:27:32.664543895Z" level=info msg="Forcibly stopping sandbox \"e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f\"" Aug 13 00:27:32.724017 containerd[2685]: 2025-08-13 00:27:32.694 [WARNING][9681] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--v7mtb-eth0", GenerateName:"coredns-674b8bbfcf-", Namespace:"kube-system", SelfLink:"", UID:"75d28fd6-77d6-4c58-8897-8ce0729a65ea", ResourceVersion:"1009", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 26, 40, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"674b8bbfcf", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-588adb827d", ContainerID:"cd71e691f7ceac8c9846ac32c539f554ff5f9ea8ed0badd72a0ca75474176af9", Pod:"coredns-674b8bbfcf-v7mtb", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.85.71/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calid5091b26cdd", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:27:32.724017 containerd[2685]: 2025-08-13 00:27:32.694 [INFO][9681] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" Aug 13 00:27:32.724017 containerd[2685]: 2025-08-13 00:27:32.694 [INFO][9681] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" iface="eth0" netns="" Aug 13 00:27:32.724017 containerd[2685]: 2025-08-13 00:27:32.694 [INFO][9681] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" Aug 13 00:27:32.724017 containerd[2685]: 2025-08-13 00:27:32.694 [INFO][9681] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" Aug 13 00:27:32.724017 containerd[2685]: 2025-08-13 00:27:32.712 [INFO][9702] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" HandleID="k8s-pod-network.e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" Workload="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--v7mtb-eth0" Aug 13 00:27:32.724017 containerd[2685]: 2025-08-13 00:27:32.712 [INFO][9702] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:32.724017 containerd[2685]: 2025-08-13 00:27:32.712 [INFO][9702] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:32.724017 containerd[2685]: 2025-08-13 00:27:32.720 [WARNING][9702] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" HandleID="k8s-pod-network.e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" Workload="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--v7mtb-eth0" Aug 13 00:27:32.724017 containerd[2685]: 2025-08-13 00:27:32.720 [INFO][9702] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" HandleID="k8s-pod-network.e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" Workload="ci--4081.3.5--a--588adb827d-k8s-coredns--674b8bbfcf--v7mtb-eth0" Aug 13 00:27:32.724017 containerd[2685]: 2025-08-13 00:27:32.721 [INFO][9702] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:32.724017 containerd[2685]: 2025-08-13 00:27:32.722 [INFO][9681] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f" Aug 13 00:27:32.724409 containerd[2685]: time="2025-08-13T00:27:32.724043700Z" level=info msg="TearDown network for sandbox \"e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f\" successfully" Aug 13 00:27:32.732044 containerd[2685]: time="2025-08-13T00:27:32.732012061Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Aug 13 00:27:32.732098 containerd[2685]: time="2025-08-13T00:27:32.732069301Z" level=info msg="RemovePodSandbox \"e5547ed3a6057dbcc7d94de7bc777f2e09746fbf0f29e991bf63125058e03a5f\" returns successfully" Aug 13 00:27:32.732427 containerd[2685]: time="2025-08-13T00:27:32.732401861Z" level=info msg="StopPodSandbox for \"4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652\"" Aug 13 00:27:32.792851 containerd[2685]: 2025-08-13 00:27:32.764 [WARNING][9731] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--588adb827d-k8s-csi--node--driver--4wkp8-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"bb51369a-5ac7-4ac1-b819-64d9c4e312d0", ResourceVersion:"1033", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 26, 52, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"8967bcb6f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-588adb827d", ContainerID:"28c2e0ac93ee78b239c852a9b7b6e26ab8f23ab9c93620b45125367839cf7097", Pod:"csi-node-driver-4wkp8", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.85.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calie22dd168197", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:27:32.792851 containerd[2685]: 2025-08-13 00:27:32.764 [INFO][9731] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" Aug 13 00:27:32.792851 containerd[2685]: 2025-08-13 00:27:32.764 [INFO][9731] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" iface="eth0" netns="" Aug 13 00:27:32.792851 containerd[2685]: 2025-08-13 00:27:32.764 [INFO][9731] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" Aug 13 00:27:32.792851 containerd[2685]: 2025-08-13 00:27:32.764 [INFO][9731] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" Aug 13 00:27:32.792851 containerd[2685]: 2025-08-13 00:27:32.781 [INFO][9753] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" HandleID="k8s-pod-network.4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" Workload="ci--4081.3.5--a--588adb827d-k8s-csi--node--driver--4wkp8-eth0" Aug 13 00:27:32.792851 containerd[2685]: 2025-08-13 00:27:32.782 [INFO][9753] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:32.792851 containerd[2685]: 2025-08-13 00:27:32.782 [INFO][9753] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:32.792851 containerd[2685]: 2025-08-13 00:27:32.789 [WARNING][9753] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" HandleID="k8s-pod-network.4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" Workload="ci--4081.3.5--a--588adb827d-k8s-csi--node--driver--4wkp8-eth0" Aug 13 00:27:32.792851 containerd[2685]: 2025-08-13 00:27:32.789 [INFO][9753] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" HandleID="k8s-pod-network.4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" Workload="ci--4081.3.5--a--588adb827d-k8s-csi--node--driver--4wkp8-eth0" Aug 13 00:27:32.792851 containerd[2685]: 2025-08-13 00:27:32.790 [INFO][9753] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:32.792851 containerd[2685]: 2025-08-13 00:27:32.791 [INFO][9731] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" Aug 13 00:27:32.792851 containerd[2685]: time="2025-08-13T00:27:32.792798467Z" level=info msg="TearDown network for sandbox \"4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652\" successfully" Aug 13 00:27:32.792851 containerd[2685]: time="2025-08-13T00:27:32.792818307Z" level=info msg="StopPodSandbox for \"4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652\" returns successfully" Aug 13 00:27:32.793333 containerd[2685]: time="2025-08-13T00:27:32.793308067Z" level=info msg="RemovePodSandbox for \"4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652\"" Aug 13 00:27:32.793359 containerd[2685]: time="2025-08-13T00:27:32.793338987Z" level=info msg="Forcibly stopping sandbox \"4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652\"" Aug 13 00:27:32.851701 containerd[2685]: 2025-08-13 00:27:32.823 [WARNING][9784] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--588adb827d-k8s-csi--node--driver--4wkp8-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"bb51369a-5ac7-4ac1-b819-64d9c4e312d0", ResourceVersion:"1033", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 26, 52, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"8967bcb6f", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-588adb827d", ContainerID:"28c2e0ac93ee78b239c852a9b7b6e26ab8f23ab9c93620b45125367839cf7097", Pod:"csi-node-driver-4wkp8", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.85.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calie22dd168197", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:27:32.851701 containerd[2685]: 2025-08-13 00:27:32.823 [INFO][9784] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" Aug 13 00:27:32.851701 containerd[2685]: 2025-08-13 00:27:32.823 [INFO][9784] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" iface="eth0" netns="" Aug 13 00:27:32.851701 containerd[2685]: 2025-08-13 00:27:32.823 [INFO][9784] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" Aug 13 00:27:32.851701 containerd[2685]: 2025-08-13 00:27:32.823 [INFO][9784] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" Aug 13 00:27:32.851701 containerd[2685]: 2025-08-13 00:27:32.840 [INFO][9806] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" HandleID="k8s-pod-network.4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" Workload="ci--4081.3.5--a--588adb827d-k8s-csi--node--driver--4wkp8-eth0" Aug 13 00:27:32.851701 containerd[2685]: 2025-08-13 00:27:32.840 [INFO][9806] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:32.851701 containerd[2685]: 2025-08-13 00:27:32.840 [INFO][9806] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:32.851701 containerd[2685]: 2025-08-13 00:27:32.848 [WARNING][9806] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" HandleID="k8s-pod-network.4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" Workload="ci--4081.3.5--a--588adb827d-k8s-csi--node--driver--4wkp8-eth0" Aug 13 00:27:32.851701 containerd[2685]: 2025-08-13 00:27:32.848 [INFO][9806] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" HandleID="k8s-pod-network.4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" Workload="ci--4081.3.5--a--588adb827d-k8s-csi--node--driver--4wkp8-eth0" Aug 13 00:27:32.851701 containerd[2685]: 2025-08-13 00:27:32.849 [INFO][9806] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:32.851701 containerd[2685]: 2025-08-13 00:27:32.850 [INFO][9784] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652" Aug 13 00:27:32.852072 containerd[2685]: time="2025-08-13T00:27:32.851738593Z" level=info msg="TearDown network for sandbox \"4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652\" successfully" Aug 13 00:27:32.860061 containerd[2685]: time="2025-08-13T00:27:32.860037474Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Aug 13 00:27:32.860105 containerd[2685]: time="2025-08-13T00:27:32.860094034Z" level=info msg="RemovePodSandbox \"4bb97cb69af1bbe1db2110e37317d94de1c49a318a1b1394a65fd567f9b5b652\" returns successfully" Aug 13 00:27:32.860491 containerd[2685]: time="2025-08-13T00:27:32.860472194Z" level=info msg="StopPodSandbox for \"2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8\"" Aug 13 00:27:32.919916 containerd[2685]: 2025-08-13 00:27:32.891 [WARNING][9837] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--4dnl6-eth0", GenerateName:"calico-apiserver-74488cd4c9-", Namespace:"calico-apiserver", SelfLink:"", UID:"0ced7fba-ab7f-4d92-941c-7095a5aea251", ResourceVersion:"969", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 26, 47, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"74488cd4c9", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-588adb827d", ContainerID:"812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6", Pod:"calico-apiserver-74488cd4c9-4dnl6", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.85.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali650bf866543", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:27:32.919916 containerd[2685]: 2025-08-13 00:27:32.891 [INFO][9837] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" Aug 13 00:27:32.919916 containerd[2685]: 2025-08-13 00:27:32.891 [INFO][9837] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" iface="eth0" netns="" Aug 13 00:27:32.919916 containerd[2685]: 2025-08-13 00:27:32.891 [INFO][9837] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" Aug 13 00:27:32.919916 containerd[2685]: 2025-08-13 00:27:32.891 [INFO][9837] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" Aug 13 00:27:32.919916 containerd[2685]: 2025-08-13 00:27:32.908 [INFO][9858] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" HandleID="k8s-pod-network.2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--4dnl6-eth0" Aug 13 00:27:32.919916 containerd[2685]: 2025-08-13 00:27:32.908 [INFO][9858] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:32.919916 containerd[2685]: 2025-08-13 00:27:32.908 [INFO][9858] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:32.919916 containerd[2685]: 2025-08-13 00:27:32.916 [WARNING][9858] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" HandleID="k8s-pod-network.2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--4dnl6-eth0" Aug 13 00:27:32.919916 containerd[2685]: 2025-08-13 00:27:32.916 [INFO][9858] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" HandleID="k8s-pod-network.2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--4dnl6-eth0" Aug 13 00:27:32.919916 containerd[2685]: 2025-08-13 00:27:32.917 [INFO][9858] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:32.919916 containerd[2685]: 2025-08-13 00:27:32.918 [INFO][9837] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" Aug 13 00:27:32.920291 containerd[2685]: time="2025-08-13T00:27:32.919948839Z" level=info msg="TearDown network for sandbox \"2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8\" successfully" Aug 13 00:27:32.920291 containerd[2685]: time="2025-08-13T00:27:32.919969119Z" level=info msg="StopPodSandbox for \"2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8\" returns successfully" Aug 13 00:27:32.920332 containerd[2685]: time="2025-08-13T00:27:32.920301399Z" level=info msg="RemovePodSandbox for \"2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8\"" Aug 13 00:27:32.920356 containerd[2685]: time="2025-08-13T00:27:32.920331639Z" level=info msg="Forcibly stopping sandbox \"2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8\"" Aug 13 00:27:32.977612 containerd[2685]: 2025-08-13 00:27:32.949 [WARNING][9888] cni-plugin/k8s.go 604: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--4dnl6-eth0", GenerateName:"calico-apiserver-74488cd4c9-", Namespace:"calico-apiserver", SelfLink:"", UID:"0ced7fba-ab7f-4d92-941c-7095a5aea251", ResourceVersion:"969", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 26, 47, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"74488cd4c9", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-588adb827d", ContainerID:"812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6", Pod:"calico-apiserver-74488cd4c9-4dnl6", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.85.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali650bf866543", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:27:32.977612 containerd[2685]: 2025-08-13 00:27:32.949 [INFO][9888] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" Aug 13 00:27:32.977612 containerd[2685]: 2025-08-13 00:27:32.949 [INFO][9888] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" iface="eth0" netns="" Aug 13 00:27:32.977612 containerd[2685]: 2025-08-13 00:27:32.949 [INFO][9888] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" Aug 13 00:27:32.977612 containerd[2685]: 2025-08-13 00:27:32.949 [INFO][9888] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" Aug 13 00:27:32.977612 containerd[2685]: 2025-08-13 00:27:32.966 [INFO][9908] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" HandleID="k8s-pod-network.2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--4dnl6-eth0" Aug 13 00:27:32.977612 containerd[2685]: 2025-08-13 00:27:32.966 [INFO][9908] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:32.977612 containerd[2685]: 2025-08-13 00:27:32.966 [INFO][9908] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:32.977612 containerd[2685]: 2025-08-13 00:27:32.974 [WARNING][9908] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" HandleID="k8s-pod-network.2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--4dnl6-eth0" Aug 13 00:27:32.977612 containerd[2685]: 2025-08-13 00:27:32.974 [INFO][9908] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" HandleID="k8s-pod-network.2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--4dnl6-eth0" Aug 13 00:27:32.977612 containerd[2685]: 2025-08-13 00:27:32.975 [INFO][9908] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:32.977612 containerd[2685]: 2025-08-13 00:27:32.976 [INFO][9888] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8" Aug 13 00:27:32.977999 containerd[2685]: time="2025-08-13T00:27:32.977655165Z" level=info msg="TearDown network for sandbox \"2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8\" successfully" Aug 13 00:27:32.988337 containerd[2685]: time="2025-08-13T00:27:32.988309046Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Aug 13 00:27:32.988581 containerd[2685]: time="2025-08-13T00:27:32.988371046Z" level=info msg="RemovePodSandbox \"2ad8170b7298bccb85f482b70adf1ac71b32846b2506cf075b8699512de979e8\" returns successfully" Aug 13 00:27:32.988728 containerd[2685]: time="2025-08-13T00:27:32.988705486Z" level=info msg="StopPodSandbox for \"74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef\"" Aug 13 00:27:33.058387 containerd[2685]: 2025-08-13 00:27:33.017 [WARNING][9939] cni-plugin/k8s.go 598: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-whisker--67db74799b--hr8ld-eth0" Aug 13 00:27:33.058387 containerd[2685]: 2025-08-13 00:27:33.018 [INFO][9939] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" Aug 13 00:27:33.058387 containerd[2685]: 2025-08-13 00:27:33.018 [INFO][9939] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" iface="eth0" netns="" Aug 13 00:27:33.058387 containerd[2685]: 2025-08-13 00:27:33.018 [INFO][9939] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" Aug 13 00:27:33.058387 containerd[2685]: 2025-08-13 00:27:33.018 [INFO][9939] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" Aug 13 00:27:33.058387 containerd[2685]: 2025-08-13 00:27:33.035 [INFO][9959] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" HandleID="k8s-pod-network.74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" Workload="ci--4081.3.5--a--588adb827d-k8s-whisker--67db74799b--hr8ld-eth0" Aug 13 00:27:33.058387 containerd[2685]: 2025-08-13 00:27:33.035 [INFO][9959] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:33.058387 containerd[2685]: 2025-08-13 00:27:33.036 [INFO][9959] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:33.058387 containerd[2685]: 2025-08-13 00:27:33.051 [WARNING][9959] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" HandleID="k8s-pod-network.74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" Workload="ci--4081.3.5--a--588adb827d-k8s-whisker--67db74799b--hr8ld-eth0" Aug 13 00:27:33.058387 containerd[2685]: 2025-08-13 00:27:33.051 [INFO][9959] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" HandleID="k8s-pod-network.74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" Workload="ci--4081.3.5--a--588adb827d-k8s-whisker--67db74799b--hr8ld-eth0" Aug 13 00:27:33.058387 containerd[2685]: 2025-08-13 00:27:33.055 [INFO][9959] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:33.058387 containerd[2685]: 2025-08-13 00:27:33.057 [INFO][9939] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" Aug 13 00:27:33.058387 containerd[2685]: time="2025-08-13T00:27:33.058373812Z" level=info msg="TearDown network for sandbox \"74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef\" successfully" Aug 13 00:27:33.058752 containerd[2685]: time="2025-08-13T00:27:33.058398772Z" level=info msg="StopPodSandbox for \"74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef\" returns successfully" Aug 13 00:27:33.058801 containerd[2685]: time="2025-08-13T00:27:33.058782212Z" level=info msg="RemovePodSandbox for \"74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef\"" Aug 13 00:27:33.058830 containerd[2685]: time="2025-08-13T00:27:33.058808212Z" level=info msg="Forcibly stopping sandbox \"74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef\"" Aug 13 00:27:33.117808 containerd[2685]: 2025-08-13 00:27:33.088 [WARNING][9988] cni-plugin/k8s.go 598: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-whisker--67db74799b--hr8ld-eth0" Aug 13 00:27:33.117808 containerd[2685]: 2025-08-13 00:27:33.088 [INFO][9988] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" Aug 13 00:27:33.117808 containerd[2685]: 2025-08-13 00:27:33.088 [INFO][9988] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" iface="eth0" netns="" Aug 13 00:27:33.117808 containerd[2685]: 2025-08-13 00:27:33.089 [INFO][9988] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" Aug 13 00:27:33.117808 containerd[2685]: 2025-08-13 00:27:33.089 [INFO][9988] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" Aug 13 00:27:33.117808 containerd[2685]: 2025-08-13 00:27:33.106 [INFO][10009] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" HandleID="k8s-pod-network.74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" Workload="ci--4081.3.5--a--588adb827d-k8s-whisker--67db74799b--hr8ld-eth0" Aug 13 00:27:33.117808 containerd[2685]: 2025-08-13 00:27:33.106 [INFO][10009] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:33.117808 containerd[2685]: 2025-08-13 00:27:33.106 [INFO][10009] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:33.117808 containerd[2685]: 2025-08-13 00:27:33.114 [WARNING][10009] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" HandleID="k8s-pod-network.74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" Workload="ci--4081.3.5--a--588adb827d-k8s-whisker--67db74799b--hr8ld-eth0" Aug 13 00:27:33.117808 containerd[2685]: 2025-08-13 00:27:33.114 [INFO][10009] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" HandleID="k8s-pod-network.74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" Workload="ci--4081.3.5--a--588adb827d-k8s-whisker--67db74799b--hr8ld-eth0" Aug 13 00:27:33.117808 containerd[2685]: 2025-08-13 00:27:33.115 [INFO][10009] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:33.117808 containerd[2685]: 2025-08-13 00:27:33.116 [INFO][9988] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef" Aug 13 00:27:33.118048 containerd[2685]: time="2025-08-13T00:27:33.117846698Z" level=info msg="TearDown network for sandbox \"74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef\" successfully" Aug 13 00:27:33.119559 containerd[2685]: time="2025-08-13T00:27:33.119535458Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Aug 13 00:27:33.119600 containerd[2685]: time="2025-08-13T00:27:33.119587018Z" level=info msg="RemovePodSandbox \"74b68a34928d7629fb1179f3d8ef40ec03937b7f657ce47572980b626e9151ef\" returns successfully" Aug 13 00:27:36.549630 kubelet[4148]: I0813 00:27:36.549553 4148 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Aug 13 00:27:43.197615 kubelet[4148]: I0813 00:27:43.197557 4148 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Aug 13 00:27:44.950521 kubelet[4148]: I0813 00:27:44.950469 4148 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Aug 13 00:27:44.993219 systemd[1]: Created slice kubepods-besteffort-podceff020b_e8dd_451b_aa5c_3b8685806de0.slice - libcontainer container kubepods-besteffort-podceff020b_e8dd_451b_aa5c_3b8685806de0.slice. Aug 13 00:27:45.034629 kubelet[4148]: I0813 00:27:45.034599 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/ceff020b-e8dd-451b-aa5c-3b8685806de0-calico-apiserver-certs\") pod \"calico-apiserver-54c5fc9686-wt7t4\" (UID: \"ceff020b-e8dd-451b-aa5c-3b8685806de0\") " pod="calico-apiserver/calico-apiserver-54c5fc9686-wt7t4" Aug 13 00:27:45.034691 kubelet[4148]: I0813 00:27:45.034635 4148 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9t9q4\" (UniqueName: \"kubernetes.io/projected/ceff020b-e8dd-451b-aa5c-3b8685806de0-kube-api-access-9t9q4\") pod \"calico-apiserver-54c5fc9686-wt7t4\" (UID: \"ceff020b-e8dd-451b-aa5c-3b8685806de0\") " pod="calico-apiserver/calico-apiserver-54c5fc9686-wt7t4" Aug 13 00:27:45.141965 containerd[2685]: time="2025-08-13T00:27:45.141932270Z" level=info msg="StopContainer for \"7d1057fb03c5afd1c6c40fbde2f435dd8914ced368b563fcf13dc6e7b062cd11\" with timeout 30 (s)" Aug 13 00:27:45.142288 containerd[2685]: time="2025-08-13T00:27:45.142264831Z" level=info msg="Stop container \"7d1057fb03c5afd1c6c40fbde2f435dd8914ced368b563fcf13dc6e7b062cd11\" with signal terminated" Aug 13 00:27:45.149598 systemd[1]: cri-containerd-7d1057fb03c5afd1c6c40fbde2f435dd8914ced368b563fcf13dc6e7b062cd11.scope: Deactivated successfully. Aug 13 00:27:45.163601 containerd[2685]: time="2025-08-13T00:27:45.163548351Z" level=info msg="shim disconnected" id=7d1057fb03c5afd1c6c40fbde2f435dd8914ced368b563fcf13dc6e7b062cd11 namespace=k8s.io Aug 13 00:27:45.163645 containerd[2685]: time="2025-08-13T00:27:45.163602391Z" level=warning msg="cleaning up after shim disconnected" id=7d1057fb03c5afd1c6c40fbde2f435dd8914ced368b563fcf13dc6e7b062cd11 namespace=k8s.io Aug 13 00:27:45.163645 containerd[2685]: time="2025-08-13T00:27:45.163612751Z" level=info msg="cleaning up dead shim" namespace=k8s.io Aug 13 00:27:45.166117 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-7d1057fb03c5afd1c6c40fbde2f435dd8914ced368b563fcf13dc6e7b062cd11-rootfs.mount: Deactivated successfully. Aug 13 00:27:45.295515 containerd[2685]: time="2025-08-13T00:27:45.295470477Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-54c5fc9686-wt7t4,Uid:ceff020b-e8dd-451b-aa5c-3b8685806de0,Namespace:calico-apiserver,Attempt:0,}" Aug 13 00:27:45.378404 containerd[2685]: time="2025-08-13T00:27:45.378363760Z" level=info msg="StopContainer for \"7d1057fb03c5afd1c6c40fbde2f435dd8914ced368b563fcf13dc6e7b062cd11\" returns successfully" Aug 13 00:27:45.378850 containerd[2685]: time="2025-08-13T00:27:45.378822560Z" level=info msg="StopPodSandbox for \"812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6\"" Aug 13 00:27:45.378880 containerd[2685]: time="2025-08-13T00:27:45.378860040Z" level=info msg="Container to stop \"7d1057fb03c5afd1c6c40fbde2f435dd8914ced368b563fcf13dc6e7b062cd11\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Aug 13 00:27:45.383997 systemd[1]: cri-containerd-812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6.scope: Deactivated successfully. Aug 13 00:27:45.389170 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6-shm.mount: Deactivated successfully. Aug 13 00:27:45.396847 containerd[2685]: time="2025-08-13T00:27:45.396784001Z" level=info msg="shim disconnected" id=812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6 namespace=k8s.io Aug 13 00:27:45.396847 containerd[2685]: time="2025-08-13T00:27:45.396833321Z" level=warning msg="cleaning up after shim disconnected" id=812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6 namespace=k8s.io Aug 13 00:27:45.396847 containerd[2685]: time="2025-08-13T00:27:45.396841481Z" level=info msg="cleaning up dead shim" namespace=k8s.io Aug 13 00:27:45.398765 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6-rootfs.mount: Deactivated successfully. Aug 13 00:27:45.452312 systemd-networkd[2588]: cali9ea9a403241: Link UP Aug 13 00:27:45.452616 systemd-networkd[2588]: cali9ea9a403241: Gained carrier Aug 13 00:27:45.459919 systemd-networkd[2588]: cali650bf866543: Link DOWN Aug 13 00:27:45.459925 systemd-networkd[2588]: cali650bf866543: Lost carrier Aug 13 00:27:45.460199 containerd[2685]: 2025-08-13 00:27:45.402 [INFO][10122] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--wt7t4-eth0 calico-apiserver-54c5fc9686- calico-apiserver ceff020b-e8dd-451b-aa5c-3b8685806de0 1163 0 2025-08-13 00:27:44 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:54c5fc9686 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4081.3.5-a-588adb827d calico-apiserver-54c5fc9686-wt7t4 eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali9ea9a403241 [] [] }} ContainerID="4ea8a9c67b1d68443c17e79cbab1eb226d8b8cab21c26bef78491d69e30f6df9" Namespace="calico-apiserver" Pod="calico-apiserver-54c5fc9686-wt7t4" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--wt7t4-" Aug 13 00:27:45.460199 containerd[2685]: 2025-08-13 00:27:45.402 [INFO][10122] cni-plugin/k8s.go 74: Extracted identifiers for CmdAddK8s ContainerID="4ea8a9c67b1d68443c17e79cbab1eb226d8b8cab21c26bef78491d69e30f6df9" Namespace="calico-apiserver" Pod="calico-apiserver-54c5fc9686-wt7t4" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--wt7t4-eth0" Aug 13 00:27:45.460199 containerd[2685]: 2025-08-13 00:27:45.422 [INFO][10180] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="4ea8a9c67b1d68443c17e79cbab1eb226d8b8cab21c26bef78491d69e30f6df9" HandleID="k8s-pod-network.4ea8a9c67b1d68443c17e79cbab1eb226d8b8cab21c26bef78491d69e30f6df9" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--wt7t4-eth0" Aug 13 00:27:45.460199 containerd[2685]: 2025-08-13 00:27:45.422 [INFO][10180] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="4ea8a9c67b1d68443c17e79cbab1eb226d8b8cab21c26bef78491d69e30f6df9" HandleID="k8s-pod-network.4ea8a9c67b1d68443c17e79cbab1eb226d8b8cab21c26bef78491d69e30f6df9" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--wt7t4-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000363e00), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4081.3.5-a-588adb827d", "pod":"calico-apiserver-54c5fc9686-wt7t4", "timestamp":"2025-08-13 00:27:45.422245802 +0000 UTC"}, Hostname:"ci-4081.3.5-a-588adb827d", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Aug 13 00:27:45.460199 containerd[2685]: 2025-08-13 00:27:45.422 [INFO][10180] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:45.460199 containerd[2685]: 2025-08-13 00:27:45.422 [INFO][10180] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:45.460199 containerd[2685]: 2025-08-13 00:27:45.422 [INFO][10180] ipam/ipam.go 110: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4081.3.5-a-588adb827d' Aug 13 00:27:45.460199 containerd[2685]: 2025-08-13 00:27:45.430 [INFO][10180] ipam/ipam.go 691: Looking up existing affinities for host handle="k8s-pod-network.4ea8a9c67b1d68443c17e79cbab1eb226d8b8cab21c26bef78491d69e30f6df9" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:45.460199 containerd[2685]: 2025-08-13 00:27:45.433 [INFO][10180] ipam/ipam.go 394: Looking up existing affinities for host host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:45.460199 containerd[2685]: 2025-08-13 00:27:45.436 [INFO][10180] ipam/ipam.go 511: Trying affinity for 192.168.85.64/26 host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:45.460199 containerd[2685]: 2025-08-13 00:27:45.437 [INFO][10180] ipam/ipam.go 158: Attempting to load block cidr=192.168.85.64/26 host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:45.460199 containerd[2685]: 2025-08-13 00:27:45.438 [INFO][10180] ipam/ipam.go 235: Affinity is confirmed and block has been loaded cidr=192.168.85.64/26 host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:45.460199 containerd[2685]: 2025-08-13 00:27:45.439 [INFO][10180] ipam/ipam.go 1220: Attempting to assign 1 addresses from block block=192.168.85.64/26 handle="k8s-pod-network.4ea8a9c67b1d68443c17e79cbab1eb226d8b8cab21c26bef78491d69e30f6df9" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:45.460199 containerd[2685]: 2025-08-13 00:27:45.440 [INFO][10180] ipam/ipam.go 1764: Creating new handle: k8s-pod-network.4ea8a9c67b1d68443c17e79cbab1eb226d8b8cab21c26bef78491d69e30f6df9 Aug 13 00:27:45.460199 containerd[2685]: 2025-08-13 00:27:45.442 [INFO][10180] ipam/ipam.go 1243: Writing block in order to claim IPs block=192.168.85.64/26 handle="k8s-pod-network.4ea8a9c67b1d68443c17e79cbab1eb226d8b8cab21c26bef78491d69e30f6df9" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:45.460199 containerd[2685]: 2025-08-13 00:27:45.447 [INFO][10180] ipam/ipam.go 1256: Successfully claimed IPs: [192.168.85.74/26] block=192.168.85.64/26 handle="k8s-pod-network.4ea8a9c67b1d68443c17e79cbab1eb226d8b8cab21c26bef78491d69e30f6df9" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:45.460199 containerd[2685]: 2025-08-13 00:27:45.447 [INFO][10180] ipam/ipam.go 878: Auto-assigned 1 out of 1 IPv4s: [192.168.85.74/26] handle="k8s-pod-network.4ea8a9c67b1d68443c17e79cbab1eb226d8b8cab21c26bef78491d69e30f6df9" host="ci-4081.3.5-a-588adb827d" Aug 13 00:27:45.460199 containerd[2685]: 2025-08-13 00:27:45.447 [INFO][10180] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:45.460199 containerd[2685]: 2025-08-13 00:27:45.447 [INFO][10180] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.85.74/26] IPv6=[] ContainerID="4ea8a9c67b1d68443c17e79cbab1eb226d8b8cab21c26bef78491d69e30f6df9" HandleID="k8s-pod-network.4ea8a9c67b1d68443c17e79cbab1eb226d8b8cab21c26bef78491d69e30f6df9" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--wt7t4-eth0" Aug 13 00:27:45.460698 containerd[2685]: 2025-08-13 00:27:45.448 [INFO][10122] cni-plugin/k8s.go 418: Populated endpoint ContainerID="4ea8a9c67b1d68443c17e79cbab1eb226d8b8cab21c26bef78491d69e30f6df9" Namespace="calico-apiserver" Pod="calico-apiserver-54c5fc9686-wt7t4" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--wt7t4-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--wt7t4-eth0", GenerateName:"calico-apiserver-54c5fc9686-", Namespace:"calico-apiserver", SelfLink:"", UID:"ceff020b-e8dd-451b-aa5c-3b8685806de0", ResourceVersion:"1163", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 27, 44, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"54c5fc9686", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-588adb827d", ContainerID:"", Pod:"calico-apiserver-54c5fc9686-wt7t4", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.85.74/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali9ea9a403241", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:27:45.460698 containerd[2685]: 2025-08-13 00:27:45.448 [INFO][10122] cni-plugin/k8s.go 419: Calico CNI using IPs: [192.168.85.74/32] ContainerID="4ea8a9c67b1d68443c17e79cbab1eb226d8b8cab21c26bef78491d69e30f6df9" Namespace="calico-apiserver" Pod="calico-apiserver-54c5fc9686-wt7t4" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--wt7t4-eth0" Aug 13 00:27:45.460698 containerd[2685]: 2025-08-13 00:27:45.448 [INFO][10122] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali9ea9a403241 ContainerID="4ea8a9c67b1d68443c17e79cbab1eb226d8b8cab21c26bef78491d69e30f6df9" Namespace="calico-apiserver" Pod="calico-apiserver-54c5fc9686-wt7t4" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--wt7t4-eth0" Aug 13 00:27:45.460698 containerd[2685]: 2025-08-13 00:27:45.453 [INFO][10122] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="4ea8a9c67b1d68443c17e79cbab1eb226d8b8cab21c26bef78491d69e30f6df9" Namespace="calico-apiserver" Pod="calico-apiserver-54c5fc9686-wt7t4" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--wt7t4-eth0" Aug 13 00:27:45.460698 containerd[2685]: 2025-08-13 00:27:45.453 [INFO][10122] cni-plugin/k8s.go 446: Added Mac, interface name, and active container ID to endpoint ContainerID="4ea8a9c67b1d68443c17e79cbab1eb226d8b8cab21c26bef78491d69e30f6df9" Namespace="calico-apiserver" Pod="calico-apiserver-54c5fc9686-wt7t4" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--wt7t4-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--wt7t4-eth0", GenerateName:"calico-apiserver-54c5fc9686-", Namespace:"calico-apiserver", SelfLink:"", UID:"ceff020b-e8dd-451b-aa5c-3b8685806de0", ResourceVersion:"1163", Generation:0, CreationTimestamp:time.Date(2025, time.August, 13, 0, 27, 44, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"54c5fc9686", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4081.3.5-a-588adb827d", ContainerID:"4ea8a9c67b1d68443c17e79cbab1eb226d8b8cab21c26bef78491d69e30f6df9", Pod:"calico-apiserver-54c5fc9686-wt7t4", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.85.74/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali9ea9a403241", MAC:"a6:40:f7:25:ec:2e", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil), QoSControls:(*v3.QoSControls)(nil)}} Aug 13 00:27:45.460698 containerd[2685]: 2025-08-13 00:27:45.459 [INFO][10122] cni-plugin/k8s.go 532: Wrote updated endpoint to datastore ContainerID="4ea8a9c67b1d68443c17e79cbab1eb226d8b8cab21c26bef78491d69e30f6df9" Namespace="calico-apiserver" Pod="calico-apiserver-54c5fc9686-wt7t4" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--54c5fc9686--wt7t4-eth0" Aug 13 00:27:45.473358 containerd[2685]: time="2025-08-13T00:27:45.473291124Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 00:27:45.473358 containerd[2685]: time="2025-08-13T00:27:45.473349644Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 00:27:45.473464 containerd[2685]: time="2025-08-13T00:27:45.473361764Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:27:45.473464 containerd[2685]: time="2025-08-13T00:27:45.473441484Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 00:27:45.511394 systemd[1]: Started cri-containerd-4ea8a9c67b1d68443c17e79cbab1eb226d8b8cab21c26bef78491d69e30f6df9.scope - libcontainer container 4ea8a9c67b1d68443c17e79cbab1eb226d8b8cab21c26bef78491d69e30f6df9. Aug 13 00:27:45.535389 containerd[2685]: time="2025-08-13T00:27:45.535354847Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-54c5fc9686-wt7t4,Uid:ceff020b-e8dd-451b-aa5c-3b8685806de0,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"4ea8a9c67b1d68443c17e79cbab1eb226d8b8cab21c26bef78491d69e30f6df9\"" Aug 13 00:27:45.550899 containerd[2685]: time="2025-08-13T00:27:45.550819567Z" level=info msg="CreateContainer within sandbox \"4ea8a9c67b1d68443c17e79cbab1eb226d8b8cab21c26bef78491d69e30f6df9\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Aug 13 00:27:45.551214 containerd[2685]: 2025-08-13 00:27:45.459 [INFO][10212] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" Aug 13 00:27:45.551214 containerd[2685]: 2025-08-13 00:27:45.459 [INFO][10212] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" iface="eth0" netns="/var/run/netns/cni-3dc1b9e8-5302-265f-8754-f46a747fb1bc" Aug 13 00:27:45.551214 containerd[2685]: 2025-08-13 00:27:45.459 [INFO][10212] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" iface="eth0" netns="/var/run/netns/cni-3dc1b9e8-5302-265f-8754-f46a747fb1bc" Aug 13 00:27:45.551214 containerd[2685]: 2025-08-13 00:27:45.506 [INFO][10212] cni-plugin/dataplane_linux.go 604: Deleted device in netns. ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" after=47.054282ms iface="eth0" netns="/var/run/netns/cni-3dc1b9e8-5302-265f-8754-f46a747fb1bc" Aug 13 00:27:45.551214 containerd[2685]: 2025-08-13 00:27:45.506 [INFO][10212] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" Aug 13 00:27:45.551214 containerd[2685]: 2025-08-13 00:27:45.506 [INFO][10212] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" Aug 13 00:27:45.551214 containerd[2685]: 2025-08-13 00:27:45.523 [INFO][10275] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" HandleID="k8s-pod-network.812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--4dnl6-eth0" Aug 13 00:27:45.551214 containerd[2685]: 2025-08-13 00:27:45.523 [INFO][10275] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:45.551214 containerd[2685]: 2025-08-13 00:27:45.523 [INFO][10275] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:45.551214 containerd[2685]: 2025-08-13 00:27:45.547 [INFO][10275] ipam/ipam_plugin.go 431: Released address using handleID ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" HandleID="k8s-pod-network.812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--4dnl6-eth0" Aug 13 00:27:45.551214 containerd[2685]: 2025-08-13 00:27:45.547 [INFO][10275] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" HandleID="k8s-pod-network.812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--4dnl6-eth0" Aug 13 00:27:45.551214 containerd[2685]: 2025-08-13 00:27:45.548 [INFO][10275] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:45.551214 containerd[2685]: 2025-08-13 00:27:45.549 [INFO][10212] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" Aug 13 00:27:45.551554 containerd[2685]: time="2025-08-13T00:27:45.551472528Z" level=info msg="TearDown network for sandbox \"812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6\" successfully" Aug 13 00:27:45.551554 containerd[2685]: time="2025-08-13T00:27:45.551494528Z" level=info msg="StopPodSandbox for \"812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6\" returns successfully" Aug 13 00:27:45.555222 containerd[2685]: time="2025-08-13T00:27:45.555194328Z" level=info msg="CreateContainer within sandbox \"4ea8a9c67b1d68443c17e79cbab1eb226d8b8cab21c26bef78491d69e30f6df9\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"c26de5df27b78f5b7873e9e3931d82f2467050c5c375fe9fa1ad0648a63fcd81\"" Aug 13 00:27:45.556011 containerd[2685]: time="2025-08-13T00:27:45.555987088Z" level=info msg="StartContainer for \"c26de5df27b78f5b7873e9e3931d82f2467050c5c375fe9fa1ad0648a63fcd81\"" Aug 13 00:27:45.587355 systemd[1]: Started cri-containerd-c26de5df27b78f5b7873e9e3931d82f2467050c5c375fe9fa1ad0648a63fcd81.scope - libcontainer container c26de5df27b78f5b7873e9e3931d82f2467050c5c375fe9fa1ad0648a63fcd81. Aug 13 00:27:45.611861 containerd[2685]: time="2025-08-13T00:27:45.611832210Z" level=info msg="StartContainer for \"c26de5df27b78f5b7873e9e3931d82f2467050c5c375fe9fa1ad0648a63fcd81\" returns successfully" Aug 13 00:27:45.640154 kubelet[4148]: I0813 00:27:45.640120 4148 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvxdb\" (UniqueName: \"kubernetes.io/projected/0ced7fba-ab7f-4d92-941c-7095a5aea251-kube-api-access-wvxdb\") pod \"0ced7fba-ab7f-4d92-941c-7095a5aea251\" (UID: \"0ced7fba-ab7f-4d92-941c-7095a5aea251\") " Aug 13 00:27:45.640271 kubelet[4148]: I0813 00:27:45.640186 4148 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/0ced7fba-ab7f-4d92-941c-7095a5aea251-calico-apiserver-certs\") pod \"0ced7fba-ab7f-4d92-941c-7095a5aea251\" (UID: \"0ced7fba-ab7f-4d92-941c-7095a5aea251\") " Aug 13 00:27:45.642349 kubelet[4148]: I0813 00:27:45.642320 4148 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ced7fba-ab7f-4d92-941c-7095a5aea251-kube-api-access-wvxdb" (OuterVolumeSpecName: "kube-api-access-wvxdb") pod "0ced7fba-ab7f-4d92-941c-7095a5aea251" (UID: "0ced7fba-ab7f-4d92-941c-7095a5aea251"). InnerVolumeSpecName "kube-api-access-wvxdb". PluginName "kubernetes.io/projected", VolumeGIDValue "" Aug 13 00:27:45.642425 kubelet[4148]: I0813 00:27:45.642400 4148 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ced7fba-ab7f-4d92-941c-7095a5aea251-calico-apiserver-certs" (OuterVolumeSpecName: "calico-apiserver-certs") pod "0ced7fba-ab7f-4d92-941c-7095a5aea251" (UID: "0ced7fba-ab7f-4d92-941c-7095a5aea251"). InnerVolumeSpecName "calico-apiserver-certs". PluginName "kubernetes.io/secret", VolumeGIDValue "" Aug 13 00:27:45.740768 kubelet[4148]: I0813 00:27:45.740743 4148 reconciler_common.go:299] "Volume detached for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/0ced7fba-ab7f-4d92-941c-7095a5aea251-calico-apiserver-certs\") on node \"ci-4081.3.5-a-588adb827d\" DevicePath \"\"" Aug 13 00:27:45.740768 kubelet[4148]: I0813 00:27:45.740766 4148 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-wvxdb\" (UniqueName: \"kubernetes.io/projected/0ced7fba-ab7f-4d92-941c-7095a5aea251-kube-api-access-wvxdb\") on node \"ci-4081.3.5-a-588adb827d\" DevicePath \"\"" Aug 13 00:27:45.996254 systemd[1]: Removed slice kubepods-besteffort-pod0ced7fba_ab7f_4d92_941c_7095a5aea251.slice - libcontainer container kubepods-besteffort-pod0ced7fba_ab7f_4d92_941c_7095a5aea251.slice. Aug 13 00:27:46.144644 kubelet[4148]: I0813 00:27:46.144609 4148 scope.go:117] "RemoveContainer" containerID="7d1057fb03c5afd1c6c40fbde2f435dd8914ced368b563fcf13dc6e7b062cd11" Aug 13 00:27:46.145782 containerd[2685]: time="2025-08-13T00:27:46.145757072Z" level=info msg="RemoveContainer for \"7d1057fb03c5afd1c6c40fbde2f435dd8914ced368b563fcf13dc6e7b062cd11\"" Aug 13 00:27:46.149019 containerd[2685]: time="2025-08-13T00:27:46.147869432Z" level=info msg="RemoveContainer for \"7d1057fb03c5afd1c6c40fbde2f435dd8914ced368b563fcf13dc6e7b062cd11\" returns successfully" Aug 13 00:27:46.149019 containerd[2685]: time="2025-08-13T00:27:46.148719632Z" level=error msg="ContainerStatus for \"7d1057fb03c5afd1c6c40fbde2f435dd8914ced368b563fcf13dc6e7b062cd11\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"7d1057fb03c5afd1c6c40fbde2f435dd8914ced368b563fcf13dc6e7b062cd11\": not found" Aug 13 00:27:46.149119 kubelet[4148]: I0813 00:27:46.148427 4148 scope.go:117] "RemoveContainer" containerID="7d1057fb03c5afd1c6c40fbde2f435dd8914ced368b563fcf13dc6e7b062cd11" Aug 13 00:27:46.149119 kubelet[4148]: E0813 00:27:46.148843 4148 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"7d1057fb03c5afd1c6c40fbde2f435dd8914ced368b563fcf13dc6e7b062cd11\": not found" containerID="7d1057fb03c5afd1c6c40fbde2f435dd8914ced368b563fcf13dc6e7b062cd11" Aug 13 00:27:46.149119 kubelet[4148]: I0813 00:27:46.148865 4148 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"7d1057fb03c5afd1c6c40fbde2f435dd8914ced368b563fcf13dc6e7b062cd11"} err="failed to get container status \"7d1057fb03c5afd1c6c40fbde2f435dd8914ced368b563fcf13dc6e7b062cd11\": rpc error: code = NotFound desc = an error occurred when try to find container \"7d1057fb03c5afd1c6c40fbde2f435dd8914ced368b563fcf13dc6e7b062cd11\": not found" Aug 13 00:27:46.166976 kubelet[4148]: I0813 00:27:46.166935 4148 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-54c5fc9686-wt7t4" podStartSLOduration=2.166923193 podStartE2EDuration="2.166923193s" podCreationTimestamp="2025-08-13 00:27:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 00:27:46.166293553 +0000 UTC m=+74.251250692" watchObservedRunningTime="2025-08-13 00:27:46.166923193 +0000 UTC m=+74.251880332" Aug 13 00:27:46.389755 systemd[1]: run-netns-cni\x2d3dc1b9e8\x2d5302\x2d265f\x2d8754\x2df46a747fb1bc.mount: Deactivated successfully. Aug 13 00:27:46.389839 systemd[1]: var-lib-kubelet-pods-0ced7fba\x2dab7f\x2d4d92\x2d941c\x2d7095a5aea251-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dwvxdb.mount: Deactivated successfully. Aug 13 00:27:46.389897 systemd[1]: var-lib-kubelet-pods-0ced7fba\x2dab7f\x2d4d92\x2d941c\x2d7095a5aea251-volumes-kubernetes.io\x7esecret-calico\x2dapiserver\x2dcerts.mount: Deactivated successfully. Aug 13 00:27:46.582343 systemd-networkd[2588]: cali9ea9a403241: Gained IPv6LL Aug 13 00:27:46.700703 containerd[2685]: time="2025-08-13T00:27:46.700600293Z" level=info msg="StopContainer for \"05d1a463839387260db0c040a26786524b1847fa73ef06c63092a8f10b27a4bd\" with timeout 30 (s)" Aug 13 00:27:46.701012 containerd[2685]: time="2025-08-13T00:27:46.700969653Z" level=info msg="Stop container \"05d1a463839387260db0c040a26786524b1847fa73ef06c63092a8f10b27a4bd\" with signal terminated" Aug 13 00:27:46.710611 systemd[1]: cri-containerd-05d1a463839387260db0c040a26786524b1847fa73ef06c63092a8f10b27a4bd.scope: Deactivated successfully. Aug 13 00:27:46.710917 systemd[1]: cri-containerd-05d1a463839387260db0c040a26786524b1847fa73ef06c63092a8f10b27a4bd.scope: Consumed 1.648s CPU time. Aug 13 00:27:46.724705 containerd[2685]: time="2025-08-13T00:27:46.724654614Z" level=info msg="shim disconnected" id=05d1a463839387260db0c040a26786524b1847fa73ef06c63092a8f10b27a4bd namespace=k8s.io Aug 13 00:27:46.724705 containerd[2685]: time="2025-08-13T00:27:46.724701534Z" level=warning msg="cleaning up after shim disconnected" id=05d1a463839387260db0c040a26786524b1847fa73ef06c63092a8f10b27a4bd namespace=k8s.io Aug 13 00:27:46.724705 containerd[2685]: time="2025-08-13T00:27:46.724709414Z" level=info msg="cleaning up dead shim" namespace=k8s.io Aug 13 00:27:46.726677 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-05d1a463839387260db0c040a26786524b1847fa73ef06c63092a8f10b27a4bd-rootfs.mount: Deactivated successfully. Aug 13 00:27:46.737461 containerd[2685]: time="2025-08-13T00:27:46.737428335Z" level=info msg="StopContainer for \"05d1a463839387260db0c040a26786524b1847fa73ef06c63092a8f10b27a4bd\" returns successfully" Aug 13 00:27:46.737859 containerd[2685]: time="2025-08-13T00:27:46.737835775Z" level=info msg="StopPodSandbox for \"6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881\"" Aug 13 00:27:46.737892 containerd[2685]: time="2025-08-13T00:27:46.737869815Z" level=info msg="Container to stop \"05d1a463839387260db0c040a26786524b1847fa73ef06c63092a8f10b27a4bd\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Aug 13 00:27:46.740240 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881-shm.mount: Deactivated successfully. Aug 13 00:27:46.743243 systemd[1]: cri-containerd-6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881.scope: Deactivated successfully. Aug 13 00:27:46.756432 containerd[2685]: time="2025-08-13T00:27:46.756386176Z" level=info msg="shim disconnected" id=6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881 namespace=k8s.io Aug 13 00:27:46.756432 containerd[2685]: time="2025-08-13T00:27:46.756429336Z" level=warning msg="cleaning up after shim disconnected" id=6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881 namespace=k8s.io Aug 13 00:27:46.756505 containerd[2685]: time="2025-08-13T00:27:46.756437816Z" level=info msg="cleaning up dead shim" namespace=k8s.io Aug 13 00:27:46.758395 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881-rootfs.mount: Deactivated successfully. Aug 13 00:27:46.855382 systemd-networkd[2588]: cali3c1d63c3857: Link DOWN Aug 13 00:27:46.855390 systemd-networkd[2588]: cali3c1d63c3857: Lost carrier Aug 13 00:27:46.937467 containerd[2685]: 2025-08-13 00:27:46.854 [INFO][10467] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" Aug 13 00:27:46.937467 containerd[2685]: 2025-08-13 00:27:46.854 [INFO][10467] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" iface="eth0" netns="/var/run/netns/cni-5eaa146f-c326-f151-ecf7-0005f12d46d9" Aug 13 00:27:46.937467 containerd[2685]: 2025-08-13 00:27:46.854 [INFO][10467] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" iface="eth0" netns="/var/run/netns/cni-5eaa146f-c326-f151-ecf7-0005f12d46d9" Aug 13 00:27:46.937467 containerd[2685]: 2025-08-13 00:27:46.890 [INFO][10467] cni-plugin/dataplane_linux.go 604: Deleted device in netns. ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" after=35.602202ms iface="eth0" netns="/var/run/netns/cni-5eaa146f-c326-f151-ecf7-0005f12d46d9" Aug 13 00:27:46.937467 containerd[2685]: 2025-08-13 00:27:46.890 [INFO][10467] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" Aug 13 00:27:46.937467 containerd[2685]: 2025-08-13 00:27:46.890 [INFO][10467] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" Aug 13 00:27:46.937467 containerd[2685]: 2025-08-13 00:27:46.907 [INFO][10494] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" HandleID="k8s-pod-network.6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--h5l86-eth0" Aug 13 00:27:46.937467 containerd[2685]: 2025-08-13 00:27:46.907 [INFO][10494] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:27:46.937467 containerd[2685]: 2025-08-13 00:27:46.907 [INFO][10494] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:27:46.937467 containerd[2685]: 2025-08-13 00:27:46.933 [INFO][10494] ipam/ipam_plugin.go 431: Released address using handleID ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" HandleID="k8s-pod-network.6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--h5l86-eth0" Aug 13 00:27:46.937467 containerd[2685]: 2025-08-13 00:27:46.933 [INFO][10494] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" HandleID="k8s-pod-network.6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--h5l86-eth0" Aug 13 00:27:46.937467 containerd[2685]: 2025-08-13 00:27:46.934 [INFO][10494] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:27:46.937467 containerd[2685]: 2025-08-13 00:27:46.936 [INFO][10467] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" Aug 13 00:27:46.937888 containerd[2685]: time="2025-08-13T00:27:46.937758263Z" level=info msg="TearDown network for sandbox \"6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881\" successfully" Aug 13 00:27:46.937888 containerd[2685]: time="2025-08-13T00:27:46.937784423Z" level=info msg="StopPodSandbox for \"6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881\" returns successfully" Aug 13 00:27:46.940083 systemd[1]: run-netns-cni\x2d5eaa146f\x2dc326\x2df151\x2decf7\x2d0005f12d46d9.mount: Deactivated successfully. Aug 13 00:27:47.049206 kubelet[4148]: I0813 00:27:47.049183 4148 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/517c2f94-3455-4677-a3ec-3a8771258060-calico-apiserver-certs\") pod \"517c2f94-3455-4677-a3ec-3a8771258060\" (UID: \"517c2f94-3455-4677-a3ec-3a8771258060\") " Aug 13 00:27:47.049345 kubelet[4148]: I0813 00:27:47.049223 4148 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbmgz\" (UniqueName: \"kubernetes.io/projected/517c2f94-3455-4677-a3ec-3a8771258060-kube-api-access-pbmgz\") pod \"517c2f94-3455-4677-a3ec-3a8771258060\" (UID: \"517c2f94-3455-4677-a3ec-3a8771258060\") " Aug 13 00:27:47.051456 kubelet[4148]: I0813 00:27:47.051434 4148 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/517c2f94-3455-4677-a3ec-3a8771258060-kube-api-access-pbmgz" (OuterVolumeSpecName: "kube-api-access-pbmgz") pod "517c2f94-3455-4677-a3ec-3a8771258060" (UID: "517c2f94-3455-4677-a3ec-3a8771258060"). InnerVolumeSpecName "kube-api-access-pbmgz". PluginName "kubernetes.io/projected", VolumeGIDValue "" Aug 13 00:27:47.051520 kubelet[4148]: I0813 00:27:47.051496 4148 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/517c2f94-3455-4677-a3ec-3a8771258060-calico-apiserver-certs" (OuterVolumeSpecName: "calico-apiserver-certs") pod "517c2f94-3455-4677-a3ec-3a8771258060" (UID: "517c2f94-3455-4677-a3ec-3a8771258060"). InnerVolumeSpecName "calico-apiserver-certs". PluginName "kubernetes.io/secret", VolumeGIDValue "" Aug 13 00:27:47.149425 kubelet[4148]: I0813 00:27:47.149400 4148 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-pbmgz\" (UniqueName: \"kubernetes.io/projected/517c2f94-3455-4677-a3ec-3a8771258060-kube-api-access-pbmgz\") on node \"ci-4081.3.5-a-588adb827d\" DevicePath \"\"" Aug 13 00:27:47.149425 kubelet[4148]: I0813 00:27:47.149419 4148 reconciler_common.go:299] "Volume detached for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/517c2f94-3455-4677-a3ec-3a8771258060-calico-apiserver-certs\") on node \"ci-4081.3.5-a-588adb827d\" DevicePath \"\"" Aug 13 00:27:47.149827 kubelet[4148]: I0813 00:27:47.149436 4148 scope.go:117] "RemoveContainer" containerID="05d1a463839387260db0c040a26786524b1847fa73ef06c63092a8f10b27a4bd" Aug 13 00:27:47.150286 containerd[2685]: time="2025-08-13T00:27:47.150262351Z" level=info msg="RemoveContainer for \"05d1a463839387260db0c040a26786524b1847fa73ef06c63092a8f10b27a4bd\"" Aug 13 00:27:47.152028 containerd[2685]: time="2025-08-13T00:27:47.152008031Z" level=info msg="RemoveContainer for \"05d1a463839387260db0c040a26786524b1847fa73ef06c63092a8f10b27a4bd\" returns successfully" Aug 13 00:27:47.152132 kubelet[4148]: I0813 00:27:47.152116 4148 scope.go:117] "RemoveContainer" containerID="05d1a463839387260db0c040a26786524b1847fa73ef06c63092a8f10b27a4bd" Aug 13 00:27:47.152316 containerd[2685]: time="2025-08-13T00:27:47.152283711Z" level=error msg="ContainerStatus for \"05d1a463839387260db0c040a26786524b1847fa73ef06c63092a8f10b27a4bd\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"05d1a463839387260db0c040a26786524b1847fa73ef06c63092a8f10b27a4bd\": not found" Aug 13 00:27:47.152409 kubelet[4148]: E0813 00:27:47.152392 4148 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"05d1a463839387260db0c040a26786524b1847fa73ef06c63092a8f10b27a4bd\": not found" containerID="05d1a463839387260db0c040a26786524b1847fa73ef06c63092a8f10b27a4bd" Aug 13 00:27:47.152433 kubelet[4148]: I0813 00:27:47.152419 4148 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"05d1a463839387260db0c040a26786524b1847fa73ef06c63092a8f10b27a4bd"} err="failed to get container status \"05d1a463839387260db0c040a26786524b1847fa73ef06c63092a8f10b27a4bd\": rpc error: code = NotFound desc = an error occurred when try to find container \"05d1a463839387260db0c040a26786524b1847fa73ef06c63092a8f10b27a4bd\": not found" Aug 13 00:27:47.152939 systemd[1]: Removed slice kubepods-besteffort-pod517c2f94_3455_4677_a3ec_3a8771258060.slice - libcontainer container kubepods-besteffort-pod517c2f94_3455_4677_a3ec_3a8771258060.slice. Aug 13 00:27:47.153019 systemd[1]: kubepods-besteffort-pod517c2f94_3455_4677_a3ec_3a8771258060.slice: Consumed 1.663s CPU time. Aug 13 00:27:47.388960 systemd[1]: var-lib-kubelet-pods-517c2f94\x2d3455\x2d4677\x2da3ec\x2d3a8771258060-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dpbmgz.mount: Deactivated successfully. Aug 13 00:27:47.389050 systemd[1]: var-lib-kubelet-pods-517c2f94\x2d3455\x2d4677\x2da3ec\x2d3a8771258060-volumes-kubernetes.io\x7esecret-calico\x2dapiserver\x2dcerts.mount: Deactivated successfully. Aug 13 00:27:47.990944 kubelet[4148]: I0813 00:27:47.990905 4148 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ced7fba-ab7f-4d92-941c-7095a5aea251" path="/var/lib/kubelet/pods/0ced7fba-ab7f-4d92-941c-7095a5aea251/volumes" Aug 13 00:27:47.991251 kubelet[4148]: I0813 00:27:47.991235 4148 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="517c2f94-3455-4677-a3ec-3a8771258060" path="/var/lib/kubelet/pods/517c2f94-3455-4677-a3ec-3a8771258060/volumes" Aug 13 00:28:33.121461 containerd[2685]: time="2025-08-13T00:28:33.121414387Z" level=info msg="StopPodSandbox for \"6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881\"" Aug 13 00:28:33.180107 containerd[2685]: 2025-08-13 00:28:33.150 [WARNING][10752] cni-plugin/k8s.go 598: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--h5l86-eth0" Aug 13 00:28:33.180107 containerd[2685]: 2025-08-13 00:28:33.150 [INFO][10752] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" Aug 13 00:28:33.180107 containerd[2685]: 2025-08-13 00:28:33.150 [INFO][10752] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" iface="eth0" netns="" Aug 13 00:28:33.180107 containerd[2685]: 2025-08-13 00:28:33.150 [INFO][10752] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" Aug 13 00:28:33.180107 containerd[2685]: 2025-08-13 00:28:33.150 [INFO][10752] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" Aug 13 00:28:33.180107 containerd[2685]: 2025-08-13 00:28:33.168 [INFO][10771] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" HandleID="k8s-pod-network.6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--h5l86-eth0" Aug 13 00:28:33.180107 containerd[2685]: 2025-08-13 00:28:33.168 [INFO][10771] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:28:33.180107 containerd[2685]: 2025-08-13 00:28:33.168 [INFO][10771] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:28:33.180107 containerd[2685]: 2025-08-13 00:28:33.176 [WARNING][10771] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" HandleID="k8s-pod-network.6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--h5l86-eth0" Aug 13 00:28:33.180107 containerd[2685]: 2025-08-13 00:28:33.176 [INFO][10771] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" HandleID="k8s-pod-network.6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--h5l86-eth0" Aug 13 00:28:33.180107 containerd[2685]: 2025-08-13 00:28:33.177 [INFO][10771] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:28:33.180107 containerd[2685]: 2025-08-13 00:28:33.178 [INFO][10752] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" Aug 13 00:28:33.180511 containerd[2685]: time="2025-08-13T00:28:33.180157864Z" level=info msg="TearDown network for sandbox \"6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881\" successfully" Aug 13 00:28:33.180511 containerd[2685]: time="2025-08-13T00:28:33.180195104Z" level=info msg="StopPodSandbox for \"6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881\" returns successfully" Aug 13 00:28:33.180619 containerd[2685]: time="2025-08-13T00:28:33.180593305Z" level=info msg="RemovePodSandbox for \"6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881\"" Aug 13 00:28:33.180647 containerd[2685]: time="2025-08-13T00:28:33.180627425Z" level=info msg="Forcibly stopping sandbox \"6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881\"" Aug 13 00:28:33.238922 containerd[2685]: 2025-08-13 00:28:33.210 [WARNING][10802] cni-plugin/k8s.go 598: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--h5l86-eth0" Aug 13 00:28:33.238922 containerd[2685]: 2025-08-13 00:28:33.210 [INFO][10802] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" Aug 13 00:28:33.238922 containerd[2685]: 2025-08-13 00:28:33.210 [INFO][10802] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" iface="eth0" netns="" Aug 13 00:28:33.238922 containerd[2685]: 2025-08-13 00:28:33.210 [INFO][10802] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" Aug 13 00:28:33.238922 containerd[2685]: 2025-08-13 00:28:33.210 [INFO][10802] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" Aug 13 00:28:33.238922 containerd[2685]: 2025-08-13 00:28:33.227 [INFO][10824] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" HandleID="k8s-pod-network.6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--h5l86-eth0" Aug 13 00:28:33.238922 containerd[2685]: 2025-08-13 00:28:33.227 [INFO][10824] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:28:33.238922 containerd[2685]: 2025-08-13 00:28:33.228 [INFO][10824] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:28:33.238922 containerd[2685]: 2025-08-13 00:28:33.235 [WARNING][10824] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" HandleID="k8s-pod-network.6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--h5l86-eth0" Aug 13 00:28:33.238922 containerd[2685]: 2025-08-13 00:28:33.235 [INFO][10824] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" HandleID="k8s-pod-network.6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--h5l86-eth0" Aug 13 00:28:33.238922 containerd[2685]: 2025-08-13 00:28:33.236 [INFO][10824] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:28:33.238922 containerd[2685]: 2025-08-13 00:28:33.237 [INFO][10802] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881" Aug 13 00:28:33.239163 containerd[2685]: time="2025-08-13T00:28:33.238961861Z" level=info msg="TearDown network for sandbox \"6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881\" successfully" Aug 13 00:28:33.240800 containerd[2685]: time="2025-08-13T00:28:33.240769025Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Aug 13 00:28:33.240860 containerd[2685]: time="2025-08-13T00:28:33.240847865Z" level=info msg="RemovePodSandbox \"6d050f94dd892d5e43d5575b113de051c4693943df91e424082a4d5cb7c0b881\" returns successfully" Aug 13 00:28:33.241190 containerd[2685]: time="2025-08-13T00:28:33.241166065Z" level=info msg="StopPodSandbox for \"812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6\"" Aug 13 00:28:33.302101 containerd[2685]: 2025-08-13 00:28:33.270 [WARNING][10854] cni-plugin/k8s.go 598: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--4dnl6-eth0" Aug 13 00:28:33.302101 containerd[2685]: 2025-08-13 00:28:33.270 [INFO][10854] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" Aug 13 00:28:33.302101 containerd[2685]: 2025-08-13 00:28:33.271 [INFO][10854] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" iface="eth0" netns="" Aug 13 00:28:33.302101 containerd[2685]: 2025-08-13 00:28:33.271 [INFO][10854] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" Aug 13 00:28:33.302101 containerd[2685]: 2025-08-13 00:28:33.271 [INFO][10854] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" Aug 13 00:28:33.302101 containerd[2685]: 2025-08-13 00:28:33.288 [INFO][10876] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" HandleID="k8s-pod-network.812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--4dnl6-eth0" Aug 13 00:28:33.302101 containerd[2685]: 2025-08-13 00:28:33.288 [INFO][10876] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:28:33.302101 containerd[2685]: 2025-08-13 00:28:33.288 [INFO][10876] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:28:33.302101 containerd[2685]: 2025-08-13 00:28:33.298 [WARNING][10876] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" HandleID="k8s-pod-network.812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--4dnl6-eth0" Aug 13 00:28:33.302101 containerd[2685]: 2025-08-13 00:28:33.298 [INFO][10876] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" HandleID="k8s-pod-network.812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--4dnl6-eth0" Aug 13 00:28:33.302101 containerd[2685]: 2025-08-13 00:28:33.299 [INFO][10876] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:28:33.302101 containerd[2685]: 2025-08-13 00:28:33.300 [INFO][10854] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" Aug 13 00:28:33.302459 containerd[2685]: time="2025-08-13T00:28:33.302125107Z" level=info msg="TearDown network for sandbox \"812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6\" successfully" Aug 13 00:28:33.302459 containerd[2685]: time="2025-08-13T00:28:33.302148427Z" level=info msg="StopPodSandbox for \"812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6\" returns successfully" Aug 13 00:28:33.302503 containerd[2685]: time="2025-08-13T00:28:33.302455307Z" level=info msg="RemovePodSandbox for \"812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6\"" Aug 13 00:28:33.302503 containerd[2685]: time="2025-08-13T00:28:33.302483627Z" level=info msg="Forcibly stopping sandbox \"812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6\"" Aug 13 00:28:33.361729 containerd[2685]: 2025-08-13 00:28:33.333 [WARNING][10904] cni-plugin/k8s.go 598: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" WorkloadEndpoint="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--4dnl6-eth0" Aug 13 00:28:33.361729 containerd[2685]: 2025-08-13 00:28:33.333 [INFO][10904] cni-plugin/k8s.go 640: Cleaning up netns ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" Aug 13 00:28:33.361729 containerd[2685]: 2025-08-13 00:28:33.333 [INFO][10904] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" iface="eth0" netns="" Aug 13 00:28:33.361729 containerd[2685]: 2025-08-13 00:28:33.333 [INFO][10904] cni-plugin/k8s.go 647: Releasing IP address(es) ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" Aug 13 00:28:33.361729 containerd[2685]: 2025-08-13 00:28:33.333 [INFO][10904] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" Aug 13 00:28:33.361729 containerd[2685]: 2025-08-13 00:28:33.350 [INFO][10925] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" HandleID="k8s-pod-network.812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--4dnl6-eth0" Aug 13 00:28:33.361729 containerd[2685]: 2025-08-13 00:28:33.350 [INFO][10925] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Aug 13 00:28:33.361729 containerd[2685]: 2025-08-13 00:28:33.350 [INFO][10925] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Aug 13 00:28:33.361729 containerd[2685]: 2025-08-13 00:28:33.358 [WARNING][10925] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" HandleID="k8s-pod-network.812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--4dnl6-eth0" Aug 13 00:28:33.361729 containerd[2685]: 2025-08-13 00:28:33.358 [INFO][10925] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" HandleID="k8s-pod-network.812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" Workload="ci--4081.3.5--a--588adb827d-k8s-calico--apiserver--74488cd4c9--4dnl6-eth0" Aug 13 00:28:33.361729 containerd[2685]: 2025-08-13 00:28:33.359 [INFO][10925] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Aug 13 00:28:33.361729 containerd[2685]: 2025-08-13 00:28:33.360 [INFO][10904] cni-plugin/k8s.go 653: Teardown processing complete. ContainerID="812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6" Aug 13 00:28:33.361973 containerd[2685]: time="2025-08-13T00:28:33.361762865Z" level=info msg="TearDown network for sandbox \"812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6\" successfully" Aug 13 00:28:33.363568 containerd[2685]: time="2025-08-13T00:28:33.363534149Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Aug 13 00:28:33.363622 containerd[2685]: time="2025-08-13T00:28:33.363596509Z" level=info msg="RemovePodSandbox \"812286c1f8a36f70ce315543286cb405db470fb30c47fe5382713e849d523ad6\" returns successfully" Aug 13 00:28:35.633188 update_engine[2678]: I20250813 00:28:35.632963 2678 prefs.cc:52] certificate-report-to-send-update not present in /var/lib/update_engine/prefs Aug 13 00:28:35.633188 update_engine[2678]: I20250813 00:28:35.633024 2678 prefs.cc:52] certificate-report-to-send-download not present in /var/lib/update_engine/prefs Aug 13 00:28:35.633523 update_engine[2678]: I20250813 00:28:35.633252 2678 prefs.cc:52] aleph-version not present in /var/lib/update_engine/prefs Aug 13 00:28:35.633595 update_engine[2678]: I20250813 00:28:35.633584 2678 omaha_request_params.cc:62] Current group set to lts Aug 13 00:28:35.633677 update_engine[2678]: I20250813 00:28:35.633662 2678 update_attempter.cc:499] Already updated boot flags. Skipping. Aug 13 00:28:35.633677 update_engine[2678]: I20250813 00:28:35.633672 2678 update_attempter.cc:643] Scheduling an action processor start. Aug 13 00:28:35.633729 update_engine[2678]: I20250813 00:28:35.633687 2678 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction Aug 13 00:28:35.633729 update_engine[2678]: I20250813 00:28:35.633711 2678 prefs.cc:52] previous-version not present in /var/lib/update_engine/prefs Aug 13 00:28:35.633769 update_engine[2678]: I20250813 00:28:35.633759 2678 omaha_request_action.cc:271] Posting an Omaha request to disabled Aug 13 00:28:35.633790 update_engine[2678]: I20250813 00:28:35.633766 2678 omaha_request_action.cc:272] Request: Aug 13 00:28:35.633790 update_engine[2678]: Aug 13 00:28:35.633790 update_engine[2678]: Aug 13 00:28:35.633790 update_engine[2678]: Aug 13 00:28:35.633790 update_engine[2678]: Aug 13 00:28:35.633790 update_engine[2678]: Aug 13 00:28:35.633790 update_engine[2678]: Aug 13 00:28:35.633790 update_engine[2678]: Aug 13 00:28:35.633790 update_engine[2678]: Aug 13 00:28:35.633790 update_engine[2678]: I20250813 00:28:35.633772 2678 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Aug 13 00:28:35.633968 locksmithd[2709]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_CHECKING_FOR_UPDATE" NewVersion=0.0.0 NewSize=0 Aug 13 00:28:35.634727 update_engine[2678]: I20250813 00:28:35.634709 2678 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Aug 13 00:28:35.634969 update_engine[2678]: I20250813 00:28:35.634949 2678 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Aug 13 00:28:35.635439 update_engine[2678]: E20250813 00:28:35.635414 2678 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Aug 13 00:28:35.635563 update_engine[2678]: I20250813 00:28:35.635545 2678 libcurl_http_fetcher.cc:283] No HTTP response, retry 1 Aug 13 00:28:45.593781 update_engine[2678]: I20250813 00:28:45.593608 2678 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Aug 13 00:28:45.594270 update_engine[2678]: I20250813 00:28:45.593873 2678 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Aug 13 00:28:45.594270 update_engine[2678]: I20250813 00:28:45.594053 2678 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Aug 13 00:28:45.594717 update_engine[2678]: E20250813 00:28:45.594689 2678 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Aug 13 00:28:45.594853 update_engine[2678]: I20250813 00:28:45.594834 2678 libcurl_http_fetcher.cc:283] No HTTP response, retry 2 Aug 13 00:28:55.594248 update_engine[2678]: I20250813 00:28:55.593928 2678 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Aug 13 00:28:55.594248 update_engine[2678]: I20250813 00:28:55.594213 2678 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Aug 13 00:28:55.594617 update_engine[2678]: I20250813 00:28:55.594411 2678 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Aug 13 00:28:55.595072 update_engine[2678]: E20250813 00:28:55.595053 2678 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Aug 13 00:28:55.595105 update_engine[2678]: I20250813 00:28:55.595093 2678 libcurl_http_fetcher.cc:283] No HTTP response, retry 3 Aug 13 00:29:05.594199 update_engine[2678]: I20250813 00:29:05.594136 2678 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Aug 13 00:29:05.594602 update_engine[2678]: I20250813 00:29:05.594409 2678 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Aug 13 00:29:05.594602 update_engine[2678]: I20250813 00:29:05.594584 2678 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Aug 13 00:29:05.595208 update_engine[2678]: E20250813 00:29:05.595191 2678 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Aug 13 00:29:05.595246 update_engine[2678]: I20250813 00:29:05.595225 2678 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded Aug 13 00:29:05.595246 update_engine[2678]: I20250813 00:29:05.595239 2678 omaha_request_action.cc:617] Omaha request response: Aug 13 00:29:05.595318 update_engine[2678]: E20250813 00:29:05.595306 2678 omaha_request_action.cc:636] Omaha request network transfer failed. Aug 13 00:29:05.595341 update_engine[2678]: I20250813 00:29:05.595322 2678 action_processor.cc:68] ActionProcessor::ActionComplete: OmahaRequestAction action failed. Aborting processing. Aug 13 00:29:05.595341 update_engine[2678]: I20250813 00:29:05.595328 2678 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Aug 13 00:29:05.595341 update_engine[2678]: I20250813 00:29:05.595332 2678 update_attempter.cc:306] Processing Done. Aug 13 00:29:05.595403 update_engine[2678]: E20250813 00:29:05.595345 2678 update_attempter.cc:619] Update failed. Aug 13 00:29:05.595403 update_engine[2678]: I20250813 00:29:05.595350 2678 utils.cc:600] Converting error code 2000 to kActionCodeOmahaErrorInHTTPResponse Aug 13 00:29:05.595403 update_engine[2678]: I20250813 00:29:05.595355 2678 payload_state.cc:97] Updating payload state for error code: 37 (kActionCodeOmahaErrorInHTTPResponse) Aug 13 00:29:05.595403 update_engine[2678]: I20250813 00:29:05.595359 2678 payload_state.cc:103] Ignoring failures until we get a valid Omaha response. Aug 13 00:29:05.595482 update_engine[2678]: I20250813 00:29:05.595415 2678 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction Aug 13 00:29:05.595482 update_engine[2678]: I20250813 00:29:05.595434 2678 omaha_request_action.cc:271] Posting an Omaha request to disabled Aug 13 00:29:05.595482 update_engine[2678]: I20250813 00:29:05.595439 2678 omaha_request_action.cc:272] Request: Aug 13 00:29:05.595482 update_engine[2678]: Aug 13 00:29:05.595482 update_engine[2678]: Aug 13 00:29:05.595482 update_engine[2678]: Aug 13 00:29:05.595482 update_engine[2678]: Aug 13 00:29:05.595482 update_engine[2678]: Aug 13 00:29:05.595482 update_engine[2678]: Aug 13 00:29:05.595482 update_engine[2678]: I20250813 00:29:05.595444 2678 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Aug 13 00:29:05.595658 update_engine[2678]: I20250813 00:29:05.595551 2678 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Aug 13 00:29:05.595680 locksmithd[2709]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_REPORTING_ERROR_EVENT" NewVersion=0.0.0 NewSize=0 Aug 13 00:29:05.595863 update_engine[2678]: I20250813 00:29:05.595682 2678 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Aug 13 00:29:05.596281 update_engine[2678]: E20250813 00:29:05.596264 2678 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Aug 13 00:29:05.596309 update_engine[2678]: I20250813 00:29:05.596298 2678 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded Aug 13 00:29:05.596335 update_engine[2678]: I20250813 00:29:05.596307 2678 omaha_request_action.cc:617] Omaha request response: Aug 13 00:29:05.596335 update_engine[2678]: I20250813 00:29:05.596313 2678 action_processor.cc:65] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Aug 13 00:29:05.596335 update_engine[2678]: I20250813 00:29:05.596318 2678 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Aug 13 00:29:05.596335 update_engine[2678]: I20250813 00:29:05.596323 2678 update_attempter.cc:306] Processing Done. Aug 13 00:29:05.596335 update_engine[2678]: I20250813 00:29:05.596326 2678 update_attempter.cc:310] Error event sent. Aug 13 00:29:05.596335 update_engine[2678]: I20250813 00:29:05.596333 2678 update_check_scheduler.cc:74] Next update check in 48m58s Aug 13 00:29:05.596474 locksmithd[2709]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_IDLE" NewVersion=0.0.0 NewSize=0 Aug 13 00:35:39.600833 systemd[1]: Started sshd@9-147.75.53.22:22-147.75.109.163:53742.service - OpenSSH per-connection server daemon (147.75.109.163:53742). Aug 13 00:35:39.901594 sshd[12886]: Accepted publickey for core from 147.75.109.163 port 53742 ssh2: RSA SHA256:u7vn2r+eDvbQsVz87adSciVbwQCBuSV2wy8SWfSsl1k Aug 13 00:35:39.902673 sshd[12886]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:35:39.905995 systemd-logind[2668]: New session 12 of user core. Aug 13 00:35:39.919340 systemd[1]: Started session-12.scope - Session 12 of User core. Aug 13 00:35:40.180143 sshd[12886]: pam_unix(sshd:session): session closed for user core Aug 13 00:35:40.183148 systemd[1]: sshd@9-147.75.53.22:22-147.75.109.163:53742.service: Deactivated successfully. Aug 13 00:35:40.185461 systemd[1]: session-12.scope: Deactivated successfully. Aug 13 00:35:40.186005 systemd-logind[2668]: Session 12 logged out. Waiting for processes to exit. Aug 13 00:35:40.186608 systemd-logind[2668]: Removed session 12. Aug 13 00:35:45.236807 systemd[1]: Started sshd@10-147.75.53.22:22-147.75.109.163:53752.service - OpenSSH per-connection server daemon (147.75.109.163:53752). Aug 13 00:35:45.537553 sshd[12964]: Accepted publickey for core from 147.75.109.163 port 53752 ssh2: RSA SHA256:u7vn2r+eDvbQsVz87adSciVbwQCBuSV2wy8SWfSsl1k Aug 13 00:35:45.538678 sshd[12964]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:35:45.541586 systemd-logind[2668]: New session 13 of user core. Aug 13 00:35:45.555324 systemd[1]: Started session-13.scope - Session 13 of User core. Aug 13 00:35:45.807108 sshd[12964]: pam_unix(sshd:session): session closed for user core Aug 13 00:35:45.810021 systemd[1]: sshd@10-147.75.53.22:22-147.75.109.163:53752.service: Deactivated successfully. Aug 13 00:35:45.811731 systemd[1]: session-13.scope: Deactivated successfully. Aug 13 00:35:45.812281 systemd-logind[2668]: Session 13 logged out. Waiting for processes to exit. Aug 13 00:35:45.812867 systemd-logind[2668]: Removed session 13. Aug 13 00:35:45.856686 systemd[1]: Started sshd@11-147.75.53.22:22-147.75.109.163:53768.service - OpenSSH per-connection server daemon (147.75.109.163:53768). Aug 13 00:35:46.129293 sshd[13002]: Accepted publickey for core from 147.75.109.163 port 53768 ssh2: RSA SHA256:u7vn2r+eDvbQsVz87adSciVbwQCBuSV2wy8SWfSsl1k Aug 13 00:35:46.130433 sshd[13002]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:35:46.133553 systemd-logind[2668]: New session 14 of user core. Aug 13 00:35:46.145399 systemd[1]: Started session-14.scope - Session 14 of User core. Aug 13 00:35:46.410948 sshd[13002]: pam_unix(sshd:session): session closed for user core Aug 13 00:35:46.413902 systemd[1]: sshd@11-147.75.53.22:22-147.75.109.163:53768.service: Deactivated successfully. Aug 13 00:35:46.415623 systemd[1]: session-14.scope: Deactivated successfully. Aug 13 00:35:46.416142 systemd-logind[2668]: Session 14 logged out. Waiting for processes to exit. Aug 13 00:35:46.416740 systemd-logind[2668]: Removed session 14. Aug 13 00:35:46.465708 systemd[1]: Started sshd@12-147.75.53.22:22-147.75.109.163:53782.service - OpenSSH per-connection server daemon (147.75.109.163:53782). Aug 13 00:35:46.764025 sshd[13037]: Accepted publickey for core from 147.75.109.163 port 53782 ssh2: RSA SHA256:u7vn2r+eDvbQsVz87adSciVbwQCBuSV2wy8SWfSsl1k Aug 13 00:35:46.765094 sshd[13037]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:35:46.768032 systemd-logind[2668]: New session 15 of user core. Aug 13 00:35:46.785347 systemd[1]: Started session-15.scope - Session 15 of User core. Aug 13 00:35:47.033729 sshd[13037]: pam_unix(sshd:session): session closed for user core Aug 13 00:35:47.036586 systemd[1]: sshd@12-147.75.53.22:22-147.75.109.163:53782.service: Deactivated successfully. Aug 13 00:35:47.038846 systemd[1]: session-15.scope: Deactivated successfully. Aug 13 00:35:47.039384 systemd-logind[2668]: Session 15 logged out. Waiting for processes to exit. Aug 13 00:35:47.039922 systemd-logind[2668]: Removed session 15. Aug 13 00:35:52.086793 systemd[1]: Started sshd@13-147.75.53.22:22-147.75.109.163:44052.service - OpenSSH per-connection server daemon (147.75.109.163:44052). Aug 13 00:35:52.387265 sshd[13171]: Accepted publickey for core from 147.75.109.163 port 44052 ssh2: RSA SHA256:u7vn2r+eDvbQsVz87adSciVbwQCBuSV2wy8SWfSsl1k Aug 13 00:35:52.388319 sshd[13171]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:35:52.391197 systemd-logind[2668]: New session 16 of user core. Aug 13 00:35:52.400336 systemd[1]: Started session-16.scope - Session 16 of User core. Aug 13 00:35:52.663409 sshd[13171]: pam_unix(sshd:session): session closed for user core Aug 13 00:35:52.666215 systemd[1]: sshd@13-147.75.53.22:22-147.75.109.163:44052.service: Deactivated successfully. Aug 13 00:35:52.667898 systemd[1]: session-16.scope: Deactivated successfully. Aug 13 00:35:52.668443 systemd-logind[2668]: Session 16 logged out. Waiting for processes to exit. Aug 13 00:35:52.669005 systemd-logind[2668]: Removed session 16. Aug 13 00:35:57.716801 systemd[1]: Started sshd@14-147.75.53.22:22-147.75.109.163:44064.service - OpenSSH per-connection server daemon (147.75.109.163:44064). Aug 13 00:35:58.017554 sshd[13210]: Accepted publickey for core from 147.75.109.163 port 44064 ssh2: RSA SHA256:u7vn2r+eDvbQsVz87adSciVbwQCBuSV2wy8SWfSsl1k Aug 13 00:35:58.018597 sshd[13210]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:35:58.021534 systemd-logind[2668]: New session 17 of user core. Aug 13 00:35:58.031340 systemd[1]: Started session-17.scope - Session 17 of User core. Aug 13 00:35:58.289893 sshd[13210]: pam_unix(sshd:session): session closed for user core Aug 13 00:35:58.292047 systemd[1]: sshd@14-147.75.53.22:22-147.75.109.163:44064.service: Deactivated successfully. Aug 13 00:35:58.293663 systemd[1]: session-17.scope: Deactivated successfully. Aug 13 00:35:58.294893 systemd-logind[2668]: Session 17 logged out. Waiting for processes to exit. Aug 13 00:35:58.295558 systemd-logind[2668]: Removed session 17. Aug 13 00:36:03.345822 systemd[1]: Started sshd@15-147.75.53.22:22-147.75.109.163:51724.service - OpenSSH per-connection server daemon (147.75.109.163:51724). Aug 13 00:36:03.647467 sshd[13239]: Accepted publickey for core from 147.75.109.163 port 51724 ssh2: RSA SHA256:u7vn2r+eDvbQsVz87adSciVbwQCBuSV2wy8SWfSsl1k Aug 13 00:36:03.648592 sshd[13239]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:36:03.651827 systemd-logind[2668]: New session 18 of user core. Aug 13 00:36:03.665337 systemd[1]: Started session-18.scope - Session 18 of User core. Aug 13 00:36:03.919915 sshd[13239]: pam_unix(sshd:session): session closed for user core Aug 13 00:36:03.922790 systemd[1]: sshd@15-147.75.53.22:22-147.75.109.163:51724.service: Deactivated successfully. Aug 13 00:36:03.924572 systemd[1]: session-18.scope: Deactivated successfully. Aug 13 00:36:03.925731 systemd-logind[2668]: Session 18 logged out. Waiting for processes to exit. Aug 13 00:36:03.926301 systemd-logind[2668]: Removed session 18. Aug 13 00:36:03.975793 systemd[1]: Started sshd@16-147.75.53.22:22-147.75.109.163:51730.service - OpenSSH per-connection server daemon (147.75.109.163:51730). Aug 13 00:36:04.279977 sshd[13272]: Accepted publickey for core from 147.75.109.163 port 51730 ssh2: RSA SHA256:u7vn2r+eDvbQsVz87adSciVbwQCBuSV2wy8SWfSsl1k Aug 13 00:36:04.281072 sshd[13272]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:36:04.283958 systemd-logind[2668]: New session 19 of user core. Aug 13 00:36:04.300402 systemd[1]: Started session-19.scope - Session 19 of User core. Aug 13 00:36:04.595422 sshd[13272]: pam_unix(sshd:session): session closed for user core Aug 13 00:36:04.598334 systemd[1]: sshd@16-147.75.53.22:22-147.75.109.163:51730.service: Deactivated successfully. Aug 13 00:36:04.600609 systemd[1]: session-19.scope: Deactivated successfully. Aug 13 00:36:04.601164 systemd-logind[2668]: Session 19 logged out. Waiting for processes to exit. Aug 13 00:36:04.601801 systemd-logind[2668]: Removed session 19. Aug 13 00:36:04.646827 systemd[1]: Started sshd@17-147.75.53.22:22-147.75.109.163:51740.service - OpenSSH per-connection server daemon (147.75.109.163:51740). Aug 13 00:36:04.921478 sshd[13303]: Accepted publickey for core from 147.75.109.163 port 51740 ssh2: RSA SHA256:u7vn2r+eDvbQsVz87adSciVbwQCBuSV2wy8SWfSsl1k Aug 13 00:36:04.922563 sshd[13303]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:36:04.925563 systemd-logind[2668]: New session 20 of user core. Aug 13 00:36:04.936397 systemd[1]: Started session-20.scope - Session 20 of User core. Aug 13 00:36:05.653527 sshd[13303]: pam_unix(sshd:session): session closed for user core Aug 13 00:36:05.656435 systemd[1]: sshd@17-147.75.53.22:22-147.75.109.163:51740.service: Deactivated successfully. Aug 13 00:36:05.658994 systemd[1]: session-20.scope: Deactivated successfully. Aug 13 00:36:05.659604 systemd-logind[2668]: Session 20 logged out. Waiting for processes to exit. Aug 13 00:36:05.660212 systemd-logind[2668]: Removed session 20. Aug 13 00:36:05.709847 systemd[1]: Started sshd@18-147.75.53.22:22-147.75.109.163:51742.service - OpenSSH per-connection server daemon (147.75.109.163:51742). Aug 13 00:36:06.007415 sshd[13361]: Accepted publickey for core from 147.75.109.163 port 51742 ssh2: RSA SHA256:u7vn2r+eDvbQsVz87adSciVbwQCBuSV2wy8SWfSsl1k Aug 13 00:36:06.008676 sshd[13361]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:36:06.011579 systemd-logind[2668]: New session 21 of user core. Aug 13 00:36:06.022365 systemd[1]: Started session-21.scope - Session 21 of User core. Aug 13 00:36:06.377633 sshd[13361]: pam_unix(sshd:session): session closed for user core Aug 13 00:36:06.380604 systemd[1]: sshd@18-147.75.53.22:22-147.75.109.163:51742.service: Deactivated successfully. Aug 13 00:36:06.382923 systemd[1]: session-21.scope: Deactivated successfully. Aug 13 00:36:06.383458 systemd-logind[2668]: Session 21 logged out. Waiting for processes to exit. Aug 13 00:36:06.384041 systemd-logind[2668]: Removed session 21. Aug 13 00:36:06.434755 systemd[1]: Started sshd@19-147.75.53.22:22-147.75.109.163:51758.service - OpenSSH per-connection server daemon (147.75.109.163:51758). Aug 13 00:36:06.731999 sshd[13411]: Accepted publickey for core from 147.75.109.163 port 51758 ssh2: RSA SHA256:u7vn2r+eDvbQsVz87adSciVbwQCBuSV2wy8SWfSsl1k Aug 13 00:36:06.733058 sshd[13411]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:36:06.736115 systemd-logind[2668]: New session 22 of user core. Aug 13 00:36:06.745336 systemd[1]: Started session-22.scope - Session 22 of User core. Aug 13 00:36:06.999709 sshd[13411]: pam_unix(sshd:session): session closed for user core Aug 13 00:36:07.002512 systemd[1]: sshd@19-147.75.53.22:22-147.75.109.163:51758.service: Deactivated successfully. Aug 13 00:36:07.004846 systemd[1]: session-22.scope: Deactivated successfully. Aug 13 00:36:07.005379 systemd-logind[2668]: Session 22 logged out. Waiting for processes to exit. Aug 13 00:36:07.005908 systemd-logind[2668]: Removed session 22. Aug 13 00:36:12.054764 systemd[1]: Started sshd@20-147.75.53.22:22-147.75.109.163:55098.service - OpenSSH per-connection server daemon (147.75.109.163:55098). Aug 13 00:36:12.355140 sshd[13452]: Accepted publickey for core from 147.75.109.163 port 55098 ssh2: RSA SHA256:u7vn2r+eDvbQsVz87adSciVbwQCBuSV2wy8SWfSsl1k Aug 13 00:36:12.356296 sshd[13452]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:36:12.359387 systemd-logind[2668]: New session 23 of user core. Aug 13 00:36:12.373397 systemd[1]: Started session-23.scope - Session 23 of User core. Aug 13 00:36:12.623647 sshd[13452]: pam_unix(sshd:session): session closed for user core Aug 13 00:36:12.626736 systemd[1]: sshd@20-147.75.53.22:22-147.75.109.163:55098.service: Deactivated successfully. Aug 13 00:36:12.628437 systemd[1]: session-23.scope: Deactivated successfully. Aug 13 00:36:12.628966 systemd-logind[2668]: Session 23 logged out. Waiting for processes to exit. Aug 13 00:36:12.629547 systemd-logind[2668]: Removed session 23. Aug 13 00:36:17.676722 systemd[1]: Started sshd@21-147.75.53.22:22-147.75.109.163:55102.service - OpenSSH per-connection server daemon (147.75.109.163:55102). Aug 13 00:36:17.977143 sshd[13527]: Accepted publickey for core from 147.75.109.163 port 55102 ssh2: RSA SHA256:u7vn2r+eDvbQsVz87adSciVbwQCBuSV2wy8SWfSsl1k Aug 13 00:36:17.978251 sshd[13527]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 00:36:17.981486 systemd-logind[2668]: New session 24 of user core. Aug 13 00:36:17.994392 systemd[1]: Started session-24.scope - Session 24 of User core. Aug 13 00:36:18.247197 sshd[13527]: pam_unix(sshd:session): session closed for user core Aug 13 00:36:18.250269 systemd[1]: sshd@21-147.75.53.22:22-147.75.109.163:55102.service: Deactivated successfully. Aug 13 00:36:18.252582 systemd[1]: session-24.scope: Deactivated successfully. Aug 13 00:36:18.253134 systemd-logind[2668]: Session 24 logged out. Waiting for processes to exit. Aug 13 00:36:18.253770 systemd-logind[2668]: Removed session 24.